mirror of
https://github.com/edk2-porting/linux-next.git
synced 2024-12-23 20:53:53 +08:00
769848c038
It is often known at allocation time whether a page may be migrated or not. This patch adds a flag called __GFP_MOVABLE and a new mask called GFP_HIGH_MOVABLE. Allocations using the __GFP_MOVABLE can be either migrated using the page migration mechanism or reclaimed by syncing with backing storage and discarding. An API function very similar to alloc_zeroed_user_highpage() is added for __GFP_MOVABLE allocations called alloc_zeroed_user_highpage_movable(). The flags used by alloc_zeroed_user_highpage() are not changed because it would change the semantics of an existing API. After this patch is applied there are no in-kernel users of alloc_zeroed_user_highpage() so it probably should be marked deprecated if this patch is merged. Note that this patch includes a minor cleanup to the use of __GFP_ZERO in shmem.c to keep all flag modifications to inode->mapping in the shmem_dir_alloc() helper function. This clean-up suggestion is courtesy of Hugh Dickens. Additional credit goes to Christoph Lameter and Linus Torvalds for shaping the concept. Credit to Hugh Dickens for catching issues with shmem swap vector and ramfs allocations. [akpm@linux-foundation.org: build fix] [hugh@veritas.com: __GFP_ZERO cleanup] Signed-off-by: Mel Gorman <mel@csn.ul.ie> Cc: Andy Whitcroft <apw@shadowen.org> Cc: Christoph Lameter <clameter@sgi.com> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
232 lines
6.3 KiB
C
232 lines
6.3 KiB
C
#ifndef _ASM_IA64_PAGE_H
|
|
#define _ASM_IA64_PAGE_H
|
|
/*
|
|
* Pagetable related stuff.
|
|
*
|
|
* Copyright (C) 1998, 1999, 2002 Hewlett-Packard Co
|
|
* David Mosberger-Tang <davidm@hpl.hp.com>
|
|
*/
|
|
|
|
# ifdef __KERNEL__
|
|
|
|
#include <asm/intrinsics.h>
|
|
#include <asm/types.h>
|
|
|
|
/*
|
|
* The top three bits of an IA64 address are its Region Number.
|
|
* Different regions are assigned to different purposes.
|
|
*/
|
|
#define RGN_SHIFT (61)
|
|
#define RGN_BASE(r) (__IA64_UL_CONST(r)<<RGN_SHIFT)
|
|
#define RGN_BITS (RGN_BASE(-1))
|
|
|
|
#define RGN_KERNEL 7 /* Identity mapped region */
|
|
#define RGN_UNCACHED 6 /* Identity mapped I/O region */
|
|
#define RGN_GATE 5 /* Gate page, Kernel text, etc */
|
|
#define RGN_HPAGE 4 /* For Huge TLB pages */
|
|
|
|
/*
|
|
* PAGE_SHIFT determines the actual kernel page size.
|
|
*/
|
|
#if defined(CONFIG_IA64_PAGE_SIZE_4KB)
|
|
# define PAGE_SHIFT 12
|
|
#elif defined(CONFIG_IA64_PAGE_SIZE_8KB)
|
|
# define PAGE_SHIFT 13
|
|
#elif defined(CONFIG_IA64_PAGE_SIZE_16KB)
|
|
# define PAGE_SHIFT 14
|
|
#elif defined(CONFIG_IA64_PAGE_SIZE_64KB)
|
|
# define PAGE_SHIFT 16
|
|
#else
|
|
# error Unsupported page size!
|
|
#endif
|
|
|
|
#define PAGE_SIZE (__IA64_UL_CONST(1) << PAGE_SHIFT)
|
|
#define PAGE_MASK (~(PAGE_SIZE - 1))
|
|
#define PAGE_ALIGN(addr) (((addr) + PAGE_SIZE - 1) & PAGE_MASK)
|
|
|
|
#define PERCPU_PAGE_SHIFT 16 /* log2() of max. size of per-CPU area */
|
|
#define PERCPU_PAGE_SIZE (__IA64_UL_CONST(1) << PERCPU_PAGE_SHIFT)
|
|
|
|
|
|
#ifdef CONFIG_HUGETLB_PAGE
|
|
# define HPAGE_REGION_BASE RGN_BASE(RGN_HPAGE)
|
|
# define HPAGE_SHIFT hpage_shift
|
|
# define HPAGE_SHIFT_DEFAULT 28 /* check ia64 SDM for architecture supported size */
|
|
# define HPAGE_SIZE (__IA64_UL_CONST(1) << HPAGE_SHIFT)
|
|
# define HPAGE_MASK (~(HPAGE_SIZE - 1))
|
|
|
|
# define HAVE_ARCH_HUGETLB_UNMAPPED_AREA
|
|
# define ARCH_HAS_HUGEPAGE_ONLY_RANGE
|
|
# define ARCH_HAS_PREPARE_HUGEPAGE_RANGE
|
|
# define ARCH_HAS_HUGETLB_FREE_PGD_RANGE
|
|
#endif /* CONFIG_HUGETLB_PAGE */
|
|
|
|
#ifdef __ASSEMBLY__
|
|
# define __pa(x) ((x) - PAGE_OFFSET)
|
|
# define __va(x) ((x) + PAGE_OFFSET)
|
|
#else /* !__ASSEMBLY */
|
|
# define STRICT_MM_TYPECHECKS
|
|
|
|
extern void clear_page (void *page);
|
|
extern void copy_page (void *to, void *from);
|
|
|
|
/*
|
|
* clear_user_page() and copy_user_page() can't be inline functions because
|
|
* flush_dcache_page() can't be defined until later...
|
|
*/
|
|
#define clear_user_page(addr, vaddr, page) \
|
|
do { \
|
|
clear_page(addr); \
|
|
flush_dcache_page(page); \
|
|
} while (0)
|
|
|
|
#define copy_user_page(to, from, vaddr, page) \
|
|
do { \
|
|
copy_page((to), (from)); \
|
|
flush_dcache_page(page); \
|
|
} while (0)
|
|
|
|
|
|
#define __alloc_zeroed_user_highpage(movableflags, vma, vaddr) \
|
|
({ \
|
|
struct page *page = alloc_page_vma( \
|
|
GFP_HIGHUSER | __GFP_ZERO | movableflags, vma, vaddr); \
|
|
if (page) \
|
|
flush_dcache_page(page); \
|
|
page; \
|
|
})
|
|
|
|
#define __HAVE_ARCH_ALLOC_ZEROED_USER_HIGHPAGE
|
|
|
|
#define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT)
|
|
|
|
#ifdef CONFIG_VIRTUAL_MEM_MAP
|
|
extern int ia64_pfn_valid (unsigned long pfn);
|
|
#else
|
|
# define ia64_pfn_valid(pfn) 1
|
|
#endif
|
|
|
|
#ifdef CONFIG_VIRTUAL_MEM_MAP
|
|
extern struct page *vmem_map;
|
|
#ifdef CONFIG_DISCONTIGMEM
|
|
# define page_to_pfn(page) ((unsigned long) (page - vmem_map))
|
|
# define pfn_to_page(pfn) (vmem_map + (pfn))
|
|
#else
|
|
# include <asm-generic/memory_model.h>
|
|
#endif
|
|
#else
|
|
# include <asm-generic/memory_model.h>
|
|
#endif
|
|
|
|
#ifdef CONFIG_FLATMEM
|
|
# define pfn_valid(pfn) (((pfn) < max_mapnr) && ia64_pfn_valid(pfn))
|
|
#elif defined(CONFIG_DISCONTIGMEM)
|
|
extern unsigned long min_low_pfn;
|
|
extern unsigned long max_low_pfn;
|
|
# define pfn_valid(pfn) (((pfn) >= min_low_pfn) && ((pfn) < max_low_pfn) && ia64_pfn_valid(pfn))
|
|
#endif
|
|
|
|
#define page_to_phys(page) (page_to_pfn(page) << PAGE_SHIFT)
|
|
#define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT)
|
|
#define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT)
|
|
|
|
typedef union ia64_va {
|
|
struct {
|
|
unsigned long off : 61; /* intra-region offset */
|
|
unsigned long reg : 3; /* region number */
|
|
} f;
|
|
unsigned long l;
|
|
void *p;
|
|
} ia64_va;
|
|
|
|
/*
|
|
* Note: These macros depend on the fact that PAGE_OFFSET has all
|
|
* region bits set to 1 and all other bits set to zero. They are
|
|
* expressed in this way to ensure they result in a single "dep"
|
|
* instruction.
|
|
*/
|
|
#define __pa(x) ({ia64_va _v; _v.l = (long) (x); _v.f.reg = 0; _v.l;})
|
|
#define __va(x) ({ia64_va _v; _v.l = (long) (x); _v.f.reg = -1; _v.p;})
|
|
|
|
#define REGION_NUMBER(x) ({ia64_va _v; _v.l = (long) (x); _v.f.reg;})
|
|
#define REGION_OFFSET(x) ({ia64_va _v; _v.l = (long) (x); _v.f.off;})
|
|
|
|
#ifdef CONFIG_HUGETLB_PAGE
|
|
# define htlbpage_to_page(x) (((unsigned long) REGION_NUMBER(x) << 61) \
|
|
| (REGION_OFFSET(x) >> (HPAGE_SHIFT-PAGE_SHIFT)))
|
|
# define HUGETLB_PAGE_ORDER (HPAGE_SHIFT - PAGE_SHIFT)
|
|
# define is_hugepage_only_range(mm, addr, len) \
|
|
(REGION_NUMBER(addr) == RGN_HPAGE || \
|
|
REGION_NUMBER((addr)+(len)-1) == RGN_HPAGE)
|
|
extern unsigned int hpage_shift;
|
|
#endif
|
|
|
|
static __inline__ int
|
|
get_order (unsigned long size)
|
|
{
|
|
long double d = size - 1;
|
|
long order;
|
|
|
|
order = ia64_getf_exp(d);
|
|
order = order - PAGE_SHIFT - 0xffff + 1;
|
|
if (order < 0)
|
|
order = 0;
|
|
return order;
|
|
}
|
|
|
|
#endif /* !__ASSEMBLY__ */
|
|
|
|
#ifdef STRICT_MM_TYPECHECKS
|
|
/*
|
|
* These are used to make use of C type-checking..
|
|
*/
|
|
typedef struct { unsigned long pte; } pte_t;
|
|
typedef struct { unsigned long pmd; } pmd_t;
|
|
#ifdef CONFIG_PGTABLE_4
|
|
typedef struct { unsigned long pud; } pud_t;
|
|
#endif
|
|
typedef struct { unsigned long pgd; } pgd_t;
|
|
typedef struct { unsigned long pgprot; } pgprot_t;
|
|
|
|
# define pte_val(x) ((x).pte)
|
|
# define pmd_val(x) ((x).pmd)
|
|
#ifdef CONFIG_PGTABLE_4
|
|
# define pud_val(x) ((x).pud)
|
|
#endif
|
|
# define pgd_val(x) ((x).pgd)
|
|
# define pgprot_val(x) ((x).pgprot)
|
|
|
|
# define __pte(x) ((pte_t) { (x) } )
|
|
# define __pgprot(x) ((pgprot_t) { (x) } )
|
|
|
|
#else /* !STRICT_MM_TYPECHECKS */
|
|
/*
|
|
* .. while these make it easier on the compiler
|
|
*/
|
|
# ifndef __ASSEMBLY__
|
|
typedef unsigned long pte_t;
|
|
typedef unsigned long pmd_t;
|
|
typedef unsigned long pgd_t;
|
|
typedef unsigned long pgprot_t;
|
|
# endif
|
|
|
|
# define pte_val(x) (x)
|
|
# define pmd_val(x) (x)
|
|
# define pgd_val(x) (x)
|
|
# define pgprot_val(x) (x)
|
|
|
|
# define __pte(x) (x)
|
|
# define __pgd(x) (x)
|
|
# define __pgprot(x) (x)
|
|
#endif /* !STRICT_MM_TYPECHECKS */
|
|
|
|
#define PAGE_OFFSET RGN_BASE(RGN_KERNEL)
|
|
|
|
#define VM_DATA_DEFAULT_FLAGS (VM_READ | VM_WRITE | \
|
|
VM_MAYREAD | VM_MAYWRITE | VM_MAYEXEC | \
|
|
(((current->personality & READ_IMPLIES_EXEC) != 0) \
|
|
? VM_EXEC : 0))
|
|
|
|
# endif /* __KERNEL__ */
|
|
#endif /* _ASM_IA64_PAGE_H */
|