mirror of
https://github.com/edk2-porting/linux-next.git
synced 2024-12-23 04:34:11 +08:00
0889eba5b3
x86_64 uses 2M page table entries to map its 1-1 kernel space. We also implement the virtual memmap using 2M page table entries. So there is no additional runtime overhead over FLATMEM, initialisation is slightly more complex. As FLATMEM still references memory to obtain the mem_map pointer and SPARSEMEM_VMEMMAP uses a compile time constant, SPARSEMEM_VMEMMAP should be superior. With this SPARSEMEM becomes the most efficient way of handling virt_to_page, pfn_to_page and friends for UP, SMP and NUMA on x86_64. [apw@shadowen.org: code resplit, style fixups] [apw@shadowen.org: vmemmap x86_64: ensure end of section memmap is initialised] Signed-off-by: Christoph Lameter <clameter@sgi.com> Signed-off-by: Andy Whitcroft <apw@shadowen.org> Acked-by: Mel Gorman <mel@csn.ul.ie> Cc: Andi Kleen <ak@suse.de> Cc: KAMEZAWA Hiroyuki <kamezawa.hiroyu@jp.fujitsu.com> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
145 lines
4.2 KiB
C
145 lines
4.2 KiB
C
#ifndef _X86_64_PAGE_H
|
|
#define _X86_64_PAGE_H
|
|
|
|
#include <linux/const.h>
|
|
|
|
/* PAGE_SHIFT determines the page size */
|
|
#define PAGE_SHIFT 12
|
|
#define PAGE_SIZE (_AC(1,UL) << PAGE_SHIFT)
|
|
#define PAGE_MASK (~(PAGE_SIZE-1))
|
|
#define PHYSICAL_PAGE_MASK (~(PAGE_SIZE-1) & __PHYSICAL_MASK)
|
|
|
|
#define THREAD_ORDER 1
|
|
#define THREAD_SIZE (PAGE_SIZE << THREAD_ORDER)
|
|
#define CURRENT_MASK (~(THREAD_SIZE-1))
|
|
|
|
#define EXCEPTION_STACK_ORDER 0
|
|
#define EXCEPTION_STKSZ (PAGE_SIZE << EXCEPTION_STACK_ORDER)
|
|
|
|
#define DEBUG_STACK_ORDER (EXCEPTION_STACK_ORDER + 1)
|
|
#define DEBUG_STKSZ (PAGE_SIZE << DEBUG_STACK_ORDER)
|
|
|
|
#define IRQSTACK_ORDER 2
|
|
#define IRQSTACKSIZE (PAGE_SIZE << IRQSTACK_ORDER)
|
|
|
|
#define STACKFAULT_STACK 1
|
|
#define DOUBLEFAULT_STACK 2
|
|
#define NMI_STACK 3
|
|
#define DEBUG_STACK 4
|
|
#define MCE_STACK 5
|
|
#define N_EXCEPTION_STACKS 5 /* hw limit: 7 */
|
|
|
|
#define LARGE_PAGE_MASK (~(LARGE_PAGE_SIZE-1))
|
|
#define LARGE_PAGE_SIZE (_AC(1,UL) << PMD_SHIFT)
|
|
|
|
#define HPAGE_SHIFT PMD_SHIFT
|
|
#define HPAGE_SIZE (_AC(1,UL) << HPAGE_SHIFT)
|
|
#define HPAGE_MASK (~(HPAGE_SIZE - 1))
|
|
#define HUGETLB_PAGE_ORDER (HPAGE_SHIFT - PAGE_SHIFT)
|
|
|
|
#ifdef __KERNEL__
|
|
#ifndef __ASSEMBLY__
|
|
|
|
extern unsigned long end_pfn;
|
|
|
|
void clear_page(void *);
|
|
void copy_page(void *, void *);
|
|
|
|
#define clear_user_page(page, vaddr, pg) clear_page(page)
|
|
#define copy_user_page(to, from, vaddr, pg) copy_page(to, from)
|
|
|
|
#define __alloc_zeroed_user_highpage(movableflags, vma, vaddr) \
|
|
alloc_page_vma(GFP_HIGHUSER | __GFP_ZERO | movableflags, vma, vaddr)
|
|
#define __HAVE_ARCH_ALLOC_ZEROED_USER_HIGHPAGE
|
|
/*
|
|
* These are used to make use of C type-checking..
|
|
*/
|
|
typedef struct { unsigned long pte; } pte_t;
|
|
typedef struct { unsigned long pmd; } pmd_t;
|
|
typedef struct { unsigned long pud; } pud_t;
|
|
typedef struct { unsigned long pgd; } pgd_t;
|
|
#define PTE_MASK PHYSICAL_PAGE_MASK
|
|
|
|
typedef struct { unsigned long pgprot; } pgprot_t;
|
|
|
|
extern unsigned long phys_base;
|
|
|
|
#define pte_val(x) ((x).pte)
|
|
#define pmd_val(x) ((x).pmd)
|
|
#define pud_val(x) ((x).pud)
|
|
#define pgd_val(x) ((x).pgd)
|
|
#define pgprot_val(x) ((x).pgprot)
|
|
|
|
#define __pte(x) ((pte_t) { (x) } )
|
|
#define __pmd(x) ((pmd_t) { (x) } )
|
|
#define __pud(x) ((pud_t) { (x) } )
|
|
#define __pgd(x) ((pgd_t) { (x) } )
|
|
#define __pgprot(x) ((pgprot_t) { (x) } )
|
|
|
|
#endif /* !__ASSEMBLY__ */
|
|
|
|
#define __PHYSICAL_START CONFIG_PHYSICAL_START
|
|
#define __KERNEL_ALIGN 0x200000
|
|
|
|
/*
|
|
* Make sure kernel is aligned to 2MB address. Catching it at compile
|
|
* time is better. Change your config file and compile the kernel
|
|
* for a 2MB aligned address (CONFIG_PHYSICAL_START)
|
|
*/
|
|
#if (CONFIG_PHYSICAL_START % __KERNEL_ALIGN) != 0
|
|
#error "CONFIG_PHYSICAL_START must be a multiple of 2MB"
|
|
#endif
|
|
|
|
#define __START_KERNEL (__START_KERNEL_map + __PHYSICAL_START)
|
|
#define __START_KERNEL_map _AC(0xffffffff80000000, UL)
|
|
#define __PAGE_OFFSET _AC(0xffff810000000000, UL)
|
|
|
|
/* to align the pointer to the (next) page boundary */
|
|
#define PAGE_ALIGN(addr) (((addr)+PAGE_SIZE-1)&PAGE_MASK)
|
|
|
|
/* See Documentation/x86_64/mm.txt for a description of the memory map. */
|
|
#define __PHYSICAL_MASK_SHIFT 46
|
|
#define __PHYSICAL_MASK ((_AC(1,UL) << __PHYSICAL_MASK_SHIFT) - 1)
|
|
#define __VIRTUAL_MASK_SHIFT 48
|
|
#define __VIRTUAL_MASK ((_AC(1,UL) << __VIRTUAL_MASK_SHIFT) - 1)
|
|
|
|
#define KERNEL_TEXT_SIZE (40*1024*1024)
|
|
#define KERNEL_TEXT_START _AC(0xffffffff80000000, UL)
|
|
#define PAGE_OFFSET __PAGE_OFFSET
|
|
|
|
#ifndef __ASSEMBLY__
|
|
|
|
#include <asm/bug.h>
|
|
|
|
extern unsigned long __phys_addr(unsigned long);
|
|
|
|
#endif /* __ASSEMBLY__ */
|
|
|
|
#define __pa(x) __phys_addr((unsigned long)(x))
|
|
#define __pa_symbol(x) __phys_addr((unsigned long)(x))
|
|
|
|
#define __va(x) ((void *)((unsigned long)(x)+PAGE_OFFSET))
|
|
#define __boot_va(x) __va(x)
|
|
#define __boot_pa(x) __pa(x)
|
|
#ifdef CONFIG_FLATMEM
|
|
#define pfn_valid(pfn) ((pfn) < end_pfn)
|
|
#endif
|
|
|
|
#define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT)
|
|
#define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT)
|
|
#define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT)
|
|
|
|
#define VM_DATA_DEFAULT_FLAGS \
|
|
(((current->personality & READ_IMPLIES_EXEC) ? VM_EXEC : 0 ) | \
|
|
VM_READ | VM_WRITE | VM_MAYREAD | VM_MAYWRITE | VM_MAYEXEC)
|
|
|
|
#define __HAVE_ARCH_GATE_AREA 1
|
|
#define vmemmap ((struct page *)VMEMMAP_START)
|
|
|
|
#include <asm-generic/memory_model.h>
|
|
#include <asm-generic/page.h>
|
|
|
|
#endif /* __KERNEL__ */
|
|
|
|
#endif /* _X86_64_PAGE_H */
|