mirror of
https://mirrors.bfsu.edu.cn/git/linux.git
synced 2024-11-18 17:54:13 +08:00
163 lines
5.7 KiB
Plaintext
163 lines
5.7 KiB
Plaintext
|
|
||
|
====================
|
||
|
HIGH MEMORY HANDLING
|
||
|
====================
|
||
|
|
||
|
By: Peter Zijlstra <a.p.zijlstra@chello.nl>
|
||
|
|
||
|
Contents:
|
||
|
|
||
|
(*) What is high memory?
|
||
|
|
||
|
(*) Temporary virtual mappings.
|
||
|
|
||
|
(*) Using kmap_atomic.
|
||
|
|
||
|
(*) Cost of temporary mappings.
|
||
|
|
||
|
(*) i386 PAE.
|
||
|
|
||
|
|
||
|
====================
|
||
|
WHAT IS HIGH MEMORY?
|
||
|
====================
|
||
|
|
||
|
High memory (highmem) is used when the size of physical memory approaches or
|
||
|
exceeds the maximum size of virtual memory. At that point it becomes
|
||
|
impossible for the kernel to keep all of the available physical memory mapped
|
||
|
at all times. This means the kernel needs to start using temporary mappings of
|
||
|
the pieces of physical memory that it wants to access.
|
||
|
|
||
|
The part of (physical) memory not covered by a permanent mapping is what we
|
||
|
refer to as 'highmem'. There are various architecture dependent constraints on
|
||
|
where exactly that border lies.
|
||
|
|
||
|
In the i386 arch, for example, we choose to map the kernel into every process's
|
||
|
VM space so that we don't have to pay the full TLB invalidation costs for
|
||
|
kernel entry/exit. This means the available virtual memory space (4GiB on
|
||
|
i386) has to be divided between user and kernel space.
|
||
|
|
||
|
The traditional split for architectures using this approach is 3:1, 3GiB for
|
||
|
userspace and the top 1GiB for kernel space:
|
||
|
|
||
|
+--------+ 0xffffffff
|
||
|
| Kernel |
|
||
|
+--------+ 0xc0000000
|
||
|
| |
|
||
|
| User |
|
||
|
| |
|
||
|
+--------+ 0x00000000
|
||
|
|
||
|
This means that the kernel can at most map 1GiB of physical memory at any one
|
||
|
time, but because we need virtual address space for other things - including
|
||
|
temporary maps to access the rest of the physical memory - the actual direct
|
||
|
map will typically be less (usually around ~896MiB).
|
||
|
|
||
|
Other architectures that have mm context tagged TLBs can have separate kernel
|
||
|
and user maps. Some hardware (like some ARMs), however, have limited virtual
|
||
|
space when they use mm context tags.
|
||
|
|
||
|
|
||
|
==========================
|
||
|
TEMPORARY VIRTUAL MAPPINGS
|
||
|
==========================
|
||
|
|
||
|
The kernel contains several ways of creating temporary mappings:
|
||
|
|
||
|
(*) vmap(). This can be used to make a long duration mapping of multiple
|
||
|
physical pages into a contiguous virtual space. It needs global
|
||
|
synchronization to unmap.
|
||
|
|
||
|
(*) kmap(). This permits a short duration mapping of a single page. It needs
|
||
|
global synchronization, but is amortized somewhat. It is also prone to
|
||
|
deadlocks when using in a nested fashion, and so it is not recommended for
|
||
|
new code.
|
||
|
|
||
|
(*) kmap_atomic(). This permits a very short duration mapping of a single
|
||
|
page. Since the mapping is restricted to the CPU that issued it, it
|
||
|
performs well, but the issuing task is therefore required to stay on that
|
||
|
CPU until it has finished, lest some other task displace its mappings.
|
||
|
|
||
|
kmap_atomic() may also be used by interrupt contexts, since it is does not
|
||
|
sleep and the caller may not sleep until after kunmap_atomic() is called.
|
||
|
|
||
|
It may be assumed that k[un]map_atomic() won't fail.
|
||
|
|
||
|
|
||
|
=================
|
||
|
USING KMAP_ATOMIC
|
||
|
=================
|
||
|
|
||
|
When and where to use kmap_atomic() is straightforward. It is used when code
|
||
|
wants to access the contents of a page that might be allocated from high memory
|
||
|
(see __GFP_HIGHMEM), for example a page in the pagecache. The API has two
|
||
|
functions, and they can be used in a manner similar to the following:
|
||
|
|
||
|
/* Find the page of interest. */
|
||
|
struct page *page = find_get_page(mapping, offset);
|
||
|
|
||
|
/* Gain access to the contents of that page. */
|
||
|
void *vaddr = kmap_atomic(page);
|
||
|
|
||
|
/* Do something to the contents of that page. */
|
||
|
memset(vaddr, 0, PAGE_SIZE);
|
||
|
|
||
|
/* Unmap that page. */
|
||
|
kunmap_atomic(vaddr);
|
||
|
|
||
|
Note that the kunmap_atomic() call takes the result of the kmap_atomic() call
|
||
|
not the argument.
|
||
|
|
||
|
If you need to map two pages because you want to copy from one page to
|
||
|
another you need to keep the kmap_atomic calls strictly nested, like:
|
||
|
|
||
|
vaddr1 = kmap_atomic(page1);
|
||
|
vaddr2 = kmap_atomic(page2);
|
||
|
|
||
|
memcpy(vaddr1, vaddr2, PAGE_SIZE);
|
||
|
|
||
|
kunmap_atomic(vaddr2);
|
||
|
kunmap_atomic(vaddr1);
|
||
|
|
||
|
|
||
|
==========================
|
||
|
COST OF TEMPORARY MAPPINGS
|
||
|
==========================
|
||
|
|
||
|
The cost of creating temporary mappings can be quite high. The arch has to
|
||
|
manipulate the kernel's page tables, the data TLB and/or the MMU's registers.
|
||
|
|
||
|
If CONFIG_HIGHMEM is not set, then the kernel will try and create a mapping
|
||
|
simply with a bit of arithmetic that will convert the page struct address into
|
||
|
a pointer to the page contents rather than juggling mappings about. In such a
|
||
|
case, the unmap operation may be a null operation.
|
||
|
|
||
|
If CONFIG_MMU is not set, then there can be no temporary mappings and no
|
||
|
highmem. In such a case, the arithmetic approach will also be used.
|
||
|
|
||
|
|
||
|
========
|
||
|
i386 PAE
|
||
|
========
|
||
|
|
||
|
The i386 arch, under some circumstances, will permit you to stick up to 64GiB
|
||
|
of RAM into your 32-bit machine. This has a number of consequences:
|
||
|
|
||
|
(*) Linux needs a page-frame structure for each page in the system and the
|
||
|
pageframes need to live in the permanent mapping, which means:
|
||
|
|
||
|
(*) you can have 896M/sizeof(struct page) page-frames at most; with struct
|
||
|
page being 32-bytes that would end up being something in the order of 112G
|
||
|
worth of pages; the kernel, however, needs to store more than just
|
||
|
page-frames in that memory...
|
||
|
|
||
|
(*) PAE makes your page tables larger - which slows the system down as more
|
||
|
data has to be accessed to traverse in TLB fills and the like. One
|
||
|
advantage is that PAE has more PTE bits and can provide advanced features
|
||
|
like NX and PAT.
|
||
|
|
||
|
The general recommendation is that you don't use more than 8GiB on a 32-bit
|
||
|
machine - although more might work for you and your workload, you're pretty
|
||
|
much on your own - don't expect kernel developers to really care much if things
|
||
|
come apart.
|