Merge branch 'lmb-to-memblock' of git://git.kernel.org/pub/scm/linux/kernel/git/benh/powerpc

* 'lmb-to-memblock' of git://git.kernel.org/pub/scm/linux/kernel/git/benh/powerpc:
  lmb: rename to memblock
This commit is contained in:
Linus Torvalds 2010-07-14 17:27:44 -07:00
commit 6f7dd68b75
72 changed files with 1025 additions and 1025 deletions

View File

@ -1265,7 +1265,7 @@ and is between 256 and 4096 characters. It is defined in the file
If there are multiple matching configurations changing If there are multiple matching configurations changing
the same attribute, the last one is used. the same attribute, the last one is used.
lmb=debug [KNL] Enable lmb debug messages. memblock=debug [KNL] Enable memblock debug messages.
load_ramdisk= [RAM] List of ramdisks to load from floppy load_ramdisk= [RAM] List of ramdisks to load from floppy
See Documentation/blockdev/ramdisk.txt. See Documentation/blockdev/ramdisk.txt.

View File

@ -5,7 +5,7 @@ mainmenu "Linux/Microblaze Kernel Configuration"
config MICROBLAZE config MICROBLAZE
def_bool y def_bool y
select HAVE_LMB select HAVE_MEMBLOCK
select HAVE_FUNCTION_TRACER select HAVE_FUNCTION_TRACER
select HAVE_FUNCTION_TRACE_MCOUNT_TEST select HAVE_FUNCTION_TRACE_MCOUNT_TEST
select HAVE_FUNCTION_GRAPH_TRACER select HAVE_FUNCTION_GRAPH_TRACER

View File

@ -6,12 +6,12 @@
* for more details. * for more details.
*/ */
#ifndef _ASM_MICROBLAZE_LMB_H #ifndef _ASM_MICROBLAZE_MEMBLOCK_H
#define _ASM_MICROBLAZE_LMB_H #define _ASM_MICROBLAZE_MEMBLOCK_H
/* LMB limit is OFF */ /* MEMBLOCK limit is OFF */
#define LMB_REAL_LIMIT 0xFFFFFFFF #define MEMBLOCK_REAL_LIMIT 0xFFFFFFFF
#endif /* _ASM_MICROBLAZE_LMB_H */ #endif /* _ASM_MICROBLAZE_MEMBLOCK_H */

View File

@ -29,7 +29,7 @@
#include <linux/kexec.h> #include <linux/kexec.h>
#include <linux/debugfs.h> #include <linux/debugfs.h>
#include <linux/irq.h> #include <linux/irq.h>
#include <linux/lmb.h> #include <linux/memblock.h>
#include <asm/prom.h> #include <asm/prom.h>
#include <asm/page.h> #include <asm/page.h>
@ -49,12 +49,12 @@ void __init early_init_dt_scan_chosen_arch(unsigned long node)
void __init early_init_dt_add_memory_arch(u64 base, u64 size) void __init early_init_dt_add_memory_arch(u64 base, u64 size)
{ {
lmb_add(base, size); memblock_add(base, size);
} }
u64 __init early_init_dt_alloc_memory_arch(u64 size, u64 align) u64 __init early_init_dt_alloc_memory_arch(u64 size, u64 align)
{ {
return lmb_alloc(size, align); return memblock_alloc(size, align);
} }
#ifdef CONFIG_EARLY_PRINTK #ifdef CONFIG_EARLY_PRINTK
@ -104,8 +104,8 @@ void __init early_init_devtree(void *params)
*/ */
of_scan_flat_dt(early_init_dt_scan_chosen, NULL); of_scan_flat_dt(early_init_dt_scan_chosen, NULL);
/* Scan memory nodes and rebuild LMBs */ /* Scan memory nodes and rebuild MEMBLOCKs */
lmb_init(); memblock_init();
of_scan_flat_dt(early_init_dt_scan_root, NULL); of_scan_flat_dt(early_init_dt_scan_root, NULL);
of_scan_flat_dt(early_init_dt_scan_memory, NULL); of_scan_flat_dt(early_init_dt_scan_memory, NULL);
@ -113,9 +113,9 @@ void __init early_init_devtree(void *params)
strlcpy(boot_command_line, cmd_line, COMMAND_LINE_SIZE); strlcpy(boot_command_line, cmd_line, COMMAND_LINE_SIZE);
parse_early_param(); parse_early_param();
lmb_analyze(); memblock_analyze();
pr_debug("Phys. mem: %lx\n", (unsigned long) lmb_phys_mem_size()); pr_debug("Phys. mem: %lx\n", (unsigned long) memblock_phys_mem_size());
pr_debug(" <- early_init_devtree()\n"); pr_debug(" <- early_init_devtree()\n");
} }

View File

@ -10,7 +10,7 @@
#include <linux/bootmem.h> #include <linux/bootmem.h>
#include <linux/init.h> #include <linux/init.h>
#include <linux/kernel.h> #include <linux/kernel.h>
#include <linux/lmb.h> #include <linux/memblock.h>
#include <linux/mm.h> /* mem_init */ #include <linux/mm.h> /* mem_init */
#include <linux/initrd.h> #include <linux/initrd.h>
#include <linux/pagemap.h> #include <linux/pagemap.h>
@ -76,10 +76,10 @@ void __init setup_memory(void)
u32 kernel_align_start, kernel_align_size; u32 kernel_align_start, kernel_align_size;
/* Find main memory where is the kernel */ /* Find main memory where is the kernel */
for (i = 0; i < lmb.memory.cnt; i++) { for (i = 0; i < memblock.memory.cnt; i++) {
memory_start = (u32) lmb.memory.region[i].base; memory_start = (u32) memblock.memory.region[i].base;
memory_end = (u32) lmb.memory.region[i].base memory_end = (u32) memblock.memory.region[i].base
+ (u32) lmb.memory.region[i].size; + (u32) memblock.memory.region[i].size;
if ((memory_start <= (u32)_text) && if ((memory_start <= (u32)_text) &&
((u32)_text <= memory_end)) { ((u32)_text <= memory_end)) {
memory_size = memory_end - memory_start; memory_size = memory_end - memory_start;
@ -100,7 +100,7 @@ void __init setup_memory(void)
kernel_align_start = PAGE_DOWN((u32)_text); kernel_align_start = PAGE_DOWN((u32)_text);
/* ALIGN can be remove because _end in vmlinux.lds.S is align */ /* ALIGN can be remove because _end in vmlinux.lds.S is align */
kernel_align_size = PAGE_UP((u32)klimit) - kernel_align_start; kernel_align_size = PAGE_UP((u32)klimit) - kernel_align_start;
lmb_reserve(kernel_align_start, kernel_align_size); memblock_reserve(kernel_align_start, kernel_align_size);
printk(KERN_INFO "%s: kernel addr=0x%08x-0x%08x size=0x%08x\n", printk(KERN_INFO "%s: kernel addr=0x%08x-0x%08x size=0x%08x\n",
__func__, kernel_align_start, kernel_align_start __func__, kernel_align_start, kernel_align_start
+ kernel_align_size, kernel_align_size); + kernel_align_size, kernel_align_size);
@ -141,18 +141,18 @@ void __init setup_memory(void)
map_size = init_bootmem_node(&contig_page_data, map_size = init_bootmem_node(&contig_page_data,
PFN_UP(TOPHYS((u32)klimit)), min_low_pfn, max_low_pfn); PFN_UP(TOPHYS((u32)klimit)), min_low_pfn, max_low_pfn);
#endif #endif
lmb_reserve(PFN_UP(TOPHYS((u32)klimit)) << PAGE_SHIFT, map_size); memblock_reserve(PFN_UP(TOPHYS((u32)klimit)) << PAGE_SHIFT, map_size);
/* free bootmem is whole main memory */ /* free bootmem is whole main memory */
free_bootmem(memory_start, memory_size); free_bootmem(memory_start, memory_size);
/* reserve allocate blocks */ /* reserve allocate blocks */
for (i = 0; i < lmb.reserved.cnt; i++) { for (i = 0; i < memblock.reserved.cnt; i++) {
pr_debug("reserved %d - 0x%08x-0x%08x\n", i, pr_debug("reserved %d - 0x%08x-0x%08x\n", i,
(u32) lmb.reserved.region[i].base, (u32) memblock.reserved.region[i].base,
(u32) lmb_size_bytes(&lmb.reserved, i)); (u32) memblock_size_bytes(&memblock.reserved, i));
reserve_bootmem(lmb.reserved.region[i].base, reserve_bootmem(memblock.reserved.region[i].base,
lmb_size_bytes(&lmb.reserved, i) - 1, BOOTMEM_DEFAULT); memblock_size_bytes(&memblock.reserved, i) - 1, BOOTMEM_DEFAULT);
} }
#ifdef CONFIG_MMU #ifdef CONFIG_MMU
init_bootmem_done = 1; init_bootmem_done = 1;
@ -235,7 +235,7 @@ static void mm_cmdline_setup(void)
if (maxmem && memory_size > maxmem) { if (maxmem && memory_size > maxmem) {
memory_size = maxmem; memory_size = maxmem;
memory_end = memory_start + memory_size; memory_end = memory_start + memory_size;
lmb.memory.region[0].size = memory_size; memblock.memory.region[0].size = memory_size;
} }
} }
} }
@ -273,19 +273,19 @@ asmlinkage void __init mmu_init(void)
{ {
unsigned int kstart, ksize; unsigned int kstart, ksize;
if (!lmb.reserved.cnt) { if (!memblock.reserved.cnt) {
printk(KERN_EMERG "Error memory count\n"); printk(KERN_EMERG "Error memory count\n");
machine_restart(NULL); machine_restart(NULL);
} }
if ((u32) lmb.memory.region[0].size < 0x1000000) { if ((u32) memblock.memory.region[0].size < 0x1000000) {
printk(KERN_EMERG "Memory must be greater than 16MB\n"); printk(KERN_EMERG "Memory must be greater than 16MB\n");
machine_restart(NULL); machine_restart(NULL);
} }
/* Find main memory where the kernel is */ /* Find main memory where the kernel is */
memory_start = (u32) lmb.memory.region[0].base; memory_start = (u32) memblock.memory.region[0].base;
memory_end = (u32) lmb.memory.region[0].base + memory_end = (u32) memblock.memory.region[0].base +
(u32) lmb.memory.region[0].size; (u32) memblock.memory.region[0].size;
memory_size = memory_end - memory_start; memory_size = memory_end - memory_start;
mm_cmdline_setup(); /* FIXME parse args from command line - not used */ mm_cmdline_setup(); /* FIXME parse args from command line - not used */
@ -297,7 +297,7 @@ asmlinkage void __init mmu_init(void)
kstart = __pa(CONFIG_KERNEL_START); /* kernel start */ kstart = __pa(CONFIG_KERNEL_START); /* kernel start */
/* kernel size */ /* kernel size */
ksize = PAGE_ALIGN(((u32)_end - (u32)CONFIG_KERNEL_START)); ksize = PAGE_ALIGN(((u32)_end - (u32)CONFIG_KERNEL_START));
lmb_reserve(kstart, ksize); memblock_reserve(kstart, ksize);
#if defined(CONFIG_BLK_DEV_INITRD) #if defined(CONFIG_BLK_DEV_INITRD)
/* Remove the init RAM disk from the available memory. */ /* Remove the init RAM disk from the available memory. */
@ -335,7 +335,7 @@ void __init *early_get_page(void)
* Mem start + 32MB -> here is limit * Mem start + 32MB -> here is limit
* because of mem mapping from head.S * because of mem mapping from head.S
*/ */
p = __va(lmb_alloc_base(PAGE_SIZE, PAGE_SIZE, p = __va(memblock_alloc_base(PAGE_SIZE, PAGE_SIZE,
memory_start + 0x2000000)); memory_start + 0x2000000));
} }
return p; return p;

View File

@ -132,7 +132,7 @@ config PPC
select HAVE_ARCH_KGDB select HAVE_ARCH_KGDB
select HAVE_KRETPROBES select HAVE_KRETPROBES
select HAVE_ARCH_TRACEHOOK select HAVE_ARCH_TRACEHOOK
select HAVE_LMB select HAVE_MEMBLOCK
select HAVE_DMA_ATTRS select HAVE_DMA_ATTRS
select HAVE_DMA_API_DEBUG select HAVE_DMA_API_DEBUG
select USE_GENERIC_SMP_HELPERS if SMP select USE_GENERIC_SMP_HELPERS if SMP

View File

@ -12,7 +12,7 @@
* 2 of the License, or (at your option) any later version. * 2 of the License, or (at your option) any later version.
*/ */
#include <linux/lmb.h> #include <linux/memblock.h>
#include <asm/types.h> #include <asm/types.h>
#include <asm/page.h> #include <asm/page.h>

View File

@ -1,15 +0,0 @@
#ifndef _ASM_POWERPC_LMB_H
#define _ASM_POWERPC_LMB_H
#include <asm/udbg.h>
#define LMB_DBG(fmt...) udbg_printf(fmt)
#ifdef CONFIG_PPC32
extern phys_addr_t lowmem_end_addr;
#define LMB_REAL_LIMIT lowmem_end_addr
#else
#define LMB_REAL_LIMIT 0
#endif
#endif /* _ASM_POWERPC_LMB_H */

View File

@ -0,0 +1,15 @@
#ifndef _ASM_POWERPC_MEMBLOCK_H
#define _ASM_POWERPC_MEMBLOCK_H
#include <asm/udbg.h>
#define MEMBLOCK_DBG(fmt...) udbg_printf(fmt)
#ifdef CONFIG_PPC32
extern phys_addr_t lowmem_end_addr;
#define MEMBLOCK_REAL_LIMIT lowmem_end_addr
#else
#define MEMBLOCK_REAL_LIMIT 0
#endif
#endif /* _ASM_POWERPC_MEMBLOCK_H */

View File

@ -7,7 +7,7 @@
#include <linux/string.h> #include <linux/string.h>
#include <linux/init.h> #include <linux/init.h>
#include <linux/module.h> #include <linux/module.h>
#include <linux/lmb.h> #include <linux/memblock.h>
#include <asm/sections.h> #include <asm/sections.h>
#include <asm/prom.h> #include <asm/prom.h>

View File

@ -24,7 +24,7 @@
#include <linux/init.h> #include <linux/init.h>
#include <linux/irq.h> #include <linux/irq.h>
#include <linux/types.h> #include <linux/types.h>
#include <linux/lmb.h> #include <linux/memblock.h>
#include <asm/processor.h> #include <asm/processor.h>
#include <asm/machdep.h> #include <asm/machdep.h>

View File

@ -13,7 +13,7 @@
#include <linux/crash_dump.h> #include <linux/crash_dump.h>
#include <linux/bootmem.h> #include <linux/bootmem.h>
#include <linux/lmb.h> #include <linux/memblock.h>
#include <asm/code-patching.h> #include <asm/code-patching.h>
#include <asm/kdump.h> #include <asm/kdump.h>
#include <asm/prom.h> #include <asm/prom.h>
@ -33,7 +33,7 @@ unsigned long long elfcorehdr_addr = ELFCORE_ADDR_MAX;
#ifndef CONFIG_RELOCATABLE #ifndef CONFIG_RELOCATABLE
void __init reserve_kdump_trampoline(void) void __init reserve_kdump_trampoline(void)
{ {
lmb_reserve(0, KDUMP_RESERVE_LIMIT); memblock_reserve(0, KDUMP_RESERVE_LIMIT);
} }
static void __init create_trampoline(unsigned long addr) static void __init create_trampoline(unsigned long addr)

View File

@ -71,7 +71,7 @@ static int ppc_swiotlb_bus_notify(struct notifier_block *nb,
sd->max_direct_dma_addr = 0; sd->max_direct_dma_addr = 0;
/* May need to bounce if the device can't address all of DRAM */ /* May need to bounce if the device can't address all of DRAM */
if ((dma_get_mask(dev) + 1) < lmb_end_of_DRAM()) if ((dma_get_mask(dev) + 1) < memblock_end_of_DRAM())
set_dma_ops(dev, &swiotlb_dma_ops); set_dma_ops(dev, &swiotlb_dma_ops);
return NOTIFY_DONE; return NOTIFY_DONE;

View File

@ -9,7 +9,7 @@
#include <linux/dma-mapping.h> #include <linux/dma-mapping.h>
#include <linux/dma-debug.h> #include <linux/dma-debug.h>
#include <linux/gfp.h> #include <linux/gfp.h>
#include <linux/lmb.h> #include <linux/memblock.h>
#include <asm/bug.h> #include <asm/bug.h>
#include <asm/abs_addr.h> #include <asm/abs_addr.h>
@ -89,7 +89,7 @@ static int dma_direct_dma_supported(struct device *dev, u64 mask)
/* Could be improved so platforms can set the limit in case /* Could be improved so platforms can set the limit in case
* they have limited DMA windows * they have limited DMA windows
*/ */
return mask >= (lmb_end_of_DRAM() - 1); return mask >= (memblock_end_of_DRAM() - 1);
#else #else
return 1; return 1;
#endif #endif

View File

@ -12,7 +12,7 @@
#include <linux/kexec.h> #include <linux/kexec.h>
#include <linux/reboot.h> #include <linux/reboot.h>
#include <linux/threads.h> #include <linux/threads.h>
#include <linux/lmb.h> #include <linux/memblock.h>
#include <linux/of.h> #include <linux/of.h>
#include <asm/machdep.h> #include <asm/machdep.h>
#include <asm/prom.h> #include <asm/prom.h>
@ -66,11 +66,11 @@ void __init reserve_crashkernel(void)
unsigned long long crash_size, crash_base; unsigned long long crash_size, crash_base;
int ret; int ret;
/* this is necessary because of lmb_phys_mem_size() */ /* this is necessary because of memblock_phys_mem_size() */
lmb_analyze(); memblock_analyze();
/* use common parsing */ /* use common parsing */
ret = parse_crashkernel(boot_command_line, lmb_phys_mem_size(), ret = parse_crashkernel(boot_command_line, memblock_phys_mem_size(),
&crash_size, &crash_base); &crash_size, &crash_base);
if (ret == 0 && crash_size > 0) { if (ret == 0 && crash_size > 0) {
crashk_res.start = crash_base; crashk_res.start = crash_base;
@ -133,9 +133,9 @@ void __init reserve_crashkernel(void)
"for crashkernel (System RAM: %ldMB)\n", "for crashkernel (System RAM: %ldMB)\n",
(unsigned long)(crash_size >> 20), (unsigned long)(crash_size >> 20),
(unsigned long)(crashk_res.start >> 20), (unsigned long)(crashk_res.start >> 20),
(unsigned long)(lmb_phys_mem_size() >> 20)); (unsigned long)(memblock_phys_mem_size() >> 20));
lmb_reserve(crashk_res.start, crash_size); memblock_reserve(crashk_res.start, crash_size);
} }
int overlaps_crashkernel(unsigned long start, unsigned long size) int overlaps_crashkernel(unsigned long start, unsigned long size)

View File

@ -9,7 +9,7 @@
#include <linux/threads.h> #include <linux/threads.h>
#include <linux/module.h> #include <linux/module.h>
#include <linux/lmb.h> #include <linux/memblock.h>
#include <asm/firmware.h> #include <asm/firmware.h>
#include <asm/lppaca.h> #include <asm/lppaca.h>
@ -117,7 +117,7 @@ void __init allocate_pacas(void)
* the first segment. On iSeries they must be within the area mapped * the first segment. On iSeries they must be within the area mapped
* by the HV, which is HvPagesToMap * HVPAGESIZE bytes. * by the HV, which is HvPagesToMap * HVPAGESIZE bytes.
*/ */
limit = min(0x10000000ULL, lmb.rmo_size); limit = min(0x10000000ULL, memblock.rmo_size);
if (firmware_has_feature(FW_FEATURE_ISERIES)) if (firmware_has_feature(FW_FEATURE_ISERIES))
limit = min(limit, HvPagesToMap * HVPAGESIZE); limit = min(limit, HvPagesToMap * HVPAGESIZE);
@ -128,7 +128,7 @@ void __init allocate_pacas(void)
paca_size = PAGE_ALIGN(sizeof(struct paca_struct) * nr_cpus); paca_size = PAGE_ALIGN(sizeof(struct paca_struct) * nr_cpus);
paca = __va(lmb_alloc_base(paca_size, PAGE_SIZE, limit)); paca = __va(memblock_alloc_base(paca_size, PAGE_SIZE, limit));
memset(paca, 0, paca_size); memset(paca, 0, paca_size);
printk(KERN_DEBUG "Allocated %u bytes for %d pacas at %p\n", printk(KERN_DEBUG "Allocated %u bytes for %d pacas at %p\n",
@ -148,7 +148,7 @@ void __init free_unused_pacas(void)
if (new_size >= paca_size) if (new_size >= paca_size)
return; return;
lmb_free(__pa(paca) + new_size, paca_size - new_size); memblock_free(__pa(paca) + new_size, paca_size - new_size);
printk(KERN_DEBUG "Freed %u bytes for unused pacas\n", printk(KERN_DEBUG "Freed %u bytes for unused pacas\n",
paca_size - new_size); paca_size - new_size);

View File

@ -31,7 +31,7 @@
#include <linux/kexec.h> #include <linux/kexec.h>
#include <linux/debugfs.h> #include <linux/debugfs.h>
#include <linux/irq.h> #include <linux/irq.h>
#include <linux/lmb.h> #include <linux/memblock.h>
#include <asm/prom.h> #include <asm/prom.h>
#include <asm/rtas.h> #include <asm/rtas.h>
@ -98,7 +98,7 @@ static void __init move_device_tree(void)
if ((memory_limit && (start + size) > memory_limit) || if ((memory_limit && (start + size) > memory_limit) ||
overlaps_crashkernel(start, size)) { overlaps_crashkernel(start, size)) {
p = __va(lmb_alloc_base(size, PAGE_SIZE, lmb.rmo_size)); p = __va(memblock_alloc_base(size, PAGE_SIZE, memblock.rmo_size));
memcpy(p, initial_boot_params, size); memcpy(p, initial_boot_params, size);
initial_boot_params = (struct boot_param_header *)p; initial_boot_params = (struct boot_param_header *)p;
DBG("Moved device tree to 0x%p\n", p); DBG("Moved device tree to 0x%p\n", p);
@ -411,13 +411,13 @@ static int __init early_init_dt_scan_drconf_memory(unsigned long node)
{ {
__be32 *dm, *ls, *usm; __be32 *dm, *ls, *usm;
unsigned long l, n, flags; unsigned long l, n, flags;
u64 base, size, lmb_size; u64 base, size, memblock_size;
unsigned int is_kexec_kdump = 0, rngs; unsigned int is_kexec_kdump = 0, rngs;
ls = of_get_flat_dt_prop(node, "ibm,lmb-size", &l); ls = of_get_flat_dt_prop(node, "ibm,memblock-size", &l);
if (ls == NULL || l < dt_root_size_cells * sizeof(__be32)) if (ls == NULL || l < dt_root_size_cells * sizeof(__be32))
return 0; return 0;
lmb_size = dt_mem_next_cell(dt_root_size_cells, &ls); memblock_size = dt_mem_next_cell(dt_root_size_cells, &ls);
dm = of_get_flat_dt_prop(node, "ibm,dynamic-memory", &l); dm = of_get_flat_dt_prop(node, "ibm,dynamic-memory", &l);
if (dm == NULL || l < sizeof(__be32)) if (dm == NULL || l < sizeof(__be32))
@ -442,11 +442,11 @@ static int __init early_init_dt_scan_drconf_memory(unsigned long node)
or if the block is not assigned to this partition (0x8) */ or if the block is not assigned to this partition (0x8) */
if ((flags & 0x80) || !(flags & 0x8)) if ((flags & 0x80) || !(flags & 0x8))
continue; continue;
size = lmb_size; size = memblock_size;
rngs = 1; rngs = 1;
if (is_kexec_kdump) { if (is_kexec_kdump) {
/* /*
* For each lmb in ibm,dynamic-memory, a corresponding * For each memblock in ibm,dynamic-memory, a corresponding
* entry in linux,drconf-usable-memory property contains * entry in linux,drconf-usable-memory property contains
* a counter 'p' followed by 'p' (base, size) duple. * a counter 'p' followed by 'p' (base, size) duple.
* Now read the counter from * Now read the counter from
@ -469,10 +469,10 @@ static int __init early_init_dt_scan_drconf_memory(unsigned long node)
if ((base + size) > 0x80000000ul) if ((base + size) > 0x80000000ul)
size = 0x80000000ul - base; size = 0x80000000ul - base;
} }
lmb_add(base, size); memblock_add(base, size);
} while (--rngs); } while (--rngs);
} }
lmb_dump_all(); memblock_dump_all();
return 0; return 0;
} }
#else #else
@ -501,14 +501,14 @@ void __init early_init_dt_add_memory_arch(u64 base, u64 size)
} }
#endif #endif
lmb_add(base, size); memblock_add(base, size);
memstart_addr = min((u64)memstart_addr, base); memstart_addr = min((u64)memstart_addr, base);
} }
u64 __init early_init_dt_alloc_memory_arch(u64 size, u64 align) u64 __init early_init_dt_alloc_memory_arch(u64 size, u64 align)
{ {
return lmb_alloc(size, align); return memblock_alloc(size, align);
} }
#ifdef CONFIG_BLK_DEV_INITRD #ifdef CONFIG_BLK_DEV_INITRD
@ -534,12 +534,12 @@ static void __init early_reserve_mem(void)
/* before we do anything, lets reserve the dt blob */ /* before we do anything, lets reserve the dt blob */
self_base = __pa((unsigned long)initial_boot_params); self_base = __pa((unsigned long)initial_boot_params);
self_size = initial_boot_params->totalsize; self_size = initial_boot_params->totalsize;
lmb_reserve(self_base, self_size); memblock_reserve(self_base, self_size);
#ifdef CONFIG_BLK_DEV_INITRD #ifdef CONFIG_BLK_DEV_INITRD
/* then reserve the initrd, if any */ /* then reserve the initrd, if any */
if (initrd_start && (initrd_end > initrd_start)) if (initrd_start && (initrd_end > initrd_start))
lmb_reserve(__pa(initrd_start), initrd_end - initrd_start); memblock_reserve(__pa(initrd_start), initrd_end - initrd_start);
#endif /* CONFIG_BLK_DEV_INITRD */ #endif /* CONFIG_BLK_DEV_INITRD */
#ifdef CONFIG_PPC32 #ifdef CONFIG_PPC32
@ -560,7 +560,7 @@ static void __init early_reserve_mem(void)
if (base_32 == self_base && size_32 == self_size) if (base_32 == self_base && size_32 == self_size)
continue; continue;
DBG("reserving: %x -> %x\n", base_32, size_32); DBG("reserving: %x -> %x\n", base_32, size_32);
lmb_reserve(base_32, size_32); memblock_reserve(base_32, size_32);
} }
return; return;
} }
@ -571,7 +571,7 @@ static void __init early_reserve_mem(void)
if (size == 0) if (size == 0)
break; break;
DBG("reserving: %llx -> %llx\n", base, size); DBG("reserving: %llx -> %llx\n", base, size);
lmb_reserve(base, size); memblock_reserve(base, size);
} }
} }
@ -594,7 +594,7 @@ static inline unsigned long phyp_dump_calculate_reserve_size(void)
return phyp_dump_info->reserve_bootvar; return phyp_dump_info->reserve_bootvar;
/* divide by 20 to get 5% of value */ /* divide by 20 to get 5% of value */
tmp = lmb_end_of_DRAM(); tmp = memblock_end_of_DRAM();
do_div(tmp, 20); do_div(tmp, 20);
/* round it down in multiples of 256 */ /* round it down in multiples of 256 */
@ -633,11 +633,11 @@ static void __init phyp_dump_reserve_mem(void)
if (phyp_dump_info->phyp_dump_is_active) { if (phyp_dump_info->phyp_dump_is_active) {
/* Reserve *everything* above RMR.Area freed by userland tools*/ /* Reserve *everything* above RMR.Area freed by userland tools*/
base = variable_reserve_size; base = variable_reserve_size;
size = lmb_end_of_DRAM() - base; size = memblock_end_of_DRAM() - base;
/* XXX crashed_ram_end is wrong, since it may be beyond /* XXX crashed_ram_end is wrong, since it may be beyond
* the memory_limit, it will need to be adjusted. */ * the memory_limit, it will need to be adjusted. */
lmb_reserve(base, size); memblock_reserve(base, size);
phyp_dump_info->init_reserve_start = base; phyp_dump_info->init_reserve_start = base;
phyp_dump_info->init_reserve_size = size; phyp_dump_info->init_reserve_size = size;
@ -645,8 +645,8 @@ static void __init phyp_dump_reserve_mem(void)
size = phyp_dump_info->cpu_state_size + size = phyp_dump_info->cpu_state_size +
phyp_dump_info->hpte_region_size + phyp_dump_info->hpte_region_size +
variable_reserve_size; variable_reserve_size;
base = lmb_end_of_DRAM() - size; base = memblock_end_of_DRAM() - size;
lmb_reserve(base, size); memblock_reserve(base, size);
phyp_dump_info->init_reserve_start = base; phyp_dump_info->init_reserve_start = base;
phyp_dump_info->init_reserve_size = size; phyp_dump_info->init_reserve_size = size;
} }
@ -681,8 +681,8 @@ void __init early_init_devtree(void *params)
*/ */
of_scan_flat_dt(early_init_dt_scan_chosen, NULL); of_scan_flat_dt(early_init_dt_scan_chosen, NULL);
/* Scan memory nodes and rebuild LMBs */ /* Scan memory nodes and rebuild MEMBLOCKs */
lmb_init(); memblock_init();
of_scan_flat_dt(early_init_dt_scan_root, NULL); of_scan_flat_dt(early_init_dt_scan_root, NULL);
of_scan_flat_dt(early_init_dt_scan_memory_ppc, NULL); of_scan_flat_dt(early_init_dt_scan_memory_ppc, NULL);
@ -690,11 +690,11 @@ void __init early_init_devtree(void *params)
strlcpy(boot_command_line, cmd_line, COMMAND_LINE_SIZE); strlcpy(boot_command_line, cmd_line, COMMAND_LINE_SIZE);
parse_early_param(); parse_early_param();
/* Reserve LMB regions used by kernel, initrd, dt, etc... */ /* Reserve MEMBLOCK regions used by kernel, initrd, dt, etc... */
lmb_reserve(PHYSICAL_START, __pa(klimit) - PHYSICAL_START); memblock_reserve(PHYSICAL_START, __pa(klimit) - PHYSICAL_START);
/* If relocatable, reserve first 32k for interrupt vectors etc. */ /* If relocatable, reserve first 32k for interrupt vectors etc. */
if (PHYSICAL_START > MEMORY_START) if (PHYSICAL_START > MEMORY_START)
lmb_reserve(MEMORY_START, 0x8000); memblock_reserve(MEMORY_START, 0x8000);
reserve_kdump_trampoline(); reserve_kdump_trampoline();
reserve_crashkernel(); reserve_crashkernel();
early_reserve_mem(); early_reserve_mem();
@ -706,17 +706,17 @@ void __init early_init_devtree(void *params)
/* Ensure that total memory size is page-aligned, because /* Ensure that total memory size is page-aligned, because
* otherwise mark_bootmem() gets upset. */ * otherwise mark_bootmem() gets upset. */
lmb_analyze(); memblock_analyze();
memsize = lmb_phys_mem_size(); memsize = memblock_phys_mem_size();
if ((memsize & PAGE_MASK) != memsize) if ((memsize & PAGE_MASK) != memsize)
limit = memsize & PAGE_MASK; limit = memsize & PAGE_MASK;
} }
lmb_enforce_memory_limit(limit); memblock_enforce_memory_limit(limit);
lmb_analyze(); memblock_analyze();
lmb_dump_all(); memblock_dump_all();
DBG("Phys. mem: %llx\n", lmb_phys_mem_size()); DBG("Phys. mem: %llx\n", memblock_phys_mem_size());
/* We may need to relocate the flat tree, do it now. /* We may need to relocate the flat tree, do it now.
* FIXME .. and the initrd too? */ * FIXME .. and the initrd too? */

View File

@ -22,7 +22,7 @@
#include <linux/smp.h> #include <linux/smp.h>
#include <linux/completion.h> #include <linux/completion.h>
#include <linux/cpumask.h> #include <linux/cpumask.h>
#include <linux/lmb.h> #include <linux/memblock.h>
#include <linux/slab.h> #include <linux/slab.h>
#include <asm/prom.h> #include <asm/prom.h>
@ -934,11 +934,11 @@ void __init rtas_initialize(void)
*/ */
#ifdef CONFIG_PPC64 #ifdef CONFIG_PPC64
if (machine_is(pseries) && firmware_has_feature(FW_FEATURE_LPAR)) { if (machine_is(pseries) && firmware_has_feature(FW_FEATURE_LPAR)) {
rtas_region = min(lmb.rmo_size, RTAS_INSTANTIATE_MAX); rtas_region = min(memblock.rmo_size, RTAS_INSTANTIATE_MAX);
ibm_suspend_me_token = rtas_token("ibm,suspend-me"); ibm_suspend_me_token = rtas_token("ibm,suspend-me");
} }
#endif #endif
rtas_rmo_buf = lmb_alloc_base(RTAS_RMOBUF_MAX, PAGE_SIZE, rtas_region); rtas_rmo_buf = memblock_alloc_base(RTAS_RMOBUF_MAX, PAGE_SIZE, rtas_region);
#ifdef CONFIG_RTAS_ERROR_LOGGING #ifdef CONFIG_RTAS_ERROR_LOGGING
rtas_last_error_token = rtas_token("rtas-last-error"); rtas_last_error_token = rtas_token("rtas-last-error");

View File

@ -33,7 +33,7 @@
#include <linux/serial_8250.h> #include <linux/serial_8250.h>
#include <linux/debugfs.h> #include <linux/debugfs.h>
#include <linux/percpu.h> #include <linux/percpu.h>
#include <linux/lmb.h> #include <linux/memblock.h>
#include <linux/of_platform.h> #include <linux/of_platform.h>
#include <asm/io.h> #include <asm/io.h>
#include <asm/paca.h> #include <asm/paca.h>

View File

@ -16,7 +16,7 @@
#include <linux/root_dev.h> #include <linux/root_dev.h>
#include <linux/cpu.h> #include <linux/cpu.h>
#include <linux/console.h> #include <linux/console.h>
#include <linux/lmb.h> #include <linux/memblock.h>
#include <asm/io.h> #include <asm/io.h>
#include <asm/prom.h> #include <asm/prom.h>
@ -246,12 +246,12 @@ static void __init irqstack_early_init(void)
unsigned int i; unsigned int i;
/* interrupt stacks must be in lowmem, we get that for free on ppc32 /* interrupt stacks must be in lowmem, we get that for free on ppc32
* as the lmb is limited to lowmem by LMB_REAL_LIMIT */ * as the memblock is limited to lowmem by MEMBLOCK_REAL_LIMIT */
for_each_possible_cpu(i) { for_each_possible_cpu(i) {
softirq_ctx[i] = (struct thread_info *) softirq_ctx[i] = (struct thread_info *)
__va(lmb_alloc(THREAD_SIZE, THREAD_SIZE)); __va(memblock_alloc(THREAD_SIZE, THREAD_SIZE));
hardirq_ctx[i] = (struct thread_info *) hardirq_ctx[i] = (struct thread_info *)
__va(lmb_alloc(THREAD_SIZE, THREAD_SIZE)); __va(memblock_alloc(THREAD_SIZE, THREAD_SIZE));
} }
} }
@ -261,15 +261,15 @@ static void __init exc_lvl_early_init(void)
unsigned int i; unsigned int i;
/* interrupt stacks must be in lowmem, we get that for free on ppc32 /* interrupt stacks must be in lowmem, we get that for free on ppc32
* as the lmb is limited to lowmem by LMB_REAL_LIMIT */ * as the memblock is limited to lowmem by MEMBLOCK_REAL_LIMIT */
for_each_possible_cpu(i) { for_each_possible_cpu(i) {
critirq_ctx[i] = (struct thread_info *) critirq_ctx[i] = (struct thread_info *)
__va(lmb_alloc(THREAD_SIZE, THREAD_SIZE)); __va(memblock_alloc(THREAD_SIZE, THREAD_SIZE));
#ifdef CONFIG_BOOKE #ifdef CONFIG_BOOKE
dbgirq_ctx[i] = (struct thread_info *) dbgirq_ctx[i] = (struct thread_info *)
__va(lmb_alloc(THREAD_SIZE, THREAD_SIZE)); __va(memblock_alloc(THREAD_SIZE, THREAD_SIZE));
mcheckirq_ctx[i] = (struct thread_info *) mcheckirq_ctx[i] = (struct thread_info *)
__va(lmb_alloc(THREAD_SIZE, THREAD_SIZE)); __va(memblock_alloc(THREAD_SIZE, THREAD_SIZE));
#endif #endif
} }
} }

View File

@ -34,7 +34,7 @@
#include <linux/bootmem.h> #include <linux/bootmem.h>
#include <linux/pci.h> #include <linux/pci.h>
#include <linux/lockdep.h> #include <linux/lockdep.h>
#include <linux/lmb.h> #include <linux/memblock.h>
#include <asm/io.h> #include <asm/io.h>
#include <asm/kdump.h> #include <asm/kdump.h>
#include <asm/prom.h> #include <asm/prom.h>
@ -158,7 +158,7 @@ static void __init setup_paca(struct paca_struct *new_paca)
* the CPU that ignores the top 2 bits of the address in real * the CPU that ignores the top 2 bits of the address in real
* mode so we can access kernel globals normally provided we * mode so we can access kernel globals normally provided we
* only toy with things in the RMO region. From here, we do * only toy with things in the RMO region. From here, we do
* some early parsing of the device-tree to setup out LMB * some early parsing of the device-tree to setup out MEMBLOCK
* data structures, and allocate & initialize the hash table * data structures, and allocate & initialize the hash table
* and segment tables so we can start running with translation * and segment tables so we can start running with translation
* enabled. * enabled.
@ -404,7 +404,7 @@ void __init setup_system(void)
printk("-----------------------------------------------------\n"); printk("-----------------------------------------------------\n");
printk("ppc64_pft_size = 0x%llx\n", ppc64_pft_size); printk("ppc64_pft_size = 0x%llx\n", ppc64_pft_size);
printk("physicalMemorySize = 0x%llx\n", lmb_phys_mem_size()); printk("physicalMemorySize = 0x%llx\n", memblock_phys_mem_size());
if (ppc64_caches.dline_size != 0x80) if (ppc64_caches.dline_size != 0x80)
printk("ppc64_caches.dcache_line_size = 0x%x\n", printk("ppc64_caches.dcache_line_size = 0x%x\n",
ppc64_caches.dline_size); ppc64_caches.dline_size);
@ -443,10 +443,10 @@ static void __init irqstack_early_init(void)
*/ */
for_each_possible_cpu(i) { for_each_possible_cpu(i) {
softirq_ctx[i] = (struct thread_info *) softirq_ctx[i] = (struct thread_info *)
__va(lmb_alloc_base(THREAD_SIZE, __va(memblock_alloc_base(THREAD_SIZE,
THREAD_SIZE, limit)); THREAD_SIZE, limit));
hardirq_ctx[i] = (struct thread_info *) hardirq_ctx[i] = (struct thread_info *)
__va(lmb_alloc_base(THREAD_SIZE, __va(memblock_alloc_base(THREAD_SIZE,
THREAD_SIZE, limit)); THREAD_SIZE, limit));
} }
} }
@ -458,11 +458,11 @@ static void __init exc_lvl_early_init(void)
for_each_possible_cpu(i) { for_each_possible_cpu(i) {
critirq_ctx[i] = (struct thread_info *) critirq_ctx[i] = (struct thread_info *)
__va(lmb_alloc(THREAD_SIZE, THREAD_SIZE)); __va(memblock_alloc(THREAD_SIZE, THREAD_SIZE));
dbgirq_ctx[i] = (struct thread_info *) dbgirq_ctx[i] = (struct thread_info *)
__va(lmb_alloc(THREAD_SIZE, THREAD_SIZE)); __va(memblock_alloc(THREAD_SIZE, THREAD_SIZE));
mcheckirq_ctx[i] = (struct thread_info *) mcheckirq_ctx[i] = (struct thread_info *)
__va(lmb_alloc(THREAD_SIZE, THREAD_SIZE)); __va(memblock_alloc(THREAD_SIZE, THREAD_SIZE));
} }
} }
#else #else
@ -487,11 +487,11 @@ static void __init emergency_stack_init(void)
* bringup, we need to get at them in real mode. This means they * bringup, we need to get at them in real mode. This means they
* must also be within the RMO region. * must also be within the RMO region.
*/ */
limit = min(slb0_limit(), lmb.rmo_size); limit = min(slb0_limit(), memblock.rmo_size);
for_each_possible_cpu(i) { for_each_possible_cpu(i) {
unsigned long sp; unsigned long sp;
sp = lmb_alloc_base(THREAD_SIZE, THREAD_SIZE, limit); sp = memblock_alloc_base(THREAD_SIZE, THREAD_SIZE, limit);
sp += THREAD_SIZE; sp += THREAD_SIZE;
paca[i].emergency_sp = __va(sp); paca[i].emergency_sp = __va(sp);
} }

View File

@ -22,7 +22,7 @@
#include <linux/elf.h> #include <linux/elf.h>
#include <linux/security.h> #include <linux/security.h>
#include <linux/bootmem.h> #include <linux/bootmem.h>
#include <linux/lmb.h> #include <linux/memblock.h>
#include <asm/pgtable.h> #include <asm/pgtable.h>
#include <asm/system.h> #include <asm/system.h>
@ -734,7 +734,7 @@ static int __init vdso_init(void)
vdso_data->platform = machine_is(iseries) ? 0x200 : 0x100; vdso_data->platform = machine_is(iseries) ? 0x200 : 0x100;
if (firmware_has_feature(FW_FEATURE_LPAR)) if (firmware_has_feature(FW_FEATURE_LPAR))
vdso_data->platform |= 1; vdso_data->platform |= 1;
vdso_data->physicalMemorySize = lmb_phys_mem_size(); vdso_data->physicalMemorySize = memblock_phys_mem_size();
vdso_data->dcache_size = ppc64_caches.dsize; vdso_data->dcache_size = ppc64_caches.dsize;
vdso_data->dcache_line_size = ppc64_caches.dline_size; vdso_data->dcache_line_size = ppc64_caches.dline_size;
vdso_data->icache_size = ppc64_caches.isize; vdso_data->icache_size = ppc64_caches.isize;

View File

@ -135,7 +135,7 @@ unsigned long __init mmu_mapin_ram(unsigned long top)
/* If the size of RAM is not an exact power of two, we may not /* If the size of RAM is not an exact power of two, we may not
* have covered RAM in its entirety with 16 and 4 MiB * have covered RAM in its entirety with 16 and 4 MiB
* pages. Consequently, restrict the top end of RAM currently * pages. Consequently, restrict the top end of RAM currently
* allocable so that calls to the LMB to allocate PTEs for "tail" * allocable so that calls to the MEMBLOCK to allocate PTEs for "tail"
* coverage with normal-sized pages (or other reasons) do not * coverage with normal-sized pages (or other reasons) do not
* attempt to allocate outside the allowed range. * attempt to allocate outside the allowed range.
*/ */

View File

@ -31,7 +31,7 @@
#include <linux/cache.h> #include <linux/cache.h>
#include <linux/init.h> #include <linux/init.h>
#include <linux/signal.h> #include <linux/signal.h>
#include <linux/lmb.h> #include <linux/memblock.h>
#include <asm/processor.h> #include <asm/processor.h>
#include <asm/pgtable.h> #include <asm/pgtable.h>
@ -384,8 +384,8 @@ static int __init htab_dt_scan_hugepage_blocks(unsigned long node,
printk(KERN_INFO "Huge page(16GB) memory: " printk(KERN_INFO "Huge page(16GB) memory: "
"addr = 0x%lX size = 0x%lX pages = %d\n", "addr = 0x%lX size = 0x%lX pages = %d\n",
phys_addr, block_size, expected_pages); phys_addr, block_size, expected_pages);
if (phys_addr + (16 * GB) <= lmb_end_of_DRAM()) { if (phys_addr + (16 * GB) <= memblock_end_of_DRAM()) {
lmb_reserve(phys_addr, block_size * expected_pages); memblock_reserve(phys_addr, block_size * expected_pages);
add_gpage(phys_addr, block_size, expected_pages); add_gpage(phys_addr, block_size, expected_pages);
} }
return 0; return 0;
@ -458,7 +458,7 @@ static void __init htab_init_page_sizes(void)
* and we have at least 1G of RAM at boot * and we have at least 1G of RAM at boot
*/ */
if (mmu_psize_defs[MMU_PAGE_16M].shift && if (mmu_psize_defs[MMU_PAGE_16M].shift &&
lmb_phys_mem_size() >= 0x40000000) memblock_phys_mem_size() >= 0x40000000)
mmu_vmemmap_psize = MMU_PAGE_16M; mmu_vmemmap_psize = MMU_PAGE_16M;
else if (mmu_psize_defs[MMU_PAGE_64K].shift) else if (mmu_psize_defs[MMU_PAGE_64K].shift)
mmu_vmemmap_psize = MMU_PAGE_64K; mmu_vmemmap_psize = MMU_PAGE_64K;
@ -520,7 +520,7 @@ static unsigned long __init htab_get_table_size(void)
return 1UL << ppc64_pft_size; return 1UL << ppc64_pft_size;
/* round mem_size up to next power of 2 */ /* round mem_size up to next power of 2 */
mem_size = lmb_phys_mem_size(); mem_size = memblock_phys_mem_size();
rnd_mem_size = 1UL << __ilog2(mem_size); rnd_mem_size = 1UL << __ilog2(mem_size);
if (rnd_mem_size < mem_size) if (rnd_mem_size < mem_size)
rnd_mem_size <<= 1; rnd_mem_size <<= 1;
@ -627,7 +627,7 @@ static void __init htab_initialize(void)
else else
limit = 0; limit = 0;
table = lmb_alloc_base(htab_size_bytes, htab_size_bytes, limit); table = memblock_alloc_base(htab_size_bytes, htab_size_bytes, limit);
DBG("Hash table allocated at %lx, size: %lx\n", table, DBG("Hash table allocated at %lx, size: %lx\n", table,
htab_size_bytes); htab_size_bytes);
@ -647,9 +647,9 @@ static void __init htab_initialize(void)
prot = pgprot_val(PAGE_KERNEL); prot = pgprot_val(PAGE_KERNEL);
#ifdef CONFIG_DEBUG_PAGEALLOC #ifdef CONFIG_DEBUG_PAGEALLOC
linear_map_hash_count = lmb_end_of_DRAM() >> PAGE_SHIFT; linear_map_hash_count = memblock_end_of_DRAM() >> PAGE_SHIFT;
linear_map_hash_slots = __va(lmb_alloc_base(linear_map_hash_count, linear_map_hash_slots = __va(memblock_alloc_base(linear_map_hash_count,
1, lmb.rmo_size)); 1, memblock.rmo_size));
memset(linear_map_hash_slots, 0, linear_map_hash_count); memset(linear_map_hash_slots, 0, linear_map_hash_count);
#endif /* CONFIG_DEBUG_PAGEALLOC */ #endif /* CONFIG_DEBUG_PAGEALLOC */
@ -659,16 +659,16 @@ static void __init htab_initialize(void)
*/ */
/* create bolted the linear mapping in the hash table */ /* create bolted the linear mapping in the hash table */
for (i=0; i < lmb.memory.cnt; i++) { for (i=0; i < memblock.memory.cnt; i++) {
base = (unsigned long)__va(lmb.memory.region[i].base); base = (unsigned long)__va(memblock.memory.region[i].base);
size = lmb.memory.region[i].size; size = memblock.memory.region[i].size;
DBG("creating mapping for region: %lx..%lx (prot: %lx)\n", DBG("creating mapping for region: %lx..%lx (prot: %lx)\n",
base, size, prot); base, size, prot);
#ifdef CONFIG_U3_DART #ifdef CONFIG_U3_DART
/* Do not map the DART space. Fortunately, it will be aligned /* Do not map the DART space. Fortunately, it will be aligned
* in such a way that it will not cross two lmb regions and * in such a way that it will not cross two memblock regions and
* will fit within a single 16Mb page. * will fit within a single 16Mb page.
* The DART space is assumed to be a full 16Mb region even if * The DART space is assumed to be a full 16Mb region even if
* we only use 2Mb of that space. We will use more of it later * we only use 2Mb of that space. We will use more of it later

View File

@ -30,7 +30,7 @@
#include <linux/highmem.h> #include <linux/highmem.h>
#include <linux/initrd.h> #include <linux/initrd.h>
#include <linux/pagemap.h> #include <linux/pagemap.h>
#include <linux/lmb.h> #include <linux/memblock.h>
#include <linux/gfp.h> #include <linux/gfp.h>
#include <asm/pgalloc.h> #include <asm/pgalloc.h>
@ -136,17 +136,17 @@ void __init MMU_init(void)
/* parse args from command line */ /* parse args from command line */
MMU_setup(); MMU_setup();
if (lmb.memory.cnt > 1) { if (memblock.memory.cnt > 1) {
#ifndef CONFIG_WII #ifndef CONFIG_WII
lmb.memory.cnt = 1; memblock.memory.cnt = 1;
lmb_analyze(); memblock_analyze();
printk(KERN_WARNING "Only using first contiguous memory region"); printk(KERN_WARNING "Only using first contiguous memory region");
#else #else
wii_memory_fixups(); wii_memory_fixups();
#endif #endif
} }
total_lowmem = total_memory = lmb_end_of_DRAM() - memstart_addr; total_lowmem = total_memory = memblock_end_of_DRAM() - memstart_addr;
lowmem_end_addr = memstart_addr + total_lowmem; lowmem_end_addr = memstart_addr + total_lowmem;
#ifdef CONFIG_FSL_BOOKE #ifdef CONFIG_FSL_BOOKE
@ -161,8 +161,8 @@ void __init MMU_init(void)
lowmem_end_addr = memstart_addr + total_lowmem; lowmem_end_addr = memstart_addr + total_lowmem;
#ifndef CONFIG_HIGHMEM #ifndef CONFIG_HIGHMEM
total_memory = total_lowmem; total_memory = total_lowmem;
lmb_enforce_memory_limit(lowmem_end_addr); memblock_enforce_memory_limit(lowmem_end_addr);
lmb_analyze(); memblock_analyze();
#endif /* CONFIG_HIGHMEM */ #endif /* CONFIG_HIGHMEM */
} }
@ -200,7 +200,7 @@ void __init *early_get_page(void)
if (init_bootmem_done) { if (init_bootmem_done) {
p = alloc_bootmem_pages(PAGE_SIZE); p = alloc_bootmem_pages(PAGE_SIZE);
} else { } else {
p = __va(lmb_alloc_base(PAGE_SIZE, PAGE_SIZE, p = __va(memblock_alloc_base(PAGE_SIZE, PAGE_SIZE,
__initial_memory_limit_addr)); __initial_memory_limit_addr));
} }
return p; return p;

View File

@ -40,7 +40,7 @@
#include <linux/nodemask.h> #include <linux/nodemask.h>
#include <linux/module.h> #include <linux/module.h>
#include <linux/poison.h> #include <linux/poison.h>
#include <linux/lmb.h> #include <linux/memblock.h>
#include <linux/hugetlb.h> #include <linux/hugetlb.h>
#include <linux/slab.h> #include <linux/slab.h>

View File

@ -32,7 +32,7 @@
#include <linux/initrd.h> #include <linux/initrd.h>
#include <linux/pagemap.h> #include <linux/pagemap.h>
#include <linux/suspend.h> #include <linux/suspend.h>
#include <linux/lmb.h> #include <linux/memblock.h>
#include <linux/hugetlb.h> #include <linux/hugetlb.h>
#include <asm/pgalloc.h> #include <asm/pgalloc.h>
@ -83,13 +83,13 @@ int page_is_ram(unsigned long pfn)
#else #else
unsigned long paddr = (pfn << PAGE_SHIFT); unsigned long paddr = (pfn << PAGE_SHIFT);
int i; int i;
for (i=0; i < lmb.memory.cnt; i++) { for (i=0; i < memblock.memory.cnt; i++) {
unsigned long base; unsigned long base;
base = lmb.memory.region[i].base; base = memblock.memory.region[i].base;
if ((paddr >= base) && if ((paddr >= base) &&
(paddr < (base + lmb.memory.region[i].size))) { (paddr < (base + memblock.memory.region[i].size))) {
return 1; return 1;
} }
} }
@ -142,14 +142,14 @@ int arch_add_memory(int nid, u64 start, u64 size)
/* /*
* walk_memory_resource() needs to make sure there is no holes in a given * walk_memory_resource() needs to make sure there is no holes in a given
* memory range. PPC64 does not maintain the memory layout in /proc/iomem. * memory range. PPC64 does not maintain the memory layout in /proc/iomem.
* Instead it maintains it in lmb.memory structures. Walk through the * Instead it maintains it in memblock.memory structures. Walk through the
* memory regions, find holes and callback for contiguous regions. * memory regions, find holes and callback for contiguous regions.
*/ */
int int
walk_system_ram_range(unsigned long start_pfn, unsigned long nr_pages, walk_system_ram_range(unsigned long start_pfn, unsigned long nr_pages,
void *arg, int (*func)(unsigned long, unsigned long, void *)) void *arg, int (*func)(unsigned long, unsigned long, void *))
{ {
struct lmb_property res; struct memblock_property res;
unsigned long pfn, len; unsigned long pfn, len;
u64 end; u64 end;
int ret = -1; int ret = -1;
@ -158,7 +158,7 @@ walk_system_ram_range(unsigned long start_pfn, unsigned long nr_pages,
res.size = (u64) nr_pages << PAGE_SHIFT; res.size = (u64) nr_pages << PAGE_SHIFT;
end = res.base + res.size - 1; end = res.base + res.size - 1;
while ((res.base < end) && (lmb_find(&res) >= 0)) { while ((res.base < end) && (memblock_find(&res) >= 0)) {
pfn = (unsigned long)(res.base >> PAGE_SHIFT); pfn = (unsigned long)(res.base >> PAGE_SHIFT);
len = (unsigned long)(res.size >> PAGE_SHIFT); len = (unsigned long)(res.size >> PAGE_SHIFT);
ret = (*func)(pfn, len, arg); ret = (*func)(pfn, len, arg);
@ -184,8 +184,8 @@ void __init do_init_bootmem(void)
unsigned long total_pages; unsigned long total_pages;
int boot_mapsize; int boot_mapsize;
max_low_pfn = max_pfn = lmb_end_of_DRAM() >> PAGE_SHIFT; max_low_pfn = max_pfn = memblock_end_of_DRAM() >> PAGE_SHIFT;
total_pages = (lmb_end_of_DRAM() - memstart_addr) >> PAGE_SHIFT; total_pages = (memblock_end_of_DRAM() - memstart_addr) >> PAGE_SHIFT;
#ifdef CONFIG_HIGHMEM #ifdef CONFIG_HIGHMEM
total_pages = total_lowmem >> PAGE_SHIFT; total_pages = total_lowmem >> PAGE_SHIFT;
max_low_pfn = lowmem_end_addr >> PAGE_SHIFT; max_low_pfn = lowmem_end_addr >> PAGE_SHIFT;
@ -198,16 +198,16 @@ void __init do_init_bootmem(void)
*/ */
bootmap_pages = bootmem_bootmap_pages(total_pages); bootmap_pages = bootmem_bootmap_pages(total_pages);
start = lmb_alloc(bootmap_pages << PAGE_SHIFT, PAGE_SIZE); start = memblock_alloc(bootmap_pages << PAGE_SHIFT, PAGE_SIZE);
min_low_pfn = MEMORY_START >> PAGE_SHIFT; min_low_pfn = MEMORY_START >> PAGE_SHIFT;
boot_mapsize = init_bootmem_node(NODE_DATA(0), start >> PAGE_SHIFT, min_low_pfn, max_low_pfn); boot_mapsize = init_bootmem_node(NODE_DATA(0), start >> PAGE_SHIFT, min_low_pfn, max_low_pfn);
/* Add active regions with valid PFNs */ /* Add active regions with valid PFNs */
for (i = 0; i < lmb.memory.cnt; i++) { for (i = 0; i < memblock.memory.cnt; i++) {
unsigned long start_pfn, end_pfn; unsigned long start_pfn, end_pfn;
start_pfn = lmb.memory.region[i].base >> PAGE_SHIFT; start_pfn = memblock.memory.region[i].base >> PAGE_SHIFT;
end_pfn = start_pfn + lmb_size_pages(&lmb.memory, i); end_pfn = start_pfn + memblock_size_pages(&memblock.memory, i);
add_active_range(0, start_pfn, end_pfn); add_active_range(0, start_pfn, end_pfn);
} }
@ -218,17 +218,17 @@ void __init do_init_bootmem(void)
free_bootmem_with_active_regions(0, lowmem_end_addr >> PAGE_SHIFT); free_bootmem_with_active_regions(0, lowmem_end_addr >> PAGE_SHIFT);
/* reserve the sections we're already using */ /* reserve the sections we're already using */
for (i = 0; i < lmb.reserved.cnt; i++) { for (i = 0; i < memblock.reserved.cnt; i++) {
unsigned long addr = lmb.reserved.region[i].base + unsigned long addr = memblock.reserved.region[i].base +
lmb_size_bytes(&lmb.reserved, i) - 1; memblock_size_bytes(&memblock.reserved, i) - 1;
if (addr < lowmem_end_addr) if (addr < lowmem_end_addr)
reserve_bootmem(lmb.reserved.region[i].base, reserve_bootmem(memblock.reserved.region[i].base,
lmb_size_bytes(&lmb.reserved, i), memblock_size_bytes(&memblock.reserved, i),
BOOTMEM_DEFAULT); BOOTMEM_DEFAULT);
else if (lmb.reserved.region[i].base < lowmem_end_addr) { else if (memblock.reserved.region[i].base < lowmem_end_addr) {
unsigned long adjusted_size = lowmem_end_addr - unsigned long adjusted_size = lowmem_end_addr -
lmb.reserved.region[i].base; memblock.reserved.region[i].base;
reserve_bootmem(lmb.reserved.region[i].base, reserve_bootmem(memblock.reserved.region[i].base,
adjusted_size, BOOTMEM_DEFAULT); adjusted_size, BOOTMEM_DEFAULT);
} }
} }
@ -236,9 +236,9 @@ void __init do_init_bootmem(void)
free_bootmem_with_active_regions(0, max_pfn); free_bootmem_with_active_regions(0, max_pfn);
/* reserve the sections we're already using */ /* reserve the sections we're already using */
for (i = 0; i < lmb.reserved.cnt; i++) for (i = 0; i < memblock.reserved.cnt; i++)
reserve_bootmem(lmb.reserved.region[i].base, reserve_bootmem(memblock.reserved.region[i].base,
lmb_size_bytes(&lmb.reserved, i), memblock_size_bytes(&memblock.reserved, i),
BOOTMEM_DEFAULT); BOOTMEM_DEFAULT);
#endif #endif
@ -251,20 +251,20 @@ void __init do_init_bootmem(void)
/* mark pages that don't exist as nosave */ /* mark pages that don't exist as nosave */
static int __init mark_nonram_nosave(void) static int __init mark_nonram_nosave(void)
{ {
unsigned long lmb_next_region_start_pfn, unsigned long memblock_next_region_start_pfn,
lmb_region_max_pfn; memblock_region_max_pfn;
int i; int i;
for (i = 0; i < lmb.memory.cnt - 1; i++) { for (i = 0; i < memblock.memory.cnt - 1; i++) {
lmb_region_max_pfn = memblock_region_max_pfn =
(lmb.memory.region[i].base >> PAGE_SHIFT) + (memblock.memory.region[i].base >> PAGE_SHIFT) +
(lmb.memory.region[i].size >> PAGE_SHIFT); (memblock.memory.region[i].size >> PAGE_SHIFT);
lmb_next_region_start_pfn = memblock_next_region_start_pfn =
lmb.memory.region[i+1].base >> PAGE_SHIFT; memblock.memory.region[i+1].base >> PAGE_SHIFT;
if (lmb_region_max_pfn < lmb_next_region_start_pfn) if (memblock_region_max_pfn < memblock_next_region_start_pfn)
register_nosave_region(lmb_region_max_pfn, register_nosave_region(memblock_region_max_pfn,
lmb_next_region_start_pfn); memblock_next_region_start_pfn);
} }
return 0; return 0;
@ -275,8 +275,8 @@ static int __init mark_nonram_nosave(void)
*/ */
void __init paging_init(void) void __init paging_init(void)
{ {
unsigned long total_ram = lmb_phys_mem_size(); unsigned long total_ram = memblock_phys_mem_size();
phys_addr_t top_of_ram = lmb_end_of_DRAM(); phys_addr_t top_of_ram = memblock_end_of_DRAM();
unsigned long max_zone_pfns[MAX_NR_ZONES]; unsigned long max_zone_pfns[MAX_NR_ZONES];
#ifdef CONFIG_PPC32 #ifdef CONFIG_PPC32
@ -327,7 +327,7 @@ void __init mem_init(void)
swiotlb_init(1); swiotlb_init(1);
#endif #endif
num_physpages = lmb.memory.size >> PAGE_SHIFT; num_physpages = memblock.memory.size >> PAGE_SHIFT;
high_memory = (void *) __va(max_low_pfn * PAGE_SIZE); high_memory = (void *) __va(max_low_pfn * PAGE_SIZE);
#ifdef CONFIG_NEED_MULTIPLE_NODES #ifdef CONFIG_NEED_MULTIPLE_NODES
@ -364,7 +364,7 @@ void __init mem_init(void)
highmem_mapnr = lowmem_end_addr >> PAGE_SHIFT; highmem_mapnr = lowmem_end_addr >> PAGE_SHIFT;
for (pfn = highmem_mapnr; pfn < max_mapnr; ++pfn) { for (pfn = highmem_mapnr; pfn < max_mapnr; ++pfn) {
struct page *page = pfn_to_page(pfn); struct page *page = pfn_to_page(pfn);
if (lmb_is_reserved(pfn << PAGE_SHIFT)) if (memblock_is_reserved(pfn << PAGE_SHIFT))
continue; continue;
ClearPageReserved(page); ClearPageReserved(page);
init_page_count(page); init_page_count(page);

View File

@ -17,7 +17,7 @@
#include <linux/nodemask.h> #include <linux/nodemask.h>
#include <linux/cpu.h> #include <linux/cpu.h>
#include <linux/notifier.h> #include <linux/notifier.h>
#include <linux/lmb.h> #include <linux/memblock.h>
#include <linux/of.h> #include <linux/of.h>
#include <linux/pfn.h> #include <linux/pfn.h>
#include <asm/sparsemem.h> #include <asm/sparsemem.h>
@ -351,7 +351,7 @@ struct of_drconf_cell {
#define DRCONF_MEM_RESERVED 0x00000080 #define DRCONF_MEM_RESERVED 0x00000080
/* /*
* Read the next lmb list entry from the ibm,dynamic-memory property * Read the next memblock list entry from the ibm,dynamic-memory property
* and return the information in the provided of_drconf_cell structure. * and return the information in the provided of_drconf_cell structure.
*/ */
static void read_drconf_cell(struct of_drconf_cell *drmem, const u32 **cellp) static void read_drconf_cell(struct of_drconf_cell *drmem, const u32 **cellp)
@ -372,8 +372,8 @@ static void read_drconf_cell(struct of_drconf_cell *drmem, const u32 **cellp)
/* /*
* Retreive and validate the ibm,dynamic-memory property of the device tree. * Retreive and validate the ibm,dynamic-memory property of the device tree.
* *
* The layout of the ibm,dynamic-memory property is a number N of lmb * The layout of the ibm,dynamic-memory property is a number N of memblock
* list entries followed by N lmb list entries. Each lmb list entry * list entries followed by N memblock list entries. Each memblock list entry
* contains information as layed out in the of_drconf_cell struct above. * contains information as layed out in the of_drconf_cell struct above.
*/ */
static int of_get_drconf_memory(struct device_node *memory, const u32 **dm) static int of_get_drconf_memory(struct device_node *memory, const u32 **dm)
@ -398,15 +398,15 @@ static int of_get_drconf_memory(struct device_node *memory, const u32 **dm)
} }
/* /*
* Retreive and validate the ibm,lmb-size property for drconf memory * Retreive and validate the ibm,memblock-size property for drconf memory
* from the device tree. * from the device tree.
*/ */
static u64 of_get_lmb_size(struct device_node *memory) static u64 of_get_memblock_size(struct device_node *memory)
{ {
const u32 *prop; const u32 *prop;
u32 len; u32 len;
prop = of_get_property(memory, "ibm,lmb-size", &len); prop = of_get_property(memory, "ibm,memblock-size", &len);
if (!prop || len < sizeof(unsigned int)) if (!prop || len < sizeof(unsigned int))
return 0; return 0;
@ -540,19 +540,19 @@ static unsigned long __init numa_enforce_memory_limit(unsigned long start,
unsigned long size) unsigned long size)
{ {
/* /*
* We use lmb_end_of_DRAM() in here instead of memory_limit because * We use memblock_end_of_DRAM() in here instead of memory_limit because
* we've already adjusted it for the limit and it takes care of * we've already adjusted it for the limit and it takes care of
* having memory holes below the limit. Also, in the case of * having memory holes below the limit. Also, in the case of
* iommu_is_off, memory_limit is not set but is implicitly enforced. * iommu_is_off, memory_limit is not set but is implicitly enforced.
*/ */
if (start + size <= lmb_end_of_DRAM()) if (start + size <= memblock_end_of_DRAM())
return size; return size;
if (start >= lmb_end_of_DRAM()) if (start >= memblock_end_of_DRAM())
return 0; return 0;
return lmb_end_of_DRAM() - start; return memblock_end_of_DRAM() - start;
} }
/* /*
@ -562,7 +562,7 @@ static unsigned long __init numa_enforce_memory_limit(unsigned long start,
static inline int __init read_usm_ranges(const u32 **usm) static inline int __init read_usm_ranges(const u32 **usm)
{ {
/* /*
* For each lmb in ibm,dynamic-memory a corresponding * For each memblock in ibm,dynamic-memory a corresponding
* entry in linux,drconf-usable-memory property contains * entry in linux,drconf-usable-memory property contains
* a counter followed by that many (base, size) duple. * a counter followed by that many (base, size) duple.
* read the counter from linux,drconf-usable-memory * read the counter from linux,drconf-usable-memory
@ -578,7 +578,7 @@ static void __init parse_drconf_memory(struct device_node *memory)
{ {
const u32 *dm, *usm; const u32 *dm, *usm;
unsigned int n, rc, ranges, is_kexec_kdump = 0; unsigned int n, rc, ranges, is_kexec_kdump = 0;
unsigned long lmb_size, base, size, sz; unsigned long memblock_size, base, size, sz;
int nid; int nid;
struct assoc_arrays aa; struct assoc_arrays aa;
@ -586,8 +586,8 @@ static void __init parse_drconf_memory(struct device_node *memory)
if (!n) if (!n)
return; return;
lmb_size = of_get_lmb_size(memory); memblock_size = of_get_memblock_size(memory);
if (!lmb_size) if (!memblock_size)
return; return;
rc = of_get_assoc_arrays(memory, &aa); rc = of_get_assoc_arrays(memory, &aa);
@ -611,7 +611,7 @@ static void __init parse_drconf_memory(struct device_node *memory)
continue; continue;
base = drmem.base_addr; base = drmem.base_addr;
size = lmb_size; size = memblock_size;
ranges = 1; ranges = 1;
if (is_kexec_kdump) { if (is_kexec_kdump) {
@ -731,7 +731,7 @@ new_range:
} }
/* /*
* Now do the same thing for each LMB listed in the ibm,dynamic-memory * Now do the same thing for each MEMBLOCK listed in the ibm,dynamic-memory
* property in the ibm,dynamic-reconfiguration-memory node. * property in the ibm,dynamic-reconfiguration-memory node.
*/ */
memory = of_find_node_by_path("/ibm,dynamic-reconfiguration-memory"); memory = of_find_node_by_path("/ibm,dynamic-reconfiguration-memory");
@ -743,8 +743,8 @@ new_range:
static void __init setup_nonnuma(void) static void __init setup_nonnuma(void)
{ {
unsigned long top_of_ram = lmb_end_of_DRAM(); unsigned long top_of_ram = memblock_end_of_DRAM();
unsigned long total_ram = lmb_phys_mem_size(); unsigned long total_ram = memblock_phys_mem_size();
unsigned long start_pfn, end_pfn; unsigned long start_pfn, end_pfn;
unsigned int i, nid = 0; unsigned int i, nid = 0;
@ -753,9 +753,9 @@ static void __init setup_nonnuma(void)
printk(KERN_DEBUG "Memory hole size: %ldMB\n", printk(KERN_DEBUG "Memory hole size: %ldMB\n",
(top_of_ram - total_ram) >> 20); (top_of_ram - total_ram) >> 20);
for (i = 0; i < lmb.memory.cnt; ++i) { for (i = 0; i < memblock.memory.cnt; ++i) {
start_pfn = lmb.memory.region[i].base >> PAGE_SHIFT; start_pfn = memblock.memory.region[i].base >> PAGE_SHIFT;
end_pfn = start_pfn + lmb_size_pages(&lmb.memory, i); end_pfn = start_pfn + memblock_size_pages(&memblock.memory, i);
fake_numa_create_new_node(end_pfn, &nid); fake_numa_create_new_node(end_pfn, &nid);
add_active_range(nid, start_pfn, end_pfn); add_active_range(nid, start_pfn, end_pfn);
@ -813,7 +813,7 @@ static void __init dump_numa_memory_topology(void)
count = 0; count = 0;
for (i = 0; i < lmb_end_of_DRAM(); for (i = 0; i < memblock_end_of_DRAM();
i += (1 << SECTION_SIZE_BITS)) { i += (1 << SECTION_SIZE_BITS)) {
if (early_pfn_to_nid(i >> PAGE_SHIFT) == node) { if (early_pfn_to_nid(i >> PAGE_SHIFT) == node) {
if (count == 0) if (count == 0)
@ -833,7 +833,7 @@ static void __init dump_numa_memory_topology(void)
} }
/* /*
* Allocate some memory, satisfying the lmb or bootmem allocator where * Allocate some memory, satisfying the memblock or bootmem allocator where
* required. nid is the preferred node and end is the physical address of * required. nid is the preferred node and end is the physical address of
* the highest address in the node. * the highest address in the node.
* *
@ -847,11 +847,11 @@ static void __init *careful_zallocation(int nid, unsigned long size,
int new_nid; int new_nid;
unsigned long ret_paddr; unsigned long ret_paddr;
ret_paddr = __lmb_alloc_base(size, align, end_pfn << PAGE_SHIFT); ret_paddr = __memblock_alloc_base(size, align, end_pfn << PAGE_SHIFT);
/* retry over all memory */ /* retry over all memory */
if (!ret_paddr) if (!ret_paddr)
ret_paddr = __lmb_alloc_base(size, align, lmb_end_of_DRAM()); ret_paddr = __memblock_alloc_base(size, align, memblock_end_of_DRAM());
if (!ret_paddr) if (!ret_paddr)
panic("numa.c: cannot allocate %lu bytes for node %d", panic("numa.c: cannot allocate %lu bytes for node %d",
@ -861,14 +861,14 @@ static void __init *careful_zallocation(int nid, unsigned long size,
/* /*
* We initialize the nodes in numeric order: 0, 1, 2... * We initialize the nodes in numeric order: 0, 1, 2...
* and hand over control from the LMB allocator to the * and hand over control from the MEMBLOCK allocator to the
* bootmem allocator. If this function is called for * bootmem allocator. If this function is called for
* node 5, then we know that all nodes <5 are using the * node 5, then we know that all nodes <5 are using the
* bootmem allocator instead of the LMB allocator. * bootmem allocator instead of the MEMBLOCK allocator.
* *
* So, check the nid from which this allocation came * So, check the nid from which this allocation came
* and double check to see if we need to use bootmem * and double check to see if we need to use bootmem
* instead of the LMB. We don't free the LMB memory * instead of the MEMBLOCK. We don't free the MEMBLOCK memory
* since it would be useless. * since it would be useless.
*/ */
new_nid = early_pfn_to_nid(ret_paddr >> PAGE_SHIFT); new_nid = early_pfn_to_nid(ret_paddr >> PAGE_SHIFT);
@ -893,9 +893,9 @@ static void mark_reserved_regions_for_nid(int nid)
struct pglist_data *node = NODE_DATA(nid); struct pglist_data *node = NODE_DATA(nid);
int i; int i;
for (i = 0; i < lmb.reserved.cnt; i++) { for (i = 0; i < memblock.reserved.cnt; i++) {
unsigned long physbase = lmb.reserved.region[i].base; unsigned long physbase = memblock.reserved.region[i].base;
unsigned long size = lmb.reserved.region[i].size; unsigned long size = memblock.reserved.region[i].size;
unsigned long start_pfn = physbase >> PAGE_SHIFT; unsigned long start_pfn = physbase >> PAGE_SHIFT;
unsigned long end_pfn = PFN_UP(physbase + size); unsigned long end_pfn = PFN_UP(physbase + size);
struct node_active_region node_ar; struct node_active_region node_ar;
@ -903,7 +903,7 @@ static void mark_reserved_regions_for_nid(int nid)
node->node_spanned_pages; node->node_spanned_pages;
/* /*
* Check to make sure that this lmb.reserved area is * Check to make sure that this memblock.reserved area is
* within the bounds of the node that we care about. * within the bounds of the node that we care about.
* Checking the nid of the start and end points is not * Checking the nid of the start and end points is not
* sufficient because the reserved area could span the * sufficient because the reserved area could span the
@ -961,7 +961,7 @@ void __init do_init_bootmem(void)
int nid; int nid;
min_low_pfn = 0; min_low_pfn = 0;
max_low_pfn = lmb_end_of_DRAM() >> PAGE_SHIFT; max_low_pfn = memblock_end_of_DRAM() >> PAGE_SHIFT;
max_pfn = max_low_pfn; max_pfn = max_low_pfn;
if (parse_numa_properties()) if (parse_numa_properties())
@ -1038,7 +1038,7 @@ void __init paging_init(void)
{ {
unsigned long max_zone_pfns[MAX_NR_ZONES]; unsigned long max_zone_pfns[MAX_NR_ZONES];
memset(max_zone_pfns, 0, sizeof(max_zone_pfns)); memset(max_zone_pfns, 0, sizeof(max_zone_pfns));
max_zone_pfns[ZONE_DMA] = lmb_end_of_DRAM() >> PAGE_SHIFT; max_zone_pfns[ZONE_DMA] = memblock_end_of_DRAM() >> PAGE_SHIFT;
free_area_init_nodes(max_zone_pfns); free_area_init_nodes(max_zone_pfns);
} }
@ -1072,7 +1072,7 @@ static int hot_add_drconf_scn_to_nid(struct device_node *memory,
{ {
const u32 *dm; const u32 *dm;
unsigned int drconf_cell_cnt, rc; unsigned int drconf_cell_cnt, rc;
unsigned long lmb_size; unsigned long memblock_size;
struct assoc_arrays aa; struct assoc_arrays aa;
int nid = -1; int nid = -1;
@ -1080,8 +1080,8 @@ static int hot_add_drconf_scn_to_nid(struct device_node *memory,
if (!drconf_cell_cnt) if (!drconf_cell_cnt)
return -1; return -1;
lmb_size = of_get_lmb_size(memory); memblock_size = of_get_memblock_size(memory);
if (!lmb_size) if (!memblock_size)
return -1; return -1;
rc = of_get_assoc_arrays(memory, &aa); rc = of_get_assoc_arrays(memory, &aa);
@ -1100,7 +1100,7 @@ static int hot_add_drconf_scn_to_nid(struct device_node *memory,
continue; continue;
if ((scn_addr < drmem.base_addr) if ((scn_addr < drmem.base_addr)
|| (scn_addr >= (drmem.base_addr + lmb_size))) || (scn_addr >= (drmem.base_addr + memblock_size)))
continue; continue;
nid = of_drconf_to_nid_single(&drmem, &aa); nid = of_drconf_to_nid_single(&drmem, &aa);
@ -1113,7 +1113,7 @@ static int hot_add_drconf_scn_to_nid(struct device_node *memory,
/* /*
* Find the node associated with a hot added memory section for memory * Find the node associated with a hot added memory section for memory
* represented in the device tree as a node (i.e. memory@XXXX) for * represented in the device tree as a node (i.e. memory@XXXX) for
* each lmb. * each memblock.
*/ */
int hot_add_node_scn_to_nid(unsigned long scn_addr) int hot_add_node_scn_to_nid(unsigned long scn_addr)
{ {
@ -1154,8 +1154,8 @@ int hot_add_node_scn_to_nid(unsigned long scn_addr)
/* /*
* Find the node associated with a hot added memory section. Section * Find the node associated with a hot added memory section. Section
* corresponds to a SPARSEMEM section, not an LMB. It is assumed that * corresponds to a SPARSEMEM section, not an MEMBLOCK. It is assumed that
* sections are fully contained within a single LMB. * sections are fully contained within a single MEMBLOCK.
*/ */
int hot_add_scn_to_nid(unsigned long scn_addr) int hot_add_scn_to_nid(unsigned long scn_addr)
{ {

View File

@ -26,7 +26,7 @@
#include <linux/vmalloc.h> #include <linux/vmalloc.h>
#include <linux/init.h> #include <linux/init.h>
#include <linux/highmem.h> #include <linux/highmem.h>
#include <linux/lmb.h> #include <linux/memblock.h>
#include <linux/slab.h> #include <linux/slab.h>
#include <asm/pgtable.h> #include <asm/pgtable.h>
@ -198,7 +198,7 @@ __ioremap_caller(phys_addr_t addr, unsigned long size, unsigned long flags,
* mem_init() sets high_memory so only do the check after that. * mem_init() sets high_memory so only do the check after that.
*/ */
if (mem_init_done && (p < virt_to_phys(high_memory)) && if (mem_init_done && (p < virt_to_phys(high_memory)) &&
!(__allow_ioremap_reserved && lmb_is_region_reserved(p, size))) { !(__allow_ioremap_reserved && memblock_is_region_reserved(p, size))) {
printk("__ioremap(): phys addr 0x%llx is RAM lr %p\n", printk("__ioremap(): phys addr 0x%llx is RAM lr %p\n",
(unsigned long long)p, __builtin_return_address(0)); (unsigned long long)p, __builtin_return_address(0));
return NULL; return NULL;
@ -331,7 +331,7 @@ void __init mapin_ram(void)
s = mmu_mapin_ram(top); s = mmu_mapin_ram(top);
__mapin_ram_chunk(s, top); __mapin_ram_chunk(s, top);
top = lmb_end_of_DRAM(); top = memblock_end_of_DRAM();
s = wii_mmu_mapin_mem2(top); s = wii_mmu_mapin_mem2(top);
__mapin_ram_chunk(s, top); __mapin_ram_chunk(s, top);
} }

View File

@ -34,7 +34,7 @@
#include <linux/vmalloc.h> #include <linux/vmalloc.h>
#include <linux/init.h> #include <linux/init.h>
#include <linux/bootmem.h> #include <linux/bootmem.h>
#include <linux/lmb.h> #include <linux/memblock.h>
#include <linux/slab.h> #include <linux/slab.h>
#include <asm/pgalloc.h> #include <asm/pgalloc.h>
@ -67,7 +67,7 @@ static void *early_alloc_pgtable(unsigned long size)
if (init_bootmem_done) if (init_bootmem_done)
pt = __alloc_bootmem(size, size, __pa(MAX_DMA_ADDRESS)); pt = __alloc_bootmem(size, size, __pa(MAX_DMA_ADDRESS));
else else
pt = __va(lmb_alloc_base(size, size, pt = __va(memblock_alloc_base(size, size,
__pa(MAX_DMA_ADDRESS))); __pa(MAX_DMA_ADDRESS)));
memset(pt, 0, size); memset(pt, 0, size);

View File

@ -26,7 +26,7 @@
#include <linux/mm.h> #include <linux/mm.h>
#include <linux/init.h> #include <linux/init.h>
#include <linux/highmem.h> #include <linux/highmem.h>
#include <linux/lmb.h> #include <linux/memblock.h>
#include <asm/prom.h> #include <asm/prom.h>
#include <asm/mmu.h> #include <asm/mmu.h>
@ -223,7 +223,7 @@ void __init MMU_init_hw(void)
* Find some memory for the hash table. * Find some memory for the hash table.
*/ */
if ( ppc_md.progress ) ppc_md.progress("hash:find piece", 0x322); if ( ppc_md.progress ) ppc_md.progress("hash:find piece", 0x322);
Hash = __va(lmb_alloc_base(Hash_size, Hash_size, Hash = __va(memblock_alloc_base(Hash_size, Hash_size,
__initial_memory_limit_addr)); __initial_memory_limit_addr));
cacheable_memzero(Hash, Hash_size); cacheable_memzero(Hash, Hash_size);
_SDR1 = __pa(Hash) | SDR1_LOW_BITS; _SDR1 = __pa(Hash) | SDR1_LOW_BITS;

View File

@ -12,7 +12,7 @@
* 2 of the License, or (at your option) any later version. * 2 of the License, or (at your option) any later version.
*/ */
#include <linux/lmb.h> #include <linux/memblock.h>
#include <asm/pgtable.h> #include <asm/pgtable.h>
#include <asm/mmu.h> #include <asm/mmu.h>
@ -252,7 +252,7 @@ void __init stabs_alloc(void)
if (cpu == 0) if (cpu == 0)
continue; /* stab for CPU 0 is statically allocated */ continue; /* stab for CPU 0 is statically allocated */
newstab = lmb_alloc_base(HW_PAGE_SIZE, HW_PAGE_SIZE, newstab = memblock_alloc_base(HW_PAGE_SIZE, HW_PAGE_SIZE,
1<<SID_SHIFT); 1<<SID_SHIFT);
newstab = (unsigned long)__va(newstab); newstab = (unsigned long)__va(newstab);

View File

@ -34,7 +34,7 @@
#include <linux/pagemap.h> #include <linux/pagemap.h>
#include <linux/preempt.h> #include <linux/preempt.h>
#include <linux/spinlock.h> #include <linux/spinlock.h>
#include <linux/lmb.h> #include <linux/memblock.h>
#include <asm/tlbflush.h> #include <asm/tlbflush.h>
#include <asm/tlb.h> #include <asm/tlb.h>
@ -426,7 +426,7 @@ static void __early_init_mmu(int boot_cpu)
/* Set the global containing the top of the linear mapping /* Set the global containing the top of the linear mapping
* for use by the TLB miss code * for use by the TLB miss code
*/ */
linear_map_top = lmb_end_of_DRAM(); linear_map_top = memblock_end_of_DRAM();
/* A sync won't hurt us after mucking around with /* A sync won't hurt us after mucking around with
* the MMU configuration * the MMU configuration

View File

@ -16,7 +16,7 @@
#include <linux/kdev_t.h> #include <linux/kdev_t.h>
#include <linux/delay.h> #include <linux/delay.h>
#include <linux/interrupt.h> #include <linux/interrupt.h>
#include <linux/lmb.h> #include <linux/memblock.h>
#include <asm/system.h> #include <asm/system.h>
#include <asm/time.h> #include <asm/time.h>
@ -100,7 +100,7 @@ void __init corenet_ds_setup_arch(void)
#endif #endif
#ifdef CONFIG_SWIOTLB #ifdef CONFIG_SWIOTLB
if (lmb_end_of_DRAM() > max) { if (memblock_end_of_DRAM() > max) {
ppc_swiotlb_enable = 1; ppc_swiotlb_enable = 1;
set_pci_dma_ops(&swiotlb_dma_ops); set_pci_dma_ops(&swiotlb_dma_ops);
ppc_md.pci_dma_dev_setup = pci_dma_dev_setup_swiotlb; ppc_md.pci_dma_dev_setup = pci_dma_dev_setup_swiotlb;

View File

@ -17,7 +17,7 @@
#include <linux/seq_file.h> #include <linux/seq_file.h>
#include <linux/interrupt.h> #include <linux/interrupt.h>
#include <linux/of_platform.h> #include <linux/of_platform.h>
#include <linux/lmb.h> #include <linux/memblock.h>
#include <asm/system.h> #include <asm/system.h>
#include <asm/time.h> #include <asm/time.h>
@ -94,7 +94,7 @@ static void __init mpc8536_ds_setup_arch(void)
#endif #endif
#ifdef CONFIG_SWIOTLB #ifdef CONFIG_SWIOTLB
if (lmb_end_of_DRAM() > max) { if (memblock_end_of_DRAM() > max) {
ppc_swiotlb_enable = 1; ppc_swiotlb_enable = 1;
set_pci_dma_ops(&swiotlb_dma_ops); set_pci_dma_ops(&swiotlb_dma_ops);
ppc_md.pci_dma_dev_setup = pci_dma_dev_setup_swiotlb; ppc_md.pci_dma_dev_setup = pci_dma_dev_setup_swiotlb;

View File

@ -20,7 +20,7 @@
#include <linux/seq_file.h> #include <linux/seq_file.h>
#include <linux/interrupt.h> #include <linux/interrupt.h>
#include <linux/of_platform.h> #include <linux/of_platform.h>
#include <linux/lmb.h> #include <linux/memblock.h>
#include <asm/system.h> #include <asm/system.h>
#include <asm/time.h> #include <asm/time.h>
@ -190,7 +190,7 @@ static void __init mpc85xx_ds_setup_arch(void)
#endif #endif
#ifdef CONFIG_SWIOTLB #ifdef CONFIG_SWIOTLB
if (lmb_end_of_DRAM() > max) { if (memblock_end_of_DRAM() > max) {
ppc_swiotlb_enable = 1; ppc_swiotlb_enable = 1;
set_pci_dma_ops(&swiotlb_dma_ops); set_pci_dma_ops(&swiotlb_dma_ops);
ppc_md.pci_dma_dev_setup = pci_dma_dev_setup_swiotlb; ppc_md.pci_dma_dev_setup = pci_dma_dev_setup_swiotlb;

View File

@ -33,7 +33,7 @@
#include <linux/of_platform.h> #include <linux/of_platform.h>
#include <linux/of_device.h> #include <linux/of_device.h>
#include <linux/phy.h> #include <linux/phy.h>
#include <linux/lmb.h> #include <linux/memblock.h>
#include <asm/system.h> #include <asm/system.h>
#include <asm/atomic.h> #include <asm/atomic.h>
@ -325,7 +325,7 @@ static void __init mpc85xx_mds_setup_arch(void)
#endif /* CONFIG_QUICC_ENGINE */ #endif /* CONFIG_QUICC_ENGINE */
#ifdef CONFIG_SWIOTLB #ifdef CONFIG_SWIOTLB
if (lmb_end_of_DRAM() > max) { if (memblock_end_of_DRAM() > max) {
ppc_swiotlb_enable = 1; ppc_swiotlb_enable = 1;
set_pci_dma_ops(&swiotlb_dma_ops); set_pci_dma_ops(&swiotlb_dma_ops);
ppc_md.pci_dma_dev_setup = pci_dma_dev_setup_swiotlb; ppc_md.pci_dma_dev_setup = pci_dma_dev_setup_swiotlb;

View File

@ -19,7 +19,7 @@
#include <linux/delay.h> #include <linux/delay.h>
#include <linux/seq_file.h> #include <linux/seq_file.h>
#include <linux/of_platform.h> #include <linux/of_platform.h>
#include <linux/lmb.h> #include <linux/memblock.h>
#include <asm/system.h> #include <asm/system.h>
#include <asm/time.h> #include <asm/time.h>
@ -103,7 +103,7 @@ mpc86xx_hpcn_setup_arch(void)
#endif #endif
#ifdef CONFIG_SWIOTLB #ifdef CONFIG_SWIOTLB
if (lmb_end_of_DRAM() > max) { if (memblock_end_of_DRAM() > max) {
ppc_swiotlb_enable = 1; ppc_swiotlb_enable = 1;
set_pci_dma_ops(&swiotlb_dma_ops); set_pci_dma_ops(&swiotlb_dma_ops);
ppc_md.pci_dma_dev_setup = pci_dma_dev_setup_swiotlb; ppc_md.pci_dma_dev_setup = pci_dma_dev_setup_swiotlb;

View File

@ -29,7 +29,7 @@
#include <linux/of.h> #include <linux/of.h>
#include <linux/of_platform.h> #include <linux/of_platform.h>
#include <linux/slab.h> #include <linux/slab.h>
#include <linux/lmb.h> #include <linux/memblock.h>
#include <asm/prom.h> #include <asm/prom.h>
#include <asm/iommu.h> #include <asm/iommu.h>
@ -845,10 +845,10 @@ static int __init cell_iommu_init_disabled(void)
/* If we found a DMA window, we check if it's big enough to enclose /* If we found a DMA window, we check if it's big enough to enclose
* all of physical memory. If not, we force enable IOMMU * all of physical memory. If not, we force enable IOMMU
*/ */
if (np && size < lmb_end_of_DRAM()) { if (np && size < memblock_end_of_DRAM()) {
printk(KERN_WARNING "iommu: force-enabled, dma window" printk(KERN_WARNING "iommu: force-enabled, dma window"
" (%ldMB) smaller than total memory (%lldMB)\n", " (%ldMB) smaller than total memory (%lldMB)\n",
size >> 20, lmb_end_of_DRAM() >> 20); size >> 20, memblock_end_of_DRAM() >> 20);
return -ENODEV; return -ENODEV;
} }
@ -1064,7 +1064,7 @@ static int __init cell_iommu_fixed_mapping_init(void)
} }
fbase = _ALIGN_UP(fbase, 1 << IO_SEGMENT_SHIFT); fbase = _ALIGN_UP(fbase, 1 << IO_SEGMENT_SHIFT);
fsize = lmb_phys_mem_size(); fsize = memblock_phys_mem_size();
if ((fbase + fsize) <= 0x800000000ul) if ((fbase + fsize) <= 0x800000000ul)
hbase = 0; /* use the device tree window */ hbase = 0; /* use the device tree window */
@ -1169,7 +1169,7 @@ static int __init cell_iommu_init(void)
* Note: should we make sure we have the IOMMU actually disabled ? * Note: should we make sure we have the IOMMU actually disabled ?
*/ */
if (iommu_is_off || if (iommu_is_off ||
(!iommu_force_on && lmb_end_of_DRAM() <= 0x80000000ull)) (!iommu_force_on && memblock_end_of_DRAM() <= 0x80000000ull))
if (cell_iommu_init_disabled() == 0) if (cell_iommu_init_disabled() == 0)
goto bail; goto bail;

View File

@ -20,7 +20,7 @@
#include <linux/seq_file.h> #include <linux/seq_file.h>
#include <linux/kexec.h> #include <linux/kexec.h>
#include <linux/of_platform.h> #include <linux/of_platform.h>
#include <linux/lmb.h> #include <linux/memblock.h>
#include <mm/mmu_decl.h> #include <mm/mmu_decl.h>
#include <asm/io.h> #include <asm/io.h>
@ -65,7 +65,7 @@ static int __init page_aligned(unsigned long x)
void __init wii_memory_fixups(void) void __init wii_memory_fixups(void)
{ {
struct lmb_property *p = lmb.memory.region; struct memblock_property *p = memblock.memory.region;
/* /*
* This is part of a workaround to allow the use of two * This is part of a workaround to allow the use of two
@ -77,7 +77,7 @@ void __init wii_memory_fixups(void)
* between both ranges. * between both ranges.
*/ */
BUG_ON(lmb.memory.cnt != 2); BUG_ON(memblock.memory.cnt != 2);
BUG_ON(!page_aligned(p[0].base) || !page_aligned(p[1].base)); BUG_ON(!page_aligned(p[0].base) || !page_aligned(p[1].base));
p[0].size = _ALIGN_DOWN(p[0].size, PAGE_SIZE); p[0].size = _ALIGN_DOWN(p[0].size, PAGE_SIZE);
@ -92,11 +92,11 @@ void __init wii_memory_fixups(void)
p[0].size += wii_hole_size + p[1].size; p[0].size += wii_hole_size + p[1].size;
lmb.memory.cnt = 1; memblock.memory.cnt = 1;
lmb_analyze(); memblock_analyze();
/* reserve the hole */ /* reserve the hole */
lmb_reserve(wii_hole_start, wii_hole_size); memblock_reserve(wii_hole_start, wii_hole_size);
/* allow ioremapping the address space in the hole */ /* allow ioremapping the address space in the hole */
__allow_ioremap_reserved = 1; __allow_ioremap_reserved = 1;

View File

@ -41,7 +41,7 @@
#include <linux/smp.h> #include <linux/smp.h>
#include <linux/bitops.h> #include <linux/bitops.h>
#include <linux/of_device.h> #include <linux/of_device.h>
#include <linux/lmb.h> #include <linux/memblock.h>
#include <asm/processor.h> #include <asm/processor.h>
#include <asm/sections.h> #include <asm/sections.h>

View File

@ -204,7 +204,7 @@ int __init iob_init(struct device_node *dn)
pr_debug(" -> %s\n", __func__); pr_debug(" -> %s\n", __func__);
/* Allocate a spare page to map all invalid IOTLB pages. */ /* Allocate a spare page to map all invalid IOTLB pages. */
tmp = lmb_alloc(IOBMAP_PAGE_SIZE, IOBMAP_PAGE_SIZE); tmp = memblock_alloc(IOBMAP_PAGE_SIZE, IOBMAP_PAGE_SIZE);
if (!tmp) if (!tmp)
panic("IOBMAP: Cannot allocate spare page!"); panic("IOBMAP: Cannot allocate spare page!");
/* Empty l1 is marked invalid */ /* Empty l1 is marked invalid */
@ -275,7 +275,7 @@ void __init alloc_iobmap_l2(void)
return; return;
#endif #endif
/* For 2G space, 8x64 pages (2^21 bytes) is max total l2 size */ /* For 2G space, 8x64 pages (2^21 bytes) is max total l2 size */
iob_l2_base = (u32 *)abs_to_virt(lmb_alloc_base(1UL<<21, 1UL<<21, 0x80000000)); iob_l2_base = (u32 *)abs_to_virt(memblock_alloc_base(1UL<<21, 1UL<<21, 0x80000000));
printk(KERN_INFO "IOBMAP L2 allocated at: %p\n", iob_l2_base); printk(KERN_INFO "IOBMAP L2 allocated at: %p\n", iob_l2_base);
} }

View File

@ -51,7 +51,7 @@
#include <linux/suspend.h> #include <linux/suspend.h>
#include <linux/of_device.h> #include <linux/of_device.h>
#include <linux/of_platform.h> #include <linux/of_platform.h>
#include <linux/lmb.h> #include <linux/memblock.h>
#include <asm/reg.h> #include <asm/reg.h>
#include <asm/sections.h> #include <asm/sections.h>
@ -619,7 +619,7 @@ static int __init pmac_probe(void)
* driver needs that. We have to allocate it now. We allocate 4k * driver needs that. We have to allocate it now. We allocate 4k
* (1 small page) for now. * (1 small page) for now.
*/ */
smu_cmdbuf_abs = lmb_alloc_base(4096, 4096, 0x80000000UL); smu_cmdbuf_abs = memblock_alloc_base(4096, 4096, 0x80000000UL);
#endif /* CONFIG_PMAC_SMU */ #endif /* CONFIG_PMAC_SMU */
return 1; return 1;

View File

@ -19,7 +19,7 @@
*/ */
#include <linux/kernel.h> #include <linux/kernel.h>
#include <linux/lmb.h> #include <linux/memblock.h>
#include <asm/machdep.h> #include <asm/machdep.h>
#include <asm/prom.h> #include <asm/prom.h>

View File

@ -21,7 +21,7 @@
#include <linux/kernel.h> #include <linux/kernel.h>
#include <linux/module.h> #include <linux/module.h>
#include <linux/memory_hotplug.h> #include <linux/memory_hotplug.h>
#include <linux/lmb.h> #include <linux/memblock.h>
#include <linux/slab.h> #include <linux/slab.h>
#include <asm/cell-regs.h> #include <asm/cell-regs.h>
@ -318,8 +318,8 @@ static int __init ps3_mm_add_memory(void)
return result; return result;
} }
lmb_add(start_addr, map.r1.size); memblock_add(start_addr, map.r1.size);
lmb_analyze(); memblock_analyze();
result = online_pages(start_pfn, nr_pages); result = online_pages(start_pfn, nr_pages);

View File

@ -24,7 +24,7 @@
#include <linux/fs.h> #include <linux/fs.h>
#include <linux/syscalls.h> #include <linux/syscalls.h>
#include <linux/ctype.h> #include <linux/ctype.h>
#include <linux/lmb.h> #include <linux/memblock.h>
#include <linux/of.h> #include <linux/of.h>
#include <linux/slab.h> #include <linux/slab.h>
@ -723,7 +723,7 @@ static void os_area_queue_work(void)
* flash to a high address in the boot memory region and then puts that RAM * flash to a high address in the boot memory region and then puts that RAM
* address and the byte count into the repository for retrieval by the guest. * address and the byte count into the repository for retrieval by the guest.
* We copy the data we want into a static variable and allow the memory setup * We copy the data we want into a static variable and allow the memory setup
* by the HV to be claimed by the lmb manager. * by the HV to be claimed by the memblock manager.
* *
* The os area mirror will not be available to a second stage kernel, and * The os area mirror will not be available to a second stage kernel, and
* the header verify will fail. In this case, the saved_params values will * the header verify will fail. In this case, the saved_params values will

View File

@ -10,14 +10,14 @@
*/ */
#include <linux/of.h> #include <linux/of.h>
#include <linux/lmb.h> #include <linux/memblock.h>
#include <linux/vmalloc.h> #include <linux/vmalloc.h>
#include <asm/firmware.h> #include <asm/firmware.h>
#include <asm/machdep.h> #include <asm/machdep.h>
#include <asm/pSeries_reconfig.h> #include <asm/pSeries_reconfig.h>
#include <asm/sparsemem.h> #include <asm/sparsemem.h>
static int pseries_remove_lmb(unsigned long base, unsigned int lmb_size) static int pseries_remove_memblock(unsigned long base, unsigned int memblock_size)
{ {
unsigned long start, start_pfn; unsigned long start, start_pfn;
struct zone *zone; struct zone *zone;
@ -26,7 +26,7 @@ static int pseries_remove_lmb(unsigned long base, unsigned int lmb_size)
start_pfn = base >> PAGE_SHIFT; start_pfn = base >> PAGE_SHIFT;
if (!pfn_valid(start_pfn)) { if (!pfn_valid(start_pfn)) {
lmb_remove(base, lmb_size); memblock_remove(base, memblock_size);
return 0; return 0;
} }
@ -41,20 +41,20 @@ static int pseries_remove_lmb(unsigned long base, unsigned int lmb_size)
* to sysfs "state" file and we can't remove sysfs entries * to sysfs "state" file and we can't remove sysfs entries
* while writing to it. So we have to defer it to here. * while writing to it. So we have to defer it to here.
*/ */
ret = __remove_pages(zone, start_pfn, lmb_size >> PAGE_SHIFT); ret = __remove_pages(zone, start_pfn, memblock_size >> PAGE_SHIFT);
if (ret) if (ret)
return ret; return ret;
/* /*
* Update memory regions for memory remove * Update memory regions for memory remove
*/ */
lmb_remove(base, lmb_size); memblock_remove(base, memblock_size);
/* /*
* Remove htab bolted mappings for this section of memory * Remove htab bolted mappings for this section of memory
*/ */
start = (unsigned long)__va(base); start = (unsigned long)__va(base);
ret = remove_section_mapping(start, start + lmb_size); ret = remove_section_mapping(start, start + memblock_size);
/* Ensure all vmalloc mappings are flushed in case they also /* Ensure all vmalloc mappings are flushed in case they also
* hit that section of memory * hit that section of memory
@ -69,7 +69,7 @@ static int pseries_remove_memory(struct device_node *np)
const char *type; const char *type;
const unsigned int *regs; const unsigned int *regs;
unsigned long base; unsigned long base;
unsigned int lmb_size; unsigned int memblock_size;
int ret = -EINVAL; int ret = -EINVAL;
/* /*
@ -80,16 +80,16 @@ static int pseries_remove_memory(struct device_node *np)
return 0; return 0;
/* /*
* Find the bae address and size of the lmb * Find the bae address and size of the memblock
*/ */
regs = of_get_property(np, "reg", NULL); regs = of_get_property(np, "reg", NULL);
if (!regs) if (!regs)
return ret; return ret;
base = *(unsigned long *)regs; base = *(unsigned long *)regs;
lmb_size = regs[3]; memblock_size = regs[3];
ret = pseries_remove_lmb(base, lmb_size); ret = pseries_remove_memblock(base, memblock_size);
return ret; return ret;
} }
@ -98,7 +98,7 @@ static int pseries_add_memory(struct device_node *np)
const char *type; const char *type;
const unsigned int *regs; const unsigned int *regs;
unsigned long base; unsigned long base;
unsigned int lmb_size; unsigned int memblock_size;
int ret = -EINVAL; int ret = -EINVAL;
/* /*
@ -109,43 +109,43 @@ static int pseries_add_memory(struct device_node *np)
return 0; return 0;
/* /*
* Find the base and size of the lmb * Find the base and size of the memblock
*/ */
regs = of_get_property(np, "reg", NULL); regs = of_get_property(np, "reg", NULL);
if (!regs) if (!regs)
return ret; return ret;
base = *(unsigned long *)regs; base = *(unsigned long *)regs;
lmb_size = regs[3]; memblock_size = regs[3];
/* /*
* Update memory region to represent the memory add * Update memory region to represent the memory add
*/ */
ret = lmb_add(base, lmb_size); ret = memblock_add(base, memblock_size);
return (ret < 0) ? -EINVAL : 0; return (ret < 0) ? -EINVAL : 0;
} }
static int pseries_drconf_memory(unsigned long *base, unsigned int action) static int pseries_drconf_memory(unsigned long *base, unsigned int action)
{ {
struct device_node *np; struct device_node *np;
const unsigned long *lmb_size; const unsigned long *memblock_size;
int rc; int rc;
np = of_find_node_by_path("/ibm,dynamic-reconfiguration-memory"); np = of_find_node_by_path("/ibm,dynamic-reconfiguration-memory");
if (!np) if (!np)
return -EINVAL; return -EINVAL;
lmb_size = of_get_property(np, "ibm,lmb-size", NULL); memblock_size = of_get_property(np, "ibm,memblock-size", NULL);
if (!lmb_size) { if (!memblock_size) {
of_node_put(np); of_node_put(np);
return -EINVAL; return -EINVAL;
} }
if (action == PSERIES_DRCONF_MEM_ADD) { if (action == PSERIES_DRCONF_MEM_ADD) {
rc = lmb_add(*base, *lmb_size); rc = memblock_add(*base, *memblock_size);
rc = (rc < 0) ? -EINVAL : 0; rc = (rc < 0) ? -EINVAL : 0;
} else if (action == PSERIES_DRCONF_MEM_REMOVE) { } else if (action == PSERIES_DRCONF_MEM_REMOVE) {
rc = pseries_remove_lmb(*base, *lmb_size); rc = pseries_remove_memblock(*base, *memblock_size);
} else { } else {
rc = -EINVAL; rc = -EINVAL;
} }

View File

@ -66,7 +66,7 @@ static int tce_build_pSeries(struct iommu_table *tbl, long index,
tcep = ((u64 *)tbl->it_base) + index; tcep = ((u64 *)tbl->it_base) + index;
while (npages--) { while (npages--) {
/* can't move this out since we might cross LMB boundary */ /* can't move this out since we might cross MEMBLOCK boundary */
rpn = (virt_to_abs(uaddr)) >> TCE_SHIFT; rpn = (virt_to_abs(uaddr)) >> TCE_SHIFT;
*tcep = proto_tce | (rpn & TCE_RPN_MASK) << TCE_RPN_SHIFT; *tcep = proto_tce | (rpn & TCE_RPN_MASK) << TCE_RPN_SHIFT;

View File

@ -255,12 +255,12 @@ void invalidate_last_dump(struct phyp_dump_header *ph, unsigned long addr)
/* ------------------------------------------------- */ /* ------------------------------------------------- */
/** /**
* release_memory_range -- release memory previously lmb_reserved * release_memory_range -- release memory previously memblock_reserved
* @start_pfn: starting physical frame number * @start_pfn: starting physical frame number
* @nr_pages: number of pages to free. * @nr_pages: number of pages to free.
* *
* This routine will release memory that had been previously * This routine will release memory that had been previously
* lmb_reserved in early boot. The released memory becomes * memblock_reserved in early boot. The released memory becomes
* available for genreal use. * available for genreal use.
*/ */
static void release_memory_range(unsigned long start_pfn, static void release_memory_range(unsigned long start_pfn,

View File

@ -36,7 +36,7 @@
#include <linux/dma-mapping.h> #include <linux/dma-mapping.h>
#include <linux/vmalloc.h> #include <linux/vmalloc.h>
#include <linux/suspend.h> #include <linux/suspend.h>
#include <linux/lmb.h> #include <linux/memblock.h>
#include <linux/gfp.h> #include <linux/gfp.h>
#include <asm/io.h> #include <asm/io.h>
#include <asm/prom.h> #include <asm/prom.h>
@ -232,7 +232,7 @@ static int __init dart_init(struct device_node *dart_node)
* that to work around what looks like a problem with the HT bridge * that to work around what looks like a problem with the HT bridge
* prefetching into invalid pages and corrupting data * prefetching into invalid pages and corrupting data
*/ */
tmp = lmb_alloc(DART_PAGE_SIZE, DART_PAGE_SIZE); tmp = memblock_alloc(DART_PAGE_SIZE, DART_PAGE_SIZE);
dart_emptyval = DARTMAP_VALID | ((tmp >> DART_PAGE_SHIFT) & dart_emptyval = DARTMAP_VALID | ((tmp >> DART_PAGE_SHIFT) &
DARTMAP_RPNMASK); DARTMAP_RPNMASK);
@ -407,7 +407,7 @@ void __init alloc_dart_table(void)
if (iommu_is_off) if (iommu_is_off)
return; return;
if (!iommu_force_on && lmb_end_of_DRAM() <= 0x40000000ull) if (!iommu_force_on && memblock_end_of_DRAM() <= 0x40000000ull)
return; return;
/* 512 pages (2MB) is max DART tablesize. */ /* 512 pages (2MB) is max DART tablesize. */
@ -416,7 +416,7 @@ void __init alloc_dart_table(void)
* will blow up an entire large page anyway in the kernel mapping * will blow up an entire large page anyway in the kernel mapping
*/ */
dart_tablebase = (unsigned long) dart_tablebase = (unsigned long)
abs_to_virt(lmb_alloc_base(1UL<<24, 1UL<<24, 0x80000000L)); abs_to_virt(memblock_alloc_base(1UL<<24, 1UL<<24, 0x80000000L));
printk(KERN_INFO "DART table allocated at: %lx\n", dart_tablebase); printk(KERN_INFO "DART table allocated at: %lx\n", dart_tablebase);
} }

View File

@ -23,7 +23,7 @@
#include <linux/string.h> #include <linux/string.h>
#include <linux/init.h> #include <linux/init.h>
#include <linux/bootmem.h> #include <linux/bootmem.h>
#include <linux/lmb.h> #include <linux/memblock.h>
#include <linux/log2.h> #include <linux/log2.h>
#include <linux/slab.h> #include <linux/slab.h>
@ -190,7 +190,7 @@ static void __init setup_pci_atmu(struct pci_controller *hose,
pr_info("%s: PCICSRBAR @ 0x%x\n", name, pcicsrbar); pr_info("%s: PCICSRBAR @ 0x%x\n", name, pcicsrbar);
/* Setup inbound mem window */ /* Setup inbound mem window */
mem = lmb_end_of_DRAM(); mem = memblock_end_of_DRAM();
sz = min(mem, paddr_lo); sz = min(mem, paddr_lo);
mem_log = __ilog2_u64(sz); mem_log = __ilog2_u64(sz);

View File

@ -10,7 +10,7 @@ config SUPERH
select EMBEDDED select EMBEDDED
select HAVE_CLK select HAVE_CLK
select HAVE_IDE if HAS_IOPORT select HAVE_IDE if HAS_IOPORT
select HAVE_LMB select HAVE_MEMBLOCK
select HAVE_OPROFILE select HAVE_OPROFILE
select HAVE_GENERIC_DMA_COHERENT select HAVE_GENERIC_DMA_COHERENT
select HAVE_ARCH_TRACEHOOK select HAVE_ARCH_TRACEHOOK

View File

@ -1,6 +0,0 @@
#ifndef __ASM_SH_LMB_H
#define __ASM_SH_LMB_H
#define LMB_REAL_LIMIT 0
#endif /* __ASM_SH_LMB_H */

View File

@ -0,0 +1,6 @@
#ifndef __ASM_SH_MEMBLOCK_H
#define __ASM_SH_MEMBLOCK_H
#define MEMBLOCK_REAL_LIMIT 0
#endif /* __ASM_SH_MEMBLOCK_H */

View File

@ -15,7 +15,7 @@
#include <linux/numa.h> #include <linux/numa.h>
#include <linux/ftrace.h> #include <linux/ftrace.h>
#include <linux/suspend.h> #include <linux/suspend.h>
#include <linux/lmb.h> #include <linux/memblock.h>
#include <asm/pgtable.h> #include <asm/pgtable.h>
#include <asm/pgalloc.h> #include <asm/pgalloc.h>
#include <asm/mmu_context.h> #include <asm/mmu_context.h>
@ -157,10 +157,10 @@ void __init reserve_crashkernel(void)
unsigned long long crash_size, crash_base; unsigned long long crash_size, crash_base;
int ret; int ret;
/* this is necessary because of lmb_phys_mem_size() */ /* this is necessary because of memblock_phys_mem_size() */
lmb_analyze(); memblock_analyze();
ret = parse_crashkernel(boot_command_line, lmb_phys_mem_size(), ret = parse_crashkernel(boot_command_line, memblock_phys_mem_size(),
&crash_size, &crash_base); &crash_size, &crash_base);
if (ret == 0 && crash_size > 0) { if (ret == 0 && crash_size > 0) {
crashk_res.start = crash_base; crashk_res.start = crash_base;
@ -172,14 +172,14 @@ void __init reserve_crashkernel(void)
crash_size = PAGE_ALIGN(crashk_res.end - crashk_res.start + 1); crash_size = PAGE_ALIGN(crashk_res.end - crashk_res.start + 1);
if (!crashk_res.start) { if (!crashk_res.start) {
unsigned long max = lmb_end_of_DRAM() - memory_limit; unsigned long max = memblock_end_of_DRAM() - memory_limit;
crashk_res.start = __lmb_alloc_base(crash_size, PAGE_SIZE, max); crashk_res.start = __memblock_alloc_base(crash_size, PAGE_SIZE, max);
if (!crashk_res.start) { if (!crashk_res.start) {
pr_err("crashkernel allocation failed\n"); pr_err("crashkernel allocation failed\n");
goto disable; goto disable;
} }
} else { } else {
ret = lmb_reserve(crashk_res.start, crash_size); ret = memblock_reserve(crashk_res.start, crash_size);
if (unlikely(ret < 0)) { if (unlikely(ret < 0)) {
pr_err("crashkernel reservation failed - " pr_err("crashkernel reservation failed - "
"memory is in use\n"); "memory is in use\n");
@ -192,7 +192,7 @@ void __init reserve_crashkernel(void)
/* /*
* Crash kernel trumps memory limit * Crash kernel trumps memory limit
*/ */
if ((lmb_end_of_DRAM() - memory_limit) <= crashk_res.end) { if ((memblock_end_of_DRAM() - memory_limit) <= crashk_res.end) {
memory_limit = 0; memory_limit = 0;
pr_info("Disabled memory limit for crashkernel\n"); pr_info("Disabled memory limit for crashkernel\n");
} }
@ -201,7 +201,7 @@ void __init reserve_crashkernel(void)
"for crashkernel (System RAM: %ldMB)\n", "for crashkernel (System RAM: %ldMB)\n",
(unsigned long)(crash_size >> 20), (unsigned long)(crash_size >> 20),
(unsigned long)(crashk_res.start), (unsigned long)(crashk_res.start),
(unsigned long)(lmb_phys_mem_size() >> 20)); (unsigned long)(memblock_phys_mem_size() >> 20));
return; return;

View File

@ -30,7 +30,7 @@
#include <linux/clk.h> #include <linux/clk.h>
#include <linux/delay.h> #include <linux/delay.h>
#include <linux/platform_device.h> #include <linux/platform_device.h>
#include <linux/lmb.h> #include <linux/memblock.h>
#include <asm/uaccess.h> #include <asm/uaccess.h>
#include <asm/io.h> #include <asm/io.h>
#include <asm/page.h> #include <asm/page.h>
@ -141,10 +141,10 @@ void __init check_for_initrd(void)
goto disable; goto disable;
} }
if (unlikely(end > lmb_end_of_DRAM())) { if (unlikely(end > memblock_end_of_DRAM())) {
pr_err("initrd extends beyond end of memory " pr_err("initrd extends beyond end of memory "
"(0x%08lx > 0x%08lx)\ndisabling initrd\n", "(0x%08lx > 0x%08lx)\ndisabling initrd\n",
end, (unsigned long)lmb_end_of_DRAM()); end, (unsigned long)memblock_end_of_DRAM());
goto disable; goto disable;
} }
@ -161,7 +161,7 @@ void __init check_for_initrd(void)
initrd_start = (unsigned long)__va(__pa(start)); initrd_start = (unsigned long)__va(__pa(start));
initrd_end = initrd_start + INITRD_SIZE; initrd_end = initrd_start + INITRD_SIZE;
lmb_reserve(__pa(initrd_start), INITRD_SIZE); memblock_reserve(__pa(initrd_start), INITRD_SIZE);
return; return;

View File

@ -16,7 +16,7 @@
#include <linux/pagemap.h> #include <linux/pagemap.h>
#include <linux/percpu.h> #include <linux/percpu.h>
#include <linux/io.h> #include <linux/io.h>
#include <linux/lmb.h> #include <linux/memblock.h>
#include <linux/dma-mapping.h> #include <linux/dma-mapping.h>
#include <asm/mmu_context.h> #include <asm/mmu_context.h>
#include <asm/mmzone.h> #include <asm/mmzone.h>
@ -33,7 +33,7 @@ pgd_t swapper_pg_dir[PTRS_PER_PGD];
void __init generic_mem_init(void) void __init generic_mem_init(void)
{ {
lmb_add(__MEMORY_START, __MEMORY_SIZE); memblock_add(__MEMORY_START, __MEMORY_SIZE);
} }
void __init __weak plat_mem_setup(void) void __init __weak plat_mem_setup(void)
@ -176,12 +176,12 @@ void __init allocate_pgdat(unsigned int nid)
get_pfn_range_for_nid(nid, &start_pfn, &end_pfn); get_pfn_range_for_nid(nid, &start_pfn, &end_pfn);
#ifdef CONFIG_NEED_MULTIPLE_NODES #ifdef CONFIG_NEED_MULTIPLE_NODES
phys = __lmb_alloc_base(sizeof(struct pglist_data), phys = __memblock_alloc_base(sizeof(struct pglist_data),
SMP_CACHE_BYTES, end_pfn << PAGE_SHIFT); SMP_CACHE_BYTES, end_pfn << PAGE_SHIFT);
/* Retry with all of system memory */ /* Retry with all of system memory */
if (!phys) if (!phys)
phys = __lmb_alloc_base(sizeof(struct pglist_data), phys = __memblock_alloc_base(sizeof(struct pglist_data),
SMP_CACHE_BYTES, lmb_end_of_DRAM()); SMP_CACHE_BYTES, memblock_end_of_DRAM());
if (!phys) if (!phys)
panic("Can't allocate pgdat for node %d\n", nid); panic("Can't allocate pgdat for node %d\n", nid);
@ -212,7 +212,7 @@ static void __init bootmem_init_one_node(unsigned int nid)
total_pages = bootmem_bootmap_pages(p->node_spanned_pages); total_pages = bootmem_bootmap_pages(p->node_spanned_pages);
paddr = lmb_alloc(total_pages << PAGE_SHIFT, PAGE_SIZE); paddr = memblock_alloc(total_pages << PAGE_SHIFT, PAGE_SIZE);
if (!paddr) if (!paddr)
panic("Can't allocate bootmap for nid[%d]\n", nid); panic("Can't allocate bootmap for nid[%d]\n", nid);
@ -227,9 +227,9 @@ static void __init bootmem_init_one_node(unsigned int nid)
*/ */
if (nid == 0) { if (nid == 0) {
/* Reserve the sections we're already using. */ /* Reserve the sections we're already using. */
for (i = 0; i < lmb.reserved.cnt; i++) for (i = 0; i < memblock.reserved.cnt; i++)
reserve_bootmem(lmb.reserved.region[i].base, reserve_bootmem(memblock.reserved.region[i].base,
lmb_size_bytes(&lmb.reserved, i), memblock_size_bytes(&memblock.reserved, i),
BOOTMEM_DEFAULT); BOOTMEM_DEFAULT);
} }
@ -241,10 +241,10 @@ static void __init do_init_bootmem(void)
int i; int i;
/* Add active regions with valid PFNs. */ /* Add active regions with valid PFNs. */
for (i = 0; i < lmb.memory.cnt; i++) { for (i = 0; i < memblock.memory.cnt; i++) {
unsigned long start_pfn, end_pfn; unsigned long start_pfn, end_pfn;
start_pfn = lmb.memory.region[i].base >> PAGE_SHIFT; start_pfn = memblock.memory.region[i].base >> PAGE_SHIFT;
end_pfn = start_pfn + lmb_size_pages(&lmb.memory, i); end_pfn = start_pfn + memblock_size_pages(&memblock.memory, i);
__add_active_range(0, start_pfn, end_pfn); __add_active_range(0, start_pfn, end_pfn);
} }
@ -276,7 +276,7 @@ static void __init early_reserve_mem(void)
* this catches the (definitely buggy) case of us accidentally * this catches the (definitely buggy) case of us accidentally
* initializing the bootmem allocator with an invalid RAM area. * initializing the bootmem allocator with an invalid RAM area.
*/ */
lmb_reserve(__MEMORY_START + CONFIG_ZERO_PAGE_OFFSET, memblock_reserve(__MEMORY_START + CONFIG_ZERO_PAGE_OFFSET,
(PFN_PHYS(start_pfn) + PAGE_SIZE - 1) - (PFN_PHYS(start_pfn) + PAGE_SIZE - 1) -
(__MEMORY_START + CONFIG_ZERO_PAGE_OFFSET)); (__MEMORY_START + CONFIG_ZERO_PAGE_OFFSET));
@ -284,7 +284,7 @@ static void __init early_reserve_mem(void)
* Reserve physical pages below CONFIG_ZERO_PAGE_OFFSET. * Reserve physical pages below CONFIG_ZERO_PAGE_OFFSET.
*/ */
if (CONFIG_ZERO_PAGE_OFFSET != 0) if (CONFIG_ZERO_PAGE_OFFSET != 0)
lmb_reserve(__MEMORY_START, CONFIG_ZERO_PAGE_OFFSET); memblock_reserve(__MEMORY_START, CONFIG_ZERO_PAGE_OFFSET);
/* /*
* Handle additional early reservations * Handle additional early reservations
@ -299,27 +299,27 @@ void __init paging_init(void)
unsigned long vaddr, end; unsigned long vaddr, end;
int nid; int nid;
lmb_init(); memblock_init();
sh_mv.mv_mem_init(); sh_mv.mv_mem_init();
early_reserve_mem(); early_reserve_mem();
lmb_enforce_memory_limit(memory_limit); memblock_enforce_memory_limit(memory_limit);
lmb_analyze(); memblock_analyze();
lmb_dump_all(); memblock_dump_all();
/* /*
* Determine low and high memory ranges: * Determine low and high memory ranges:
*/ */
max_low_pfn = max_pfn = lmb_end_of_DRAM() >> PAGE_SHIFT; max_low_pfn = max_pfn = memblock_end_of_DRAM() >> PAGE_SHIFT;
min_low_pfn = __MEMORY_START >> PAGE_SHIFT; min_low_pfn = __MEMORY_START >> PAGE_SHIFT;
nodes_clear(node_online_map); nodes_clear(node_online_map);
memory_start = (unsigned long)__va(__MEMORY_START); memory_start = (unsigned long)__va(__MEMORY_START);
memory_end = memory_start + (memory_limit ?: lmb_phys_mem_size()); memory_end = memory_start + (memory_limit ?: memblock_phys_mem_size());
uncached_init(); uncached_init();
pmb_init(); pmb_init();

View File

@ -9,7 +9,7 @@
*/ */
#include <linux/module.h> #include <linux/module.h>
#include <linux/bootmem.h> #include <linux/bootmem.h>
#include <linux/lmb.h> #include <linux/memblock.h>
#include <linux/mm.h> #include <linux/mm.h>
#include <linux/numa.h> #include <linux/numa.h>
#include <linux/pfn.h> #include <linux/pfn.h>
@ -39,12 +39,12 @@ void __init setup_bootmem_node(int nid, unsigned long start, unsigned long end)
pmb_bolt_mapping((unsigned long)__va(start), start, end - start, pmb_bolt_mapping((unsigned long)__va(start), start, end - start,
PAGE_KERNEL); PAGE_KERNEL);
lmb_add(start, end - start); memblock_add(start, end - start);
__add_active_range(nid, start_pfn, end_pfn); __add_active_range(nid, start_pfn, end_pfn);
/* Node-local pgdat */ /* Node-local pgdat */
NODE_DATA(nid) = __va(lmb_alloc_base(sizeof(struct pglist_data), NODE_DATA(nid) = __va(memblock_alloc_base(sizeof(struct pglist_data),
SMP_CACHE_BYTES, end)); SMP_CACHE_BYTES, end));
memset(NODE_DATA(nid), 0, sizeof(struct pglist_data)); memset(NODE_DATA(nid), 0, sizeof(struct pglist_data));
@ -54,7 +54,7 @@ void __init setup_bootmem_node(int nid, unsigned long start, unsigned long end)
/* Node-local bootmap */ /* Node-local bootmap */
bootmap_pages = bootmem_bootmap_pages(end_pfn - start_pfn); bootmap_pages = bootmem_bootmap_pages(end_pfn - start_pfn);
bootmem_paddr = lmb_alloc_base(bootmap_pages << PAGE_SHIFT, bootmem_paddr = memblock_alloc_base(bootmap_pages << PAGE_SHIFT,
PAGE_SIZE, end); PAGE_SIZE, end);
init_bootmem_node(NODE_DATA(nid), bootmem_paddr >> PAGE_SHIFT, init_bootmem_node(NODE_DATA(nid), bootmem_paddr >> PAGE_SHIFT,
start_pfn, end_pfn); start_pfn, end_pfn);

View File

@ -42,7 +42,7 @@ config SPARC64
select HAVE_FUNCTION_TRACE_MCOUNT_TEST select HAVE_FUNCTION_TRACE_MCOUNT_TEST
select HAVE_KRETPROBES select HAVE_KRETPROBES
select HAVE_KPROBES select HAVE_KPROBES
select HAVE_LMB select HAVE_MEMBLOCK
select HAVE_SYSCALL_WRAPPERS select HAVE_SYSCALL_WRAPPERS
select HAVE_DYNAMIC_FTRACE select HAVE_DYNAMIC_FTRACE
select HAVE_FTRACE_MCOUNT_RECORD select HAVE_FTRACE_MCOUNT_RECORD

View File

@ -1,10 +0,0 @@
#ifndef _SPARC64_LMB_H
#define _SPARC64_LMB_H
#include <asm/oplib.h>
#define LMB_DBG(fmt...) prom_printf(fmt)
#define LMB_REAL_LIMIT 0
#endif /* !(_SPARC64_LMB_H) */

View File

@ -0,0 +1,10 @@
#ifndef _SPARC64_MEMBLOCK_H
#define _SPARC64_MEMBLOCK_H
#include <asm/oplib.h>
#define MEMBLOCK_DBG(fmt...) prom_printf(fmt)
#define MEMBLOCK_REAL_LIMIT 0
#endif /* !(_SPARC64_MEMBLOCK_H) */

View File

@ -4,7 +4,7 @@
*/ */
#include <linux/kernel.h> #include <linux/kernel.h>
#include <linux/types.h> #include <linux/types.h>
#include <linux/lmb.h> #include <linux/memblock.h>
#include <linux/log2.h> #include <linux/log2.h>
#include <linux/list.h> #include <linux/list.h>
#include <linux/slab.h> #include <linux/slab.h>
@ -86,7 +86,7 @@ static void mdesc_handle_init(struct mdesc_handle *hp,
hp->handle_size = handle_size; hp->handle_size = handle_size;
} }
static struct mdesc_handle * __init mdesc_lmb_alloc(unsigned int mdesc_size) static struct mdesc_handle * __init mdesc_memblock_alloc(unsigned int mdesc_size)
{ {
unsigned int handle_size, alloc_size; unsigned int handle_size, alloc_size;
struct mdesc_handle *hp; struct mdesc_handle *hp;
@ -97,7 +97,7 @@ static struct mdesc_handle * __init mdesc_lmb_alloc(unsigned int mdesc_size)
mdesc_size); mdesc_size);
alloc_size = PAGE_ALIGN(handle_size); alloc_size = PAGE_ALIGN(handle_size);
paddr = lmb_alloc(alloc_size, PAGE_SIZE); paddr = memblock_alloc(alloc_size, PAGE_SIZE);
hp = NULL; hp = NULL;
if (paddr) { if (paddr) {
@ -107,7 +107,7 @@ static struct mdesc_handle * __init mdesc_lmb_alloc(unsigned int mdesc_size)
return hp; return hp;
} }
static void mdesc_lmb_free(struct mdesc_handle *hp) static void mdesc_memblock_free(struct mdesc_handle *hp)
{ {
unsigned int alloc_size; unsigned int alloc_size;
unsigned long start; unsigned long start;
@ -120,9 +120,9 @@ static void mdesc_lmb_free(struct mdesc_handle *hp)
free_bootmem_late(start, alloc_size); free_bootmem_late(start, alloc_size);
} }
static struct mdesc_mem_ops lmb_mdesc_ops = { static struct mdesc_mem_ops memblock_mdesc_ops = {
.alloc = mdesc_lmb_alloc, .alloc = mdesc_memblock_alloc,
.free = mdesc_lmb_free, .free = mdesc_memblock_free,
}; };
static struct mdesc_handle *mdesc_kmalloc(unsigned int mdesc_size) static struct mdesc_handle *mdesc_kmalloc(unsigned int mdesc_size)
@ -914,7 +914,7 @@ void __init sun4v_mdesc_init(void)
printk("MDESC: Size is %lu bytes.\n", len); printk("MDESC: Size is %lu bytes.\n", len);
hp = mdesc_alloc(len, &lmb_mdesc_ops); hp = mdesc_alloc(len, &memblock_mdesc_ops);
if (hp == NULL) { if (hp == NULL) {
prom_printf("MDESC: alloc of %lu bytes failed.\n", len); prom_printf("MDESC: alloc of %lu bytes failed.\n", len);
prom_halt(); prom_halt();

View File

@ -20,7 +20,7 @@
#include <linux/string.h> #include <linux/string.h>
#include <linux/mm.h> #include <linux/mm.h>
#include <linux/module.h> #include <linux/module.h>
#include <linux/lmb.h> #include <linux/memblock.h>
#include <linux/of_device.h> #include <linux/of_device.h>
#include <asm/prom.h> #include <asm/prom.h>
@ -34,7 +34,7 @@
void * __init prom_early_alloc(unsigned long size) void * __init prom_early_alloc(unsigned long size)
{ {
unsigned long paddr = lmb_alloc(size, SMP_CACHE_BYTES); unsigned long paddr = memblock_alloc(size, SMP_CACHE_BYTES);
void *ret; void *ret;
if (!paddr) { if (!paddr) {

View File

@ -23,7 +23,7 @@
#include <linux/cache.h> #include <linux/cache.h>
#include <linux/sort.h> #include <linux/sort.h>
#include <linux/percpu.h> #include <linux/percpu.h>
#include <linux/lmb.h> #include <linux/memblock.h>
#include <linux/mmzone.h> #include <linux/mmzone.h>
#include <linux/gfp.h> #include <linux/gfp.h>
@ -726,7 +726,7 @@ static void __init find_ramdisk(unsigned long phys_base)
initrd_start = ramdisk_image; initrd_start = ramdisk_image;
initrd_end = ramdisk_image + sparc_ramdisk_size; initrd_end = ramdisk_image + sparc_ramdisk_size;
lmb_reserve(initrd_start, sparc_ramdisk_size); memblock_reserve(initrd_start, sparc_ramdisk_size);
initrd_start += PAGE_OFFSET; initrd_start += PAGE_OFFSET;
initrd_end += PAGE_OFFSET; initrd_end += PAGE_OFFSET;
@ -822,7 +822,7 @@ static void __init allocate_node_data(int nid)
struct pglist_data *p; struct pglist_data *p;
#ifdef CONFIG_NEED_MULTIPLE_NODES #ifdef CONFIG_NEED_MULTIPLE_NODES
paddr = lmb_alloc_nid(sizeof(struct pglist_data), paddr = memblock_alloc_nid(sizeof(struct pglist_data),
SMP_CACHE_BYTES, nid, nid_range); SMP_CACHE_BYTES, nid, nid_range);
if (!paddr) { if (!paddr) {
prom_printf("Cannot allocate pglist_data for nid[%d]\n", nid); prom_printf("Cannot allocate pglist_data for nid[%d]\n", nid);
@ -843,7 +843,7 @@ static void __init allocate_node_data(int nid)
if (p->node_spanned_pages) { if (p->node_spanned_pages) {
num_pages = bootmem_bootmap_pages(p->node_spanned_pages); num_pages = bootmem_bootmap_pages(p->node_spanned_pages);
paddr = lmb_alloc_nid(num_pages << PAGE_SHIFT, PAGE_SIZE, nid, paddr = memblock_alloc_nid(num_pages << PAGE_SHIFT, PAGE_SIZE, nid,
nid_range); nid_range);
if (!paddr) { if (!paddr) {
prom_printf("Cannot allocate bootmap for nid[%d]\n", prom_printf("Cannot allocate bootmap for nid[%d]\n",
@ -974,11 +974,11 @@ static void __init add_node_ranges(void)
{ {
int i; int i;
for (i = 0; i < lmb.memory.cnt; i++) { for (i = 0; i < memblock.memory.cnt; i++) {
unsigned long size = lmb_size_bytes(&lmb.memory, i); unsigned long size = memblock_size_bytes(&memblock.memory, i);
unsigned long start, end; unsigned long start, end;
start = lmb.memory.region[i].base; start = memblock.memory.region[i].base;
end = start + size; end = start + size;
while (start < end) { while (start < end) {
unsigned long this_end; unsigned long this_end;
@ -1010,7 +1010,7 @@ static int __init grab_mlgroups(struct mdesc_handle *md)
if (!count) if (!count)
return -ENOENT; return -ENOENT;
paddr = lmb_alloc(count * sizeof(struct mdesc_mlgroup), paddr = memblock_alloc(count * sizeof(struct mdesc_mlgroup),
SMP_CACHE_BYTES); SMP_CACHE_BYTES);
if (!paddr) if (!paddr)
return -ENOMEM; return -ENOMEM;
@ -1051,7 +1051,7 @@ static int __init grab_mblocks(struct mdesc_handle *md)
if (!count) if (!count)
return -ENOENT; return -ENOENT;
paddr = lmb_alloc(count * sizeof(struct mdesc_mblock), paddr = memblock_alloc(count * sizeof(struct mdesc_mblock),
SMP_CACHE_BYTES); SMP_CACHE_BYTES);
if (!paddr) if (!paddr)
return -ENOMEM; return -ENOMEM;
@ -1279,8 +1279,8 @@ static int bootmem_init_numa(void)
static void __init bootmem_init_nonnuma(void) static void __init bootmem_init_nonnuma(void)
{ {
unsigned long top_of_ram = lmb_end_of_DRAM(); unsigned long top_of_ram = memblock_end_of_DRAM();
unsigned long total_ram = lmb_phys_mem_size(); unsigned long total_ram = memblock_phys_mem_size();
unsigned int i; unsigned int i;
numadbg("bootmem_init_nonnuma()\n"); numadbg("bootmem_init_nonnuma()\n");
@ -1292,15 +1292,15 @@ static void __init bootmem_init_nonnuma(void)
init_node_masks_nonnuma(); init_node_masks_nonnuma();
for (i = 0; i < lmb.memory.cnt; i++) { for (i = 0; i < memblock.memory.cnt; i++) {
unsigned long size = lmb_size_bytes(&lmb.memory, i); unsigned long size = memblock_size_bytes(&memblock.memory, i);
unsigned long start_pfn, end_pfn; unsigned long start_pfn, end_pfn;
if (!size) if (!size)
continue; continue;
start_pfn = lmb.memory.region[i].base >> PAGE_SHIFT; start_pfn = memblock.memory.region[i].base >> PAGE_SHIFT;
end_pfn = start_pfn + lmb_size_pages(&lmb.memory, i); end_pfn = start_pfn + memblock_size_pages(&memblock.memory, i);
add_active_range(0, start_pfn, end_pfn); add_active_range(0, start_pfn, end_pfn);
} }
@ -1338,9 +1338,9 @@ static void __init trim_reserved_in_node(int nid)
numadbg(" trim_reserved_in_node(%d)\n", nid); numadbg(" trim_reserved_in_node(%d)\n", nid);
for (i = 0; i < lmb.reserved.cnt; i++) { for (i = 0; i < memblock.reserved.cnt; i++) {
unsigned long start = lmb.reserved.region[i].base; unsigned long start = memblock.reserved.region[i].base;
unsigned long size = lmb_size_bytes(&lmb.reserved, i); unsigned long size = memblock_size_bytes(&memblock.reserved, i);
unsigned long end = start + size; unsigned long end = start + size;
reserve_range_in_node(nid, start, end); reserve_range_in_node(nid, start, end);
@ -1384,7 +1384,7 @@ static unsigned long __init bootmem_init(unsigned long phys_base)
unsigned long end_pfn; unsigned long end_pfn;
int nid; int nid;
end_pfn = lmb_end_of_DRAM() >> PAGE_SHIFT; end_pfn = memblock_end_of_DRAM() >> PAGE_SHIFT;
max_pfn = max_low_pfn = end_pfn; max_pfn = max_low_pfn = end_pfn;
min_low_pfn = (phys_base >> PAGE_SHIFT); min_low_pfn = (phys_base >> PAGE_SHIFT);
@ -1734,7 +1734,7 @@ void __init paging_init(void)
sun4v_ktsb_init(); sun4v_ktsb_init();
} }
lmb_init(); memblock_init();
/* Find available physical memory... /* Find available physical memory...
* *
@ -1752,17 +1752,17 @@ void __init paging_init(void)
phys_base = 0xffffffffffffffffUL; phys_base = 0xffffffffffffffffUL;
for (i = 0; i < pavail_ents; i++) { for (i = 0; i < pavail_ents; i++) {
phys_base = min(phys_base, pavail[i].phys_addr); phys_base = min(phys_base, pavail[i].phys_addr);
lmb_add(pavail[i].phys_addr, pavail[i].reg_size); memblock_add(pavail[i].phys_addr, pavail[i].reg_size);
} }
lmb_reserve(kern_base, kern_size); memblock_reserve(kern_base, kern_size);
find_ramdisk(phys_base); find_ramdisk(phys_base);
lmb_enforce_memory_limit(cmdline_memory_size); memblock_enforce_memory_limit(cmdline_memory_size);
lmb_analyze(); memblock_analyze();
lmb_dump_all(); memblock_dump_all();
set_bit(0, mmu_context_bmap); set_bit(0, mmu_context_bmap);
@ -1816,8 +1816,8 @@ void __init paging_init(void)
*/ */
for_each_possible_cpu(i) { for_each_possible_cpu(i) {
/* XXX Use node local allocations... XXX */ /* XXX Use node local allocations... XXX */
softirq_stack[i] = __va(lmb_alloc(THREAD_SIZE, THREAD_SIZE)); softirq_stack[i] = __va(memblock_alloc(THREAD_SIZE, THREAD_SIZE));
hardirq_stack[i] = __va(lmb_alloc(THREAD_SIZE, THREAD_SIZE)); hardirq_stack[i] = __va(memblock_alloc(THREAD_SIZE, THREAD_SIZE));
} }
/* Setup bootmem... */ /* Setup bootmem... */

View File

@ -1,89 +0,0 @@
#ifndef _LINUX_LMB_H
#define _LINUX_LMB_H
#ifdef __KERNEL__
/*
* Logical memory blocks.
*
* Copyright (C) 2001 Peter Bergner, IBM Corp.
*
* This program is free software; you can redistribute it and/or
* modify it under the terms of the GNU General Public License
* as published by the Free Software Foundation; either version
* 2 of the License, or (at your option) any later version.
*/
#include <linux/init.h>
#include <linux/mm.h>
#define MAX_LMB_REGIONS 128
struct lmb_property {
u64 base;
u64 size;
};
struct lmb_region {
unsigned long cnt;
u64 size;
struct lmb_property region[MAX_LMB_REGIONS+1];
};
struct lmb {
unsigned long debug;
u64 rmo_size;
struct lmb_region memory;
struct lmb_region reserved;
};
extern struct lmb lmb;
extern void __init lmb_init(void);
extern void __init lmb_analyze(void);
extern long lmb_add(u64 base, u64 size);
extern long lmb_remove(u64 base, u64 size);
extern long __init lmb_free(u64 base, u64 size);
extern long __init lmb_reserve(u64 base, u64 size);
extern u64 __init lmb_alloc_nid(u64 size, u64 align, int nid,
u64 (*nid_range)(u64, u64, int *));
extern u64 __init lmb_alloc(u64 size, u64 align);
extern u64 __init lmb_alloc_base(u64 size,
u64, u64 max_addr);
extern u64 __init __lmb_alloc_base(u64 size,
u64 align, u64 max_addr);
extern u64 __init lmb_phys_mem_size(void);
extern u64 lmb_end_of_DRAM(void);
extern void __init lmb_enforce_memory_limit(u64 memory_limit);
extern int __init lmb_is_reserved(u64 addr);
extern int lmb_is_region_reserved(u64 base, u64 size);
extern int lmb_find(struct lmb_property *res);
extern void lmb_dump_all(void);
static inline u64
lmb_size_bytes(struct lmb_region *type, unsigned long region_nr)
{
return type->region[region_nr].size;
}
static inline u64
lmb_size_pages(struct lmb_region *type, unsigned long region_nr)
{
return lmb_size_bytes(type, region_nr) >> PAGE_SHIFT;
}
static inline u64
lmb_start_pfn(struct lmb_region *type, unsigned long region_nr)
{
return type->region[region_nr].base >> PAGE_SHIFT;
}
static inline u64
lmb_end_pfn(struct lmb_region *type, unsigned long region_nr)
{
return lmb_start_pfn(type, region_nr) +
lmb_size_pages(type, region_nr);
}
#include <asm/lmb.h>
#endif /* __KERNEL__ */
#endif /* _LINUX_LMB_H */

89
include/linux/memblock.h Normal file
View File

@ -0,0 +1,89 @@
#ifndef _LINUX_MEMBLOCK_H
#define _LINUX_MEMBLOCK_H
#ifdef __KERNEL__
/*
* Logical memory blocks.
*
* Copyright (C) 2001 Peter Bergner, IBM Corp.
*
* This program is free software; you can redistribute it and/or
* modify it under the terms of the GNU General Public License
* as published by the Free Software Foundation; either version
* 2 of the License, or (at your option) any later version.
*/
#include <linux/init.h>
#include <linux/mm.h>
#define MAX_MEMBLOCK_REGIONS 128
struct memblock_property {
u64 base;
u64 size;
};
struct memblock_region {
unsigned long cnt;
u64 size;
struct memblock_property region[MAX_MEMBLOCK_REGIONS+1];
};
struct memblock {
unsigned long debug;
u64 rmo_size;
struct memblock_region memory;
struct memblock_region reserved;
};
extern struct memblock memblock;
extern void __init memblock_init(void);
extern void __init memblock_analyze(void);
extern long memblock_add(u64 base, u64 size);
extern long memblock_remove(u64 base, u64 size);
extern long __init memblock_free(u64 base, u64 size);
extern long __init memblock_reserve(u64 base, u64 size);
extern u64 __init memblock_alloc_nid(u64 size, u64 align, int nid,
u64 (*nid_range)(u64, u64, int *));
extern u64 __init memblock_alloc(u64 size, u64 align);
extern u64 __init memblock_alloc_base(u64 size,
u64, u64 max_addr);
extern u64 __init __memblock_alloc_base(u64 size,
u64 align, u64 max_addr);
extern u64 __init memblock_phys_mem_size(void);
extern u64 memblock_end_of_DRAM(void);
extern void __init memblock_enforce_memory_limit(u64 memory_limit);
extern int __init memblock_is_reserved(u64 addr);
extern int memblock_is_region_reserved(u64 base, u64 size);
extern int memblock_find(struct memblock_property *res);
extern void memblock_dump_all(void);
static inline u64
memblock_size_bytes(struct memblock_region *type, unsigned long region_nr)
{
return type->region[region_nr].size;
}
static inline u64
memblock_size_pages(struct memblock_region *type, unsigned long region_nr)
{
return memblock_size_bytes(type, region_nr) >> PAGE_SHIFT;
}
static inline u64
memblock_start_pfn(struct memblock_region *type, unsigned long region_nr)
{
return type->region[region_nr].base >> PAGE_SHIFT;
}
static inline u64
memblock_end_pfn(struct memblock_region *type, unsigned long region_nr)
{
return memblock_start_pfn(type, region_nr) +
memblock_size_pages(type, region_nr);
}
#include <asm/memblock.h>
#endif /* __KERNEL__ */
#endif /* _LINUX_MEMBLOCK_H */

View File

@ -181,9 +181,6 @@ config HAS_DMA
config CHECK_SIGNATURE config CHECK_SIGNATURE
bool bool
config HAVE_LMB
boolean
config CPUMASK_OFFSTACK config CPUMASK_OFFSTACK
bool "Force CPU masks off stack" if DEBUG_PER_CPU_MAPS bool "Force CPU masks off stack" if DEBUG_PER_CPU_MAPS
help help

View File

@ -89,8 +89,6 @@ obj-$(CONFIG_CPU_NOTIFIER_ERROR_INJECT) += cpu-notifier-error-inject.o
lib-$(CONFIG_GENERIC_BUG) += bug.o lib-$(CONFIG_GENERIC_BUG) += bug.o
obj-$(CONFIG_HAVE_LMB) += lmb.o
obj-$(CONFIG_HAVE_ARCH_TRACEHOOK) += syscall.o obj-$(CONFIG_HAVE_ARCH_TRACEHOOK) += syscall.o
obj-$(CONFIG_DYNAMIC_DEBUG) += dynamic_debug.o obj-$(CONFIG_DYNAMIC_DEBUG) += dynamic_debug.o

541
lib/lmb.c
View File

@ -1,541 +0,0 @@
/*
* Procedures for maintaining information about logical memory blocks.
*
* Peter Bergner, IBM Corp. June 2001.
* Copyright (C) 2001 Peter Bergner.
*
* This program is free software; you can redistribute it and/or
* modify it under the terms of the GNU General Public License
* as published by the Free Software Foundation; either version
* 2 of the License, or (at your option) any later version.
*/
#include <linux/kernel.h>
#include <linux/init.h>
#include <linux/bitops.h>
#include <linux/lmb.h>
#define LMB_ALLOC_ANYWHERE 0
struct lmb lmb;
static int lmb_debug;
static int __init early_lmb(char *p)
{
if (p && strstr(p, "debug"))
lmb_debug = 1;
return 0;
}
early_param("lmb", early_lmb);
static void lmb_dump(struct lmb_region *region, char *name)
{
unsigned long long base, size;
int i;
pr_info(" %s.cnt = 0x%lx\n", name, region->cnt);
for (i = 0; i < region->cnt; i++) {
base = region->region[i].base;
size = region->region[i].size;
pr_info(" %s[0x%x]\t0x%016llx - 0x%016llx, 0x%llx bytes\n",
name, i, base, base + size - 1, size);
}
}
void lmb_dump_all(void)
{
if (!lmb_debug)
return;
pr_info("LMB configuration:\n");
pr_info(" rmo_size = 0x%llx\n", (unsigned long long)lmb.rmo_size);
pr_info(" memory.size = 0x%llx\n", (unsigned long long)lmb.memory.size);
lmb_dump(&lmb.memory, "memory");
lmb_dump(&lmb.reserved, "reserved");
}
static unsigned long lmb_addrs_overlap(u64 base1, u64 size1, u64 base2,
u64 size2)
{
return ((base1 < (base2 + size2)) && (base2 < (base1 + size1)));
}
static long lmb_addrs_adjacent(u64 base1, u64 size1, u64 base2, u64 size2)
{
if (base2 == base1 + size1)
return 1;
else if (base1 == base2 + size2)
return -1;
return 0;
}
static long lmb_regions_adjacent(struct lmb_region *rgn,
unsigned long r1, unsigned long r2)
{
u64 base1 = rgn->region[r1].base;
u64 size1 = rgn->region[r1].size;
u64 base2 = rgn->region[r2].base;
u64 size2 = rgn->region[r2].size;
return lmb_addrs_adjacent(base1, size1, base2, size2);
}
static void lmb_remove_region(struct lmb_region *rgn, unsigned long r)
{
unsigned long i;
for (i = r; i < rgn->cnt - 1; i++) {
rgn->region[i].base = rgn->region[i + 1].base;
rgn->region[i].size = rgn->region[i + 1].size;
}
rgn->cnt--;
}
/* Assumption: base addr of region 1 < base addr of region 2 */
static void lmb_coalesce_regions(struct lmb_region *rgn,
unsigned long r1, unsigned long r2)
{
rgn->region[r1].size += rgn->region[r2].size;
lmb_remove_region(rgn, r2);
}
void __init lmb_init(void)
{
/* Create a dummy zero size LMB which will get coalesced away later.
* This simplifies the lmb_add() code below...
*/
lmb.memory.region[0].base = 0;
lmb.memory.region[0].size = 0;
lmb.memory.cnt = 1;
/* Ditto. */
lmb.reserved.region[0].base = 0;
lmb.reserved.region[0].size = 0;
lmb.reserved.cnt = 1;
}
void __init lmb_analyze(void)
{
int i;
lmb.memory.size = 0;
for (i = 0; i < lmb.memory.cnt; i++)
lmb.memory.size += lmb.memory.region[i].size;
}
static long lmb_add_region(struct lmb_region *rgn, u64 base, u64 size)
{
unsigned long coalesced = 0;
long adjacent, i;
if ((rgn->cnt == 1) && (rgn->region[0].size == 0)) {
rgn->region[0].base = base;
rgn->region[0].size = size;
return 0;
}
/* First try and coalesce this LMB with another. */
for (i = 0; i < rgn->cnt; i++) {
u64 rgnbase = rgn->region[i].base;
u64 rgnsize = rgn->region[i].size;
if ((rgnbase == base) && (rgnsize == size))
/* Already have this region, so we're done */
return 0;
adjacent = lmb_addrs_adjacent(base, size, rgnbase, rgnsize);
if (adjacent > 0) {
rgn->region[i].base -= size;
rgn->region[i].size += size;
coalesced++;
break;
} else if (adjacent < 0) {
rgn->region[i].size += size;
coalesced++;
break;
}
}
if ((i < rgn->cnt - 1) && lmb_regions_adjacent(rgn, i, i+1)) {
lmb_coalesce_regions(rgn, i, i+1);
coalesced++;
}
if (coalesced)
return coalesced;
if (rgn->cnt >= MAX_LMB_REGIONS)
return -1;
/* Couldn't coalesce the LMB, so add it to the sorted table. */
for (i = rgn->cnt - 1; i >= 0; i--) {
if (base < rgn->region[i].base) {
rgn->region[i+1].base = rgn->region[i].base;
rgn->region[i+1].size = rgn->region[i].size;
} else {
rgn->region[i+1].base = base;
rgn->region[i+1].size = size;
break;
}
}
if (base < rgn->region[0].base) {
rgn->region[0].base = base;
rgn->region[0].size = size;
}
rgn->cnt++;
return 0;
}
long lmb_add(u64 base, u64 size)
{
struct lmb_region *_rgn = &lmb.memory;
/* On pSeries LPAR systems, the first LMB is our RMO region. */
if (base == 0)
lmb.rmo_size = size;
return lmb_add_region(_rgn, base, size);
}
static long __lmb_remove(struct lmb_region *rgn, u64 base, u64 size)
{
u64 rgnbegin, rgnend;
u64 end = base + size;
int i;
rgnbegin = rgnend = 0; /* supress gcc warnings */
/* Find the region where (base, size) belongs to */
for (i=0; i < rgn->cnt; i++) {
rgnbegin = rgn->region[i].base;
rgnend = rgnbegin + rgn->region[i].size;
if ((rgnbegin <= base) && (end <= rgnend))
break;
}
/* Didn't find the region */
if (i == rgn->cnt)
return -1;
/* Check to see if we are removing entire region */
if ((rgnbegin == base) && (rgnend == end)) {
lmb_remove_region(rgn, i);
return 0;
}
/* Check to see if region is matching at the front */
if (rgnbegin == base) {
rgn->region[i].base = end;
rgn->region[i].size -= size;
return 0;
}
/* Check to see if the region is matching at the end */
if (rgnend == end) {
rgn->region[i].size -= size;
return 0;
}
/*
* We need to split the entry - adjust the current one to the
* beginging of the hole and add the region after hole.
*/
rgn->region[i].size = base - rgn->region[i].base;
return lmb_add_region(rgn, end, rgnend - end);
}
long lmb_remove(u64 base, u64 size)
{
return __lmb_remove(&lmb.memory, base, size);
}
long __init lmb_free(u64 base, u64 size)
{
return __lmb_remove(&lmb.reserved, base, size);
}
long __init lmb_reserve(u64 base, u64 size)
{
struct lmb_region *_rgn = &lmb.reserved;
BUG_ON(0 == size);
return lmb_add_region(_rgn, base, size);
}
long lmb_overlaps_region(struct lmb_region *rgn, u64 base, u64 size)
{
unsigned long i;
for (i = 0; i < rgn->cnt; i++) {
u64 rgnbase = rgn->region[i].base;
u64 rgnsize = rgn->region[i].size;
if (lmb_addrs_overlap(base, size, rgnbase, rgnsize))
break;
}
return (i < rgn->cnt) ? i : -1;
}
static u64 lmb_align_down(u64 addr, u64 size)
{
return addr & ~(size - 1);
}
static u64 lmb_align_up(u64 addr, u64 size)
{
return (addr + (size - 1)) & ~(size - 1);
}
static u64 __init lmb_alloc_nid_unreserved(u64 start, u64 end,
u64 size, u64 align)
{
u64 base, res_base;
long j;
base = lmb_align_down((end - size), align);
while (start <= base) {
j = lmb_overlaps_region(&lmb.reserved, base, size);
if (j < 0) {
/* this area isn't reserved, take it */
if (lmb_add_region(&lmb.reserved, base, size) < 0)
base = ~(u64)0;
return base;
}
res_base = lmb.reserved.region[j].base;
if (res_base < size)
break;
base = lmb_align_down(res_base - size, align);
}
return ~(u64)0;
}
static u64 __init lmb_alloc_nid_region(struct lmb_property *mp,
u64 (*nid_range)(u64, u64, int *),
u64 size, u64 align, int nid)
{
u64 start, end;
start = mp->base;
end = start + mp->size;
start = lmb_align_up(start, align);
while (start < end) {
u64 this_end;
int this_nid;
this_end = nid_range(start, end, &this_nid);
if (this_nid == nid) {
u64 ret = lmb_alloc_nid_unreserved(start, this_end,
size, align);
if (ret != ~(u64)0)
return ret;
}
start = this_end;
}
return ~(u64)0;
}
u64 __init lmb_alloc_nid(u64 size, u64 align, int nid,
u64 (*nid_range)(u64 start, u64 end, int *nid))
{
struct lmb_region *mem = &lmb.memory;
int i;
BUG_ON(0 == size);
size = lmb_align_up(size, align);
for (i = 0; i < mem->cnt; i++) {
u64 ret = lmb_alloc_nid_region(&mem->region[i],
nid_range,
size, align, nid);
if (ret != ~(u64)0)
return ret;
}
return lmb_alloc(size, align);
}
u64 __init lmb_alloc(u64 size, u64 align)
{
return lmb_alloc_base(size, align, LMB_ALLOC_ANYWHERE);
}
u64 __init lmb_alloc_base(u64 size, u64 align, u64 max_addr)
{
u64 alloc;
alloc = __lmb_alloc_base(size, align, max_addr);
if (alloc == 0)
panic("ERROR: Failed to allocate 0x%llx bytes below 0x%llx.\n",
(unsigned long long) size, (unsigned long long) max_addr);
return alloc;
}
u64 __init __lmb_alloc_base(u64 size, u64 align, u64 max_addr)
{
long i, j;
u64 base = 0;
u64 res_base;
BUG_ON(0 == size);
size = lmb_align_up(size, align);
/* On some platforms, make sure we allocate lowmem */
/* Note that LMB_REAL_LIMIT may be LMB_ALLOC_ANYWHERE */
if (max_addr == LMB_ALLOC_ANYWHERE)
max_addr = LMB_REAL_LIMIT;
for (i = lmb.memory.cnt - 1; i >= 0; i--) {
u64 lmbbase = lmb.memory.region[i].base;
u64 lmbsize = lmb.memory.region[i].size;
if (lmbsize < size)
continue;
if (max_addr == LMB_ALLOC_ANYWHERE)
base = lmb_align_down(lmbbase + lmbsize - size, align);
else if (lmbbase < max_addr) {
base = min(lmbbase + lmbsize, max_addr);
base = lmb_align_down(base - size, align);
} else
continue;
while (base && lmbbase <= base) {
j = lmb_overlaps_region(&lmb.reserved, base, size);
if (j < 0) {
/* this area isn't reserved, take it */
if (lmb_add_region(&lmb.reserved, base, size) < 0)
return 0;
return base;
}
res_base = lmb.reserved.region[j].base;
if (res_base < size)
break;
base = lmb_align_down(res_base - size, align);
}
}
return 0;
}
/* You must call lmb_analyze() before this. */
u64 __init lmb_phys_mem_size(void)
{
return lmb.memory.size;
}
u64 lmb_end_of_DRAM(void)
{
int idx = lmb.memory.cnt - 1;
return (lmb.memory.region[idx].base + lmb.memory.region[idx].size);
}
/* You must call lmb_analyze() after this. */
void __init lmb_enforce_memory_limit(u64 memory_limit)
{
unsigned long i;
u64 limit;
struct lmb_property *p;
if (!memory_limit)
return;
/* Truncate the lmb regions to satisfy the memory limit. */
limit = memory_limit;
for (i = 0; i < lmb.memory.cnt; i++) {
if (limit > lmb.memory.region[i].size) {
limit -= lmb.memory.region[i].size;
continue;
}
lmb.memory.region[i].size = limit;
lmb.memory.cnt = i + 1;
break;
}
if (lmb.memory.region[0].size < lmb.rmo_size)
lmb.rmo_size = lmb.memory.region[0].size;
memory_limit = lmb_end_of_DRAM();
/* And truncate any reserves above the limit also. */
for (i = 0; i < lmb.reserved.cnt; i++) {
p = &lmb.reserved.region[i];
if (p->base > memory_limit)
p->size = 0;
else if ((p->base + p->size) > memory_limit)
p->size = memory_limit - p->base;
if (p->size == 0) {
lmb_remove_region(&lmb.reserved, i);
i--;
}
}
}
int __init lmb_is_reserved(u64 addr)
{
int i;
for (i = 0; i < lmb.reserved.cnt; i++) {
u64 upper = lmb.reserved.region[i].base +
lmb.reserved.region[i].size - 1;
if ((addr >= lmb.reserved.region[i].base) && (addr <= upper))
return 1;
}
return 0;
}
int lmb_is_region_reserved(u64 base, u64 size)
{
return lmb_overlaps_region(&lmb.reserved, base, size);
}
/*
* Given a <base, len>, find which memory regions belong to this range.
* Adjust the request and return a contiguous chunk.
*/
int lmb_find(struct lmb_property *res)
{
int i;
u64 rstart, rend;
rstart = res->base;
rend = rstart + res->size - 1;
for (i = 0; i < lmb.memory.cnt; i++) {
u64 start = lmb.memory.region[i].base;
u64 end = start + lmb.memory.region[i].size - 1;
if (start > rend)
return -1;
if ((end >= rstart) && (start < rend)) {
/* adjust the request */
if (rstart < start)
rstart = start;
if (rend > end)
rend = end;
res->base = rstart;
res->size = rend - rstart + 1;
return 0;
}
}
return -1;
}

View File

@ -128,6 +128,9 @@ config SPARSEMEM_VMEMMAP
pfn_to_page and page_to_pfn operations. This is the most pfn_to_page and page_to_pfn operations. This is the most
efficient option when sufficient kernel resources are available. efficient option when sufficient kernel resources are available.
config HAVE_MEMBLOCK
boolean
# eventually, we can have this option just 'select SPARSEMEM' # eventually, we can have this option just 'select SPARSEMEM'
config MEMORY_HOTPLUG config MEMORY_HOTPLUG
bool "Allow for memory hot-add" bool "Allow for memory hot-add"

View File

@ -15,6 +15,8 @@ obj-y := bootmem.o filemap.o mempool.o oom_kill.o fadvise.o \
$(mmu-y) $(mmu-y)
obj-y += init-mm.o obj-y += init-mm.o
obj-$(CONFIG_HAVE_MEMBLOCK) += memblock.o
obj-$(CONFIG_BOUNCE) += bounce.o obj-$(CONFIG_BOUNCE) += bounce.o
obj-$(CONFIG_SWAP) += page_io.o swap_state.o swapfile.o thrash.o obj-$(CONFIG_SWAP) += page_io.o swap_state.o swapfile.o thrash.o
obj-$(CONFIG_HAS_DMA) += dmapool.o obj-$(CONFIG_HAS_DMA) += dmapool.o

541
mm/memblock.c Normal file
View File

@ -0,0 +1,541 @@
/*
* Procedures for maintaining information about logical memory blocks.
*
* Peter Bergner, IBM Corp. June 2001.
* Copyright (C) 2001 Peter Bergner.
*
* This program is free software; you can redistribute it and/or
* modify it under the terms of the GNU General Public License
* as published by the Free Software Foundation; either version
* 2 of the License, or (at your option) any later version.
*/
#include <linux/kernel.h>
#include <linux/init.h>
#include <linux/bitops.h>
#include <linux/memblock.h>
#define MEMBLOCK_ALLOC_ANYWHERE 0
struct memblock memblock;
static int memblock_debug;
static int __init early_memblock(char *p)
{
if (p && strstr(p, "debug"))
memblock_debug = 1;
return 0;
}
early_param("memblock", early_memblock);
static void memblock_dump(struct memblock_region *region, char *name)
{
unsigned long long base, size;
int i;
pr_info(" %s.cnt = 0x%lx\n", name, region->cnt);
for (i = 0; i < region->cnt; i++) {
base = region->region[i].base;
size = region->region[i].size;
pr_info(" %s[0x%x]\t0x%016llx - 0x%016llx, 0x%llx bytes\n",
name, i, base, base + size - 1, size);
}
}
void memblock_dump_all(void)
{
if (!memblock_debug)
return;
pr_info("MEMBLOCK configuration:\n");
pr_info(" rmo_size = 0x%llx\n", (unsigned long long)memblock.rmo_size);
pr_info(" memory.size = 0x%llx\n", (unsigned long long)memblock.memory.size);
memblock_dump(&memblock.memory, "memory");
memblock_dump(&memblock.reserved, "reserved");
}
static unsigned long memblock_addrs_overlap(u64 base1, u64 size1, u64 base2,
u64 size2)
{
return ((base1 < (base2 + size2)) && (base2 < (base1 + size1)));
}
static long memblock_addrs_adjacent(u64 base1, u64 size1, u64 base2, u64 size2)
{
if (base2 == base1 + size1)
return 1;
else if (base1 == base2 + size2)
return -1;
return 0;
}
static long memblock_regions_adjacent(struct memblock_region *rgn,
unsigned long r1, unsigned long r2)
{
u64 base1 = rgn->region[r1].base;
u64 size1 = rgn->region[r1].size;
u64 base2 = rgn->region[r2].base;
u64 size2 = rgn->region[r2].size;
return memblock_addrs_adjacent(base1, size1, base2, size2);
}
static void memblock_remove_region(struct memblock_region *rgn, unsigned long r)
{
unsigned long i;
for (i = r; i < rgn->cnt - 1; i++) {
rgn->region[i].base = rgn->region[i + 1].base;
rgn->region[i].size = rgn->region[i + 1].size;
}
rgn->cnt--;
}
/* Assumption: base addr of region 1 < base addr of region 2 */
static void memblock_coalesce_regions(struct memblock_region *rgn,
unsigned long r1, unsigned long r2)
{
rgn->region[r1].size += rgn->region[r2].size;
memblock_remove_region(rgn, r2);
}
void __init memblock_init(void)
{
/* Create a dummy zero size MEMBLOCK which will get coalesced away later.
* This simplifies the memblock_add() code below...
*/
memblock.memory.region[0].base = 0;
memblock.memory.region[0].size = 0;
memblock.memory.cnt = 1;
/* Ditto. */
memblock.reserved.region[0].base = 0;
memblock.reserved.region[0].size = 0;
memblock.reserved.cnt = 1;
}
void __init memblock_analyze(void)
{
int i;
memblock.memory.size = 0;
for (i = 0; i < memblock.memory.cnt; i++)
memblock.memory.size += memblock.memory.region[i].size;
}
static long memblock_add_region(struct memblock_region *rgn, u64 base, u64 size)
{
unsigned long coalesced = 0;
long adjacent, i;
if ((rgn->cnt == 1) && (rgn->region[0].size == 0)) {
rgn->region[0].base = base;
rgn->region[0].size = size;
return 0;
}
/* First try and coalesce this MEMBLOCK with another. */
for (i = 0; i < rgn->cnt; i++) {
u64 rgnbase = rgn->region[i].base;
u64 rgnsize = rgn->region[i].size;
if ((rgnbase == base) && (rgnsize == size))
/* Already have this region, so we're done */
return 0;
adjacent = memblock_addrs_adjacent(base, size, rgnbase, rgnsize);
if (adjacent > 0) {
rgn->region[i].base -= size;
rgn->region[i].size += size;
coalesced++;
break;
} else if (adjacent < 0) {
rgn->region[i].size += size;
coalesced++;
break;
}
}
if ((i < rgn->cnt - 1) && memblock_regions_adjacent(rgn, i, i+1)) {
memblock_coalesce_regions(rgn, i, i+1);
coalesced++;
}
if (coalesced)
return coalesced;
if (rgn->cnt >= MAX_MEMBLOCK_REGIONS)
return -1;
/* Couldn't coalesce the MEMBLOCK, so add it to the sorted table. */
for (i = rgn->cnt - 1; i >= 0; i--) {
if (base < rgn->region[i].base) {
rgn->region[i+1].base = rgn->region[i].base;
rgn->region[i+1].size = rgn->region[i].size;
} else {
rgn->region[i+1].base = base;
rgn->region[i+1].size = size;
break;
}
}
if (base < rgn->region[0].base) {
rgn->region[0].base = base;
rgn->region[0].size = size;
}
rgn->cnt++;
return 0;
}
long memblock_add(u64 base, u64 size)
{
struct memblock_region *_rgn = &memblock.memory;
/* On pSeries LPAR systems, the first MEMBLOCK is our RMO region. */
if (base == 0)
memblock.rmo_size = size;
return memblock_add_region(_rgn, base, size);
}
static long __memblock_remove(struct memblock_region *rgn, u64 base, u64 size)
{
u64 rgnbegin, rgnend;
u64 end = base + size;
int i;
rgnbegin = rgnend = 0; /* supress gcc warnings */
/* Find the region where (base, size) belongs to */
for (i=0; i < rgn->cnt; i++) {
rgnbegin = rgn->region[i].base;
rgnend = rgnbegin + rgn->region[i].size;
if ((rgnbegin <= base) && (end <= rgnend))
break;
}
/* Didn't find the region */
if (i == rgn->cnt)
return -1;
/* Check to see if we are removing entire region */
if ((rgnbegin == base) && (rgnend == end)) {
memblock_remove_region(rgn, i);
return 0;
}
/* Check to see if region is matching at the front */
if (rgnbegin == base) {
rgn->region[i].base = end;
rgn->region[i].size -= size;
return 0;
}
/* Check to see if the region is matching at the end */
if (rgnend == end) {
rgn->region[i].size -= size;
return 0;
}
/*
* We need to split the entry - adjust the current one to the
* beginging of the hole and add the region after hole.
*/
rgn->region[i].size = base - rgn->region[i].base;
return memblock_add_region(rgn, end, rgnend - end);
}
long memblock_remove(u64 base, u64 size)
{
return __memblock_remove(&memblock.memory, base, size);
}
long __init memblock_free(u64 base, u64 size)
{
return __memblock_remove(&memblock.reserved, base, size);
}
long __init memblock_reserve(u64 base, u64 size)
{
struct memblock_region *_rgn = &memblock.reserved;
BUG_ON(0 == size);
return memblock_add_region(_rgn, base, size);
}
long memblock_overlaps_region(struct memblock_region *rgn, u64 base, u64 size)
{
unsigned long i;
for (i = 0; i < rgn->cnt; i++) {
u64 rgnbase = rgn->region[i].base;
u64 rgnsize = rgn->region[i].size;
if (memblock_addrs_overlap(base, size, rgnbase, rgnsize))
break;
}
return (i < rgn->cnt) ? i : -1;
}
static u64 memblock_align_down(u64 addr, u64 size)
{
return addr & ~(size - 1);
}
static u64 memblock_align_up(u64 addr, u64 size)
{
return (addr + (size - 1)) & ~(size - 1);
}
static u64 __init memblock_alloc_nid_unreserved(u64 start, u64 end,
u64 size, u64 align)
{
u64 base, res_base;
long j;
base = memblock_align_down((end - size), align);
while (start <= base) {
j = memblock_overlaps_region(&memblock.reserved, base, size);
if (j < 0) {
/* this area isn't reserved, take it */
if (memblock_add_region(&memblock.reserved, base, size) < 0)
base = ~(u64)0;
return base;
}
res_base = memblock.reserved.region[j].base;
if (res_base < size)
break;
base = memblock_align_down(res_base - size, align);
}
return ~(u64)0;
}
static u64 __init memblock_alloc_nid_region(struct memblock_property *mp,
u64 (*nid_range)(u64, u64, int *),
u64 size, u64 align, int nid)
{
u64 start, end;
start = mp->base;
end = start + mp->size;
start = memblock_align_up(start, align);
while (start < end) {
u64 this_end;
int this_nid;
this_end = nid_range(start, end, &this_nid);
if (this_nid == nid) {
u64 ret = memblock_alloc_nid_unreserved(start, this_end,
size, align);
if (ret != ~(u64)0)
return ret;
}
start = this_end;
}
return ~(u64)0;
}
u64 __init memblock_alloc_nid(u64 size, u64 align, int nid,
u64 (*nid_range)(u64 start, u64 end, int *nid))
{
struct memblock_region *mem = &memblock.memory;
int i;
BUG_ON(0 == size);
size = memblock_align_up(size, align);
for (i = 0; i < mem->cnt; i++) {
u64 ret = memblock_alloc_nid_region(&mem->region[i],
nid_range,
size, align, nid);
if (ret != ~(u64)0)
return ret;
}
return memblock_alloc(size, align);
}
u64 __init memblock_alloc(u64 size, u64 align)
{
return memblock_alloc_base(size, align, MEMBLOCK_ALLOC_ANYWHERE);
}
u64 __init memblock_alloc_base(u64 size, u64 align, u64 max_addr)
{
u64 alloc;
alloc = __memblock_alloc_base(size, align, max_addr);
if (alloc == 0)
panic("ERROR: Failed to allocate 0x%llx bytes below 0x%llx.\n",
(unsigned long long) size, (unsigned long long) max_addr);
return alloc;
}
u64 __init __memblock_alloc_base(u64 size, u64 align, u64 max_addr)
{
long i, j;
u64 base = 0;
u64 res_base;
BUG_ON(0 == size);
size = memblock_align_up(size, align);
/* On some platforms, make sure we allocate lowmem */
/* Note that MEMBLOCK_REAL_LIMIT may be MEMBLOCK_ALLOC_ANYWHERE */
if (max_addr == MEMBLOCK_ALLOC_ANYWHERE)
max_addr = MEMBLOCK_REAL_LIMIT;
for (i = memblock.memory.cnt - 1; i >= 0; i--) {
u64 memblockbase = memblock.memory.region[i].base;
u64 memblocksize = memblock.memory.region[i].size;
if (memblocksize < size)
continue;
if (max_addr == MEMBLOCK_ALLOC_ANYWHERE)
base = memblock_align_down(memblockbase + memblocksize - size, align);
else if (memblockbase < max_addr) {
base = min(memblockbase + memblocksize, max_addr);
base = memblock_align_down(base - size, align);
} else
continue;
while (base && memblockbase <= base) {
j = memblock_overlaps_region(&memblock.reserved, base, size);
if (j < 0) {
/* this area isn't reserved, take it */
if (memblock_add_region(&memblock.reserved, base, size) < 0)
return 0;
return base;
}
res_base = memblock.reserved.region[j].base;
if (res_base < size)
break;
base = memblock_align_down(res_base - size, align);
}
}
return 0;
}
/* You must call memblock_analyze() before this. */
u64 __init memblock_phys_mem_size(void)
{
return memblock.memory.size;
}
u64 memblock_end_of_DRAM(void)
{
int idx = memblock.memory.cnt - 1;
return (memblock.memory.region[idx].base + memblock.memory.region[idx].size);
}
/* You must call memblock_analyze() after this. */
void __init memblock_enforce_memory_limit(u64 memory_limit)
{
unsigned long i;
u64 limit;
struct memblock_property *p;
if (!memory_limit)
return;
/* Truncate the memblock regions to satisfy the memory limit. */
limit = memory_limit;
for (i = 0; i < memblock.memory.cnt; i++) {
if (limit > memblock.memory.region[i].size) {
limit -= memblock.memory.region[i].size;
continue;
}
memblock.memory.region[i].size = limit;
memblock.memory.cnt = i + 1;
break;
}
if (memblock.memory.region[0].size < memblock.rmo_size)
memblock.rmo_size = memblock.memory.region[0].size;
memory_limit = memblock_end_of_DRAM();
/* And truncate any reserves above the limit also. */
for (i = 0; i < memblock.reserved.cnt; i++) {
p = &memblock.reserved.region[i];
if (p->base > memory_limit)
p->size = 0;
else if ((p->base + p->size) > memory_limit)
p->size = memory_limit - p->base;
if (p->size == 0) {
memblock_remove_region(&memblock.reserved, i);
i--;
}
}
}
int __init memblock_is_reserved(u64 addr)
{
int i;
for (i = 0; i < memblock.reserved.cnt; i++) {
u64 upper = memblock.reserved.region[i].base +
memblock.reserved.region[i].size - 1;
if ((addr >= memblock.reserved.region[i].base) && (addr <= upper))
return 1;
}
return 0;
}
int memblock_is_region_reserved(u64 base, u64 size)
{
return memblock_overlaps_region(&memblock.reserved, base, size);
}
/*
* Given a <base, len>, find which memory regions belong to this range.
* Adjust the request and return a contiguous chunk.
*/
int memblock_find(struct memblock_property *res)
{
int i;
u64 rstart, rend;
rstart = res->base;
rend = rstart + res->size - 1;
for (i = 0; i < memblock.memory.cnt; i++) {
u64 start = memblock.memory.region[i].base;
u64 end = start + memblock.memory.region[i].size - 1;
if (start > rend)
return -1;
if ((end >= rstart) && (start < rend)) {
/* adjust the request */
if (rstart < start)
rstart = start;
if (rend > end)
rend = end;
res->base = rstart;
res->size = rend - rstart + 1;
return 0;
}
}
return -1;
}