mirror of
https://mirrors.bfsu.edu.cn/git/linux.git
synced 2024-11-17 01:04:19 +08:00
2c6e6db41f
This attached patch significantly shrinks boot memory allocation on ia64. It does this by not allocating per_cpu areas for cpus that can never exist. In the case where acpi does not have any numa node description of the cpus, I defaulted to assigning the first 32 round-robin on the known nodes.. For the !CONFIG_ACPI I used for_each_possible_cpu(). Signed-off-by: Robin Holt <holt@sgi.com> Signed-off-by: Tony Luck <tony.luck@intel.com>
168 lines
4.8 KiB
C
168 lines
4.8 KiB
C
/*
|
|
* asm-ia64/acpi.h
|
|
*
|
|
* Copyright (C) 1999 VA Linux Systems
|
|
* Copyright (C) 1999 Walt Drummond <drummond@valinux.com>
|
|
* Copyright (C) 2000,2001 J.I. Lee <jung-ik.lee@intel.com>
|
|
* Copyright (C) 2001,2002 Paul Diefenbaugh <paul.s.diefenbaugh@intel.com>
|
|
*
|
|
* ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
|
*
|
|
* This program is free software; you can redistribute it and/or modify
|
|
* it under the terms of the GNU General Public License as published by
|
|
* the Free Software Foundation; either version 2 of the License, or
|
|
* (at your option) any later version.
|
|
*
|
|
* This program is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
* GNU General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU General Public License
|
|
* along with this program; if not, write to the Free Software
|
|
* Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
|
|
*
|
|
* ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
|
*/
|
|
|
|
#ifndef _ASM_ACPI_H
|
|
#define _ASM_ACPI_H
|
|
|
|
#ifdef __KERNEL__
|
|
|
|
#include <acpi/pdc_intel.h>
|
|
|
|
#include <linux/init.h>
|
|
#include <linux/numa.h>
|
|
#include <asm/system.h>
|
|
#include <asm/numa.h>
|
|
|
|
#define COMPILER_DEPENDENT_INT64 long
|
|
#define COMPILER_DEPENDENT_UINT64 unsigned long
|
|
|
|
/*
|
|
* Calling conventions:
|
|
*
|
|
* ACPI_SYSTEM_XFACE - Interfaces to host OS (handlers, threads)
|
|
* ACPI_EXTERNAL_XFACE - External ACPI interfaces
|
|
* ACPI_INTERNAL_XFACE - Internal ACPI interfaces
|
|
* ACPI_INTERNAL_VAR_XFACE - Internal variable-parameter list interfaces
|
|
*/
|
|
#define ACPI_SYSTEM_XFACE
|
|
#define ACPI_EXTERNAL_XFACE
|
|
#define ACPI_INTERNAL_XFACE
|
|
#define ACPI_INTERNAL_VAR_XFACE
|
|
|
|
/* Asm macros */
|
|
|
|
#define ACPI_ASM_MACROS
|
|
#define BREAKPOINT3
|
|
#define ACPI_DISABLE_IRQS() local_irq_disable()
|
|
#define ACPI_ENABLE_IRQS() local_irq_enable()
|
|
#define ACPI_FLUSH_CPU_CACHE()
|
|
|
|
static inline int
|
|
ia64_acpi_acquire_global_lock (unsigned int *lock)
|
|
{
|
|
unsigned int old, new, val;
|
|
do {
|
|
old = *lock;
|
|
new = (((old & ~0x3) + 2) + ((old >> 1) & 0x1));
|
|
val = ia64_cmpxchg4_acq(lock, new, old);
|
|
} while (unlikely (val != old));
|
|
return (new < 3) ? -1 : 0;
|
|
}
|
|
|
|
static inline int
|
|
ia64_acpi_release_global_lock (unsigned int *lock)
|
|
{
|
|
unsigned int old, new, val;
|
|
do {
|
|
old = *lock;
|
|
new = old & ~0x3;
|
|
val = ia64_cmpxchg4_acq(lock, new, old);
|
|
} while (unlikely (val != old));
|
|
return old & 0x1;
|
|
}
|
|
|
|
#define ACPI_ACQUIRE_GLOBAL_LOCK(facs, Acq) \
|
|
((Acq) = ia64_acpi_acquire_global_lock(&facs->global_lock))
|
|
|
|
#define ACPI_RELEASE_GLOBAL_LOCK(facs, Acq) \
|
|
((Acq) = ia64_acpi_release_global_lock(&facs->global_lock))
|
|
|
|
#define acpi_disabled 0 /* ACPI always enabled on IA64 */
|
|
#define acpi_noirq 0 /* ACPI always enabled on IA64 */
|
|
#define acpi_pci_disabled 0 /* ACPI PCI always enabled on IA64 */
|
|
#define acpi_strict 1 /* no ACPI spec workarounds on IA64 */
|
|
#define acpi_processor_cstate_check(x) (x) /* no idle limits on IA64 :) */
|
|
static inline void disable_acpi(void) { }
|
|
|
|
const char *acpi_get_sysname (void);
|
|
int acpi_request_vector (u32 int_type);
|
|
int acpi_gsi_to_irq (u32 gsi, unsigned int *irq);
|
|
|
|
/* routines for saving/restoring kernel state */
|
|
extern int acpi_save_state_mem(void);
|
|
extern void acpi_restore_state_mem(void);
|
|
extern unsigned long acpi_wakeup_address;
|
|
|
|
/*
|
|
* Record the cpei override flag and current logical cpu. This is
|
|
* useful for CPU removal.
|
|
*/
|
|
extern unsigned int can_cpei_retarget(void);
|
|
extern unsigned int is_cpu_cpei_target(unsigned int cpu);
|
|
extern void set_cpei_target_cpu(unsigned int cpu);
|
|
extern unsigned int get_cpei_target_cpu(void);
|
|
extern void prefill_possible_map(void);
|
|
#ifdef CONFIG_ACPI_HOTPLUG_CPU
|
|
extern int additional_cpus;
|
|
#else
|
|
#define additional_cpus 0
|
|
#endif
|
|
|
|
#ifdef CONFIG_ACPI_NUMA
|
|
#if MAX_NUMNODES > 256
|
|
#define MAX_PXM_DOMAINS MAX_NUMNODES
|
|
#else
|
|
#define MAX_PXM_DOMAINS (256)
|
|
#endif
|
|
extern int __devinitdata pxm_to_nid_map[MAX_PXM_DOMAINS];
|
|
extern int __initdata nid_to_pxm_map[MAX_NUMNODES];
|
|
#endif
|
|
|
|
#define acpi_unlazy_tlb(x)
|
|
|
|
#ifdef CONFIG_ACPI_NUMA
|
|
extern cpumask_t early_cpu_possible_map;
|
|
#define for_each_possible_early_cpu(cpu) \
|
|
for_each_cpu_mask((cpu), early_cpu_possible_map)
|
|
|
|
static inline void per_cpu_scan_finalize(int min_cpus, int reserve_cpus)
|
|
{
|
|
int low_cpu, high_cpu;
|
|
int cpu;
|
|
int next_nid = 0;
|
|
|
|
low_cpu = cpus_weight(early_cpu_possible_map);
|
|
|
|
high_cpu = max(low_cpu, min_cpus);
|
|
high_cpu = min(high_cpu + reserve_cpus, NR_CPUS);
|
|
|
|
for (cpu = low_cpu; cpu < high_cpu; cpu++) {
|
|
cpu_set(cpu, early_cpu_possible_map);
|
|
if (node_cpuid[cpu].nid == NUMA_NO_NODE) {
|
|
node_cpuid[cpu].nid = next_nid;
|
|
next_nid++;
|
|
if (next_nid >= num_online_nodes())
|
|
next_nid = 0;
|
|
}
|
|
}
|
|
}
|
|
#endif /* CONFIG_ACPI_NUMA */
|
|
|
|
#endif /*__KERNEL__*/
|
|
|
|
#endif /*_ASM_ACPI_H*/
|