mirror of
https://mirrors.bfsu.edu.cn/git/linux.git
synced 2024-11-16 08:44:21 +08:00
aa95387774
The CPU hotplug locking was quite messy, with a recursive lock to handle the fact that both the actual up/down sequence wanted to protect itself from being re-entered, but the callbacks that it called also tended to want to protect themselves from CPU events. This splits the lock into two (one to serialize the whole hotplug sequence, the other to protect against the CPU present bitmaps changing). The latter still allows recursive usage because some subsystems (ondemand policy for cpufreq at least) had already gotten too used to the lax locking, but the locking mistakes are hopefully now less fundamental, and we now warn about recursive lock usage when we see it, in the hope that it can be fixed. Signed-off-by: Linus Torvalds <torvalds@osdl.org>
226 lines
4.9 KiB
C
226 lines
4.9 KiB
C
/* CPU control.
|
|
* (C) 2001, 2002, 2003, 2004 Rusty Russell
|
|
*
|
|
* This code is licenced under the GPL.
|
|
*/
|
|
#include <linux/proc_fs.h>
|
|
#include <linux/smp.h>
|
|
#include <linux/init.h>
|
|
#include <linux/notifier.h>
|
|
#include <linux/sched.h>
|
|
#include <linux/unistd.h>
|
|
#include <linux/cpu.h>
|
|
#include <linux/module.h>
|
|
#include <linux/kthread.h>
|
|
#include <linux/stop_machine.h>
|
|
#include <linux/mutex.h>
|
|
|
|
/* This protects CPUs going up and down... */
|
|
static DEFINE_MUTEX(cpu_add_remove_lock);
|
|
static DEFINE_MUTEX(cpu_bitmask_lock);
|
|
|
|
static __cpuinitdata BLOCKING_NOTIFIER_HEAD(cpu_chain);
|
|
|
|
#ifdef CONFIG_HOTPLUG_CPU
|
|
|
|
/* Crappy recursive lock-takers in cpufreq! Complain loudly about idiots */
|
|
static struct task_struct *recursive;
|
|
static int recursive_depth;
|
|
|
|
void lock_cpu_hotplug(void)
|
|
{
|
|
struct task_struct *tsk = current;
|
|
|
|
if (tsk == recursive) {
|
|
static int warnings = 10;
|
|
if (warnings) {
|
|
printk(KERN_ERR "Lukewarm IQ detected in hotplug locking\n");
|
|
WARN_ON(1);
|
|
warnings--;
|
|
}
|
|
recursive_depth++;
|
|
return;
|
|
}
|
|
mutex_lock(&cpu_bitmask_lock);
|
|
recursive = tsk;
|
|
}
|
|
EXPORT_SYMBOL_GPL(lock_cpu_hotplug);
|
|
|
|
void unlock_cpu_hotplug(void)
|
|
{
|
|
WARN_ON(recursive != current);
|
|
if (recursive_depth) {
|
|
recursive_depth--;
|
|
return;
|
|
}
|
|
mutex_unlock(&cpu_bitmask_lock);
|
|
recursive = NULL;
|
|
}
|
|
EXPORT_SYMBOL_GPL(unlock_cpu_hotplug);
|
|
|
|
#endif /* CONFIG_HOTPLUG_CPU */
|
|
|
|
/* Need to know about CPUs going up/down? */
|
|
int __cpuinit register_cpu_notifier(struct notifier_block *nb)
|
|
{
|
|
return blocking_notifier_chain_register(&cpu_chain, nb);
|
|
}
|
|
|
|
#ifdef CONFIG_HOTPLUG_CPU
|
|
|
|
EXPORT_SYMBOL(register_cpu_notifier);
|
|
|
|
void unregister_cpu_notifier(struct notifier_block *nb)
|
|
{
|
|
blocking_notifier_chain_unregister(&cpu_chain, nb);
|
|
}
|
|
EXPORT_SYMBOL(unregister_cpu_notifier);
|
|
|
|
static inline void check_for_tasks(int cpu)
|
|
{
|
|
struct task_struct *p;
|
|
|
|
write_lock_irq(&tasklist_lock);
|
|
for_each_process(p) {
|
|
if (task_cpu(p) == cpu &&
|
|
(!cputime_eq(p->utime, cputime_zero) ||
|
|
!cputime_eq(p->stime, cputime_zero)))
|
|
printk(KERN_WARNING "Task %s (pid = %d) is on cpu %d\
|
|
(state = %ld, flags = %lx) \n",
|
|
p->comm, p->pid, cpu, p->state, p->flags);
|
|
}
|
|
write_unlock_irq(&tasklist_lock);
|
|
}
|
|
|
|
/* Take this CPU down. */
|
|
static int take_cpu_down(void *unused)
|
|
{
|
|
int err;
|
|
|
|
/* Ensure this CPU doesn't handle any more interrupts. */
|
|
err = __cpu_disable();
|
|
if (err < 0)
|
|
return err;
|
|
|
|
/* Force idle task to run as soon as we yield: it should
|
|
immediately notice cpu is offline and die quickly. */
|
|
sched_idle_next();
|
|
return 0;
|
|
}
|
|
|
|
int cpu_down(unsigned int cpu)
|
|
{
|
|
int err;
|
|
struct task_struct *p;
|
|
cpumask_t old_allowed, tmp;
|
|
|
|
mutex_lock(&cpu_add_remove_lock);
|
|
if (num_online_cpus() == 1) {
|
|
err = -EBUSY;
|
|
goto out;
|
|
}
|
|
|
|
if (!cpu_online(cpu)) {
|
|
err = -EINVAL;
|
|
goto out;
|
|
}
|
|
|
|
err = blocking_notifier_call_chain(&cpu_chain, CPU_DOWN_PREPARE,
|
|
(void *)(long)cpu);
|
|
if (err == NOTIFY_BAD) {
|
|
printk("%s: attempt to take down CPU %u failed\n",
|
|
__FUNCTION__, cpu);
|
|
err = -EINVAL;
|
|
goto out;
|
|
}
|
|
|
|
/* Ensure that we are not runnable on dying cpu */
|
|
old_allowed = current->cpus_allowed;
|
|
tmp = CPU_MASK_ALL;
|
|
cpu_clear(cpu, tmp);
|
|
set_cpus_allowed(current, tmp);
|
|
|
|
mutex_lock(&cpu_bitmask_lock);
|
|
p = __stop_machine_run(take_cpu_down, NULL, cpu);
|
|
mutex_unlock(&cpu_bitmask_lock);
|
|
|
|
if (IS_ERR(p)) {
|
|
/* CPU didn't die: tell everyone. Can't complain. */
|
|
if (blocking_notifier_call_chain(&cpu_chain, CPU_DOWN_FAILED,
|
|
(void *)(long)cpu) == NOTIFY_BAD)
|
|
BUG();
|
|
|
|
err = PTR_ERR(p);
|
|
goto out_allowed;
|
|
}
|
|
|
|
if (cpu_online(cpu))
|
|
goto out_thread;
|
|
|
|
/* Wait for it to sleep (leaving idle task). */
|
|
while (!idle_cpu(cpu))
|
|
yield();
|
|
|
|
/* This actually kills the CPU. */
|
|
__cpu_die(cpu);
|
|
|
|
/* Move it here so it can run. */
|
|
kthread_bind(p, get_cpu());
|
|
put_cpu();
|
|
|
|
/* CPU is completely dead: tell everyone. Too late to complain. */
|
|
if (blocking_notifier_call_chain(&cpu_chain, CPU_DEAD,
|
|
(void *)(long)cpu) == NOTIFY_BAD)
|
|
BUG();
|
|
|
|
check_for_tasks(cpu);
|
|
|
|
out_thread:
|
|
err = kthread_stop(p);
|
|
out_allowed:
|
|
set_cpus_allowed(current, old_allowed);
|
|
out:
|
|
mutex_unlock(&cpu_add_remove_lock);
|
|
return err;
|
|
}
|
|
#endif /*CONFIG_HOTPLUG_CPU*/
|
|
|
|
int __devinit cpu_up(unsigned int cpu)
|
|
{
|
|
int ret;
|
|
void *hcpu = (void *)(long)cpu;
|
|
|
|
mutex_lock(&cpu_add_remove_lock);
|
|
if (cpu_online(cpu) || !cpu_present(cpu)) {
|
|
ret = -EINVAL;
|
|
goto out;
|
|
}
|
|
|
|
ret = blocking_notifier_call_chain(&cpu_chain, CPU_UP_PREPARE, hcpu);
|
|
if (ret == NOTIFY_BAD) {
|
|
printk("%s: attempt to bring up CPU %u failed\n",
|
|
__FUNCTION__, cpu);
|
|
ret = -EINVAL;
|
|
goto out_notify;
|
|
}
|
|
|
|
/* Arch-specific enabling code. */
|
|
mutex_lock(&cpu_bitmask_lock);
|
|
ret = __cpu_up(cpu);
|
|
mutex_unlock(&cpu_bitmask_lock);
|
|
if (ret != 0)
|
|
goto out_notify;
|
|
BUG_ON(!cpu_online(cpu));
|
|
|
|
/* Now call notifier in preparation. */
|
|
blocking_notifier_call_chain(&cpu_chain, CPU_ONLINE, hcpu);
|
|
|
|
out_notify:
|
|
if (ret != 0)
|
|
blocking_notifier_call_chain(&cpu_chain,
|
|
CPU_UP_CANCELED, hcpu);
|
|
out:
|
|
mutex_unlock(&cpu_add_remove_lock);
|
|
return ret;
|
|
}
|