2019-05-19 20:08:55 +08:00
|
|
|
// SPDX-License-Identifier: GPL-2.0-only
|
2009-01-10 04:27:08 +08:00
|
|
|
/*
|
|
|
|
* Uniprocessor-only support functions. The counterpart to kernel/smp.c
|
|
|
|
*/
|
|
|
|
|
2009-01-12 23:04:37 +08:00
|
|
|
#include <linux/interrupt.h>
|
2009-01-10 04:27:08 +08:00
|
|
|
#include <linux/kernel.h>
|
2011-05-24 02:51:41 +08:00
|
|
|
#include <linux/export.h>
|
2009-01-10 04:27:08 +08:00
|
|
|
#include <linux/smp.h>
|
2016-08-29 14:48:43 +08:00
|
|
|
#include <linux/hypervisor.h>
|
2009-01-10 04:27:08 +08:00
|
|
|
|
|
|
|
int smp_call_function_single(int cpu, void (*func) (void *info), void *info,
|
|
|
|
int wait)
|
|
|
|
{
|
2013-09-12 05:23:25 +08:00
|
|
|
unsigned long flags;
|
|
|
|
|
2020-02-05 22:34:09 +08:00
|
|
|
if (cpu != 0)
|
|
|
|
return -ENXIO;
|
2009-01-11 12:15:21 +08:00
|
|
|
|
2013-09-12 05:23:25 +08:00
|
|
|
local_irq_save(flags);
|
|
|
|
func(info);
|
|
|
|
local_irq_restore(flags);
|
2009-01-11 12:15:21 +08:00
|
|
|
|
2009-01-10 04:27:08 +08:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(smp_call_function_single);
|
2013-09-12 05:23:24 +08:00
|
|
|
|
2023-08-31 14:31:28 +08:00
|
|
|
int smp_call_function_single_async(int cpu, call_single_data_t *csd)
|
2013-11-15 06:32:08 +08:00
|
|
|
{
|
|
|
|
unsigned long flags;
|
|
|
|
|
|
|
|
local_irq_save(flags);
|
|
|
|
csd->func(csd->info);
|
|
|
|
local_irq_restore(flags);
|
2014-02-24 23:39:57 +08:00
|
|
|
return 0;
|
2013-11-15 06:32:08 +08:00
|
|
|
}
|
2014-02-24 23:40:02 +08:00
|
|
|
EXPORT_SYMBOL(smp_call_function_single_async);
|
2013-11-15 06:32:08 +08:00
|
|
|
|
2013-09-12 05:23:24 +08:00
|
|
|
/*
|
|
|
|
* Preemption is disabled here to make sure the cond_func is called under the
|
2021-05-07 09:06:33 +08:00
|
|
|
* same conditions in UP and SMP.
|
2013-09-12 05:23:24 +08:00
|
|
|
*/
|
2020-01-17 17:01:35 +08:00
|
|
|
void on_each_cpu_cond_mask(smp_cond_func_t cond_func, smp_call_func_t func,
|
2020-01-17 17:01:37 +08:00
|
|
|
void *info, bool wait, const struct cpumask *mask)
|
2013-09-12 05:23:24 +08:00
|
|
|
{
|
|
|
|
unsigned long flags;
|
|
|
|
|
|
|
|
preempt_disable();
|
2021-02-21 07:17:12 +08:00
|
|
|
if ((!cond_func || cond_func(0, info)) && cpumask_test_cpu(0, mask)) {
|
2013-09-12 05:23:24 +08:00
|
|
|
local_irq_save(flags);
|
|
|
|
func(info);
|
|
|
|
local_irq_restore(flags);
|
|
|
|
}
|
|
|
|
preempt_enable();
|
|
|
|
}
|
2018-09-26 11:58:41 +08:00
|
|
|
EXPORT_SYMBOL(on_each_cpu_cond_mask);
|
|
|
|
|
2016-08-29 14:48:44 +08:00
|
|
|
int smp_call_on_cpu(unsigned int cpu, int (*func)(void *), void *par, bool phys)
|
|
|
|
{
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
if (cpu != 0)
|
|
|
|
return -ENXIO;
|
|
|
|
|
|
|
|
if (phys)
|
|
|
|
hypervisor_pin_vcpu(0);
|
|
|
|
ret = func(par);
|
|
|
|
if (phys)
|
|
|
|
hypervisor_pin_vcpu(-1);
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(smp_call_on_cpu);
|