2014-05-05 21:33:50 +08:00
|
|
|
/*
|
2015-09-04 16:17:25 +08:00
|
|
|
* Generic OPP helper interface for CPU device
|
2014-05-05 21:33:50 +08:00
|
|
|
*
|
|
|
|
* Copyright (C) 2009-2014 Texas Instruments Incorporated.
|
|
|
|
* Nishanth Menon
|
|
|
|
* Romit Dasgupta
|
|
|
|
* Kevin Hilman
|
|
|
|
*
|
|
|
|
* This program is free software; you can redistribute it and/or modify
|
|
|
|
* it under the terms of the GNU General Public License version 2 as
|
|
|
|
* published by the Free Software Foundation.
|
|
|
|
*/
|
2015-10-17 12:15:18 +08:00
|
|
|
|
|
|
|
#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
|
|
|
|
|
2015-09-04 16:17:26 +08:00
|
|
|
#include <linux/cpu.h>
|
2014-05-05 21:33:50 +08:00
|
|
|
#include <linux/cpufreq.h>
|
|
|
|
#include <linux/err.h>
|
|
|
|
#include <linux/errno.h>
|
|
|
|
#include <linux/export.h>
|
2015-09-04 16:17:26 +08:00
|
|
|
#include <linux/of.h>
|
2014-05-05 21:33:50 +08:00
|
|
|
#include <linux/slab.h>
|
|
|
|
|
2015-09-04 16:17:26 +08:00
|
|
|
#include "opp.h"
|
|
|
|
|
2015-09-04 16:17:25 +08:00
|
|
|
#ifdef CONFIG_CPU_FREQ
|
2015-09-04 16:17:26 +08:00
|
|
|
|
2014-05-05 21:33:50 +08:00
|
|
|
/**
|
|
|
|
* dev_pm_opp_init_cpufreq_table() - create a cpufreq table for a device
|
|
|
|
* @dev: device for which we do this operation
|
|
|
|
* @table: Cpufreq table returned back to caller
|
|
|
|
*
|
|
|
|
* Generate a cpufreq table for a provided device- this assumes that the
|
|
|
|
* opp list is already initialized and ready for usage.
|
|
|
|
*
|
|
|
|
* This function allocates required memory for the cpufreq table. It is
|
|
|
|
* expected that the caller does the required maintenance such as freeing
|
|
|
|
* the table as required.
|
|
|
|
*
|
|
|
|
* Returns -EINVAL for bad pointers, -ENODEV if the device is not found, -ENOMEM
|
|
|
|
* if no memory available for the operation (table is not populated), returns 0
|
|
|
|
* if successful and table is populated.
|
|
|
|
*
|
|
|
|
* WARNING: It is important for the callers to ensure refreshing their copy of
|
|
|
|
* the table if any of the mentioned functions have been invoked in the interim.
|
|
|
|
*
|
|
|
|
* Locking: The internal device_opp and opp structures are RCU protected.
|
|
|
|
* Since we just use the regular accessor functions to access the internal data
|
|
|
|
* structures, we use RCU read lock inside this function. As a result, users of
|
|
|
|
* this function DONOT need to use explicit locks for invoking.
|
|
|
|
*/
|
|
|
|
int dev_pm_opp_init_cpufreq_table(struct device *dev,
|
|
|
|
struct cpufreq_frequency_table **table)
|
|
|
|
{
|
|
|
|
struct dev_pm_opp *opp;
|
|
|
|
struct cpufreq_frequency_table *freq_table = NULL;
|
|
|
|
int i, max_opps, ret = 0;
|
|
|
|
unsigned long rate;
|
|
|
|
|
|
|
|
rcu_read_lock();
|
|
|
|
|
|
|
|
max_opps = dev_pm_opp_get_opp_count(dev);
|
|
|
|
if (max_opps <= 0) {
|
|
|
|
ret = max_opps ? max_opps : -ENODATA;
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
2014-09-05 11:08:30 +08:00
|
|
|
freq_table = kcalloc((max_opps + 1), sizeof(*freq_table), GFP_ATOMIC);
|
2014-05-05 21:33:50 +08:00
|
|
|
if (!freq_table) {
|
|
|
|
ret = -ENOMEM;
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
|
|
|
for (i = 0, rate = 0; i < max_opps; i++, rate++) {
|
|
|
|
/* find next rate */
|
|
|
|
opp = dev_pm_opp_find_freq_ceil(dev, &rate);
|
|
|
|
if (IS_ERR(opp)) {
|
|
|
|
ret = PTR_ERR(opp);
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
freq_table[i].driver_data = i;
|
|
|
|
freq_table[i].frequency = rate / 1000;
|
2015-07-29 18:53:08 +08:00
|
|
|
|
|
|
|
/* Is Boost/turbo opp ? */
|
|
|
|
if (dev_pm_opp_is_turbo(opp))
|
|
|
|
freq_table[i].flags = CPUFREQ_BOOST_FREQ;
|
2014-05-05 21:33:50 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
freq_table[i].driver_data = i;
|
|
|
|
freq_table[i].frequency = CPUFREQ_TABLE_END;
|
|
|
|
|
|
|
|
*table = &freq_table[0];
|
|
|
|
|
|
|
|
out:
|
|
|
|
rcu_read_unlock();
|
|
|
|
if (ret)
|
|
|
|
kfree(freq_table);
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(dev_pm_opp_init_cpufreq_table);
|
|
|
|
|
|
|
|
/**
|
|
|
|
* dev_pm_opp_free_cpufreq_table() - free the cpufreq table
|
|
|
|
* @dev: device for which we do this operation
|
|
|
|
* @table: table to free
|
|
|
|
*
|
|
|
|
* Free up the table allocated by dev_pm_opp_init_cpufreq_table
|
|
|
|
*/
|
|
|
|
void dev_pm_opp_free_cpufreq_table(struct device *dev,
|
|
|
|
struct cpufreq_frequency_table **table)
|
|
|
|
{
|
|
|
|
if (!table)
|
|
|
|
return;
|
|
|
|
|
|
|
|
kfree(*table);
|
|
|
|
*table = NULL;
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(dev_pm_opp_free_cpufreq_table);
|
2015-09-04 16:17:25 +08:00
|
|
|
#endif /* CONFIG_CPU_FREQ */
|
2015-09-04 16:17:26 +08:00
|
|
|
|
|
|
|
/* Required only for V1 bindings, as v2 can manage it from DT itself */
|
|
|
|
int dev_pm_opp_set_sharing_cpus(struct device *cpu_dev, cpumask_var_t cpumask)
|
|
|
|
{
|
|
|
|
struct device_list_opp *list_dev;
|
|
|
|
struct device_opp *dev_opp;
|
|
|
|
struct device *dev;
|
|
|
|
int cpu, ret = 0;
|
|
|
|
|
PM / OPP: Protect updates to list_dev with mutex
dev_opp_list_lock is used everywhere to protect device and OPP lists,
but dev_pm_opp_set_sharing_cpus() is missed somehow. And instead we used
rcu-lock, which wouldn't help here as we are adding a new list_dev.
This also fixes a problem where we have called kzalloc(..., GFP_KERNEL)
from within rcu-lock, which isn't allowed as kzalloc can sleep when
called with GFP_KERNEL.
With CONFIG_DEBUG_ATOMIC_SLEEP set, we get following lockdep-splat:
include/linux/rcupdate.h:578 Illegal context switch in RCU read-side critical section!
other info that might help us debug this:
rcu_scheduler_active = 1, debug_locks = 0
5 locks held by swapper/0/1:
#0: (&dev->mutex){......}, at: [<c02f68f4>] __driver_attach+0x48/0x98
#1: (&dev->mutex){......}, at: [<c02f6904>] __driver_attach+0x58/0x98
#2: (cpu_hotplug.lock){++++++}, at: [<c00249d0>] get_online_cpus+0x40/0xb0
#3: (subsys mutex#5){+.+.+.}, at: [<c02f4f8c>] subsys_interface_register+0x44/0xdc
#4: (rcu_read_lock){......}, at: [<c0305c80>] dev_pm_opp_set_sharing_cpus+0x0/0x1e4
stack backtrace:
CPU: 1 PID: 1 Comm: swapper/0 Tainted: G W 4.3.0-rc7-00047-g81f5932958a8 #59
Hardware name: SAMSUNG EXYNOS (Flattened Device Tree)
[<c0016874>] (unwind_backtrace) from [<c001355c>] (show_stack+0x10/0x14)
[<c001355c>] (show_stack) from [<c022553c>] (dump_stack+0x94/0xbc)
[<c022553c>] (dump_stack) from [<c004904c>] (___might_sleep+0x24c/0x298)
[<c004904c>] (___might_sleep) from [<c00f07e4>] (kmem_cache_alloc+0xe8/0x164)
[<c00f07e4>] (kmem_cache_alloc) from [<c0305354>] (_add_list_dev+0x30/0x58)
[<c0305354>] (_add_list_dev) from [<c0305d50>] (dev_pm_opp_set_sharing_cpus+0xd0/0x1e4)
[<c0305d50>] (dev_pm_opp_set_sharing_cpus) from [<c040eda4>] (cpufreq_init+0x4cc/0x62c)
[<c040eda4>] (cpufreq_init) from [<c040a964>] (cpufreq_online+0xbc/0x73c)
[<c040a964>] (cpufreq_online) from [<c02f4fe0>] (subsys_interface_register+0x98/0xdc)
[<c02f4fe0>] (subsys_interface_register) from [<c040a640>] (cpufreq_register_driver+0x110/0x17c)
[<c040a640>] (cpufreq_register_driver) from [<c040ef64>] (dt_cpufreq_probe+0x60/0x8c)
[<c040ef64>] (dt_cpufreq_probe) from [<c02f8084>] (platform_drv_probe+0x44/0xa4)
[<c02f8084>] (platform_drv_probe) from [<c02f67c0>] (driver_probe_device+0x208/0x2f4)
[<c02f67c0>] (driver_probe_device) from [<c02f6940>] (__driver_attach+0x94/0x98)
[<c02f6940>] (__driver_attach) from [<c02f4c1c>] (bus_for_each_dev+0x68/0x9c)
Reported-by: Michael Turquette <mturquette@baylibre.com>
Reviewed-by: Stephen Boyd <sboyd@codeaurora.org>
Signed-off-by: Viresh Kumar <viresh.kumar@linaro.org>
Cc: 4.3 <stable@vger.kernel.org> # 4.3
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2015-11-05 16:51:19 +08:00
|
|
|
mutex_lock(&dev_opp_list_lock);
|
2015-09-04 16:17:26 +08:00
|
|
|
|
|
|
|
dev_opp = _find_device_opp(cpu_dev);
|
|
|
|
if (IS_ERR(dev_opp)) {
|
|
|
|
ret = -EINVAL;
|
PM / OPP: Protect updates to list_dev with mutex
dev_opp_list_lock is used everywhere to protect device and OPP lists,
but dev_pm_opp_set_sharing_cpus() is missed somehow. And instead we used
rcu-lock, which wouldn't help here as we are adding a new list_dev.
This also fixes a problem where we have called kzalloc(..., GFP_KERNEL)
from within rcu-lock, which isn't allowed as kzalloc can sleep when
called with GFP_KERNEL.
With CONFIG_DEBUG_ATOMIC_SLEEP set, we get following lockdep-splat:
include/linux/rcupdate.h:578 Illegal context switch in RCU read-side critical section!
other info that might help us debug this:
rcu_scheduler_active = 1, debug_locks = 0
5 locks held by swapper/0/1:
#0: (&dev->mutex){......}, at: [<c02f68f4>] __driver_attach+0x48/0x98
#1: (&dev->mutex){......}, at: [<c02f6904>] __driver_attach+0x58/0x98
#2: (cpu_hotplug.lock){++++++}, at: [<c00249d0>] get_online_cpus+0x40/0xb0
#3: (subsys mutex#5){+.+.+.}, at: [<c02f4f8c>] subsys_interface_register+0x44/0xdc
#4: (rcu_read_lock){......}, at: [<c0305c80>] dev_pm_opp_set_sharing_cpus+0x0/0x1e4
stack backtrace:
CPU: 1 PID: 1 Comm: swapper/0 Tainted: G W 4.3.0-rc7-00047-g81f5932958a8 #59
Hardware name: SAMSUNG EXYNOS (Flattened Device Tree)
[<c0016874>] (unwind_backtrace) from [<c001355c>] (show_stack+0x10/0x14)
[<c001355c>] (show_stack) from [<c022553c>] (dump_stack+0x94/0xbc)
[<c022553c>] (dump_stack) from [<c004904c>] (___might_sleep+0x24c/0x298)
[<c004904c>] (___might_sleep) from [<c00f07e4>] (kmem_cache_alloc+0xe8/0x164)
[<c00f07e4>] (kmem_cache_alloc) from [<c0305354>] (_add_list_dev+0x30/0x58)
[<c0305354>] (_add_list_dev) from [<c0305d50>] (dev_pm_opp_set_sharing_cpus+0xd0/0x1e4)
[<c0305d50>] (dev_pm_opp_set_sharing_cpus) from [<c040eda4>] (cpufreq_init+0x4cc/0x62c)
[<c040eda4>] (cpufreq_init) from [<c040a964>] (cpufreq_online+0xbc/0x73c)
[<c040a964>] (cpufreq_online) from [<c02f4fe0>] (subsys_interface_register+0x98/0xdc)
[<c02f4fe0>] (subsys_interface_register) from [<c040a640>] (cpufreq_register_driver+0x110/0x17c)
[<c040a640>] (cpufreq_register_driver) from [<c040ef64>] (dt_cpufreq_probe+0x60/0x8c)
[<c040ef64>] (dt_cpufreq_probe) from [<c02f8084>] (platform_drv_probe+0x44/0xa4)
[<c02f8084>] (platform_drv_probe) from [<c02f67c0>] (driver_probe_device+0x208/0x2f4)
[<c02f67c0>] (driver_probe_device) from [<c02f6940>] (__driver_attach+0x94/0x98)
[<c02f6940>] (__driver_attach) from [<c02f4c1c>] (bus_for_each_dev+0x68/0x9c)
Reported-by: Michael Turquette <mturquette@baylibre.com>
Reviewed-by: Stephen Boyd <sboyd@codeaurora.org>
Signed-off-by: Viresh Kumar <viresh.kumar@linaro.org>
Cc: 4.3 <stable@vger.kernel.org> # 4.3
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2015-11-05 16:51:19 +08:00
|
|
|
goto unlock;
|
2015-09-04 16:17:26 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
for_each_cpu(cpu, cpumask) {
|
|
|
|
if (cpu == cpu_dev->id)
|
|
|
|
continue;
|
|
|
|
|
|
|
|
dev = get_cpu_device(cpu);
|
|
|
|
if (!dev) {
|
|
|
|
dev_err(cpu_dev, "%s: failed to get cpu%d device\n",
|
|
|
|
__func__, cpu);
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
|
|
|
list_dev = _add_list_dev(dev, dev_opp);
|
|
|
|
if (!list_dev) {
|
|
|
|
dev_err(dev, "%s: failed to add list-dev for cpu%d device\n",
|
|
|
|
__func__, cpu);
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
}
|
PM / OPP: Protect updates to list_dev with mutex
dev_opp_list_lock is used everywhere to protect device and OPP lists,
but dev_pm_opp_set_sharing_cpus() is missed somehow. And instead we used
rcu-lock, which wouldn't help here as we are adding a new list_dev.
This also fixes a problem where we have called kzalloc(..., GFP_KERNEL)
from within rcu-lock, which isn't allowed as kzalloc can sleep when
called with GFP_KERNEL.
With CONFIG_DEBUG_ATOMIC_SLEEP set, we get following lockdep-splat:
include/linux/rcupdate.h:578 Illegal context switch in RCU read-side critical section!
other info that might help us debug this:
rcu_scheduler_active = 1, debug_locks = 0
5 locks held by swapper/0/1:
#0: (&dev->mutex){......}, at: [<c02f68f4>] __driver_attach+0x48/0x98
#1: (&dev->mutex){......}, at: [<c02f6904>] __driver_attach+0x58/0x98
#2: (cpu_hotplug.lock){++++++}, at: [<c00249d0>] get_online_cpus+0x40/0xb0
#3: (subsys mutex#5){+.+.+.}, at: [<c02f4f8c>] subsys_interface_register+0x44/0xdc
#4: (rcu_read_lock){......}, at: [<c0305c80>] dev_pm_opp_set_sharing_cpus+0x0/0x1e4
stack backtrace:
CPU: 1 PID: 1 Comm: swapper/0 Tainted: G W 4.3.0-rc7-00047-g81f5932958a8 #59
Hardware name: SAMSUNG EXYNOS (Flattened Device Tree)
[<c0016874>] (unwind_backtrace) from [<c001355c>] (show_stack+0x10/0x14)
[<c001355c>] (show_stack) from [<c022553c>] (dump_stack+0x94/0xbc)
[<c022553c>] (dump_stack) from [<c004904c>] (___might_sleep+0x24c/0x298)
[<c004904c>] (___might_sleep) from [<c00f07e4>] (kmem_cache_alloc+0xe8/0x164)
[<c00f07e4>] (kmem_cache_alloc) from [<c0305354>] (_add_list_dev+0x30/0x58)
[<c0305354>] (_add_list_dev) from [<c0305d50>] (dev_pm_opp_set_sharing_cpus+0xd0/0x1e4)
[<c0305d50>] (dev_pm_opp_set_sharing_cpus) from [<c040eda4>] (cpufreq_init+0x4cc/0x62c)
[<c040eda4>] (cpufreq_init) from [<c040a964>] (cpufreq_online+0xbc/0x73c)
[<c040a964>] (cpufreq_online) from [<c02f4fe0>] (subsys_interface_register+0x98/0xdc)
[<c02f4fe0>] (subsys_interface_register) from [<c040a640>] (cpufreq_register_driver+0x110/0x17c)
[<c040a640>] (cpufreq_register_driver) from [<c040ef64>] (dt_cpufreq_probe+0x60/0x8c)
[<c040ef64>] (dt_cpufreq_probe) from [<c02f8084>] (platform_drv_probe+0x44/0xa4)
[<c02f8084>] (platform_drv_probe) from [<c02f67c0>] (driver_probe_device+0x208/0x2f4)
[<c02f67c0>] (driver_probe_device) from [<c02f6940>] (__driver_attach+0x94/0x98)
[<c02f6940>] (__driver_attach) from [<c02f4c1c>] (bus_for_each_dev+0x68/0x9c)
Reported-by: Michael Turquette <mturquette@baylibre.com>
Reviewed-by: Stephen Boyd <sboyd@codeaurora.org>
Signed-off-by: Viresh Kumar <viresh.kumar@linaro.org>
Cc: 4.3 <stable@vger.kernel.org> # 4.3
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2015-11-05 16:51:19 +08:00
|
|
|
unlock:
|
|
|
|
mutex_unlock(&dev_opp_list_lock);
|
2015-09-04 16:17:26 +08:00
|
|
|
|
2015-11-05 16:51:18 +08:00
|
|
|
return ret;
|
2015-09-04 16:17:26 +08:00
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(dev_pm_opp_set_sharing_cpus);
|
|
|
|
|
|
|
|
#ifdef CONFIG_OF
|
|
|
|
void dev_pm_opp_of_cpumask_remove_table(cpumask_var_t cpumask)
|
|
|
|
{
|
|
|
|
struct device *cpu_dev;
|
|
|
|
int cpu;
|
|
|
|
|
|
|
|
WARN_ON(cpumask_empty(cpumask));
|
|
|
|
|
|
|
|
for_each_cpu(cpu, cpumask) {
|
|
|
|
cpu_dev = get_cpu_device(cpu);
|
|
|
|
if (!cpu_dev) {
|
|
|
|
pr_err("%s: failed to get cpu%d device\n", __func__,
|
|
|
|
cpu);
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
|
|
|
dev_pm_opp_of_remove_table(cpu_dev);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(dev_pm_opp_of_cpumask_remove_table);
|
|
|
|
|
|
|
|
int dev_pm_opp_of_cpumask_add_table(cpumask_var_t cpumask)
|
|
|
|
{
|
|
|
|
struct device *cpu_dev;
|
|
|
|
int cpu, ret = 0;
|
|
|
|
|
|
|
|
WARN_ON(cpumask_empty(cpumask));
|
|
|
|
|
|
|
|
for_each_cpu(cpu, cpumask) {
|
|
|
|
cpu_dev = get_cpu_device(cpu);
|
|
|
|
if (!cpu_dev) {
|
|
|
|
pr_err("%s: failed to get cpu%d device\n", __func__,
|
|
|
|
cpu);
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
|
|
|
ret = dev_pm_opp_of_add_table(cpu_dev);
|
|
|
|
if (ret) {
|
|
|
|
pr_err("%s: couldn't find opp table for cpu:%d, %d\n",
|
|
|
|
__func__, cpu, ret);
|
|
|
|
|
|
|
|
/* Free all other OPPs */
|
|
|
|
dev_pm_opp_of_cpumask_remove_table(cpumask);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(dev_pm_opp_of_cpumask_add_table);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Works only for OPP v2 bindings.
|
|
|
|
*
|
|
|
|
* Returns -ENOENT if operating-points-v2 bindings aren't supported.
|
|
|
|
*/
|
|
|
|
int dev_pm_opp_of_get_sharing_cpus(struct device *cpu_dev, cpumask_var_t cpumask)
|
|
|
|
{
|
|
|
|
struct device_node *np, *tmp_np;
|
|
|
|
struct device *tcpu_dev;
|
|
|
|
int cpu, ret = 0;
|
|
|
|
|
|
|
|
/* Get OPP descriptor node */
|
|
|
|
np = _of_get_opp_desc_node(cpu_dev);
|
|
|
|
if (!np) {
|
2015-09-22 00:26:02 +08:00
|
|
|
dev_dbg(cpu_dev, "%s: Couldn't find cpu_dev node.\n", __func__);
|
2015-09-04 16:17:26 +08:00
|
|
|
return -ENOENT;
|
|
|
|
}
|
|
|
|
|
2015-12-28 21:06:17 +08:00
|
|
|
cpumask_set_cpu(cpu_dev->id, cpumask);
|
|
|
|
|
2015-09-04 16:17:26 +08:00
|
|
|
/* OPPs are shared ? */
|
|
|
|
if (!of_property_read_bool(np, "opp-shared"))
|
|
|
|
goto put_cpu_node;
|
|
|
|
|
|
|
|
for_each_possible_cpu(cpu) {
|
|
|
|
if (cpu == cpu_dev->id)
|
|
|
|
continue;
|
|
|
|
|
|
|
|
tcpu_dev = get_cpu_device(cpu);
|
|
|
|
if (!tcpu_dev) {
|
|
|
|
dev_err(cpu_dev, "%s: failed to get cpu%d device\n",
|
|
|
|
__func__, cpu);
|
|
|
|
ret = -ENODEV;
|
|
|
|
goto put_cpu_node;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Get OPP descriptor node */
|
|
|
|
tmp_np = _of_get_opp_desc_node(tcpu_dev);
|
|
|
|
if (!tmp_np) {
|
2015-09-22 00:26:02 +08:00
|
|
|
dev_err(tcpu_dev, "%s: Couldn't find tcpu_dev node.\n",
|
|
|
|
__func__);
|
|
|
|
ret = -ENOENT;
|
2015-09-04 16:17:26 +08:00
|
|
|
goto put_cpu_node;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* CPUs are sharing opp node */
|
|
|
|
if (np == tmp_np)
|
|
|
|
cpumask_set_cpu(cpu, cpumask);
|
|
|
|
|
|
|
|
of_node_put(tmp_np);
|
|
|
|
}
|
|
|
|
|
|
|
|
put_cpu_node:
|
|
|
|
of_node_put(np);
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(dev_pm_opp_of_get_sharing_cpus);
|
|
|
|
#endif
|