mirror of
https://mirrors.bfsu.edu.cn/git/linux.git
synced 2024-12-15 15:04:27 +08:00
2874c5fd28
Based on 1 normalized pattern(s): this program is free software you can redistribute it and or modify it under the terms of the gnu general public license as published by the free software foundation either version 2 of the license or at your option any later version extracted by the scancode license scanner the SPDX license identifier GPL-2.0-or-later has been chosen to replace the boilerplate/reference in 3029 file(s). Signed-off-by: Thomas Gleixner <tglx@linutronix.de> Reviewed-by: Allison Randal <allison@lohutok.net> Cc: linux-spdx@vger.kernel.org Link: https://lkml.kernel.org/r/20190527070032.746973796@linutronix.de Signed-off-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
179 lines
4.2 KiB
C
179 lines
4.2 KiB
C
// SPDX-License-Identifier: GPL-2.0-or-later
|
|
/*
|
|
* Copyright (C) 2014 Imagination Technologies
|
|
* Author: Paul Burton <paul.burton@mips.com>
|
|
*/
|
|
|
|
#include <linux/cpu_pm.h>
|
|
#include <linux/cpuidle.h>
|
|
#include <linux/init.h>
|
|
|
|
#include <asm/idle.h>
|
|
#include <asm/pm-cps.h>
|
|
|
|
/* Enumeration of the various idle states this driver may enter */
|
|
enum cps_idle_state {
|
|
STATE_WAIT = 0, /* MIPS wait instruction, coherent */
|
|
STATE_NC_WAIT, /* MIPS wait instruction, non-coherent */
|
|
STATE_CLOCK_GATED, /* Core clock gated */
|
|
STATE_POWER_GATED, /* Core power gated */
|
|
STATE_COUNT
|
|
};
|
|
|
|
static int cps_nc_enter(struct cpuidle_device *dev,
|
|
struct cpuidle_driver *drv, int index)
|
|
{
|
|
enum cps_pm_state pm_state;
|
|
int err;
|
|
|
|
/*
|
|
* At least one core must remain powered up & clocked in order for the
|
|
* system to have any hope of functioning.
|
|
*
|
|
* TODO: don't treat core 0 specially, just prevent the final core
|
|
* TODO: remap interrupt affinity temporarily
|
|
*/
|
|
if (cpus_are_siblings(0, dev->cpu) && (index > STATE_NC_WAIT))
|
|
index = STATE_NC_WAIT;
|
|
|
|
/* Select the appropriate cps_pm_state */
|
|
switch (index) {
|
|
case STATE_NC_WAIT:
|
|
pm_state = CPS_PM_NC_WAIT;
|
|
break;
|
|
case STATE_CLOCK_GATED:
|
|
pm_state = CPS_PM_CLOCK_GATED;
|
|
break;
|
|
case STATE_POWER_GATED:
|
|
pm_state = CPS_PM_POWER_GATED;
|
|
break;
|
|
default:
|
|
BUG();
|
|
return -EINVAL;
|
|
}
|
|
|
|
/* Notify listeners the CPU is about to power down */
|
|
if ((pm_state == CPS_PM_POWER_GATED) && cpu_pm_enter())
|
|
return -EINTR;
|
|
|
|
/* Enter that state */
|
|
err = cps_pm_enter_state(pm_state);
|
|
|
|
/* Notify listeners the CPU is back up */
|
|
if (pm_state == CPS_PM_POWER_GATED)
|
|
cpu_pm_exit();
|
|
|
|
return err ?: index;
|
|
}
|
|
|
|
static struct cpuidle_driver cps_driver = {
|
|
.name = "cpc_cpuidle",
|
|
.owner = THIS_MODULE,
|
|
.states = {
|
|
[STATE_WAIT] = MIPS_CPUIDLE_WAIT_STATE,
|
|
[STATE_NC_WAIT] = {
|
|
.enter = cps_nc_enter,
|
|
.exit_latency = 200,
|
|
.target_residency = 450,
|
|
.name = "nc-wait",
|
|
.desc = "non-coherent MIPS wait",
|
|
},
|
|
[STATE_CLOCK_GATED] = {
|
|
.enter = cps_nc_enter,
|
|
.exit_latency = 300,
|
|
.target_residency = 700,
|
|
.flags = CPUIDLE_FLAG_TIMER_STOP,
|
|
.name = "clock-gated",
|
|
.desc = "core clock gated",
|
|
},
|
|
[STATE_POWER_GATED] = {
|
|
.enter = cps_nc_enter,
|
|
.exit_latency = 600,
|
|
.target_residency = 1000,
|
|
.flags = CPUIDLE_FLAG_TIMER_STOP,
|
|
.name = "power-gated",
|
|
.desc = "core power gated",
|
|
},
|
|
},
|
|
.state_count = STATE_COUNT,
|
|
.safe_state_index = 0,
|
|
};
|
|
|
|
static void __init cps_cpuidle_unregister(void)
|
|
{
|
|
int cpu;
|
|
struct cpuidle_device *device;
|
|
|
|
for_each_possible_cpu(cpu) {
|
|
device = &per_cpu(cpuidle_dev, cpu);
|
|
cpuidle_unregister_device(device);
|
|
}
|
|
|
|
cpuidle_unregister_driver(&cps_driver);
|
|
}
|
|
|
|
static int __init cps_cpuidle_init(void)
|
|
{
|
|
int err, cpu, i;
|
|
struct cpuidle_device *device;
|
|
|
|
/* Detect supported states */
|
|
if (!cps_pm_support_state(CPS_PM_POWER_GATED))
|
|
cps_driver.state_count = STATE_CLOCK_GATED + 1;
|
|
if (!cps_pm_support_state(CPS_PM_CLOCK_GATED))
|
|
cps_driver.state_count = STATE_NC_WAIT + 1;
|
|
if (!cps_pm_support_state(CPS_PM_NC_WAIT))
|
|
cps_driver.state_count = STATE_WAIT + 1;
|
|
|
|
/* Inform the user if some states are unavailable */
|
|
if (cps_driver.state_count < STATE_COUNT) {
|
|
pr_info("cpuidle-cps: limited to ");
|
|
switch (cps_driver.state_count - 1) {
|
|
case STATE_WAIT:
|
|
pr_cont("coherent wait\n");
|
|
break;
|
|
case STATE_NC_WAIT:
|
|
pr_cont("non-coherent wait\n");
|
|
break;
|
|
case STATE_CLOCK_GATED:
|
|
pr_cont("clock gating\n");
|
|
break;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* Set the coupled flag on the appropriate states if this system
|
|
* requires it.
|
|
*/
|
|
if (coupled_coherence)
|
|
for (i = STATE_NC_WAIT; i < cps_driver.state_count; i++)
|
|
cps_driver.states[i].flags |= CPUIDLE_FLAG_COUPLED;
|
|
|
|
err = cpuidle_register_driver(&cps_driver);
|
|
if (err) {
|
|
pr_err("Failed to register CPS cpuidle driver\n");
|
|
return err;
|
|
}
|
|
|
|
for_each_possible_cpu(cpu) {
|
|
device = &per_cpu(cpuidle_dev, cpu);
|
|
device->cpu = cpu;
|
|
#ifdef CONFIG_ARCH_NEEDS_CPU_IDLE_COUPLED
|
|
cpumask_copy(&device->coupled_cpus, &cpu_sibling_map[cpu]);
|
|
#endif
|
|
|
|
err = cpuidle_register_device(device);
|
|
if (err) {
|
|
pr_err("Failed to register CPU%d cpuidle device\n",
|
|
cpu);
|
|
goto err_out;
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
err_out:
|
|
cps_cpuidle_unregister();
|
|
return err;
|
|
}
|
|
device_initcall(cps_cpuidle_init);
|