mirror of
https://mirrors.bfsu.edu.cn/git/linux.git
synced 2024-11-16 00:34:20 +08:00
sched: clean up __normal_prio() position
clean up: move __normal_prio() in head of normal_prio(). no code changed. Signed-off-by: Ingo Molnar <mingo@elte.hu>
This commit is contained in:
parent
71f8bd4600
commit
14531189f0
@ -671,35 +671,6 @@ static inline void resched_task(struct task_struct *p)
|
||||
|
||||
#include "sched_stats.h"
|
||||
|
||||
/*
|
||||
* __normal_prio - return the priority that is based on the static
|
||||
* priority but is modified by bonuses/penalties.
|
||||
*
|
||||
* We scale the actual sleep average [0 .... MAX_SLEEP_AVG]
|
||||
* into the -5 ... 0 ... +5 bonus/penalty range.
|
||||
*
|
||||
* We use 25% of the full 0...39 priority range so that:
|
||||
*
|
||||
* 1) nice +19 interactive tasks do not preempt nice 0 CPU hogs.
|
||||
* 2) nice -20 CPU hogs do not get preempted by nice 0 tasks.
|
||||
*
|
||||
* Both properties are important to certain workloads.
|
||||
*/
|
||||
|
||||
static inline int __normal_prio(struct task_struct *p)
|
||||
{
|
||||
int bonus, prio;
|
||||
|
||||
bonus = CURRENT_BONUS(p) - MAX_BONUS / 2;
|
||||
|
||||
prio = p->static_prio - bonus;
|
||||
if (prio < MAX_RT_PRIO)
|
||||
prio = MAX_RT_PRIO;
|
||||
if (prio > MAX_PRIO-1)
|
||||
prio = MAX_PRIO-1;
|
||||
return prio;
|
||||
}
|
||||
|
||||
/*
|
||||
* To aid in avoiding the subversion of "niceness" due to uneven distribution
|
||||
* of tasks with abnormal "nice" values across CPUs the contribution that
|
||||
@ -802,6 +773,35 @@ enqueue_task_head(struct task_struct *p, struct prio_array *array)
|
||||
p->array = array;
|
||||
}
|
||||
|
||||
/*
|
||||
* __normal_prio - return the priority that is based on the static
|
||||
* priority but is modified by bonuses/penalties.
|
||||
*
|
||||
* We scale the actual sleep average [0 .... MAX_SLEEP_AVG]
|
||||
* into the -5 ... 0 ... +5 bonus/penalty range.
|
||||
*
|
||||
* We use 25% of the full 0...39 priority range so that:
|
||||
*
|
||||
* 1) nice +19 interactive tasks do not preempt nice 0 CPU hogs.
|
||||
* 2) nice -20 CPU hogs do not get preempted by nice 0 tasks.
|
||||
*
|
||||
* Both properties are important to certain workloads.
|
||||
*/
|
||||
|
||||
static inline int __normal_prio(struct task_struct *p)
|
||||
{
|
||||
int bonus, prio;
|
||||
|
||||
bonus = CURRENT_BONUS(p) - MAX_BONUS / 2;
|
||||
|
||||
prio = p->static_prio - bonus;
|
||||
if (prio < MAX_RT_PRIO)
|
||||
prio = MAX_RT_PRIO;
|
||||
if (prio > MAX_PRIO-1)
|
||||
prio = MAX_PRIO-1;
|
||||
return prio;
|
||||
}
|
||||
|
||||
/*
|
||||
* Calculate the expected normal priority: i.e. priority
|
||||
* without taking RT-inheritance into account. Might be
|
||||
|
Loading…
Reference in New Issue
Block a user