mirror of
https://mirrors.bfsu.edu.cn/git/linux.git
synced 2024-11-11 04:18:39 +08:00
08a36a4854
There is no architecture-specific code or data left that generic <linux/vtime.h> needs to know about. Thus, avoid the inclusion of <asm/vtime.h> header. Signed-off-by: Alexander Gordeev <agordeev@linux.ibm.com> Signed-off-by: Ingo Molnar <mingo@kernel.org> Reviewed-by: Frederic Weisbecker <frederic@kernel.org> Acked-by: Nicholas Piggin <npiggin@gmail.com> Link: https://lore.kernel.org/r/f7cd245668b9ae61a55184871aec494ec9199c4a.1712760275.git.agordeev@linux.ibm.com
158 lines
4.5 KiB
C
158 lines
4.5 KiB
C
/* SPDX-License-Identifier: GPL-2.0 */
|
|
#ifndef _LINUX_KERNEL_VTIME_H
|
|
#define _LINUX_KERNEL_VTIME_H
|
|
|
|
#include <linux/context_tracking_state.h>
|
|
#include <linux/sched.h>
|
|
|
|
/*
|
|
* Common vtime APIs
|
|
*/
|
|
#ifdef CONFIG_VIRT_CPU_ACCOUNTING
|
|
extern void vtime_account_kernel(struct task_struct *tsk);
|
|
extern void vtime_account_idle(struct task_struct *tsk);
|
|
#endif /* !CONFIG_VIRT_CPU_ACCOUNTING */
|
|
|
|
#ifdef CONFIG_VIRT_CPU_ACCOUNTING_GEN
|
|
extern void vtime_user_enter(struct task_struct *tsk);
|
|
extern void vtime_user_exit(struct task_struct *tsk);
|
|
extern void vtime_guest_enter(struct task_struct *tsk);
|
|
extern void vtime_guest_exit(struct task_struct *tsk);
|
|
extern void vtime_init_idle(struct task_struct *tsk, int cpu);
|
|
#else /* !CONFIG_VIRT_CPU_ACCOUNTING_GEN */
|
|
static inline void vtime_user_enter(struct task_struct *tsk) { }
|
|
static inline void vtime_user_exit(struct task_struct *tsk) { }
|
|
static inline void vtime_guest_enter(struct task_struct *tsk) { }
|
|
static inline void vtime_guest_exit(struct task_struct *tsk) { }
|
|
static inline void vtime_init_idle(struct task_struct *tsk, int cpu) { }
|
|
#endif
|
|
|
|
#ifdef CONFIG_VIRT_CPU_ACCOUNTING_NATIVE
|
|
extern void vtime_account_irq(struct task_struct *tsk, unsigned int offset);
|
|
extern void vtime_account_softirq(struct task_struct *tsk);
|
|
extern void vtime_account_hardirq(struct task_struct *tsk);
|
|
extern void vtime_flush(struct task_struct *tsk);
|
|
#else /* !CONFIG_VIRT_CPU_ACCOUNTING_NATIVE */
|
|
static inline void vtime_account_irq(struct task_struct *tsk, unsigned int offset) { }
|
|
static inline void vtime_account_softirq(struct task_struct *tsk) { }
|
|
static inline void vtime_account_hardirq(struct task_struct *tsk) { }
|
|
static inline void vtime_flush(struct task_struct *tsk) { }
|
|
#endif
|
|
|
|
/*
|
|
* vtime_accounting_enabled_this_cpu() definitions/declarations
|
|
*/
|
|
#if defined(CONFIG_VIRT_CPU_ACCOUNTING_NATIVE)
|
|
|
|
static inline bool vtime_accounting_enabled_this_cpu(void) { return true; }
|
|
extern void vtime_task_switch(struct task_struct *prev);
|
|
|
|
static __always_inline void vtime_account_guest_enter(void)
|
|
{
|
|
vtime_account_kernel(current);
|
|
current->flags |= PF_VCPU;
|
|
}
|
|
|
|
static __always_inline void vtime_account_guest_exit(void)
|
|
{
|
|
vtime_account_kernel(current);
|
|
current->flags &= ~PF_VCPU;
|
|
}
|
|
|
|
#elif defined(CONFIG_VIRT_CPU_ACCOUNTING_GEN)
|
|
|
|
/*
|
|
* Checks if vtime is enabled on some CPU. Cputime readers want to be careful
|
|
* in that case and compute the tickless cputime.
|
|
* For now vtime state is tied to context tracking. We might want to decouple
|
|
* those later if necessary.
|
|
*/
|
|
static inline bool vtime_accounting_enabled(void)
|
|
{
|
|
return context_tracking_enabled();
|
|
}
|
|
|
|
static inline bool vtime_accounting_enabled_cpu(int cpu)
|
|
{
|
|
return context_tracking_enabled_cpu(cpu);
|
|
}
|
|
|
|
static inline bool vtime_accounting_enabled_this_cpu(void)
|
|
{
|
|
return context_tracking_enabled_this_cpu();
|
|
}
|
|
|
|
extern void vtime_task_switch_generic(struct task_struct *prev);
|
|
|
|
static inline void vtime_task_switch(struct task_struct *prev)
|
|
{
|
|
if (vtime_accounting_enabled_this_cpu())
|
|
vtime_task_switch_generic(prev);
|
|
}
|
|
|
|
static __always_inline void vtime_account_guest_enter(void)
|
|
{
|
|
if (vtime_accounting_enabled_this_cpu())
|
|
vtime_guest_enter(current);
|
|
else
|
|
current->flags |= PF_VCPU;
|
|
}
|
|
|
|
static __always_inline void vtime_account_guest_exit(void)
|
|
{
|
|
if (vtime_accounting_enabled_this_cpu())
|
|
vtime_guest_exit(current);
|
|
else
|
|
current->flags &= ~PF_VCPU;
|
|
}
|
|
|
|
#else /* !CONFIG_VIRT_CPU_ACCOUNTING */
|
|
|
|
static inline bool vtime_accounting_enabled_this_cpu(void) { return false; }
|
|
static inline void vtime_task_switch(struct task_struct *prev) { }
|
|
|
|
static __always_inline void vtime_account_guest_enter(void)
|
|
{
|
|
current->flags |= PF_VCPU;
|
|
}
|
|
|
|
static __always_inline void vtime_account_guest_exit(void)
|
|
{
|
|
current->flags &= ~PF_VCPU;
|
|
}
|
|
|
|
#endif
|
|
|
|
|
|
#ifdef CONFIG_IRQ_TIME_ACCOUNTING
|
|
extern void irqtime_account_irq(struct task_struct *tsk, unsigned int offset);
|
|
#else
|
|
static inline void irqtime_account_irq(struct task_struct *tsk, unsigned int offset) { }
|
|
#endif
|
|
|
|
static inline void account_softirq_enter(struct task_struct *tsk)
|
|
{
|
|
vtime_account_irq(tsk, SOFTIRQ_OFFSET);
|
|
irqtime_account_irq(tsk, SOFTIRQ_OFFSET);
|
|
}
|
|
|
|
static inline void account_softirq_exit(struct task_struct *tsk)
|
|
{
|
|
vtime_account_softirq(tsk);
|
|
irqtime_account_irq(tsk, 0);
|
|
}
|
|
|
|
static inline void account_hardirq_enter(struct task_struct *tsk)
|
|
{
|
|
vtime_account_irq(tsk, HARDIRQ_OFFSET);
|
|
irqtime_account_irq(tsk, HARDIRQ_OFFSET);
|
|
}
|
|
|
|
static inline void account_hardirq_exit(struct task_struct *tsk)
|
|
{
|
|
vtime_account_hardirq(tsk);
|
|
irqtime_account_irq(tsk, 0);
|
|
}
|
|
|
|
#endif /* _LINUX_KERNEL_VTIME_H */
|