mirror of
https://mirrors.bfsu.edu.cn/git/linux.git
synced 2024-12-05 01:54:09 +08:00
timekeeping: Provide multi-timestamp accessor to NMI safe timekeeper
printk wants to store various timestamps (MONOTONIC, REALTIME, BOOTTIME) to make correlation of dmesg from several systems easier. Provide an interface to retrieve all three timestamps in one go. There are some caveats: 1) Boot time and late sleep time injection Boot time is a racy access on 32bit systems if the sleep time injection happens late during resume and not in timekeeping_resume(). That could be avoided by expanding struct tk_read_base with boot offset for 32bit and adding more overhead to the update. As this is a hard to observe once per resume event which can be filtered with reasonable effort using the accurate mono/real timestamps, it's probably not worth the trouble. Aside of that it might be possible on 32 and 64 bit to observe the following when the sleep time injection happens late: CPU 0 CPU 1 timekeeping_resume() ktime_get_fast_timestamps() mono, real = __ktime_get_real_fast() inject_sleep_time() update boot offset boot = mono + bootoffset; That means that boot time already has the sleep time adjustment, but real time does not. On the next readout both are in sync again. Preventing this for 64bit is not really feasible without destroying the careful cache layout of the timekeeper because the sequence count and struct tk_read_base would then need two cache lines instead of one. 2) Suspend/resume timestamps Access to the time keeper clock source is disabled accross the innermost steps of suspend/resume. The accessors still work, but the timestamps are frozen until time keeping is resumed which happens very early. For regular suspend/resume there is no observable difference vs. sched clock, but it might affect some of the nasty low level debug printks. OTOH, access to sched clock is not guaranteed accross suspend/resume on all systems either so it depends on the hardware in use. If that turns out to be a real problem then this could be mitigated by using sched clock in a similar way as during early boot. But it's not as trivial as on early boot because it needs some careful protection against the clock monotonic timestamp jumping backwards on resume. Signed-off-by: Thomas Gleixner <tglx@linutronix.de> Tested-by: Petr Mladek <pmladek@suse.com> Link: https://lore.kernel.org/r/20200814115512.159981360@linutronix.de
This commit is contained in:
parent
71419b30ca
commit
e2d977c9f1
@ -222,6 +222,18 @@ extern bool timekeeping_rtc_skipresume(void);
|
|||||||
|
|
||||||
extern void timekeeping_inject_sleeptime64(const struct timespec64 *delta);
|
extern void timekeeping_inject_sleeptime64(const struct timespec64 *delta);
|
||||||
|
|
||||||
|
/*
|
||||||
|
* struct ktime_timestanps - Simultaneous mono/boot/real timestamps
|
||||||
|
* @mono: Monotonic timestamp
|
||||||
|
* @boot: Boottime timestamp
|
||||||
|
* @real: Realtime timestamp
|
||||||
|
*/
|
||||||
|
struct ktime_timestamps {
|
||||||
|
u64 mono;
|
||||||
|
u64 boot;
|
||||||
|
u64 real;
|
||||||
|
};
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* struct system_time_snapshot - simultaneous raw/real time capture with
|
* struct system_time_snapshot - simultaneous raw/real time capture with
|
||||||
* counter value
|
* counter value
|
||||||
@ -280,6 +292,9 @@ extern int get_device_system_crosststamp(
|
|||||||
*/
|
*/
|
||||||
extern void ktime_get_snapshot(struct system_time_snapshot *systime_snapshot);
|
extern void ktime_get_snapshot(struct system_time_snapshot *systime_snapshot);
|
||||||
|
|
||||||
|
/* NMI safe mono/boot/realtime timestamps */
|
||||||
|
extern void ktime_get_fast_timestamps(struct ktime_timestamps *snap);
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Persistent clock related interfaces
|
* Persistent clock related interfaces
|
||||||
*/
|
*/
|
||||||
|
@ -530,29 +530,29 @@ u64 notrace ktime_get_boot_fast_ns(void)
|
|||||||
}
|
}
|
||||||
EXPORT_SYMBOL_GPL(ktime_get_boot_fast_ns);
|
EXPORT_SYMBOL_GPL(ktime_get_boot_fast_ns);
|
||||||
|
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* See comment for __ktime_get_fast_ns() vs. timestamp ordering
|
* See comment for __ktime_get_fast_ns() vs. timestamp ordering
|
||||||
*/
|
*/
|
||||||
static __always_inline u64 __ktime_get_real_fast_ns(struct tk_fast *tkf)
|
static __always_inline u64 __ktime_get_real_fast(struct tk_fast *tkf, u64 *mono)
|
||||||
{
|
{
|
||||||
struct tk_read_base *tkr;
|
struct tk_read_base *tkr;
|
||||||
|
u64 basem, baser, delta;
|
||||||
unsigned int seq;
|
unsigned int seq;
|
||||||
u64 now;
|
|
||||||
|
|
||||||
do {
|
do {
|
||||||
seq = raw_read_seqcount_latch(&tkf->seq);
|
seq = raw_read_seqcount_latch(&tkf->seq);
|
||||||
tkr = tkf->base + (seq & 0x01);
|
tkr = tkf->base + (seq & 0x01);
|
||||||
now = ktime_to_ns(tkr->base_real);
|
basem = ktime_to_ns(tkr->base);
|
||||||
|
baser = ktime_to_ns(tkr->base_real);
|
||||||
|
|
||||||
now += timekeeping_delta_to_ns(tkr,
|
delta = timekeeping_delta_to_ns(tkr,
|
||||||
clocksource_delta(
|
clocksource_delta(tk_clock_read(tkr),
|
||||||
tk_clock_read(tkr),
|
tkr->cycle_last, tkr->mask));
|
||||||
tkr->cycle_last,
|
|
||||||
tkr->mask));
|
|
||||||
} while (read_seqcount_retry(&tkf->seq, seq));
|
} while (read_seqcount_retry(&tkf->seq, seq));
|
||||||
|
|
||||||
return now;
|
if (mono)
|
||||||
|
*mono = basem + delta;
|
||||||
|
return baser + delta;
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
@ -560,10 +560,64 @@ static __always_inline u64 __ktime_get_real_fast_ns(struct tk_fast *tkf)
|
|||||||
*/
|
*/
|
||||||
u64 ktime_get_real_fast_ns(void)
|
u64 ktime_get_real_fast_ns(void)
|
||||||
{
|
{
|
||||||
return __ktime_get_real_fast_ns(&tk_fast_mono);
|
return __ktime_get_real_fast(&tk_fast_mono, NULL);
|
||||||
}
|
}
|
||||||
EXPORT_SYMBOL_GPL(ktime_get_real_fast_ns);
|
EXPORT_SYMBOL_GPL(ktime_get_real_fast_ns);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* ktime_get_fast_timestamps: - NMI safe timestamps
|
||||||
|
* @snapshot: Pointer to timestamp storage
|
||||||
|
*
|
||||||
|
* Stores clock monotonic, boottime and realtime timestamps.
|
||||||
|
*
|
||||||
|
* Boot time is a racy access on 32bit systems if the sleep time injection
|
||||||
|
* happens late during resume and not in timekeeping_resume(). That could
|
||||||
|
* be avoided by expanding struct tk_read_base with boot offset for 32bit
|
||||||
|
* and adding more overhead to the update. As this is a hard to observe
|
||||||
|
* once per resume event which can be filtered with reasonable effort using
|
||||||
|
* the accurate mono/real timestamps, it's probably not worth the trouble.
|
||||||
|
*
|
||||||
|
* Aside of that it might be possible on 32 and 64 bit to observe the
|
||||||
|
* following when the sleep time injection happens late:
|
||||||
|
*
|
||||||
|
* CPU 0 CPU 1
|
||||||
|
* timekeeping_resume()
|
||||||
|
* ktime_get_fast_timestamps()
|
||||||
|
* mono, real = __ktime_get_real_fast()
|
||||||
|
* inject_sleep_time()
|
||||||
|
* update boot offset
|
||||||
|
* boot = mono + bootoffset;
|
||||||
|
*
|
||||||
|
* That means that boot time already has the sleep time adjustment, but
|
||||||
|
* real time does not. On the next readout both are in sync again.
|
||||||
|
*
|
||||||
|
* Preventing this for 64bit is not really feasible without destroying the
|
||||||
|
* careful cache layout of the timekeeper because the sequence count and
|
||||||
|
* struct tk_read_base would then need two cache lines instead of one.
|
||||||
|
*
|
||||||
|
* Access to the time keeper clock source is disabled accross the innermost
|
||||||
|
* steps of suspend/resume. The accessors still work, but the timestamps
|
||||||
|
* are frozen until time keeping is resumed which happens very early.
|
||||||
|
*
|
||||||
|
* For regular suspend/resume there is no observable difference vs. sched
|
||||||
|
* clock, but it might affect some of the nasty low level debug printks.
|
||||||
|
*
|
||||||
|
* OTOH, access to sched clock is not guaranteed accross suspend/resume on
|
||||||
|
* all systems either so it depends on the hardware in use.
|
||||||
|
*
|
||||||
|
* If that turns out to be a real problem then this could be mitigated by
|
||||||
|
* using sched clock in a similar way as during early boot. But it's not as
|
||||||
|
* trivial as on early boot because it needs some careful protection
|
||||||
|
* against the clock monotonic timestamp jumping backwards on resume.
|
||||||
|
*/
|
||||||
|
void ktime_get_fast_timestamps(struct ktime_timestamps *snapshot)
|
||||||
|
{
|
||||||
|
struct timekeeper *tk = &tk_core.timekeeper;
|
||||||
|
|
||||||
|
snapshot->real = __ktime_get_real_fast(&tk_fast_mono, &snapshot->mono);
|
||||||
|
snapshot->boot = snapshot->mono + ktime_to_ns(data_race(tk->offs_boot));
|
||||||
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* halt_fast_timekeeper - Prevent fast timekeeper from accessing clocksource.
|
* halt_fast_timekeeper - Prevent fast timekeeper from accessing clocksource.
|
||||||
* @tk: Timekeeper to snapshot.
|
* @tk: Timekeeper to snapshot.
|
||||||
|
Loading…
Reference in New Issue
Block a user