mirror of
https://mirrors.bfsu.edu.cn/git/linux.git
synced 2024-12-25 12:04:46 +08:00
5727347180
On ARM, the PMU does not stop counting after an overflow and therefore
IRQ latency affects the new counter value read by the kernel. This is
significant for non-sampling runs where it is possible for the new value
to overtake the previous one, causing the delta to be out by up to
max_period events.
Commit a737823d
("ARM: 6835/1: perf: ensure overflows aren't missed due
to IRQ latency") attempted to fix this problem by allowing interrupt
handlers to pass an overflow flag to the event update function, causing
the overflow calculation to assume that the counter passed through zero
when going from prev to new. Unfortunately, this doesn't work when
overflow occurs on the perf_task_tick path because we have the flag
cleared and end up computing a large negative delta.
This patch removes the overflow flag from armpmu_event_update and
instead limits the sample_period to half of the max_period for
non-sampling profiling runs.
Cc: <stable@vger.kernel.org>
Signed-off-by: Ming Lei <ming.lei@canonical.com>
Signed-off-by: Will Deacon <will.deacon@arm.com>
Signed-off-by: Russell King <rmk+kernel@arm.linux.org.uk>
146 lines
3.7 KiB
C
146 lines
3.7 KiB
C
/*
|
|
* linux/arch/arm/include/asm/pmu.h
|
|
*
|
|
* Copyright (C) 2009 picoChip Designs Ltd, Jamie Iles
|
|
*
|
|
* This program is free software; you can redistribute it and/or modify
|
|
* it under the terms of the GNU General Public License version 2 as
|
|
* published by the Free Software Foundation.
|
|
*
|
|
*/
|
|
|
|
#ifndef __ARM_PMU_H__
|
|
#define __ARM_PMU_H__
|
|
|
|
#include <linux/interrupt.h>
|
|
#include <linux/perf_event.h>
|
|
|
|
/*
|
|
* Types of PMUs that can be accessed directly and require mutual
|
|
* exclusion between profiling tools.
|
|
*/
|
|
enum arm_pmu_type {
|
|
ARM_PMU_DEVICE_CPU = 0,
|
|
ARM_NUM_PMU_DEVICES,
|
|
};
|
|
|
|
/*
|
|
* struct arm_pmu_platdata - ARM PMU platform data
|
|
*
|
|
* @handle_irq: an optional handler which will be called from the
|
|
* interrupt and passed the address of the low level handler,
|
|
* and can be used to implement any platform specific handling
|
|
* before or after calling it.
|
|
* @enable_irq: an optional handler which will be called after
|
|
* request_irq and be used to handle some platform specific
|
|
* irq enablement
|
|
* @disable_irq: an optional handler which will be called before
|
|
* free_irq and be used to handle some platform specific
|
|
* irq disablement
|
|
*/
|
|
struct arm_pmu_platdata {
|
|
irqreturn_t (*handle_irq)(int irq, void *dev,
|
|
irq_handler_t pmu_handler);
|
|
void (*enable_irq)(int irq);
|
|
void (*disable_irq)(int irq);
|
|
};
|
|
|
|
#ifdef CONFIG_CPU_HAS_PMU
|
|
|
|
/**
|
|
* reserve_pmu() - reserve the hardware performance counters
|
|
*
|
|
* Reserve the hardware performance counters in the system for exclusive use.
|
|
* Returns 0 on success or -EBUSY if the lock is already held.
|
|
*/
|
|
extern int
|
|
reserve_pmu(enum arm_pmu_type type);
|
|
|
|
/**
|
|
* release_pmu() - Relinquish control of the performance counters
|
|
*
|
|
* Release the performance counters and allow someone else to use them.
|
|
*/
|
|
extern void
|
|
release_pmu(enum arm_pmu_type type);
|
|
|
|
#else /* CONFIG_CPU_HAS_PMU */
|
|
|
|
#include <linux/err.h>
|
|
|
|
static inline int
|
|
reserve_pmu(enum arm_pmu_type type)
|
|
{
|
|
return -ENODEV;
|
|
}
|
|
|
|
static inline void
|
|
release_pmu(enum arm_pmu_type type) { }
|
|
|
|
#endif /* CONFIG_CPU_HAS_PMU */
|
|
|
|
#ifdef CONFIG_HW_PERF_EVENTS
|
|
|
|
/* The events for a given PMU register set. */
|
|
struct pmu_hw_events {
|
|
/*
|
|
* The events that are active on the PMU for the given index.
|
|
*/
|
|
struct perf_event **events;
|
|
|
|
/*
|
|
* A 1 bit for an index indicates that the counter is being used for
|
|
* an event. A 0 means that the counter can be used.
|
|
*/
|
|
unsigned long *used_mask;
|
|
|
|
/*
|
|
* Hardware lock to serialize accesses to PMU registers. Needed for the
|
|
* read/modify/write sequences.
|
|
*/
|
|
raw_spinlock_t pmu_lock;
|
|
};
|
|
|
|
struct arm_pmu {
|
|
struct pmu pmu;
|
|
enum arm_perf_pmu_ids id;
|
|
enum arm_pmu_type type;
|
|
cpumask_t active_irqs;
|
|
const char *name;
|
|
irqreturn_t (*handle_irq)(int irq_num, void *dev);
|
|
void (*enable)(struct hw_perf_event *evt, int idx);
|
|
void (*disable)(struct hw_perf_event *evt, int idx);
|
|
int (*get_event_idx)(struct pmu_hw_events *hw_events,
|
|
struct hw_perf_event *hwc);
|
|
int (*set_event_filter)(struct hw_perf_event *evt,
|
|
struct perf_event_attr *attr);
|
|
u32 (*read_counter)(int idx);
|
|
void (*write_counter)(int idx, u32 val);
|
|
void (*start)(void);
|
|
void (*stop)(void);
|
|
void (*reset)(void *);
|
|
int (*map_event)(struct perf_event *event);
|
|
int num_events;
|
|
atomic_t active_events;
|
|
struct mutex reserve_mutex;
|
|
u64 max_period;
|
|
struct platform_device *plat_device;
|
|
struct pmu_hw_events *(*get_hw_events)(void);
|
|
};
|
|
|
|
#define to_arm_pmu(p) (container_of(p, struct arm_pmu, pmu))
|
|
|
|
int __init armpmu_register(struct arm_pmu *armpmu, char *name, int type);
|
|
|
|
u64 armpmu_event_update(struct perf_event *event,
|
|
struct hw_perf_event *hwc,
|
|
int idx);
|
|
|
|
int armpmu_event_set_period(struct perf_event *event,
|
|
struct hw_perf_event *hwc,
|
|
int idx);
|
|
|
|
#endif /* CONFIG_HW_PERF_EVENTS */
|
|
|
|
#endif /* __ARM_PMU_H__ */
|