mirror of
https://mirrors.bfsu.edu.cn/git/linux.git
synced 2024-12-18 08:35:08 +08:00
cb9c448c66
Incorporates the new performance counter reservation system in oprofile. Also cleans up a lot of the initialization code. The code original zero'd out every register associated with performance counters regardless if those registers were used or not. This causes issues with the nmi watchdog. Now oprofile tries to reserve registers and gives up if it can't get them. Cc: levon@movementarian.org Cc: oprofile-list@lists.sf.net Signed-off-by: Don Zickus <dzickus@redhat.com> Signed-off-by: Andi Kleen <ak@suse.de>
181 lines
4.4 KiB
C
181 lines
4.4 KiB
C
/**
|
|
* @file op_model_athlon.h
|
|
* athlon / K7 model-specific MSR operations
|
|
*
|
|
* @remark Copyright 2002 OProfile authors
|
|
* @remark Read the file COPYING
|
|
*
|
|
* @author John Levon
|
|
* @author Philippe Elie
|
|
* @author Graydon Hoare
|
|
*/
|
|
|
|
#include <linux/oprofile.h>
|
|
#include <asm/ptrace.h>
|
|
#include <asm/msr.h>
|
|
#include <asm/nmi.h>
|
|
|
|
#include "op_x86_model.h"
|
|
#include "op_counter.h"
|
|
|
|
#define NUM_COUNTERS 4
|
|
#define NUM_CONTROLS 4
|
|
|
|
#define CTR_IS_RESERVED(msrs,c) (msrs->counters[(c)].addr ? 1 : 0)
|
|
#define CTR_READ(l,h,msrs,c) do {rdmsr(msrs->counters[(c)].addr, (l), (h));} while (0)
|
|
#define CTR_WRITE(l,msrs,c) do {wrmsr(msrs->counters[(c)].addr, -(unsigned int)(l), -1);} while (0)
|
|
#define CTR_OVERFLOWED(n) (!((n) & (1U<<31)))
|
|
|
|
#define CTRL_IS_RESERVED(msrs,c) (msrs->controls[(c)].addr ? 1 : 0)
|
|
#define CTRL_READ(l,h,msrs,c) do {rdmsr(msrs->controls[(c)].addr, (l), (h));} while (0)
|
|
#define CTRL_WRITE(l,h,msrs,c) do {wrmsr(msrs->controls[(c)].addr, (l), (h));} while (0)
|
|
#define CTRL_SET_ACTIVE(n) (n |= (1<<22))
|
|
#define CTRL_SET_INACTIVE(n) (n &= ~(1<<22))
|
|
#define CTRL_CLEAR(x) (x &= (1<<21))
|
|
#define CTRL_SET_ENABLE(val) (val |= 1<<20)
|
|
#define CTRL_SET_USR(val,u) (val |= ((u & 1) << 16))
|
|
#define CTRL_SET_KERN(val,k) (val |= ((k & 1) << 17))
|
|
#define CTRL_SET_UM(val, m) (val |= (m << 8))
|
|
#define CTRL_SET_EVENT(val, e) (val |= e)
|
|
|
|
static unsigned long reset_value[NUM_COUNTERS];
|
|
|
|
static void athlon_fill_in_addresses(struct op_msrs * const msrs)
|
|
{
|
|
int i;
|
|
|
|
for (i=0; i < NUM_COUNTERS; i++) {
|
|
if (reserve_perfctr_nmi(MSR_K7_PERFCTR0 + i))
|
|
msrs->counters[i].addr = MSR_K7_PERFCTR0 + i;
|
|
else
|
|
msrs->counters[i].addr = 0;
|
|
}
|
|
|
|
for (i=0; i < NUM_CONTROLS; i++) {
|
|
if (reserve_evntsel_nmi(MSR_K7_EVNTSEL0 + i))
|
|
msrs->controls[i].addr = MSR_K7_EVNTSEL0 + i;
|
|
else
|
|
msrs->controls[i].addr = 0;
|
|
}
|
|
}
|
|
|
|
|
|
static void athlon_setup_ctrs(struct op_msrs const * const msrs)
|
|
{
|
|
unsigned int low, high;
|
|
int i;
|
|
|
|
/* clear all counters */
|
|
for (i = 0 ; i < NUM_CONTROLS; ++i) {
|
|
if (unlikely(!CTRL_IS_RESERVED(msrs,i)))
|
|
continue;
|
|
CTRL_READ(low, high, msrs, i);
|
|
CTRL_CLEAR(low);
|
|
CTRL_WRITE(low, high, msrs, i);
|
|
}
|
|
|
|
/* avoid a false detection of ctr overflows in NMI handler */
|
|
for (i = 0; i < NUM_COUNTERS; ++i) {
|
|
if (unlikely(!CTR_IS_RESERVED(msrs,i)))
|
|
continue;
|
|
CTR_WRITE(1, msrs, i);
|
|
}
|
|
|
|
/* enable active counters */
|
|
for (i = 0; i < NUM_COUNTERS; ++i) {
|
|
if ((counter_config[i].enabled) && (CTR_IS_RESERVED(msrs,i))) {
|
|
reset_value[i] = counter_config[i].count;
|
|
|
|
CTR_WRITE(counter_config[i].count, msrs, i);
|
|
|
|
CTRL_READ(low, high, msrs, i);
|
|
CTRL_CLEAR(low);
|
|
CTRL_SET_ENABLE(low);
|
|
CTRL_SET_USR(low, counter_config[i].user);
|
|
CTRL_SET_KERN(low, counter_config[i].kernel);
|
|
CTRL_SET_UM(low, counter_config[i].unit_mask);
|
|
CTRL_SET_EVENT(low, counter_config[i].event);
|
|
CTRL_WRITE(low, high, msrs, i);
|
|
} else {
|
|
reset_value[i] = 0;
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
static int athlon_check_ctrs(struct pt_regs * const regs,
|
|
struct op_msrs const * const msrs)
|
|
{
|
|
unsigned int low, high;
|
|
int i;
|
|
|
|
for (i = 0 ; i < NUM_COUNTERS; ++i) {
|
|
if (!reset_value[i])
|
|
continue;
|
|
CTR_READ(low, high, msrs, i);
|
|
if (CTR_OVERFLOWED(low)) {
|
|
oprofile_add_sample(regs, i);
|
|
CTR_WRITE(reset_value[i], msrs, i);
|
|
}
|
|
}
|
|
|
|
/* See op_model_ppro.c */
|
|
return 1;
|
|
}
|
|
|
|
|
|
static void athlon_start(struct op_msrs const * const msrs)
|
|
{
|
|
unsigned int low, high;
|
|
int i;
|
|
for (i = 0 ; i < NUM_COUNTERS ; ++i) {
|
|
if (reset_value[i]) {
|
|
CTRL_READ(low, high, msrs, i);
|
|
CTRL_SET_ACTIVE(low);
|
|
CTRL_WRITE(low, high, msrs, i);
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
static void athlon_stop(struct op_msrs const * const msrs)
|
|
{
|
|
unsigned int low,high;
|
|
int i;
|
|
|
|
/* Subtle: stop on all counters to avoid race with
|
|
* setting our pm callback */
|
|
for (i = 0 ; i < NUM_COUNTERS ; ++i) {
|
|
if (!reset_value[i])
|
|
continue;
|
|
CTRL_READ(low, high, msrs, i);
|
|
CTRL_SET_INACTIVE(low);
|
|
CTRL_WRITE(low, high, msrs, i);
|
|
}
|
|
}
|
|
|
|
static void athlon_shutdown(struct op_msrs const * const msrs)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0 ; i < NUM_COUNTERS ; ++i) {
|
|
if (CTR_IS_RESERVED(msrs,i))
|
|
release_perfctr_nmi(MSR_K7_PERFCTR0 + i);
|
|
}
|
|
for (i = 0 ; i < NUM_CONTROLS ; ++i) {
|
|
if (CTRL_IS_RESERVED(msrs,i))
|
|
release_evntsel_nmi(MSR_K7_EVNTSEL0 + i);
|
|
}
|
|
}
|
|
|
|
struct op_x86_model_spec const op_athlon_spec = {
|
|
.num_counters = NUM_COUNTERS,
|
|
.num_controls = NUM_CONTROLS,
|
|
.fill_in_addresses = &athlon_fill_in_addresses,
|
|
.setup_ctrs = &athlon_setup_ctrs,
|
|
.check_ctrs = &athlon_check_ctrs,
|
|
.start = &athlon_start,
|
|
.stop = &athlon_stop,
|
|
.shutdown = &athlon_shutdown
|
|
};
|