mirror of
https://github.com/edk2-porting/linux-next.git
synced 2024-12-22 20:23:57 +08:00
18f2190d79
Add PPU event-based and cycle-based profiling support to Oprofile for Cell. Oprofile is expected to collect data on all CPUs simultaneously. However, there is one set of performance counters per node. There are two hardware threads or virtual CPUs on each node. Hence, OProfile must multiplex in time the performance counter collection on the two virtual CPUs. The multiplexing of the performance counters is done by a virtual counter routine. Initially, the counters are configured to collect data on the even CPUs in the system, one CPU per node. In order to capture the PC for the virtual CPU when the performance counter interrupt occurs (the specified number of events between samples has occurred), the even processors are configured to handle the performance counter interrupts for their node. The virtual counter routine is called via a kernel timer after the virtual sample time. The routine stops the counters, saves the current counts, loads the last counts for the other virtual CPU on the node, sets interrupts to be handled by the other virtual CPU and restarts the counters, the virtual timer routine is scheduled to run again. The virtual sample time is kept relatively small to make sure sampling occurs on both CPUs on the node with a relatively small granularity. Whenever the counters overflow, the performance counter interrupt is called to collect the PC for the CPU where data is being collected. The oprofile driver relies on a firmware RTAS call to setup the debug bus to route the desired signals to the performance counter hardware to be counted. The RTAS call must set the routing registers appropriately in each of the islands to pass the signals down the debug bus as well as routing the signals from a particular island onto the bus. There is a second firmware RTAS call to reset the debug bus to the non pass thru state when the counters are not in use. Signed-off-by: Carl Love <carll@us.ibm.com> Signed-off-by: Maynard Johnson <mpjohn@us.ibm.com> Signed-off-by: Arnd Bergmann <arnd.bergmann@de.ibm.com> Signed-off-by: Paul Mackerras <paulus@samba.org>
216 lines
4.0 KiB
C
216 lines
4.0 KiB
C
/*
|
|
* Copyright (C) 2004 Anton Blanchard <anton@au.ibm.com>, IBM
|
|
*
|
|
* Based on alpha version.
|
|
*
|
|
* This program is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU General Public License
|
|
* as published by the Free Software Foundation; either version
|
|
* 2 of the License, or (at your option) any later version.
|
|
*/
|
|
|
|
#ifndef _ASM_POWERPC_OPROFILE_IMPL_H
|
|
#define _ASM_POWERPC_OPROFILE_IMPL_H
|
|
#ifdef __KERNEL__
|
|
|
|
#define OP_MAX_COUNTER 8
|
|
|
|
/* Per-counter configuration as set via oprofilefs. */
|
|
struct op_counter_config {
|
|
unsigned long enabled;
|
|
unsigned long event;
|
|
unsigned long count;
|
|
/* Classic doesn't support per-counter user/kernel selection */
|
|
unsigned long kernel;
|
|
unsigned long user;
|
|
unsigned long unit_mask;
|
|
};
|
|
|
|
/* System-wide configuration as set via oprofilefs. */
|
|
struct op_system_config {
|
|
#ifdef CONFIG_PPC64
|
|
unsigned long mmcr0;
|
|
unsigned long mmcr1;
|
|
unsigned long mmcra;
|
|
#endif
|
|
unsigned long enable_kernel;
|
|
unsigned long enable_user;
|
|
};
|
|
|
|
/* Per-arch configuration */
|
|
struct op_powerpc_model {
|
|
void (*reg_setup) (struct op_counter_config *,
|
|
struct op_system_config *,
|
|
int num_counters);
|
|
void (*cpu_setup) (struct op_counter_config *);
|
|
void (*start) (struct op_counter_config *);
|
|
void (*global_start) (struct op_counter_config *);
|
|
void (*stop) (void);
|
|
void (*global_stop) (void);
|
|
void (*handle_interrupt) (struct pt_regs *,
|
|
struct op_counter_config *);
|
|
int num_counters;
|
|
};
|
|
|
|
extern struct op_powerpc_model op_model_fsl_booke;
|
|
extern struct op_powerpc_model op_model_rs64;
|
|
extern struct op_powerpc_model op_model_power4;
|
|
extern struct op_powerpc_model op_model_7450;
|
|
extern struct op_powerpc_model op_model_cell;
|
|
|
|
#ifndef CONFIG_FSL_BOOKE
|
|
|
|
/* All the classic PPC parts use these */
|
|
static inline unsigned int ctr_read(unsigned int i)
|
|
{
|
|
switch(i) {
|
|
case 0:
|
|
return mfspr(SPRN_PMC1);
|
|
case 1:
|
|
return mfspr(SPRN_PMC2);
|
|
case 2:
|
|
return mfspr(SPRN_PMC3);
|
|
case 3:
|
|
return mfspr(SPRN_PMC4);
|
|
case 4:
|
|
return mfspr(SPRN_PMC5);
|
|
case 5:
|
|
return mfspr(SPRN_PMC6);
|
|
|
|
/* No PPC32 chip has more than 6 so far */
|
|
#ifdef CONFIG_PPC64
|
|
case 6:
|
|
return mfspr(SPRN_PMC7);
|
|
case 7:
|
|
return mfspr(SPRN_PMC8);
|
|
#endif
|
|
default:
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
static inline void ctr_write(unsigned int i, unsigned int val)
|
|
{
|
|
switch(i) {
|
|
case 0:
|
|
mtspr(SPRN_PMC1, val);
|
|
break;
|
|
case 1:
|
|
mtspr(SPRN_PMC2, val);
|
|
break;
|
|
case 2:
|
|
mtspr(SPRN_PMC3, val);
|
|
break;
|
|
case 3:
|
|
mtspr(SPRN_PMC4, val);
|
|
break;
|
|
case 4:
|
|
mtspr(SPRN_PMC5, val);
|
|
break;
|
|
case 5:
|
|
mtspr(SPRN_PMC6, val);
|
|
break;
|
|
|
|
/* No PPC32 chip has more than 6, yet */
|
|
#ifdef CONFIG_PPC64
|
|
case 6:
|
|
mtspr(SPRN_PMC7, val);
|
|
break;
|
|
case 7:
|
|
mtspr(SPRN_PMC8, val);
|
|
break;
|
|
#endif
|
|
default:
|
|
break;
|
|
}
|
|
}
|
|
#else /* CONFIG_FSL_BOOKE */
|
|
static inline u32 get_pmlca(int ctr)
|
|
{
|
|
u32 pmlca;
|
|
|
|
switch (ctr) {
|
|
case 0:
|
|
pmlca = mfpmr(PMRN_PMLCA0);
|
|
break;
|
|
case 1:
|
|
pmlca = mfpmr(PMRN_PMLCA1);
|
|
break;
|
|
case 2:
|
|
pmlca = mfpmr(PMRN_PMLCA2);
|
|
break;
|
|
case 3:
|
|
pmlca = mfpmr(PMRN_PMLCA3);
|
|
break;
|
|
default:
|
|
panic("Bad ctr number\n");
|
|
}
|
|
|
|
return pmlca;
|
|
}
|
|
|
|
static inline void set_pmlca(int ctr, u32 pmlca)
|
|
{
|
|
switch (ctr) {
|
|
case 0:
|
|
mtpmr(PMRN_PMLCA0, pmlca);
|
|
break;
|
|
case 1:
|
|
mtpmr(PMRN_PMLCA1, pmlca);
|
|
break;
|
|
case 2:
|
|
mtpmr(PMRN_PMLCA2, pmlca);
|
|
break;
|
|
case 3:
|
|
mtpmr(PMRN_PMLCA3, pmlca);
|
|
break;
|
|
default:
|
|
panic("Bad ctr number\n");
|
|
}
|
|
}
|
|
|
|
static inline unsigned int ctr_read(unsigned int i)
|
|
{
|
|
switch(i) {
|
|
case 0:
|
|
return mfpmr(PMRN_PMC0);
|
|
case 1:
|
|
return mfpmr(PMRN_PMC1);
|
|
case 2:
|
|
return mfpmr(PMRN_PMC2);
|
|
case 3:
|
|
return mfpmr(PMRN_PMC3);
|
|
default:
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
static inline void ctr_write(unsigned int i, unsigned int val)
|
|
{
|
|
switch(i) {
|
|
case 0:
|
|
mtpmr(PMRN_PMC0, val);
|
|
break;
|
|
case 1:
|
|
mtpmr(PMRN_PMC1, val);
|
|
break;
|
|
case 2:
|
|
mtpmr(PMRN_PMC2, val);
|
|
break;
|
|
case 3:
|
|
mtpmr(PMRN_PMC3, val);
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
}
|
|
|
|
|
|
#endif /* CONFIG_FSL_BOOKE */
|
|
|
|
|
|
extern void op_powerpc_backtrace(struct pt_regs * const regs, unsigned int depth);
|
|
|
|
#endif /* __KERNEL__ */
|
|
#endif /* _ASM_POWERPC_OPROFILE_IMPL_H */
|