2017-04-18 21:46:11 +08:00
|
|
|
#include <errno.h>
|
2017-04-18 02:23:08 +08:00
|
|
|
#include <inttypes.h>
|
2012-09-17 16:31:14 +08:00
|
|
|
#include <math.h>
|
|
|
|
#include "stat.h"
|
2015-06-26 17:29:16 +08:00
|
|
|
#include "evlist.h"
|
2015-06-04 21:50:55 +08:00
|
|
|
#include "evsel.h"
|
2015-06-26 17:29:16 +08:00
|
|
|
#include "thread_map.h"
|
2012-09-17 16:31:14 +08:00
|
|
|
|
|
|
|
void update_stats(struct stats *stats, u64 val)
|
|
|
|
{
|
|
|
|
double delta;
|
|
|
|
|
|
|
|
stats->n++;
|
|
|
|
delta = val - stats->mean;
|
|
|
|
stats->mean += delta / stats->n;
|
|
|
|
stats->M2 += delta*(val - stats->mean);
|
2013-08-03 04:05:40 +08:00
|
|
|
|
|
|
|
if (val > stats->max)
|
|
|
|
stats->max = val;
|
|
|
|
|
|
|
|
if (val < stats->min)
|
|
|
|
stats->min = val;
|
2012-09-17 16:31:14 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
double avg_stats(struct stats *stats)
|
|
|
|
{
|
|
|
|
return stats->mean;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* http://en.wikipedia.org/wiki/Algorithms_for_calculating_variance
|
|
|
|
*
|
|
|
|
* (\Sum n_i^2) - ((\Sum n_i)^2)/n
|
|
|
|
* s^2 = -------------------------------
|
|
|
|
* n - 1
|
|
|
|
*
|
|
|
|
* http://en.wikipedia.org/wiki/Stddev
|
|
|
|
*
|
|
|
|
* The std dev of the mean is related to the std dev by:
|
|
|
|
*
|
|
|
|
* s
|
|
|
|
* s_mean = -------
|
|
|
|
* sqrt(n)
|
|
|
|
*
|
|
|
|
*/
|
|
|
|
double stddev_stats(struct stats *stats)
|
|
|
|
{
|
|
|
|
double variance, variance_mean;
|
|
|
|
|
2013-05-26 08:24:48 +08:00
|
|
|
if (stats->n < 2)
|
2012-09-17 16:31:14 +08:00
|
|
|
return 0.0;
|
|
|
|
|
|
|
|
variance = stats->M2 / (stats->n - 1);
|
|
|
|
variance_mean = variance / stats->n;
|
|
|
|
|
|
|
|
return sqrt(variance_mean);
|
|
|
|
}
|
|
|
|
|
|
|
|
double rel_stddev_stats(double stddev, double avg)
|
|
|
|
{
|
|
|
|
double pct = 0.0;
|
|
|
|
|
|
|
|
if (avg)
|
|
|
|
pct = 100.0 * stddev/avg;
|
|
|
|
|
|
|
|
return pct;
|
|
|
|
}
|
2015-06-04 21:50:55 +08:00
|
|
|
|
|
|
|
bool __perf_evsel_stat__is(struct perf_evsel *evsel,
|
|
|
|
enum perf_stat_evsel_id id)
|
|
|
|
{
|
2017-10-27 01:22:34 +08:00
|
|
|
struct perf_stat_evsel *ps = evsel->stats;
|
2015-06-04 21:50:55 +08:00
|
|
|
|
|
|
|
return ps->id == id;
|
|
|
|
}
|
|
|
|
|
|
|
|
#define ID(id, name) [PERF_STAT_EVSEL_ID__##id] = #name
|
|
|
|
static const char *id_str[PERF_STAT_EVSEL_ID__MAX] = {
|
2015-06-03 22:25:52 +08:00
|
|
|
ID(NONE, x),
|
|
|
|
ID(CYCLES_IN_TX, cpu/cycles-t/),
|
|
|
|
ID(TRANSACTION_START, cpu/tx-start/),
|
|
|
|
ID(ELISION_START, cpu/el-start/),
|
|
|
|
ID(CYCLES_IN_TX_CP, cpu/cycles-ct/),
|
2016-05-25 03:52:37 +08:00
|
|
|
ID(TOPDOWN_TOTAL_SLOTS, topdown-total-slots),
|
|
|
|
ID(TOPDOWN_SLOTS_ISSUED, topdown-slots-issued),
|
|
|
|
ID(TOPDOWN_SLOTS_RETIRED, topdown-slots-retired),
|
|
|
|
ID(TOPDOWN_FETCH_BUBBLES, topdown-fetch-bubbles),
|
|
|
|
ID(TOPDOWN_RECOVERY_BUBBLES, topdown-recovery-bubbles),
|
2017-05-27 03:05:38 +08:00
|
|
|
ID(SMI_NUM, msr/smi/),
|
|
|
|
ID(APERF, msr/aperf/),
|
2015-06-04 21:50:55 +08:00
|
|
|
};
|
|
|
|
#undef ID
|
|
|
|
|
|
|
|
void perf_stat_evsel_id_init(struct perf_evsel *evsel)
|
|
|
|
{
|
2017-10-27 01:22:34 +08:00
|
|
|
struct perf_stat_evsel *ps = evsel->stats;
|
2015-06-04 21:50:55 +08:00
|
|
|
int i;
|
|
|
|
|
|
|
|
/* ps->id is 0 hence PERF_STAT_EVSEL_ID__NONE by default */
|
|
|
|
|
|
|
|
for (i = 0; i < PERF_STAT_EVSEL_ID__MAX; i++) {
|
|
|
|
if (!strcmp(perf_evsel__name(evsel), id_str[i])) {
|
|
|
|
ps->id = i;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
2015-06-14 16:19:26 +08:00
|
|
|
|
2016-01-20 19:56:35 +08:00
|
|
|
static void perf_evsel__reset_stat_priv(struct perf_evsel *evsel)
|
2015-06-26 17:29:14 +08:00
|
|
|
{
|
|
|
|
int i;
|
2017-10-27 01:22:34 +08:00
|
|
|
struct perf_stat_evsel *ps = evsel->stats;
|
2015-06-26 17:29:14 +08:00
|
|
|
|
|
|
|
for (i = 0; i < 3; i++)
|
|
|
|
init_stats(&ps->res_stats[i]);
|
|
|
|
|
|
|
|
perf_stat_evsel_id_init(evsel);
|
|
|
|
}
|
|
|
|
|
2016-01-20 19:56:35 +08:00
|
|
|
static int perf_evsel__alloc_stat_priv(struct perf_evsel *evsel)
|
2015-06-26 17:29:14 +08:00
|
|
|
{
|
2017-10-27 01:22:34 +08:00
|
|
|
evsel->stats = zalloc(sizeof(struct perf_stat_evsel));
|
|
|
|
if (evsel->stats == NULL)
|
2015-06-26 17:29:14 +08:00
|
|
|
return -ENOMEM;
|
|
|
|
perf_evsel__reset_stat_priv(evsel);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2016-01-20 19:56:35 +08:00
|
|
|
static void perf_evsel__free_stat_priv(struct perf_evsel *evsel)
|
2015-06-26 17:29:14 +08:00
|
|
|
{
|
2017-10-27 01:22:34 +08:00
|
|
|
struct perf_stat_evsel *ps = evsel->stats;
|
2017-07-26 20:02:05 +08:00
|
|
|
|
|
|
|
if (ps)
|
|
|
|
free(ps->group_data);
|
2017-10-27 01:22:34 +08:00
|
|
|
zfree(&evsel->stats);
|
2015-06-26 17:29:14 +08:00
|
|
|
}
|
2015-06-26 17:29:15 +08:00
|
|
|
|
2016-01-20 19:56:35 +08:00
|
|
|
static int perf_evsel__alloc_prev_raw_counts(struct perf_evsel *evsel,
|
|
|
|
int ncpus, int nthreads)
|
2015-06-26 17:29:15 +08:00
|
|
|
{
|
|
|
|
struct perf_counts *counts;
|
|
|
|
|
|
|
|
counts = perf_counts__new(ncpus, nthreads);
|
|
|
|
if (counts)
|
|
|
|
evsel->prev_raw_counts = counts;
|
|
|
|
|
|
|
|
return counts ? 0 : -ENOMEM;
|
|
|
|
}
|
|
|
|
|
2016-01-20 19:56:35 +08:00
|
|
|
static void perf_evsel__free_prev_raw_counts(struct perf_evsel *evsel)
|
2015-06-26 17:29:15 +08:00
|
|
|
{
|
|
|
|
perf_counts__delete(evsel->prev_raw_counts);
|
|
|
|
evsel->prev_raw_counts = NULL;
|
|
|
|
}
|
2015-06-26 17:29:16 +08:00
|
|
|
|
2016-01-20 19:56:35 +08:00
|
|
|
static int perf_evsel__alloc_stats(struct perf_evsel *evsel, bool alloc_raw)
|
2015-06-26 17:29:17 +08:00
|
|
|
{
|
|
|
|
int ncpus = perf_evsel__nr_cpus(evsel);
|
|
|
|
int nthreads = thread_map__nr(evsel->threads);
|
|
|
|
|
|
|
|
if (perf_evsel__alloc_stat_priv(evsel) < 0 ||
|
|
|
|
perf_evsel__alloc_counts(evsel, ncpus, nthreads) < 0 ||
|
|
|
|
(alloc_raw && perf_evsel__alloc_prev_raw_counts(evsel, ncpus, nthreads) < 0))
|
|
|
|
return -ENOMEM;
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2015-06-26 17:29:16 +08:00
|
|
|
int perf_evlist__alloc_stats(struct perf_evlist *evlist, bool alloc_raw)
|
|
|
|
{
|
|
|
|
struct perf_evsel *evsel;
|
|
|
|
|
2016-06-23 22:26:15 +08:00
|
|
|
evlist__for_each_entry(evlist, evsel) {
|
2015-06-26 17:29:17 +08:00
|
|
|
if (perf_evsel__alloc_stats(evsel, alloc_raw))
|
2015-06-26 17:29:16 +08:00
|
|
|
goto out_free;
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
out_free:
|
|
|
|
perf_evlist__free_stats(evlist);
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
|
|
|
|
void perf_evlist__free_stats(struct perf_evlist *evlist)
|
|
|
|
{
|
|
|
|
struct perf_evsel *evsel;
|
|
|
|
|
2016-06-23 22:26:15 +08:00
|
|
|
evlist__for_each_entry(evlist, evsel) {
|
2015-06-26 17:29:16 +08:00
|
|
|
perf_evsel__free_stat_priv(evsel);
|
|
|
|
perf_evsel__free_counts(evsel);
|
|
|
|
perf_evsel__free_prev_raw_counts(evsel);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
void perf_evlist__reset_stats(struct perf_evlist *evlist)
|
|
|
|
{
|
|
|
|
struct perf_evsel *evsel;
|
|
|
|
|
2016-06-23 22:26:15 +08:00
|
|
|
evlist__for_each_entry(evlist, evsel) {
|
2015-06-26 17:29:16 +08:00
|
|
|
perf_evsel__reset_stat_priv(evsel);
|
|
|
|
perf_evsel__reset_counts(evsel);
|
|
|
|
}
|
|
|
|
}
|
2015-07-21 20:31:27 +08:00
|
|
|
|
|
|
|
static void zero_per_pkg(struct perf_evsel *counter)
|
|
|
|
{
|
|
|
|
if (counter->per_pkg_mask)
|
|
|
|
memset(counter->per_pkg_mask, 0, MAX_NR_CPUS);
|
|
|
|
}
|
|
|
|
|
2015-09-03 21:23:40 +08:00
|
|
|
static int check_per_pkg(struct perf_evsel *counter,
|
|
|
|
struct perf_counts_values *vals, int cpu, bool *skip)
|
2015-07-21 20:31:27 +08:00
|
|
|
{
|
|
|
|
unsigned long *mask = counter->per_pkg_mask;
|
|
|
|
struct cpu_map *cpus = perf_evsel__cpus(counter);
|
|
|
|
int s;
|
|
|
|
|
|
|
|
*skip = false;
|
|
|
|
|
|
|
|
if (!counter->per_pkg)
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
if (cpu_map__empty(cpus))
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
if (!mask) {
|
|
|
|
mask = zalloc(MAX_NR_CPUS);
|
|
|
|
if (!mask)
|
|
|
|
return -ENOMEM;
|
|
|
|
|
|
|
|
counter->per_pkg_mask = mask;
|
|
|
|
}
|
|
|
|
|
2015-09-03 21:23:40 +08:00
|
|
|
/*
|
|
|
|
* we do not consider an event that has not run as a good
|
|
|
|
* instance to mark a package as used (skip=1). Otherwise
|
|
|
|
* we may run into a situation where the first CPU in a package
|
|
|
|
* is not running anything, yet the second is, and this function
|
|
|
|
* would mark the package as used after the first CPU and would
|
|
|
|
* not read the values from the second CPU.
|
|
|
|
*/
|
|
|
|
if (!(vals->run && vals->ena))
|
|
|
|
return 0;
|
|
|
|
|
2015-10-16 18:41:15 +08:00
|
|
|
s = cpu_map__get_socket(cpus, cpu, NULL);
|
2015-07-21 20:31:27 +08:00
|
|
|
if (s < 0)
|
|
|
|
return -1;
|
|
|
|
|
|
|
|
*skip = test_and_set_bit(s, mask) == 1;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int
|
|
|
|
process_counter_values(struct perf_stat_config *config, struct perf_evsel *evsel,
|
|
|
|
int cpu, int thread,
|
|
|
|
struct perf_counts_values *count)
|
|
|
|
{
|
|
|
|
struct perf_counts_values *aggr = &evsel->counts->aggr;
|
|
|
|
static struct perf_counts_values zero;
|
|
|
|
bool skip = false;
|
|
|
|
|
2015-09-03 21:23:40 +08:00
|
|
|
if (check_per_pkg(evsel, count, cpu, &skip)) {
|
2015-07-21 20:31:27 +08:00
|
|
|
pr_err("failed to read per-pkg counter\n");
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (skip)
|
|
|
|
count = &zero;
|
|
|
|
|
|
|
|
switch (config->aggr_mode) {
|
|
|
|
case AGGR_THREAD:
|
|
|
|
case AGGR_CORE:
|
|
|
|
case AGGR_SOCKET:
|
|
|
|
case AGGR_NONE:
|
|
|
|
if (!evsel->snapshot)
|
|
|
|
perf_evsel__compute_deltas(evsel, cpu, thread, count);
|
|
|
|
perf_counts_values__scale(count, config->scale, NULL);
|
|
|
|
if (config->aggr_mode == AGGR_NONE)
|
2017-01-24 05:42:56 +08:00
|
|
|
perf_stat__update_shadow_stats(evsel, count->val, cpu);
|
perf stat: Make --per-thread update shadow stats to show metrics
We should support this because it would allow easily to collect metrics
for different threads in applications.
Original patch from posted by Jin Yao in here [1].
1. Current output, for example:
root@skl:/tmp# perf stat --per-thread -p 21623
^C
Performance counter stats for process id '21623':
vmstat-21623 0.517479 task-clock (msec) # 0.000 CPUs utilized
vmstat-21623 1 context-switches
vmstat-21623 0 cpu-migrations
vmstat-21623 0 page-faults
vmstat-21623 461,306 cycles
vmstat-21623 630,724 instructions
vmstat-21623 136,265 branches
vmstat-21623 2,520 branch-misses
1.444020756 seconds time elapsed
root@skl:/tmp# perf stat --per-thread --metrics ipc -p 21623
^C
Performance counter stats for process id '21623':
vmstat-21623 631,185 inst_retired.any
vmstat-21623 605,893 cpu_clk_unhalted.thread
1.415679293 seconds time elapsed
2. With this patch, the result would be:
root@skl:/tmp# perf stat --per-thread -p 21623
^C
Performance counter stats for process id '21623':
vmstat-21623 0.533759 task-clock (msec) # 0.000 CPUs utilized
vmstat-21623 1 context-switches # 0.002 M/sec
vmstat-21623 0 cpu-migrations # 0.000 K/sec
vmstat-21623 0 page-faults # 0.000 K/sec
vmstat-21623 473,896 cycles # 0.888 GHz
vmstat-21623 631,072 instructions # 1.33 insn per cycle
vmstat-21623 136,307 branches # 255.372 M/sec
vmstat-21623 2,524 branch-misses # 1.85% of all branches
1.544862861 seconds time elapsed
root@skl:/tmp# perf stat --per-thread --metrics ipc -p 21623
^C
Performance counter stats for process id '21623':
vmstat-21623 1,259,104 inst_retired.any # 1.2 IPC
vmstat-21623 1,056,756 cpu_clk_unhalted.thread
2.040954502 seconds time elapsed
[1] https://marc.info/?l=linux-kernel&m=150777054620511&w=2
Originally-from: Jin Yao <yao.jin@linux.intel.com>
Signed-off-by: Jiri Olsa <jolsa@kernel.org>
Cc: Andi Kleen <andi@firstfloor.org>
Cc: Changbin Du <changbin.du@intel.com>
Cc: David Ahern <dsahern@gmail.com>
Cc: Namhyung Kim <namhyung@kernel.org>
Cc: Peter Zijlstra <peterz@infradead.org>
Cc: Wang Nan <wangnan0@huawei.com>
Link: http://lkml.kernel.org/n/tip-tr8ntktxmy4qc5769ajg5u6c@git.kernel.org
Signed-off-by: Arnaldo Carvalho de Melo <acme@redhat.com>
2017-10-12 19:21:01 +08:00
|
|
|
if (config->aggr_mode == AGGR_THREAD)
|
|
|
|
perf_stat__update_shadow_stats(evsel, count->val, 0);
|
2015-07-21 20:31:27 +08:00
|
|
|
break;
|
|
|
|
case AGGR_GLOBAL:
|
|
|
|
aggr->val += count->val;
|
|
|
|
if (config->scale) {
|
|
|
|
aggr->ena += count->ena;
|
|
|
|
aggr->run += count->run;
|
|
|
|
}
|
2015-10-16 18:41:04 +08:00
|
|
|
case AGGR_UNSET:
|
2015-07-21 20:31:27 +08:00
|
|
|
default:
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int process_counter_maps(struct perf_stat_config *config,
|
|
|
|
struct perf_evsel *counter)
|
|
|
|
{
|
|
|
|
int nthreads = thread_map__nr(counter->threads);
|
|
|
|
int ncpus = perf_evsel__nr_cpus(counter);
|
|
|
|
int cpu, thread;
|
|
|
|
|
|
|
|
if (counter->system_wide)
|
|
|
|
nthreads = 1;
|
|
|
|
|
|
|
|
for (thread = 0; thread < nthreads; thread++) {
|
|
|
|
for (cpu = 0; cpu < ncpus; cpu++) {
|
|
|
|
if (process_counter_values(config, counter, cpu, thread,
|
|
|
|
perf_counts(counter->counts, cpu, thread)))
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
int perf_stat_process_counter(struct perf_stat_config *config,
|
|
|
|
struct perf_evsel *counter)
|
|
|
|
{
|
|
|
|
struct perf_counts_values *aggr = &counter->counts->aggr;
|
2017-10-27 01:22:34 +08:00
|
|
|
struct perf_stat_evsel *ps = counter->stats;
|
2015-07-21 20:31:27 +08:00
|
|
|
u64 *count = counter->counts->aggr.values;
|
|
|
|
int i, ret;
|
|
|
|
|
|
|
|
aggr->val = aggr->ena = aggr->run = 0;
|
|
|
|
|
perf stat: Fix interval output values
We broke interval data displays with commit:
3f416f22d1e2 ("perf stat: Do not clean event's private stats")
This commit removed stats cleaning, which is important for '-r' option
to carry counters data over the whole run. But it's necessary to clean
it for interval mode, otherwise the displayed value is avg of all
previous values.
Before:
$ perf stat -e cycles -a -I 1000 record
# time counts unit events
1.000240796 75,216,287 cycles
2.000512791 107,823,524 cycles
$ perf stat report
# time counts unit events
1.000240796 75,216,287 cycles
2.000512791 91,519,906 cycles
Now:
$ perf stat report
# time counts unit events
1.000240796 75,216,287 cycles
2.000512791 107,823,524 cycles
Notice the second value being bigger (91,.. < 107,..).
This could be easily verified by using perf script which displays raw
stat data:
$ perf script
CPU THREAD VAL ENA RUN TIME EVENT
0 -1 23855779 1000209530 1000209530 1000240796 cycles
1 -1 33340397 1000224964 1000224964 1000240796 cycles
2 -1 15835415 1000226695 1000226695 1000240796 cycles
3 -1 2184696 1000228245 1000228245 1000240796 cycles
0 -1 97014312 2000514533 2000514533 2000512791 cycles
1 -1 46121497 2000543795 2000543795 2000512791 cycles
2 -1 32269530 2000543566 2000543566 2000512791 cycles
3 -1 7634472 2000544108 2000544108 2000512791 cycles
The sum of the first 4 values is the first interval aggregated value:
23855779 + 33340397 + 15835415 + 2184696 = 75,216,287
The sum of the second 4 values minus first value is the second interval
aggregated value:
97014312 + 46121497 + 32269530 + 7634472 - 75216287 = 107,823,524
Signed-off-by: Jiri Olsa <jolsa@kernel.org>
Cc: Andi Kleen <ak@linux.intel.com>
Cc: David Ahern <dsahern@gmail.com>
Cc: Namhyung Kim <namhyung@kernel.org>
Cc: Peter Zijlstra <a.p.zijlstra@chello.nl>
Cc: Stephane Eranian <eranian@google.com>
Link: http://lkml.kernel.org/r/1454485436-20639-1-git-send-email-jolsa@kernel.org
Signed-off-by: Arnaldo Carvalho de Melo <acme@redhat.com>
2016-02-03 15:43:56 +08:00
|
|
|
/*
|
|
|
|
* We calculate counter's data every interval,
|
|
|
|
* and the display code shows ps->res_stats
|
|
|
|
* avg value. We need to zero the stats for
|
|
|
|
* interval mode, otherwise overall avg running
|
|
|
|
* averages will be shown for each interval.
|
|
|
|
*/
|
|
|
|
if (config->interval)
|
|
|
|
init_stats(ps->res_stats);
|
|
|
|
|
2015-07-21 20:31:27 +08:00
|
|
|
if (counter->per_pkg)
|
|
|
|
zero_per_pkg(counter);
|
|
|
|
|
|
|
|
ret = process_counter_maps(config, counter);
|
|
|
|
if (ret)
|
|
|
|
return ret;
|
|
|
|
|
|
|
|
if (config->aggr_mode != AGGR_GLOBAL)
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
if (!counter->snapshot)
|
|
|
|
perf_evsel__compute_deltas(counter, -1, -1, aggr);
|
|
|
|
perf_counts_values__scale(aggr, config->scale, &counter->counts->scaled);
|
|
|
|
|
|
|
|
for (i = 0; i < 3; i++)
|
|
|
|
update_stats(&ps->res_stats[i], count[i]);
|
|
|
|
|
2017-02-17 16:17:38 +08:00
|
|
|
if (verbose > 0) {
|
2015-07-21 20:31:27 +08:00
|
|
|
fprintf(config->output, "%s: %" PRIu64 " %" PRIu64 " %" PRIu64 "\n",
|
|
|
|
perf_evsel__name(counter), count[0], count[1], count[2]);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Save the full runtime - to allow normalization during printout:
|
|
|
|
*/
|
2017-01-24 05:42:56 +08:00
|
|
|
perf_stat__update_shadow_stats(counter, *count, 0);
|
2015-07-21 20:31:27 +08:00
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
2015-10-25 22:51:32 +08:00
|
|
|
|
|
|
|
int perf_event__process_stat_event(struct perf_tool *tool __maybe_unused,
|
|
|
|
union perf_event *event,
|
|
|
|
struct perf_session *session)
|
|
|
|
{
|
|
|
|
struct perf_counts_values count;
|
|
|
|
struct stat_event *st = &event->stat;
|
|
|
|
struct perf_evsel *counter;
|
|
|
|
|
|
|
|
count.val = st->val;
|
|
|
|
count.ena = st->ena;
|
|
|
|
count.run = st->run;
|
|
|
|
|
|
|
|
counter = perf_evlist__id2evsel(session->evlist, st->id);
|
|
|
|
if (!counter) {
|
|
|
|
pr_err("Failed to resolve counter for stat event.\n");
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
|
|
|
|
*perf_counts(counter->counts, st->cpu, st->thread) = count;
|
|
|
|
counter->supported = true;
|
|
|
|
return 0;
|
|
|
|
}
|
2015-10-25 22:51:35 +08:00
|
|
|
|
|
|
|
size_t perf_event__fprintf_stat(union perf_event *event, FILE *fp)
|
|
|
|
{
|
|
|
|
struct stat_event *st = (struct stat_event *) event;
|
|
|
|
size_t ret;
|
|
|
|
|
|
|
|
ret = fprintf(fp, "\n... id %" PRIu64 ", cpu %d, thread %d\n",
|
|
|
|
st->id, st->cpu, st->thread);
|
|
|
|
ret += fprintf(fp, "... value %" PRIu64 ", enabled %" PRIu64 ", running %" PRIu64 "\n",
|
|
|
|
st->val, st->ena, st->run);
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
size_t perf_event__fprintf_stat_round(union perf_event *event, FILE *fp)
|
|
|
|
{
|
|
|
|
struct stat_round_event *rd = (struct stat_round_event *)event;
|
|
|
|
size_t ret;
|
|
|
|
|
|
|
|
ret = fprintf(fp, "\n... time %" PRIu64 ", type %s\n", rd->time,
|
|
|
|
rd->type == PERF_STAT_ROUND_TYPE__FINAL ? "FINAL" : "INTERVAL");
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
size_t perf_event__fprintf_stat_config(union perf_event *event, FILE *fp)
|
|
|
|
{
|
|
|
|
struct perf_stat_config sc;
|
|
|
|
size_t ret;
|
|
|
|
|
|
|
|
perf_event__read_stat_config(&sc, &event->stat_config);
|
|
|
|
|
|
|
|
ret = fprintf(fp, "\n");
|
|
|
|
ret += fprintf(fp, "... aggr_mode %d\n", sc.aggr_mode);
|
|
|
|
ret += fprintf(fp, "... scale %d\n", sc.scale);
|
|
|
|
ret += fprintf(fp, "... interval %u\n", sc.interval);
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|