mirror of
https://mirrors.bfsu.edu.cn/git/linux.git
synced 2024-12-01 08:04:22 +08:00
63cd02d84b
Before this change, the '--graph-funcs', '--nograph-funcs' and '--trace-funcs' options didn't work as expected when the <func> doesn't exist. Because the kernel side hid possible errors. $ sudo ./perf ftrace -a --graph-depth 1 --graph-funcs abcdefg 0) 0.140 us | rcu_all_qs(); 3) 0.304 us | mutex_unlock(); 0) 0.153 us | find_vma(); 3) 0.088 us | __fsnotify_parent(); 0) 6.145 us | handle_mm_fault(); 3) 0.089 us | fsnotify(); 3) 0.161 us | __sb_end_write(); 3) 0.710 us | SyS_close(); 3) 7.848 us | exit_to_usermode_loop(); On the example above, I specified the function filter 'abcdefg' but all functions are enabled. The expected result is for all functions to be filtered, since there is no such function ('abcdefg') The original fix is to make the kernel support '\0' as end of string: https://lkml.org/lkml/2018/1/16/116 But above fix cannot be compatible with old kernels. Then Namhyung Kim suggest adding a space after function name. This patch will append an '\n' when write tracing file. After this fix, the perf will report correct error state. Also let it print an error if reset_tracing_files() fails. Committer testing: Now it prints: # perf ftrace -a --graph-depth 1 --graph-funcs abcdefg failed to set tracing filters # And for an existing function: # perf ftrace -a --graph-depth 1 --graph-funcs SyS_open 3) | SyS_open() { 3) ! 494.899 us | } 0) + 23.910 us | SyS_open(); 1) + 17.115 us | SyS_open(); 1) + 13.900 us | SyS_open(); ------------------------------------------ 3) qemu-sy-2817 => pickup-1290 ------------------------------------------ 3) + 20.021 us | SyS_open(); # Signed-off-by: Changbin Du <changbin.du@intel.com> Tested-by: Arnaldo Carvalho de Melo <acme@redhat.com> Cc: Jiri Olsa <jolsa@redhat.com> Cc: Namhyung Kim <namhyung@kernel.org> Cc: Peter Zijlstra <peterz@infradead.org> Link: http://lkml.kernel.org/r/1519007609-14551-1-git-send-email-changbin.du@intel.com Signed-off-by: Arnaldo Carvalho de Melo <acme@redhat.com>
522 lines
11 KiB
C
522 lines
11 KiB
C
/*
|
|
* builtin-ftrace.c
|
|
*
|
|
* Copyright (c) 2013 LG Electronics, Namhyung Kim <namhyung@kernel.org>
|
|
*
|
|
* Released under the GPL v2.
|
|
*/
|
|
|
|
#include "builtin.h"
|
|
#include "perf.h"
|
|
|
|
#include <errno.h>
|
|
#include <unistd.h>
|
|
#include <signal.h>
|
|
#include <fcntl.h>
|
|
#include <poll.h>
|
|
|
|
#include "debug.h"
|
|
#include <subcmd/parse-options.h>
|
|
#include <api/fs/tracing_path.h>
|
|
#include "evlist.h"
|
|
#include "target.h"
|
|
#include "cpumap.h"
|
|
#include "thread_map.h"
|
|
#include "util/config.h"
|
|
|
|
|
|
#define DEFAULT_TRACER "function_graph"
|
|
|
|
struct perf_ftrace {
|
|
struct perf_evlist *evlist;
|
|
struct target target;
|
|
const char *tracer;
|
|
struct list_head filters;
|
|
struct list_head notrace;
|
|
struct list_head graph_funcs;
|
|
struct list_head nograph_funcs;
|
|
int graph_depth;
|
|
};
|
|
|
|
struct filter_entry {
|
|
struct list_head list;
|
|
char name[];
|
|
};
|
|
|
|
static bool done;
|
|
|
|
static void sig_handler(int sig __maybe_unused)
|
|
{
|
|
done = true;
|
|
}
|
|
|
|
/*
|
|
* perf_evlist__prepare_workload will send a SIGUSR1 if the fork fails, since
|
|
* we asked by setting its exec_error to the function below,
|
|
* ftrace__workload_exec_failed_signal.
|
|
*
|
|
* XXX We need to handle this more appropriately, emitting an error, etc.
|
|
*/
|
|
static void ftrace__workload_exec_failed_signal(int signo __maybe_unused,
|
|
siginfo_t *info __maybe_unused,
|
|
void *ucontext __maybe_unused)
|
|
{
|
|
/* workload_exec_errno = info->si_value.sival_int; */
|
|
done = true;
|
|
}
|
|
|
|
static int __write_tracing_file(const char *name, const char *val, bool append)
|
|
{
|
|
char *file;
|
|
int fd, ret = -1;
|
|
ssize_t size = strlen(val);
|
|
int flags = O_WRONLY;
|
|
char errbuf[512];
|
|
char *val_copy;
|
|
|
|
file = get_tracing_file(name);
|
|
if (!file) {
|
|
pr_debug("cannot get tracing file: %s\n", name);
|
|
return -1;
|
|
}
|
|
|
|
if (append)
|
|
flags |= O_APPEND;
|
|
else
|
|
flags |= O_TRUNC;
|
|
|
|
fd = open(file, flags);
|
|
if (fd < 0) {
|
|
pr_debug("cannot open tracing file: %s: %s\n",
|
|
name, str_error_r(errno, errbuf, sizeof(errbuf)));
|
|
goto out;
|
|
}
|
|
|
|
/*
|
|
* Copy the original value and append a '\n'. Without this,
|
|
* the kernel can hide possible errors.
|
|
*/
|
|
val_copy = strdup(val);
|
|
if (!val_copy)
|
|
goto out_close;
|
|
val_copy[size] = '\n';
|
|
|
|
if (write(fd, val_copy, size + 1) == size + 1)
|
|
ret = 0;
|
|
else
|
|
pr_debug("write '%s' to tracing/%s failed: %s\n",
|
|
val, name, str_error_r(errno, errbuf, sizeof(errbuf)));
|
|
|
|
free(val_copy);
|
|
out_close:
|
|
close(fd);
|
|
out:
|
|
put_tracing_file(file);
|
|
return ret;
|
|
}
|
|
|
|
static int write_tracing_file(const char *name, const char *val)
|
|
{
|
|
return __write_tracing_file(name, val, false);
|
|
}
|
|
|
|
static int append_tracing_file(const char *name, const char *val)
|
|
{
|
|
return __write_tracing_file(name, val, true);
|
|
}
|
|
|
|
static int reset_tracing_cpu(void);
|
|
static void reset_tracing_filters(void);
|
|
|
|
static int reset_tracing_files(struct perf_ftrace *ftrace __maybe_unused)
|
|
{
|
|
if (write_tracing_file("tracing_on", "0") < 0)
|
|
return -1;
|
|
|
|
if (write_tracing_file("current_tracer", "nop") < 0)
|
|
return -1;
|
|
|
|
if (write_tracing_file("set_ftrace_pid", " ") < 0)
|
|
return -1;
|
|
|
|
if (reset_tracing_cpu() < 0)
|
|
return -1;
|
|
|
|
if (write_tracing_file("max_graph_depth", "0") < 0)
|
|
return -1;
|
|
|
|
reset_tracing_filters();
|
|
return 0;
|
|
}
|
|
|
|
static int set_tracing_pid(struct perf_ftrace *ftrace)
|
|
{
|
|
int i;
|
|
char buf[16];
|
|
|
|
if (target__has_cpu(&ftrace->target))
|
|
return 0;
|
|
|
|
for (i = 0; i < thread_map__nr(ftrace->evlist->threads); i++) {
|
|
scnprintf(buf, sizeof(buf), "%d",
|
|
ftrace->evlist->threads->map[i]);
|
|
if (append_tracing_file("set_ftrace_pid", buf) < 0)
|
|
return -1;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static int set_tracing_cpumask(struct cpu_map *cpumap)
|
|
{
|
|
char *cpumask;
|
|
size_t mask_size;
|
|
int ret;
|
|
int last_cpu;
|
|
|
|
last_cpu = cpu_map__cpu(cpumap, cpumap->nr - 1);
|
|
mask_size = (last_cpu + 3) / 4 + 1;
|
|
mask_size += last_cpu / 32; /* ',' is needed for every 32th cpus */
|
|
|
|
cpumask = malloc(mask_size);
|
|
if (cpumask == NULL) {
|
|
pr_debug("failed to allocate cpu mask\n");
|
|
return -1;
|
|
}
|
|
|
|
cpu_map__snprint_mask(cpumap, cpumask, mask_size);
|
|
|
|
ret = write_tracing_file("tracing_cpumask", cpumask);
|
|
|
|
free(cpumask);
|
|
return ret;
|
|
}
|
|
|
|
static int set_tracing_cpu(struct perf_ftrace *ftrace)
|
|
{
|
|
struct cpu_map *cpumap = ftrace->evlist->cpus;
|
|
|
|
if (!target__has_cpu(&ftrace->target))
|
|
return 0;
|
|
|
|
return set_tracing_cpumask(cpumap);
|
|
}
|
|
|
|
static int reset_tracing_cpu(void)
|
|
{
|
|
struct cpu_map *cpumap = cpu_map__new(NULL);
|
|
int ret;
|
|
|
|
ret = set_tracing_cpumask(cpumap);
|
|
cpu_map__put(cpumap);
|
|
return ret;
|
|
}
|
|
|
|
static int __set_tracing_filter(const char *filter_file, struct list_head *funcs)
|
|
{
|
|
struct filter_entry *pos;
|
|
|
|
list_for_each_entry(pos, funcs, list) {
|
|
if (append_tracing_file(filter_file, pos->name) < 0)
|
|
return -1;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int set_tracing_filters(struct perf_ftrace *ftrace)
|
|
{
|
|
int ret;
|
|
|
|
ret = __set_tracing_filter("set_ftrace_filter", &ftrace->filters);
|
|
if (ret < 0)
|
|
return ret;
|
|
|
|
ret = __set_tracing_filter("set_ftrace_notrace", &ftrace->notrace);
|
|
if (ret < 0)
|
|
return ret;
|
|
|
|
ret = __set_tracing_filter("set_graph_function", &ftrace->graph_funcs);
|
|
if (ret < 0)
|
|
return ret;
|
|
|
|
/* old kernels do not have this filter */
|
|
__set_tracing_filter("set_graph_notrace", &ftrace->nograph_funcs);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static void reset_tracing_filters(void)
|
|
{
|
|
write_tracing_file("set_ftrace_filter", " ");
|
|
write_tracing_file("set_ftrace_notrace", " ");
|
|
write_tracing_file("set_graph_function", " ");
|
|
write_tracing_file("set_graph_notrace", " ");
|
|
}
|
|
|
|
static int set_tracing_depth(struct perf_ftrace *ftrace)
|
|
{
|
|
char buf[16];
|
|
|
|
if (ftrace->graph_depth == 0)
|
|
return 0;
|
|
|
|
if (ftrace->graph_depth < 0) {
|
|
pr_err("invalid graph depth: %d\n", ftrace->graph_depth);
|
|
return -1;
|
|
}
|
|
|
|
snprintf(buf, sizeof(buf), "%d", ftrace->graph_depth);
|
|
|
|
if (write_tracing_file("max_graph_depth", buf) < 0)
|
|
return -1;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int __cmd_ftrace(struct perf_ftrace *ftrace, int argc, const char **argv)
|
|
{
|
|
char *trace_file;
|
|
int trace_fd;
|
|
char buf[4096];
|
|
struct pollfd pollfd = {
|
|
.events = POLLIN,
|
|
};
|
|
|
|
if (geteuid() != 0) {
|
|
pr_err("ftrace only works for root!\n");
|
|
return -1;
|
|
}
|
|
|
|
signal(SIGINT, sig_handler);
|
|
signal(SIGUSR1, sig_handler);
|
|
signal(SIGCHLD, sig_handler);
|
|
signal(SIGPIPE, sig_handler);
|
|
|
|
if (reset_tracing_files(ftrace) < 0) {
|
|
pr_err("failed to reset ftrace\n");
|
|
goto out;
|
|
}
|
|
|
|
/* reset ftrace buffer */
|
|
if (write_tracing_file("trace", "0") < 0)
|
|
goto out;
|
|
|
|
if (argc && perf_evlist__prepare_workload(ftrace->evlist,
|
|
&ftrace->target, argv, false,
|
|
ftrace__workload_exec_failed_signal) < 0) {
|
|
goto out;
|
|
}
|
|
|
|
if (set_tracing_pid(ftrace) < 0) {
|
|
pr_err("failed to set ftrace pid\n");
|
|
goto out_reset;
|
|
}
|
|
|
|
if (set_tracing_cpu(ftrace) < 0) {
|
|
pr_err("failed to set tracing cpumask\n");
|
|
goto out_reset;
|
|
}
|
|
|
|
if (set_tracing_filters(ftrace) < 0) {
|
|
pr_err("failed to set tracing filters\n");
|
|
goto out_reset;
|
|
}
|
|
|
|
if (set_tracing_depth(ftrace) < 0) {
|
|
pr_err("failed to set graph depth\n");
|
|
goto out_reset;
|
|
}
|
|
|
|
if (write_tracing_file("current_tracer", ftrace->tracer) < 0) {
|
|
pr_err("failed to set current_tracer to %s\n", ftrace->tracer);
|
|
goto out_reset;
|
|
}
|
|
|
|
setup_pager();
|
|
|
|
trace_file = get_tracing_file("trace_pipe");
|
|
if (!trace_file) {
|
|
pr_err("failed to open trace_pipe\n");
|
|
goto out_reset;
|
|
}
|
|
|
|
trace_fd = open(trace_file, O_RDONLY);
|
|
|
|
put_tracing_file(trace_file);
|
|
|
|
if (trace_fd < 0) {
|
|
pr_err("failed to open trace_pipe\n");
|
|
goto out_reset;
|
|
}
|
|
|
|
fcntl(trace_fd, F_SETFL, O_NONBLOCK);
|
|
pollfd.fd = trace_fd;
|
|
|
|
if (write_tracing_file("tracing_on", "1") < 0) {
|
|
pr_err("can't enable tracing\n");
|
|
goto out_close_fd;
|
|
}
|
|
|
|
perf_evlist__start_workload(ftrace->evlist);
|
|
|
|
while (!done) {
|
|
if (poll(&pollfd, 1, -1) < 0)
|
|
break;
|
|
|
|
if (pollfd.revents & POLLIN) {
|
|
int n = read(trace_fd, buf, sizeof(buf));
|
|
if (n < 0)
|
|
break;
|
|
if (fwrite(buf, n, 1, stdout) != 1)
|
|
break;
|
|
}
|
|
}
|
|
|
|
write_tracing_file("tracing_on", "0");
|
|
|
|
/* read remaining buffer contents */
|
|
while (true) {
|
|
int n = read(trace_fd, buf, sizeof(buf));
|
|
if (n <= 0)
|
|
break;
|
|
if (fwrite(buf, n, 1, stdout) != 1)
|
|
break;
|
|
}
|
|
|
|
out_close_fd:
|
|
close(trace_fd);
|
|
out_reset:
|
|
reset_tracing_files(ftrace);
|
|
out:
|
|
return done ? 0 : -1;
|
|
}
|
|
|
|
static int perf_ftrace_config(const char *var, const char *value, void *cb)
|
|
{
|
|
struct perf_ftrace *ftrace = cb;
|
|
|
|
if (!strstarts(var, "ftrace."))
|
|
return 0;
|
|
|
|
if (strcmp(var, "ftrace.tracer"))
|
|
return -1;
|
|
|
|
if (!strcmp(value, "function_graph") ||
|
|
!strcmp(value, "function")) {
|
|
ftrace->tracer = value;
|
|
return 0;
|
|
}
|
|
|
|
pr_err("Please select \"function_graph\" (default) or \"function\"\n");
|
|
return -1;
|
|
}
|
|
|
|
static int parse_filter_func(const struct option *opt, const char *str,
|
|
int unset __maybe_unused)
|
|
{
|
|
struct list_head *head = opt->value;
|
|
struct filter_entry *entry;
|
|
|
|
entry = malloc(sizeof(*entry) + strlen(str) + 1);
|
|
if (entry == NULL)
|
|
return -ENOMEM;
|
|
|
|
strcpy(entry->name, str);
|
|
list_add_tail(&entry->list, head);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void delete_filter_func(struct list_head *head)
|
|
{
|
|
struct filter_entry *pos, *tmp;
|
|
|
|
list_for_each_entry_safe(pos, tmp, head, list) {
|
|
list_del(&pos->list);
|
|
free(pos);
|
|
}
|
|
}
|
|
|
|
int cmd_ftrace(int argc, const char **argv)
|
|
{
|
|
int ret;
|
|
struct perf_ftrace ftrace = {
|
|
.tracer = DEFAULT_TRACER,
|
|
.target = { .uid = UINT_MAX, },
|
|
};
|
|
const char * const ftrace_usage[] = {
|
|
"perf ftrace [<options>] [<command>]",
|
|
"perf ftrace [<options>] -- <command> [<options>]",
|
|
NULL
|
|
};
|
|
const struct option ftrace_options[] = {
|
|
OPT_STRING('t', "tracer", &ftrace.tracer, "tracer",
|
|
"tracer to use: function_graph(default) or function"),
|
|
OPT_STRING('p', "pid", &ftrace.target.pid, "pid",
|
|
"trace on existing process id"),
|
|
OPT_INCR('v', "verbose", &verbose,
|
|
"be more verbose"),
|
|
OPT_BOOLEAN('a', "all-cpus", &ftrace.target.system_wide,
|
|
"system-wide collection from all CPUs"),
|
|
OPT_STRING('C', "cpu", &ftrace.target.cpu_list, "cpu",
|
|
"list of cpus to monitor"),
|
|
OPT_CALLBACK('T', "trace-funcs", &ftrace.filters, "func",
|
|
"trace given functions only", parse_filter_func),
|
|
OPT_CALLBACK('N', "notrace-funcs", &ftrace.notrace, "func",
|
|
"do not trace given functions", parse_filter_func),
|
|
OPT_CALLBACK('G', "graph-funcs", &ftrace.graph_funcs, "func",
|
|
"Set graph filter on given functions", parse_filter_func),
|
|
OPT_CALLBACK('g', "nograph-funcs", &ftrace.nograph_funcs, "func",
|
|
"Set nograph filter on given functions", parse_filter_func),
|
|
OPT_INTEGER('D', "graph-depth", &ftrace.graph_depth,
|
|
"Max depth for function graph tracer"),
|
|
OPT_END()
|
|
};
|
|
|
|
INIT_LIST_HEAD(&ftrace.filters);
|
|
INIT_LIST_HEAD(&ftrace.notrace);
|
|
INIT_LIST_HEAD(&ftrace.graph_funcs);
|
|
INIT_LIST_HEAD(&ftrace.nograph_funcs);
|
|
|
|
ret = perf_config(perf_ftrace_config, &ftrace);
|
|
if (ret < 0)
|
|
return -1;
|
|
|
|
argc = parse_options(argc, argv, ftrace_options, ftrace_usage,
|
|
PARSE_OPT_STOP_AT_NON_OPTION);
|
|
if (!argc && target__none(&ftrace.target))
|
|
usage_with_options(ftrace_usage, ftrace_options);
|
|
|
|
ret = target__validate(&ftrace.target);
|
|
if (ret) {
|
|
char errbuf[512];
|
|
|
|
target__strerror(&ftrace.target, ret, errbuf, 512);
|
|
pr_err("%s\n", errbuf);
|
|
goto out_delete_filters;
|
|
}
|
|
|
|
ftrace.evlist = perf_evlist__new();
|
|
if (ftrace.evlist == NULL) {
|
|
ret = -ENOMEM;
|
|
goto out_delete_filters;
|
|
}
|
|
|
|
ret = perf_evlist__create_maps(ftrace.evlist, &ftrace.target);
|
|
if (ret < 0)
|
|
goto out_delete_evlist;
|
|
|
|
ret = __cmd_ftrace(&ftrace, argc, argv);
|
|
|
|
out_delete_evlist:
|
|
perf_evlist__delete(ftrace.evlist);
|
|
|
|
out_delete_filters:
|
|
delete_filter_func(&ftrace.filters);
|
|
delete_filter_func(&ftrace.notrace);
|
|
delete_filter_func(&ftrace.graph_funcs);
|
|
delete_filter_func(&ftrace.nograph_funcs);
|
|
|
|
return ret;
|
|
}
|