mirror of
https://mirrors.bfsu.edu.cn/git/linux.git
synced 2024-12-15 06:55:13 +08:00
c8b5f2c96d
The tools so far have been using the strerror_r() GNU variant, that returns a string, be it the buffer passed or something else. But that, besides being tricky in cases where we expect that the function using strerror_r() returns the error formatted in a provided buffer (we have to check if it returned something else and copy that instead), breaks the build on systems not using glibc, like Alpine Linux, where musl libc is used. So, introduce yet another wrapper, str_error_r(), that has the GNU interface, but uses the portable XSI variant of strerror_r(), so that users rest asured that the provided buffer is used and it is what is returned. Cc: Adrian Hunter <adrian.hunter@intel.com> Cc: David Ahern <dsahern@gmail.com> Cc: Jiri Olsa <jolsa@kernel.org> Cc: Namhyung Kim <namhyung@kernel.org> Cc: Wang Nan <wangnan0@huawei.com> Link: http://lkml.kernel.org/n/tip-d4t42fnf48ytlk8rjxs822tf@git.kernel.org Signed-off-by: Arnaldo Carvalho de Melo <acme@redhat.com>
116 lines
2.9 KiB
C
116 lines
2.9 KiB
C
#include <api/fs/fs.h>
|
|
#include <linux/err.h>
|
|
#include "evsel.h"
|
|
#include "tests.h"
|
|
#include "thread_map.h"
|
|
#include "cpumap.h"
|
|
#include "debug.h"
|
|
#include "stat.h"
|
|
|
|
int test__openat_syscall_event_on_all_cpus(int subtest __maybe_unused)
|
|
{
|
|
int err = -1, fd, cpu;
|
|
struct cpu_map *cpus;
|
|
struct perf_evsel *evsel;
|
|
unsigned int nr_openat_calls = 111, i;
|
|
cpu_set_t cpu_set;
|
|
struct thread_map *threads = thread_map__new(-1, getpid(), UINT_MAX);
|
|
char sbuf[STRERR_BUFSIZE];
|
|
char errbuf[BUFSIZ];
|
|
|
|
if (threads == NULL) {
|
|
pr_debug("thread_map__new\n");
|
|
return -1;
|
|
}
|
|
|
|
cpus = cpu_map__new(NULL);
|
|
if (cpus == NULL) {
|
|
pr_debug("cpu_map__new\n");
|
|
goto out_thread_map_delete;
|
|
}
|
|
|
|
CPU_ZERO(&cpu_set);
|
|
|
|
evsel = perf_evsel__newtp("syscalls", "sys_enter_openat");
|
|
if (IS_ERR(evsel)) {
|
|
tracing_path__strerror_open_tp(errno, errbuf, sizeof(errbuf), "syscalls", "sys_enter_openat");
|
|
pr_debug("%s\n", errbuf);
|
|
goto out_thread_map_delete;
|
|
}
|
|
|
|
if (perf_evsel__open(evsel, cpus, threads) < 0) {
|
|
pr_debug("failed to open counter: %s, "
|
|
"tweak /proc/sys/kernel/perf_event_paranoid?\n",
|
|
str_error_r(errno, sbuf, sizeof(sbuf)));
|
|
goto out_evsel_delete;
|
|
}
|
|
|
|
for (cpu = 0; cpu < cpus->nr; ++cpu) {
|
|
unsigned int ncalls = nr_openat_calls + cpu;
|
|
/*
|
|
* XXX eventually lift this restriction in a way that
|
|
* keeps perf building on older glibc installations
|
|
* without CPU_ALLOC. 1024 cpus in 2010 still seems
|
|
* a reasonable upper limit tho :-)
|
|
*/
|
|
if (cpus->map[cpu] >= CPU_SETSIZE) {
|
|
pr_debug("Ignoring CPU %d\n", cpus->map[cpu]);
|
|
continue;
|
|
}
|
|
|
|
CPU_SET(cpus->map[cpu], &cpu_set);
|
|
if (sched_setaffinity(0, sizeof(cpu_set), &cpu_set) < 0) {
|
|
pr_debug("sched_setaffinity() failed on CPU %d: %s ",
|
|
cpus->map[cpu],
|
|
str_error_r(errno, sbuf, sizeof(sbuf)));
|
|
goto out_close_fd;
|
|
}
|
|
for (i = 0; i < ncalls; ++i) {
|
|
fd = openat(0, "/etc/passwd", O_RDONLY);
|
|
close(fd);
|
|
}
|
|
CPU_CLR(cpus->map[cpu], &cpu_set);
|
|
}
|
|
|
|
/*
|
|
* Here we need to explicitly preallocate the counts, as if
|
|
* we use the auto allocation it will allocate just for 1 cpu,
|
|
* as we start by cpu 0.
|
|
*/
|
|
if (perf_evsel__alloc_counts(evsel, cpus->nr, 1) < 0) {
|
|
pr_debug("perf_evsel__alloc_counts(ncpus=%d)\n", cpus->nr);
|
|
goto out_close_fd;
|
|
}
|
|
|
|
err = 0;
|
|
|
|
for (cpu = 0; cpu < cpus->nr; ++cpu) {
|
|
unsigned int expected;
|
|
|
|
if (cpus->map[cpu] >= CPU_SETSIZE)
|
|
continue;
|
|
|
|
if (perf_evsel__read_on_cpu(evsel, cpu, 0) < 0) {
|
|
pr_debug("perf_evsel__read_on_cpu\n");
|
|
err = -1;
|
|
break;
|
|
}
|
|
|
|
expected = nr_openat_calls + cpu;
|
|
if (perf_counts(evsel->counts, cpu, 0)->val != expected) {
|
|
pr_debug("perf_evsel__read_on_cpu: expected to intercept %d calls on cpu %d, got %" PRIu64 "\n",
|
|
expected, cpus->map[cpu], perf_counts(evsel->counts, cpu, 0)->val);
|
|
err = -1;
|
|
}
|
|
}
|
|
|
|
perf_evsel__free_counts(evsel);
|
|
out_close_fd:
|
|
perf_evsel__close_fd(evsel, 1, threads->nr);
|
|
out_evsel_delete:
|
|
perf_evsel__delete(evsel);
|
|
out_thread_map_delete:
|
|
thread_map__put(threads);
|
|
return err;
|
|
}
|