mirror of
https://mirrors.bfsu.edu.cn/git/linux.git
synced 2024-11-15 00:04:15 +08:00
5ab6d715c3
Introduce functions to access struct maps. These functions reduce the number of places reference counting is necessary. While tidying APIs do some small const-ification, in particlar to unwind_libunwind_ops. Committer notes: Fixed up tools/perf/util/unwind-libunwind.c: - return ops->get_entries(cb, arg, thread, data, max_stack); + return ops->get_entries(cb, arg, thread, data, max_stack, best_effort); Signed-off-by: Ian Rogers <irogers@google.com> Cc: Adrian Hunter <adrian.hunter@intel.com> Cc: Alexander Shishkin <alexander.shishkin@linux.intel.com> Cc: Alexey Bayduraev <alexey.v.bayduraev@linux.intel.com> Cc: Andi Kleen <ak@linux.intel.com> Cc: Andrew Morton <akpm@linux-foundation.org> Cc: Andy Shevchenko <andriy.shevchenko@linux.intel.com> Cc: Darren Hart <dvhart@infradead.org> Cc: Davidlohr Bueso <dave@stgolabs.net> Cc: Dmitriy Vyukov <dvyukov@google.com> Cc: Eric Dumazet <edumazet@google.com> Cc: German Gomez <german.gomez@arm.com> Cc: Hao Luo <haoluo@google.com> Cc: Ingo Molnar <mingo@redhat.com> Cc: James Clark <james.clark@arm.com> Cc: Jiri Olsa <jolsa@kernel.org> Cc: John Garry <john.g.garry@oracle.com> Cc: Kajol Jain <kjain@linux.ibm.com> Cc: Kan Liang <kan.liang@linux.intel.com> Cc: Leo Yan <leo.yan@linaro.org> Cc: Madhavan Srinivasan <maddy@linux.ibm.com> Cc: Mark Rutland <mark.rutland@arm.com> Cc: Masami Hiramatsu <mhiramat@kernel.org> Cc: Miaoqian Lin <linmq006@gmail.com> Cc: Namhyung Kim <namhyung@kernel.org> Cc: Peter Zijlstra <peterz@infradead.org> Cc: Riccardo Mancini <rickyman7@gmail.com> Cc: Shunsuke Nakamura <nakamura.shun@fujitsu.com> Cc: Song Liu <song@kernel.org> Cc: Stephane Eranian <eranian@google.com> Cc: Stephen Brennan <stephen.s.brennan@oracle.com> Cc: Steven Rostedt (VMware) <rostedt@goodmis.org> Cc: Thomas Gleixner <tglx@linutronix.de> Cc: Thomas Richter <tmricht@linux.ibm.com> Cc: Yury Norov <yury.norov@gmail.com> Link: https://lore.kernel.org/r/20230320212248.1175731-2-irogers@google.com Signed-off-by: Arnaldo Carvalho de Melo <acme@redhat.com>
469 lines
9.3 KiB
C
469 lines
9.3 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
#include <errno.h>
|
|
#include <stdlib.h>
|
|
#include <linux/zalloc.h>
|
|
#include "debug.h"
|
|
#include "dso.h"
|
|
#include "map.h"
|
|
#include "maps.h"
|
|
#include "thread.h"
|
|
#include "ui/ui.h"
|
|
#include "unwind.h"
|
|
|
|
static void maps__init(struct maps *maps, struct machine *machine)
|
|
{
|
|
maps->entries = RB_ROOT;
|
|
init_rwsem(maps__lock(maps));
|
|
maps->machine = machine;
|
|
maps->last_search_by_name = NULL;
|
|
maps->nr_maps = 0;
|
|
maps->maps_by_name = NULL;
|
|
refcount_set(&maps->refcnt, 1);
|
|
}
|
|
|
|
static void __maps__free_maps_by_name(struct maps *maps)
|
|
{
|
|
/*
|
|
* Free everything to try to do it from the rbtree in the next search
|
|
*/
|
|
zfree(&maps->maps_by_name);
|
|
maps->nr_maps_allocated = 0;
|
|
}
|
|
|
|
static int __maps__insert(struct maps *maps, struct map *map)
|
|
{
|
|
struct rb_node **p = &maps__entries(maps)->rb_node;
|
|
struct rb_node *parent = NULL;
|
|
const u64 ip = map->start;
|
|
struct map_rb_node *m, *new_rb_node;
|
|
|
|
new_rb_node = malloc(sizeof(*new_rb_node));
|
|
if (!new_rb_node)
|
|
return -ENOMEM;
|
|
|
|
RB_CLEAR_NODE(&new_rb_node->rb_node);
|
|
new_rb_node->map = map;
|
|
|
|
while (*p != NULL) {
|
|
parent = *p;
|
|
m = rb_entry(parent, struct map_rb_node, rb_node);
|
|
if (ip < m->map->start)
|
|
p = &(*p)->rb_left;
|
|
else
|
|
p = &(*p)->rb_right;
|
|
}
|
|
|
|
rb_link_node(&new_rb_node->rb_node, parent, p);
|
|
rb_insert_color(&new_rb_node->rb_node, maps__entries(maps));
|
|
map__get(map);
|
|
return 0;
|
|
}
|
|
|
|
int maps__insert(struct maps *maps, struct map *map)
|
|
{
|
|
int err;
|
|
|
|
down_write(maps__lock(maps));
|
|
err = __maps__insert(maps, map);
|
|
if (err)
|
|
goto out;
|
|
|
|
++maps->nr_maps;
|
|
|
|
if (map->dso && map->dso->kernel) {
|
|
struct kmap *kmap = map__kmap(map);
|
|
|
|
if (kmap)
|
|
kmap->kmaps = maps;
|
|
else
|
|
pr_err("Internal error: kernel dso with non kernel map\n");
|
|
}
|
|
|
|
|
|
/*
|
|
* If we already performed some search by name, then we need to add the just
|
|
* inserted map and resort.
|
|
*/
|
|
if (maps__maps_by_name(maps)) {
|
|
if (maps__nr_maps(maps) > maps->nr_maps_allocated) {
|
|
int nr_allocate = maps__nr_maps(maps) * 2;
|
|
struct map **maps_by_name = realloc(maps__maps_by_name(maps),
|
|
nr_allocate * sizeof(map));
|
|
|
|
if (maps_by_name == NULL) {
|
|
__maps__free_maps_by_name(maps);
|
|
err = -ENOMEM;
|
|
goto out;
|
|
}
|
|
|
|
maps->maps_by_name = maps_by_name;
|
|
maps->nr_maps_allocated = nr_allocate;
|
|
}
|
|
maps__maps_by_name(maps)[maps__nr_maps(maps) - 1] = map;
|
|
__maps__sort_by_name(maps);
|
|
}
|
|
out:
|
|
up_write(maps__lock(maps));
|
|
return err;
|
|
}
|
|
|
|
static void __maps__remove(struct maps *maps, struct map_rb_node *rb_node)
|
|
{
|
|
rb_erase_init(&rb_node->rb_node, maps__entries(maps));
|
|
map__put(rb_node->map);
|
|
free(rb_node);
|
|
}
|
|
|
|
void maps__remove(struct maps *maps, struct map *map)
|
|
{
|
|
struct map_rb_node *rb_node;
|
|
|
|
down_write(maps__lock(maps));
|
|
if (maps->last_search_by_name == map)
|
|
maps->last_search_by_name = NULL;
|
|
|
|
rb_node = maps__find_node(maps, map);
|
|
assert(rb_node->map == map);
|
|
__maps__remove(maps, rb_node);
|
|
--maps->nr_maps;
|
|
if (maps__maps_by_name(maps))
|
|
__maps__free_maps_by_name(maps);
|
|
up_write(maps__lock(maps));
|
|
}
|
|
|
|
static void __maps__purge(struct maps *maps)
|
|
{
|
|
struct map_rb_node *pos, *next;
|
|
|
|
maps__for_each_entry_safe(maps, pos, next) {
|
|
rb_erase_init(&pos->rb_node, maps__entries(maps));
|
|
map__put(pos->map);
|
|
free(pos);
|
|
}
|
|
}
|
|
|
|
static void maps__exit(struct maps *maps)
|
|
{
|
|
down_write(maps__lock(maps));
|
|
__maps__purge(maps);
|
|
up_write(maps__lock(maps));
|
|
}
|
|
|
|
bool maps__empty(struct maps *maps)
|
|
{
|
|
return !maps__first(maps);
|
|
}
|
|
|
|
struct maps *maps__new(struct machine *machine)
|
|
{
|
|
struct maps *maps = zalloc(sizeof(*maps));
|
|
|
|
if (maps != NULL)
|
|
maps__init(maps, machine);
|
|
|
|
return maps;
|
|
}
|
|
|
|
void maps__delete(struct maps *maps)
|
|
{
|
|
maps__exit(maps);
|
|
unwind__finish_access(maps);
|
|
free(maps);
|
|
}
|
|
|
|
struct maps *maps__get(struct maps *maps)
|
|
{
|
|
if (maps)
|
|
refcount_inc(&maps->refcnt);
|
|
|
|
return maps;
|
|
}
|
|
|
|
void maps__put(struct maps *maps)
|
|
{
|
|
if (maps && refcount_dec_and_test(&maps->refcnt))
|
|
maps__delete(maps);
|
|
}
|
|
|
|
struct symbol *maps__find_symbol(struct maps *maps, u64 addr, struct map **mapp)
|
|
{
|
|
struct map *map = maps__find(maps, addr);
|
|
|
|
/* Ensure map is loaded before using map->map_ip */
|
|
if (map != NULL && map__load(map) >= 0) {
|
|
if (mapp != NULL)
|
|
*mapp = map;
|
|
return map__find_symbol(map, map->map_ip(map, addr));
|
|
}
|
|
|
|
return NULL;
|
|
}
|
|
|
|
struct symbol *maps__find_symbol_by_name(struct maps *maps, const char *name, struct map **mapp)
|
|
{
|
|
struct symbol *sym;
|
|
struct map_rb_node *pos;
|
|
|
|
down_read(maps__lock(maps));
|
|
|
|
maps__for_each_entry(maps, pos) {
|
|
sym = map__find_symbol_by_name(pos->map, name);
|
|
|
|
if (sym == NULL)
|
|
continue;
|
|
if (!map__contains_symbol(pos->map, sym)) {
|
|
sym = NULL;
|
|
continue;
|
|
}
|
|
if (mapp != NULL)
|
|
*mapp = pos->map;
|
|
goto out;
|
|
}
|
|
|
|
sym = NULL;
|
|
out:
|
|
up_read(maps__lock(maps));
|
|
return sym;
|
|
}
|
|
|
|
int maps__find_ams(struct maps *maps, struct addr_map_symbol *ams)
|
|
{
|
|
if (ams->addr < ams->ms.map->start || ams->addr >= ams->ms.map->end) {
|
|
if (maps == NULL)
|
|
return -1;
|
|
ams->ms.map = maps__find(maps, ams->addr);
|
|
if (ams->ms.map == NULL)
|
|
return -1;
|
|
}
|
|
|
|
ams->al_addr = ams->ms.map->map_ip(ams->ms.map, ams->addr);
|
|
ams->ms.sym = map__find_symbol(ams->ms.map, ams->al_addr);
|
|
|
|
return ams->ms.sym ? 0 : -1;
|
|
}
|
|
|
|
size_t maps__fprintf(struct maps *maps, FILE *fp)
|
|
{
|
|
size_t printed = 0;
|
|
struct map_rb_node *pos;
|
|
|
|
down_read(maps__lock(maps));
|
|
|
|
maps__for_each_entry(maps, pos) {
|
|
printed += fprintf(fp, "Map:");
|
|
printed += map__fprintf(pos->map, fp);
|
|
if (verbose > 2) {
|
|
printed += dso__fprintf(pos->map->dso, fp);
|
|
printed += fprintf(fp, "--\n");
|
|
}
|
|
}
|
|
|
|
up_read(maps__lock(maps));
|
|
|
|
return printed;
|
|
}
|
|
|
|
int maps__fixup_overlappings(struct maps *maps, struct map *map, FILE *fp)
|
|
{
|
|
struct rb_root *root;
|
|
struct rb_node *next, *first;
|
|
int err = 0;
|
|
|
|
down_write(maps__lock(maps));
|
|
|
|
root = maps__entries(maps);
|
|
|
|
/*
|
|
* Find first map where end > map->start.
|
|
* Same as find_vma() in kernel.
|
|
*/
|
|
next = root->rb_node;
|
|
first = NULL;
|
|
while (next) {
|
|
struct map_rb_node *pos = rb_entry(next, struct map_rb_node, rb_node);
|
|
|
|
if (pos->map->end > map->start) {
|
|
first = next;
|
|
if (pos->map->start <= map->start)
|
|
break;
|
|
next = next->rb_left;
|
|
} else
|
|
next = next->rb_right;
|
|
}
|
|
|
|
next = first;
|
|
while (next) {
|
|
struct map_rb_node *pos = rb_entry(next, struct map_rb_node, rb_node);
|
|
next = rb_next(&pos->rb_node);
|
|
|
|
/*
|
|
* Stop if current map starts after map->end.
|
|
* Maps are ordered by start: next will not overlap for sure.
|
|
*/
|
|
if (pos->map->start >= map->end)
|
|
break;
|
|
|
|
if (verbose >= 2) {
|
|
|
|
if (use_browser) {
|
|
pr_debug("overlapping maps in %s (disable tui for more info)\n",
|
|
map->dso->name);
|
|
} else {
|
|
fputs("overlapping maps:\n", fp);
|
|
map__fprintf(map, fp);
|
|
map__fprintf(pos->map, fp);
|
|
}
|
|
}
|
|
|
|
rb_erase_init(&pos->rb_node, root);
|
|
/*
|
|
* Now check if we need to create new maps for areas not
|
|
* overlapped by the new map:
|
|
*/
|
|
if (map->start > pos->map->start) {
|
|
struct map *before = map__clone(pos->map);
|
|
|
|
if (before == NULL) {
|
|
err = -ENOMEM;
|
|
goto put_map;
|
|
}
|
|
|
|
before->end = map->start;
|
|
err = __maps__insert(maps, before);
|
|
if (err)
|
|
goto put_map;
|
|
|
|
if (verbose >= 2 && !use_browser)
|
|
map__fprintf(before, fp);
|
|
map__put(before);
|
|
}
|
|
|
|
if (map->end < pos->map->end) {
|
|
struct map *after = map__clone(pos->map);
|
|
|
|
if (after == NULL) {
|
|
err = -ENOMEM;
|
|
goto put_map;
|
|
}
|
|
|
|
after->start = map->end;
|
|
after->pgoff += map->end - pos->map->start;
|
|
assert(pos->map->map_ip(pos->map, map->end) ==
|
|
after->map_ip(after, map->end));
|
|
err = __maps__insert(maps, after);
|
|
if (err)
|
|
goto put_map;
|
|
|
|
if (verbose >= 2 && !use_browser)
|
|
map__fprintf(after, fp);
|
|
map__put(after);
|
|
}
|
|
put_map:
|
|
map__put(pos->map);
|
|
|
|
if (err)
|
|
goto out;
|
|
}
|
|
|
|
err = 0;
|
|
out:
|
|
up_write(maps__lock(maps));
|
|
return err;
|
|
}
|
|
|
|
/*
|
|
* XXX This should not really _copy_ te maps, but refcount them.
|
|
*/
|
|
int maps__clone(struct thread *thread, struct maps *parent)
|
|
{
|
|
struct maps *maps = thread->maps;
|
|
int err;
|
|
struct map_rb_node *rb_node;
|
|
|
|
down_read(maps__lock(parent));
|
|
|
|
maps__for_each_entry(parent, rb_node) {
|
|
struct map *new = map__clone(rb_node->map);
|
|
|
|
if (new == NULL) {
|
|
err = -ENOMEM;
|
|
goto out_unlock;
|
|
}
|
|
|
|
err = unwind__prepare_access(maps, new, NULL);
|
|
if (err)
|
|
goto out_unlock;
|
|
|
|
err = maps__insert(maps, new);
|
|
if (err)
|
|
goto out_unlock;
|
|
|
|
map__put(new);
|
|
}
|
|
|
|
err = 0;
|
|
out_unlock:
|
|
up_read(maps__lock(parent));
|
|
return err;
|
|
}
|
|
|
|
struct map_rb_node *maps__find_node(struct maps *maps, struct map *map)
|
|
{
|
|
struct map_rb_node *rb_node;
|
|
|
|
maps__for_each_entry(maps, rb_node) {
|
|
if (rb_node->map == map)
|
|
return rb_node;
|
|
}
|
|
return NULL;
|
|
}
|
|
|
|
struct map *maps__find(struct maps *maps, u64 ip)
|
|
{
|
|
struct rb_node *p;
|
|
struct map_rb_node *m;
|
|
|
|
|
|
down_read(maps__lock(maps));
|
|
|
|
p = maps__entries(maps)->rb_node;
|
|
while (p != NULL) {
|
|
m = rb_entry(p, struct map_rb_node, rb_node);
|
|
if (ip < m->map->start)
|
|
p = p->rb_left;
|
|
else if (ip >= m->map->end)
|
|
p = p->rb_right;
|
|
else
|
|
goto out;
|
|
}
|
|
|
|
m = NULL;
|
|
out:
|
|
up_read(maps__lock(maps));
|
|
return m ? m->map : NULL;
|
|
}
|
|
|
|
struct map_rb_node *maps__first(struct maps *maps)
|
|
{
|
|
struct rb_node *first = rb_first(maps__entries(maps));
|
|
|
|
if (first)
|
|
return rb_entry(first, struct map_rb_node, rb_node);
|
|
return NULL;
|
|
}
|
|
|
|
struct map_rb_node *map_rb_node__next(struct map_rb_node *node)
|
|
{
|
|
struct rb_node *next;
|
|
|
|
if (!node)
|
|
return NULL;
|
|
|
|
next = rb_next(&node->rb_node);
|
|
|
|
if (!next)
|
|
return NULL;
|
|
|
|
return rb_entry(next, struct map_rb_node, rb_node);
|
|
}
|