| // SPDX-License-Identifier: GPL-2.0 |
| #include <errno.h> |
| #include <stdlib.h> |
| #include <linux/zalloc.h> |
| #include "debug.h" |
| #include "dso.h" |
| #include "map.h" |
| #include "maps.h" |
| #include "rwsem.h" |
| #include "thread.h" |
| #include "ui/ui.h" |
| #include "unwind.h" |
| #include <internal/rc_check.h> |
| |
| /* |
| * Locking/sorting note: |
| * |
| * Sorting is done with the write lock, iteration and binary searching happens |
| * under the read lock requiring being sorted. There is a race between sorting |
| * releasing the write lock and acquiring the read lock for iteration/searching |
| * where another thread could insert and break the sorting of the maps. In |
| * practice inserting maps should be rare meaning that the race shouldn't lead |
| * to live lock. Removal of maps doesn't break being sorted. |
| */ |
| |
| DECLARE_RC_STRUCT(maps) { |
| struct rw_semaphore lock; |
| /** |
| * @maps_by_address: array of maps sorted by their starting address if |
| * maps_by_address_sorted is true. |
| */ |
| struct map **maps_by_address; |
| /** |
| * @maps_by_name: optional array of maps sorted by their dso name if |
| * maps_by_name_sorted is true. |
| */ |
| struct map **maps_by_name; |
| struct machine *machine; |
| #ifdef HAVE_LIBUNWIND_SUPPORT |
| void *addr_space; |
| const struct unwind_libunwind_ops *unwind_libunwind_ops; |
| #endif |
| refcount_t refcnt; |
| /** |
| * @nr_maps: number of maps_by_address, and possibly maps_by_name, |
| * entries that contain maps. |
| */ |
| unsigned int nr_maps; |
| /** |
| * @nr_maps_allocated: number of entries in maps_by_address and possibly |
| * maps_by_name. |
| */ |
| unsigned int nr_maps_allocated; |
| /** |
| * @last_search_by_name_idx: cache of last found by name entry's index |
| * as frequent searches for the same dso name are common. |
| */ |
| unsigned int last_search_by_name_idx; |
| /** @maps_by_address_sorted: is maps_by_address sorted. */ |
| bool maps_by_address_sorted; |
| /** @maps_by_name_sorted: is maps_by_name sorted. */ |
| bool maps_by_name_sorted; |
| /** @ends_broken: does the map contain a map where end values are unset/unsorted? */ |
| bool ends_broken; |
| }; |
| |
| static void check_invariants(const struct maps *maps __maybe_unused) |
| { |
| #ifndef NDEBUG |
| assert(RC_CHK_ACCESS(maps)->nr_maps <= RC_CHK_ACCESS(maps)->nr_maps_allocated); |
| for (unsigned int i = 0; i < RC_CHK_ACCESS(maps)->nr_maps; i++) { |
| struct map *map = RC_CHK_ACCESS(maps)->maps_by_address[i]; |
| |
| /* Check map is well-formed. */ |
| assert(map__end(map) == 0 || map__start(map) <= map__end(map)); |
| /* Expect at least 1 reference count. */ |
| assert(refcount_read(map__refcnt(map)) > 0); |
| |
| if (map__dso(map) && map__dso(map)->kernel) |
| assert(RC_CHK_EQUAL(map__kmap(map)->kmaps, maps)); |
| |
| if (i > 0) { |
| struct map *prev = RC_CHK_ACCESS(maps)->maps_by_address[i - 1]; |
| |
| /* If addresses are sorted... */ |
| if (RC_CHK_ACCESS(maps)->maps_by_address_sorted) { |
| /* Maps should be in start address order. */ |
| assert(map__start(prev) <= map__start(map)); |
| /* |
| * If the ends of maps aren't broken (during |
| * construction) then they should be ordered |
| * too. |
| */ |
| if (!RC_CHK_ACCESS(maps)->ends_broken) { |
| assert(map__end(prev) <= map__end(map)); |
| assert(map__end(prev) <= map__start(map) || |
| map__start(prev) == map__start(map)); |
| } |
| } |
| } |
| } |
| if (RC_CHK_ACCESS(maps)->maps_by_name) { |
| for (unsigned int i = 0; i < RC_CHK_ACCESS(maps)->nr_maps; i++) { |
| struct map *map = RC_CHK_ACCESS(maps)->maps_by_name[i]; |
| |
| /* |
| * Maps by name maps should be in maps_by_address, so |
| * the reference count should be higher. |
| */ |
| assert(refcount_read(map__refcnt(map)) > 1); |
| } |
| } |
| #endif |
| } |
| |
| static struct map **maps__maps_by_address(const struct maps *maps) |
| { |
| return RC_CHK_ACCESS(maps)->maps_by_address; |
| } |
| |
| static void maps__set_maps_by_address(struct maps *maps, struct map **new) |
| { |
| RC_CHK_ACCESS(maps)->maps_by_address = new; |
| |
| } |
| |
| static struct map ***maps__maps_by_name_addr(struct maps *maps) |
| { |
| return &RC_CHK_ACCESS(maps)->maps_by_name; |
| } |
| |
| static void maps__set_nr_maps_allocated(struct maps *maps, unsigned int nr_maps_allocated) |
| { |
| RC_CHK_ACCESS(maps)->nr_maps_allocated = nr_maps_allocated; |
| } |
| |
| static void maps__set_nr_maps(struct maps *maps, unsigned int nr_maps) |
| { |
| RC_CHK_ACCESS(maps)->nr_maps = nr_maps; |
| } |
| |
| /* Not in the header, to aid reference counting. */ |
| static struct map **maps__maps_by_name(const struct maps *maps) |
| { |
| return RC_CHK_ACCESS(maps)->maps_by_name; |
| |
| } |
| |
| static void maps__set_maps_by_name(struct maps *maps, struct map **new) |
| { |
| RC_CHK_ACCESS(maps)->maps_by_name = new; |
| |
| } |
| |
| static bool maps__maps_by_address_sorted(const struct maps *maps) |
| { |
| return RC_CHK_ACCESS(maps)->maps_by_address_sorted; |
| } |
| |
| static void maps__set_maps_by_address_sorted(struct maps *maps, bool value) |
| { |
| RC_CHK_ACCESS(maps)->maps_by_address_sorted = value; |
| } |
| |
| static bool maps__maps_by_name_sorted(const struct maps *maps) |
| { |
| return RC_CHK_ACCESS(maps)->maps_by_name_sorted; |
| } |
| |
| static void maps__set_maps_by_name_sorted(struct maps *maps, bool value) |
| { |
| RC_CHK_ACCESS(maps)->maps_by_name_sorted = value; |
| } |
| |
| struct machine *maps__machine(const struct maps *maps) |
| { |
| return RC_CHK_ACCESS(maps)->machine; |
| } |
| |
| unsigned int maps__nr_maps(const struct maps *maps) |
| { |
| return RC_CHK_ACCESS(maps)->nr_maps; |
| } |
| |
| refcount_t *maps__refcnt(struct maps *maps) |
| { |
| return &RC_CHK_ACCESS(maps)->refcnt; |
| } |
| |
| #ifdef HAVE_LIBUNWIND_SUPPORT |
| void *maps__addr_space(const struct maps *maps) |
| { |
| return RC_CHK_ACCESS(maps)->addr_space; |
| } |
| |
| void maps__set_addr_space(struct maps *maps, void *addr_space) |
| { |
| RC_CHK_ACCESS(maps)->addr_space = addr_space; |
| } |
| |
| const struct unwind_libunwind_ops *maps__unwind_libunwind_ops(const struct maps *maps) |
| { |
| return RC_CHK_ACCESS(maps)->unwind_libunwind_ops; |
| } |
| |
| void maps__set_unwind_libunwind_ops(struct maps *maps, const struct unwind_libunwind_ops *ops) |
| { |
| RC_CHK_ACCESS(maps)->unwind_libunwind_ops = ops; |
| } |
| #endif |
| |
| static struct rw_semaphore *maps__lock(struct maps *maps) |
| { |
| /* |
| * When the lock is acquired or released the maps invariants should |
| * hold. |
| */ |
| check_invariants(maps); |
| return &RC_CHK_ACCESS(maps)->lock; |
| } |
| |
| static void maps__init(struct maps *maps, struct machine *machine) |
| { |
| init_rwsem(maps__lock(maps)); |
| RC_CHK_ACCESS(maps)->maps_by_address = NULL; |
| RC_CHK_ACCESS(maps)->maps_by_name = NULL; |
| RC_CHK_ACCESS(maps)->machine = machine; |
| #ifdef HAVE_LIBUNWIND_SUPPORT |
| RC_CHK_ACCESS(maps)->addr_space = NULL; |
| RC_CHK_ACCESS(maps)->unwind_libunwind_ops = NULL; |
| #endif |
| refcount_set(maps__refcnt(maps), 1); |
| RC_CHK_ACCESS(maps)->nr_maps = 0; |
| RC_CHK_ACCESS(maps)->nr_maps_allocated = 0; |
| RC_CHK_ACCESS(maps)->last_search_by_name_idx = 0; |
| RC_CHK_ACCESS(maps)->maps_by_address_sorted = true; |
| RC_CHK_ACCESS(maps)->maps_by_name_sorted = false; |
| } |
| |
| static void maps__exit(struct maps *maps) |
| { |
| struct map **maps_by_address = maps__maps_by_address(maps); |
| struct map **maps_by_name = maps__maps_by_name(maps); |
| |
| for (unsigned int i = 0; i < maps__nr_maps(maps); i++) { |
| map__zput(maps_by_address[i]); |
| if (maps_by_name) |
| map__zput(maps_by_name[i]); |
| } |
| zfree(&maps_by_address); |
| zfree(&maps_by_name); |
| unwind__finish_access(maps); |
| } |
| |
| struct maps *maps__new(struct machine *machine) |
| { |
| struct maps *result; |
| RC_STRUCT(maps) *maps = zalloc(sizeof(*maps)); |
| |
| if (ADD_RC_CHK(result, maps)) |
| maps__init(result, machine); |
| |
| return result; |
| } |
| |
| static void maps__delete(struct maps *maps) |
| { |
| maps__exit(maps); |
| RC_CHK_FREE(maps); |
| } |
| |
| struct maps *maps__get(struct maps *maps) |
| { |
| struct maps *result; |
| |
| if (RC_CHK_GET(result, maps)) |
| refcount_inc(maps__refcnt(maps)); |
| |
| return result; |
| } |
| |
| void maps__put(struct maps *maps) |
| { |
| if (maps && refcount_dec_and_test(maps__refcnt(maps))) |
| maps__delete(maps); |
| else |
| RC_CHK_PUT(maps); |
| } |
| |
| static void __maps__free_maps_by_name(struct maps *maps) |
| { |
| /* |
| * Free everything to try to do it from the rbtree in the next search |
| */ |
| for (unsigned int i = 0; i < maps__nr_maps(maps); i++) |
| map__put(maps__maps_by_name(maps)[i]); |
| |
| zfree(&RC_CHK_ACCESS(maps)->maps_by_name); |
| } |
| |
| static int map__start_cmp(const void *a, const void *b) |
| { |
| const struct map *map_a = *(const struct map * const *)a; |
| const struct map *map_b = *(const struct map * const *)b; |
| u64 map_a_start = map__start(map_a); |
| u64 map_b_start = map__start(map_b); |
| |
| if (map_a_start == map_b_start) { |
| u64 map_a_end = map__end(map_a); |
| u64 map_b_end = map__end(map_b); |
| |
| if (map_a_end == map_b_end) { |
| /* Ensure maps with the same addresses have a fixed order. */ |
| if (RC_CHK_ACCESS(map_a) == RC_CHK_ACCESS(map_b)) |
| return 0; |
| return (intptr_t)RC_CHK_ACCESS(map_a) > (intptr_t)RC_CHK_ACCESS(map_b) |
| ? 1 : -1; |
| } |
| return map_a_end > map_b_end ? 1 : -1; |
| } |
| return map_a_start > map_b_start ? 1 : -1; |
| } |
| |
| static void __maps__sort_by_address(struct maps *maps) |
| { |
| if (maps__maps_by_address_sorted(maps)) |
| return; |
| |
| qsort(maps__maps_by_address(maps), |
| maps__nr_maps(maps), |
| sizeof(struct map *), |
| map__start_cmp); |
| maps__set_maps_by_address_sorted(maps, true); |
| } |
| |
| static void maps__sort_by_address(struct maps *maps) |
| { |
| down_write(maps__lock(maps)); |
| __maps__sort_by_address(maps); |
| up_write(maps__lock(maps)); |
| } |
| |
| static int map__strcmp(const void *a, const void *b) |
| { |
| const struct map *map_a = *(const struct map * const *)a; |
| const struct map *map_b = *(const struct map * const *)b; |
| const struct dso *dso_a = map__dso(map_a); |
| const struct dso *dso_b = map__dso(map_b); |
| int ret = strcmp(dso_a->short_name, dso_b->short_name); |
| |
| if (ret == 0 && RC_CHK_ACCESS(map_a) != RC_CHK_ACCESS(map_b)) { |
| /* Ensure distinct but name equal maps have an order. */ |
| return map__start_cmp(a, b); |
| } |
| return ret; |
| } |
| |
| static int maps__sort_by_name(struct maps *maps) |
| { |
| int err = 0; |
| down_write(maps__lock(maps)); |
| if (!maps__maps_by_name_sorted(maps)) { |
| struct map **maps_by_name = maps__maps_by_name(maps); |
| |
| if (!maps_by_name) { |
| maps_by_name = malloc(RC_CHK_ACCESS(maps)->nr_maps_allocated * |
| sizeof(*maps_by_name)); |
| if (!maps_by_name) |
| err = -ENOMEM; |
| else { |
| struct map **maps_by_address = maps__maps_by_address(maps); |
| unsigned int n = maps__nr_maps(maps); |
| |
| maps__set_maps_by_name(maps, maps_by_name); |
| for (unsigned int i = 0; i < n; i++) |
| maps_by_name[i] = map__get(maps_by_address[i]); |
| } |
| } |
| if (!err) { |
| qsort(maps_by_name, |
| maps__nr_maps(maps), |
| sizeof(struct map *), |
| map__strcmp); |
| maps__set_maps_by_name_sorted(maps, true); |
| } |
| } |
| up_write(maps__lock(maps)); |
| return err; |
| } |
| |
| static unsigned int maps__by_address_index(const struct maps *maps, const struct map *map) |
| { |
| struct map **maps_by_address = maps__maps_by_address(maps); |
| |
| if (maps__maps_by_address_sorted(maps)) { |
| struct map **mapp = |
| bsearch(&map, maps__maps_by_address(maps), maps__nr_maps(maps), |
| sizeof(*mapp), map__start_cmp); |
| |
| if (mapp) |
| return mapp - maps_by_address; |
| } else { |
| for (unsigned int i = 0; i < maps__nr_maps(maps); i++) { |
| if (RC_CHK_ACCESS(maps_by_address[i]) == RC_CHK_ACCESS(map)) |
| return i; |
| } |
| } |
| pr_err("Map missing from maps"); |
| return -1; |
| } |
| |
| static unsigned int maps__by_name_index(const struct maps *maps, const struct map *map) |
| { |
| struct map **maps_by_name = maps__maps_by_name(maps); |
| |
| if (maps__maps_by_name_sorted(maps)) { |
| struct map **mapp = |
| bsearch(&map, maps_by_name, maps__nr_maps(maps), |
| sizeof(*mapp), map__strcmp); |
| |
| if (mapp) |
| return mapp - maps_by_name; |
| } else { |
| for (unsigned int i = 0; i < maps__nr_maps(maps); i++) { |
| if (RC_CHK_ACCESS(maps_by_name[i]) == RC_CHK_ACCESS(map)) |
| return i; |
| } |
| } |
| pr_err("Map missing from maps"); |
| return -1; |
| } |
| |
| static int __maps__insert(struct maps *maps, struct map *new) |
| { |
| struct map **maps_by_address = maps__maps_by_address(maps); |
| struct map **maps_by_name = maps__maps_by_name(maps); |
| const struct dso *dso = map__dso(new); |
| unsigned int nr_maps = maps__nr_maps(maps); |
| unsigned int nr_allocate = RC_CHK_ACCESS(maps)->nr_maps_allocated; |
| |
| if (nr_maps + 1 > nr_allocate) { |
| nr_allocate = !nr_allocate ? 32 : nr_allocate * 2; |
| |
| maps_by_address = realloc(maps_by_address, nr_allocate * sizeof(new)); |
| if (!maps_by_address) |
| return -ENOMEM; |
| |
| maps__set_maps_by_address(maps, maps_by_address); |
| if (maps_by_name) { |
| maps_by_name = realloc(maps_by_name, nr_allocate * sizeof(new)); |
| if (!maps_by_name) { |
| /* |
| * If by name fails, just disable by name and it will |
| * recompute next time it is required. |
| */ |
| __maps__free_maps_by_name(maps); |
| } |
| maps__set_maps_by_name(maps, maps_by_name); |
| } |
| RC_CHK_ACCESS(maps)->nr_maps_allocated = nr_allocate; |
| } |
| /* Insert the value at the end. */ |
| maps_by_address[nr_maps] = map__get(new); |
| if (maps_by_name) |
| maps_by_name[nr_maps] = map__get(new); |
| |
| nr_maps++; |
| RC_CHK_ACCESS(maps)->nr_maps = nr_maps; |
| |
| /* |
| * Recompute if things are sorted. If things are inserted in a sorted |
| * manner, for example by processing /proc/pid/maps, then no |
| * sorting/resorting will be necessary. |
| */ |
| if (nr_maps == 1) { |
| /* If there's just 1 entry then maps are sorted. */ |
| maps__set_maps_by_address_sorted(maps, true); |
| maps__set_maps_by_name_sorted(maps, maps_by_name != NULL); |
| } else { |
| /* Sorted if maps were already sorted and this map starts after the last one. */ |
| maps__set_maps_by_address_sorted(maps, |
| maps__maps_by_address_sorted(maps) && |
| map__end(maps_by_address[nr_maps - 2]) <= map__start(new)); |
| maps__set_maps_by_name_sorted(maps, false); |
| } |
| if (map__end(new) < map__start(new)) |
| RC_CHK_ACCESS(maps)->ends_broken = true; |
| if (dso && dso->kernel) { |
| struct kmap *kmap = map__kmap(new); |
| |
| if (kmap) |
| kmap->kmaps = maps; |
| else |
| pr_err("Internal error: kernel dso with non kernel map\n"); |
| } |
| return 0; |
| } |
| |
| int maps__insert(struct maps *maps, struct map *map) |
| { |
| int ret; |
| |
| down_write(maps__lock(maps)); |
| ret = __maps__insert(maps, map); |
| up_write(maps__lock(maps)); |
| return ret; |
| } |
| |
| static void __maps__remove(struct maps *maps, struct map *map) |
| { |
| struct map **maps_by_address = maps__maps_by_address(maps); |
| struct map **maps_by_name = maps__maps_by_name(maps); |
| unsigned int nr_maps = maps__nr_maps(maps); |
| unsigned int address_idx; |
| |
| /* Slide later mappings over the one to remove */ |
| address_idx = maps__by_address_index(maps, map); |
| map__put(maps_by_address[address_idx]); |
| memmove(&maps_by_address[address_idx], |
| &maps_by_address[address_idx + 1], |
| (nr_maps - address_idx - 1) * sizeof(*maps_by_address)); |
| |
| if (maps_by_name) { |
| unsigned int name_idx = maps__by_name_index(maps, map); |
| |
| map__put(maps_by_name[name_idx]); |
| memmove(&maps_by_name[name_idx], |
| &maps_by_name[name_idx + 1], |
| (nr_maps - name_idx - 1) * sizeof(*maps_by_name)); |
| } |
| |
| --RC_CHK_ACCESS(maps)->nr_maps; |
| } |
| |
| void maps__remove(struct maps *maps, struct map *map) |
| { |
| down_write(maps__lock(maps)); |
| __maps__remove(maps, map); |
| up_write(maps__lock(maps)); |
| } |
| |
| bool maps__empty(struct maps *maps) |
| { |
| bool res; |
| |
| down_read(maps__lock(maps)); |
| res = maps__nr_maps(maps) == 0; |
| up_read(maps__lock(maps)); |
| |
| return res; |
| } |
| |
| bool maps__equal(struct maps *a, struct maps *b) |
| { |
| return RC_CHK_EQUAL(a, b); |
| } |
| |
| int maps__for_each_map(struct maps *maps, int (*cb)(struct map *map, void *data), void *data) |
| { |
| bool done = false; |
| int ret = 0; |
| |
| /* See locking/sorting note. */ |
| while (!done) { |
| down_read(maps__lock(maps)); |
| if (maps__maps_by_address_sorted(maps)) { |
| /* |
| * maps__for_each_map callbacks may buggily/unsafely |
| * insert into maps_by_address. Deliberately reload |
| * maps__nr_maps and maps_by_address on each iteration |
| * to avoid using memory freed by maps__insert growing |
| * the array - this may cause maps to be skipped or |
| * repeated. |
| */ |
| for (unsigned int i = 0; i < maps__nr_maps(maps); i++) { |
| struct map **maps_by_address = maps__maps_by_address(maps); |
| struct map *map = maps_by_address[i]; |
| |
| ret = cb(map, data); |
| if (ret) |
| break; |
| } |
| done = true; |
| } |
| up_read(maps__lock(maps)); |
| if (!done) |
| maps__sort_by_address(maps); |
| } |
| return ret; |
| } |
| |
| void maps__remove_maps(struct maps *maps, bool (*cb)(struct map *map, void *data), void *data) |
| { |
| struct map **maps_by_address; |
| |
| down_write(maps__lock(maps)); |
| |
| maps_by_address = maps__maps_by_address(maps); |
| for (unsigned int i = 0; i < maps__nr_maps(maps);) { |
| if (cb(maps_by_address[i], data)) |
| __maps__remove(maps, maps_by_address[i]); |
| else |
| i++; |
| } |
| up_write(maps__lock(maps)); |
| } |
| |
| struct symbol *maps__find_symbol(struct maps *maps, u64 addr, struct map **mapp) |
| { |
| struct map *map = maps__find(maps, addr); |
| struct symbol *result = NULL; |
| |
| /* Ensure map is loaded before using map->map_ip */ |
| if (map != NULL && map__load(map) >= 0) |
| result = map__find_symbol(map, map__map_ip(map, addr)); |
| |
| if (mapp) |
| *mapp = map; |
| else |
| map__put(map); |
| |
| return result; |
| } |
| |
| struct maps__find_symbol_by_name_args { |
| struct map **mapp; |
| const char *name; |
| struct symbol *sym; |
| }; |
| |
| static int maps__find_symbol_by_name_cb(struct map *map, void *data) |
| { |
| struct maps__find_symbol_by_name_args *args = data; |
| |
| args->sym = map__find_symbol_by_name(map, args->name); |
| if (!args->sym) |
| return 0; |
| |
| if (!map__contains_symbol(map, args->sym)) { |
| args->sym = NULL; |
| return 0; |
| } |
| |
| if (args->mapp != NULL) |
| *args->mapp = map__get(map); |
| return 1; |
| } |
| |
| struct symbol *maps__find_symbol_by_name(struct maps *maps, const char *name, struct map **mapp) |
| { |
| struct maps__find_symbol_by_name_args args = { |
| .mapp = mapp, |
| .name = name, |
| .sym = NULL, |
| }; |
| |
| maps__for_each_map(maps, maps__find_symbol_by_name_cb, &args); |
| return args.sym; |
| } |
| |
| int maps__find_ams(struct maps *maps, struct addr_map_symbol *ams) |
| { |
| if (ams->addr < map__start(ams->ms.map) || ams->addr >= map__end(ams->ms.map)) { |
| if (maps == NULL) |
| return -1; |
| ams->ms.map = maps__find(maps, ams->addr); |
| if (ams->ms.map == NULL) |
| return -1; |
| } |
| |
| ams->al_addr = map__map_ip(ams->ms.map, ams->addr); |
| ams->ms.sym = map__find_symbol(ams->ms.map, ams->al_addr); |
| |
| return ams->ms.sym ? 0 : -1; |
| } |
| |
| struct maps__fprintf_args { |
| FILE *fp; |
| size_t printed; |
| }; |
| |
| static int maps__fprintf_cb(struct map *map, void *data) |
| { |
| struct maps__fprintf_args *args = data; |
| |
| args->printed += fprintf(args->fp, "Map:"); |
| args->printed += map__fprintf(map, args->fp); |
| if (verbose > 2) { |
| args->printed += dso__fprintf(map__dso(map), args->fp); |
| args->printed += fprintf(args->fp, "--\n"); |
| } |
| return 0; |
| } |
| |
| size_t maps__fprintf(struct maps *maps, FILE *fp) |
| { |
| struct maps__fprintf_args args = { |
| .fp = fp, |
| .printed = 0, |
| }; |
| |
| maps__for_each_map(maps, maps__fprintf_cb, &args); |
| |
| return args.printed; |
| } |
| |
| /* |
| * Find first map where end > map->start. |
| * Same as find_vma() in kernel. |
| */ |
| static unsigned int first_ending_after(struct maps *maps, const struct map *map) |
| { |
| struct map **maps_by_address = maps__maps_by_address(maps); |
| int low = 0, high = (int)maps__nr_maps(maps) - 1, first = high + 1; |
| |
| assert(maps__maps_by_address_sorted(maps)); |
| if (low <= high && map__end(maps_by_address[0]) > map__start(map)) |
| return 0; |
| |
| while (low <= high) { |
| int mid = (low + high) / 2; |
| struct map *pos = maps_by_address[mid]; |
| |
| if (map__end(pos) > map__start(map)) { |
| first = mid; |
| if (map__start(pos) <= map__start(map)) { |
| /* Entry overlaps map. */ |
| break; |
| } |
| high = mid - 1; |
| } else |
| low = mid + 1; |
| } |
| return first; |
| } |
| |
| /* |
| * Adds new to maps, if new overlaps existing entries then the existing maps are |
| * adjusted or removed so that new fits without overlapping any entries. |
| */ |
| static int __maps__fixup_overlap_and_insert(struct maps *maps, struct map *new) |
| { |
| struct map **maps_by_address; |
| int err = 0; |
| FILE *fp = debug_file(); |
| |
| sort_again: |
| if (!maps__maps_by_address_sorted(maps)) |
| __maps__sort_by_address(maps); |
| |
| maps_by_address = maps__maps_by_address(maps); |
| /* |
| * Iterate through entries where the end of the existing entry is |
| * greater-than the new map's start. |
| */ |
| for (unsigned int i = first_ending_after(maps, new); i < maps__nr_maps(maps); ) { |
| struct map *pos = maps_by_address[i]; |
| struct map *before = NULL, *after = NULL; |
| |
| /* |
| * Stop if current map starts after map->end. |
| * Maps are ordered by start: next will not overlap for sure. |
| */ |
| if (map__start(pos) >= map__end(new)) |
| break; |
| |
| if (use_browser) { |
| pr_debug("overlapping maps in %s (disable tui for more info)\n", |
| map__dso(new)->name); |
| } else if (verbose >= 2) { |
| pr_debug("overlapping maps:\n"); |
| map__fprintf(new, fp); |
| map__fprintf(pos, fp); |
| } |
| |
| /* |
| * Now check if we need to create new maps for areas not |
| * overlapped by the new map: |
| */ |
| if (map__start(new) > map__start(pos)) { |
| /* Map starts within existing map. Need to shorten the existing map. */ |
| before = map__clone(pos); |
| |
| if (before == NULL) { |
| err = -ENOMEM; |
| goto out_err; |
| } |
| map__set_end(before, map__start(new)); |
| |
| if (verbose >= 2 && !use_browser) |
| map__fprintf(before, fp); |
| } |
| if (map__end(new) < map__end(pos)) { |
| /* The new map isn't as long as the existing map. */ |
| after = map__clone(pos); |
| |
| if (after == NULL) { |
| map__zput(before); |
| err = -ENOMEM; |
| goto out_err; |
| } |
| |
| map__set_start(after, map__end(new)); |
| map__add_pgoff(after, map__end(new) - map__start(pos)); |
| assert(map__map_ip(pos, map__end(new)) == |
| map__map_ip(after, map__end(new))); |
| |
| if (verbose >= 2 && !use_browser) |
| map__fprintf(after, fp); |
| } |
| /* |
| * If adding one entry, for `before` or `after`, we can replace |
| * the existing entry. If both `before` and `after` are |
| * necessary than an insert is needed. If the existing entry |
| * entirely overlaps the existing entry it can just be removed. |
| */ |
| if (before) { |
| map__put(maps_by_address[i]); |
| maps_by_address[i] = before; |
| /* Maps are still ordered, go to next one. */ |
| i++; |
| if (after) { |
| __maps__insert(maps, after); |
| map__put(after); |
| if (!maps__maps_by_address_sorted(maps)) { |
| /* |
| * Sorting broken so invariants don't |
| * hold, sort and go again. |
| */ |
| goto sort_again; |
| } |
| /* |
| * Maps are still ordered, skip after and go to |
| * next one (terminate loop). |
| */ |
| i++; |
| } |
| } else if (after) { |
| map__put(maps_by_address[i]); |
| maps_by_address[i] = after; |
| /* Maps are ordered, go to next one. */ |
| i++; |
| } else { |
| __maps__remove(maps, pos); |
| /* |
| * Maps are ordered but no need to increase `i` as the |
| * later maps were moved down. |
| */ |
| } |
| check_invariants(maps); |
| } |
| /* Add the map. */ |
| __maps__insert(maps, new); |
| out_err: |
| return err; |
| } |
| |
| int maps__fixup_overlap_and_insert(struct maps *maps, struct map *new) |
| { |
| int err; |
| |
| down_write(maps__lock(maps)); |
| err = __maps__fixup_overlap_and_insert(maps, new); |
| up_write(maps__lock(maps)); |
| return err; |
| } |
| |
| int maps__copy_from(struct maps *dest, struct maps *parent) |
| { |
| /* Note, if struct map were immutable then cloning could use ref counts. */ |
| struct map **parent_maps_by_address; |
| int err = 0; |
| unsigned int n; |
| |
| down_write(maps__lock(dest)); |
| down_read(maps__lock(parent)); |
| |
| parent_maps_by_address = maps__maps_by_address(parent); |
| n = maps__nr_maps(parent); |
| if (maps__nr_maps(dest) == 0) { |
| /* No existing mappings so just copy from parent to avoid reallocs in insert. */ |
| unsigned int nr_maps_allocated = RC_CHK_ACCESS(parent)->nr_maps_allocated; |
| struct map **dest_maps_by_address = |
| malloc(nr_maps_allocated * sizeof(struct map *)); |
| struct map **dest_maps_by_name = NULL; |
| |
| if (!dest_maps_by_address) |
| err = -ENOMEM; |
| else { |
| if (maps__maps_by_name(parent)) { |
| dest_maps_by_name = |
| malloc(nr_maps_allocated * sizeof(struct map *)); |
| } |
| |
| RC_CHK_ACCESS(dest)->maps_by_address = dest_maps_by_address; |
| RC_CHK_ACCESS(dest)->maps_by_name = dest_maps_by_name; |
| RC_CHK_ACCESS(dest)->nr_maps_allocated = nr_maps_allocated; |
| } |
| |
| for (unsigned int i = 0; !err && i < n; i++) { |
| struct map *pos = parent_maps_by_address[i]; |
| struct map *new = map__clone(pos); |
| |
| if (!new) |
| err = -ENOMEM; |
| else { |
| err = unwind__prepare_access(dest, new, NULL); |
| if (!err) { |
| dest_maps_by_address[i] = new; |
| if (dest_maps_by_name) |
| dest_maps_by_name[i] = map__get(new); |
| RC_CHK_ACCESS(dest)->nr_maps = i + 1; |
| } |
| } |
| if (err) |
| map__put(new); |
| } |
| maps__set_maps_by_address_sorted(dest, maps__maps_by_address_sorted(parent)); |
| if (!err) { |
| RC_CHK_ACCESS(dest)->last_search_by_name_idx = |
| RC_CHK_ACCESS(parent)->last_search_by_name_idx; |
| maps__set_maps_by_name_sorted(dest, |
| dest_maps_by_name && |
| maps__maps_by_name_sorted(parent)); |
| } else { |
| RC_CHK_ACCESS(dest)->last_search_by_name_idx = 0; |
| maps__set_maps_by_name_sorted(dest, false); |
| } |
| } else { |
| /* Unexpected copying to a maps containing entries. */ |
| for (unsigned int i = 0; !err && i < n; i++) { |
| struct map *pos = parent_maps_by_address[i]; |
| struct map *new = map__clone(pos); |
| |
| if (!new) |
| err = -ENOMEM; |
| else { |
| err = unwind__prepare_access(dest, new, NULL); |
| if (!err) |
| err = __maps__insert(dest, new); |
| } |
| map__put(new); |
| } |
| } |
| up_read(maps__lock(parent)); |
| up_write(maps__lock(dest)); |
| return err; |
| } |
| |
| static int map__addr_cmp(const void *key, const void *entry) |
| { |
| const u64 ip = *(const u64 *)key; |
| const struct map *map = *(const struct map * const *)entry; |
| |
| if (ip < map__start(map)) |
| return -1; |
| if (ip >= map__end(map)) |
| return 1; |
| return 0; |
| } |
| |
| struct map *maps__find(struct maps *maps, u64 ip) |
| { |
| struct map *result = NULL; |
| bool done = false; |
| |
| /* See locking/sorting note. */ |
| while (!done) { |
| down_read(maps__lock(maps)); |
| if (maps__maps_by_address_sorted(maps)) { |
| struct map **mapp = |
| bsearch(&ip, maps__maps_by_address(maps), maps__nr_maps(maps), |
| sizeof(*mapp), map__addr_cmp); |
| |
| if (mapp) |
| result = map__get(*mapp); |
| done = true; |
| } |
| up_read(maps__lock(maps)); |
| if (!done) |
| maps__sort_by_address(maps); |
| } |
| return result; |
| } |
| |
| static int map__strcmp_name(const void *name, const void *b) |
| { |
| const struct dso *dso = map__dso(*(const struct map **)b); |
| |
| return strcmp(name, dso->short_name); |
| } |
| |
| struct map *maps__find_by_name(struct maps *maps, const char *name) |
| { |
| struct map *result = NULL; |
| bool done = false; |
| |
| /* See locking/sorting note. */ |
| while (!done) { |
| unsigned int i; |
| |
| down_read(maps__lock(maps)); |
| |
| /* First check last found entry. */ |
| i = RC_CHK_ACCESS(maps)->last_search_by_name_idx; |
| if (i < maps__nr_maps(maps) && maps__maps_by_name(maps)) { |
| struct dso *dso = map__dso(maps__maps_by_name(maps)[i]); |
| |
| if (dso && strcmp(dso->short_name, name) == 0) { |
| result = map__get(maps__maps_by_name(maps)[i]); |
| done = true; |
| } |
| } |
| |
| /* Second search sorted array. */ |
| if (!done && maps__maps_by_name_sorted(maps)) { |
| struct map **mapp = |
| bsearch(name, maps__maps_by_name(maps), maps__nr_maps(maps), |
| sizeof(*mapp), map__strcmp_name); |
| |
| if (mapp) { |
| result = map__get(*mapp); |
| i = mapp - maps__maps_by_name(maps); |
| RC_CHK_ACCESS(maps)->last_search_by_name_idx = i; |
| } |
| done = true; |
| } |
| up_read(maps__lock(maps)); |
| if (!done) { |
| /* Sort and retry binary search. */ |
| if (maps__sort_by_name(maps)) { |
| /* |
| * Memory allocation failed do linear search |
| * through address sorted maps. |
| */ |
| struct map **maps_by_address; |
| unsigned int n; |
| |
| down_read(maps__lock(maps)); |
| maps_by_address = maps__maps_by_address(maps); |
| n = maps__nr_maps(maps); |
| for (i = 0; i < n; i++) { |
| struct map *pos = maps_by_address[i]; |
| struct dso *dso = map__dso(pos); |
| |
| if (dso && strcmp(dso->short_name, name) == 0) { |
| result = map__get(pos); |
| break; |
| } |
| } |
| up_read(maps__lock(maps)); |
| done = true; |
| } |
| } |
| } |
| return result; |
| } |
| |
| struct map *maps__find_next_entry(struct maps *maps, struct map *map) |
| { |
| unsigned int i; |
| struct map *result = NULL; |
| |
| down_read(maps__lock(maps)); |
| i = maps__by_address_index(maps, map); |
| if (i < maps__nr_maps(maps)) |
| result = map__get(maps__maps_by_address(maps)[i]); |
| |
| up_read(maps__lock(maps)); |
| return result; |
| } |
| |
| void maps__fixup_end(struct maps *maps) |
| { |
| struct map **maps_by_address; |
| unsigned int n; |
| |
| down_write(maps__lock(maps)); |
| if (!maps__maps_by_address_sorted(maps)) |
| __maps__sort_by_address(maps); |
| |
| maps_by_address = maps__maps_by_address(maps); |
| n = maps__nr_maps(maps); |
| for (unsigned int i = 1; i < n; i++) { |
| struct map *prev = maps_by_address[i - 1]; |
| struct map *curr = maps_by_address[i]; |
| |
| if (!map__end(prev) || map__end(prev) > map__start(curr)) |
| map__set_end(prev, map__start(curr)); |
| } |
| |
| /* |
| * We still haven't the actual symbols, so guess the |
| * last map final address. |
| */ |
| if (n > 0 && !map__end(maps_by_address[n - 1])) |
| map__set_end(maps_by_address[n - 1], ~0ULL); |
| |
| RC_CHK_ACCESS(maps)->ends_broken = false; |
| |
| up_write(maps__lock(maps)); |
| } |
| |
| /* |
| * Merges map into maps by splitting the new map within the existing map |
| * regions. |
| */ |
| int maps__merge_in(struct maps *kmaps, struct map *new_map) |
| { |
| unsigned int first_after_, kmaps__nr_maps; |
| struct map **kmaps_maps_by_address; |
| struct map **merged_maps_by_address; |
| unsigned int merged_nr_maps_allocated; |
| |
| /* First try under a read lock. */ |
| while (true) { |
| down_read(maps__lock(kmaps)); |
| if (maps__maps_by_address_sorted(kmaps)) |
| break; |
| |
| up_read(maps__lock(kmaps)); |
| |
| /* First after binary search requires sorted maps. Sort and try again. */ |
| maps__sort_by_address(kmaps); |
| } |
| first_after_ = first_ending_after(kmaps, new_map); |
| kmaps_maps_by_address = maps__maps_by_address(kmaps); |
| |
| if (first_after_ >= maps__nr_maps(kmaps) || |
| map__start(kmaps_maps_by_address[first_after_]) >= map__end(new_map)) { |
| /* No overlap so regular insert suffices. */ |
| up_read(maps__lock(kmaps)); |
| return maps__insert(kmaps, new_map); |
| } |
| up_read(maps__lock(kmaps)); |
| |
| /* Plain insert with a read-lock failed, try again now with the write lock. */ |
| down_write(maps__lock(kmaps)); |
| if (!maps__maps_by_address_sorted(kmaps)) |
| __maps__sort_by_address(kmaps); |
| |
| first_after_ = first_ending_after(kmaps, new_map); |
| kmaps_maps_by_address = maps__maps_by_address(kmaps); |
| kmaps__nr_maps = maps__nr_maps(kmaps); |
| |
| if (first_after_ >= kmaps__nr_maps || |
| map__start(kmaps_maps_by_address[first_after_]) >= map__end(new_map)) { |
| /* No overlap so regular insert suffices. */ |
| int ret = __maps__insert(kmaps, new_map); |
| up_write(maps__lock(kmaps)); |
| return ret; |
| } |
| /* Array to merge into, possibly 1 more for the sake of new_map. */ |
| merged_nr_maps_allocated = RC_CHK_ACCESS(kmaps)->nr_maps_allocated; |
| if (kmaps__nr_maps + 1 == merged_nr_maps_allocated) |
| merged_nr_maps_allocated++; |
| |
| merged_maps_by_address = malloc(merged_nr_maps_allocated * sizeof(*merged_maps_by_address)); |
| if (!merged_maps_by_address) { |
| up_write(maps__lock(kmaps)); |
| return -ENOMEM; |
| } |
| maps__set_maps_by_address(kmaps, merged_maps_by_address); |
| maps__set_maps_by_address_sorted(kmaps, true); |
| zfree(maps__maps_by_name_addr(kmaps)); |
| maps__set_maps_by_name_sorted(kmaps, true); |
| maps__set_nr_maps_allocated(kmaps, merged_nr_maps_allocated); |
| |
| /* Copy entries before the new_map that can't overlap. */ |
| for (unsigned int i = 0; i < first_after_; i++) |
| merged_maps_by_address[i] = map__get(kmaps_maps_by_address[i]); |
| |
| maps__set_nr_maps(kmaps, first_after_); |
| |
| /* Add the new map, it will be split when the later overlapping mappings are added. */ |
| __maps__insert(kmaps, new_map); |
| |
| /* Insert mappings after new_map, splitting new_map in the process. */ |
| for (unsigned int i = first_after_; i < kmaps__nr_maps; i++) |
| __maps__fixup_overlap_and_insert(kmaps, kmaps_maps_by_address[i]); |
| |
| /* Copy the maps from merged into kmaps. */ |
| for (unsigned int i = 0; i < kmaps__nr_maps; i++) |
| map__zput(kmaps_maps_by_address[i]); |
| |
| free(kmaps_maps_by_address); |
| up_write(maps__lock(kmaps)); |
| return 0; |
| } |
| |
| void maps__load_first(struct maps *maps) |
| { |
| down_read(maps__lock(maps)); |
| |
| if (maps__nr_maps(maps) > 0) |
| map__load(maps__maps_by_address(maps)[0]); |
| |
| up_read(maps__lock(maps)); |
| } |