| // SPDX-License-Identifier: GPL-2.0 |
| /* |
| * Copyright (C) 2009-2011, Frederic Weisbecker <fweisbec@gmail.com> |
| * |
| * Handle the callchains from the stream in an ad-hoc radix tree and then |
| * sort them in an rbtree. |
| * |
| * Using a radix for code path provides a fast retrieval and factorizes |
| * memory use. Also that lets us use the paths in a hierarchical graph view. |
| * |
| */ |
| |
| #include <inttypes.h> |
| #include <stdlib.h> |
| #include <stdio.h> |
| #include <stdbool.h> |
| #include <errno.h> |
| #include <math.h> |
| #include <linux/string.h> |
| #include <linux/zalloc.h> |
| |
| #include "asm/bug.h" |
| |
| #include "debug.h" |
| #include "dso.h" |
| #include "event.h" |
| #include "hist.h" |
| #include "sort.h" |
| #include "machine.h" |
| #include "map.h" |
| #include "callchain.h" |
| #include "branch.h" |
| #include "symbol.h" |
| #include "util.h" |
| #include "../perf.h" |
| |
| #define CALLCHAIN_PARAM_DEFAULT \ |
| .mode = CHAIN_GRAPH_ABS, \ |
| .min_percent = 0.5, \ |
| .order = ORDER_CALLEE, \ |
| .key = CCKEY_FUNCTION, \ |
| .value = CCVAL_PERCENT, \ |
| |
| struct callchain_param callchain_param = { |
| CALLCHAIN_PARAM_DEFAULT |
| }; |
| |
| /* |
| * Are there any events usind DWARF callchains? |
| * |
| * I.e. |
| * |
| * -e cycles/call-graph=dwarf/ |
| */ |
| bool dwarf_callchain_users; |
| |
| struct callchain_param callchain_param_default = { |
| CALLCHAIN_PARAM_DEFAULT |
| }; |
| |
| __thread struct callchain_cursor callchain_cursor; |
| |
| int parse_callchain_record_opt(const char *arg, struct callchain_param *param) |
| { |
| return parse_callchain_record(arg, param); |
| } |
| |
| static int parse_callchain_mode(const char *value) |
| { |
| if (!strncmp(value, "graph", strlen(value))) { |
| callchain_param.mode = CHAIN_GRAPH_ABS; |
| return 0; |
| } |
| if (!strncmp(value, "flat", strlen(value))) { |
| callchain_param.mode = CHAIN_FLAT; |
| return 0; |
| } |
| if (!strncmp(value, "fractal", strlen(value))) { |
| callchain_param.mode = CHAIN_GRAPH_REL; |
| return 0; |
| } |
| if (!strncmp(value, "folded", strlen(value))) { |
| callchain_param.mode = CHAIN_FOLDED; |
| return 0; |
| } |
| return -1; |
| } |
| |
| static int parse_callchain_order(const char *value) |
| { |
| if (!strncmp(value, "caller", strlen(value))) { |
| callchain_param.order = ORDER_CALLER; |
| callchain_param.order_set = true; |
| return 0; |
| } |
| if (!strncmp(value, "callee", strlen(value))) { |
| callchain_param.order = ORDER_CALLEE; |
| callchain_param.order_set = true; |
| return 0; |
| } |
| return -1; |
| } |
| |
| static int parse_callchain_sort_key(const char *value) |
| { |
| if (!strncmp(value, "function", strlen(value))) { |
| callchain_param.key = CCKEY_FUNCTION; |
| return 0; |
| } |
| if (!strncmp(value, "address", strlen(value))) { |
| callchain_param.key = CCKEY_ADDRESS; |
| return 0; |
| } |
| if (!strncmp(value, "srcline", strlen(value))) { |
| callchain_param.key = CCKEY_SRCLINE; |
| return 0; |
| } |
| if (!strncmp(value, "branch", strlen(value))) { |
| callchain_param.branch_callstack = 1; |
| return 0; |
| } |
| return -1; |
| } |
| |
| static int parse_callchain_value(const char *value) |
| { |
| if (!strncmp(value, "percent", strlen(value))) { |
| callchain_param.value = CCVAL_PERCENT; |
| return 0; |
| } |
| if (!strncmp(value, "period", strlen(value))) { |
| callchain_param.value = CCVAL_PERIOD; |
| return 0; |
| } |
| if (!strncmp(value, "count", strlen(value))) { |
| callchain_param.value = CCVAL_COUNT; |
| return 0; |
| } |
| return -1; |
| } |
| |
| static int get_stack_size(const char *str, unsigned long *_size) |
| { |
| char *endptr; |
| unsigned long size; |
| unsigned long max_size = round_down(USHRT_MAX, sizeof(u64)); |
| |
| size = strtoul(str, &endptr, 0); |
| |
| do { |
| if (*endptr) |
| break; |
| |
| size = round_up(size, sizeof(u64)); |
| if (!size || size > max_size) |
| break; |
| |
| *_size = size; |
| return 0; |
| |
| } while (0); |
| |
| pr_err("callchain: Incorrect stack dump size (max %ld): %s\n", |
| max_size, str); |
| return -1; |
| } |
| |
| static int |
| __parse_callchain_report_opt(const char *arg, bool allow_record_opt) |
| { |
| char *tok; |
| char *endptr, *saveptr = NULL; |
| bool minpcnt_set = false; |
| bool record_opt_set = false; |
| bool try_stack_size = false; |
| |
| callchain_param.enabled = true; |
| symbol_conf.use_callchain = true; |
| |
| if (!arg) |
| return 0; |
| |
| while ((tok = strtok_r((char *)arg, ",", &saveptr)) != NULL) { |
| if (!strncmp(tok, "none", strlen(tok))) { |
| callchain_param.mode = CHAIN_NONE; |
| callchain_param.enabled = false; |
| symbol_conf.use_callchain = false; |
| return 0; |
| } |
| |
| if (!parse_callchain_mode(tok) || |
| !parse_callchain_order(tok) || |
| !parse_callchain_sort_key(tok) || |
| !parse_callchain_value(tok)) { |
| /* parsing ok - move on to the next */ |
| try_stack_size = false; |
| goto next; |
| } else if (allow_record_opt && !record_opt_set) { |
| if (parse_callchain_record(tok, &callchain_param)) |
| goto try_numbers; |
| |
| /* assume that number followed by 'dwarf' is stack size */ |
| if (callchain_param.record_mode == CALLCHAIN_DWARF) |
| try_stack_size = true; |
| |
| record_opt_set = true; |
| goto next; |
| } |
| |
| try_numbers: |
| if (try_stack_size) { |
| unsigned long size = 0; |
| |
| if (get_stack_size(tok, &size) < 0) |
| return -1; |
| callchain_param.dump_size = size; |
| try_stack_size = false; |
| } else if (!minpcnt_set) { |
| /* try to get the min percent */ |
| callchain_param.min_percent = strtod(tok, &endptr); |
| if (tok == endptr) |
| return -1; |
| minpcnt_set = true; |
| } else { |
| /* try print limit at last */ |
| callchain_param.print_limit = strtoul(tok, &endptr, 0); |
| if (tok == endptr) |
| return -1; |
| } |
| next: |
| arg = NULL; |
| } |
| |
| if (callchain_register_param(&callchain_param) < 0) { |
| pr_err("Can't register callchain params\n"); |
| return -1; |
| } |
| return 0; |
| } |
| |
| int parse_callchain_report_opt(const char *arg) |
| { |
| return __parse_callchain_report_opt(arg, false); |
| } |
| |
| int parse_callchain_top_opt(const char *arg) |
| { |
| return __parse_callchain_report_opt(arg, true); |
| } |
| |
| int parse_callchain_record(const char *arg, struct callchain_param *param) |
| { |
| char *tok, *name, *saveptr = NULL; |
| char *buf; |
| int ret = -1; |
| |
| /* We need buffer that we know we can write to. */ |
| buf = malloc(strlen(arg) + 1); |
| if (!buf) |
| return -ENOMEM; |
| |
| strcpy(buf, arg); |
| |
| tok = strtok_r((char *)buf, ",", &saveptr); |
| name = tok ? : (char *)buf; |
| |
| do { |
| /* Framepointer style */ |
| if (!strncmp(name, "fp", sizeof("fp"))) { |
| ret = 0; |
| param->record_mode = CALLCHAIN_FP; |
| |
| tok = strtok_r(NULL, ",", &saveptr); |
| if (tok) { |
| unsigned long size; |
| |
| size = strtoul(tok, &name, 0); |
| if (size < (unsigned) sysctl__max_stack()) |
| param->max_stack = size; |
| } |
| break; |
| |
| /* Dwarf style */ |
| } else if (!strncmp(name, "dwarf", sizeof("dwarf"))) { |
| const unsigned long default_stack_dump_size = 8192; |
| |
| ret = 0; |
| param->record_mode = CALLCHAIN_DWARF; |
| param->dump_size = default_stack_dump_size; |
| dwarf_callchain_users = true; |
| |
| tok = strtok_r(NULL, ",", &saveptr); |
| if (tok) { |
| unsigned long size = 0; |
| |
| ret = get_stack_size(tok, &size); |
| param->dump_size = size; |
| } |
| } else if (!strncmp(name, "lbr", sizeof("lbr"))) { |
| if (!strtok_r(NULL, ",", &saveptr)) { |
| param->record_mode = CALLCHAIN_LBR; |
| ret = 0; |
| } else |
| pr_err("callchain: No more arguments " |
| "needed for --call-graph lbr\n"); |
| break; |
| } else { |
| pr_err("callchain: Unknown --call-graph option " |
| "value: %s\n", arg); |
| break; |
| } |
| |
| } while (0); |
| |
| free(buf); |
| return ret; |
| } |
| |
| int perf_callchain_config(const char *var, const char *value) |
| { |
| char *endptr; |
| |
| if (!strstarts(var, "call-graph.")) |
| return 0; |
| var += sizeof("call-graph.") - 1; |
| |
| if (!strcmp(var, "record-mode")) |
| return parse_callchain_record_opt(value, &callchain_param); |
| if (!strcmp(var, "dump-size")) { |
| unsigned long size = 0; |
| int ret; |
| |
| ret = get_stack_size(value, &size); |
| callchain_param.dump_size = size; |
| |
| return ret; |
| } |
| if (!strcmp(var, "print-type")){ |
| int ret; |
| ret = parse_callchain_mode(value); |
| if (ret == -1) |
| pr_err("Invalid callchain mode: %s\n", value); |
| return ret; |
| } |
| if (!strcmp(var, "order")){ |
| int ret; |
| ret = parse_callchain_order(value); |
| if (ret == -1) |
| pr_err("Invalid callchain order: %s\n", value); |
| return ret; |
| } |
| if (!strcmp(var, "sort-key")){ |
| int ret; |
| ret = parse_callchain_sort_key(value); |
| if (ret == -1) |
| pr_err("Invalid callchain sort key: %s\n", value); |
| return ret; |
| } |
| if (!strcmp(var, "threshold")) { |
| callchain_param.min_percent = strtod(value, &endptr); |
| if (value == endptr) { |
| pr_err("Invalid callchain threshold: %s\n", value); |
| return -1; |
| } |
| } |
| if (!strcmp(var, "print-limit")) { |
| callchain_param.print_limit = strtod(value, &endptr); |
| if (value == endptr) { |
| pr_err("Invalid callchain print limit: %s\n", value); |
| return -1; |
| } |
| } |
| |
| return 0; |
| } |
| |
| static void |
| rb_insert_callchain(struct rb_root *root, struct callchain_node *chain, |
| enum chain_mode mode) |
| { |
| struct rb_node **p = &root->rb_node; |
| struct rb_node *parent = NULL; |
| struct callchain_node *rnode; |
| u64 chain_cumul = callchain_cumul_hits(chain); |
| |
| while (*p) { |
| u64 rnode_cumul; |
| |
| parent = *p; |
| rnode = rb_entry(parent, struct callchain_node, rb_node); |
| rnode_cumul = callchain_cumul_hits(rnode); |
| |
| switch (mode) { |
| case CHAIN_FLAT: |
| case CHAIN_FOLDED: |
| if (rnode->hit < chain->hit) |
| p = &(*p)->rb_left; |
| else |
| p = &(*p)->rb_right; |
| break; |
| case CHAIN_GRAPH_ABS: /* Falldown */ |
| case CHAIN_GRAPH_REL: |
| if (rnode_cumul < chain_cumul) |
| p = &(*p)->rb_left; |
| else |
| p = &(*p)->rb_right; |
| break; |
| case CHAIN_NONE: |
| default: |
| break; |
| } |
| } |
| |
| rb_link_node(&chain->rb_node, parent, p); |
| rb_insert_color(&chain->rb_node, root); |
| } |
| |
| static void |
| __sort_chain_flat(struct rb_root *rb_root, struct callchain_node *node, |
| u64 min_hit) |
| { |
| struct rb_node *n; |
| struct callchain_node *child; |
| |
| n = rb_first(&node->rb_root_in); |
| while (n) { |
| child = rb_entry(n, struct callchain_node, rb_node_in); |
| n = rb_next(n); |
| |
| __sort_chain_flat(rb_root, child, min_hit); |
| } |
| |
| if (node->hit && node->hit >= min_hit) |
| rb_insert_callchain(rb_root, node, CHAIN_FLAT); |
| } |
| |
| /* |
| * Once we get every callchains from the stream, we can now |
| * sort them by hit |
| */ |
| static void |
| sort_chain_flat(struct rb_root *rb_root, struct callchain_root *root, |
| u64 min_hit, struct callchain_param *param __maybe_unused) |
| { |
| *rb_root = RB_ROOT; |
| __sort_chain_flat(rb_root, &root->node, min_hit); |
| } |
| |
| static void __sort_chain_graph_abs(struct callchain_node *node, |
| u64 min_hit) |
| { |
| struct rb_node *n; |
| struct callchain_node *child; |
| |
| node->rb_root = RB_ROOT; |
| n = rb_first(&node->rb_root_in); |
| |
| while (n) { |
| child = rb_entry(n, struct callchain_node, rb_node_in); |
| n = rb_next(n); |
| |
| __sort_chain_graph_abs(child, min_hit); |
| if (callchain_cumul_hits(child) >= min_hit) |
| rb_insert_callchain(&node->rb_root, child, |
| CHAIN_GRAPH_ABS); |
| } |
| } |
| |
| static void |
| sort_chain_graph_abs(struct rb_root *rb_root, struct callchain_root *chain_root, |
| u64 min_hit, struct callchain_param *param __maybe_unused) |
| { |
| __sort_chain_graph_abs(&chain_root->node, min_hit); |
| rb_root->rb_node = chain_root->node.rb_root.rb_node; |
| } |
| |
| static void __sort_chain_graph_rel(struct callchain_node *node, |
| double min_percent) |
| { |
| struct rb_node *n; |
| struct callchain_node *child; |
| u64 min_hit; |
| |
| node->rb_root = RB_ROOT; |
| min_hit = ceil(node->children_hit * min_percent); |
| |
| n = rb_first(&node->rb_root_in); |
| while (n) { |
| child = rb_entry(n, struct callchain_node, rb_node_in); |
| n = rb_next(n); |
| |
| __sort_chain_graph_rel(child, min_percent); |
| if (callchain_cumul_hits(child) >= min_hit) |
| rb_insert_callchain(&node->rb_root, child, |
| CHAIN_GRAPH_REL); |
| } |
| } |
| |
| static void |
| sort_chain_graph_rel(struct rb_root *rb_root, struct callchain_root *chain_root, |
| u64 min_hit __maybe_unused, struct callchain_param *param) |
| { |
| __sort_chain_graph_rel(&chain_root->node, param->min_percent / 100.0); |
| rb_root->rb_node = chain_root->node.rb_root.rb_node; |
| } |
| |
| int callchain_register_param(struct callchain_param *param) |
| { |
| switch (param->mode) { |
| case CHAIN_GRAPH_ABS: |
| param->sort = sort_chain_graph_abs; |
| break; |
| case CHAIN_GRAPH_REL: |
| param->sort = sort_chain_graph_rel; |
| break; |
| case CHAIN_FLAT: |
| case CHAIN_FOLDED: |
| param->sort = sort_chain_flat; |
| break; |
| case CHAIN_NONE: |
| default: |
| return -1; |
| } |
| return 0; |
| } |
| |
| /* |
| * Create a child for a parent. If inherit_children, then the new child |
| * will become the new parent of it's parent children |
| */ |
| static struct callchain_node * |
| create_child(struct callchain_node *parent, bool inherit_children) |
| { |
| struct callchain_node *new; |
| |
| new = zalloc(sizeof(*new)); |
| if (!new) { |
| perror("not enough memory to create child for code path tree"); |
| return NULL; |
| } |
| new->parent = parent; |
| INIT_LIST_HEAD(&new->val); |
| INIT_LIST_HEAD(&new->parent_val); |
| |
| if (inherit_children) { |
| struct rb_node *n; |
| struct callchain_node *child; |
| |
| new->rb_root_in = parent->rb_root_in; |
| parent->rb_root_in = RB_ROOT; |
| |
| n = rb_first(&new->rb_root_in); |
| while (n) { |
| child = rb_entry(n, struct callchain_node, rb_node_in); |
| child->parent = new; |
| n = rb_next(n); |
| } |
| |
| /* make it the first child */ |
| rb_link_node(&new->rb_node_in, NULL, &parent->rb_root_in.rb_node); |
| rb_insert_color(&new->rb_node_in, &parent->rb_root_in); |
| } |
| |
| return new; |
| } |
| |
| |
| /* |
| * Fill the node with callchain values |
| */ |
| static int |
| fill_node(struct callchain_node *node, struct callchain_cursor *cursor) |
| { |
| struct callchain_cursor_node *cursor_node; |
| |
| node->val_nr = cursor->nr - cursor->pos; |
| if (!node->val_nr) |
| pr_warning("Warning: empty node in callchain tree\n"); |
| |
| cursor_node = callchain_cursor_current(cursor); |
| |
| while (cursor_node) { |
| struct callchain_list *call; |
| |
| call = zalloc(sizeof(*call)); |
| if (!call) { |
| perror("not enough memory for the code path tree"); |
| return -1; |
| } |
| call->ip = cursor_node->ip; |
| call->ms = cursor_node->ms; |
| call->ms.map = map__get(call->ms.map); |
| call->srcline = cursor_node->srcline; |
| |
| if (cursor_node->branch) { |
| call->branch_count = 1; |
| |
| if (cursor_node->branch_from) { |
| /* |
| * branch_from is set with value somewhere else |
| * to imply it's "to" of a branch. |
| */ |
| call->brtype_stat.branch_to = true; |
| |
| if (cursor_node->branch_flags.predicted) |
| call->predicted_count = 1; |
| |
| if (cursor_node->branch_flags.abort) |
| call->abort_count = 1; |
| |
| branch_type_count(&call->brtype_stat, |
| &cursor_node->branch_flags, |
| cursor_node->branch_from, |
| cursor_node->ip); |
| } else { |
| /* |
| * It's "from" of a branch |
| */ |
| call->brtype_stat.branch_to = false; |
| call->cycles_count = |
| cursor_node->branch_flags.cycles; |
| call->iter_count = cursor_node->nr_loop_iter; |
| call->iter_cycles = cursor_node->iter_cycles; |
| } |
| } |
| |
| list_add_tail(&call->list, &node->val); |
| |
| callchain_cursor_advance(cursor); |
| cursor_node = callchain_cursor_current(cursor); |
| } |
| return 0; |
| } |
| |
| static struct callchain_node * |
| add_child(struct callchain_node *parent, |
| struct callchain_cursor *cursor, |
| u64 period) |
| { |
| struct callchain_node *new; |
| |
| new = create_child(parent, false); |
| if (new == NULL) |
| return NULL; |
| |
| if (fill_node(new, cursor) < 0) { |
| struct callchain_list *call, *tmp; |
| |
| list_for_each_entry_safe(call, tmp, &new->val, list) { |
| list_del_init(&call->list); |
| map__zput(call->ms.map); |
| free(call); |
| } |
| free(new); |
| return NULL; |
| } |
| |
| new->children_hit = 0; |
| new->hit = period; |
| new->children_count = 0; |
| new->count = 1; |
| return new; |
| } |
| |
| enum match_result { |
| MATCH_ERROR = -1, |
| MATCH_EQ, |
| MATCH_LT, |
| MATCH_GT, |
| }; |
| |
| static enum match_result match_chain_strings(const char *left, |
| const char *right) |
| { |
| enum match_result ret = MATCH_EQ; |
| int cmp; |
| |
| if (left && right) |
| cmp = strcmp(left, right); |
| else if (!left && right) |
| cmp = 1; |
| else if (left && !right) |
| cmp = -1; |
| else |
| return MATCH_ERROR; |
| |
| if (cmp != 0) |
| ret = cmp < 0 ? MATCH_LT : MATCH_GT; |
| |
| return ret; |
| } |
| |
| /* |
| * We need to always use relative addresses because we're aggregating |
| * callchains from multiple threads, i.e. different address spaces, so |
| * comparing absolute addresses make no sense as a symbol in a DSO may end up |
| * in a different address when used in a different binary or even the same |
| * binary but with some sort of address randomization technique, thus we need |
| * to compare just relative addresses. -acme |
| */ |
| static enum match_result match_chain_dso_addresses(struct map *left_map, u64 left_ip, |
| struct map *right_map, u64 right_ip) |
| { |
| struct dso *left_dso = left_map ? map__dso(left_map) : NULL; |
| struct dso *right_dso = right_map ? map__dso(right_map) : NULL; |
| |
| if (left_dso != right_dso) |
| return left_dso < right_dso ? MATCH_LT : MATCH_GT; |
| |
| if (left_ip != right_ip) |
| return left_ip < right_ip ? MATCH_LT : MATCH_GT; |
| |
| return MATCH_EQ; |
| } |
| |
| static enum match_result match_chain(struct callchain_cursor_node *node, |
| struct callchain_list *cnode) |
| { |
| enum match_result match = MATCH_ERROR; |
| |
| switch (callchain_param.key) { |
| case CCKEY_SRCLINE: |
| match = match_chain_strings(cnode->srcline, node->srcline); |
| if (match != MATCH_ERROR) |
| break; |
| /* otherwise fall-back to symbol-based comparison below */ |
| fallthrough; |
| case CCKEY_FUNCTION: |
| if (node->ms.sym && cnode->ms.sym) { |
| /* |
| * Compare inlined frames based on their symbol name |
| * because different inlined frames will have the same |
| * symbol start. Otherwise do a faster comparison based |
| * on the symbol start address. |
| */ |
| if (cnode->ms.sym->inlined || node->ms.sym->inlined) { |
| match = match_chain_strings(cnode->ms.sym->name, |
| node->ms.sym->name); |
| if (match != MATCH_ERROR) |
| break; |
| } else { |
| match = match_chain_dso_addresses(cnode->ms.map, cnode->ms.sym->start, |
| node->ms.map, node->ms.sym->start); |
| break; |
| } |
| } |
| /* otherwise fall-back to IP-based comparison below */ |
| fallthrough; |
| case CCKEY_ADDRESS: |
| default: |
| match = match_chain_dso_addresses(cnode->ms.map, cnode->ip, node->ms.map, node->ip); |
| break; |
| } |
| |
| if (match == MATCH_EQ && node->branch) { |
| cnode->branch_count++; |
| |
| if (node->branch_from) { |
| /* |
| * It's "to" of a branch |
| */ |
| cnode->brtype_stat.branch_to = true; |
| |
| if (node->branch_flags.predicted) |
| cnode->predicted_count++; |
| |
| if (node->branch_flags.abort) |
| cnode->abort_count++; |
| |
| branch_type_count(&cnode->brtype_stat, |
| &node->branch_flags, |
| node->branch_from, |
| node->ip); |
| } else { |
| /* |
| * It's "from" of a branch |
| */ |
| cnode->brtype_stat.branch_to = false; |
| cnode->cycles_count += node->branch_flags.cycles; |
| cnode->iter_count += node->nr_loop_iter; |
| cnode->iter_cycles += node->iter_cycles; |
| cnode->from_count++; |
| } |
| } |
| |
| return match; |
| } |
| |
| /* |
| * Split the parent in two parts (a new child is created) and |
| * give a part of its callchain to the created child. |
| * Then create another child to host the given callchain of new branch |
| */ |
| static int |
| split_add_child(struct callchain_node *parent, |
| struct callchain_cursor *cursor, |
| struct callchain_list *to_split, |
| u64 idx_parents, u64 idx_local, u64 period) |
| { |
| struct callchain_node *new; |
| struct list_head *old_tail; |
| unsigned int idx_total = idx_parents + idx_local; |
| |
| /* split */ |
| new = create_child(parent, true); |
| if (new == NULL) |
| return -1; |
| |
| /* split the callchain and move a part to the new child */ |
| old_tail = parent->val.prev; |
| list_del_range(&to_split->list, old_tail); |
| new->val.next = &to_split->list; |
| new->val.prev = old_tail; |
| to_split->list.prev = &new->val; |
| old_tail->next = &new->val; |
| |
| /* split the hits */ |
| new->hit = parent->hit; |
| new->children_hit = parent->children_hit; |
| parent->children_hit = callchain_cumul_hits(new); |
| new->val_nr = parent->val_nr - idx_local; |
| parent->val_nr = idx_local; |
| new->count = parent->count; |
| new->children_count = parent->children_count; |
| parent->children_count = callchain_cumul_counts(new); |
| |
| /* create a new child for the new branch if any */ |
| if (idx_total < cursor->nr) { |
| struct callchain_node *first; |
| struct callchain_list *cnode; |
| struct callchain_cursor_node *node; |
| struct rb_node *p, **pp; |
| |
| parent->hit = 0; |
| parent->children_hit += period; |
| parent->count = 0; |
| parent->children_count += 1; |
| |
| node = callchain_cursor_current(cursor); |
| new = add_child(parent, cursor, period); |
| if (new == NULL) |
| return -1; |
| |
| /* |
| * This is second child since we moved parent's children |
| * to new (first) child above. |
| */ |
| p = parent->rb_root_in.rb_node; |
| first = rb_entry(p, struct callchain_node, rb_node_in); |
| cnode = list_first_entry(&first->val, struct callchain_list, |
| list); |
| |
| if (match_chain(node, cnode) == MATCH_LT) |
| pp = &p->rb_left; |
| else |
| pp = &p->rb_right; |
| |
| rb_link_node(&new->rb_node_in, p, pp); |
| rb_insert_color(&new->rb_node_in, &parent->rb_root_in); |
| } else { |
| parent->hit = period; |
| parent->count = 1; |
| } |
| return 0; |
| } |
| |
| static enum match_result |
| append_chain(struct callchain_node *root, |
| struct callchain_cursor *cursor, |
| u64 period); |
| |
| static int |
| append_chain_children(struct callchain_node *root, |
| struct callchain_cursor *cursor, |
| u64 period) |
| { |
| struct callchain_node *rnode; |
| struct callchain_cursor_node *node; |
| struct rb_node **p = &root->rb_root_in.rb_node; |
| struct rb_node *parent = NULL; |
| |
| node = callchain_cursor_current(cursor); |
| if (!node) |
| return -1; |
| |
| /* lookup in children */ |
| while (*p) { |
| enum match_result ret; |
| |
| parent = *p; |
| rnode = rb_entry(parent, struct callchain_node, rb_node_in); |
| |
| /* If at least first entry matches, rely to children */ |
| ret = append_chain(rnode, cursor, period); |
| if (ret == MATCH_EQ) |
| goto inc_children_hit; |
| if (ret == MATCH_ERROR) |
| return -1; |
| |
| if (ret == MATCH_LT) |
| p = &parent->rb_left; |
| else |
| p = &parent->rb_right; |
| } |
| /* nothing in children, add to the current node */ |
| rnode = add_child(root, cursor, period); |
| if (rnode == NULL) |
| return -1; |
| |
| rb_link_node(&rnode->rb_node_in, parent, p); |
| rb_insert_color(&rnode->rb_node_in, &root->rb_root_in); |
| |
| inc_children_hit: |
| root->children_hit += period; |
| root->children_count++; |
| return 0; |
| } |
| |
| static enum match_result |
| append_chain(struct callchain_node *root, |
| struct callchain_cursor *cursor, |
| u64 period) |
| { |
| struct callchain_list *cnode; |
| u64 start = cursor->pos; |
| bool found = false; |
| u64 matches; |
| enum match_result cmp = MATCH_ERROR; |
| |
| /* |
| * Lookup in the current node |
| * If we have a symbol, then compare the start to match |
| * anywhere inside a function, unless function |
| * mode is disabled. |
| */ |
| list_for_each_entry(cnode, &root->val, list) { |
| struct callchain_cursor_node *node; |
| |
| node = callchain_cursor_current(cursor); |
| if (!node) |
| break; |
| |
| cmp = match_chain(node, cnode); |
| if (cmp != MATCH_EQ) |
| break; |
| |
| found = true; |
| |
| callchain_cursor_advance(cursor); |
| } |
| |
| /* matches not, relay no the parent */ |
| if (!found) { |
| WARN_ONCE(cmp == MATCH_ERROR, "Chain comparison error\n"); |
| return cmp; |
| } |
| |
| matches = cursor->pos - start; |
| |
| /* we match only a part of the node. Split it and add the new chain */ |
| if (matches < root->val_nr) { |
| if (split_add_child(root, cursor, cnode, start, matches, |
| period) < 0) |
| return MATCH_ERROR; |
| |
| return MATCH_EQ; |
| } |
| |
| /* we match 100% of the path, increment the hit */ |
| if (matches == root->val_nr && cursor->pos == cursor->nr) { |
| root->hit += period; |
| root->count++; |
| return MATCH_EQ; |
| } |
| |
| /* We match the node and still have a part remaining */ |
| if (append_chain_children(root, cursor, period) < 0) |
| return MATCH_ERROR; |
| |
| return MATCH_EQ; |
| } |
| |
| int callchain_append(struct callchain_root *root, |
| struct callchain_cursor *cursor, |
| u64 period) |
| { |
| if (!cursor->nr) |
| return 0; |
| |
| callchain_cursor_commit(cursor); |
| |
| if (append_chain_children(&root->node, cursor, period) < 0) |
| return -1; |
| |
| if (cursor->nr > root->max_depth) |
| root->max_depth = cursor->nr; |
| |
| return 0; |
| } |
| |
| static int |
| merge_chain_branch(struct callchain_cursor *cursor, |
| struct callchain_node *dst, struct callchain_node *src) |
| { |
| struct callchain_cursor_node **old_last = cursor->last; |
| struct callchain_node *child; |
| struct callchain_list *list, *next_list; |
| struct rb_node *n; |
| int old_pos = cursor->nr; |
| int err = 0; |
| |
| list_for_each_entry_safe(list, next_list, &src->val, list) { |
| callchain_cursor_append(cursor, list->ip, &list->ms, |
| false, NULL, 0, 0, 0, list->srcline); |
| list_del_init(&list->list); |
| map__zput(list->ms.map); |
| free(list); |
| } |
| |
| if (src->hit) { |
| callchain_cursor_commit(cursor); |
| if (append_chain_children(dst, cursor, src->hit) < 0) |
| return -1; |
| } |
| |
| n = rb_first(&src->rb_root_in); |
| while (n) { |
| child = container_of(n, struct callchain_node, rb_node_in); |
| n = rb_next(n); |
| rb_erase(&child->rb_node_in, &src->rb_root_in); |
| |
| err = merge_chain_branch(cursor, dst, child); |
| if (err) |
| break; |
| |
| free(child); |
| } |
| |
| cursor->nr = old_pos; |
| cursor->last = old_last; |
| |
| return err; |
| } |
| |
| int callchain_merge(struct callchain_cursor *cursor, |
| struct callchain_root *dst, struct callchain_root *src) |
| { |
| return merge_chain_branch(cursor, &dst->node, &src->node); |
| } |
| |
| int callchain_cursor_append(struct callchain_cursor *cursor, |
| u64 ip, struct map_symbol *ms, |
| bool branch, struct branch_flags *flags, |
| int nr_loop_iter, u64 iter_cycles, u64 branch_from, |
| const char *srcline) |
| { |
| struct callchain_cursor_node *node = *cursor->last; |
| |
| if (!node) { |
| node = calloc(1, sizeof(*node)); |
| if (!node) |
| return -ENOMEM; |
| |
| *cursor->last = node; |
| } |
| |
| node->ip = ip; |
| map__zput(node->ms.map); |
| node->ms = *ms; |
| node->ms.map = map__get(node->ms.map); |
| node->branch = branch; |
| node->nr_loop_iter = nr_loop_iter; |
| node->iter_cycles = iter_cycles; |
| node->srcline = srcline; |
| |
| if (flags) |
| memcpy(&node->branch_flags, flags, |
| sizeof(struct branch_flags)); |
| |
| node->branch_from = branch_from; |
| cursor->nr++; |
| |
| cursor->last = &node->next; |
| |
| return 0; |
| } |
| |
| int sample__resolve_callchain(struct perf_sample *sample, |
| struct callchain_cursor *cursor, struct symbol **parent, |
| struct evsel *evsel, struct addr_location *al, |
| int max_stack) |
| { |
| if (sample->callchain == NULL && !symbol_conf.show_branchflag_count) |
| return 0; |
| |
| if (symbol_conf.use_callchain || symbol_conf.cumulate_callchain || |
| perf_hpp_list.parent || symbol_conf.show_branchflag_count) { |
| return thread__resolve_callchain(al->thread, cursor, evsel, sample, |
| parent, al, max_stack); |
| } |
| return 0; |
| } |
| |
| int hist_entry__append_callchain(struct hist_entry *he, struct perf_sample *sample) |
| { |
| if ((!symbol_conf.use_callchain || sample->callchain == NULL) && |
| !symbol_conf.show_branchflag_count) |
| return 0; |
| return callchain_append(he->callchain, &callchain_cursor, sample->period); |
| } |
| |
| int fill_callchain_info(struct addr_location *al, struct callchain_cursor_node *node, |
| bool hide_unresolved) |
| { |
| struct machine *machine = maps__machine(node->ms.maps); |
| |
| al->maps = node->ms.maps; |
| map__put(al->map); |
| al->map = map__get(node->ms.map); |
| al->sym = node->ms.sym; |
| al->srcline = node->srcline; |
| al->addr = node->ip; |
| |
| if (al->sym == NULL) { |
| if (hide_unresolved) |
| return 0; |
| if (al->map == NULL) |
| goto out; |
| } |
| if (al->maps == machine__kernel_maps(machine)) { |
| if (machine__is_host(machine)) { |
| al->cpumode = PERF_RECORD_MISC_KERNEL; |
| al->level = 'k'; |
| } else { |
| al->cpumode = PERF_RECORD_MISC_GUEST_KERNEL; |
| al->level = 'g'; |
| } |
| } else { |
| if (machine__is_host(machine)) { |
| al->cpumode = PERF_RECORD_MISC_USER; |
| al->level = '.'; |
| } else if (perf_guest) { |
| al->cpumode = PERF_RECORD_MISC_GUEST_USER; |
| al->level = 'u'; |
| } else { |
| al->cpumode = PERF_RECORD_MISC_HYPERVISOR; |
| al->level = 'H'; |
| } |
| } |
| |
| out: |
| return 1; |
| } |
| |
| char *callchain_list__sym_name(struct callchain_list *cl, |
| char *bf, size_t bfsize, bool show_dso) |
| { |
| bool show_addr = callchain_param.key == CCKEY_ADDRESS; |
| bool show_srcline = show_addr || callchain_param.key == CCKEY_SRCLINE; |
| int printed; |
| |
| if (cl->ms.sym) { |
| const char *inlined = cl->ms.sym->inlined ? " (inlined)" : ""; |
| |
| if (show_srcline && cl->srcline) |
| printed = scnprintf(bf, bfsize, "%s %s%s", |
| cl->ms.sym->name, cl->srcline, |
| inlined); |
| else |
| printed = scnprintf(bf, bfsize, "%s%s", |
| cl->ms.sym->name, inlined); |
| } else |
| printed = scnprintf(bf, bfsize, "%#" PRIx64, cl->ip); |
| |
| if (show_dso) |
| scnprintf(bf + printed, bfsize - printed, " %s", |
| cl->ms.map ? |
| map__dso(cl->ms.map)->short_name : |
| "unknown"); |
| |
| return bf; |
| } |
| |
| char *callchain_node__scnprintf_value(struct callchain_node *node, |
| char *bf, size_t bfsize, u64 total) |
| { |
| double percent = 0.0; |
| u64 period = callchain_cumul_hits(node); |
| unsigned count = callchain_cumul_counts(node); |
| |
| if (callchain_param.mode == CHAIN_FOLDED) { |
| period = node->hit; |
| count = node->count; |
| } |
| |
| switch (callchain_param.value) { |
| case CCVAL_PERIOD: |
| scnprintf(bf, bfsize, "%"PRIu64, period); |
| break; |
| case CCVAL_COUNT: |
| scnprintf(bf, bfsize, "%u", count); |
| break; |
| case CCVAL_PERCENT: |
| default: |
| if (total) |
| percent = period * 100.0 / total; |
| scnprintf(bf, bfsize, "%.2f%%", percent); |
| break; |
| } |
| return bf; |
| } |
| |
| int callchain_node__fprintf_value(struct callchain_node *node, |
| FILE *fp, u64 total) |
| { |
| double percent = 0.0; |
| u64 period = callchain_cumul_hits(node); |
| unsigned count = callchain_cumul_counts(node); |
| |
| if (callchain_param.mode == CHAIN_FOLDED) { |
| period = node->hit; |
| count = node->count; |
| } |
| |
| switch (callchain_param.value) { |
| case CCVAL_PERIOD: |
| return fprintf(fp, "%"PRIu64, period); |
| case CCVAL_COUNT: |
| return fprintf(fp, "%u", count); |
| case CCVAL_PERCENT: |
| default: |
| if (total) |
| percent = period * 100.0 / total; |
| return percent_color_fprintf(fp, "%.2f%%", percent); |
| } |
| return 0; |
| } |
| |
| static void callchain_counts_value(struct callchain_node *node, |
| u64 *branch_count, u64 *predicted_count, |
| u64 *abort_count, u64 *cycles_count) |
| { |
| struct callchain_list *clist; |
| |
| list_for_each_entry(clist, &node->val, list) { |
| if (branch_count) |
| *branch_count += clist->branch_count; |
| |
| if (predicted_count) |
| *predicted_count += clist->predicted_count; |
| |
| if (abort_count) |
| *abort_count += clist->abort_count; |
| |
| if (cycles_count) |
| *cycles_count += clist->cycles_count; |
| } |
| } |
| |
| static int callchain_node_branch_counts_cumul(struct callchain_node *node, |
| u64 *branch_count, |
| u64 *predicted_count, |
| u64 *abort_count, |
| u64 *cycles_count) |
| { |
| struct callchain_node *child; |
| struct rb_node *n; |
| |
| n = rb_first(&node->rb_root_in); |
| while (n) { |
| child = rb_entry(n, struct callchain_node, rb_node_in); |
| n = rb_next(n); |
| |
| callchain_node_branch_counts_cumul(child, branch_count, |
| predicted_count, |
| abort_count, |
| cycles_count); |
| |
| callchain_counts_value(child, branch_count, |
| predicted_count, abort_count, |
| cycles_count); |
| } |
| |
| return 0; |
| } |
| |
| int callchain_branch_counts(struct callchain_root *root, |
| u64 *branch_count, u64 *predicted_count, |
| u64 *abort_count, u64 *cycles_count) |
| { |
| if (branch_count) |
| *branch_count = 0; |
| |
| if (predicted_count) |
| *predicted_count = 0; |
| |
| if (abort_count) |
| *abort_count = 0; |
| |
| if (cycles_count) |
| *cycles_count = 0; |
| |
| return callchain_node_branch_counts_cumul(&root->node, |
| branch_count, |
| predicted_count, |
| abort_count, |
| cycles_count); |
| } |
| |
| static int count_pri64_printf(int idx, const char *str, u64 value, char *bf, int bfsize) |
| { |
| return scnprintf(bf, bfsize, "%s%s:%" PRId64 "", (idx) ? " " : " (", str, value); |
| } |
| |
| static int count_float_printf(int idx, const char *str, float value, |
| char *bf, int bfsize, float threshold) |
| { |
| if (threshold != 0.0 && value < threshold) |
| return 0; |
| |
| return scnprintf(bf, bfsize, "%s%s:%.1f%%", (idx) ? " " : " (", str, value); |
| } |
| |
| static int branch_to_str(char *bf, int bfsize, |
| u64 branch_count, u64 predicted_count, |
| u64 abort_count, |
| struct branch_type_stat *brtype_stat) |
| { |
| int printed, i = 0; |
| |
| printed = branch_type_str(brtype_stat, bf, bfsize); |
| if (printed) |
| i++; |
| |
| if (predicted_count < branch_count) { |
| printed += count_float_printf(i++, "predicted", |
| predicted_count * 100.0 / branch_count, |
| bf + printed, bfsize - printed, 0.0); |
| } |
| |
| if (abort_count) { |
| printed += count_float_printf(i++, "abort", |
| abort_count * 100.0 / branch_count, |
| bf + printed, bfsize - printed, 0.1); |
| } |
| |
| if (i) |
| printed += scnprintf(bf + printed, bfsize - printed, ")"); |
| |
| return printed; |
| } |
| |
| static int branch_from_str(char *bf, int bfsize, |
| u64 branch_count, |
| u64 cycles_count, u64 iter_count, |
| u64 iter_cycles, u64 from_count) |
| { |
| int printed = 0, i = 0; |
| u64 cycles, v = 0; |
| |
| cycles = cycles_count / branch_count; |
| if (cycles) { |
| printed += count_pri64_printf(i++, "cycles", |
| cycles, |
| bf + printed, bfsize - printed); |
| } |
| |
| if (iter_count && from_count) { |
| v = iter_count / from_count; |
| if (v) { |
| printed += count_pri64_printf(i++, "iter", |
| v, bf + printed, bfsize - printed); |
| |
| printed += count_pri64_printf(i++, "avg_cycles", |
| iter_cycles / iter_count, |
| bf + printed, bfsize - printed); |
| } |
| } |
| |
| if (i) |
| printed += scnprintf(bf + printed, bfsize - printed, ")"); |
| |
| return printed; |
| } |
| |
| static int counts_str_build(char *bf, int bfsize, |
| u64 branch_count, u64 predicted_count, |
| u64 abort_count, u64 cycles_count, |
| u64 iter_count, u64 iter_cycles, |
| u64 from_count, |
| struct branch_type_stat *brtype_stat) |
| { |
| int printed; |
| |
| if (branch_count == 0) |
| return scnprintf(bf, bfsize, " (calltrace)"); |
| |
| if (brtype_stat->branch_to) { |
| printed = branch_to_str(bf, bfsize, branch_count, |
| predicted_count, abort_count, brtype_stat); |
| } else { |
| printed = branch_from_str(bf, bfsize, branch_count, |
| cycles_count, iter_count, iter_cycles, |
| from_count); |
| } |
| |
| if (!printed) |
| bf[0] = 0; |
| |
| return printed; |
| } |
| |
| static int callchain_counts_printf(FILE *fp, char *bf, int bfsize, |
| u64 branch_count, u64 predicted_count, |
| u64 abort_count, u64 cycles_count, |
| u64 iter_count, u64 iter_cycles, |
| u64 from_count, |
| struct branch_type_stat *brtype_stat) |
| { |
| char str[256]; |
| |
| counts_str_build(str, sizeof(str), branch_count, |
| predicted_count, abort_count, cycles_count, |
| iter_count, iter_cycles, from_count, brtype_stat); |
| |
| if (fp) |
| return fprintf(fp, "%s", str); |
| |
| return scnprintf(bf, bfsize, "%s", str); |
| } |
| |
| int callchain_list_counts__printf_value(struct callchain_list *clist, |
| FILE *fp, char *bf, int bfsize) |
| { |
| u64 branch_count, predicted_count; |
| u64 abort_count, cycles_count; |
| u64 iter_count, iter_cycles; |
| u64 from_count; |
| |
| branch_count = clist->branch_count; |
| predicted_count = clist->predicted_count; |
| abort_count = clist->abort_count; |
| cycles_count = clist->cycles_count; |
| iter_count = clist->iter_count; |
| iter_cycles = clist->iter_cycles; |
| from_count = clist->from_count; |
| |
| return callchain_counts_printf(fp, bf, bfsize, branch_count, |
| predicted_count, abort_count, |
| cycles_count, iter_count, iter_cycles, |
| from_count, &clist->brtype_stat); |
| } |
| |
| static void free_callchain_node(struct callchain_node *node) |
| { |
| struct callchain_list *list, *tmp; |
| struct callchain_node *child; |
| struct rb_node *n; |
| |
| list_for_each_entry_safe(list, tmp, &node->parent_val, list) { |
| list_del_init(&list->list); |
| map__zput(list->ms.map); |
| free(list); |
| } |
| |
| list_for_each_entry_safe(list, tmp, &node->val, list) { |
| list_del_init(&list->list); |
| map__zput(list->ms.map); |
| free(list); |
| } |
| |
| n = rb_first(&node->rb_root_in); |
| while (n) { |
| child = container_of(n, struct callchain_node, rb_node_in); |
| n = rb_next(n); |
| rb_erase(&child->rb_node_in, &node->rb_root_in); |
| |
| free_callchain_node(child); |
| free(child); |
| } |
| } |
| |
| void free_callchain(struct callchain_root *root) |
| { |
| if (!symbol_conf.use_callchain) |
| return; |
| |
| free_callchain_node(&root->node); |
| } |
| |
| static u64 decay_callchain_node(struct callchain_node *node) |
| { |
| struct callchain_node *child; |
| struct rb_node *n; |
| u64 child_hits = 0; |
| |
| n = rb_first(&node->rb_root_in); |
| while (n) { |
| child = container_of(n, struct callchain_node, rb_node_in); |
| |
| child_hits += decay_callchain_node(child); |
| n = rb_next(n); |
| } |
| |
| node->hit = (node->hit * 7) / 8; |
| node->children_hit = child_hits; |
| |
| return node->hit; |
| } |
| |
| void decay_callchain(struct callchain_root *root) |
| { |
| if (!symbol_conf.use_callchain) |
| return; |
| |
| decay_callchain_node(&root->node); |
| } |
| |
| int callchain_node__make_parent_list(struct callchain_node *node) |
| { |
| struct callchain_node *parent = node->parent; |
| struct callchain_list *chain, *new; |
| LIST_HEAD(head); |
| |
| while (parent) { |
| list_for_each_entry_reverse(chain, &parent->val, list) { |
| new = malloc(sizeof(*new)); |
| if (new == NULL) |
| goto out; |
| *new = *chain; |
| new->has_children = false; |
| new->ms.map = map__get(new->ms.map); |
| list_add_tail(&new->list, &head); |
| } |
| parent = parent->parent; |
| } |
| |
| list_for_each_entry_safe_reverse(chain, new, &head, list) |
| list_move_tail(&chain->list, &node->parent_val); |
| |
| if (!list_empty(&node->parent_val)) { |
| chain = list_first_entry(&node->parent_val, struct callchain_list, list); |
| chain->has_children = rb_prev(&node->rb_node) || rb_next(&node->rb_node); |
| |
| chain = list_first_entry(&node->val, struct callchain_list, list); |
| chain->has_children = false; |
| } |
| return 0; |
| |
| out: |
| list_for_each_entry_safe(chain, new, &head, list) { |
| list_del_init(&chain->list); |
| map__zput(chain->ms.map); |
| free(chain); |
| } |
| return -ENOMEM; |
| } |
| |
| int callchain_cursor__copy(struct callchain_cursor *dst, |
| struct callchain_cursor *src) |
| { |
| int rc = 0; |
| |
| callchain_cursor_reset(dst); |
| callchain_cursor_commit(src); |
| |
| while (true) { |
| struct callchain_cursor_node *node; |
| |
| node = callchain_cursor_current(src); |
| if (node == NULL) |
| break; |
| |
| rc = callchain_cursor_append(dst, node->ip, &node->ms, |
| node->branch, &node->branch_flags, |
| node->nr_loop_iter, |
| node->iter_cycles, |
| node->branch_from, node->srcline); |
| if (rc) |
| break; |
| |
| callchain_cursor_advance(src); |
| } |
| |
| return rc; |
| } |
| |
| /* |
| * Initialize a cursor before adding entries inside, but keep |
| * the previously allocated entries as a cache. |
| */ |
| void callchain_cursor_reset(struct callchain_cursor *cursor) |
| { |
| struct callchain_cursor_node *node; |
| |
| cursor->nr = 0; |
| cursor->last = &cursor->first; |
| |
| for (node = cursor->first; node != NULL; node = node->next) |
| map__zput(node->ms.map); |
| } |
| |
| void callchain_param_setup(u64 sample_type, const char *arch) |
| { |
| if (symbol_conf.use_callchain || symbol_conf.cumulate_callchain) { |
| if ((sample_type & PERF_SAMPLE_REGS_USER) && |
| (sample_type & PERF_SAMPLE_STACK_USER)) { |
| callchain_param.record_mode = CALLCHAIN_DWARF; |
| dwarf_callchain_users = true; |
| } else if (sample_type & PERF_SAMPLE_BRANCH_STACK) |
| callchain_param.record_mode = CALLCHAIN_LBR; |
| else |
| callchain_param.record_mode = CALLCHAIN_FP; |
| } |
| |
| /* |
| * It's necessary to use libunwind to reliably determine the caller of |
| * a leaf function on aarch64, as otherwise we cannot know whether to |
| * start from the LR or FP. |
| * |
| * Always starting from the LR can result in duplicate or entirely |
| * erroneous entries. Always skipping the LR and starting from the FP |
| * can result in missing entries. |
| */ |
| if (callchain_param.record_mode == CALLCHAIN_FP && !strcmp(arch, "arm64")) |
| dwarf_callchain_users = true; |
| } |
| |
| static bool chain_match(struct callchain_list *base_chain, |
| struct callchain_list *pair_chain) |
| { |
| enum match_result match; |
| |
| match = match_chain_strings(base_chain->srcline, |
| pair_chain->srcline); |
| if (match != MATCH_ERROR) |
| return match == MATCH_EQ; |
| |
| match = match_chain_dso_addresses(base_chain->ms.map, |
| base_chain->ip, |
| pair_chain->ms.map, |
| pair_chain->ip); |
| |
| return match == MATCH_EQ; |
| } |
| |
| bool callchain_cnode_matched(struct callchain_node *base_cnode, |
| struct callchain_node *pair_cnode) |
| { |
| struct callchain_list *base_chain, *pair_chain; |
| bool match = false; |
| |
| pair_chain = list_first_entry(&pair_cnode->val, |
| struct callchain_list, |
| list); |
| |
| list_for_each_entry(base_chain, &base_cnode->val, list) { |
| if (&pair_chain->list == &pair_cnode->val) |
| return false; |
| |
| if (!base_chain->srcline || !pair_chain->srcline) { |
| pair_chain = list_next_entry(pair_chain, list); |
| continue; |
| } |
| |
| match = chain_match(base_chain, pair_chain); |
| if (!match) |
| return false; |
| |
| pair_chain = list_next_entry(pair_chain, list); |
| } |
| |
| /* |
| * Say chain1 is ABC, chain2 is ABCD, we consider they are |
| * not fully matched. |
| */ |
| if (pair_chain && (&pair_chain->list != &pair_cnode->val)) |
| return false; |
| |
| return match; |
| } |
| |
| static u64 count_callchain_hits(struct hist_entry *he) |
| { |
| struct rb_root *root = &he->sorted_chain; |
| struct rb_node *rb_node = rb_first(root); |
| struct callchain_node *node; |
| u64 chain_hits = 0; |
| |
| while (rb_node) { |
| node = rb_entry(rb_node, struct callchain_node, rb_node); |
| chain_hits += node->hit; |
| rb_node = rb_next(rb_node); |
| } |
| |
| return chain_hits; |
| } |
| |
| u64 callchain_total_hits(struct hists *hists) |
| { |
| struct rb_node *next = rb_first_cached(&hists->entries); |
| u64 chain_hits = 0; |
| |
| while (next) { |
| struct hist_entry *he = rb_entry(next, struct hist_entry, |
| rb_node); |
| |
| chain_hits += count_callchain_hits(he); |
| next = rb_next(&he->rb_node); |
| } |
| |
| return chain_hits; |
| } |
| |
| s64 callchain_avg_cycles(struct callchain_node *cnode) |
| { |
| struct callchain_list *chain; |
| s64 cycles = 0; |
| |
| list_for_each_entry(chain, &cnode->val, list) { |
| if (chain->srcline && chain->branch_count) |
| cycles += chain->cycles_count / chain->branch_count; |
| } |
| |
| return cycles; |
| } |