| // SPDX-License-Identifier: GPL-2.0 |
| #include "util/debug.h" |
| #include "util/dso.h" |
| #include "util/event.h" |
| #include "util/map.h" |
| #include "util/symbol.h" |
| #include "util/sort.h" |
| #include "util/evsel.h" |
| #include "util/evlist.h" |
| #include "util/machine.h" |
| #include "util/parse-events.h" |
| #include "util/thread.h" |
| #include "tests/tests.h" |
| #include "tests/hists_common.h" |
| #include <linux/kernel.h> |
| |
| struct sample { |
| u32 pid; |
| u64 ip; |
| struct thread *thread; |
| struct map *map; |
| struct symbol *sym; |
| }; |
| |
| /* For the numbers, see hists_common.c */ |
| static struct sample fake_samples[] = { |
| /* perf [kernel] schedule() */ |
| { .pid = FAKE_PID_PERF1, .ip = FAKE_IP_KERNEL_SCHEDULE, }, |
| /* perf [perf] main() */ |
| { .pid = FAKE_PID_PERF1, .ip = FAKE_IP_PERF_MAIN, }, |
| /* perf [perf] cmd_record() */ |
| { .pid = FAKE_PID_PERF1, .ip = FAKE_IP_PERF_CMD_RECORD, }, |
| /* perf [libc] malloc() */ |
| { .pid = FAKE_PID_PERF1, .ip = FAKE_IP_LIBC_MALLOC, }, |
| /* perf [libc] free() */ |
| { .pid = FAKE_PID_PERF1, .ip = FAKE_IP_LIBC_FREE, }, |
| /* perf [perf] main() */ |
| { .pid = FAKE_PID_PERF2, .ip = FAKE_IP_PERF_MAIN, }, |
| /* perf [kernel] page_fault() */ |
| { .pid = FAKE_PID_PERF2, .ip = FAKE_IP_KERNEL_PAGE_FAULT, }, |
| /* bash [bash] main() */ |
| { .pid = FAKE_PID_BASH, .ip = FAKE_IP_BASH_MAIN, }, |
| /* bash [bash] xmalloc() */ |
| { .pid = FAKE_PID_BASH, .ip = FAKE_IP_BASH_XMALLOC, }, |
| /* bash [kernel] page_fault() */ |
| { .pid = FAKE_PID_BASH, .ip = FAKE_IP_KERNEL_PAGE_FAULT, }, |
| }; |
| |
| /* |
| * Will be cast to struct ip_callchain which has all 64 bit entries |
| * of nr and ips[]. |
| */ |
| static u64 fake_callchains[][10] = { |
| /* schedule => run_command => main */ |
| { 3, FAKE_IP_KERNEL_SCHEDULE, FAKE_IP_PERF_RUN_COMMAND, FAKE_IP_PERF_MAIN, }, |
| /* main */ |
| { 1, FAKE_IP_PERF_MAIN, }, |
| /* cmd_record => run_command => main */ |
| { 3, FAKE_IP_PERF_CMD_RECORD, FAKE_IP_PERF_RUN_COMMAND, FAKE_IP_PERF_MAIN, }, |
| /* malloc => cmd_record => run_command => main */ |
| { 4, FAKE_IP_LIBC_MALLOC, FAKE_IP_PERF_CMD_RECORD, FAKE_IP_PERF_RUN_COMMAND, |
| FAKE_IP_PERF_MAIN, }, |
| /* free => cmd_record => run_command => main */ |
| { 4, FAKE_IP_LIBC_FREE, FAKE_IP_PERF_CMD_RECORD, FAKE_IP_PERF_RUN_COMMAND, |
| FAKE_IP_PERF_MAIN, }, |
| /* main */ |
| { 1, FAKE_IP_PERF_MAIN, }, |
| /* page_fault => sys_perf_event_open => run_command => main */ |
| { 4, FAKE_IP_KERNEL_PAGE_FAULT, FAKE_IP_KERNEL_SYS_PERF_EVENT_OPEN, |
| FAKE_IP_PERF_RUN_COMMAND, FAKE_IP_PERF_MAIN, }, |
| /* main */ |
| { 1, FAKE_IP_BASH_MAIN, }, |
| /* xmalloc => malloc => xmalloc => malloc => xmalloc => main */ |
| { 6, FAKE_IP_BASH_XMALLOC, FAKE_IP_LIBC_MALLOC, FAKE_IP_BASH_XMALLOC, |
| FAKE_IP_LIBC_MALLOC, FAKE_IP_BASH_XMALLOC, FAKE_IP_BASH_MAIN, }, |
| /* page_fault => malloc => main */ |
| { 3, FAKE_IP_KERNEL_PAGE_FAULT, FAKE_IP_LIBC_MALLOC, FAKE_IP_BASH_MAIN, }, |
| }; |
| |
| static int add_hist_entries(struct hists *hists, struct machine *machine) |
| { |
| struct addr_location al; |
| struct evsel *evsel = hists_to_evsel(hists); |
| struct perf_sample sample = { .period = 1000, }; |
| size_t i; |
| |
| addr_location__init(&al); |
| for (i = 0; i < ARRAY_SIZE(fake_samples); i++) { |
| struct hist_entry_iter iter = { |
| .evsel = evsel, |
| .sample = &sample, |
| .hide_unresolved = false, |
| }; |
| |
| if (symbol_conf.cumulate_callchain) |
| iter.ops = &hist_iter_cumulative; |
| else |
| iter.ops = &hist_iter_normal; |
| |
| sample.cpumode = PERF_RECORD_MISC_USER; |
| sample.pid = fake_samples[i].pid; |
| sample.tid = fake_samples[i].pid; |
| sample.ip = fake_samples[i].ip; |
| sample.callchain = (struct ip_callchain *)fake_callchains[i]; |
| |
| if (machine__resolve(machine, &al, &sample) < 0) |
| goto out; |
| |
| if (hist_entry_iter__add(&iter, &al, sysctl_perf_event_max_stack, |
| NULL) < 0) { |
| goto out; |
| } |
| |
| thread__put(fake_samples[i].thread); |
| fake_samples[i].thread = thread__get(al.thread); |
| map__put(fake_samples[i].map); |
| fake_samples[i].map = map__get(al.map); |
| fake_samples[i].sym = al.sym; |
| } |
| |
| addr_location__exit(&al); |
| return TEST_OK; |
| |
| out: |
| pr_debug("Not enough memory for adding a hist entry\n"); |
| addr_location__exit(&al); |
| return TEST_FAIL; |
| } |
| |
| static void del_hist_entries(struct hists *hists) |
| { |
| struct hist_entry *he; |
| struct rb_root_cached *root_in; |
| struct rb_root_cached *root_out; |
| struct rb_node *node; |
| |
| if (hists__has(hists, need_collapse)) |
| root_in = &hists->entries_collapsed; |
| else |
| root_in = hists->entries_in; |
| |
| root_out = &hists->entries; |
| |
| while (!RB_EMPTY_ROOT(&root_out->rb_root)) { |
| node = rb_first_cached(root_out); |
| |
| he = rb_entry(node, struct hist_entry, rb_node); |
| rb_erase_cached(node, root_out); |
| rb_erase_cached(&he->rb_node_in, root_in); |
| hist_entry__delete(he); |
| } |
| } |
| |
| static void put_fake_samples(void) |
| { |
| size_t i; |
| |
| for (i = 0; i < ARRAY_SIZE(fake_samples); i++) { |
| map__zput(fake_samples[i].map); |
| thread__zput(fake_samples[i].thread); |
| } |
| } |
| |
| typedef int (*test_fn_t)(struct evsel *, struct machine *); |
| |
| #define COMM(he) (thread__comm_str(he->thread)) |
| #define DSO(he) (map__dso(he->ms.map)->short_name) |
| #define SYM(he) (he->ms.sym->name) |
| #define CPU(he) (he->cpu) |
| #define DEPTH(he) (he->callchain->max_depth) |
| #define CDSO(cl) (map__dso(cl->ms.map)->short_name) |
| #define CSYM(cl) (cl->ms.sym->name) |
| |
| struct result { |
| u64 children; |
| u64 self; |
| const char *comm; |
| const char *dso; |
| const char *sym; |
| }; |
| |
| struct callchain_result { |
| u64 nr; |
| struct { |
| const char *dso; |
| const char *sym; |
| } node[10]; |
| }; |
| |
| static int do_test(struct hists *hists, struct result *expected, size_t nr_expected, |
| struct callchain_result *expected_callchain, size_t nr_callchain) |
| { |
| char buf[32]; |
| size_t i, c; |
| struct hist_entry *he; |
| struct rb_root *root; |
| struct rb_node *node; |
| struct callchain_node *cnode; |
| struct callchain_list *clist; |
| |
| /* |
| * adding and deleting hist entries must be done outside of this |
| * function since TEST_ASSERT_VAL() returns in case of failure. |
| */ |
| hists__collapse_resort(hists, NULL); |
| evsel__output_resort(hists_to_evsel(hists), NULL); |
| |
| if (verbose > 2) { |
| pr_info("use callchain: %d, cumulate callchain: %d\n", |
| symbol_conf.use_callchain, |
| symbol_conf.cumulate_callchain); |
| print_hists_out(hists); |
| } |
| |
| root = &hists->entries.rb_root; |
| for (node = rb_first(root), i = 0; |
| node && (he = rb_entry(node, struct hist_entry, rb_node)); |
| node = rb_next(node), i++) { |
| scnprintf(buf, sizeof(buf), "Invalid hist entry #%zd", i); |
| |
| TEST_ASSERT_VAL("Incorrect number of hist entry", |
| i < nr_expected); |
| TEST_ASSERT_VAL(buf, he->stat.period == expected[i].self && |
| !strcmp(COMM(he), expected[i].comm) && |
| !strcmp(DSO(he), expected[i].dso) && |
| !strcmp(SYM(he), expected[i].sym)); |
| |
| if (symbol_conf.cumulate_callchain) |
| TEST_ASSERT_VAL(buf, he->stat_acc->period == expected[i].children); |
| |
| if (!symbol_conf.use_callchain) |
| continue; |
| |
| /* check callchain entries */ |
| root = &he->callchain->node.rb_root; |
| |
| TEST_ASSERT_VAL("callchains expected", !RB_EMPTY_ROOT(root)); |
| cnode = rb_entry(rb_first(root), struct callchain_node, rb_node); |
| |
| c = 0; |
| list_for_each_entry(clist, &cnode->val, list) { |
| scnprintf(buf, sizeof(buf), "Invalid callchain entry #%zd/%zd", i, c); |
| |
| TEST_ASSERT_VAL("Incorrect number of callchain entry", |
| c < expected_callchain[i].nr); |
| TEST_ASSERT_VAL(buf, |
| !strcmp(CDSO(clist), expected_callchain[i].node[c].dso) && |
| !strcmp(CSYM(clist), expected_callchain[i].node[c].sym)); |
| c++; |
| } |
| /* TODO: handle multiple child nodes properly */ |
| TEST_ASSERT_VAL("Incorrect number of callchain entry", |
| c <= expected_callchain[i].nr); |
| } |
| TEST_ASSERT_VAL("Incorrect number of hist entry", |
| i == nr_expected); |
| TEST_ASSERT_VAL("Incorrect number of callchain entry", |
| !symbol_conf.use_callchain || nr_expected == nr_callchain); |
| return 0; |
| } |
| |
| /* NO callchain + NO children */ |
| static int test1(struct evsel *evsel, struct machine *machine) |
| { |
| int err; |
| struct hists *hists = evsel__hists(evsel); |
| /* |
| * expected output: |
| * |
| * Overhead Command Shared Object Symbol |
| * ======== ======= ============= ============== |
| * 20.00% perf perf [.] main |
| * 10.00% bash [kernel] [k] page_fault |
| * 10.00% bash bash [.] main |
| * 10.00% bash bash [.] xmalloc |
| * 10.00% perf [kernel] [k] page_fault |
| * 10.00% perf [kernel] [k] schedule |
| * 10.00% perf libc [.] free |
| * 10.00% perf libc [.] malloc |
| * 10.00% perf perf [.] cmd_record |
| */ |
| struct result expected[] = { |
| { 0, 2000, "perf", "perf", "main" }, |
| { 0, 1000, "bash", "[kernel]", "page_fault" }, |
| { 0, 1000, "bash", "bash", "main" }, |
| { 0, 1000, "bash", "bash", "xmalloc" }, |
| { 0, 1000, "perf", "[kernel]", "page_fault" }, |
| { 0, 1000, "perf", "[kernel]", "schedule" }, |
| { 0, 1000, "perf", "libc", "free" }, |
| { 0, 1000, "perf", "libc", "malloc" }, |
| { 0, 1000, "perf", "perf", "cmd_record" }, |
| }; |
| |
| symbol_conf.use_callchain = false; |
| symbol_conf.cumulate_callchain = false; |
| evsel__reset_sample_bit(evsel, CALLCHAIN); |
| |
| setup_sorting(NULL); |
| callchain_register_param(&callchain_param); |
| |
| err = add_hist_entries(hists, machine); |
| if (err < 0) |
| goto out; |
| |
| err = do_test(hists, expected, ARRAY_SIZE(expected), NULL, 0); |
| |
| out: |
| del_hist_entries(hists); |
| reset_output_field(); |
| return err; |
| } |
| |
| /* callchain + NO children */ |
| static int test2(struct evsel *evsel, struct machine *machine) |
| { |
| int err; |
| struct hists *hists = evsel__hists(evsel); |
| /* |
| * expected output: |
| * |
| * Overhead Command Shared Object Symbol |
| * ======== ======= ============= ============== |
| * 20.00% perf perf [.] main |
| * | |
| * --- main |
| * |
| * 10.00% bash [kernel] [k] page_fault |
| * | |
| * --- page_fault |
| * malloc |
| * main |
| * |
| * 10.00% bash bash [.] main |
| * | |
| * --- main |
| * |
| * 10.00% bash bash [.] xmalloc |
| * | |
| * --- xmalloc |
| * malloc |
| * xmalloc <--- NOTE: there's a cycle |
| * malloc |
| * xmalloc |
| * main |
| * |
| * 10.00% perf [kernel] [k] page_fault |
| * | |
| * --- page_fault |
| * sys_perf_event_open |
| * run_command |
| * main |
| * |
| * 10.00% perf [kernel] [k] schedule |
| * | |
| * --- schedule |
| * run_command |
| * main |
| * |
| * 10.00% perf libc [.] free |
| * | |
| * --- free |
| * cmd_record |
| * run_command |
| * main |
| * |
| * 10.00% perf libc [.] malloc |
| * | |
| * --- malloc |
| * cmd_record |
| * run_command |
| * main |
| * |
| * 10.00% perf perf [.] cmd_record |
| * | |
| * --- cmd_record |
| * run_command |
| * main |
| * |
| */ |
| struct result expected[] = { |
| { 0, 2000, "perf", "perf", "main" }, |
| { 0, 1000, "bash", "[kernel]", "page_fault" }, |
| { 0, 1000, "bash", "bash", "main" }, |
| { 0, 1000, "bash", "bash", "xmalloc" }, |
| { 0, 1000, "perf", "[kernel]", "page_fault" }, |
| { 0, 1000, "perf", "[kernel]", "schedule" }, |
| { 0, 1000, "perf", "libc", "free" }, |
| { 0, 1000, "perf", "libc", "malloc" }, |
| { 0, 1000, "perf", "perf", "cmd_record" }, |
| }; |
| struct callchain_result expected_callchain[] = { |
| { |
| 1, { { "perf", "main" }, }, |
| }, |
| { |
| 3, { { "[kernel]", "page_fault" }, |
| { "libc", "malloc" }, |
| { "bash", "main" }, }, |
| }, |
| { |
| 1, { { "bash", "main" }, }, |
| }, |
| { |
| 6, { { "bash", "xmalloc" }, |
| { "libc", "malloc" }, |
| { "bash", "xmalloc" }, |
| { "libc", "malloc" }, |
| { "bash", "xmalloc" }, |
| { "bash", "main" }, }, |
| }, |
| { |
| 4, { { "[kernel]", "page_fault" }, |
| { "[kernel]", "sys_perf_event_open" }, |
| { "perf", "run_command" }, |
| { "perf", "main" }, }, |
| }, |
| { |
| 3, { { "[kernel]", "schedule" }, |
| { "perf", "run_command" }, |
| { "perf", "main" }, }, |
| }, |
| { |
| 4, { { "libc", "free" }, |
| { "perf", "cmd_record" }, |
| { "perf", "run_command" }, |
| { "perf", "main" }, }, |
| }, |
| { |
| 4, { { "libc", "malloc" }, |
| { "perf", "cmd_record" }, |
| { "perf", "run_command" }, |
| { "perf", "main" }, }, |
| }, |
| { |
| 3, { { "perf", "cmd_record" }, |
| { "perf", "run_command" }, |
| { "perf", "main" }, }, |
| }, |
| }; |
| |
| symbol_conf.use_callchain = true; |
| symbol_conf.cumulate_callchain = false; |
| evsel__set_sample_bit(evsel, CALLCHAIN); |
| |
| setup_sorting(NULL); |
| callchain_register_param(&callchain_param); |
| |
| err = add_hist_entries(hists, machine); |
| if (err < 0) |
| goto out; |
| |
| err = do_test(hists, expected, ARRAY_SIZE(expected), |
| expected_callchain, ARRAY_SIZE(expected_callchain)); |
| |
| out: |
| del_hist_entries(hists); |
| reset_output_field(); |
| return err; |
| } |
| |
| /* NO callchain + children */ |
| static int test3(struct evsel *evsel, struct machine *machine) |
| { |
| int err; |
| struct hists *hists = evsel__hists(evsel); |
| /* |
| * expected output: |
| * |
| * Children Self Command Shared Object Symbol |
| * ======== ======== ======= ============= ======================= |
| * 70.00% 20.00% perf perf [.] main |
| * 50.00% 0.00% perf perf [.] run_command |
| * 30.00% 10.00% bash bash [.] main |
| * 30.00% 10.00% perf perf [.] cmd_record |
| * 20.00% 0.00% bash libc [.] malloc |
| * 10.00% 10.00% bash [kernel] [k] page_fault |
| * 10.00% 10.00% bash bash [.] xmalloc |
| * 10.00% 10.00% perf [kernel] [k] page_fault |
| * 10.00% 10.00% perf libc [.] malloc |
| * 10.00% 10.00% perf [kernel] [k] schedule |
| * 10.00% 10.00% perf libc [.] free |
| * 10.00% 0.00% perf [kernel] [k] sys_perf_event_open |
| */ |
| struct result expected[] = { |
| { 7000, 2000, "perf", "perf", "main" }, |
| { 5000, 0, "perf", "perf", "run_command" }, |
| { 3000, 1000, "bash", "bash", "main" }, |
| { 3000, 1000, "perf", "perf", "cmd_record" }, |
| { 2000, 0, "bash", "libc", "malloc" }, |
| { 1000, 1000, "bash", "[kernel]", "page_fault" }, |
| { 1000, 1000, "bash", "bash", "xmalloc" }, |
| { 1000, 1000, "perf", "[kernel]", "page_fault" }, |
| { 1000, 1000, "perf", "[kernel]", "schedule" }, |
| { 1000, 1000, "perf", "libc", "free" }, |
| { 1000, 1000, "perf", "libc", "malloc" }, |
| { 1000, 0, "perf", "[kernel]", "sys_perf_event_open" }, |
| }; |
| |
| symbol_conf.use_callchain = false; |
| symbol_conf.cumulate_callchain = true; |
| evsel__reset_sample_bit(evsel, CALLCHAIN); |
| |
| setup_sorting(NULL); |
| callchain_register_param(&callchain_param); |
| |
| err = add_hist_entries(hists, machine); |
| if (err < 0) |
| goto out; |
| |
| err = do_test(hists, expected, ARRAY_SIZE(expected), NULL, 0); |
| |
| out: |
| del_hist_entries(hists); |
| reset_output_field(); |
| return err; |
| } |
| |
| /* callchain + children */ |
| static int test4(struct evsel *evsel, struct machine *machine) |
| { |
| int err; |
| struct hists *hists = evsel__hists(evsel); |
| /* |
| * expected output: |
| * |
| * Children Self Command Shared Object Symbol |
| * ======== ======== ======= ============= ======================= |
| * 70.00% 20.00% perf perf [.] main |
| * | |
| * --- main |
| * |
| * 50.00% 0.00% perf perf [.] run_command |
| * | |
| * --- run_command |
| * main |
| * |
| * 30.00% 10.00% bash bash [.] main |
| * | |
| * --- main |
| * |
| * 30.00% 10.00% perf perf [.] cmd_record |
| * | |
| * --- cmd_record |
| * run_command |
| * main |
| * |
| * 20.00% 0.00% bash libc [.] malloc |
| * | |
| * --- malloc |
| * | |
| * |--50.00%-- xmalloc |
| * | main |
| * --50.00%-- main |
| * |
| * 10.00% 10.00% bash [kernel] [k] page_fault |
| * | |
| * --- page_fault |
| * malloc |
| * main |
| * |
| * 10.00% 10.00% bash bash [.] xmalloc |
| * | |
| * --- xmalloc |
| * malloc |
| * xmalloc <--- NOTE: there's a cycle |
| * malloc |
| * xmalloc |
| * main |
| * |
| * 10.00% 0.00% perf [kernel] [k] sys_perf_event_open |
| * | |
| * --- sys_perf_event_open |
| * run_command |
| * main |
| * |
| * 10.00% 10.00% perf [kernel] [k] page_fault |
| * | |
| * --- page_fault |
| * sys_perf_event_open |
| * run_command |
| * main |
| * |
| * 10.00% 10.00% perf [kernel] [k] schedule |
| * | |
| * --- schedule |
| * run_command |
| * main |
| * |
| * 10.00% 10.00% perf libc [.] free |
| * | |
| * --- free |
| * cmd_record |
| * run_command |
| * main |
| * |
| * 10.00% 10.00% perf libc [.] malloc |
| * | |
| * --- malloc |
| * cmd_record |
| * run_command |
| * main |
| * |
| */ |
| struct result expected[] = { |
| { 7000, 2000, "perf", "perf", "main" }, |
| { 5000, 0, "perf", "perf", "run_command" }, |
| { 3000, 1000, "bash", "bash", "main" }, |
| { 3000, 1000, "perf", "perf", "cmd_record" }, |
| { 2000, 0, "bash", "libc", "malloc" }, |
| { 1000, 1000, "bash", "[kernel]", "page_fault" }, |
| { 1000, 1000, "bash", "bash", "xmalloc" }, |
| { 1000, 0, "perf", "[kernel]", "sys_perf_event_open" }, |
| { 1000, 1000, "perf", "[kernel]", "page_fault" }, |
| { 1000, 1000, "perf", "[kernel]", "schedule" }, |
| { 1000, 1000, "perf", "libc", "free" }, |
| { 1000, 1000, "perf", "libc", "malloc" }, |
| }; |
| struct callchain_result expected_callchain[] = { |
| { |
| 1, { { "perf", "main" }, }, |
| }, |
| { |
| 2, { { "perf", "run_command" }, |
| { "perf", "main" }, }, |
| }, |
| { |
| 1, { { "bash", "main" }, }, |
| }, |
| { |
| 3, { { "perf", "cmd_record" }, |
| { "perf", "run_command" }, |
| { "perf", "main" }, }, |
| }, |
| { |
| 4, { { "libc", "malloc" }, |
| { "bash", "xmalloc" }, |
| { "bash", "main" }, |
| { "bash", "main" }, }, |
| }, |
| { |
| 3, { { "[kernel]", "page_fault" }, |
| { "libc", "malloc" }, |
| { "bash", "main" }, }, |
| }, |
| { |
| 6, { { "bash", "xmalloc" }, |
| { "libc", "malloc" }, |
| { "bash", "xmalloc" }, |
| { "libc", "malloc" }, |
| { "bash", "xmalloc" }, |
| { "bash", "main" }, }, |
| }, |
| { |
| 3, { { "[kernel]", "sys_perf_event_open" }, |
| { "perf", "run_command" }, |
| { "perf", "main" }, }, |
| }, |
| { |
| 4, { { "[kernel]", "page_fault" }, |
| { "[kernel]", "sys_perf_event_open" }, |
| { "perf", "run_command" }, |
| { "perf", "main" }, }, |
| }, |
| { |
| 3, { { "[kernel]", "schedule" }, |
| { "perf", "run_command" }, |
| { "perf", "main" }, }, |
| }, |
| { |
| 4, { { "libc", "free" }, |
| { "perf", "cmd_record" }, |
| { "perf", "run_command" }, |
| { "perf", "main" }, }, |
| }, |
| { |
| 4, { { "libc", "malloc" }, |
| { "perf", "cmd_record" }, |
| { "perf", "run_command" }, |
| { "perf", "main" }, }, |
| }, |
| }; |
| |
| symbol_conf.use_callchain = true; |
| symbol_conf.cumulate_callchain = true; |
| evsel__set_sample_bit(evsel, CALLCHAIN); |
| |
| setup_sorting(NULL); |
| |
| callchain_param = callchain_param_default; |
| callchain_register_param(&callchain_param); |
| |
| err = add_hist_entries(hists, machine); |
| if (err < 0) |
| goto out; |
| |
| err = do_test(hists, expected, ARRAY_SIZE(expected), |
| expected_callchain, ARRAY_SIZE(expected_callchain)); |
| |
| out: |
| del_hist_entries(hists); |
| reset_output_field(); |
| return err; |
| } |
| |
| static int test__hists_cumulate(struct test_suite *test __maybe_unused, int subtest __maybe_unused) |
| { |
| int err = TEST_FAIL; |
| struct machines machines; |
| struct machine *machine; |
| struct evsel *evsel; |
| struct evlist *evlist = evlist__new(); |
| size_t i; |
| test_fn_t testcases[] = { |
| test1, |
| test2, |
| test3, |
| test4, |
| }; |
| |
| TEST_ASSERT_VAL("No memory", evlist); |
| |
| err = parse_event(evlist, "cpu-clock"); |
| if (err) |
| goto out; |
| err = TEST_FAIL; |
| |
| machines__init(&machines); |
| |
| /* setup threads/dso/map/symbols also */ |
| machine = setup_fake_machine(&machines); |
| if (!machine) |
| goto out; |
| |
| if (verbose > 1) |
| machine__fprintf(machine, stderr); |
| |
| evsel = evlist__first(evlist); |
| |
| for (i = 0; i < ARRAY_SIZE(testcases); i++) { |
| err = testcases[i](evsel, machine); |
| if (err < 0) |
| break; |
| } |
| |
| out: |
| /* tear down everything */ |
| evlist__delete(evlist); |
| machines__exit(&machines); |
| put_fake_samples(); |
| |
| return err; |
| } |
| |
| DEFINE_SUITE("Cumulate child hist entries", hists_cumulate); |