| // SPDX-License-Identifier: GPL-2.0 |
| /* |
| * builtin-test.c |
| * |
| * Builtin regression testing command: ever growing number of sanity tests |
| */ |
| #include <fcntl.h> |
| #include <errno.h> |
| #include <unistd.h> |
| #include <string.h> |
| #include <stdlib.h> |
| #include <sys/types.h> |
| #include <dirent.h> |
| #include <sys/wait.h> |
| #include <sys/stat.h> |
| #include "builtin.h" |
| #include "hist.h" |
| #include "intlist.h" |
| #include "tests.h" |
| #include "debug.h" |
| #include "color.h" |
| #include <subcmd/parse-options.h> |
| #include "string2.h" |
| #include "symbol.h" |
| #include "util/rlimit.h" |
| #include <linux/kernel.h> |
| #include <linux/string.h> |
| #include <subcmd/exec-cmd.h> |
| #include <linux/zalloc.h> |
| |
| static bool dont_fork; |
| |
| struct test __weak arch_tests[] = { |
| { |
| .func = NULL, |
| }, |
| }; |
| |
| static struct test generic_tests[] = { |
| { |
| .desc = "vmlinux symtab matches kallsyms", |
| .func = test__vmlinux_matches_kallsyms, |
| }, |
| { |
| .desc = "Detect openat syscall event", |
| .func = test__openat_syscall_event, |
| }, |
| { |
| .desc = "Detect openat syscall event on all cpus", |
| .func = test__openat_syscall_event_on_all_cpus, |
| }, |
| { |
| .desc = "Read samples using the mmap interface", |
| .func = test__basic_mmap, |
| }, |
| { |
| .desc = "Test data source output", |
| .func = test__mem, |
| }, |
| { |
| .desc = "Parse event definition strings", |
| .func = test__parse_events, |
| }, |
| { |
| .desc = "Simple expression parser", |
| .func = test__expr, |
| }, |
| { |
| .desc = "PERF_RECORD_* events & perf_sample fields", |
| .func = test__PERF_RECORD, |
| }, |
| { |
| .desc = "Parse perf pmu format", |
| .func = test__pmu, |
| }, |
| { |
| .desc = "PMU events", |
| .func = test__pmu_events, |
| .subtest = { |
| .skip_if_fail = false, |
| .get_nr = test__pmu_events_subtest_get_nr, |
| .get_desc = test__pmu_events_subtest_get_desc, |
| .skip_reason = test__pmu_events_subtest_skip_reason, |
| }, |
| |
| }, |
| { |
| .desc = "DSO data read", |
| .func = test__dso_data, |
| }, |
| { |
| .desc = "DSO data cache", |
| .func = test__dso_data_cache, |
| }, |
| { |
| .desc = "DSO data reopen", |
| .func = test__dso_data_reopen, |
| }, |
| { |
| .desc = "Roundtrip evsel->name", |
| .func = test__perf_evsel__roundtrip_name_test, |
| }, |
| { |
| .desc = "Parse sched tracepoints fields", |
| .func = test__perf_evsel__tp_sched_test, |
| }, |
| { |
| .desc = "syscalls:sys_enter_openat event fields", |
| .func = test__syscall_openat_tp_fields, |
| }, |
| { |
| .desc = "Setup struct perf_event_attr", |
| .func = test__attr, |
| }, |
| { |
| .desc = "Match and link multiple hists", |
| .func = test__hists_link, |
| }, |
| { |
| .desc = "'import perf' in python", |
| .func = test__python_use, |
| }, |
| { |
| .desc = "Breakpoint overflow signal handler", |
| .func = test__bp_signal, |
| .is_supported = test__bp_signal_is_supported, |
| }, |
| { |
| .desc = "Breakpoint overflow sampling", |
| .func = test__bp_signal_overflow, |
| .is_supported = test__bp_signal_is_supported, |
| }, |
| { |
| .desc = "Breakpoint accounting", |
| .func = test__bp_accounting, |
| .is_supported = test__bp_account_is_supported, |
| }, |
| { |
| .desc = "Watchpoint", |
| .func = test__wp, |
| .is_supported = test__wp_is_supported, |
| .subtest = { |
| .skip_if_fail = false, |
| .get_nr = test__wp_subtest_get_nr, |
| .get_desc = test__wp_subtest_get_desc, |
| .skip_reason = test__wp_subtest_skip_reason, |
| }, |
| }, |
| { |
| .desc = "Number of exit events of a simple workload", |
| .func = test__task_exit, |
| }, |
| { |
| .desc = "Software clock events period values", |
| .func = test__sw_clock_freq, |
| }, |
| { |
| .desc = "Object code reading", |
| .func = test__code_reading, |
| }, |
| { |
| .desc = "Sample parsing", |
| .func = test__sample_parsing, |
| }, |
| { |
| .desc = "Use a dummy software event to keep tracking", |
| .func = test__keep_tracking, |
| }, |
| { |
| .desc = "Parse with no sample_id_all bit set", |
| .func = test__parse_no_sample_id_all, |
| }, |
| { |
| .desc = "Filter hist entries", |
| .func = test__hists_filter, |
| }, |
| { |
| .desc = "Lookup mmap thread", |
| .func = test__mmap_thread_lookup, |
| }, |
| { |
| .desc = "Share thread maps", |
| .func = test__thread_maps_share, |
| }, |
| { |
| .desc = "Sort output of hist entries", |
| .func = test__hists_output, |
| }, |
| { |
| .desc = "Cumulate child hist entries", |
| .func = test__hists_cumulate, |
| }, |
| { |
| .desc = "Track with sched_switch", |
| .func = test__switch_tracking, |
| }, |
| { |
| .desc = "Filter fds with revents mask in a fdarray", |
| .func = test__fdarray__filter, |
| }, |
| { |
| .desc = "Add fd to a fdarray, making it autogrow", |
| .func = test__fdarray__add, |
| }, |
| { |
| .desc = "kmod_path__parse", |
| .func = test__kmod_path__parse, |
| }, |
| { |
| .desc = "Thread map", |
| .func = test__thread_map, |
| }, |
| { |
| .desc = "LLVM search and compile", |
| .func = test__llvm, |
| .subtest = { |
| .skip_if_fail = true, |
| .get_nr = test__llvm_subtest_get_nr, |
| .get_desc = test__llvm_subtest_get_desc, |
| }, |
| }, |
| { |
| .desc = "Session topology", |
| .func = test__session_topology, |
| }, |
| { |
| .desc = "BPF filter", |
| .func = test__bpf, |
| .subtest = { |
| .skip_if_fail = true, |
| .get_nr = test__bpf_subtest_get_nr, |
| .get_desc = test__bpf_subtest_get_desc, |
| }, |
| }, |
| { |
| .desc = "Synthesize thread map", |
| .func = test__thread_map_synthesize, |
| }, |
| { |
| .desc = "Remove thread map", |
| .func = test__thread_map_remove, |
| }, |
| { |
| .desc = "Synthesize cpu map", |
| .func = test__cpu_map_synthesize, |
| }, |
| { |
| .desc = "Synthesize stat config", |
| .func = test__synthesize_stat_config, |
| }, |
| { |
| .desc = "Synthesize stat", |
| .func = test__synthesize_stat, |
| }, |
| { |
| .desc = "Synthesize stat round", |
| .func = test__synthesize_stat_round, |
| }, |
| { |
| .desc = "Synthesize attr update", |
| .func = test__event_update, |
| }, |
| { |
| .desc = "Event times", |
| .func = test__event_times, |
| }, |
| { |
| .desc = "Read backward ring buffer", |
| .func = test__backward_ring_buffer, |
| }, |
| { |
| .desc = "Print cpu map", |
| .func = test__cpu_map_print, |
| }, |
| { |
| .desc = "Merge cpu map", |
| .func = test__cpu_map_merge, |
| }, |
| |
| { |
| .desc = "Probe SDT events", |
| .func = test__sdt_event, |
| }, |
| { |
| .desc = "is_printable_array", |
| .func = test__is_printable_array, |
| }, |
| { |
| .desc = "Print bitmap", |
| .func = test__bitmap_print, |
| }, |
| { |
| .desc = "perf hooks", |
| .func = test__perf_hooks, |
| }, |
| { |
| .desc = "builtin clang support", |
| .func = test__clang, |
| .subtest = { |
| .skip_if_fail = true, |
| .get_nr = test__clang_subtest_get_nr, |
| .get_desc = test__clang_subtest_get_desc, |
| } |
| }, |
| { |
| .desc = "unit_number__scnprintf", |
| .func = test__unit_number__scnprint, |
| }, |
| { |
| .desc = "mem2node", |
| .func = test__mem2node, |
| }, |
| { |
| .desc = "time utils", |
| .func = test__time_utils, |
| }, |
| { |
| .desc = "Test jit_write_elf", |
| .func = test__jit_write_elf, |
| }, |
| { |
| .desc = "Test libpfm4 support", |
| .func = test__pfm, |
| .subtest = { |
| .skip_if_fail = true, |
| .get_nr = test__pfm_subtest_get_nr, |
| .get_desc = test__pfm_subtest_get_desc, |
| } |
| }, |
| { |
| .desc = "Test api io", |
| .func = test__api_io, |
| }, |
| { |
| .desc = "maps__merge_in", |
| .func = test__maps__merge_in, |
| }, |
| { |
| .desc = "Demangle Java", |
| .func = test__demangle_java, |
| }, |
| { |
| .desc = "Demangle OCaml", |
| .func = test__demangle_ocaml, |
| }, |
| { |
| .desc = "Parse and process metrics", |
| .func = test__parse_metric, |
| }, |
| { |
| .desc = "PE file support", |
| .func = test__pe_file_parsing, |
| }, |
| { |
| .desc = "Event expansion for cgroups", |
| .func = test__expand_cgroup_events, |
| }, |
| { |
| .desc = "Convert perf time to TSC", |
| .func = test__perf_time_to_tsc, |
| .is_supported = test__tsc_is_supported, |
| }, |
| { |
| .func = NULL, |
| }, |
| }; |
| |
| static struct test *tests[] = { |
| generic_tests, |
| arch_tests, |
| }; |
| |
| static bool perf_test__matches(const char *desc, int curr, int argc, const char *argv[]) |
| { |
| int i; |
| |
| if (argc == 0) |
| return true; |
| |
| for (i = 0; i < argc; ++i) { |
| char *end; |
| long nr = strtoul(argv[i], &end, 10); |
| |
| if (*end == '\0') { |
| if (nr == curr + 1) |
| return true; |
| continue; |
| } |
| |
| if (strcasestr(desc, argv[i])) |
| return true; |
| } |
| |
| return false; |
| } |
| |
| static int run_test(struct test *test, int subtest) |
| { |
| int status, err = -1, child = dont_fork ? 0 : fork(); |
| char sbuf[STRERR_BUFSIZE]; |
| |
| if (child < 0) { |
| pr_err("failed to fork test: %s\n", |
| str_error_r(errno, sbuf, sizeof(sbuf))); |
| return -1; |
| } |
| |
| if (!child) { |
| if (!dont_fork) { |
| pr_debug("test child forked, pid %d\n", getpid()); |
| |
| if (verbose <= 0) { |
| int nullfd = open("/dev/null", O_WRONLY); |
| |
| if (nullfd >= 0) { |
| close(STDERR_FILENO); |
| close(STDOUT_FILENO); |
| |
| dup2(nullfd, STDOUT_FILENO); |
| dup2(STDOUT_FILENO, STDERR_FILENO); |
| close(nullfd); |
| } |
| } else { |
| signal(SIGSEGV, sighandler_dump_stack); |
| signal(SIGFPE, sighandler_dump_stack); |
| } |
| } |
| |
| err = test->func(test, subtest); |
| if (!dont_fork) |
| exit(err); |
| } |
| |
| if (!dont_fork) { |
| wait(&status); |
| |
| if (WIFEXITED(status)) { |
| err = (signed char)WEXITSTATUS(status); |
| pr_debug("test child finished with %d\n", err); |
| } else if (WIFSIGNALED(status)) { |
| err = -1; |
| pr_debug("test child interrupted\n"); |
| } |
| } |
| |
| return err; |
| } |
| |
| #define for_each_test(j, t) \ |
| for (j = 0; j < ARRAY_SIZE(tests); j++) \ |
| for (t = &tests[j][0]; t->func; t++) |
| |
| static int test_and_print(struct test *t, bool force_skip, int subtest) |
| { |
| int err; |
| |
| if (!force_skip) { |
| pr_debug("\n--- start ---\n"); |
| err = run_test(t, subtest); |
| pr_debug("---- end ----\n"); |
| } else { |
| pr_debug("\n--- force skipped ---\n"); |
| err = TEST_SKIP; |
| } |
| |
| if (!t->subtest.get_nr) |
| pr_debug("%s:", t->desc); |
| else |
| pr_debug("%s subtest %d:", t->desc, subtest + 1); |
| |
| switch (err) { |
| case TEST_OK: |
| pr_info(" Ok\n"); |
| break; |
| case TEST_SKIP: { |
| const char *skip_reason = NULL; |
| if (t->subtest.skip_reason) |
| skip_reason = t->subtest.skip_reason(subtest); |
| if (skip_reason) |
| color_fprintf(stderr, PERF_COLOR_YELLOW, " Skip (%s)\n", skip_reason); |
| else |
| color_fprintf(stderr, PERF_COLOR_YELLOW, " Skip\n"); |
| } |
| break; |
| case TEST_FAIL: |
| default: |
| color_fprintf(stderr, PERF_COLOR_RED, " FAILED!\n"); |
| break; |
| } |
| |
| return err; |
| } |
| |
| static const char *shell_test__description(char *description, size_t size, |
| const char *path, const char *name) |
| { |
| FILE *fp; |
| char filename[PATH_MAX]; |
| |
| path__join(filename, sizeof(filename), path, name); |
| fp = fopen(filename, "r"); |
| if (!fp) |
| return NULL; |
| |
| /* Skip shebang */ |
| while (fgetc(fp) != '\n'); |
| |
| description = fgets(description, size, fp); |
| fclose(fp); |
| |
| return description ? strim(description + 1) : NULL; |
| } |
| |
| #define for_each_shell_test(entlist, nr, base, ent) \ |
| for (int __i = 0; __i < nr && (ent = entlist[__i]); __i++) \ |
| if (!is_directory(base, ent) && ent->d_name[0] != '.') |
| |
| static const char *shell_tests__dir(char *path, size_t size) |
| { |
| const char *devel_dirs[] = { "./tools/perf/tests", "./tests", }; |
| char *exec_path; |
| unsigned int i; |
| |
| for (i = 0; i < ARRAY_SIZE(devel_dirs); ++i) { |
| struct stat st; |
| if (!lstat(devel_dirs[i], &st)) { |
| scnprintf(path, size, "%s/shell", devel_dirs[i]); |
| if (!lstat(devel_dirs[i], &st)) |
| return path; |
| } |
| } |
| |
| /* Then installed path. */ |
| exec_path = get_argv_exec_path(); |
| scnprintf(path, size, "%s/tests/shell", exec_path); |
| free(exec_path); |
| return path; |
| } |
| |
| static int shell_tests__max_desc_width(void) |
| { |
| struct dirent **entlist; |
| struct dirent *ent; |
| int n_dirs, e; |
| char path_dir[PATH_MAX]; |
| const char *path = shell_tests__dir(path_dir, sizeof(path_dir)); |
| int width = 0; |
| |
| if (path == NULL) |
| return -1; |
| |
| n_dirs = scandir(path, &entlist, NULL, alphasort); |
| if (n_dirs == -1) |
| return -1; |
| |
| for_each_shell_test(entlist, n_dirs, path, ent) { |
| char bf[256]; |
| const char *desc = shell_test__description(bf, sizeof(bf), path, ent->d_name); |
| |
| if (desc) { |
| int len = strlen(desc); |
| |
| if (width < len) |
| width = len; |
| } |
| } |
| |
| for (e = 0; e < n_dirs; e++) |
| zfree(&entlist[e]); |
| free(entlist); |
| return width; |
| } |
| |
| struct shell_test { |
| const char *dir; |
| const char *file; |
| }; |
| |
| static int shell_test__run(struct test *test, int subdir __maybe_unused) |
| { |
| int err; |
| char script[PATH_MAX]; |
| struct shell_test *st = test->priv; |
| |
| path__join(script, sizeof(script) - 3, st->dir, st->file); |
| |
| if (verbose) |
| strncat(script, " -v", sizeof(script) - strlen(script) - 1); |
| |
| err = system(script); |
| if (!err) |
| return TEST_OK; |
| |
| return WEXITSTATUS(err) == 2 ? TEST_SKIP : TEST_FAIL; |
| } |
| |
| static int run_shell_tests(int argc, const char *argv[], int i, int width) |
| { |
| struct dirent **entlist; |
| struct dirent *ent; |
| int n_dirs, e; |
| char path_dir[PATH_MAX]; |
| struct shell_test st = { |
| .dir = shell_tests__dir(path_dir, sizeof(path_dir)), |
| }; |
| |
| if (st.dir == NULL) |
| return -1; |
| |
| n_dirs = scandir(st.dir, &entlist, NULL, alphasort); |
| if (n_dirs == -1) { |
| pr_err("failed to open shell test directory: %s\n", |
| st.dir); |
| return -1; |
| } |
| |
| for_each_shell_test(entlist, n_dirs, st.dir, ent) { |
| int curr = i++; |
| char desc[256]; |
| struct test test = { |
| .desc = shell_test__description(desc, sizeof(desc), st.dir, ent->d_name), |
| .func = shell_test__run, |
| .priv = &st, |
| }; |
| |
| if (!perf_test__matches(test.desc, curr, argc, argv)) |
| continue; |
| |
| st.file = ent->d_name; |
| pr_info("%2d: %-*s:", i, width, test.desc); |
| test_and_print(&test, false, -1); |
| } |
| |
| for (e = 0; e < n_dirs; e++) |
| zfree(&entlist[e]); |
| free(entlist); |
| return 0; |
| } |
| |
| static int __cmd_test(int argc, const char *argv[], struct intlist *skiplist) |
| { |
| struct test *t; |
| unsigned int j; |
| int i = 0; |
| int width = shell_tests__max_desc_width(); |
| |
| for_each_test(j, t) { |
| int len = strlen(t->desc); |
| |
| if (width < len) |
| width = len; |
| } |
| |
| for_each_test(j, t) { |
| int curr = i++, err; |
| int subi; |
| |
| if (!perf_test__matches(t->desc, curr, argc, argv)) { |
| bool skip = true; |
| int subn; |
| |
| if (!t->subtest.get_nr) |
| continue; |
| |
| subn = t->subtest.get_nr(); |
| |
| for (subi = 0; subi < subn; subi++) { |
| if (perf_test__matches(t->subtest.get_desc(subi), curr, argc, argv)) |
| skip = false; |
| } |
| |
| if (skip) |
| continue; |
| } |
| |
| if (t->is_supported && !t->is_supported()) { |
| pr_debug("%2d: %-*s: Disabled\n", i, width, t->desc); |
| continue; |
| } |
| |
| pr_info("%2d: %-*s:", i, width, t->desc); |
| |
| if (intlist__find(skiplist, i)) { |
| color_fprintf(stderr, PERF_COLOR_YELLOW, " Skip (user override)\n"); |
| continue; |
| } |
| |
| if (!t->subtest.get_nr) { |
| test_and_print(t, false, -1); |
| } else { |
| int subn = t->subtest.get_nr(); |
| /* |
| * minus 2 to align with normal testcases. |
| * For subtest we print additional '.x' in number. |
| * for example: |
| * |
| * 35: Test LLVM searching and compiling : |
| * 35.1: Basic BPF llvm compiling test : Ok |
| */ |
| int subw = width > 2 ? width - 2 : width; |
| bool skip = false; |
| |
| if (subn <= 0) { |
| color_fprintf(stderr, PERF_COLOR_YELLOW, |
| " Skip (not compiled in)\n"); |
| continue; |
| } |
| pr_info("\n"); |
| |
| for (subi = 0; subi < subn; subi++) { |
| int len = strlen(t->subtest.get_desc(subi)); |
| |
| if (subw < len) |
| subw = len; |
| } |
| |
| for (subi = 0; subi < subn; subi++) { |
| if (!perf_test__matches(t->subtest.get_desc(subi), curr, argc, argv)) |
| continue; |
| |
| pr_info("%2d.%1d: %-*s:", i, subi + 1, subw, |
| t->subtest.get_desc(subi)); |
| err = test_and_print(t, skip, subi); |
| if (err != TEST_OK && t->subtest.skip_if_fail) |
| skip = true; |
| } |
| } |
| } |
| |
| return run_shell_tests(argc, argv, i, width); |
| } |
| |
| static int perf_test__list_shell(int argc, const char **argv, int i) |
| { |
| struct dirent **entlist; |
| struct dirent *ent; |
| int n_dirs, e; |
| char path_dir[PATH_MAX]; |
| const char *path = shell_tests__dir(path_dir, sizeof(path_dir)); |
| |
| if (path == NULL) |
| return -1; |
| |
| n_dirs = scandir(path, &entlist, NULL, alphasort); |
| if (n_dirs == -1) |
| return -1; |
| |
| for_each_shell_test(entlist, n_dirs, path, ent) { |
| int curr = i++; |
| char bf[256]; |
| struct test t = { |
| .desc = shell_test__description(bf, sizeof(bf), path, ent->d_name), |
| }; |
| |
| if (!perf_test__matches(t.desc, curr, argc, argv)) |
| continue; |
| |
| pr_info("%2d: %s\n", i, t.desc); |
| |
| } |
| |
| for (e = 0; e < n_dirs; e++) |
| zfree(&entlist[e]); |
| free(entlist); |
| return 0; |
| } |
| |
| static int perf_test__list(int argc, const char **argv) |
| { |
| unsigned int j; |
| struct test *t; |
| int i = 0; |
| |
| for_each_test(j, t) { |
| int curr = i++; |
| |
| if (!perf_test__matches(t->desc, curr, argc, argv) || |
| (t->is_supported && !t->is_supported())) |
| continue; |
| |
| pr_info("%2d: %s\n", i, t->desc); |
| |
| if (t->subtest.get_nr) { |
| int subn = t->subtest.get_nr(); |
| int subi; |
| |
| for (subi = 0; subi < subn; subi++) |
| pr_info("%2d:%1d: %s\n", i, subi + 1, |
| t->subtest.get_desc(subi)); |
| } |
| } |
| |
| perf_test__list_shell(argc, argv, i); |
| |
| return 0; |
| } |
| |
| int cmd_test(int argc, const char **argv) |
| { |
| const char *test_usage[] = { |
| "perf test [<options>] [{list <test-name-fragment>|[<test-name-fragments>|<test-numbers>]}]", |
| NULL, |
| }; |
| const char *skip = NULL; |
| const struct option test_options[] = { |
| OPT_STRING('s', "skip", &skip, "tests", "tests to skip"), |
| OPT_INCR('v', "verbose", &verbose, |
| "be more verbose (show symbol address, etc)"), |
| OPT_BOOLEAN('F', "dont-fork", &dont_fork, |
| "Do not fork for testcase"), |
| OPT_END() |
| }; |
| const char * const test_subcommands[] = { "list", NULL }; |
| struct intlist *skiplist = NULL; |
| int ret = hists__init(); |
| |
| if (ret < 0) |
| return ret; |
| |
| argc = parse_options_subcommand(argc, argv, test_options, test_subcommands, test_usage, 0); |
| if (argc >= 1 && !strcmp(argv[0], "list")) |
| return perf_test__list(argc - 1, argv + 1); |
| |
| symbol_conf.priv_size = sizeof(int); |
| symbol_conf.sort_by_name = true; |
| symbol_conf.try_vmlinux_path = true; |
| |
| if (symbol__init(NULL) < 0) |
| return -1; |
| |
| if (skip != NULL) |
| skiplist = intlist__new(skip); |
| /* |
| * Tests that create BPF maps, for instance, need more than the 64K |
| * default: |
| */ |
| rlimit__bump_memlock(); |
| |
| return __cmd_test(argc, argv, skiplist); |
| } |