| // SPDX-License-Identifier: GPL-2.0 |
| #include <linux/list.h> |
| #include <linux/list_sort.h> |
| #include <linux/string.h> |
| #include <linux/zalloc.h> |
| #include <subcmd/pager.h> |
| #include <sys/types.h> |
| #include <ctype.h> |
| #include <dirent.h> |
| #include <pthread.h> |
| #include <string.h> |
| #include <unistd.h> |
| #include "cpumap.h" |
| #include "debug.h" |
| #include "evsel.h" |
| #include "pmus.h" |
| #include "pmu.h" |
| #include "print-events.h" |
| |
| /* |
| * core_pmus: A PMU belongs to core_pmus if it's name is "cpu" or it's sysfs |
| * directory contains "cpus" file. All PMUs belonging to core_pmus |
| * must have pmu->is_core=1. If there are more than one PMU in |
| * this list, perf interprets it as a heterogeneous platform. |
| * (FWIW, certain ARM platforms having heterogeneous cores uses |
| * homogeneous PMU, and thus they are treated as homogeneous |
| * platform by perf because core_pmus will have only one entry) |
| * other_pmus: All other PMUs which are not part of core_pmus list. It doesn't |
| * matter whether PMU is present per SMT-thread or outside of the |
| * core in the hw. For e.g., an instance of AMD ibs_fetch// and |
| * ibs_op// PMUs is present in each hw SMT thread, however they |
| * are captured under other_pmus. PMUs belonging to other_pmus |
| * must have pmu->is_core=0 but pmu->is_uncore could be 0 or 1. |
| */ |
| static LIST_HEAD(core_pmus); |
| static LIST_HEAD(other_pmus); |
| static bool read_sysfs_core_pmus; |
| static bool read_sysfs_all_pmus; |
| |
| static void pmu_read_sysfs(bool core_only); |
| |
| int pmu_name_len_no_suffix(const char *str, unsigned long *num) |
| { |
| int orig_len, len; |
| |
| orig_len = len = strlen(str); |
| |
| /* Non-uncore PMUs have their full length, for example, i915. */ |
| if (!strstarts(str, "uncore_")) |
| return len; |
| |
| /* |
| * Count trailing digits and '_', if '_{num}' suffix isn't present use |
| * the full length. |
| */ |
| while (len > 0 && isdigit(str[len - 1])) |
| len--; |
| |
| if (len > 0 && len != orig_len && str[len - 1] == '_') { |
| if (num) |
| *num = strtoul(&str[len], NULL, 10); |
| return len - 1; |
| } |
| return orig_len; |
| } |
| |
| void perf_pmus__destroy(void) |
| { |
| struct perf_pmu *pmu, *tmp; |
| |
| list_for_each_entry_safe(pmu, tmp, &core_pmus, list) { |
| list_del(&pmu->list); |
| |
| perf_pmu__delete(pmu); |
| } |
| list_for_each_entry_safe(pmu, tmp, &other_pmus, list) { |
| list_del(&pmu->list); |
| |
| perf_pmu__delete(pmu); |
| } |
| read_sysfs_core_pmus = false; |
| read_sysfs_all_pmus = false; |
| } |
| |
| static struct perf_pmu *pmu_find(const char *name) |
| { |
| struct perf_pmu *pmu; |
| |
| list_for_each_entry(pmu, &core_pmus, list) { |
| if (!strcmp(pmu->name, name) || |
| (pmu->alias_name && !strcmp(pmu->alias_name, name))) |
| return pmu; |
| } |
| list_for_each_entry(pmu, &other_pmus, list) { |
| if (!strcmp(pmu->name, name) || |
| (pmu->alias_name && !strcmp(pmu->alias_name, name))) |
| return pmu; |
| } |
| |
| return NULL; |
| } |
| |
| struct perf_pmu *perf_pmus__find(const char *name) |
| { |
| struct perf_pmu *pmu; |
| int dirfd; |
| bool core_pmu; |
| |
| /* |
| * Once PMU is loaded it stays in the list, |
| * so we keep us from multiple reading/parsing |
| * the pmu format definitions. |
| */ |
| pmu = pmu_find(name); |
| if (pmu) |
| return pmu; |
| |
| if (read_sysfs_all_pmus) |
| return NULL; |
| |
| core_pmu = is_pmu_core(name); |
| if (core_pmu && read_sysfs_core_pmus) |
| return NULL; |
| |
| dirfd = perf_pmu__event_source_devices_fd(); |
| pmu = perf_pmu__lookup(core_pmu ? &core_pmus : &other_pmus, dirfd, name); |
| close(dirfd); |
| |
| if (!pmu) { |
| /* |
| * Looking up an inidividual PMU failed. This may mean name is |
| * an alias, so read the PMUs from sysfs and try to find again. |
| */ |
| pmu_read_sysfs(core_pmu); |
| pmu = pmu_find(name); |
| } |
| return pmu; |
| } |
| |
| static struct perf_pmu *perf_pmu__find2(int dirfd, const char *name) |
| { |
| struct perf_pmu *pmu; |
| bool core_pmu; |
| |
| /* |
| * Once PMU is loaded it stays in the list, |
| * so we keep us from multiple reading/parsing |
| * the pmu format definitions. |
| */ |
| pmu = pmu_find(name); |
| if (pmu) |
| return pmu; |
| |
| if (read_sysfs_all_pmus) |
| return NULL; |
| |
| core_pmu = is_pmu_core(name); |
| if (core_pmu && read_sysfs_core_pmus) |
| return NULL; |
| |
| return perf_pmu__lookup(core_pmu ? &core_pmus : &other_pmus, dirfd, name); |
| } |
| |
| static int pmus_cmp(void *priv __maybe_unused, |
| const struct list_head *lhs, const struct list_head *rhs) |
| { |
| unsigned long lhs_num = 0, rhs_num = 0; |
| struct perf_pmu *lhs_pmu = container_of(lhs, struct perf_pmu, list); |
| struct perf_pmu *rhs_pmu = container_of(rhs, struct perf_pmu, list); |
| const char *lhs_pmu_name = lhs_pmu->name ?: ""; |
| const char *rhs_pmu_name = rhs_pmu->name ?: ""; |
| int lhs_pmu_name_len = pmu_name_len_no_suffix(lhs_pmu_name, &lhs_num); |
| int rhs_pmu_name_len = pmu_name_len_no_suffix(rhs_pmu_name, &rhs_num); |
| int ret = strncmp(lhs_pmu_name, rhs_pmu_name, |
| lhs_pmu_name_len < rhs_pmu_name_len ? lhs_pmu_name_len : rhs_pmu_name_len); |
| |
| if (lhs_pmu_name_len != rhs_pmu_name_len || ret != 0 || lhs_pmu_name_len == 0) |
| return ret; |
| |
| return lhs_num < rhs_num ? -1 : (lhs_num > rhs_num ? 1 : 0); |
| } |
| |
| /* Add all pmus in sysfs to pmu list: */ |
| static void pmu_read_sysfs(bool core_only) |
| { |
| int fd; |
| DIR *dir; |
| struct dirent *dent; |
| |
| if (read_sysfs_all_pmus || (core_only && read_sysfs_core_pmus)) |
| return; |
| |
| fd = perf_pmu__event_source_devices_fd(); |
| if (fd < 0) |
| return; |
| |
| dir = fdopendir(fd); |
| if (!dir) { |
| close(fd); |
| return; |
| } |
| |
| while ((dent = readdir(dir))) { |
| if (!strcmp(dent->d_name, ".") || !strcmp(dent->d_name, "..")) |
| continue; |
| if (core_only && !is_pmu_core(dent->d_name)) |
| continue; |
| /* add to static LIST_HEAD(core_pmus) or LIST_HEAD(other_pmus): */ |
| perf_pmu__find2(fd, dent->d_name); |
| } |
| |
| closedir(dir); |
| if (list_empty(&core_pmus)) { |
| if (!perf_pmu__create_placeholder_core_pmu(&core_pmus)) |
| pr_err("Failure to set up any core PMUs\n"); |
| } |
| list_sort(NULL, &core_pmus, pmus_cmp); |
| list_sort(NULL, &other_pmus, pmus_cmp); |
| if (!list_empty(&core_pmus)) { |
| read_sysfs_core_pmus = true; |
| if (!core_only) |
| read_sysfs_all_pmus = true; |
| } |
| } |
| |
| static struct perf_pmu *__perf_pmus__find_by_type(unsigned int type) |
| { |
| struct perf_pmu *pmu; |
| |
| list_for_each_entry(pmu, &core_pmus, list) { |
| if (pmu->type == type) |
| return pmu; |
| } |
| |
| list_for_each_entry(pmu, &other_pmus, list) { |
| if (pmu->type == type) |
| return pmu; |
| } |
| return NULL; |
| } |
| |
| struct perf_pmu *perf_pmus__find_by_type(unsigned int type) |
| { |
| struct perf_pmu *pmu = __perf_pmus__find_by_type(type); |
| |
| if (pmu || read_sysfs_all_pmus) |
| return pmu; |
| |
| pmu_read_sysfs(/*core_only=*/false); |
| pmu = __perf_pmus__find_by_type(type); |
| return pmu; |
| } |
| |
| /* |
| * pmu iterator: If pmu is NULL, we start at the begin, otherwise return the |
| * next pmu. Returns NULL on end. |
| */ |
| struct perf_pmu *perf_pmus__scan(struct perf_pmu *pmu) |
| { |
| bool use_core_pmus = !pmu || pmu->is_core; |
| |
| if (!pmu) { |
| pmu_read_sysfs(/*core_only=*/false); |
| pmu = list_prepare_entry(pmu, &core_pmus, list); |
| } |
| if (use_core_pmus) { |
| list_for_each_entry_continue(pmu, &core_pmus, list) |
| return pmu; |
| |
| pmu = NULL; |
| pmu = list_prepare_entry(pmu, &other_pmus, list); |
| } |
| list_for_each_entry_continue(pmu, &other_pmus, list) |
| return pmu; |
| return NULL; |
| } |
| |
| struct perf_pmu *perf_pmus__scan_core(struct perf_pmu *pmu) |
| { |
| if (!pmu) { |
| pmu_read_sysfs(/*core_only=*/true); |
| return list_first_entry_or_null(&core_pmus, typeof(*pmu), list); |
| } |
| list_for_each_entry_continue(pmu, &core_pmus, list) |
| return pmu; |
| |
| return NULL; |
| } |
| |
| static struct perf_pmu *perf_pmus__scan_skip_duplicates(struct perf_pmu *pmu) |
| { |
| bool use_core_pmus = !pmu || pmu->is_core; |
| int last_pmu_name_len = 0; |
| const char *last_pmu_name = (pmu && pmu->name) ? pmu->name : ""; |
| |
| if (!pmu) { |
| pmu_read_sysfs(/*core_only=*/false); |
| pmu = list_prepare_entry(pmu, &core_pmus, list); |
| } else |
| last_pmu_name_len = pmu_name_len_no_suffix(pmu->name ?: "", NULL); |
| |
| if (use_core_pmus) { |
| list_for_each_entry_continue(pmu, &core_pmus, list) { |
| int pmu_name_len = pmu_name_len_no_suffix(pmu->name ?: "", /*num=*/NULL); |
| |
| if (last_pmu_name_len == pmu_name_len && |
| !strncmp(last_pmu_name, pmu->name ?: "", pmu_name_len)) |
| continue; |
| |
| return pmu; |
| } |
| pmu = NULL; |
| pmu = list_prepare_entry(pmu, &other_pmus, list); |
| } |
| list_for_each_entry_continue(pmu, &other_pmus, list) { |
| int pmu_name_len = pmu_name_len_no_suffix(pmu->name ?: "", /*num=*/NULL); |
| |
| if (last_pmu_name_len == pmu_name_len && |
| !strncmp(last_pmu_name, pmu->name ?: "", pmu_name_len)) |
| continue; |
| |
| return pmu; |
| } |
| return NULL; |
| } |
| |
| const struct perf_pmu *perf_pmus__pmu_for_pmu_filter(const char *str) |
| { |
| struct perf_pmu *pmu = NULL; |
| |
| while ((pmu = perf_pmus__scan(pmu)) != NULL) { |
| if (!strcmp(pmu->name, str)) |
| return pmu; |
| /* Ignore "uncore_" prefix. */ |
| if (!strncmp(pmu->name, "uncore_", 7)) { |
| if (!strcmp(pmu->name + 7, str)) |
| return pmu; |
| } |
| /* Ignore "cpu_" prefix on Intel hybrid PMUs. */ |
| if (!strncmp(pmu->name, "cpu_", 4)) { |
| if (!strcmp(pmu->name + 4, str)) |
| return pmu; |
| } |
| } |
| return NULL; |
| } |
| |
| /** Struct for ordering events as output in perf list. */ |
| struct sevent { |
| /** PMU for event. */ |
| const struct perf_pmu *pmu; |
| const char *name; |
| const char* alias; |
| const char *scale_unit; |
| const char *desc; |
| const char *long_desc; |
| const char *encoding_desc; |
| const char *topic; |
| const char *pmu_name; |
| bool deprecated; |
| }; |
| |
| static int cmp_sevent(const void *a, const void *b) |
| { |
| const struct sevent *as = a; |
| const struct sevent *bs = b; |
| bool a_iscpu, b_iscpu; |
| int ret; |
| |
| /* Put extra events last. */ |
| if (!!as->desc != !!bs->desc) |
| return !!as->desc - !!bs->desc; |
| |
| /* Order by topics. */ |
| ret = strcmp(as->topic ?: "", bs->topic ?: ""); |
| if (ret) |
| return ret; |
| |
| /* Order CPU core events to be first */ |
| a_iscpu = as->pmu ? as->pmu->is_core : true; |
| b_iscpu = bs->pmu ? bs->pmu->is_core : true; |
| if (a_iscpu != b_iscpu) |
| return a_iscpu ? -1 : 1; |
| |
| /* Order by PMU name. */ |
| if (as->pmu != bs->pmu) { |
| ret = strcmp(as->pmu_name ?: "", bs->pmu_name ?: ""); |
| if (ret) |
| return ret; |
| } |
| |
| /* Order by event name. */ |
| return strcmp(as->name, bs->name); |
| } |
| |
| static bool pmu_alias_is_duplicate(struct sevent *a, struct sevent *b) |
| { |
| /* Different names -> never duplicates */ |
| if (strcmp(a->name ?: "//", b->name ?: "//")) |
| return false; |
| |
| /* Don't remove duplicates for different PMUs */ |
| return strcmp(a->pmu_name, b->pmu_name) == 0; |
| } |
| |
| struct events_callback_state { |
| struct sevent *aliases; |
| size_t aliases_len; |
| size_t index; |
| }; |
| |
| static int perf_pmus__print_pmu_events__callback(void *vstate, |
| struct pmu_event_info *info) |
| { |
| struct events_callback_state *state = vstate; |
| struct sevent *s; |
| |
| if (state->index >= state->aliases_len) { |
| pr_err("Unexpected event %s/%s/\n", info->pmu->name, info->name); |
| return 1; |
| } |
| s = &state->aliases[state->index]; |
| s->pmu = info->pmu; |
| #define COPY_STR(str) s->str = info->str ? strdup(info->str) : NULL |
| COPY_STR(name); |
| COPY_STR(alias); |
| COPY_STR(scale_unit); |
| COPY_STR(desc); |
| COPY_STR(long_desc); |
| COPY_STR(encoding_desc); |
| COPY_STR(topic); |
| COPY_STR(pmu_name); |
| #undef COPY_STR |
| s->deprecated = info->deprecated; |
| state->index++; |
| return 0; |
| } |
| |
| void perf_pmus__print_pmu_events(const struct print_callbacks *print_cb, void *print_state) |
| { |
| struct perf_pmu *pmu; |
| int printed = 0; |
| int len; |
| struct sevent *aliases; |
| struct events_callback_state state; |
| bool skip_duplicate_pmus = print_cb->skip_duplicate_pmus(print_state); |
| struct perf_pmu *(*scan_fn)(struct perf_pmu *); |
| |
| if (skip_duplicate_pmus) |
| scan_fn = perf_pmus__scan_skip_duplicates; |
| else |
| scan_fn = perf_pmus__scan; |
| |
| pmu = NULL; |
| len = 0; |
| while ((pmu = scan_fn(pmu)) != NULL) |
| len += perf_pmu__num_events(pmu); |
| |
| aliases = zalloc(sizeof(struct sevent) * len); |
| if (!aliases) { |
| pr_err("FATAL: not enough memory to print PMU events\n"); |
| return; |
| } |
| pmu = NULL; |
| state = (struct events_callback_state) { |
| .aliases = aliases, |
| .aliases_len = len, |
| .index = 0, |
| }; |
| while ((pmu = scan_fn(pmu)) != NULL) { |
| perf_pmu__for_each_event(pmu, skip_duplicate_pmus, &state, |
| perf_pmus__print_pmu_events__callback); |
| } |
| qsort(aliases, len, sizeof(struct sevent), cmp_sevent); |
| for (int j = 0; j < len; j++) { |
| /* Skip duplicates */ |
| if (j > 0 && pmu_alias_is_duplicate(&aliases[j], &aliases[j - 1])) |
| continue; |
| |
| print_cb->print_event(print_state, |
| aliases[j].pmu_name, |
| aliases[j].topic, |
| aliases[j].name, |
| aliases[j].alias, |
| aliases[j].scale_unit, |
| aliases[j].deprecated, |
| "Kernel PMU event", |
| aliases[j].desc, |
| aliases[j].long_desc, |
| aliases[j].encoding_desc); |
| zfree(&aliases[j].name); |
| zfree(&aliases[j].alias); |
| zfree(&aliases[j].scale_unit); |
| zfree(&aliases[j].desc); |
| zfree(&aliases[j].long_desc); |
| zfree(&aliases[j].encoding_desc); |
| zfree(&aliases[j].topic); |
| zfree(&aliases[j].pmu_name); |
| } |
| if (printed && pager_in_use()) |
| printf("\n"); |
| |
| zfree(&aliases); |
| } |
| |
| bool perf_pmus__have_event(const char *pname, const char *name) |
| { |
| struct perf_pmu *pmu = perf_pmus__find(pname); |
| |
| return pmu && perf_pmu__have_event(pmu, name); |
| } |
| |
| int perf_pmus__num_core_pmus(void) |
| { |
| static int count; |
| |
| if (!count) { |
| struct perf_pmu *pmu = NULL; |
| |
| while ((pmu = perf_pmus__scan_core(pmu)) != NULL) |
| count++; |
| } |
| return count; |
| } |
| |
| static bool __perf_pmus__supports_extended_type(void) |
| { |
| struct perf_pmu *pmu = NULL; |
| |
| if (perf_pmus__num_core_pmus() <= 1) |
| return false; |
| |
| while ((pmu = perf_pmus__scan_core(pmu)) != NULL) { |
| if (!is_event_supported(PERF_TYPE_HARDWARE, PERF_COUNT_HW_CPU_CYCLES | ((__u64)pmu->type << PERF_PMU_TYPE_SHIFT))) |
| return false; |
| } |
| |
| return true; |
| } |
| |
| static bool perf_pmus__do_support_extended_type; |
| |
| static void perf_pmus__init_supports_extended_type(void) |
| { |
| perf_pmus__do_support_extended_type = __perf_pmus__supports_extended_type(); |
| } |
| |
| bool perf_pmus__supports_extended_type(void) |
| { |
| static pthread_once_t extended_type_once = PTHREAD_ONCE_INIT; |
| |
| pthread_once(&extended_type_once, perf_pmus__init_supports_extended_type); |
| |
| return perf_pmus__do_support_extended_type; |
| } |
| |
| char *perf_pmus__default_pmu_name(void) |
| { |
| int fd; |
| DIR *dir; |
| struct dirent *dent; |
| char *result = NULL; |
| |
| if (!list_empty(&core_pmus)) |
| return strdup(list_first_entry(&core_pmus, struct perf_pmu, list)->name); |
| |
| fd = perf_pmu__event_source_devices_fd(); |
| if (fd < 0) |
| return strdup("cpu"); |
| |
| dir = fdopendir(fd); |
| if (!dir) { |
| close(fd); |
| return strdup("cpu"); |
| } |
| |
| while ((dent = readdir(dir))) { |
| if (!strcmp(dent->d_name, ".") || !strcmp(dent->d_name, "..")) |
| continue; |
| if (is_pmu_core(dent->d_name)) { |
| result = strdup(dent->d_name); |
| break; |
| } |
| } |
| |
| closedir(dir); |
| return result ?: strdup("cpu"); |
| } |
| |
| struct perf_pmu *evsel__find_pmu(const struct evsel *evsel) |
| { |
| struct perf_pmu *pmu = evsel->pmu; |
| |
| if (!pmu) { |
| pmu = perf_pmus__find_by_type(evsel->core.attr.type); |
| ((struct evsel *)evsel)->pmu = pmu; |
| } |
| return pmu; |
| } |
| |
| struct perf_pmu *perf_pmus__find_core_pmu(void) |
| { |
| return perf_pmus__scan_core(NULL); |
| } |