| // SPDX-License-Identifier: GPL-2.0-only |
| /* |
| * Copyright (c) 2017, Intel Corporation. |
| */ |
| |
| /* Manage metrics and groups of metrics from JSON files */ |
| |
| #include "metricgroup.h" |
| #include "debug.h" |
| #include "evlist.h" |
| #include "evsel.h" |
| #include "strbuf.h" |
| #include "pmu.h" |
| #include "pmus.h" |
| #include "print-events.h" |
| #include "smt.h" |
| #include "expr.h" |
| #include "rblist.h" |
| #include <string.h> |
| #include <errno.h> |
| #include "strlist.h" |
| #include <assert.h> |
| #include <linux/ctype.h> |
| #include <linux/list_sort.h> |
| #include <linux/string.h> |
| #include <linux/zalloc.h> |
| #include <perf/cpumap.h> |
| #include <subcmd/parse-options.h> |
| #include <api/fs/fs.h> |
| #include "util.h" |
| #include <asm/bug.h> |
| #include "cgroup.h" |
| #include "util/hashmap.h" |
| |
| struct metric_event *metricgroup__lookup(struct rblist *metric_events, |
| struct evsel *evsel, |
| bool create) |
| { |
| struct rb_node *nd; |
| struct metric_event me = { |
| .evsel = evsel |
| }; |
| |
| if (!metric_events) |
| return NULL; |
| |
| if (evsel && evsel->metric_leader) |
| me.evsel = evsel->metric_leader; |
| nd = rblist__find(metric_events, &me); |
| if (nd) |
| return container_of(nd, struct metric_event, nd); |
| if (create) { |
| rblist__add_node(metric_events, &me); |
| nd = rblist__find(metric_events, &me); |
| if (nd) |
| return container_of(nd, struct metric_event, nd); |
| } |
| return NULL; |
| } |
| |
| static int metric_event_cmp(struct rb_node *rb_node, const void *entry) |
| { |
| struct metric_event *a = container_of(rb_node, |
| struct metric_event, |
| nd); |
| const struct metric_event *b = entry; |
| |
| if (a->evsel == b->evsel) |
| return 0; |
| if ((char *)a->evsel < (char *)b->evsel) |
| return -1; |
| return +1; |
| } |
| |
| static struct rb_node *metric_event_new(struct rblist *rblist __maybe_unused, |
| const void *entry) |
| { |
| struct metric_event *me = malloc(sizeof(struct metric_event)); |
| |
| if (!me) |
| return NULL; |
| memcpy(me, entry, sizeof(struct metric_event)); |
| me->evsel = ((struct metric_event *)entry)->evsel; |
| me->is_default = false; |
| INIT_LIST_HEAD(&me->head); |
| return &me->nd; |
| } |
| |
| static void metric_event_delete(struct rblist *rblist __maybe_unused, |
| struct rb_node *rb_node) |
| { |
| struct metric_event *me = container_of(rb_node, struct metric_event, nd); |
| struct metric_expr *expr, *tmp; |
| |
| list_for_each_entry_safe(expr, tmp, &me->head, nd) { |
| zfree(&expr->metric_name); |
| zfree(&expr->metric_refs); |
| zfree(&expr->metric_events); |
| free(expr); |
| } |
| |
| free(me); |
| } |
| |
| static void metricgroup__rblist_init(struct rblist *metric_events) |
| { |
| rblist__init(metric_events); |
| metric_events->node_cmp = metric_event_cmp; |
| metric_events->node_new = metric_event_new; |
| metric_events->node_delete = metric_event_delete; |
| } |
| |
| void metricgroup__rblist_exit(struct rblist *metric_events) |
| { |
| rblist__exit(metric_events); |
| } |
| |
| /** |
| * The metric under construction. The data held here will be placed in a |
| * metric_expr. |
| */ |
| struct metric { |
| struct list_head nd; |
| /** |
| * The expression parse context importantly holding the IDs contained |
| * within the expression. |
| */ |
| struct expr_parse_ctx *pctx; |
| const char *pmu; |
| /** The name of the metric such as "IPC". */ |
| const char *metric_name; |
| /** Modifier on the metric such as "u" or NULL for none. */ |
| const char *modifier; |
| /** The expression to parse, for example, "instructions/cycles". */ |
| const char *metric_expr; |
| /** Optional threshold expression where zero value is green, otherwise red. */ |
| const char *metric_threshold; |
| /** |
| * The "ScaleUnit" that scales and adds a unit to the metric during |
| * output. |
| */ |
| const char *metric_unit; |
| /** |
| * Optional name of the metric group reported |
| * if the Default metric group is being processed. |
| */ |
| const char *default_metricgroup_name; |
| /** Optional null terminated array of referenced metrics. */ |
| struct metric_ref *metric_refs; |
| /** |
| * Should events of the metric be grouped? |
| */ |
| bool group_events; |
| /** |
| * Parsed events for the metric. Optional as events may be taken from a |
| * different metric whose group contains all the IDs necessary for this |
| * one. |
| */ |
| struct evlist *evlist; |
| }; |
| |
| static void metric__watchdog_constraint_hint(const char *name, bool foot) |
| { |
| static bool violate_nmi_constraint; |
| |
| if (!foot) { |
| pr_warning("Not grouping metric %s's events.\n", name); |
| violate_nmi_constraint = true; |
| return; |
| } |
| |
| if (!violate_nmi_constraint) |
| return; |
| |
| pr_warning("Try disabling the NMI watchdog to comply NO_NMI_WATCHDOG metric constraint:\n" |
| " echo 0 > /proc/sys/kernel/nmi_watchdog\n" |
| " perf stat ...\n" |
| " echo 1 > /proc/sys/kernel/nmi_watchdog\n"); |
| } |
| |
| static bool metric__group_events(const struct pmu_metric *pm) |
| { |
| switch (pm->event_grouping) { |
| case MetricNoGroupEvents: |
| return false; |
| case MetricNoGroupEventsNmi: |
| if (!sysctl__nmi_watchdog_enabled()) |
| return true; |
| metric__watchdog_constraint_hint(pm->metric_name, /*foot=*/false); |
| return false; |
| case MetricNoGroupEventsSmt: |
| return !smt_on(); |
| case MetricGroupEvents: |
| default: |
| return true; |
| } |
| } |
| |
| static void metric__free(struct metric *m) |
| { |
| if (!m) |
| return; |
| |
| zfree(&m->metric_refs); |
| expr__ctx_free(m->pctx); |
| zfree(&m->modifier); |
| evlist__delete(m->evlist); |
| free(m); |
| } |
| |
| static struct metric *metric__new(const struct pmu_metric *pm, |
| const char *modifier, |
| bool metric_no_group, |
| int runtime, |
| const char *user_requested_cpu_list, |
| bool system_wide) |
| { |
| struct metric *m; |
| |
| m = zalloc(sizeof(*m)); |
| if (!m) |
| return NULL; |
| |
| m->pctx = expr__ctx_new(); |
| if (!m->pctx) |
| goto out_err; |
| |
| m->pmu = pm->pmu ?: "cpu"; |
| m->metric_name = pm->metric_name; |
| m->default_metricgroup_name = pm->default_metricgroup_name ?: ""; |
| m->modifier = NULL; |
| if (modifier) { |
| m->modifier = strdup(modifier); |
| if (!m->modifier) |
| goto out_err; |
| } |
| m->metric_expr = pm->metric_expr; |
| m->metric_threshold = pm->metric_threshold; |
| m->metric_unit = pm->unit; |
| m->pctx->sctx.user_requested_cpu_list = NULL; |
| if (user_requested_cpu_list) { |
| m->pctx->sctx.user_requested_cpu_list = strdup(user_requested_cpu_list); |
| if (!m->pctx->sctx.user_requested_cpu_list) |
| goto out_err; |
| } |
| m->pctx->sctx.runtime = runtime; |
| m->pctx->sctx.system_wide = system_wide; |
| m->group_events = !metric_no_group && metric__group_events(pm); |
| m->metric_refs = NULL; |
| m->evlist = NULL; |
| |
| return m; |
| out_err: |
| metric__free(m); |
| return NULL; |
| } |
| |
| static bool contains_metric_id(struct evsel **metric_events, int num_events, |
| const char *metric_id) |
| { |
| int i; |
| |
| for (i = 0; i < num_events; i++) { |
| if (!strcmp(evsel__metric_id(metric_events[i]), metric_id)) |
| return true; |
| } |
| return false; |
| } |
| |
| /** |
| * setup_metric_events - Find a group of events in metric_evlist that correspond |
| * to the IDs from a parsed metric expression. |
| * @pmu: The PMU for the IDs. |
| * @ids: the metric IDs to match. |
| * @metric_evlist: the list of perf events. |
| * @out_metric_events: holds the created metric events array. |
| */ |
| static int setup_metric_events(const char *pmu, struct hashmap *ids, |
| struct evlist *metric_evlist, |
| struct evsel ***out_metric_events) |
| { |
| struct evsel **metric_events; |
| const char *metric_id; |
| struct evsel *ev; |
| size_t ids_size, matched_events, i; |
| bool all_pmus = !strcmp(pmu, "all") || perf_pmus__num_core_pmus() == 1 || !is_pmu_core(pmu); |
| |
| *out_metric_events = NULL; |
| ids_size = hashmap__size(ids); |
| |
| metric_events = calloc(ids_size + 1, sizeof(void *)); |
| if (!metric_events) |
| return -ENOMEM; |
| |
| matched_events = 0; |
| evlist__for_each_entry(metric_evlist, ev) { |
| struct expr_id_data *val_ptr; |
| |
| /* Don't match events for the wrong hybrid PMU. */ |
| if (!all_pmus && ev->pmu_name && evsel__is_hybrid(ev) && |
| strcmp(ev->pmu_name, pmu)) |
| continue; |
| /* |
| * Check for duplicate events with the same name. For |
| * example, uncore_imc/cas_count_read/ will turn into 6 |
| * events per socket on skylakex. Only the first such |
| * event is placed in metric_events. |
| */ |
| metric_id = evsel__metric_id(ev); |
| if (contains_metric_id(metric_events, matched_events, metric_id)) |
| continue; |
| /* |
| * Does this event belong to the parse context? For |
| * combined or shared groups, this metric may not care |
| * about this event. |
| */ |
| if (hashmap__find(ids, metric_id, &val_ptr)) { |
| pr_debug("Matched metric-id %s to %s\n", metric_id, evsel__name(ev)); |
| metric_events[matched_events++] = ev; |
| |
| if (matched_events >= ids_size) |
| break; |
| } |
| } |
| if (matched_events < ids_size) { |
| free(metric_events); |
| return -EINVAL; |
| } |
| for (i = 0; i < ids_size; i++) { |
| ev = metric_events[i]; |
| ev->collect_stat = true; |
| |
| /* |
| * The metric leader points to the identically named |
| * event in metric_events. |
| */ |
| ev->metric_leader = ev; |
| /* |
| * Mark two events with identical names in the same |
| * group (or globally) as being in use as uncore events |
| * may be duplicated for each pmu. Set the metric leader |
| * of such events to be the event that appears in |
| * metric_events. |
| */ |
| metric_id = evsel__metric_id(ev); |
| evlist__for_each_entry_continue(metric_evlist, ev) { |
| if (!strcmp(evsel__metric_id(ev), metric_id)) |
| ev->metric_leader = metric_events[i]; |
| } |
| } |
| *out_metric_events = metric_events; |
| return 0; |
| } |
| |
| static bool match_metric(const char *metric_or_groups, const char *sought) |
| { |
| int len; |
| char *m; |
| |
| if (!sought) |
| return false; |
| if (!strcmp(sought, "all")) |
| return true; |
| if (!metric_or_groups) |
| return !strcasecmp(sought, "No_group"); |
| len = strlen(sought); |
| if (!strncasecmp(metric_or_groups, sought, len) && |
| (metric_or_groups[len] == 0 || metric_or_groups[len] == ';')) |
| return true; |
| m = strchr(metric_or_groups, ';'); |
| return m && match_metric(m + 1, sought); |
| } |
| |
| static bool match_pm_metric(const struct pmu_metric *pm, const char *pmu, const char *metric) |
| { |
| const char *pm_pmu = pm->pmu ?: "cpu"; |
| |
| if (strcmp(pmu, "all") && strcmp(pm_pmu, pmu)) |
| return false; |
| |
| return match_metric(pm->metric_group, metric) || |
| match_metric(pm->metric_name, metric); |
| } |
| |
| /** struct mep - RB-tree node for building printing information. */ |
| struct mep { |
| /** nd - RB-tree element. */ |
| struct rb_node nd; |
| /** @metric_group: Owned metric group name, separated others with ';'. */ |
| char *metric_group; |
| const char *metric_name; |
| const char *metric_desc; |
| const char *metric_long_desc; |
| const char *metric_expr; |
| const char *metric_threshold; |
| const char *metric_unit; |
| }; |
| |
| static int mep_cmp(struct rb_node *rb_node, const void *entry) |
| { |
| struct mep *a = container_of(rb_node, struct mep, nd); |
| struct mep *b = (struct mep *)entry; |
| int ret; |
| |
| ret = strcmp(a->metric_group, b->metric_group); |
| if (ret) |
| return ret; |
| |
| return strcmp(a->metric_name, b->metric_name); |
| } |
| |
| static struct rb_node *mep_new(struct rblist *rl __maybe_unused, const void *entry) |
| { |
| struct mep *me = malloc(sizeof(struct mep)); |
| |
| if (!me) |
| return NULL; |
| |
| memcpy(me, entry, sizeof(struct mep)); |
| return &me->nd; |
| } |
| |
| static void mep_delete(struct rblist *rl __maybe_unused, |
| struct rb_node *nd) |
| { |
| struct mep *me = container_of(nd, struct mep, nd); |
| |
| zfree(&me->metric_group); |
| free(me); |
| } |
| |
| static struct mep *mep_lookup(struct rblist *groups, const char *metric_group, |
| const char *metric_name) |
| { |
| struct rb_node *nd; |
| struct mep me = { |
| .metric_group = strdup(metric_group), |
| .metric_name = metric_name, |
| }; |
| nd = rblist__find(groups, &me); |
| if (nd) { |
| free(me.metric_group); |
| return container_of(nd, struct mep, nd); |
| } |
| rblist__add_node(groups, &me); |
| nd = rblist__find(groups, &me); |
| if (nd) |
| return container_of(nd, struct mep, nd); |
| return NULL; |
| } |
| |
| static int metricgroup__add_to_mep_groups(const struct pmu_metric *pm, |
| struct rblist *groups) |
| { |
| const char *g; |
| char *omg, *mg; |
| |
| mg = strdup(pm->metric_group ?: pm->metric_name); |
| if (!mg) |
| return -ENOMEM; |
| omg = mg; |
| while ((g = strsep(&mg, ";")) != NULL) { |
| struct mep *me; |
| |
| g = skip_spaces(g); |
| if (strlen(g)) |
| me = mep_lookup(groups, g, pm->metric_name); |
| else |
| me = mep_lookup(groups, pm->metric_name, pm->metric_name); |
| |
| if (me) { |
| me->metric_desc = pm->desc; |
| me->metric_long_desc = pm->long_desc; |
| me->metric_expr = pm->metric_expr; |
| me->metric_threshold = pm->metric_threshold; |
| me->metric_unit = pm->unit; |
| } |
| } |
| free(omg); |
| |
| return 0; |
| } |
| |
| struct metricgroup_iter_data { |
| pmu_metric_iter_fn fn; |
| void *data; |
| }; |
| |
| static int metricgroup__sys_event_iter(const struct pmu_metric *pm, |
| const struct pmu_metrics_table *table, |
| void *data) |
| { |
| struct metricgroup_iter_data *d = data; |
| struct perf_pmu *pmu = NULL; |
| |
| if (!pm->metric_expr || !pm->compat) |
| return 0; |
| |
| while ((pmu = perf_pmus__scan(pmu))) { |
| |
| if (!pmu->id || !pmu_uncore_identifier_match(pm->compat, pmu->id)) |
| continue; |
| |
| return d->fn(pm, table, d->data); |
| } |
| return 0; |
| } |
| |
| static int metricgroup__add_to_mep_groups_callback(const struct pmu_metric *pm, |
| const struct pmu_metrics_table *table __maybe_unused, |
| void *vdata) |
| { |
| struct rblist *groups = vdata; |
| |
| return metricgroup__add_to_mep_groups(pm, groups); |
| } |
| |
| void metricgroup__print(const struct print_callbacks *print_cb, void *print_state) |
| { |
| struct rblist groups; |
| const struct pmu_metrics_table *table; |
| struct rb_node *node, *next; |
| |
| rblist__init(&groups); |
| groups.node_new = mep_new; |
| groups.node_cmp = mep_cmp; |
| groups.node_delete = mep_delete; |
| table = pmu_metrics_table__find(); |
| if (table) { |
| pmu_metrics_table__for_each_metric(table, |
| metricgroup__add_to_mep_groups_callback, |
| &groups); |
| } |
| { |
| struct metricgroup_iter_data data = { |
| .fn = metricgroup__add_to_mep_groups_callback, |
| .data = &groups, |
| }; |
| pmu_for_each_sys_metric(metricgroup__sys_event_iter, &data); |
| } |
| |
| for (node = rb_first_cached(&groups.entries); node; node = next) { |
| struct mep *me = container_of(node, struct mep, nd); |
| |
| print_cb->print_metric(print_state, |
| me->metric_group, |
| me->metric_name, |
| me->metric_desc, |
| me->metric_long_desc, |
| me->metric_expr, |
| me->metric_threshold, |
| me->metric_unit); |
| next = rb_next(node); |
| rblist__remove_node(&groups, node); |
| } |
| } |
| |
| static const char *code_characters = ",-=@"; |
| |
| static int encode_metric_id(struct strbuf *sb, const char *x) |
| { |
| char *c; |
| int ret = 0; |
| |
| for (; *x; x++) { |
| c = strchr(code_characters, *x); |
| if (c) { |
| ret = strbuf_addch(sb, '!'); |
| if (ret) |
| break; |
| |
| ret = strbuf_addch(sb, '0' + (c - code_characters)); |
| if (ret) |
| break; |
| } else { |
| ret = strbuf_addch(sb, *x); |
| if (ret) |
| break; |
| } |
| } |
| return ret; |
| } |
| |
| static int decode_metric_id(struct strbuf *sb, const char *x) |
| { |
| const char *orig = x; |
| size_t i; |
| char c; |
| int ret; |
| |
| for (; *x; x++) { |
| c = *x; |
| if (*x == '!') { |
| x++; |
| i = *x - '0'; |
| if (i > strlen(code_characters)) { |
| pr_err("Bad metric-id encoding in: '%s'", orig); |
| return -1; |
| } |
| c = code_characters[i]; |
| } |
| ret = strbuf_addch(sb, c); |
| if (ret) |
| return ret; |
| } |
| return 0; |
| } |
| |
| static int decode_all_metric_ids(struct evlist *perf_evlist, const char *modifier) |
| { |
| struct evsel *ev; |
| struct strbuf sb = STRBUF_INIT; |
| char *cur; |
| int ret = 0; |
| |
| evlist__for_each_entry(perf_evlist, ev) { |
| if (!ev->metric_id) |
| continue; |
| |
| ret = strbuf_setlen(&sb, 0); |
| if (ret) |
| break; |
| |
| ret = decode_metric_id(&sb, ev->metric_id); |
| if (ret) |
| break; |
| |
| free((char *)ev->metric_id); |
| ev->metric_id = strdup(sb.buf); |
| if (!ev->metric_id) { |
| ret = -ENOMEM; |
| break; |
| } |
| /* |
| * If the name is just the parsed event, use the metric-id to |
| * give a more friendly display version. |
| */ |
| if (strstr(ev->name, "metric-id=")) { |
| bool has_slash = false; |
| |
| zfree(&ev->name); |
| for (cur = strchr(sb.buf, '@') ; cur; cur = strchr(++cur, '@')) { |
| *cur = '/'; |
| has_slash = true; |
| } |
| |
| if (modifier) { |
| if (!has_slash && !strchr(sb.buf, ':')) { |
| ret = strbuf_addch(&sb, ':'); |
| if (ret) |
| break; |
| } |
| ret = strbuf_addstr(&sb, modifier); |
| if (ret) |
| break; |
| } |
| ev->name = strdup(sb.buf); |
| if (!ev->name) { |
| ret = -ENOMEM; |
| break; |
| } |
| } |
| } |
| strbuf_release(&sb); |
| return ret; |
| } |
| |
| static int metricgroup__build_event_string(struct strbuf *events, |
| const struct expr_parse_ctx *ctx, |
| const char *modifier, |
| bool group_events) |
| { |
| struct hashmap_entry *cur; |
| size_t bkt; |
| bool no_group = true, has_tool_events = false; |
| bool tool_events[PERF_TOOL_MAX] = {false}; |
| int ret = 0; |
| |
| #define RETURN_IF_NON_ZERO(x) do { if (x) return x; } while (0) |
| |
| hashmap__for_each_entry(ctx->ids, cur, bkt) { |
| const char *sep, *rsep, *id = cur->pkey; |
| enum perf_tool_event ev; |
| |
| pr_debug("found event %s\n", id); |
| |
| /* Always move tool events outside of the group. */ |
| ev = perf_tool_event__from_str(id); |
| if (ev != PERF_TOOL_NONE) { |
| has_tool_events = true; |
| tool_events[ev] = true; |
| continue; |
| } |
| /* Separate events with commas and open the group if necessary. */ |
| if (no_group) { |
| if (group_events) { |
| ret = strbuf_addch(events, '{'); |
| RETURN_IF_NON_ZERO(ret); |
| } |
| |
| no_group = false; |
| } else { |
| ret = strbuf_addch(events, ','); |
| RETURN_IF_NON_ZERO(ret); |
| } |
| /* |
| * Encode the ID as an event string. Add a qualifier for |
| * metric_id that is the original name except with characters |
| * that parse-events can't parse replaced. For example, |
| * 'msr@tsc@' gets added as msr/tsc,metric-id=msr!3tsc!3/ |
| */ |
| sep = strchr(id, '@'); |
| if (sep != NULL) { |
| ret = strbuf_add(events, id, sep - id); |
| RETURN_IF_NON_ZERO(ret); |
| ret = strbuf_addch(events, '/'); |
| RETURN_IF_NON_ZERO(ret); |
| rsep = strrchr(sep, '@'); |
| ret = strbuf_add(events, sep + 1, rsep - sep - 1); |
| RETURN_IF_NON_ZERO(ret); |
| ret = strbuf_addstr(events, ",metric-id="); |
| RETURN_IF_NON_ZERO(ret); |
| sep = rsep; |
| } else { |
| sep = strchr(id, ':'); |
| if (sep != NULL) { |
| ret = strbuf_add(events, id, sep - id); |
| RETURN_IF_NON_ZERO(ret); |
| } else { |
| ret = strbuf_addstr(events, id); |
| RETURN_IF_NON_ZERO(ret); |
| } |
| ret = strbuf_addstr(events, "/metric-id="); |
| RETURN_IF_NON_ZERO(ret); |
| } |
| ret = encode_metric_id(events, id); |
| RETURN_IF_NON_ZERO(ret); |
| ret = strbuf_addstr(events, "/"); |
| RETURN_IF_NON_ZERO(ret); |
| |
| if (sep != NULL) { |
| ret = strbuf_addstr(events, sep + 1); |
| RETURN_IF_NON_ZERO(ret); |
| } |
| if (modifier) { |
| ret = strbuf_addstr(events, modifier); |
| RETURN_IF_NON_ZERO(ret); |
| } |
| } |
| if (!no_group && group_events) { |
| ret = strbuf_addf(events, "}:W"); |
| RETURN_IF_NON_ZERO(ret); |
| } |
| if (has_tool_events) { |
| int i; |
| |
| perf_tool_event__for_each_event(i) { |
| if (tool_events[i]) { |
| if (!no_group) { |
| ret = strbuf_addch(events, ','); |
| RETURN_IF_NON_ZERO(ret); |
| } |
| no_group = false; |
| ret = strbuf_addstr(events, perf_tool_event__to_str(i)); |
| RETURN_IF_NON_ZERO(ret); |
| } |
| } |
| } |
| |
| return ret; |
| #undef RETURN_IF_NON_ZERO |
| } |
| |
| int __weak arch_get_runtimeparam(const struct pmu_metric *pm __maybe_unused) |
| { |
| return 1; |
| } |
| |
| /* |
| * A singly linked list on the stack of the names of metrics being |
| * processed. Used to identify recursion. |
| */ |
| struct visited_metric { |
| const char *name; |
| const struct visited_metric *parent; |
| }; |
| |
| struct metricgroup_add_iter_data { |
| struct list_head *metric_list; |
| const char *pmu; |
| const char *metric_name; |
| const char *modifier; |
| int *ret; |
| bool *has_match; |
| bool metric_no_group; |
| bool metric_no_threshold; |
| const char *user_requested_cpu_list; |
| bool system_wide; |
| struct metric *root_metric; |
| const struct visited_metric *visited; |
| const struct pmu_metrics_table *table; |
| }; |
| |
| static bool metricgroup__find_metric(const char *pmu, |
| const char *metric, |
| const struct pmu_metrics_table *table, |
| struct pmu_metric *pm); |
| |
| static int add_metric(struct list_head *metric_list, |
| const struct pmu_metric *pm, |
| const char *modifier, |
| bool metric_no_group, |
| bool metric_no_threshold, |
| const char *user_requested_cpu_list, |
| bool system_wide, |
| struct metric *root_metric, |
| const struct visited_metric *visited, |
| const struct pmu_metrics_table *table); |
| |
| /** |
| * resolve_metric - Locate metrics within the root metric and recursively add |
| * references to them. |
| * @metric_list: The list the metric is added to. |
| * @pmu: The PMU name to resolve metrics on, or "all" for all PMUs. |
| * @modifier: if non-null event modifiers like "u". |
| * @metric_no_group: Should events written to events be grouped "{}" or |
| * global. Grouping is the default but due to multiplexing the |
| * user may override. |
| * @user_requested_cpu_list: Command line specified CPUs to record on. |
| * @system_wide: Are events for all processes recorded. |
| * @root_metric: Metrics may reference other metrics to form a tree. In this |
| * case the root_metric holds all the IDs and a list of referenced |
| * metrics. When adding a root this argument is NULL. |
| * @visited: A singly linked list of metric names being added that is used to |
| * detect recursion. |
| * @table: The table that is searched for metrics, most commonly the table for the |
| * architecture perf is running upon. |
| */ |
| static int resolve_metric(struct list_head *metric_list, |
| const char *pmu, |
| const char *modifier, |
| bool metric_no_group, |
| bool metric_no_threshold, |
| const char *user_requested_cpu_list, |
| bool system_wide, |
| struct metric *root_metric, |
| const struct visited_metric *visited, |
| const struct pmu_metrics_table *table) |
| { |
| struct hashmap_entry *cur; |
| size_t bkt; |
| struct to_resolve { |
| /* The metric to resolve. */ |
| struct pmu_metric pm; |
| /* |
| * The key in the IDs map, this may differ from in case, |
| * etc. from pm->metric_name. |
| */ |
| const char *key; |
| } *pending = NULL; |
| int i, ret = 0, pending_cnt = 0; |
| |
| /* |
| * Iterate all the parsed IDs and if there's a matching metric and it to |
| * the pending array. |
| */ |
| hashmap__for_each_entry(root_metric->pctx->ids, cur, bkt) { |
| struct pmu_metric pm; |
| |
| if (metricgroup__find_metric(pmu, cur->pkey, table, &pm)) { |
| pending = realloc(pending, |
| (pending_cnt + 1) * sizeof(struct to_resolve)); |
| if (!pending) |
| return -ENOMEM; |
| |
| memcpy(&pending[pending_cnt].pm, &pm, sizeof(pm)); |
| pending[pending_cnt].key = cur->pkey; |
| pending_cnt++; |
| } |
| } |
| |
| /* Remove the metric IDs from the context. */ |
| for (i = 0; i < pending_cnt; i++) |
| expr__del_id(root_metric->pctx, pending[i].key); |
| |
| /* |
| * Recursively add all the metrics, IDs are added to the root metric's |
| * context. |
| */ |
| for (i = 0; i < pending_cnt; i++) { |
| ret = add_metric(metric_list, &pending[i].pm, modifier, metric_no_group, |
| metric_no_threshold, user_requested_cpu_list, system_wide, |
| root_metric, visited, table); |
| if (ret) |
| break; |
| } |
| |
| free(pending); |
| return ret; |
| } |
| |
| /** |
| * __add_metric - Add a metric to metric_list. |
| * @metric_list: The list the metric is added to. |
| * @pm: The pmu_metric containing the metric to be added. |
| * @modifier: if non-null event modifiers like "u". |
| * @metric_no_group: Should events written to events be grouped "{}" or |
| * global. Grouping is the default but due to multiplexing the |
| * user may override. |
| * @metric_no_threshold: Should threshold expressions be ignored? |
| * @runtime: A special argument for the parser only known at runtime. |
| * @user_requested_cpu_list: Command line specified CPUs to record on. |
| * @system_wide: Are events for all processes recorded. |
| * @root_metric: Metrics may reference other metrics to form a tree. In this |
| * case the root_metric holds all the IDs and a list of referenced |
| * metrics. When adding a root this argument is NULL. |
| * @visited: A singly linked list of metric names being added that is used to |
| * detect recursion. |
| * @table: The table that is searched for metrics, most commonly the table for the |
| * architecture perf is running upon. |
| */ |
| static int __add_metric(struct list_head *metric_list, |
| const struct pmu_metric *pm, |
| const char *modifier, |
| bool metric_no_group, |
| bool metric_no_threshold, |
| int runtime, |
| const char *user_requested_cpu_list, |
| bool system_wide, |
| struct metric *root_metric, |
| const struct visited_metric *visited, |
| const struct pmu_metrics_table *table) |
| { |
| const struct visited_metric *vm; |
| int ret; |
| bool is_root = !root_metric; |
| const char *expr; |
| struct visited_metric visited_node = { |
| .name = pm->metric_name, |
| .parent = visited, |
| }; |
| |
| for (vm = visited; vm; vm = vm->parent) { |
| if (!strcmp(pm->metric_name, vm->name)) { |
| pr_err("failed: recursion detected for %s\n", pm->metric_name); |
| return -1; |
| } |
| } |
| |
| if (is_root) { |
| /* |
| * This metric is the root of a tree and may reference other |
| * metrics that are added recursively. |
| */ |
| root_metric = metric__new(pm, modifier, metric_no_group, runtime, |
| user_requested_cpu_list, system_wide); |
| if (!root_metric) |
| return -ENOMEM; |
| |
| } else { |
| int cnt = 0; |
| |
| /* |
| * This metric was referenced in a metric higher in the |
| * tree. Check if the same metric is already resolved in the |
| * metric_refs list. |
| */ |
| if (root_metric->metric_refs) { |
| for (; root_metric->metric_refs[cnt].metric_name; cnt++) { |
| if (!strcmp(pm->metric_name, |
| root_metric->metric_refs[cnt].metric_name)) |
| return 0; |
| } |
| } |
| |
| /* Create reference. Need space for the entry and the terminator. */ |
| root_metric->metric_refs = realloc(root_metric->metric_refs, |
| (cnt + 2) * sizeof(struct metric_ref)); |
| if (!root_metric->metric_refs) |
| return -ENOMEM; |
| |
| /* |
| * Intentionally passing just const char pointers, |
| * from 'pe' object, so they never go away. We don't |
| * need to change them, so there's no need to create |
| * our own copy. |
| */ |
| root_metric->metric_refs[cnt].metric_name = pm->metric_name; |
| root_metric->metric_refs[cnt].metric_expr = pm->metric_expr; |
| |
| /* Null terminate array. */ |
| root_metric->metric_refs[cnt+1].metric_name = NULL; |
| root_metric->metric_refs[cnt+1].metric_expr = NULL; |
| } |
| |
| /* |
| * For both the parent and referenced metrics, we parse |
| * all the metric's IDs and add it to the root context. |
| */ |
| ret = 0; |
| expr = pm->metric_expr; |
| if (is_root && pm->metric_threshold) { |
| /* |
| * Threshold expressions are built off the actual metric. Switch |
| * to use that in case of additional necessary events. Change |
| * the visited node name to avoid this being flagged as |
| * recursion. If the threshold events are disabled, just use the |
| * metric's name as a reference. This allows metric threshold |
| * computation if there are sufficient events. |
| */ |
| assert(strstr(pm->metric_threshold, pm->metric_name)); |
| expr = metric_no_threshold ? pm->metric_name : pm->metric_threshold; |
| visited_node.name = "__threshold__"; |
| } |
| if (expr__find_ids(expr, NULL, root_metric->pctx) < 0) { |
| /* Broken metric. */ |
| ret = -EINVAL; |
| } |
| if (!ret) { |
| /* Resolve referenced metrics. */ |
| const char *pmu = pm->pmu ?: "cpu"; |
| |
| ret = resolve_metric(metric_list, pmu, modifier, metric_no_group, |
| metric_no_threshold, user_requested_cpu_list, |
| system_wide, root_metric, &visited_node, |
| table); |
| } |
| if (ret) { |
| if (is_root) |
| metric__free(root_metric); |
| |
| } else if (is_root) |
| list_add(&root_metric->nd, metric_list); |
| |
| return ret; |
| } |
| |
| struct metricgroup__find_metric_data { |
| const char *pmu; |
| const char *metric; |
| struct pmu_metric *pm; |
| }; |
| |
| static int metricgroup__find_metric_callback(const struct pmu_metric *pm, |
| const struct pmu_metrics_table *table __maybe_unused, |
| void *vdata) |
| { |
| struct metricgroup__find_metric_data *data = vdata; |
| const char *pm_pmu = pm->pmu ?: "cpu"; |
| |
| if (strcmp(data->pmu, "all") && strcmp(pm_pmu, data->pmu)) |
| return 0; |
| |
| if (!match_metric(pm->metric_name, data->metric)) |
| return 0; |
| |
| memcpy(data->pm, pm, sizeof(*pm)); |
| return 1; |
| } |
| |
| static bool metricgroup__find_metric(const char *pmu, |
| const char *metric, |
| const struct pmu_metrics_table *table, |
| struct pmu_metric *pm) |
| { |
| struct metricgroup__find_metric_data data = { |
| .pmu = pmu, |
| .metric = metric, |
| .pm = pm, |
| }; |
| |
| return pmu_metrics_table__for_each_metric(table, metricgroup__find_metric_callback, &data) |
| ? true : false; |
| } |
| |
| static int add_metric(struct list_head *metric_list, |
| const struct pmu_metric *pm, |
| const char *modifier, |
| bool metric_no_group, |
| bool metric_no_threshold, |
| const char *user_requested_cpu_list, |
| bool system_wide, |
| struct metric *root_metric, |
| const struct visited_metric *visited, |
| const struct pmu_metrics_table *table) |
| { |
| int ret = 0; |
| |
| pr_debug("metric expr %s for %s\n", pm->metric_expr, pm->metric_name); |
| |
| if (!strstr(pm->metric_expr, "?")) { |
| ret = __add_metric(metric_list, pm, modifier, metric_no_group, |
| metric_no_threshold, 0, user_requested_cpu_list, |
| system_wide, root_metric, visited, table); |
| } else { |
| int j, count; |
| |
| count = arch_get_runtimeparam(pm); |
| |
| /* This loop is added to create multiple |
| * events depend on count value and add |
| * those events to metric_list. |
| */ |
| |
| for (j = 0; j < count && !ret; j++) |
| ret = __add_metric(metric_list, pm, modifier, metric_no_group, |
| metric_no_threshold, j, user_requested_cpu_list, |
| system_wide, root_metric, visited, table); |
| } |
| |
| return ret; |
| } |
| |
| static int metricgroup__add_metric_sys_event_iter(const struct pmu_metric *pm, |
| const struct pmu_metrics_table *table __maybe_unused, |
| void *data) |
| { |
| struct metricgroup_add_iter_data *d = data; |
| int ret; |
| |
| if (!match_pm_metric(pm, d->pmu, d->metric_name)) |
| return 0; |
| |
| ret = add_metric(d->metric_list, pm, d->modifier, d->metric_no_group, |
| d->metric_no_threshold, d->user_requested_cpu_list, |
| d->system_wide, d->root_metric, d->visited, d->table); |
| if (ret) |
| goto out; |
| |
| *(d->has_match) = true; |
| |
| out: |
| *(d->ret) = ret; |
| return ret; |
| } |
| |
| /** |
| * metric_list_cmp - list_sort comparator that sorts metrics with more events to |
| * the front. tool events are excluded from the count. |
| */ |
| static int metric_list_cmp(void *priv __maybe_unused, const struct list_head *l, |
| const struct list_head *r) |
| { |
| const struct metric *left = container_of(l, struct metric, nd); |
| const struct metric *right = container_of(r, struct metric, nd); |
| struct expr_id_data *data; |
| int i, left_count, right_count; |
| |
| left_count = hashmap__size(left->pctx->ids); |
| perf_tool_event__for_each_event(i) { |
| if (!expr__get_id(left->pctx, perf_tool_event__to_str(i), &data)) |
| left_count--; |
| } |
| |
| right_count = hashmap__size(right->pctx->ids); |
| perf_tool_event__for_each_event(i) { |
| if (!expr__get_id(right->pctx, perf_tool_event__to_str(i), &data)) |
| right_count--; |
| } |
| |
| return right_count - left_count; |
| } |
| |
| /** |
| * default_metricgroup_cmp - Implements complex key for the Default metricgroup |
| * that first sorts by default_metricgroup_name, then |
| * metric_name. |
| */ |
| static int default_metricgroup_cmp(void *priv __maybe_unused, |
| const struct list_head *l, |
| const struct list_head *r) |
| { |
| const struct metric *left = container_of(l, struct metric, nd); |
| const struct metric *right = container_of(r, struct metric, nd); |
| int diff = strcmp(right->default_metricgroup_name, left->default_metricgroup_name); |
| |
| if (diff) |
| return diff; |
| |
| return strcmp(right->metric_name, left->metric_name); |
| } |
| |
| struct metricgroup__add_metric_data { |
| struct list_head *list; |
| const char *pmu; |
| const char *metric_name; |
| const char *modifier; |
| const char *user_requested_cpu_list; |
| bool metric_no_group; |
| bool metric_no_threshold; |
| bool system_wide; |
| bool has_match; |
| }; |
| |
| static int metricgroup__add_metric_callback(const struct pmu_metric *pm, |
| const struct pmu_metrics_table *table, |
| void *vdata) |
| { |
| struct metricgroup__add_metric_data *data = vdata; |
| int ret = 0; |
| |
| if (pm->metric_expr && match_pm_metric(pm, data->pmu, data->metric_name)) { |
| bool metric_no_group = data->metric_no_group || |
| match_metric(pm->metricgroup_no_group, data->metric_name); |
| |
| data->has_match = true; |
| ret = add_metric(data->list, pm, data->modifier, metric_no_group, |
| data->metric_no_threshold, data->user_requested_cpu_list, |
| data->system_wide, /*root_metric=*/NULL, |
| /*visited_metrics=*/NULL, table); |
| } |
| return ret; |
| } |
| |
| /** |
| * metricgroup__add_metric - Find and add a metric, or a metric group. |
| * @pmu: The PMU name to search for metrics on, or "all" for all PMUs. |
| * @metric_name: The name of the metric or metric group. For example, "IPC" |
| * could be the name of a metric and "TopDownL1" the name of a |
| * metric group. |
| * @modifier: if non-null event modifiers like "u". |
| * @metric_no_group: Should events written to events be grouped "{}" or |
| * global. Grouping is the default but due to multiplexing the |
| * user may override. |
| * @user_requested_cpu_list: Command line specified CPUs to record on. |
| * @system_wide: Are events for all processes recorded. |
| * @metric_list: The list that the metric or metric group are added to. |
| * @table: The table that is searched for metrics, most commonly the table for the |
| * architecture perf is running upon. |
| */ |
| static int metricgroup__add_metric(const char *pmu, const char *metric_name, const char *modifier, |
| bool metric_no_group, bool metric_no_threshold, |
| const char *user_requested_cpu_list, |
| bool system_wide, |
| struct list_head *metric_list, |
| const struct pmu_metrics_table *table) |
| { |
| LIST_HEAD(list); |
| int ret; |
| bool has_match = false; |
| |
| { |
| struct metricgroup__add_metric_data data = { |
| .list = &list, |
| .pmu = pmu, |
| .metric_name = metric_name, |
| .modifier = modifier, |
| .metric_no_group = metric_no_group, |
| .metric_no_threshold = metric_no_threshold, |
| .user_requested_cpu_list = user_requested_cpu_list, |
| .system_wide = system_wide, |
| .has_match = false, |
| }; |
| /* |
| * Iterate over all metrics seeing if metric matches either the |
| * name or group. When it does add the metric to the list. |
| */ |
| ret = pmu_metrics_table__for_each_metric(table, metricgroup__add_metric_callback, |
| &data); |
| if (ret) |
| goto out; |
| |
| has_match = data.has_match; |
| } |
| { |
| struct metricgroup_iter_data data = { |
| .fn = metricgroup__add_metric_sys_event_iter, |
| .data = (void *) &(struct metricgroup_add_iter_data) { |
| .metric_list = &list, |
| .pmu = pmu, |
| .metric_name = metric_name, |
| .modifier = modifier, |
| .metric_no_group = metric_no_group, |
| .user_requested_cpu_list = user_requested_cpu_list, |
| .system_wide = system_wide, |
| .has_match = &has_match, |
| .ret = &ret, |
| .table = table, |
| }, |
| }; |
| |
| pmu_for_each_sys_metric(metricgroup__sys_event_iter, &data); |
| } |
| /* End of pmu events. */ |
| if (!has_match) |
| ret = -EINVAL; |
| |
| out: |
| /* |
| * add to metric_list so that they can be released |
| * even if it's failed |
| */ |
| list_splice(&list, metric_list); |
| return ret; |
| } |
| |
| /** |
| * metricgroup__add_metric_list - Find and add metrics, or metric groups, |
| * specified in a list. |
| * @pmu: A pmu to restrict the metrics to, or "all" for all PMUS. |
| * @list: the list of metrics or metric groups. For example, "IPC,CPI,TopDownL1" |
| * would match the IPC and CPI metrics, and TopDownL1 would match all |
| * the metrics in the TopDownL1 group. |
| * @metric_no_group: Should events written to events be grouped "{}" or |
| * global. Grouping is the default but due to multiplexing the |
| * user may override. |
| * @user_requested_cpu_list: Command line specified CPUs to record on. |
| * @system_wide: Are events for all processes recorded. |
| * @metric_list: The list that metrics are added to. |
| * @table: The table that is searched for metrics, most commonly the table for the |
| * architecture perf is running upon. |
| */ |
| static int metricgroup__add_metric_list(const char *pmu, const char *list, |
| bool metric_no_group, |
| bool metric_no_threshold, |
| const char *user_requested_cpu_list, |
| bool system_wide, struct list_head *metric_list, |
| const struct pmu_metrics_table *table) |
| { |
| char *list_itr, *list_copy, *metric_name, *modifier; |
| int ret, count = 0; |
| |
| list_copy = strdup(list); |
| if (!list_copy) |
| return -ENOMEM; |
| list_itr = list_copy; |
| |
| while ((metric_name = strsep(&list_itr, ",")) != NULL) { |
| modifier = strchr(metric_name, ':'); |
| if (modifier) |
| *modifier++ = '\0'; |
| |
| ret = metricgroup__add_metric(pmu, metric_name, modifier, |
| metric_no_group, metric_no_threshold, |
| user_requested_cpu_list, |
| system_wide, metric_list, table); |
| if (ret == -EINVAL) |
| pr_err("Cannot find metric or group `%s'\n", metric_name); |
| |
| if (ret) |
| break; |
| |
| count++; |
| } |
| free(list_copy); |
| |
| if (!ret) { |
| /* |
| * Warn about nmi_watchdog if any parsed metrics had the |
| * NO_NMI_WATCHDOG constraint. |
| */ |
| metric__watchdog_constraint_hint(NULL, /*foot=*/true); |
| /* No metrics. */ |
| if (count == 0) |
| return -EINVAL; |
| } |
| return ret; |
| } |
| |
| static void metricgroup__free_metrics(struct list_head *metric_list) |
| { |
| struct metric *m, *tmp; |
| |
| list_for_each_entry_safe (m, tmp, metric_list, nd) { |
| list_del_init(&m->nd); |
| metric__free(m); |
| } |
| } |
| |
| /** |
| * find_tool_events - Search for the pressence of tool events in metric_list. |
| * @metric_list: List to take metrics from. |
| * @tool_events: Array of false values, indices corresponding to tool events set |
| * to true if tool event is found. |
| */ |
| static void find_tool_events(const struct list_head *metric_list, |
| bool tool_events[PERF_TOOL_MAX]) |
| { |
| struct metric *m; |
| |
| list_for_each_entry(m, metric_list, nd) { |
| int i; |
| |
| perf_tool_event__for_each_event(i) { |
| struct expr_id_data *data; |
| |
| if (!tool_events[i] && |
| !expr__get_id(m->pctx, perf_tool_event__to_str(i), &data)) |
| tool_events[i] = true; |
| } |
| } |
| } |
| |
| /** |
| * build_combined_expr_ctx - Make an expr_parse_ctx with all !group_events |
| * metric IDs, as the IDs are held in a set, |
| * duplicates will be removed. |
| * @metric_list: List to take metrics from. |
| * @combined: Out argument for result. |
| */ |
| static int build_combined_expr_ctx(const struct list_head *metric_list, |
| struct expr_parse_ctx **combined) |
| { |
| struct hashmap_entry *cur; |
| size_t bkt; |
| struct metric *m; |
| char *dup; |
| int ret; |
| |
| *combined = expr__ctx_new(); |
| if (!*combined) |
| return -ENOMEM; |
| |
| list_for_each_entry(m, metric_list, nd) { |
| if (!m->group_events && !m->modifier) { |
| hashmap__for_each_entry(m->pctx->ids, cur, bkt) { |
| dup = strdup(cur->pkey); |
| if (!dup) { |
| ret = -ENOMEM; |
| goto err_out; |
| } |
| ret = expr__add_id(*combined, dup); |
| if (ret) |
| goto err_out; |
| } |
| } |
| } |
| return 0; |
| err_out: |
| expr__ctx_free(*combined); |
| *combined = NULL; |
| return ret; |
| } |
| |
| /** |
| * parse_ids - Build the event string for the ids and parse them creating an |
| * evlist. The encoded metric_ids are decoded. |
| * @metric_no_merge: is metric sharing explicitly disabled. |
| * @fake_pmu: use a fake PMU when testing metrics not supported by the current CPU. |
| * @ids: the event identifiers parsed from a metric. |
| * @modifier: any modifiers added to the events. |
| * @group_events: should events be placed in a weak group. |
| * @tool_events: entries set true if the tool event of index could be present in |
| * the overall list of metrics. |
| * @out_evlist: the created list of events. |
| */ |
| static int parse_ids(bool metric_no_merge, bool fake_pmu, |
| struct expr_parse_ctx *ids, const char *modifier, |
| bool group_events, const bool tool_events[PERF_TOOL_MAX], |
| struct evlist **out_evlist) |
| { |
| struct parse_events_error parse_error; |
| struct evlist *parsed_evlist; |
| struct strbuf events = STRBUF_INIT; |
| int ret; |
| |
| *out_evlist = NULL; |
| if (!metric_no_merge || hashmap__size(ids->ids) == 0) { |
| bool added_event = false; |
| int i; |
| /* |
| * We may fail to share events between metrics because a tool |
| * event isn't present in one metric. For example, a ratio of |
| * cache misses doesn't need duration_time but the same events |
| * may be used for a misses per second. Events without sharing |
| * implies multiplexing, that is best avoided, so place |
| * all tool events in every group. |
| * |
| * Also, there may be no ids/events in the expression parsing |
| * context because of constant evaluation, e.g.: |
| * event1 if #smt_on else 0 |
| * Add a tool event to avoid a parse error on an empty string. |
| */ |
| perf_tool_event__for_each_event(i) { |
| if (tool_events[i]) { |
| char *tmp = strdup(perf_tool_event__to_str(i)); |
| |
| if (!tmp) |
| return -ENOMEM; |
| ids__insert(ids->ids, tmp); |
| added_event = true; |
| } |
| } |
| if (!added_event && hashmap__size(ids->ids) == 0) { |
| char *tmp = strdup("duration_time"); |
| |
| if (!tmp) |
| return -ENOMEM; |
| ids__insert(ids->ids, tmp); |
| } |
| } |
| ret = metricgroup__build_event_string(&events, ids, modifier, |
| group_events); |
| if (ret) |
| return ret; |
| |
| parsed_evlist = evlist__new(); |
| if (!parsed_evlist) { |
| ret = -ENOMEM; |
| goto err_out; |
| } |
| pr_debug("Parsing metric events '%s'\n", events.buf); |
| parse_events_error__init(&parse_error); |
| ret = __parse_events(parsed_evlist, events.buf, /*pmu_filter=*/NULL, |
| &parse_error, fake_pmu, /*warn_if_reordered=*/false, |
| /*fake_tp=*/false); |
| if (ret) { |
| parse_events_error__print(&parse_error, events.buf); |
| goto err_out; |
| } |
| ret = decode_all_metric_ids(parsed_evlist, modifier); |
| if (ret) |
| goto err_out; |
| |
| *out_evlist = parsed_evlist; |
| parsed_evlist = NULL; |
| err_out: |
| parse_events_error__exit(&parse_error); |
| evlist__delete(parsed_evlist); |
| strbuf_release(&events); |
| return ret; |
| } |
| |
| static int parse_groups(struct evlist *perf_evlist, |
| const char *pmu, const char *str, |
| bool metric_no_group, |
| bool metric_no_merge, |
| bool metric_no_threshold, |
| const char *user_requested_cpu_list, |
| bool system_wide, |
| bool fake_pmu, |
| struct rblist *metric_events_list, |
| const struct pmu_metrics_table *table) |
| { |
| struct evlist *combined_evlist = NULL; |
| LIST_HEAD(metric_list); |
| struct metric *m; |
| bool tool_events[PERF_TOOL_MAX] = {false}; |
| bool is_default = !strcmp(str, "Default"); |
| int ret; |
| |
| if (metric_events_list->nr_entries == 0) |
| metricgroup__rblist_init(metric_events_list); |
| ret = metricgroup__add_metric_list(pmu, str, metric_no_group, metric_no_threshold, |
| user_requested_cpu_list, |
| system_wide, &metric_list, table); |
| if (ret) |
| goto out; |
| |
| /* Sort metrics from largest to smallest. */ |
| list_sort(NULL, &metric_list, metric_list_cmp); |
| |
| if (!metric_no_merge) { |
| struct expr_parse_ctx *combined = NULL; |
| |
| find_tool_events(&metric_list, tool_events); |
| |
| ret = build_combined_expr_ctx(&metric_list, &combined); |
| |
| if (!ret && combined && hashmap__size(combined->ids)) { |
| ret = parse_ids(metric_no_merge, fake_pmu, combined, |
| /*modifier=*/NULL, |
| /*group_events=*/false, |
| tool_events, |
| &combined_evlist); |
| } |
| if (combined) |
| expr__ctx_free(combined); |
| |
| if (ret) |
| goto out; |
| } |
| |
| if (is_default) |
| list_sort(NULL, &metric_list, default_metricgroup_cmp); |
| |
| list_for_each_entry(m, &metric_list, nd) { |
| struct metric_event *me; |
| struct evsel **metric_events; |
| struct evlist *metric_evlist = NULL; |
| struct metric *n; |
| struct metric_expr *expr; |
| |
| if (combined_evlist && !m->group_events) { |
| metric_evlist = combined_evlist; |
| } else if (!metric_no_merge) { |
| /* |
| * See if the IDs for this metric are a subset of an |
| * earlier metric. |
| */ |
| list_for_each_entry(n, &metric_list, nd) { |
| if (m == n) |
| break; |
| |
| if (n->evlist == NULL) |
| continue; |
| |
| if ((!m->modifier && n->modifier) || |
| (m->modifier && !n->modifier) || |
| (m->modifier && n->modifier && |
| strcmp(m->modifier, n->modifier))) |
| continue; |
| |
| if ((!m->pmu && n->pmu) || |
| (m->pmu && !n->pmu) || |
| (m->pmu && n->pmu && strcmp(m->pmu, n->pmu))) |
| continue; |
| |
| if (expr__subset_of_ids(n->pctx, m->pctx)) { |
| pr_debug("Events in '%s' fully contained within '%s'\n", |
| m->metric_name, n->metric_name); |
| metric_evlist = n->evlist; |
| break; |
| } |
| |
| } |
| } |
| if (!metric_evlist) { |
| ret = parse_ids(metric_no_merge, fake_pmu, m->pctx, m->modifier, |
| m->group_events, tool_events, &m->evlist); |
| if (ret) |
| goto out; |
| |
| metric_evlist = m->evlist; |
| } |
| ret = setup_metric_events(fake_pmu ? "all" : m->pmu, m->pctx->ids, |
| metric_evlist, &metric_events); |
| if (ret) { |
| pr_err("Cannot resolve IDs for %s: %s\n", |
| m->metric_name, m->metric_expr); |
| goto out; |
| } |
| |
| me = metricgroup__lookup(metric_events_list, metric_events[0], true); |
| |
| expr = malloc(sizeof(struct metric_expr)); |
| if (!expr) { |
| ret = -ENOMEM; |
| free(metric_events); |
| goto out; |
| } |
| |
| expr->metric_refs = m->metric_refs; |
| m->metric_refs = NULL; |
| expr->metric_expr = m->metric_expr; |
| if (m->modifier) { |
| char *tmp; |
| |
| if (asprintf(&tmp, "%s:%s", m->metric_name, m->modifier) < 0) |
| expr->metric_name = NULL; |
| else |
| expr->metric_name = tmp; |
| } else |
| expr->metric_name = strdup(m->metric_name); |
| |
| if (!expr->metric_name) { |
| ret = -ENOMEM; |
| free(metric_events); |
| goto out; |
| } |
| expr->metric_threshold = m->metric_threshold; |
| expr->metric_unit = m->metric_unit; |
| expr->metric_events = metric_events; |
| expr->runtime = m->pctx->sctx.runtime; |
| expr->default_metricgroup_name = m->default_metricgroup_name; |
| me->is_default = is_default; |
| list_add(&expr->nd, &me->head); |
| } |
| |
| |
| if (combined_evlist) { |
| evlist__splice_list_tail(perf_evlist, &combined_evlist->core.entries); |
| evlist__delete(combined_evlist); |
| } |
| |
| list_for_each_entry(m, &metric_list, nd) { |
| if (m->evlist) |
| evlist__splice_list_tail(perf_evlist, &m->evlist->core.entries); |
| } |
| |
| out: |
| metricgroup__free_metrics(&metric_list); |
| return ret; |
| } |
| |
| int metricgroup__parse_groups(struct evlist *perf_evlist, |
| const char *pmu, |
| const char *str, |
| bool metric_no_group, |
| bool metric_no_merge, |
| bool metric_no_threshold, |
| const char *user_requested_cpu_list, |
| bool system_wide, |
| bool hardware_aware_grouping, |
| struct rblist *metric_events) |
| { |
| const struct pmu_metrics_table *table = pmu_metrics_table__find(); |
| |
| if (!table) |
| return -EINVAL; |
| if (hardware_aware_grouping) |
| pr_debug("Use hardware aware grouping instead of traditional metric grouping method\n"); |
| |
| return parse_groups(perf_evlist, pmu, str, metric_no_group, metric_no_merge, |
| metric_no_threshold, user_requested_cpu_list, system_wide, |
| /*fake_pmu=*/false, metric_events, table); |
| } |
| |
| int metricgroup__parse_groups_test(struct evlist *evlist, |
| const struct pmu_metrics_table *table, |
| const char *str, |
| struct rblist *metric_events) |
| { |
| return parse_groups(evlist, "all", str, |
| /*metric_no_group=*/false, |
| /*metric_no_merge=*/false, |
| /*metric_no_threshold=*/false, |
| /*user_requested_cpu_list=*/NULL, |
| /*system_wide=*/false, |
| /*fake_pmu=*/true, metric_events, table); |
| } |
| |
| struct metricgroup__has_metric_data { |
| const char *pmu; |
| const char *metric; |
| }; |
| static int metricgroup__has_metric_callback(const struct pmu_metric *pm, |
| const struct pmu_metrics_table *table __maybe_unused, |
| void *vdata) |
| { |
| struct metricgroup__has_metric_data *data = vdata; |
| |
| return match_pm_metric(pm, data->pmu, data->metric) ? 1 : 0; |
| } |
| |
| bool metricgroup__has_metric(const char *pmu, const char *metric) |
| { |
| const struct pmu_metrics_table *table = pmu_metrics_table__find(); |
| struct metricgroup__has_metric_data data = { |
| .pmu = pmu, |
| .metric = metric, |
| }; |
| |
| if (!table) |
| return false; |
| |
| return pmu_metrics_table__for_each_metric(table, metricgroup__has_metric_callback, &data) |
| ? true : false; |
| } |
| |
| static int metricgroup__topdown_max_level_callback(const struct pmu_metric *pm, |
| const struct pmu_metrics_table *table __maybe_unused, |
| void *data) |
| { |
| unsigned int *max_level = data; |
| unsigned int level; |
| const char *p = strstr(pm->metric_group ?: "", "TopdownL"); |
| |
| if (!p || p[8] == '\0') |
| return 0; |
| |
| level = p[8] - '0'; |
| if (level > *max_level) |
| *max_level = level; |
| |
| return 0; |
| } |
| |
| unsigned int metricgroups__topdown_max_level(void) |
| { |
| unsigned int max_level = 0; |
| const struct pmu_metrics_table *table = pmu_metrics_table__find(); |
| |
| if (!table) |
| return false; |
| |
| pmu_metrics_table__for_each_metric(table, metricgroup__topdown_max_level_callback, |
| &max_level); |
| return max_level; |
| } |
| |
| int metricgroup__copy_metric_events(struct evlist *evlist, struct cgroup *cgrp, |
| struct rblist *new_metric_events, |
| struct rblist *old_metric_events) |
| { |
| unsigned int i; |
| |
| for (i = 0; i < rblist__nr_entries(old_metric_events); i++) { |
| struct rb_node *nd; |
| struct metric_event *old_me, *new_me; |
| struct metric_expr *old_expr, *new_expr; |
| struct evsel *evsel; |
| size_t alloc_size; |
| int idx, nr; |
| |
| nd = rblist__entry(old_metric_events, i); |
| old_me = container_of(nd, struct metric_event, nd); |
| |
| evsel = evlist__find_evsel(evlist, old_me->evsel->core.idx); |
| if (!evsel) |
| return -EINVAL; |
| new_me = metricgroup__lookup(new_metric_events, evsel, true); |
| if (!new_me) |
| return -ENOMEM; |
| |
| pr_debug("copying metric event for cgroup '%s': %s (idx=%d)\n", |
| cgrp ? cgrp->name : "root", evsel->name, evsel->core.idx); |
| |
| list_for_each_entry(old_expr, &old_me->head, nd) { |
| new_expr = malloc(sizeof(*new_expr)); |
| if (!new_expr) |
| return -ENOMEM; |
| |
| new_expr->metric_expr = old_expr->metric_expr; |
| new_expr->metric_threshold = old_expr->metric_threshold; |
| new_expr->metric_name = strdup(old_expr->metric_name); |
| if (!new_expr->metric_name) |
| return -ENOMEM; |
| |
| new_expr->metric_unit = old_expr->metric_unit; |
| new_expr->runtime = old_expr->runtime; |
| |
| if (old_expr->metric_refs) { |
| /* calculate number of metric_events */ |
| for (nr = 0; old_expr->metric_refs[nr].metric_name; nr++) |
| continue; |
| alloc_size = sizeof(*new_expr->metric_refs); |
| new_expr->metric_refs = calloc(nr + 1, alloc_size); |
| if (!new_expr->metric_refs) { |
| free(new_expr); |
| return -ENOMEM; |
| } |
| |
| memcpy(new_expr->metric_refs, old_expr->metric_refs, |
| nr * alloc_size); |
| } else { |
| new_expr->metric_refs = NULL; |
| } |
| |
| /* calculate number of metric_events */ |
| for (nr = 0; old_expr->metric_events[nr]; nr++) |
| continue; |
| alloc_size = sizeof(*new_expr->metric_events); |
| new_expr->metric_events = calloc(nr + 1, alloc_size); |
| if (!new_expr->metric_events) { |
| zfree(&new_expr->metric_refs); |
| free(new_expr); |
| return -ENOMEM; |
| } |
| |
| /* copy evsel in the same position */ |
| for (idx = 0; idx < nr; idx++) { |
| evsel = old_expr->metric_events[idx]; |
| evsel = evlist__find_evsel(evlist, evsel->core.idx); |
| if (evsel == NULL) { |
| zfree(&new_expr->metric_events); |
| zfree(&new_expr->metric_refs); |
| free(new_expr); |
| return -EINVAL; |
| } |
| new_expr->metric_events[idx] = evsel; |
| } |
| |
| list_add(&new_expr->nd, &new_me->head); |
| } |
| } |
| return 0; |
| } |