| // SPDX-License-Identifier: GPL-2.0 |
| /* Copyright(c) 2020 Intel Corporation. All rights rsvd. */ |
| |
| #include <linux/sched/task.h> |
| #include <linux/io-64-nonatomic-lo-hi.h> |
| #include "idxd.h" |
| #include "perfmon.h" |
| |
| /* |
| * These attributes specify the bits in the config word that the perf |
| * syscall uses to pass the event ids and categories to perfmon. |
| */ |
| DEFINE_PERFMON_FORMAT_ATTR(event_category, "config:0-3"); |
| DEFINE_PERFMON_FORMAT_ATTR(event, "config:4-31"); |
| |
| /* |
| * These attributes specify the bits in the config1 word that the perf |
| * syscall uses to pass filter data to perfmon. |
| */ |
| DEFINE_PERFMON_FORMAT_ATTR(filter_wq, "config1:0-31"); |
| DEFINE_PERFMON_FORMAT_ATTR(filter_tc, "config1:32-39"); |
| DEFINE_PERFMON_FORMAT_ATTR(filter_pgsz, "config1:40-43"); |
| DEFINE_PERFMON_FORMAT_ATTR(filter_sz, "config1:44-51"); |
| DEFINE_PERFMON_FORMAT_ATTR(filter_eng, "config1:52-59"); |
| |
| #define PERFMON_FILTERS_START 2 |
| #define PERFMON_FILTERS_MAX 5 |
| |
| static struct attribute *perfmon_format_attrs[] = { |
| &format_attr_idxd_event_category.attr, |
| &format_attr_idxd_event.attr, |
| &format_attr_idxd_filter_wq.attr, |
| &format_attr_idxd_filter_tc.attr, |
| &format_attr_idxd_filter_pgsz.attr, |
| &format_attr_idxd_filter_sz.attr, |
| &format_attr_idxd_filter_eng.attr, |
| NULL, |
| }; |
| |
| static struct attribute_group perfmon_format_attr_group = { |
| .name = "format", |
| .attrs = perfmon_format_attrs, |
| }; |
| |
| static const struct attribute_group *perfmon_attr_groups[] = { |
| &perfmon_format_attr_group, |
| NULL, |
| }; |
| |
| static bool is_idxd_event(struct idxd_pmu *idxd_pmu, struct perf_event *event) |
| { |
| return &idxd_pmu->pmu == event->pmu; |
| } |
| |
| static int perfmon_collect_events(struct idxd_pmu *idxd_pmu, |
| struct perf_event *leader, |
| bool do_grp) |
| { |
| struct perf_event *event; |
| int n, max_count; |
| |
| max_count = idxd_pmu->n_counters; |
| n = idxd_pmu->n_events; |
| |
| if (n >= max_count) |
| return -EINVAL; |
| |
| if (is_idxd_event(idxd_pmu, leader)) { |
| idxd_pmu->event_list[n] = leader; |
| idxd_pmu->event_list[n]->hw.idx = n; |
| n++; |
| } |
| |
| if (!do_grp) |
| return n; |
| |
| for_each_sibling_event(event, leader) { |
| if (!is_idxd_event(idxd_pmu, event) || |
| event->state <= PERF_EVENT_STATE_OFF) |
| continue; |
| |
| if (n >= max_count) |
| return -EINVAL; |
| |
| idxd_pmu->event_list[n] = event; |
| idxd_pmu->event_list[n]->hw.idx = n; |
| n++; |
| } |
| |
| return n; |
| } |
| |
| static void perfmon_assign_hw_event(struct idxd_pmu *idxd_pmu, |
| struct perf_event *event, int idx) |
| { |
| struct idxd_device *idxd = idxd_pmu->idxd; |
| struct hw_perf_event *hwc = &event->hw; |
| |
| hwc->idx = idx; |
| hwc->config_base = ioread64(CNTRCFG_REG(idxd, idx)); |
| hwc->event_base = ioread64(CNTRCFG_REG(idxd, idx)); |
| } |
| |
| static int perfmon_assign_event(struct idxd_pmu *idxd_pmu, |
| struct perf_event *event) |
| { |
| int i; |
| |
| for (i = 0; i < IDXD_PMU_EVENT_MAX; i++) |
| if (!test_and_set_bit(i, idxd_pmu->used_mask)) |
| return i; |
| |
| return -EINVAL; |
| } |
| |
| /* |
| * Check whether there are enough counters to satisfy that all the |
| * events in the group can actually be scheduled at the same time. |
| * |
| * To do this, create a fake idxd_pmu object so the event collection |
| * and assignment functions can be used without affecting the internal |
| * state of the real idxd_pmu object. |
| */ |
| static int perfmon_validate_group(struct idxd_pmu *pmu, |
| struct perf_event *event) |
| { |
| struct perf_event *leader = event->group_leader; |
| struct idxd_pmu *fake_pmu; |
| int i, ret = 0, n, idx; |
| |
| fake_pmu = kzalloc(sizeof(*fake_pmu), GFP_KERNEL); |
| if (!fake_pmu) |
| return -ENOMEM; |
| |
| fake_pmu->pmu.name = pmu->pmu.name; |
| fake_pmu->n_counters = pmu->n_counters; |
| |
| n = perfmon_collect_events(fake_pmu, leader, true); |
| if (n < 0) { |
| ret = n; |
| goto out; |
| } |
| |
| fake_pmu->n_events = n; |
| n = perfmon_collect_events(fake_pmu, event, false); |
| if (n < 0) { |
| ret = n; |
| goto out; |
| } |
| |
| fake_pmu->n_events = n; |
| |
| for (i = 0; i < n; i++) { |
| event = fake_pmu->event_list[i]; |
| |
| idx = perfmon_assign_event(fake_pmu, event); |
| if (idx < 0) { |
| ret = idx; |
| goto out; |
| } |
| } |
| out: |
| kfree(fake_pmu); |
| |
| return ret; |
| } |
| |
| static int perfmon_pmu_event_init(struct perf_event *event) |
| { |
| struct idxd_device *idxd; |
| int ret = 0; |
| |
| idxd = event_to_idxd(event); |
| event->hw.idx = -1; |
| |
| if (event->attr.type != event->pmu->type) |
| return -ENOENT; |
| |
| /* sampling not supported */ |
| if (event->attr.sample_period) |
| return -EINVAL; |
| |
| if (event->cpu < 0) |
| return -EINVAL; |
| |
| if (event->pmu != &idxd->idxd_pmu->pmu) |
| return -EINVAL; |
| |
| event->hw.event_base = ioread64(PERFMON_TABLE_OFFSET(idxd)); |
| event->hw.config = event->attr.config; |
| |
| if (event->group_leader != event) |
| /* non-group events have themselves as leader */ |
| ret = perfmon_validate_group(idxd->idxd_pmu, event); |
| |
| return ret; |
| } |
| |
| static inline u64 perfmon_pmu_read_counter(struct perf_event *event) |
| { |
| struct hw_perf_event *hwc = &event->hw; |
| struct idxd_device *idxd; |
| int cntr = hwc->idx; |
| |
| idxd = event_to_idxd(event); |
| |
| return ioread64(CNTRDATA_REG(idxd, cntr)); |
| } |
| |
| static void perfmon_pmu_event_update(struct perf_event *event) |
| { |
| struct idxd_device *idxd = event_to_idxd(event); |
| u64 prev_raw_count, new_raw_count, delta, p, n; |
| int shift = 64 - idxd->idxd_pmu->counter_width; |
| struct hw_perf_event *hwc = &event->hw; |
| |
| prev_raw_count = local64_read(&hwc->prev_count); |
| do { |
| new_raw_count = perfmon_pmu_read_counter(event); |
| } while (!local64_try_cmpxchg(&hwc->prev_count, |
| &prev_raw_count, new_raw_count)); |
| n = (new_raw_count << shift); |
| p = (prev_raw_count << shift); |
| |
| delta = ((n - p) >> shift); |
| |
| local64_add(delta, &event->count); |
| } |
| |
| void perfmon_counter_overflow(struct idxd_device *idxd) |
| { |
| int i, n_counters, max_loop = OVERFLOW_SIZE; |
| struct perf_event *event; |
| unsigned long ovfstatus; |
| |
| n_counters = min(idxd->idxd_pmu->n_counters, OVERFLOW_SIZE); |
| |
| ovfstatus = ioread32(OVFSTATUS_REG(idxd)); |
| |
| /* |
| * While updating overflowed counters, other counters behind |
| * them could overflow and be missed in a given pass. |
| * Normally this could happen at most n_counters times, but in |
| * theory a tiny counter width could result in continual |
| * overflows and endless looping. max_loop provides a |
| * failsafe in that highly unlikely case. |
| */ |
| while (ovfstatus && max_loop--) { |
| /* Figure out which counter(s) overflowed */ |
| for_each_set_bit(i, &ovfstatus, n_counters) { |
| unsigned long ovfstatus_clear = 0; |
| |
| /* Update event->count for overflowed counter */ |
| event = idxd->idxd_pmu->event_list[i]; |
| perfmon_pmu_event_update(event); |
| /* Writing 1 to OVFSTATUS bit clears it */ |
| set_bit(i, &ovfstatus_clear); |
| iowrite32(ovfstatus_clear, OVFSTATUS_REG(idxd)); |
| } |
| |
| ovfstatus = ioread32(OVFSTATUS_REG(idxd)); |
| } |
| |
| /* |
| * Should never happen. If so, it means a counter(s) looped |
| * around twice while this handler was running. |
| */ |
| WARN_ON_ONCE(ovfstatus); |
| } |
| |
| static inline void perfmon_reset_config(struct idxd_device *idxd) |
| { |
| iowrite32(CONFIG_RESET, PERFRST_REG(idxd)); |
| iowrite32(0, OVFSTATUS_REG(idxd)); |
| iowrite32(0, PERFFRZ_REG(idxd)); |
| } |
| |
| static inline void perfmon_reset_counters(struct idxd_device *idxd) |
| { |
| iowrite32(CNTR_RESET, PERFRST_REG(idxd)); |
| } |
| |
| static inline void perfmon_reset(struct idxd_device *idxd) |
| { |
| perfmon_reset_config(idxd); |
| perfmon_reset_counters(idxd); |
| } |
| |
| static void perfmon_pmu_event_start(struct perf_event *event, int mode) |
| { |
| u32 flt_wq, flt_tc, flt_pg_sz, flt_xfer_sz, flt_eng = 0; |
| u64 cntr_cfg, cntrdata, event_enc, event_cat = 0; |
| struct hw_perf_event *hwc = &event->hw; |
| union filter_cfg flt_cfg; |
| union event_cfg event_cfg; |
| struct idxd_device *idxd; |
| int cntr; |
| |
| idxd = event_to_idxd(event); |
| |
| event->hw.idx = hwc->idx; |
| cntr = hwc->idx; |
| |
| /* Obtain event category and event value from user space */ |
| event_cfg.val = event->attr.config; |
| flt_cfg.val = event->attr.config1; |
| event_cat = event_cfg.event_cat; |
| event_enc = event_cfg.event_enc; |
| |
| /* Obtain filter configuration from user space */ |
| flt_wq = flt_cfg.wq; |
| flt_tc = flt_cfg.tc; |
| flt_pg_sz = flt_cfg.pg_sz; |
| flt_xfer_sz = flt_cfg.xfer_sz; |
| flt_eng = flt_cfg.eng; |
| |
| if (flt_wq && test_bit(FLT_WQ, &idxd->idxd_pmu->supported_filters)) |
| iowrite32(flt_wq, FLTCFG_REG(idxd, cntr, FLT_WQ)); |
| if (flt_tc && test_bit(FLT_TC, &idxd->idxd_pmu->supported_filters)) |
| iowrite32(flt_tc, FLTCFG_REG(idxd, cntr, FLT_TC)); |
| if (flt_pg_sz && test_bit(FLT_PG_SZ, &idxd->idxd_pmu->supported_filters)) |
| iowrite32(flt_pg_sz, FLTCFG_REG(idxd, cntr, FLT_PG_SZ)); |
| if (flt_xfer_sz && test_bit(FLT_XFER_SZ, &idxd->idxd_pmu->supported_filters)) |
| iowrite32(flt_xfer_sz, FLTCFG_REG(idxd, cntr, FLT_XFER_SZ)); |
| if (flt_eng && test_bit(FLT_ENG, &idxd->idxd_pmu->supported_filters)) |
| iowrite32(flt_eng, FLTCFG_REG(idxd, cntr, FLT_ENG)); |
| |
| /* Read the start value */ |
| cntrdata = ioread64(CNTRDATA_REG(idxd, cntr)); |
| local64_set(&event->hw.prev_count, cntrdata); |
| |
| /* Set counter to event/category */ |
| cntr_cfg = event_cat << CNTRCFG_CATEGORY_SHIFT; |
| cntr_cfg |= event_enc << CNTRCFG_EVENT_SHIFT; |
| /* Set interrupt on overflow and counter enable bits */ |
| cntr_cfg |= (CNTRCFG_IRQ_OVERFLOW | CNTRCFG_ENABLE); |
| |
| iowrite64(cntr_cfg, CNTRCFG_REG(idxd, cntr)); |
| } |
| |
| static void perfmon_pmu_event_stop(struct perf_event *event, int mode) |
| { |
| struct hw_perf_event *hwc = &event->hw; |
| struct idxd_device *idxd; |
| int i, cntr = hwc->idx; |
| u64 cntr_cfg; |
| |
| idxd = event_to_idxd(event); |
| |
| /* remove this event from event list */ |
| for (i = 0; i < idxd->idxd_pmu->n_events; i++) { |
| if (event != idxd->idxd_pmu->event_list[i]) |
| continue; |
| |
| for (++i; i < idxd->idxd_pmu->n_events; i++) |
| idxd->idxd_pmu->event_list[i - 1] = idxd->idxd_pmu->event_list[i]; |
| --idxd->idxd_pmu->n_events; |
| break; |
| } |
| |
| cntr_cfg = ioread64(CNTRCFG_REG(idxd, cntr)); |
| cntr_cfg &= ~CNTRCFG_ENABLE; |
| iowrite64(cntr_cfg, CNTRCFG_REG(idxd, cntr)); |
| |
| if (mode == PERF_EF_UPDATE) |
| perfmon_pmu_event_update(event); |
| |
| event->hw.idx = -1; |
| clear_bit(cntr, idxd->idxd_pmu->used_mask); |
| } |
| |
| static void perfmon_pmu_event_del(struct perf_event *event, int mode) |
| { |
| perfmon_pmu_event_stop(event, PERF_EF_UPDATE); |
| } |
| |
| static int perfmon_pmu_event_add(struct perf_event *event, int flags) |
| { |
| struct idxd_device *idxd = event_to_idxd(event); |
| struct idxd_pmu *idxd_pmu = idxd->idxd_pmu; |
| struct hw_perf_event *hwc = &event->hw; |
| int idx, n; |
| |
| n = perfmon_collect_events(idxd_pmu, event, false); |
| if (n < 0) |
| return n; |
| |
| hwc->state = PERF_HES_UPTODATE | PERF_HES_STOPPED; |
| if (!(flags & PERF_EF_START)) |
| hwc->state |= PERF_HES_ARCH; |
| |
| idx = perfmon_assign_event(idxd_pmu, event); |
| if (idx < 0) |
| return idx; |
| |
| perfmon_assign_hw_event(idxd_pmu, event, idx); |
| |
| if (flags & PERF_EF_START) |
| perfmon_pmu_event_start(event, 0); |
| |
| idxd_pmu->n_events = n; |
| |
| return 0; |
| } |
| |
| static void enable_perfmon_pmu(struct idxd_device *idxd) |
| { |
| iowrite32(COUNTER_UNFREEZE, PERFFRZ_REG(idxd)); |
| } |
| |
| static void disable_perfmon_pmu(struct idxd_device *idxd) |
| { |
| iowrite32(COUNTER_FREEZE, PERFFRZ_REG(idxd)); |
| } |
| |
| static void perfmon_pmu_enable(struct pmu *pmu) |
| { |
| struct idxd_device *idxd = pmu_to_idxd(pmu); |
| |
| enable_perfmon_pmu(idxd); |
| } |
| |
| static void perfmon_pmu_disable(struct pmu *pmu) |
| { |
| struct idxd_device *idxd = pmu_to_idxd(pmu); |
| |
| disable_perfmon_pmu(idxd); |
| } |
| |
| static void skip_filter(int i) |
| { |
| int j; |
| |
| for (j = i; j < PERFMON_FILTERS_MAX; j++) |
| perfmon_format_attrs[PERFMON_FILTERS_START + j] = |
| perfmon_format_attrs[PERFMON_FILTERS_START + j + 1]; |
| } |
| |
| static void idxd_pmu_init(struct idxd_pmu *idxd_pmu) |
| { |
| int i; |
| |
| for (i = 0 ; i < PERFMON_FILTERS_MAX; i++) { |
| if (!test_bit(i, &idxd_pmu->supported_filters)) |
| skip_filter(i); |
| } |
| |
| idxd_pmu->pmu.name = idxd_pmu->name; |
| idxd_pmu->pmu.attr_groups = perfmon_attr_groups; |
| idxd_pmu->pmu.task_ctx_nr = perf_invalid_context; |
| idxd_pmu->pmu.event_init = perfmon_pmu_event_init; |
| idxd_pmu->pmu.pmu_enable = perfmon_pmu_enable; |
| idxd_pmu->pmu.pmu_disable = perfmon_pmu_disable; |
| idxd_pmu->pmu.add = perfmon_pmu_event_add; |
| idxd_pmu->pmu.del = perfmon_pmu_event_del; |
| idxd_pmu->pmu.start = perfmon_pmu_event_start; |
| idxd_pmu->pmu.stop = perfmon_pmu_event_stop; |
| idxd_pmu->pmu.read = perfmon_pmu_event_update; |
| idxd_pmu->pmu.capabilities = PERF_PMU_CAP_NO_EXCLUDE; |
| idxd_pmu->pmu.scope = PERF_PMU_SCOPE_SYS_WIDE; |
| idxd_pmu->pmu.module = THIS_MODULE; |
| } |
| |
| void perfmon_pmu_remove(struct idxd_device *idxd) |
| { |
| if (!idxd->idxd_pmu) |
| return; |
| |
| perf_pmu_unregister(&idxd->idxd_pmu->pmu); |
| kfree(idxd->idxd_pmu); |
| idxd->idxd_pmu = NULL; |
| } |
| |
| int perfmon_pmu_init(struct idxd_device *idxd) |
| { |
| union idxd_perfcap perfcap; |
| struct idxd_pmu *idxd_pmu; |
| int rc = -ENODEV; |
| |
| /* |
| * If perfmon_offset or num_counters is 0, it means perfmon is |
| * not supported on this hardware. |
| */ |
| if (idxd->perfmon_offset == 0) |
| return -ENODEV; |
| |
| idxd_pmu = kzalloc(sizeof(*idxd_pmu), GFP_KERNEL); |
| if (!idxd_pmu) |
| return -ENOMEM; |
| |
| idxd_pmu->idxd = idxd; |
| idxd->idxd_pmu = idxd_pmu; |
| |
| if (idxd->data->type == IDXD_TYPE_DSA) { |
| rc = sprintf(idxd_pmu->name, "dsa%d", idxd->id); |
| if (rc < 0) |
| goto free; |
| } else if (idxd->data->type == IDXD_TYPE_IAX) { |
| rc = sprintf(idxd_pmu->name, "iax%d", idxd->id); |
| if (rc < 0) |
| goto free; |
| } else { |
| goto free; |
| } |
| |
| perfmon_reset(idxd); |
| |
| perfcap.bits = ioread64(PERFCAP_REG(idxd)); |
| |
| /* |
| * If total perf counter is 0, stop further registration. |
| * This is necessary in order to support driver running on |
| * guest which does not have pmon support. |
| */ |
| if (perfcap.num_perf_counter == 0) |
| goto free; |
| |
| /* A counter width of 0 means it can't count */ |
| if (perfcap.counter_width == 0) |
| goto free; |
| |
| /* Overflow interrupt and counter freeze support must be available */ |
| if (!perfcap.overflow_interrupt || !perfcap.counter_freeze) |
| goto free; |
| |
| /* Number of event categories cannot be 0 */ |
| if (perfcap.num_event_category == 0) |
| goto free; |
| |
| /* |
| * We don't support per-counter capabilities for now. |
| */ |
| if (perfcap.cap_per_counter) |
| goto free; |
| |
| idxd_pmu->n_event_categories = perfcap.num_event_category; |
| idxd_pmu->supported_event_categories = perfcap.global_event_category; |
| idxd_pmu->per_counter_caps_supported = perfcap.cap_per_counter; |
| |
| /* check filter capability. If 0, then filters are not supported */ |
| idxd_pmu->supported_filters = perfcap.filter; |
| if (perfcap.filter) |
| idxd_pmu->n_filters = hweight8(perfcap.filter); |
| |
| /* Store the total number of counters categories, and counter width */ |
| idxd_pmu->n_counters = perfcap.num_perf_counter; |
| idxd_pmu->counter_width = perfcap.counter_width; |
| |
| idxd_pmu_init(idxd_pmu); |
| |
| rc = perf_pmu_register(&idxd_pmu->pmu, idxd_pmu->name, -1); |
| if (rc) |
| goto free; |
| |
| out: |
| return rc; |
| free: |
| kfree(idxd_pmu); |
| idxd->idxd_pmu = NULL; |
| |
| goto out; |
| } |