| // SPDX-License-Identifier: GPL-2.0-only |
| /* |
| * This driver adds support for HNS3 PMU iEP device. Related perf events are |
| * bandwidth, latency, packet rate, interrupt rate etc. |
| * |
| * Copyright (C) 2022 HiSilicon Limited |
| */ |
| #include <linux/bitfield.h> |
| #include <linux/bitmap.h> |
| #include <linux/bug.h> |
| #include <linux/cpuhotplug.h> |
| #include <linux/cpumask.h> |
| #include <linux/delay.h> |
| #include <linux/device.h> |
| #include <linux/err.h> |
| #include <linux/interrupt.h> |
| #include <linux/iopoll.h> |
| #include <linux/io-64-nonatomic-hi-lo.h> |
| #include <linux/irq.h> |
| #include <linux/kernel.h> |
| #include <linux/list.h> |
| #include <linux/module.h> |
| #include <linux/pci.h> |
| #include <linux/pci-epf.h> |
| #include <linux/perf_event.h> |
| #include <linux/smp.h> |
| |
| /* registers offset address */ |
| #define HNS3_PMU_REG_GLOBAL_CTRL 0x0000 |
| #define HNS3_PMU_REG_CLOCK_FREQ 0x0020 |
| #define HNS3_PMU_REG_BDF 0x0fe0 |
| #define HNS3_PMU_REG_VERSION 0x0fe4 |
| #define HNS3_PMU_REG_DEVICE_ID 0x0fe8 |
| |
| #define HNS3_PMU_REG_EVENT_OFFSET 0x1000 |
| #define HNS3_PMU_REG_EVENT_SIZE 0x1000 |
| #define HNS3_PMU_REG_EVENT_CTRL_LOW 0x00 |
| #define HNS3_PMU_REG_EVENT_CTRL_HIGH 0x04 |
| #define HNS3_PMU_REG_EVENT_INTR_STATUS 0x08 |
| #define HNS3_PMU_REG_EVENT_INTR_MASK 0x0c |
| #define HNS3_PMU_REG_EVENT_COUNTER 0x10 |
| #define HNS3_PMU_REG_EVENT_EXT_COUNTER 0x18 |
| #define HNS3_PMU_REG_EVENT_QID_CTRL 0x28 |
| #define HNS3_PMU_REG_EVENT_QID_PARA 0x2c |
| |
| #define HNS3_PMU_FILTER_SUPPORT_GLOBAL BIT(0) |
| #define HNS3_PMU_FILTER_SUPPORT_PORT BIT(1) |
| #define HNS3_PMU_FILTER_SUPPORT_PORT_TC BIT(2) |
| #define HNS3_PMU_FILTER_SUPPORT_FUNC BIT(3) |
| #define HNS3_PMU_FILTER_SUPPORT_FUNC_QUEUE BIT(4) |
| #define HNS3_PMU_FILTER_SUPPORT_FUNC_INTR BIT(5) |
| |
| #define HNS3_PMU_FILTER_ALL_TC 0xf |
| #define HNS3_PMU_FILTER_ALL_QUEUE 0xffff |
| |
| #define HNS3_PMU_CTRL_SUBEVENT_S 4 |
| #define HNS3_PMU_CTRL_FILTER_MODE_S 24 |
| |
| #define HNS3_PMU_GLOBAL_START BIT(0) |
| |
| #define HNS3_PMU_EVENT_STATUS_RESET BIT(11) |
| #define HNS3_PMU_EVENT_EN BIT(12) |
| #define HNS3_PMU_EVENT_OVERFLOW_RESTART BIT(15) |
| |
| #define HNS3_PMU_QID_PARA_FUNC_S 0 |
| #define HNS3_PMU_QID_PARA_QUEUE_S 16 |
| |
| #define HNS3_PMU_QID_CTRL_REQ_ENABLE BIT(0) |
| #define HNS3_PMU_QID_CTRL_DONE BIT(1) |
| #define HNS3_PMU_QID_CTRL_MISS BIT(2) |
| |
| #define HNS3_PMU_INTR_MASK_OVERFLOW BIT(1) |
| |
| #define HNS3_PMU_MAX_HW_EVENTS 8 |
| |
| /* |
| * Each hardware event contains two registers (counter and ext_counter) for |
| * bandwidth, packet rate, latency and interrupt rate. These two registers will |
| * be triggered to run at the same when a hardware event is enabled. The meaning |
| * of counter and ext_counter of different event type are different, their |
| * meaning show as follow: |
| * |
| * +----------------+------------------+---------------+ |
| * | event type | counter | ext_counter | |
| * +----------------+------------------+---------------+ |
| * | bandwidth | byte number | cycle number | |
| * +----------------+------------------+---------------+ |
| * | packet rate | packet number | cycle number | |
| * +----------------+------------------+---------------+ |
| * | latency | cycle number | packet number | |
| * +----------------+------------------+---------------+ |
| * | interrupt rate | interrupt number | cycle number | |
| * +----------------+------------------+---------------+ |
| * |
| * The cycle number indicates increment of counter of hardware timer, the |
| * frequency of hardware timer can be read from hw_clk_freq file. |
| * |
| * Performance of each hardware event is calculated by: counter / ext_counter. |
| * |
| * Since processing of data is preferred to be done in userspace, we expose |
| * ext_counter as a separate event for userspace and use bit 16 to indicate it. |
| * For example, event 0x00001 and 0x10001 are actually one event for hardware |
| * because bit 0-15 are same. If the bit 16 of one event is 0 means to read |
| * counter register, otherwise means to read ext_counter register. |
| */ |
| /* bandwidth events */ |
| #define HNS3_PMU_EVT_BW_SSU_EGU_BYTE_NUM 0x00001 |
| #define HNS3_PMU_EVT_BW_SSU_EGU_TIME 0x10001 |
| #define HNS3_PMU_EVT_BW_SSU_RPU_BYTE_NUM 0x00002 |
| #define HNS3_PMU_EVT_BW_SSU_RPU_TIME 0x10002 |
| #define HNS3_PMU_EVT_BW_SSU_ROCE_BYTE_NUM 0x00003 |
| #define HNS3_PMU_EVT_BW_SSU_ROCE_TIME 0x10003 |
| #define HNS3_PMU_EVT_BW_ROCE_SSU_BYTE_NUM 0x00004 |
| #define HNS3_PMU_EVT_BW_ROCE_SSU_TIME 0x10004 |
| #define HNS3_PMU_EVT_BW_TPU_SSU_BYTE_NUM 0x00005 |
| #define HNS3_PMU_EVT_BW_TPU_SSU_TIME 0x10005 |
| #define HNS3_PMU_EVT_BW_RPU_RCBRX_BYTE_NUM 0x00006 |
| #define HNS3_PMU_EVT_BW_RPU_RCBRX_TIME 0x10006 |
| #define HNS3_PMU_EVT_BW_RCBTX_TXSCH_BYTE_NUM 0x00008 |
| #define HNS3_PMU_EVT_BW_RCBTX_TXSCH_TIME 0x10008 |
| #define HNS3_PMU_EVT_BW_WR_FBD_BYTE_NUM 0x00009 |
| #define HNS3_PMU_EVT_BW_WR_FBD_TIME 0x10009 |
| #define HNS3_PMU_EVT_BW_WR_EBD_BYTE_NUM 0x0000a |
| #define HNS3_PMU_EVT_BW_WR_EBD_TIME 0x1000a |
| #define HNS3_PMU_EVT_BW_RD_FBD_BYTE_NUM 0x0000b |
| #define HNS3_PMU_EVT_BW_RD_FBD_TIME 0x1000b |
| #define HNS3_PMU_EVT_BW_RD_EBD_BYTE_NUM 0x0000c |
| #define HNS3_PMU_EVT_BW_RD_EBD_TIME 0x1000c |
| #define HNS3_PMU_EVT_BW_RD_PAY_M0_BYTE_NUM 0x0000d |
| #define HNS3_PMU_EVT_BW_RD_PAY_M0_TIME 0x1000d |
| #define HNS3_PMU_EVT_BW_RD_PAY_M1_BYTE_NUM 0x0000e |
| #define HNS3_PMU_EVT_BW_RD_PAY_M1_TIME 0x1000e |
| #define HNS3_PMU_EVT_BW_WR_PAY_M0_BYTE_NUM 0x0000f |
| #define HNS3_PMU_EVT_BW_WR_PAY_M0_TIME 0x1000f |
| #define HNS3_PMU_EVT_BW_WR_PAY_M1_BYTE_NUM 0x00010 |
| #define HNS3_PMU_EVT_BW_WR_PAY_M1_TIME 0x10010 |
| |
| /* packet rate events */ |
| #define HNS3_PMU_EVT_PPS_IGU_SSU_PACKET_NUM 0x00100 |
| #define HNS3_PMU_EVT_PPS_IGU_SSU_TIME 0x10100 |
| #define HNS3_PMU_EVT_PPS_SSU_EGU_PACKET_NUM 0x00101 |
| #define HNS3_PMU_EVT_PPS_SSU_EGU_TIME 0x10101 |
| #define HNS3_PMU_EVT_PPS_SSU_RPU_PACKET_NUM 0x00102 |
| #define HNS3_PMU_EVT_PPS_SSU_RPU_TIME 0x10102 |
| #define HNS3_PMU_EVT_PPS_SSU_ROCE_PACKET_NUM 0x00103 |
| #define HNS3_PMU_EVT_PPS_SSU_ROCE_TIME 0x10103 |
| #define HNS3_PMU_EVT_PPS_ROCE_SSU_PACKET_NUM 0x00104 |
| #define HNS3_PMU_EVT_PPS_ROCE_SSU_TIME 0x10104 |
| #define HNS3_PMU_EVT_PPS_TPU_SSU_PACKET_NUM 0x00105 |
| #define HNS3_PMU_EVT_PPS_TPU_SSU_TIME 0x10105 |
| #define HNS3_PMU_EVT_PPS_RPU_RCBRX_PACKET_NUM 0x00106 |
| #define HNS3_PMU_EVT_PPS_RPU_RCBRX_TIME 0x10106 |
| #define HNS3_PMU_EVT_PPS_RCBTX_TPU_PACKET_NUM 0x00107 |
| #define HNS3_PMU_EVT_PPS_RCBTX_TPU_TIME 0x10107 |
| #define HNS3_PMU_EVT_PPS_RCBTX_TXSCH_PACKET_NUM 0x00108 |
| #define HNS3_PMU_EVT_PPS_RCBTX_TXSCH_TIME 0x10108 |
| #define HNS3_PMU_EVT_PPS_WR_FBD_PACKET_NUM 0x00109 |
| #define HNS3_PMU_EVT_PPS_WR_FBD_TIME 0x10109 |
| #define HNS3_PMU_EVT_PPS_WR_EBD_PACKET_NUM 0x0010a |
| #define HNS3_PMU_EVT_PPS_WR_EBD_TIME 0x1010a |
| #define HNS3_PMU_EVT_PPS_RD_FBD_PACKET_NUM 0x0010b |
| #define HNS3_PMU_EVT_PPS_RD_FBD_TIME 0x1010b |
| #define HNS3_PMU_EVT_PPS_RD_EBD_PACKET_NUM 0x0010c |
| #define HNS3_PMU_EVT_PPS_RD_EBD_TIME 0x1010c |
| #define HNS3_PMU_EVT_PPS_RD_PAY_M0_PACKET_NUM 0x0010d |
| #define HNS3_PMU_EVT_PPS_RD_PAY_M0_TIME 0x1010d |
| #define HNS3_PMU_EVT_PPS_RD_PAY_M1_PACKET_NUM 0x0010e |
| #define HNS3_PMU_EVT_PPS_RD_PAY_M1_TIME 0x1010e |
| #define HNS3_PMU_EVT_PPS_WR_PAY_M0_PACKET_NUM 0x0010f |
| #define HNS3_PMU_EVT_PPS_WR_PAY_M0_TIME 0x1010f |
| #define HNS3_PMU_EVT_PPS_WR_PAY_M1_PACKET_NUM 0x00110 |
| #define HNS3_PMU_EVT_PPS_WR_PAY_M1_TIME 0x10110 |
| #define HNS3_PMU_EVT_PPS_NICROH_TX_PRE_PACKET_NUM 0x00111 |
| #define HNS3_PMU_EVT_PPS_NICROH_TX_PRE_TIME 0x10111 |
| #define HNS3_PMU_EVT_PPS_NICROH_RX_PRE_PACKET_NUM 0x00112 |
| #define HNS3_PMU_EVT_PPS_NICROH_RX_PRE_TIME 0x10112 |
| |
| /* latency events */ |
| #define HNS3_PMU_EVT_DLY_TX_PUSH_TIME 0x00202 |
| #define HNS3_PMU_EVT_DLY_TX_PUSH_PACKET_NUM 0x10202 |
| #define HNS3_PMU_EVT_DLY_TX_TIME 0x00204 |
| #define HNS3_PMU_EVT_DLY_TX_PACKET_NUM 0x10204 |
| #define HNS3_PMU_EVT_DLY_SSU_TX_NIC_TIME 0x00206 |
| #define HNS3_PMU_EVT_DLY_SSU_TX_NIC_PACKET_NUM 0x10206 |
| #define HNS3_PMU_EVT_DLY_SSU_TX_ROCE_TIME 0x00207 |
| #define HNS3_PMU_EVT_DLY_SSU_TX_ROCE_PACKET_NUM 0x10207 |
| #define HNS3_PMU_EVT_DLY_SSU_RX_NIC_TIME 0x00208 |
| #define HNS3_PMU_EVT_DLY_SSU_RX_NIC_PACKET_NUM 0x10208 |
| #define HNS3_PMU_EVT_DLY_SSU_RX_ROCE_TIME 0x00209 |
| #define HNS3_PMU_EVT_DLY_SSU_RX_ROCE_PACKET_NUM 0x10209 |
| #define HNS3_PMU_EVT_DLY_RPU_TIME 0x0020e |
| #define HNS3_PMU_EVT_DLY_RPU_PACKET_NUM 0x1020e |
| #define HNS3_PMU_EVT_DLY_TPU_TIME 0x0020f |
| #define HNS3_PMU_EVT_DLY_TPU_PACKET_NUM 0x1020f |
| #define HNS3_PMU_EVT_DLY_RPE_TIME 0x00210 |
| #define HNS3_PMU_EVT_DLY_RPE_PACKET_NUM 0x10210 |
| #define HNS3_PMU_EVT_DLY_TPE_TIME 0x00211 |
| #define HNS3_PMU_EVT_DLY_TPE_PACKET_NUM 0x10211 |
| #define HNS3_PMU_EVT_DLY_TPE_PUSH_TIME 0x00212 |
| #define HNS3_PMU_EVT_DLY_TPE_PUSH_PACKET_NUM 0x10212 |
| #define HNS3_PMU_EVT_DLY_WR_FBD_TIME 0x00213 |
| #define HNS3_PMU_EVT_DLY_WR_FBD_PACKET_NUM 0x10213 |
| #define HNS3_PMU_EVT_DLY_WR_EBD_TIME 0x00214 |
| #define HNS3_PMU_EVT_DLY_WR_EBD_PACKET_NUM 0x10214 |
| #define HNS3_PMU_EVT_DLY_RD_FBD_TIME 0x00215 |
| #define HNS3_PMU_EVT_DLY_RD_FBD_PACKET_NUM 0x10215 |
| #define HNS3_PMU_EVT_DLY_RD_EBD_TIME 0x00216 |
| #define HNS3_PMU_EVT_DLY_RD_EBD_PACKET_NUM 0x10216 |
| #define HNS3_PMU_EVT_DLY_RD_PAY_M0_TIME 0x00217 |
| #define HNS3_PMU_EVT_DLY_RD_PAY_M0_PACKET_NUM 0x10217 |
| #define HNS3_PMU_EVT_DLY_RD_PAY_M1_TIME 0x00218 |
| #define HNS3_PMU_EVT_DLY_RD_PAY_M1_PACKET_NUM 0x10218 |
| #define HNS3_PMU_EVT_DLY_WR_PAY_M0_TIME 0x00219 |
| #define HNS3_PMU_EVT_DLY_WR_PAY_M0_PACKET_NUM 0x10219 |
| #define HNS3_PMU_EVT_DLY_WR_PAY_M1_TIME 0x0021a |
| #define HNS3_PMU_EVT_DLY_WR_PAY_M1_PACKET_NUM 0x1021a |
| #define HNS3_PMU_EVT_DLY_MSIX_WRITE_TIME 0x0021c |
| #define HNS3_PMU_EVT_DLY_MSIX_WRITE_PACKET_NUM 0x1021c |
| |
| /* interrupt rate events */ |
| #define HNS3_PMU_EVT_PPS_MSIX_NIC_INTR_NUM 0x00300 |
| #define HNS3_PMU_EVT_PPS_MSIX_NIC_TIME 0x10300 |
| |
| /* filter mode supported by each bandwidth event */ |
| #define HNS3_PMU_FILTER_BW_SSU_EGU 0x07 |
| #define HNS3_PMU_FILTER_BW_SSU_RPU 0x1f |
| #define HNS3_PMU_FILTER_BW_SSU_ROCE 0x0f |
| #define HNS3_PMU_FILTER_BW_ROCE_SSU 0x0f |
| #define HNS3_PMU_FILTER_BW_TPU_SSU 0x1f |
| #define HNS3_PMU_FILTER_BW_RPU_RCBRX 0x11 |
| #define HNS3_PMU_FILTER_BW_RCBTX_TXSCH 0x11 |
| #define HNS3_PMU_FILTER_BW_WR_FBD 0x1b |
| #define HNS3_PMU_FILTER_BW_WR_EBD 0x11 |
| #define HNS3_PMU_FILTER_BW_RD_FBD 0x01 |
| #define HNS3_PMU_FILTER_BW_RD_EBD 0x1b |
| #define HNS3_PMU_FILTER_BW_RD_PAY_M0 0x01 |
| #define HNS3_PMU_FILTER_BW_RD_PAY_M1 0x01 |
| #define HNS3_PMU_FILTER_BW_WR_PAY_M0 0x01 |
| #define HNS3_PMU_FILTER_BW_WR_PAY_M1 0x01 |
| |
| /* filter mode supported by each packet rate event */ |
| #define HNS3_PMU_FILTER_PPS_IGU_SSU 0x07 |
| #define HNS3_PMU_FILTER_PPS_SSU_EGU 0x07 |
| #define HNS3_PMU_FILTER_PPS_SSU_RPU 0x1f |
| #define HNS3_PMU_FILTER_PPS_SSU_ROCE 0x0f |
| #define HNS3_PMU_FILTER_PPS_ROCE_SSU 0x0f |
| #define HNS3_PMU_FILTER_PPS_TPU_SSU 0x1f |
| #define HNS3_PMU_FILTER_PPS_RPU_RCBRX 0x11 |
| #define HNS3_PMU_FILTER_PPS_RCBTX_TPU 0x1f |
| #define HNS3_PMU_FILTER_PPS_RCBTX_TXSCH 0x11 |
| #define HNS3_PMU_FILTER_PPS_WR_FBD 0x1b |
| #define HNS3_PMU_FILTER_PPS_WR_EBD 0x11 |
| #define HNS3_PMU_FILTER_PPS_RD_FBD 0x01 |
| #define HNS3_PMU_FILTER_PPS_RD_EBD 0x1b |
| #define HNS3_PMU_FILTER_PPS_RD_PAY_M0 0x01 |
| #define HNS3_PMU_FILTER_PPS_RD_PAY_M1 0x01 |
| #define HNS3_PMU_FILTER_PPS_WR_PAY_M0 0x01 |
| #define HNS3_PMU_FILTER_PPS_WR_PAY_M1 0x01 |
| #define HNS3_PMU_FILTER_PPS_NICROH_TX_PRE 0x01 |
| #define HNS3_PMU_FILTER_PPS_NICROH_RX_PRE 0x01 |
| |
| /* filter mode supported by each latency event */ |
| #define HNS3_PMU_FILTER_DLY_TX_PUSH 0x01 |
| #define HNS3_PMU_FILTER_DLY_TX 0x01 |
| #define HNS3_PMU_FILTER_DLY_SSU_TX_NIC 0x07 |
| #define HNS3_PMU_FILTER_DLY_SSU_TX_ROCE 0x07 |
| #define HNS3_PMU_FILTER_DLY_SSU_RX_NIC 0x07 |
| #define HNS3_PMU_FILTER_DLY_SSU_RX_ROCE 0x07 |
| #define HNS3_PMU_FILTER_DLY_RPU 0x11 |
| #define HNS3_PMU_FILTER_DLY_TPU 0x1f |
| #define HNS3_PMU_FILTER_DLY_RPE 0x01 |
| #define HNS3_PMU_FILTER_DLY_TPE 0x0b |
| #define HNS3_PMU_FILTER_DLY_TPE_PUSH 0x1b |
| #define HNS3_PMU_FILTER_DLY_WR_FBD 0x1b |
| #define HNS3_PMU_FILTER_DLY_WR_EBD 0x11 |
| #define HNS3_PMU_FILTER_DLY_RD_FBD 0x01 |
| #define HNS3_PMU_FILTER_DLY_RD_EBD 0x1b |
| #define HNS3_PMU_FILTER_DLY_RD_PAY_M0 0x01 |
| #define HNS3_PMU_FILTER_DLY_RD_PAY_M1 0x01 |
| #define HNS3_PMU_FILTER_DLY_WR_PAY_M0 0x01 |
| #define HNS3_PMU_FILTER_DLY_WR_PAY_M1 0x01 |
| #define HNS3_PMU_FILTER_DLY_MSIX_WRITE 0x01 |
| |
| /* filter mode supported by each interrupt rate event */ |
| #define HNS3_PMU_FILTER_INTR_MSIX_NIC 0x01 |
| |
| enum hns3_pmu_hw_filter_mode { |
| HNS3_PMU_HW_FILTER_GLOBAL, |
| HNS3_PMU_HW_FILTER_PORT, |
| HNS3_PMU_HW_FILTER_PORT_TC, |
| HNS3_PMU_HW_FILTER_FUNC, |
| HNS3_PMU_HW_FILTER_FUNC_QUEUE, |
| HNS3_PMU_HW_FILTER_FUNC_INTR, |
| }; |
| |
| struct hns3_pmu_event_attr { |
| u32 event; |
| u16 filter_support; |
| }; |
| |
| struct hns3_pmu { |
| struct perf_event *hw_events[HNS3_PMU_MAX_HW_EVENTS]; |
| struct hlist_node node; |
| struct pci_dev *pdev; |
| struct pmu pmu; |
| void __iomem *base; |
| int irq; |
| int on_cpu; |
| u32 identifier; |
| u32 hw_clk_freq; /* hardware clock frequency of PMU */ |
| /* maximum and minimum bdf allowed by PMU */ |
| u16 bdf_min; |
| u16 bdf_max; |
| }; |
| |
| #define to_hns3_pmu(p) (container_of((p), struct hns3_pmu, pmu)) |
| |
| #define GET_PCI_DEVFN(bdf) ((bdf) & 0xff) |
| |
| #define FILTER_CONDITION_PORT(port) ((1 << (port)) & 0xff) |
| #define FILTER_CONDITION_PORT_TC(port, tc) (((port) << 3) | ((tc) & 0x07)) |
| #define FILTER_CONDITION_FUNC_INTR(func, intr) (((intr) << 8) | (func)) |
| |
| #define HNS3_PMU_FILTER_ATTR(_name, _config, _start, _end) \ |
| static inline u64 hns3_pmu_get_##_name(struct perf_event *event) \ |
| { \ |
| return FIELD_GET(GENMASK_ULL(_end, _start), \ |
| event->attr._config); \ |
| } |
| |
| HNS3_PMU_FILTER_ATTR(subevent, config, 0, 7); |
| HNS3_PMU_FILTER_ATTR(event_type, config, 8, 15); |
| HNS3_PMU_FILTER_ATTR(ext_counter_used, config, 16, 16); |
| HNS3_PMU_FILTER_ATTR(port, config1, 0, 3); |
| HNS3_PMU_FILTER_ATTR(tc, config1, 4, 7); |
| HNS3_PMU_FILTER_ATTR(bdf, config1, 8, 23); |
| HNS3_PMU_FILTER_ATTR(queue, config1, 24, 39); |
| HNS3_PMU_FILTER_ATTR(intr, config1, 40, 51); |
| HNS3_PMU_FILTER_ATTR(global, config1, 52, 52); |
| |
| #define HNS3_BW_EVT_BYTE_NUM(_name) (&(struct hns3_pmu_event_attr) {\ |
| HNS3_PMU_EVT_BW_##_name##_BYTE_NUM, \ |
| HNS3_PMU_FILTER_BW_##_name}) |
| #define HNS3_BW_EVT_TIME(_name) (&(struct hns3_pmu_event_attr) {\ |
| HNS3_PMU_EVT_BW_##_name##_TIME, \ |
| HNS3_PMU_FILTER_BW_##_name}) |
| #define HNS3_PPS_EVT_PACKET_NUM(_name) (&(struct hns3_pmu_event_attr) {\ |
| HNS3_PMU_EVT_PPS_##_name##_PACKET_NUM, \ |
| HNS3_PMU_FILTER_PPS_##_name}) |
| #define HNS3_PPS_EVT_TIME(_name) (&(struct hns3_pmu_event_attr) {\ |
| HNS3_PMU_EVT_PPS_##_name##_TIME, \ |
| HNS3_PMU_FILTER_PPS_##_name}) |
| #define HNS3_DLY_EVT_TIME(_name) (&(struct hns3_pmu_event_attr) {\ |
| HNS3_PMU_EVT_DLY_##_name##_TIME, \ |
| HNS3_PMU_FILTER_DLY_##_name}) |
| #define HNS3_DLY_EVT_PACKET_NUM(_name) (&(struct hns3_pmu_event_attr) {\ |
| HNS3_PMU_EVT_DLY_##_name##_PACKET_NUM, \ |
| HNS3_PMU_FILTER_DLY_##_name}) |
| #define HNS3_INTR_EVT_INTR_NUM(_name) (&(struct hns3_pmu_event_attr) {\ |
| HNS3_PMU_EVT_PPS_##_name##_INTR_NUM, \ |
| HNS3_PMU_FILTER_INTR_##_name}) |
| #define HNS3_INTR_EVT_TIME(_name) (&(struct hns3_pmu_event_attr) {\ |
| HNS3_PMU_EVT_PPS_##_name##_TIME, \ |
| HNS3_PMU_FILTER_INTR_##_name}) |
| |
| static ssize_t hns3_pmu_event_show(struct device *dev, |
| struct device_attribute *attr, char *buf) |
| { |
| struct hns3_pmu_event_attr *event; |
| struct dev_ext_attribute *eattr; |
| |
| eattr = container_of(attr, struct dev_ext_attribute, attr); |
| event = eattr->var; |
| |
| return sysfs_emit(buf, "config=0x%x\n", event->event); |
| } |
| |
| static ssize_t hns3_pmu_filter_mode_show(struct device *dev, |
| struct device_attribute *attr, |
| char *buf) |
| { |
| struct hns3_pmu_event_attr *event; |
| struct dev_ext_attribute *eattr; |
| int len; |
| |
| eattr = container_of(attr, struct dev_ext_attribute, attr); |
| event = eattr->var; |
| |
| len = sysfs_emit_at(buf, 0, "filter mode supported: "); |
| if (event->filter_support & HNS3_PMU_FILTER_SUPPORT_GLOBAL) |
| len += sysfs_emit_at(buf, len, "global "); |
| if (event->filter_support & HNS3_PMU_FILTER_SUPPORT_PORT) |
| len += sysfs_emit_at(buf, len, "port "); |
| if (event->filter_support & HNS3_PMU_FILTER_SUPPORT_PORT_TC) |
| len += sysfs_emit_at(buf, len, "port-tc "); |
| if (event->filter_support & HNS3_PMU_FILTER_SUPPORT_FUNC) |
| len += sysfs_emit_at(buf, len, "func "); |
| if (event->filter_support & HNS3_PMU_FILTER_SUPPORT_FUNC_QUEUE) |
| len += sysfs_emit_at(buf, len, "func-queue "); |
| if (event->filter_support & HNS3_PMU_FILTER_SUPPORT_FUNC_INTR) |
| len += sysfs_emit_at(buf, len, "func-intr "); |
| |
| len += sysfs_emit_at(buf, len, "\n"); |
| |
| return len; |
| } |
| |
| #define HNS3_PMU_ATTR(_name, _func, _config) \ |
| (&((struct dev_ext_attribute[]) { \ |
| { __ATTR(_name, 0444, _func, NULL), (void *)_config } \ |
| })[0].attr.attr) |
| |
| #define HNS3_PMU_FORMAT_ATTR(_name, _format) \ |
| HNS3_PMU_ATTR(_name, device_show_string, _format) |
| #define HNS3_PMU_EVENT_ATTR(_name, _event) \ |
| HNS3_PMU_ATTR(_name, hns3_pmu_event_show, (void *)_event) |
| #define HNS3_PMU_FLT_MODE_ATTR(_name, _event) \ |
| HNS3_PMU_ATTR(_name, hns3_pmu_filter_mode_show, (void *)_event) |
| |
| #define HNS3_PMU_BW_EVT_PAIR(_name, _macro) \ |
| HNS3_PMU_EVENT_ATTR(_name##_byte_num, HNS3_BW_EVT_BYTE_NUM(_macro)), \ |
| HNS3_PMU_EVENT_ATTR(_name##_time, HNS3_BW_EVT_TIME(_macro)) |
| #define HNS3_PMU_PPS_EVT_PAIR(_name, _macro) \ |
| HNS3_PMU_EVENT_ATTR(_name##_packet_num, HNS3_PPS_EVT_PACKET_NUM(_macro)), \ |
| HNS3_PMU_EVENT_ATTR(_name##_time, HNS3_PPS_EVT_TIME(_macro)) |
| #define HNS3_PMU_DLY_EVT_PAIR(_name, _macro) \ |
| HNS3_PMU_EVENT_ATTR(_name##_time, HNS3_DLY_EVT_TIME(_macro)), \ |
| HNS3_PMU_EVENT_ATTR(_name##_packet_num, HNS3_DLY_EVT_PACKET_NUM(_macro)) |
| #define HNS3_PMU_INTR_EVT_PAIR(_name, _macro) \ |
| HNS3_PMU_EVENT_ATTR(_name##_intr_num, HNS3_INTR_EVT_INTR_NUM(_macro)), \ |
| HNS3_PMU_EVENT_ATTR(_name##_time, HNS3_INTR_EVT_TIME(_macro)) |
| |
| #define HNS3_PMU_BW_FLT_MODE_PAIR(_name, _macro) \ |
| HNS3_PMU_FLT_MODE_ATTR(_name##_byte_num, HNS3_BW_EVT_BYTE_NUM(_macro)), \ |
| HNS3_PMU_FLT_MODE_ATTR(_name##_time, HNS3_BW_EVT_TIME(_macro)) |
| #define HNS3_PMU_PPS_FLT_MODE_PAIR(_name, _macro) \ |
| HNS3_PMU_FLT_MODE_ATTR(_name##_packet_num, HNS3_PPS_EVT_PACKET_NUM(_macro)), \ |
| HNS3_PMU_FLT_MODE_ATTR(_name##_time, HNS3_PPS_EVT_TIME(_macro)) |
| #define HNS3_PMU_DLY_FLT_MODE_PAIR(_name, _macro) \ |
| HNS3_PMU_FLT_MODE_ATTR(_name##_time, HNS3_DLY_EVT_TIME(_macro)), \ |
| HNS3_PMU_FLT_MODE_ATTR(_name##_packet_num, HNS3_DLY_EVT_PACKET_NUM(_macro)) |
| #define HNS3_PMU_INTR_FLT_MODE_PAIR(_name, _macro) \ |
| HNS3_PMU_FLT_MODE_ATTR(_name##_intr_num, HNS3_INTR_EVT_INTR_NUM(_macro)), \ |
| HNS3_PMU_FLT_MODE_ATTR(_name##_time, HNS3_INTR_EVT_TIME(_macro)) |
| |
| static u8 hns3_pmu_hw_filter_modes[] = { |
| HNS3_PMU_HW_FILTER_GLOBAL, |
| HNS3_PMU_HW_FILTER_PORT, |
| HNS3_PMU_HW_FILTER_PORT_TC, |
| HNS3_PMU_HW_FILTER_FUNC, |
| HNS3_PMU_HW_FILTER_FUNC_QUEUE, |
| HNS3_PMU_HW_FILTER_FUNC_INTR, |
| }; |
| |
| #define HNS3_PMU_SET_HW_FILTER(_hwc, _mode) \ |
| ((_hwc)->addr_filters = (void *)&hns3_pmu_hw_filter_modes[(_mode)]) |
| |
| static ssize_t identifier_show(struct device *dev, |
| struct device_attribute *attr, char *buf) |
| { |
| struct hns3_pmu *hns3_pmu = to_hns3_pmu(dev_get_drvdata(dev)); |
| |
| return sysfs_emit(buf, "0x%x\n", hns3_pmu->identifier); |
| } |
| static DEVICE_ATTR_RO(identifier); |
| |
| static ssize_t cpumask_show(struct device *dev, struct device_attribute *attr, |
| char *buf) |
| { |
| struct hns3_pmu *hns3_pmu = to_hns3_pmu(dev_get_drvdata(dev)); |
| |
| return sysfs_emit(buf, "%d\n", hns3_pmu->on_cpu); |
| } |
| static DEVICE_ATTR_RO(cpumask); |
| |
| static ssize_t bdf_min_show(struct device *dev, struct device_attribute *attr, |
| char *buf) |
| { |
| struct hns3_pmu *hns3_pmu = to_hns3_pmu(dev_get_drvdata(dev)); |
| u16 bdf = hns3_pmu->bdf_min; |
| |
| return sysfs_emit(buf, "%02x:%02x.%x\n", PCI_BUS_NUM(bdf), |
| PCI_SLOT(bdf), PCI_FUNC(bdf)); |
| } |
| static DEVICE_ATTR_RO(bdf_min); |
| |
| static ssize_t bdf_max_show(struct device *dev, struct device_attribute *attr, |
| char *buf) |
| { |
| struct hns3_pmu *hns3_pmu = to_hns3_pmu(dev_get_drvdata(dev)); |
| u16 bdf = hns3_pmu->bdf_max; |
| |
| return sysfs_emit(buf, "%02x:%02x.%x\n", PCI_BUS_NUM(bdf), |
| PCI_SLOT(bdf), PCI_FUNC(bdf)); |
| } |
| static DEVICE_ATTR_RO(bdf_max); |
| |
| static ssize_t hw_clk_freq_show(struct device *dev, |
| struct device_attribute *attr, char *buf) |
| { |
| struct hns3_pmu *hns3_pmu = to_hns3_pmu(dev_get_drvdata(dev)); |
| |
| return sysfs_emit(buf, "%u\n", hns3_pmu->hw_clk_freq); |
| } |
| static DEVICE_ATTR_RO(hw_clk_freq); |
| |
| static struct attribute *hns3_pmu_events_attr[] = { |
| /* bandwidth events */ |
| HNS3_PMU_BW_EVT_PAIR(bw_ssu_egu, SSU_EGU), |
| HNS3_PMU_BW_EVT_PAIR(bw_ssu_rpu, SSU_RPU), |
| HNS3_PMU_BW_EVT_PAIR(bw_ssu_roce, SSU_ROCE), |
| HNS3_PMU_BW_EVT_PAIR(bw_roce_ssu, ROCE_SSU), |
| HNS3_PMU_BW_EVT_PAIR(bw_tpu_ssu, TPU_SSU), |
| HNS3_PMU_BW_EVT_PAIR(bw_rpu_rcbrx, RPU_RCBRX), |
| HNS3_PMU_BW_EVT_PAIR(bw_rcbtx_txsch, RCBTX_TXSCH), |
| HNS3_PMU_BW_EVT_PAIR(bw_wr_fbd, WR_FBD), |
| HNS3_PMU_BW_EVT_PAIR(bw_wr_ebd, WR_EBD), |
| HNS3_PMU_BW_EVT_PAIR(bw_rd_fbd, RD_FBD), |
| HNS3_PMU_BW_EVT_PAIR(bw_rd_ebd, RD_EBD), |
| HNS3_PMU_BW_EVT_PAIR(bw_rd_pay_m0, RD_PAY_M0), |
| HNS3_PMU_BW_EVT_PAIR(bw_rd_pay_m1, RD_PAY_M1), |
| HNS3_PMU_BW_EVT_PAIR(bw_wr_pay_m0, WR_PAY_M0), |
| HNS3_PMU_BW_EVT_PAIR(bw_wr_pay_m1, WR_PAY_M1), |
| |
| /* packet rate events */ |
| HNS3_PMU_PPS_EVT_PAIR(pps_igu_ssu, IGU_SSU), |
| HNS3_PMU_PPS_EVT_PAIR(pps_ssu_egu, SSU_EGU), |
| HNS3_PMU_PPS_EVT_PAIR(pps_ssu_rpu, SSU_RPU), |
| HNS3_PMU_PPS_EVT_PAIR(pps_ssu_roce, SSU_ROCE), |
| HNS3_PMU_PPS_EVT_PAIR(pps_roce_ssu, ROCE_SSU), |
| HNS3_PMU_PPS_EVT_PAIR(pps_tpu_ssu, TPU_SSU), |
| HNS3_PMU_PPS_EVT_PAIR(pps_rpu_rcbrx, RPU_RCBRX), |
| HNS3_PMU_PPS_EVT_PAIR(pps_rcbtx_tpu, RCBTX_TPU), |
| HNS3_PMU_PPS_EVT_PAIR(pps_rcbtx_txsch, RCBTX_TXSCH), |
| HNS3_PMU_PPS_EVT_PAIR(pps_wr_fbd, WR_FBD), |
| HNS3_PMU_PPS_EVT_PAIR(pps_wr_ebd, WR_EBD), |
| HNS3_PMU_PPS_EVT_PAIR(pps_rd_fbd, RD_FBD), |
| HNS3_PMU_PPS_EVT_PAIR(pps_rd_ebd, RD_EBD), |
| HNS3_PMU_PPS_EVT_PAIR(pps_rd_pay_m0, RD_PAY_M0), |
| HNS3_PMU_PPS_EVT_PAIR(pps_rd_pay_m1, RD_PAY_M1), |
| HNS3_PMU_PPS_EVT_PAIR(pps_wr_pay_m0, WR_PAY_M0), |
| HNS3_PMU_PPS_EVT_PAIR(pps_wr_pay_m1, WR_PAY_M1), |
| HNS3_PMU_PPS_EVT_PAIR(pps_intr_nicroh_tx_pre, NICROH_TX_PRE), |
| HNS3_PMU_PPS_EVT_PAIR(pps_intr_nicroh_rx_pre, NICROH_RX_PRE), |
| |
| /* latency events */ |
| HNS3_PMU_DLY_EVT_PAIR(dly_tx_push_to_mac, TX_PUSH), |
| HNS3_PMU_DLY_EVT_PAIR(dly_tx_normal_to_mac, TX), |
| HNS3_PMU_DLY_EVT_PAIR(dly_ssu_tx_th_nic, SSU_TX_NIC), |
| HNS3_PMU_DLY_EVT_PAIR(dly_ssu_tx_th_roce, SSU_TX_ROCE), |
| HNS3_PMU_DLY_EVT_PAIR(dly_ssu_rx_th_nic, SSU_RX_NIC), |
| HNS3_PMU_DLY_EVT_PAIR(dly_ssu_rx_th_roce, SSU_RX_ROCE), |
| HNS3_PMU_DLY_EVT_PAIR(dly_rpu, RPU), |
| HNS3_PMU_DLY_EVT_PAIR(dly_tpu, TPU), |
| HNS3_PMU_DLY_EVT_PAIR(dly_rpe, RPE), |
| HNS3_PMU_DLY_EVT_PAIR(dly_tpe_normal, TPE), |
| HNS3_PMU_DLY_EVT_PAIR(dly_tpe_push, TPE_PUSH), |
| HNS3_PMU_DLY_EVT_PAIR(dly_wr_fbd, WR_FBD), |
| HNS3_PMU_DLY_EVT_PAIR(dly_wr_ebd, WR_EBD), |
| HNS3_PMU_DLY_EVT_PAIR(dly_rd_fbd, RD_FBD), |
| HNS3_PMU_DLY_EVT_PAIR(dly_rd_ebd, RD_EBD), |
| HNS3_PMU_DLY_EVT_PAIR(dly_rd_pay_m0, RD_PAY_M0), |
| HNS3_PMU_DLY_EVT_PAIR(dly_rd_pay_m1, RD_PAY_M1), |
| HNS3_PMU_DLY_EVT_PAIR(dly_wr_pay_m0, WR_PAY_M0), |
| HNS3_PMU_DLY_EVT_PAIR(dly_wr_pay_m1, WR_PAY_M1), |
| HNS3_PMU_DLY_EVT_PAIR(dly_msix_write, MSIX_WRITE), |
| |
| /* interrupt rate events */ |
| HNS3_PMU_INTR_EVT_PAIR(pps_intr_msix_nic, MSIX_NIC), |
| |
| NULL |
| }; |
| |
| static struct attribute *hns3_pmu_filter_mode_attr[] = { |
| /* bandwidth events */ |
| HNS3_PMU_BW_FLT_MODE_PAIR(bw_ssu_egu, SSU_EGU), |
| HNS3_PMU_BW_FLT_MODE_PAIR(bw_ssu_rpu, SSU_RPU), |
| HNS3_PMU_BW_FLT_MODE_PAIR(bw_ssu_roce, SSU_ROCE), |
| HNS3_PMU_BW_FLT_MODE_PAIR(bw_roce_ssu, ROCE_SSU), |
| HNS3_PMU_BW_FLT_MODE_PAIR(bw_tpu_ssu, TPU_SSU), |
| HNS3_PMU_BW_FLT_MODE_PAIR(bw_rpu_rcbrx, RPU_RCBRX), |
| HNS3_PMU_BW_FLT_MODE_PAIR(bw_rcbtx_txsch, RCBTX_TXSCH), |
| HNS3_PMU_BW_FLT_MODE_PAIR(bw_wr_fbd, WR_FBD), |
| HNS3_PMU_BW_FLT_MODE_PAIR(bw_wr_ebd, WR_EBD), |
| HNS3_PMU_BW_FLT_MODE_PAIR(bw_rd_fbd, RD_FBD), |
| HNS3_PMU_BW_FLT_MODE_PAIR(bw_rd_ebd, RD_EBD), |
| HNS3_PMU_BW_FLT_MODE_PAIR(bw_rd_pay_m0, RD_PAY_M0), |
| HNS3_PMU_BW_FLT_MODE_PAIR(bw_rd_pay_m1, RD_PAY_M1), |
| HNS3_PMU_BW_FLT_MODE_PAIR(bw_wr_pay_m0, WR_PAY_M0), |
| HNS3_PMU_BW_FLT_MODE_PAIR(bw_wr_pay_m1, WR_PAY_M1), |
| |
| /* packet rate events */ |
| HNS3_PMU_PPS_FLT_MODE_PAIR(pps_igu_ssu, IGU_SSU), |
| HNS3_PMU_PPS_FLT_MODE_PAIR(pps_ssu_egu, SSU_EGU), |
| HNS3_PMU_PPS_FLT_MODE_PAIR(pps_ssu_rpu, SSU_RPU), |
| HNS3_PMU_PPS_FLT_MODE_PAIR(pps_ssu_roce, SSU_ROCE), |
| HNS3_PMU_PPS_FLT_MODE_PAIR(pps_roce_ssu, ROCE_SSU), |
| HNS3_PMU_PPS_FLT_MODE_PAIR(pps_tpu_ssu, TPU_SSU), |
| HNS3_PMU_PPS_FLT_MODE_PAIR(pps_rpu_rcbrx, RPU_RCBRX), |
| HNS3_PMU_PPS_FLT_MODE_PAIR(pps_rcbtx_tpu, RCBTX_TPU), |
| HNS3_PMU_PPS_FLT_MODE_PAIR(pps_rcbtx_txsch, RCBTX_TXSCH), |
| HNS3_PMU_PPS_FLT_MODE_PAIR(pps_wr_fbd, WR_FBD), |
| HNS3_PMU_PPS_FLT_MODE_PAIR(pps_wr_ebd, WR_EBD), |
| HNS3_PMU_PPS_FLT_MODE_PAIR(pps_rd_fbd, RD_FBD), |
| HNS3_PMU_PPS_FLT_MODE_PAIR(pps_rd_ebd, RD_EBD), |
| HNS3_PMU_PPS_FLT_MODE_PAIR(pps_rd_pay_m0, RD_PAY_M0), |
| HNS3_PMU_PPS_FLT_MODE_PAIR(pps_rd_pay_m1, RD_PAY_M1), |
| HNS3_PMU_PPS_FLT_MODE_PAIR(pps_wr_pay_m0, WR_PAY_M0), |
| HNS3_PMU_PPS_FLT_MODE_PAIR(pps_wr_pay_m1, WR_PAY_M1), |
| HNS3_PMU_PPS_FLT_MODE_PAIR(pps_intr_nicroh_tx_pre, NICROH_TX_PRE), |
| HNS3_PMU_PPS_FLT_MODE_PAIR(pps_intr_nicroh_rx_pre, NICROH_RX_PRE), |
| |
| /* latency events */ |
| HNS3_PMU_DLY_FLT_MODE_PAIR(dly_tx_push_to_mac, TX_PUSH), |
| HNS3_PMU_DLY_FLT_MODE_PAIR(dly_tx_normal_to_mac, TX), |
| HNS3_PMU_DLY_FLT_MODE_PAIR(dly_ssu_tx_th_nic, SSU_TX_NIC), |
| HNS3_PMU_DLY_FLT_MODE_PAIR(dly_ssu_tx_th_roce, SSU_TX_ROCE), |
| HNS3_PMU_DLY_FLT_MODE_PAIR(dly_ssu_rx_th_nic, SSU_RX_NIC), |
| HNS3_PMU_DLY_FLT_MODE_PAIR(dly_ssu_rx_th_roce, SSU_RX_ROCE), |
| HNS3_PMU_DLY_FLT_MODE_PAIR(dly_rpu, RPU), |
| HNS3_PMU_DLY_FLT_MODE_PAIR(dly_tpu, TPU), |
| HNS3_PMU_DLY_FLT_MODE_PAIR(dly_rpe, RPE), |
| HNS3_PMU_DLY_FLT_MODE_PAIR(dly_tpe_normal, TPE), |
| HNS3_PMU_DLY_FLT_MODE_PAIR(dly_tpe_push, TPE_PUSH), |
| HNS3_PMU_DLY_FLT_MODE_PAIR(dly_wr_fbd, WR_FBD), |
| HNS3_PMU_DLY_FLT_MODE_PAIR(dly_wr_ebd, WR_EBD), |
| HNS3_PMU_DLY_FLT_MODE_PAIR(dly_rd_fbd, RD_FBD), |
| HNS3_PMU_DLY_FLT_MODE_PAIR(dly_rd_ebd, RD_EBD), |
| HNS3_PMU_DLY_FLT_MODE_PAIR(dly_rd_pay_m0, RD_PAY_M0), |
| HNS3_PMU_DLY_FLT_MODE_PAIR(dly_rd_pay_m1, RD_PAY_M1), |
| HNS3_PMU_DLY_FLT_MODE_PAIR(dly_wr_pay_m0, WR_PAY_M0), |
| HNS3_PMU_DLY_FLT_MODE_PAIR(dly_wr_pay_m1, WR_PAY_M1), |
| HNS3_PMU_DLY_FLT_MODE_PAIR(dly_msix_write, MSIX_WRITE), |
| |
| /* interrupt rate events */ |
| HNS3_PMU_INTR_FLT_MODE_PAIR(pps_intr_msix_nic, MSIX_NIC), |
| |
| NULL |
| }; |
| |
| static struct attribute_group hns3_pmu_events_group = { |
| .name = "events", |
| .attrs = hns3_pmu_events_attr, |
| }; |
| |
| static struct attribute_group hns3_pmu_filter_mode_group = { |
| .name = "filtermode", |
| .attrs = hns3_pmu_filter_mode_attr, |
| }; |
| |
| static struct attribute *hns3_pmu_format_attr[] = { |
| HNS3_PMU_FORMAT_ATTR(subevent, "config:0-7"), |
| HNS3_PMU_FORMAT_ATTR(event_type, "config:8-15"), |
| HNS3_PMU_FORMAT_ATTR(ext_counter_used, "config:16"), |
| HNS3_PMU_FORMAT_ATTR(port, "config1:0-3"), |
| HNS3_PMU_FORMAT_ATTR(tc, "config1:4-7"), |
| HNS3_PMU_FORMAT_ATTR(bdf, "config1:8-23"), |
| HNS3_PMU_FORMAT_ATTR(queue, "config1:24-39"), |
| HNS3_PMU_FORMAT_ATTR(intr, "config1:40-51"), |
| HNS3_PMU_FORMAT_ATTR(global, "config1:52"), |
| NULL |
| }; |
| |
| static struct attribute_group hns3_pmu_format_group = { |
| .name = "format", |
| .attrs = hns3_pmu_format_attr, |
| }; |
| |
| static struct attribute *hns3_pmu_cpumask_attrs[] = { |
| &dev_attr_cpumask.attr, |
| NULL |
| }; |
| |
| static struct attribute_group hns3_pmu_cpumask_attr_group = { |
| .attrs = hns3_pmu_cpumask_attrs, |
| }; |
| |
| static struct attribute *hns3_pmu_identifier_attrs[] = { |
| &dev_attr_identifier.attr, |
| NULL |
| }; |
| |
| static struct attribute_group hns3_pmu_identifier_attr_group = { |
| .attrs = hns3_pmu_identifier_attrs, |
| }; |
| |
| static struct attribute *hns3_pmu_bdf_range_attrs[] = { |
| &dev_attr_bdf_min.attr, |
| &dev_attr_bdf_max.attr, |
| NULL |
| }; |
| |
| static struct attribute_group hns3_pmu_bdf_range_attr_group = { |
| .attrs = hns3_pmu_bdf_range_attrs, |
| }; |
| |
| static struct attribute *hns3_pmu_hw_clk_freq_attrs[] = { |
| &dev_attr_hw_clk_freq.attr, |
| NULL |
| }; |
| |
| static struct attribute_group hns3_pmu_hw_clk_freq_attr_group = { |
| .attrs = hns3_pmu_hw_clk_freq_attrs, |
| }; |
| |
| static const struct attribute_group *hns3_pmu_attr_groups[] = { |
| &hns3_pmu_events_group, |
| &hns3_pmu_filter_mode_group, |
| &hns3_pmu_format_group, |
| &hns3_pmu_cpumask_attr_group, |
| &hns3_pmu_identifier_attr_group, |
| &hns3_pmu_bdf_range_attr_group, |
| &hns3_pmu_hw_clk_freq_attr_group, |
| NULL |
| }; |
| |
| static u32 hns3_pmu_get_event(struct perf_event *event) |
| { |
| return hns3_pmu_get_ext_counter_used(event) << 16 | |
| hns3_pmu_get_event_type(event) << 8 | |
| hns3_pmu_get_subevent(event); |
| } |
| |
| static u32 hns3_pmu_get_real_event(struct perf_event *event) |
| { |
| return hns3_pmu_get_event_type(event) << 8 | |
| hns3_pmu_get_subevent(event); |
| } |
| |
| static u32 hns3_pmu_get_offset(u32 offset, u32 idx) |
| { |
| return offset + HNS3_PMU_REG_EVENT_OFFSET + |
| HNS3_PMU_REG_EVENT_SIZE * idx; |
| } |
| |
| static u32 hns3_pmu_readl(struct hns3_pmu *hns3_pmu, u32 reg_offset, u32 idx) |
| { |
| u32 offset = hns3_pmu_get_offset(reg_offset, idx); |
| |
| return readl(hns3_pmu->base + offset); |
| } |
| |
| static void hns3_pmu_writel(struct hns3_pmu *hns3_pmu, u32 reg_offset, u32 idx, |
| u32 val) |
| { |
| u32 offset = hns3_pmu_get_offset(reg_offset, idx); |
| |
| writel(val, hns3_pmu->base + offset); |
| } |
| |
| static u64 hns3_pmu_readq(struct hns3_pmu *hns3_pmu, u32 reg_offset, u32 idx) |
| { |
| u32 offset = hns3_pmu_get_offset(reg_offset, idx); |
| |
| return readq(hns3_pmu->base + offset); |
| } |
| |
| static void hns3_pmu_writeq(struct hns3_pmu *hns3_pmu, u32 reg_offset, u32 idx, |
| u64 val) |
| { |
| u32 offset = hns3_pmu_get_offset(reg_offset, idx); |
| |
| writeq(val, hns3_pmu->base + offset); |
| } |
| |
| static bool hns3_pmu_cmp_event(struct perf_event *target, |
| struct perf_event *event) |
| { |
| return hns3_pmu_get_real_event(target) == hns3_pmu_get_real_event(event); |
| } |
| |
| static int hns3_pmu_find_related_event_idx(struct hns3_pmu *hns3_pmu, |
| struct perf_event *event) |
| { |
| struct perf_event *sibling; |
| int hw_event_used = 0; |
| int idx; |
| |
| for (idx = 0; idx < HNS3_PMU_MAX_HW_EVENTS; idx++) { |
| sibling = hns3_pmu->hw_events[idx]; |
| if (!sibling) |
| continue; |
| |
| hw_event_used++; |
| |
| if (!hns3_pmu_cmp_event(sibling, event)) |
| continue; |
| |
| /* Related events is used in group */ |
| if (sibling->group_leader == event->group_leader) |
| return idx; |
| } |
| |
| /* No related event and all hardware events are used up */ |
| if (hw_event_used >= HNS3_PMU_MAX_HW_EVENTS) |
| return -EBUSY; |
| |
| /* No related event and there is extra hardware events can be use */ |
| return -ENOENT; |
| } |
| |
| static int hns3_pmu_get_event_idx(struct hns3_pmu *hns3_pmu) |
| { |
| int idx; |
| |
| for (idx = 0; idx < HNS3_PMU_MAX_HW_EVENTS; idx++) { |
| if (!hns3_pmu->hw_events[idx]) |
| return idx; |
| } |
| |
| return -EBUSY; |
| } |
| |
| static bool hns3_pmu_valid_bdf(struct hns3_pmu *hns3_pmu, u16 bdf) |
| { |
| struct pci_dev *pdev; |
| |
| if (bdf < hns3_pmu->bdf_min || bdf > hns3_pmu->bdf_max) { |
| pci_err(hns3_pmu->pdev, "Invalid EP device: %#x!\n", bdf); |
| return false; |
| } |
| |
| pdev = pci_get_domain_bus_and_slot(pci_domain_nr(hns3_pmu->pdev->bus), |
| PCI_BUS_NUM(bdf), |
| GET_PCI_DEVFN(bdf)); |
| if (!pdev) { |
| pci_err(hns3_pmu->pdev, "Nonexistent EP device: %#x!\n", bdf); |
| return false; |
| } |
| |
| pci_dev_put(pdev); |
| return true; |
| } |
| |
| static void hns3_pmu_set_qid_para(struct hns3_pmu *hns3_pmu, u32 idx, u16 bdf, |
| u16 queue) |
| { |
| u32 val; |
| |
| val = GET_PCI_DEVFN(bdf); |
| val |= (u32)queue << HNS3_PMU_QID_PARA_QUEUE_S; |
| hns3_pmu_writel(hns3_pmu, HNS3_PMU_REG_EVENT_QID_PARA, idx, val); |
| } |
| |
| static bool hns3_pmu_qid_req_start(struct hns3_pmu *hns3_pmu, u32 idx) |
| { |
| bool queue_id_valid = false; |
| u32 reg_qid_ctrl, val; |
| int err; |
| |
| /* enable queue id request */ |
| hns3_pmu_writel(hns3_pmu, HNS3_PMU_REG_EVENT_QID_CTRL, idx, |
| HNS3_PMU_QID_CTRL_REQ_ENABLE); |
| |
| reg_qid_ctrl = hns3_pmu_get_offset(HNS3_PMU_REG_EVENT_QID_CTRL, idx); |
| err = readl_poll_timeout(hns3_pmu->base + reg_qid_ctrl, val, |
| val & HNS3_PMU_QID_CTRL_DONE, 1, 1000); |
| if (err == -ETIMEDOUT) { |
| pci_err(hns3_pmu->pdev, "QID request timeout!\n"); |
| goto out; |
| } |
| |
| queue_id_valid = !(val & HNS3_PMU_QID_CTRL_MISS); |
| |
| out: |
| /* disable qid request and clear status */ |
| hns3_pmu_writel(hns3_pmu, HNS3_PMU_REG_EVENT_QID_CTRL, idx, 0); |
| |
| return queue_id_valid; |
| } |
| |
| static bool hns3_pmu_valid_queue(struct hns3_pmu *hns3_pmu, u32 idx, u16 bdf, |
| u16 queue) |
| { |
| hns3_pmu_set_qid_para(hns3_pmu, idx, bdf, queue); |
| |
| return hns3_pmu_qid_req_start(hns3_pmu, idx); |
| } |
| |
| static struct hns3_pmu_event_attr *hns3_pmu_get_pmu_event(u32 event) |
| { |
| struct hns3_pmu_event_attr *pmu_event; |
| struct dev_ext_attribute *eattr; |
| struct device_attribute *dattr; |
| struct attribute *attr; |
| u32 i; |
| |
| for (i = 0; i < ARRAY_SIZE(hns3_pmu_events_attr) - 1; i++) { |
| attr = hns3_pmu_events_attr[i]; |
| dattr = container_of(attr, struct device_attribute, attr); |
| eattr = container_of(dattr, struct dev_ext_attribute, attr); |
| pmu_event = eattr->var; |
| |
| if (event == pmu_event->event) |
| return pmu_event; |
| } |
| |
| return NULL; |
| } |
| |
| static int hns3_pmu_set_func_mode(struct perf_event *event, |
| struct hns3_pmu *hns3_pmu) |
| { |
| struct hw_perf_event *hwc = &event->hw; |
| u16 bdf = hns3_pmu_get_bdf(event); |
| |
| if (!hns3_pmu_valid_bdf(hns3_pmu, bdf)) |
| return -ENOENT; |
| |
| HNS3_PMU_SET_HW_FILTER(hwc, HNS3_PMU_HW_FILTER_FUNC); |
| |
| return 0; |
| } |
| |
| static int hns3_pmu_set_func_queue_mode(struct perf_event *event, |
| struct hns3_pmu *hns3_pmu) |
| { |
| u16 queue_id = hns3_pmu_get_queue(event); |
| struct hw_perf_event *hwc = &event->hw; |
| u16 bdf = hns3_pmu_get_bdf(event); |
| |
| if (!hns3_pmu_valid_bdf(hns3_pmu, bdf)) |
| return -ENOENT; |
| |
| if (!hns3_pmu_valid_queue(hns3_pmu, hwc->idx, bdf, queue_id)) { |
| pci_err(hns3_pmu->pdev, "Invalid queue: %u\n", queue_id); |
| return -ENOENT; |
| } |
| |
| HNS3_PMU_SET_HW_FILTER(hwc, HNS3_PMU_HW_FILTER_FUNC_QUEUE); |
| |
| return 0; |
| } |
| |
| static bool |
| hns3_pmu_is_enabled_global_mode(struct perf_event *event, |
| struct hns3_pmu_event_attr *pmu_event) |
| { |
| u8 global = hns3_pmu_get_global(event); |
| |
| if (!(pmu_event->filter_support & HNS3_PMU_FILTER_SUPPORT_GLOBAL)) |
| return false; |
| |
| return global; |
| } |
| |
| static bool hns3_pmu_is_enabled_func_mode(struct perf_event *event, |
| struct hns3_pmu_event_attr *pmu_event) |
| { |
| u16 queue_id = hns3_pmu_get_queue(event); |
| u16 bdf = hns3_pmu_get_bdf(event); |
| |
| if (!(pmu_event->filter_support & HNS3_PMU_FILTER_SUPPORT_FUNC)) |
| return false; |
| else if (queue_id != HNS3_PMU_FILTER_ALL_QUEUE) |
| return false; |
| |
| return bdf; |
| } |
| |
| static bool |
| hns3_pmu_is_enabled_func_queue_mode(struct perf_event *event, |
| struct hns3_pmu_event_attr *pmu_event) |
| { |
| u16 queue_id = hns3_pmu_get_queue(event); |
| u16 bdf = hns3_pmu_get_bdf(event); |
| |
| if (!(pmu_event->filter_support & HNS3_PMU_FILTER_SUPPORT_FUNC_QUEUE)) |
| return false; |
| else if (queue_id == HNS3_PMU_FILTER_ALL_QUEUE) |
| return false; |
| |
| return bdf; |
| } |
| |
| static bool hns3_pmu_is_enabled_port_mode(struct perf_event *event, |
| struct hns3_pmu_event_attr *pmu_event) |
| { |
| u8 tc_id = hns3_pmu_get_tc(event); |
| |
| if (!(pmu_event->filter_support & HNS3_PMU_FILTER_SUPPORT_PORT)) |
| return false; |
| |
| return tc_id == HNS3_PMU_FILTER_ALL_TC; |
| } |
| |
| static bool |
| hns3_pmu_is_enabled_port_tc_mode(struct perf_event *event, |
| struct hns3_pmu_event_attr *pmu_event) |
| { |
| u8 tc_id = hns3_pmu_get_tc(event); |
| |
| if (!(pmu_event->filter_support & HNS3_PMU_FILTER_SUPPORT_PORT_TC)) |
| return false; |
| |
| return tc_id != HNS3_PMU_FILTER_ALL_TC; |
| } |
| |
| static bool |
| hns3_pmu_is_enabled_func_intr_mode(struct perf_event *event, |
| struct hns3_pmu *hns3_pmu, |
| struct hns3_pmu_event_attr *pmu_event) |
| { |
| u16 bdf = hns3_pmu_get_bdf(event); |
| |
| if (!(pmu_event->filter_support & HNS3_PMU_FILTER_SUPPORT_FUNC_INTR)) |
| return false; |
| |
| return hns3_pmu_valid_bdf(hns3_pmu, bdf); |
| } |
| |
| static int hns3_pmu_select_filter_mode(struct perf_event *event, |
| struct hns3_pmu *hns3_pmu) |
| { |
| u32 event_id = hns3_pmu_get_event(event); |
| struct hw_perf_event *hwc = &event->hw; |
| struct hns3_pmu_event_attr *pmu_event; |
| |
| pmu_event = hns3_pmu_get_pmu_event(event_id); |
| if (!pmu_event) { |
| pci_err(hns3_pmu->pdev, "Invalid pmu event\n"); |
| return -ENOENT; |
| } |
| |
| if (hns3_pmu_is_enabled_global_mode(event, pmu_event)) { |
| HNS3_PMU_SET_HW_FILTER(hwc, HNS3_PMU_HW_FILTER_GLOBAL); |
| return 0; |
| } |
| |
| if (hns3_pmu_is_enabled_func_mode(event, pmu_event)) |
| return hns3_pmu_set_func_mode(event, hns3_pmu); |
| |
| if (hns3_pmu_is_enabled_func_queue_mode(event, pmu_event)) |
| return hns3_pmu_set_func_queue_mode(event, hns3_pmu); |
| |
| if (hns3_pmu_is_enabled_port_mode(event, pmu_event)) { |
| HNS3_PMU_SET_HW_FILTER(hwc, HNS3_PMU_HW_FILTER_PORT); |
| return 0; |
| } |
| |
| if (hns3_pmu_is_enabled_port_tc_mode(event, pmu_event)) { |
| HNS3_PMU_SET_HW_FILTER(hwc, HNS3_PMU_HW_FILTER_PORT_TC); |
| return 0; |
| } |
| |
| if (hns3_pmu_is_enabled_func_intr_mode(event, hns3_pmu, pmu_event)) { |
| HNS3_PMU_SET_HW_FILTER(hwc, HNS3_PMU_HW_FILTER_FUNC_INTR); |
| return 0; |
| } |
| |
| return -ENOENT; |
| } |
| |
| static bool hns3_pmu_validate_event_group(struct perf_event *event) |
| { |
| struct perf_event *sibling, *leader = event->group_leader; |
| struct perf_event *event_group[HNS3_PMU_MAX_HW_EVENTS]; |
| int counters = 1; |
| int num; |
| |
| event_group[0] = leader; |
| if (!is_software_event(leader)) { |
| if (leader->pmu != event->pmu) |
| return false; |
| |
| if (leader != event && !hns3_pmu_cmp_event(leader, event)) |
| event_group[counters++] = event; |
| } |
| |
| for_each_sibling_event(sibling, event->group_leader) { |
| if (is_software_event(sibling)) |
| continue; |
| |
| if (sibling->pmu != event->pmu) |
| return false; |
| |
| for (num = 0; num < counters; num++) { |
| /* |
| * If we find a related event, then it's a valid group |
| * since we don't need to allocate a new counter for it. |
| */ |
| if (hns3_pmu_cmp_event(event_group[num], sibling)) |
| break; |
| } |
| |
| /* |
| * Otherwise it's a new event but if there's no available counter, |
| * fail the check since we cannot schedule all the events in |
| * the group simultaneously. |
| */ |
| if (num == HNS3_PMU_MAX_HW_EVENTS) |
| return false; |
| |
| if (num == counters) |
| event_group[counters++] = sibling; |
| } |
| |
| return true; |
| } |
| |
| static u32 hns3_pmu_get_filter_condition(struct perf_event *event) |
| { |
| struct hw_perf_event *hwc = &event->hw; |
| u16 intr_id = hns3_pmu_get_intr(event); |
| u8 port_id = hns3_pmu_get_port(event); |
| u16 bdf = hns3_pmu_get_bdf(event); |
| u8 tc_id = hns3_pmu_get_tc(event); |
| u8 filter_mode; |
| |
| filter_mode = *(u8 *)hwc->addr_filters; |
| switch (filter_mode) { |
| case HNS3_PMU_HW_FILTER_PORT: |
| return FILTER_CONDITION_PORT(port_id); |
| case HNS3_PMU_HW_FILTER_PORT_TC: |
| return FILTER_CONDITION_PORT_TC(port_id, tc_id); |
| case HNS3_PMU_HW_FILTER_FUNC: |
| case HNS3_PMU_HW_FILTER_FUNC_QUEUE: |
| return GET_PCI_DEVFN(bdf); |
| case HNS3_PMU_HW_FILTER_FUNC_INTR: |
| return FILTER_CONDITION_FUNC_INTR(GET_PCI_DEVFN(bdf), intr_id); |
| default: |
| break; |
| } |
| |
| return 0; |
| } |
| |
| static void hns3_pmu_config_filter(struct perf_event *event) |
| { |
| struct hns3_pmu *hns3_pmu = to_hns3_pmu(event->pmu); |
| u8 event_type = hns3_pmu_get_event_type(event); |
| u8 subevent_id = hns3_pmu_get_subevent(event); |
| u16 queue_id = hns3_pmu_get_queue(event); |
| struct hw_perf_event *hwc = &event->hw; |
| u8 filter_mode = *(u8 *)hwc->addr_filters; |
| u16 bdf = hns3_pmu_get_bdf(event); |
| u32 idx = hwc->idx; |
| u32 val; |
| |
| val = event_type; |
| val |= subevent_id << HNS3_PMU_CTRL_SUBEVENT_S; |
| val |= filter_mode << HNS3_PMU_CTRL_FILTER_MODE_S; |
| val |= HNS3_PMU_EVENT_OVERFLOW_RESTART; |
| hns3_pmu_writel(hns3_pmu, HNS3_PMU_REG_EVENT_CTRL_LOW, idx, val); |
| |
| val = hns3_pmu_get_filter_condition(event); |
| hns3_pmu_writel(hns3_pmu, HNS3_PMU_REG_EVENT_CTRL_HIGH, idx, val); |
| |
| if (filter_mode == HNS3_PMU_HW_FILTER_FUNC_QUEUE) |
| hns3_pmu_set_qid_para(hns3_pmu, idx, bdf, queue_id); |
| } |
| |
| static void hns3_pmu_enable_counter(struct hns3_pmu *hns3_pmu, |
| struct hw_perf_event *hwc) |
| { |
| u32 idx = hwc->idx; |
| u32 val; |
| |
| val = hns3_pmu_readl(hns3_pmu, HNS3_PMU_REG_EVENT_CTRL_LOW, idx); |
| val |= HNS3_PMU_EVENT_EN; |
| hns3_pmu_writel(hns3_pmu, HNS3_PMU_REG_EVENT_CTRL_LOW, idx, val); |
| } |
| |
| static void hns3_pmu_disable_counter(struct hns3_pmu *hns3_pmu, |
| struct hw_perf_event *hwc) |
| { |
| u32 idx = hwc->idx; |
| u32 val; |
| |
| val = hns3_pmu_readl(hns3_pmu, HNS3_PMU_REG_EVENT_CTRL_LOW, idx); |
| val &= ~HNS3_PMU_EVENT_EN; |
| hns3_pmu_writel(hns3_pmu, HNS3_PMU_REG_EVENT_CTRL_LOW, idx, val); |
| } |
| |
| static void hns3_pmu_enable_intr(struct hns3_pmu *hns3_pmu, |
| struct hw_perf_event *hwc) |
| { |
| u32 idx = hwc->idx; |
| u32 val; |
| |
| val = hns3_pmu_readl(hns3_pmu, HNS3_PMU_REG_EVENT_INTR_MASK, idx); |
| val &= ~HNS3_PMU_INTR_MASK_OVERFLOW; |
| hns3_pmu_writel(hns3_pmu, HNS3_PMU_REG_EVENT_INTR_MASK, idx, val); |
| } |
| |
| static void hns3_pmu_disable_intr(struct hns3_pmu *hns3_pmu, |
| struct hw_perf_event *hwc) |
| { |
| u32 idx = hwc->idx; |
| u32 val; |
| |
| val = hns3_pmu_readl(hns3_pmu, HNS3_PMU_REG_EVENT_INTR_MASK, idx); |
| val |= HNS3_PMU_INTR_MASK_OVERFLOW; |
| hns3_pmu_writel(hns3_pmu, HNS3_PMU_REG_EVENT_INTR_MASK, idx, val); |
| } |
| |
| static void hns3_pmu_clear_intr_status(struct hns3_pmu *hns3_pmu, u32 idx) |
| { |
| u32 val; |
| |
| val = hns3_pmu_readl(hns3_pmu, HNS3_PMU_REG_EVENT_CTRL_LOW, idx); |
| val |= HNS3_PMU_EVENT_STATUS_RESET; |
| hns3_pmu_writel(hns3_pmu, HNS3_PMU_REG_EVENT_CTRL_LOW, idx, val); |
| |
| val = hns3_pmu_readl(hns3_pmu, HNS3_PMU_REG_EVENT_CTRL_LOW, idx); |
| val &= ~HNS3_PMU_EVENT_STATUS_RESET; |
| hns3_pmu_writel(hns3_pmu, HNS3_PMU_REG_EVENT_CTRL_LOW, idx, val); |
| } |
| |
| static u64 hns3_pmu_read_counter(struct perf_event *event) |
| { |
| struct hns3_pmu *hns3_pmu = to_hns3_pmu(event->pmu); |
| |
| return hns3_pmu_readq(hns3_pmu, event->hw.event_base, event->hw.idx); |
| } |
| |
| static void hns3_pmu_write_counter(struct perf_event *event, u64 value) |
| { |
| struct hns3_pmu *hns3_pmu = to_hns3_pmu(event->pmu); |
| u32 idx = event->hw.idx; |
| |
| hns3_pmu_writeq(hns3_pmu, HNS3_PMU_REG_EVENT_COUNTER, idx, value); |
| hns3_pmu_writeq(hns3_pmu, HNS3_PMU_REG_EVENT_EXT_COUNTER, idx, value); |
| } |
| |
| static void hns3_pmu_init_counter(struct perf_event *event) |
| { |
| struct hw_perf_event *hwc = &event->hw; |
| |
| local64_set(&hwc->prev_count, 0); |
| hns3_pmu_write_counter(event, 0); |
| } |
| |
| static int hns3_pmu_event_init(struct perf_event *event) |
| { |
| struct hns3_pmu *hns3_pmu = to_hns3_pmu(event->pmu); |
| struct hw_perf_event *hwc = &event->hw; |
| int idx; |
| int ret; |
| |
| if (event->attr.type != event->pmu->type) |
| return -ENOENT; |
| |
| /* Sampling is not supported */ |
| if (is_sampling_event(event) || event->attach_state & PERF_ATTACH_TASK) |
| return -EOPNOTSUPP; |
| |
| event->cpu = hns3_pmu->on_cpu; |
| |
| idx = hns3_pmu_get_event_idx(hns3_pmu); |
| if (idx < 0) { |
| pci_err(hns3_pmu->pdev, "Up to %u events are supported!\n", |
| HNS3_PMU_MAX_HW_EVENTS); |
| return -EBUSY; |
| } |
| |
| hwc->idx = idx; |
| |
| ret = hns3_pmu_select_filter_mode(event, hns3_pmu); |
| if (ret) { |
| pci_err(hns3_pmu->pdev, "Invalid filter, ret = %d.\n", ret); |
| return ret; |
| } |
| |
| if (!hns3_pmu_validate_event_group(event)) { |
| pci_err(hns3_pmu->pdev, "Invalid event group.\n"); |
| return -EINVAL; |
| } |
| |
| if (hns3_pmu_get_ext_counter_used(event)) |
| hwc->event_base = HNS3_PMU_REG_EVENT_EXT_COUNTER; |
| else |
| hwc->event_base = HNS3_PMU_REG_EVENT_COUNTER; |
| |
| return 0; |
| } |
| |
| static void hns3_pmu_read(struct perf_event *event) |
| { |
| struct hw_perf_event *hwc = &event->hw; |
| u64 new_cnt, prev_cnt, delta; |
| |
| do { |
| prev_cnt = local64_read(&hwc->prev_count); |
| new_cnt = hns3_pmu_read_counter(event); |
| } while (local64_cmpxchg(&hwc->prev_count, prev_cnt, new_cnt) != |
| prev_cnt); |
| |
| delta = new_cnt - prev_cnt; |
| local64_add(delta, &event->count); |
| } |
| |
| static void hns3_pmu_start(struct perf_event *event, int flags) |
| { |
| struct hns3_pmu *hns3_pmu = to_hns3_pmu(event->pmu); |
| struct hw_perf_event *hwc = &event->hw; |
| |
| if (WARN_ON_ONCE(!(hwc->state & PERF_HES_STOPPED))) |
| return; |
| |
| WARN_ON_ONCE(!(hwc->state & PERF_HES_UPTODATE)); |
| hwc->state = 0; |
| |
| hns3_pmu_config_filter(event); |
| hns3_pmu_init_counter(event); |
| hns3_pmu_enable_intr(hns3_pmu, hwc); |
| hns3_pmu_enable_counter(hns3_pmu, hwc); |
| |
| perf_event_update_userpage(event); |
| } |
| |
| static void hns3_pmu_stop(struct perf_event *event, int flags) |
| { |
| struct hns3_pmu *hns3_pmu = to_hns3_pmu(event->pmu); |
| struct hw_perf_event *hwc = &event->hw; |
| |
| hns3_pmu_disable_counter(hns3_pmu, hwc); |
| hns3_pmu_disable_intr(hns3_pmu, hwc); |
| |
| WARN_ON_ONCE(hwc->state & PERF_HES_STOPPED); |
| hwc->state |= PERF_HES_STOPPED; |
| |
| if (hwc->state & PERF_HES_UPTODATE) |
| return; |
| |
| /* Read hardware counter and update the perf counter statistics */ |
| hns3_pmu_read(event); |
| hwc->state |= PERF_HES_UPTODATE; |
| } |
| |
| static int hns3_pmu_add(struct perf_event *event, int flags) |
| { |
| struct hns3_pmu *hns3_pmu = to_hns3_pmu(event->pmu); |
| struct hw_perf_event *hwc = &event->hw; |
| int idx; |
| |
| hwc->state = PERF_HES_STOPPED | PERF_HES_UPTODATE; |
| |
| /* Check all working events to find a related event. */ |
| idx = hns3_pmu_find_related_event_idx(hns3_pmu, event); |
| if (idx < 0 && idx != -ENOENT) |
| return idx; |
| |
| /* Current event shares an enabled hardware event with related event */ |
| if (idx >= 0 && idx < HNS3_PMU_MAX_HW_EVENTS) { |
| hwc->idx = idx; |
| goto start_count; |
| } |
| |
| idx = hns3_pmu_get_event_idx(hns3_pmu); |
| if (idx < 0) |
| return idx; |
| |
| hwc->idx = idx; |
| hns3_pmu->hw_events[idx] = event; |
| |
| start_count: |
| if (flags & PERF_EF_START) |
| hns3_pmu_start(event, PERF_EF_RELOAD); |
| |
| return 0; |
| } |
| |
| static void hns3_pmu_del(struct perf_event *event, int flags) |
| { |
| struct hns3_pmu *hns3_pmu = to_hns3_pmu(event->pmu); |
| struct hw_perf_event *hwc = &event->hw; |
| |
| hns3_pmu_stop(event, PERF_EF_UPDATE); |
| hns3_pmu->hw_events[hwc->idx] = NULL; |
| perf_event_update_userpage(event); |
| } |
| |
| static void hns3_pmu_enable(struct pmu *pmu) |
| { |
| struct hns3_pmu *hns3_pmu = to_hns3_pmu(pmu); |
| u32 val; |
| |
| val = readl(hns3_pmu->base + HNS3_PMU_REG_GLOBAL_CTRL); |
| val |= HNS3_PMU_GLOBAL_START; |
| writel(val, hns3_pmu->base + HNS3_PMU_REG_GLOBAL_CTRL); |
| } |
| |
| static void hns3_pmu_disable(struct pmu *pmu) |
| { |
| struct hns3_pmu *hns3_pmu = to_hns3_pmu(pmu); |
| u32 val; |
| |
| val = readl(hns3_pmu->base + HNS3_PMU_REG_GLOBAL_CTRL); |
| val &= ~HNS3_PMU_GLOBAL_START; |
| writel(val, hns3_pmu->base + HNS3_PMU_REG_GLOBAL_CTRL); |
| } |
| |
| static int hns3_pmu_alloc_pmu(struct pci_dev *pdev, struct hns3_pmu *hns3_pmu) |
| { |
| u16 device_id; |
| char *name; |
| u32 val; |
| |
| hns3_pmu->base = pcim_iomap_table(pdev)[BAR_2]; |
| if (!hns3_pmu->base) { |
| pci_err(pdev, "ioremap failed\n"); |
| return -ENOMEM; |
| } |
| |
| hns3_pmu->hw_clk_freq = readl(hns3_pmu->base + HNS3_PMU_REG_CLOCK_FREQ); |
| |
| val = readl(hns3_pmu->base + HNS3_PMU_REG_BDF); |
| hns3_pmu->bdf_min = val & 0xffff; |
| hns3_pmu->bdf_max = val >> 16; |
| |
| val = readl(hns3_pmu->base + HNS3_PMU_REG_DEVICE_ID); |
| device_id = val & 0xffff; |
| name = devm_kasprintf(&pdev->dev, GFP_KERNEL, "hns3_pmu_sicl_%u", device_id); |
| if (!name) |
| return -ENOMEM; |
| |
| hns3_pmu->pdev = pdev; |
| hns3_pmu->on_cpu = -1; |
| hns3_pmu->identifier = readl(hns3_pmu->base + HNS3_PMU_REG_VERSION); |
| hns3_pmu->pmu = (struct pmu) { |
| .name = name, |
| .module = THIS_MODULE, |
| .parent = &pdev->dev, |
| .event_init = hns3_pmu_event_init, |
| .pmu_enable = hns3_pmu_enable, |
| .pmu_disable = hns3_pmu_disable, |
| .add = hns3_pmu_add, |
| .del = hns3_pmu_del, |
| .start = hns3_pmu_start, |
| .stop = hns3_pmu_stop, |
| .read = hns3_pmu_read, |
| .task_ctx_nr = perf_invalid_context, |
| .attr_groups = hns3_pmu_attr_groups, |
| .capabilities = PERF_PMU_CAP_NO_EXCLUDE, |
| }; |
| |
| return 0; |
| } |
| |
| static irqreturn_t hns3_pmu_irq(int irq, void *data) |
| { |
| struct hns3_pmu *hns3_pmu = data; |
| u32 intr_status, idx; |
| |
| for (idx = 0; idx < HNS3_PMU_MAX_HW_EVENTS; idx++) { |
| intr_status = hns3_pmu_readl(hns3_pmu, |
| HNS3_PMU_REG_EVENT_INTR_STATUS, |
| idx); |
| |
| /* |
| * As each counter will restart from 0 when it is overflowed, |
| * extra processing is no need, just clear interrupt status. |
| */ |
| if (intr_status) |
| hns3_pmu_clear_intr_status(hns3_pmu, idx); |
| } |
| |
| return IRQ_HANDLED; |
| } |
| |
| static int hns3_pmu_online_cpu(unsigned int cpu, struct hlist_node *node) |
| { |
| struct hns3_pmu *hns3_pmu; |
| |
| hns3_pmu = hlist_entry_safe(node, struct hns3_pmu, node); |
| if (!hns3_pmu) |
| return -ENODEV; |
| |
| if (hns3_pmu->on_cpu == -1) { |
| hns3_pmu->on_cpu = cpu; |
| irq_set_affinity(hns3_pmu->irq, cpumask_of(cpu)); |
| } |
| |
| return 0; |
| } |
| |
| static int hns3_pmu_offline_cpu(unsigned int cpu, struct hlist_node *node) |
| { |
| struct hns3_pmu *hns3_pmu; |
| unsigned int target; |
| |
| hns3_pmu = hlist_entry_safe(node, struct hns3_pmu, node); |
| if (!hns3_pmu) |
| return -ENODEV; |
| |
| /* Nothing to do if this CPU doesn't own the PMU */ |
| if (hns3_pmu->on_cpu != cpu) |
| return 0; |
| |
| /* Choose a new CPU from all online cpus */ |
| target = cpumask_any_but(cpu_online_mask, cpu); |
| if (target >= nr_cpu_ids) |
| return 0; |
| |
| perf_pmu_migrate_context(&hns3_pmu->pmu, cpu, target); |
| hns3_pmu->on_cpu = target; |
| irq_set_affinity(hns3_pmu->irq, cpumask_of(target)); |
| |
| return 0; |
| } |
| |
| static void hns3_pmu_free_irq(void *data) |
| { |
| struct pci_dev *pdev = data; |
| |
| pci_free_irq_vectors(pdev); |
| } |
| |
| static int hns3_pmu_irq_register(struct pci_dev *pdev, |
| struct hns3_pmu *hns3_pmu) |
| { |
| int irq, ret; |
| |
| ret = pci_alloc_irq_vectors(pdev, 1, 1, PCI_IRQ_MSI); |
| if (ret < 0) { |
| pci_err(pdev, "failed to enable MSI vectors, ret = %d.\n", ret); |
| return ret; |
| } |
| |
| ret = devm_add_action_or_reset(&pdev->dev, hns3_pmu_free_irq, pdev); |
| if (ret) { |
| pci_err(pdev, "failed to add free irq action, ret = %d.\n", ret); |
| return ret; |
| } |
| |
| irq = pci_irq_vector(pdev, 0); |
| ret = devm_request_irq(&pdev->dev, irq, hns3_pmu_irq, 0, |
| hns3_pmu->pmu.name, hns3_pmu); |
| if (ret) { |
| pci_err(pdev, "failed to register irq, ret = %d.\n", ret); |
| return ret; |
| } |
| |
| hns3_pmu->irq = irq; |
| |
| return 0; |
| } |
| |
| static int hns3_pmu_init_pmu(struct pci_dev *pdev, struct hns3_pmu *hns3_pmu) |
| { |
| int ret; |
| |
| ret = hns3_pmu_alloc_pmu(pdev, hns3_pmu); |
| if (ret) |
| return ret; |
| |
| ret = hns3_pmu_irq_register(pdev, hns3_pmu); |
| if (ret) |
| return ret; |
| |
| ret = cpuhp_state_add_instance(CPUHP_AP_PERF_ARM_HNS3_PMU_ONLINE, |
| &hns3_pmu->node); |
| if (ret) { |
| pci_err(pdev, "failed to register hotplug, ret = %d.\n", ret); |
| return ret; |
| } |
| |
| ret = perf_pmu_register(&hns3_pmu->pmu, hns3_pmu->pmu.name, -1); |
| if (ret) { |
| pci_err(pdev, "failed to register perf PMU, ret = %d.\n", ret); |
| cpuhp_state_remove_instance_nocalls(CPUHP_AP_PERF_ARM_HNS3_PMU_ONLINE, |
| &hns3_pmu->node); |
| } |
| |
| return ret; |
| } |
| |
| static void hns3_pmu_uninit_pmu(struct pci_dev *pdev) |
| { |
| struct hns3_pmu *hns3_pmu = pci_get_drvdata(pdev); |
| |
| perf_pmu_unregister(&hns3_pmu->pmu); |
| cpuhp_state_remove_instance_nocalls(CPUHP_AP_PERF_ARM_HNS3_PMU_ONLINE, |
| &hns3_pmu->node); |
| } |
| |
| static int hns3_pmu_init_dev(struct pci_dev *pdev) |
| { |
| int ret; |
| |
| ret = pcim_enable_device(pdev); |
| if (ret) { |
| pci_err(pdev, "failed to enable pci device, ret = %d.\n", ret); |
| return ret; |
| } |
| |
| ret = pcim_iomap_regions(pdev, BIT(BAR_2), "hns3_pmu"); |
| if (ret < 0) { |
| pci_err(pdev, "failed to request pci region, ret = %d.\n", ret); |
| return ret; |
| } |
| |
| pci_set_master(pdev); |
| |
| return 0; |
| } |
| |
| static int hns3_pmu_probe(struct pci_dev *pdev, const struct pci_device_id *id) |
| { |
| struct hns3_pmu *hns3_pmu; |
| int ret; |
| |
| hns3_pmu = devm_kzalloc(&pdev->dev, sizeof(*hns3_pmu), GFP_KERNEL); |
| if (!hns3_pmu) |
| return -ENOMEM; |
| |
| ret = hns3_pmu_init_dev(pdev); |
| if (ret) |
| return ret; |
| |
| ret = hns3_pmu_init_pmu(pdev, hns3_pmu); |
| if (ret) { |
| pci_clear_master(pdev); |
| return ret; |
| } |
| |
| pci_set_drvdata(pdev, hns3_pmu); |
| |
| return ret; |
| } |
| |
| static void hns3_pmu_remove(struct pci_dev *pdev) |
| { |
| hns3_pmu_uninit_pmu(pdev); |
| pci_clear_master(pdev); |
| pci_set_drvdata(pdev, NULL); |
| } |
| |
| static const struct pci_device_id hns3_pmu_ids[] = { |
| { PCI_DEVICE(PCI_VENDOR_ID_HUAWEI, 0xa22b) }, |
| { 0, } |
| }; |
| MODULE_DEVICE_TABLE(pci, hns3_pmu_ids); |
| |
| static struct pci_driver hns3_pmu_driver = { |
| .name = "hns3_pmu", |
| .id_table = hns3_pmu_ids, |
| .probe = hns3_pmu_probe, |
| .remove = hns3_pmu_remove, |
| }; |
| |
| static int __init hns3_pmu_module_init(void) |
| { |
| int ret; |
| |
| ret = cpuhp_setup_state_multi(CPUHP_AP_PERF_ARM_HNS3_PMU_ONLINE, |
| "AP_PERF_ARM_HNS3_PMU_ONLINE", |
| hns3_pmu_online_cpu, |
| hns3_pmu_offline_cpu); |
| if (ret) { |
| pr_err("failed to setup HNS3 PMU hotplug, ret = %d.\n", ret); |
| return ret; |
| } |
| |
| ret = pci_register_driver(&hns3_pmu_driver); |
| if (ret) { |
| pr_err("failed to register pci driver, ret = %d.\n", ret); |
| cpuhp_remove_multi_state(CPUHP_AP_PERF_ARM_HNS3_PMU_ONLINE); |
| } |
| |
| return ret; |
| } |
| module_init(hns3_pmu_module_init); |
| |
| static void __exit hns3_pmu_module_exit(void) |
| { |
| pci_unregister_driver(&hns3_pmu_driver); |
| cpuhp_remove_multi_state(CPUHP_AP_PERF_ARM_HNS3_PMU_ONLINE); |
| } |
| module_exit(hns3_pmu_module_exit); |
| |
| MODULE_DESCRIPTION("HNS3 PMU driver"); |
| MODULE_LICENSE("GPL v2"); |