| // SPDX-License-Identifier: GPL-2.0 |
| /* Marvell Octeon EP (EndPoint) VF Ethernet Driver |
| * |
| * Copyright (C) 2020 Marvell. |
| * |
| */ |
| |
| #include <linux/pci.h> |
| #include <linux/netdevice.h> |
| #include <linux/etherdevice.h> |
| |
| #include "octep_vf_config.h" |
| #include "octep_vf_main.h" |
| #include "octep_vf_regs_cn9k.h" |
| |
| /* Dump useful hardware IQ/OQ CSRs for debug purpose */ |
| static void cn93_vf_dump_q_regs(struct octep_vf_device *oct, int qno) |
| { |
| struct device *dev = &oct->pdev->dev; |
| |
| dev_info(dev, "IQ-%d register dump\n", qno); |
| dev_info(dev, "R[%d]_IN_INSTR_DBELL[0x%llx]: 0x%016llx\n", |
| qno, CN93_VF_SDP_R_IN_INSTR_DBELL(qno), |
| octep_vf_read_csr64(oct, CN93_VF_SDP_R_IN_INSTR_DBELL(qno))); |
| dev_info(dev, "R[%d]_IN_CONTROL[0x%llx]: 0x%016llx\n", |
| qno, CN93_VF_SDP_R_IN_CONTROL(qno), |
| octep_vf_read_csr64(oct, CN93_VF_SDP_R_IN_CONTROL(qno))); |
| dev_info(dev, "R[%d]_IN_ENABLE[0x%llx]: 0x%016llx\n", |
| qno, CN93_VF_SDP_R_IN_ENABLE(qno), |
| octep_vf_read_csr64(oct, CN93_VF_SDP_R_IN_ENABLE(qno))); |
| dev_info(dev, "R[%d]_IN_INSTR_BADDR[0x%llx]: 0x%016llx\n", |
| qno, CN93_VF_SDP_R_IN_INSTR_BADDR(qno), |
| octep_vf_read_csr64(oct, CN93_VF_SDP_R_IN_INSTR_BADDR(qno))); |
| dev_info(dev, "R[%d]_IN_INSTR_RSIZE[0x%llx]: 0x%016llx\n", |
| qno, CN93_VF_SDP_R_IN_INSTR_RSIZE(qno), |
| octep_vf_read_csr64(oct, CN93_VF_SDP_R_IN_INSTR_RSIZE(qno))); |
| dev_info(dev, "R[%d]_IN_CNTS[0x%llx]: 0x%016llx\n", |
| qno, CN93_VF_SDP_R_IN_CNTS(qno), |
| octep_vf_read_csr64(oct, CN93_VF_SDP_R_IN_CNTS(qno))); |
| dev_info(dev, "R[%d]_IN_INT_LEVELS[0x%llx]: 0x%016llx\n", |
| qno, CN93_VF_SDP_R_IN_INT_LEVELS(qno), |
| octep_vf_read_csr64(oct, CN93_VF_SDP_R_IN_INT_LEVELS(qno))); |
| dev_info(dev, "R[%d]_IN_PKT_CNT[0x%llx]: 0x%016llx\n", |
| qno, CN93_VF_SDP_R_IN_PKT_CNT(qno), |
| octep_vf_read_csr64(oct, CN93_VF_SDP_R_IN_PKT_CNT(qno))); |
| dev_info(dev, "R[%d]_IN_BYTE_CNT[0x%llx]: 0x%016llx\n", |
| qno, CN93_VF_SDP_R_IN_BYTE_CNT(qno), |
| octep_vf_read_csr64(oct, CN93_VF_SDP_R_IN_BYTE_CNT(qno))); |
| |
| dev_info(dev, "OQ-%d register dump\n", qno); |
| dev_info(dev, "R[%d]_OUT_SLIST_DBELL[0x%llx]: 0x%016llx\n", |
| qno, CN93_VF_SDP_R_OUT_SLIST_DBELL(qno), |
| octep_vf_read_csr64(oct, CN93_VF_SDP_R_OUT_SLIST_DBELL(qno))); |
| dev_info(dev, "R[%d]_OUT_CONTROL[0x%llx]: 0x%016llx\n", |
| qno, CN93_VF_SDP_R_OUT_CONTROL(qno), |
| octep_vf_read_csr64(oct, CN93_VF_SDP_R_OUT_CONTROL(qno))); |
| dev_info(dev, "R[%d]_OUT_ENABLE[0x%llx]: 0x%016llx\n", |
| qno, CN93_VF_SDP_R_OUT_ENABLE(qno), |
| octep_vf_read_csr64(oct, CN93_VF_SDP_R_OUT_ENABLE(qno))); |
| dev_info(dev, "R[%d]_OUT_SLIST_BADDR[0x%llx]: 0x%016llx\n", |
| qno, CN93_VF_SDP_R_OUT_SLIST_BADDR(qno), |
| octep_vf_read_csr64(oct, CN93_VF_SDP_R_OUT_SLIST_BADDR(qno))); |
| dev_info(dev, "R[%d]_OUT_SLIST_RSIZE[0x%llx]: 0x%016llx\n", |
| qno, CN93_VF_SDP_R_OUT_SLIST_RSIZE(qno), |
| octep_vf_read_csr64(oct, CN93_VF_SDP_R_OUT_SLIST_RSIZE(qno))); |
| dev_info(dev, "R[%d]_OUT_CNTS[0x%llx]: 0x%016llx\n", |
| qno, CN93_VF_SDP_R_OUT_CNTS(qno), |
| octep_vf_read_csr64(oct, CN93_VF_SDP_R_OUT_CNTS(qno))); |
| dev_info(dev, "R[%d]_OUT_INT_LEVELS[0x%llx]: 0x%016llx\n", |
| qno, CN93_VF_SDP_R_OUT_INT_LEVELS(qno), |
| octep_vf_read_csr64(oct, CN93_VF_SDP_R_OUT_INT_LEVELS(qno))); |
| dev_info(dev, "R[%d]_OUT_PKT_CNT[0x%llx]: 0x%016llx\n", |
| qno, CN93_VF_SDP_R_OUT_PKT_CNT(qno), |
| octep_vf_read_csr64(oct, CN93_VF_SDP_R_OUT_PKT_CNT(qno))); |
| dev_info(dev, "R[%d]_OUT_BYTE_CNT[0x%llx]: 0x%016llx\n", |
| qno, CN93_VF_SDP_R_OUT_BYTE_CNT(qno), |
| octep_vf_read_csr64(oct, CN93_VF_SDP_R_OUT_BYTE_CNT(qno))); |
| } |
| |
| /* Reset Hardware Tx queue */ |
| static void cn93_vf_reset_iq(struct octep_vf_device *oct, int q_no) |
| { |
| u64 val = ULL(0); |
| |
| dev_dbg(&oct->pdev->dev, "Reset VF IQ-%d\n", q_no); |
| |
| /* Disable the Tx/Instruction Ring */ |
| octep_vf_write_csr64(oct, CN93_VF_SDP_R_IN_ENABLE(q_no), val); |
| |
| /* clear the Instruction Ring packet/byte counts and doorbell CSRs */ |
| octep_vf_write_csr64(oct, CN93_VF_SDP_R_IN_INT_LEVELS(q_no), val); |
| octep_vf_write_csr64(oct, CN93_VF_SDP_R_IN_PKT_CNT(q_no), val); |
| octep_vf_write_csr64(oct, CN93_VF_SDP_R_IN_BYTE_CNT(q_no), val); |
| octep_vf_write_csr64(oct, CN93_VF_SDP_R_IN_INSTR_BADDR(q_no), val); |
| octep_vf_write_csr64(oct, CN93_VF_SDP_R_IN_INSTR_RSIZE(q_no), val); |
| |
| val = GENMASK_ULL(31, 0); |
| octep_vf_write_csr64(oct, CN93_VF_SDP_R_IN_INSTR_DBELL(q_no), val); |
| |
| val = octep_vf_read_csr64(oct, CN93_VF_SDP_R_IN_CNTS(q_no)); |
| octep_vf_write_csr64(oct, CN93_VF_SDP_R_IN_CNTS(q_no), |
| val & GENMASK_ULL(31, 0)); |
| } |
| |
| /* Reset Hardware Rx queue */ |
| static void cn93_vf_reset_oq(struct octep_vf_device *oct, int q_no) |
| { |
| u64 val = ULL(0); |
| |
| /* Disable Output (Rx) Ring */ |
| octep_vf_write_csr64(oct, CN93_VF_SDP_R_OUT_ENABLE(q_no), val); |
| |
| /* Clear count CSRs */ |
| val = octep_vf_read_csr(oct, CN93_VF_SDP_R_OUT_CNTS(q_no)); |
| octep_vf_write_csr(oct, CN93_VF_SDP_R_OUT_CNTS(q_no), val); |
| |
| octep_vf_write_csr64(oct, CN93_VF_SDP_R_OUT_PKT_CNT(q_no), GENMASK_ULL(35, 0)); |
| octep_vf_write_csr64(oct, CN93_VF_SDP_R_OUT_SLIST_DBELL(q_no), GENMASK_ULL(31, 0)); |
| } |
| |
| /* Reset all hardware Tx/Rx queues */ |
| static void octep_vf_reset_io_queues_cn93(struct octep_vf_device *oct) |
| { |
| struct pci_dev *pdev = oct->pdev; |
| int q; |
| |
| dev_dbg(&pdev->dev, "Reset OCTEP_CN93 VF IO Queues\n"); |
| |
| for (q = 0; q < CFG_GET_PORTS_ACTIVE_IO_RINGS(oct->conf); q++) { |
| cn93_vf_reset_iq(oct, q); |
| cn93_vf_reset_oq(oct, q); |
| } |
| } |
| |
| /* Initialize configuration limits and initial active config */ |
| static void octep_vf_init_config_cn93_vf(struct octep_vf_device *oct) |
| { |
| struct octep_vf_config *conf = oct->conf; |
| u64 reg_val; |
| |
| reg_val = octep_vf_read_csr64(oct, CN93_VF_SDP_R_IN_CONTROL(0)); |
| conf->ring_cfg.max_io_rings = (reg_val >> CN93_VF_R_IN_CTL_RPVF_POS) & |
| CN93_VF_R_IN_CTL_RPVF_MASK; |
| conf->ring_cfg.active_io_rings = conf->ring_cfg.max_io_rings; |
| |
| conf->iq.num_descs = OCTEP_VF_IQ_MAX_DESCRIPTORS; |
| conf->iq.instr_type = OCTEP_VF_64BYTE_INSTR; |
| conf->iq.db_min = OCTEP_VF_DB_MIN; |
| conf->iq.intr_threshold = OCTEP_VF_IQ_INTR_THRESHOLD; |
| |
| conf->oq.num_descs = OCTEP_VF_OQ_MAX_DESCRIPTORS; |
| conf->oq.buf_size = OCTEP_VF_OQ_BUF_SIZE; |
| conf->oq.refill_threshold = OCTEP_VF_OQ_REFILL_THRESHOLD; |
| conf->oq.oq_intr_pkt = OCTEP_VF_OQ_INTR_PKT_THRESHOLD; |
| conf->oq.oq_intr_time = OCTEP_VF_OQ_INTR_TIME_THRESHOLD; |
| |
| conf->msix_cfg.ioq_msix = conf->ring_cfg.active_io_rings; |
| } |
| |
| /* Setup registers for a hardware Tx Queue */ |
| static void octep_vf_setup_iq_regs_cn93(struct octep_vf_device *oct, int iq_no) |
| { |
| struct octep_vf_iq *iq = oct->iq[iq_no]; |
| u32 reset_instr_cnt; |
| u64 reg_val; |
| |
| reg_val = octep_vf_read_csr64(oct, CN93_VF_SDP_R_IN_CONTROL(iq_no)); |
| |
| /* wait for IDLE to set to 1 */ |
| if (!(reg_val & CN93_VF_R_IN_CTL_IDLE)) { |
| do { |
| reg_val = octep_vf_read_csr64(oct, CN93_VF_SDP_R_IN_CONTROL(iq_no)); |
| } while (!(reg_val & CN93_VF_R_IN_CTL_IDLE)); |
| } |
| reg_val |= CN93_VF_R_IN_CTL_RDSIZE; |
| reg_val |= CN93_VF_R_IN_CTL_IS_64B; |
| reg_val |= CN93_VF_R_IN_CTL_ESR; |
| octep_vf_write_csr64(oct, CN93_VF_SDP_R_IN_CONTROL(iq_no), reg_val); |
| |
| /* Write the start of the input queue's ring and its size */ |
| octep_vf_write_csr64(oct, CN93_VF_SDP_R_IN_INSTR_BADDR(iq_no), iq->desc_ring_dma); |
| octep_vf_write_csr64(oct, CN93_VF_SDP_R_IN_INSTR_RSIZE(iq_no), iq->max_count); |
| |
| /* Remember the doorbell & instruction count register addr for this queue */ |
| iq->doorbell_reg = oct->mmio.hw_addr + CN93_VF_SDP_R_IN_INSTR_DBELL(iq_no); |
| iq->inst_cnt_reg = oct->mmio.hw_addr + CN93_VF_SDP_R_IN_CNTS(iq_no); |
| iq->intr_lvl_reg = oct->mmio.hw_addr + CN93_VF_SDP_R_IN_INT_LEVELS(iq_no); |
| |
| /* Store the current instruction counter (used in flush_iq calculation) */ |
| reset_instr_cnt = readl(iq->inst_cnt_reg); |
| writel(reset_instr_cnt, iq->inst_cnt_reg); |
| |
| /* INTR_THRESHOLD is set to max(FFFFFFFF) to disable the INTR */ |
| reg_val = CFG_GET_IQ_INTR_THRESHOLD(oct->conf) & GENMASK_ULL(31, 0); |
| octep_vf_write_csr64(oct, CN93_VF_SDP_R_IN_INT_LEVELS(iq_no), reg_val); |
| } |
| |
| /* Setup registers for a hardware Rx Queue */ |
| static void octep_vf_setup_oq_regs_cn93(struct octep_vf_device *oct, int oq_no) |
| { |
| struct octep_vf_oq *oq = oct->oq[oq_no]; |
| u32 time_threshold = 0; |
| u64 oq_ctl = ULL(0); |
| u64 reg_val; |
| |
| reg_val = octep_vf_read_csr64(oct, CN93_VF_SDP_R_OUT_CONTROL(oq_no)); |
| |
| /* wait for IDLE to set to 1 */ |
| if (!(reg_val & CN93_VF_R_OUT_CTL_IDLE)) { |
| do { |
| reg_val = octep_vf_read_csr64(oct, CN93_VF_SDP_R_OUT_CONTROL(oq_no)); |
| } while (!(reg_val & CN93_VF_R_OUT_CTL_IDLE)); |
| } |
| |
| reg_val &= ~(CN93_VF_R_OUT_CTL_IMODE); |
| reg_val &= ~(CN93_VF_R_OUT_CTL_ROR_P); |
| reg_val &= ~(CN93_VF_R_OUT_CTL_NSR_P); |
| reg_val &= ~(CN93_VF_R_OUT_CTL_ROR_I); |
| reg_val &= ~(CN93_VF_R_OUT_CTL_NSR_I); |
| reg_val &= ~(CN93_VF_R_OUT_CTL_ES_I); |
| reg_val &= ~(CN93_VF_R_OUT_CTL_ROR_D); |
| reg_val &= ~(CN93_VF_R_OUT_CTL_NSR_D); |
| reg_val &= ~(CN93_VF_R_OUT_CTL_ES_D); |
| reg_val |= (CN93_VF_R_OUT_CTL_ES_P); |
| |
| octep_vf_write_csr64(oct, CN93_VF_SDP_R_OUT_CONTROL(oq_no), reg_val); |
| octep_vf_write_csr64(oct, CN93_VF_SDP_R_OUT_SLIST_BADDR(oq_no), oq->desc_ring_dma); |
| octep_vf_write_csr64(oct, CN93_VF_SDP_R_OUT_SLIST_RSIZE(oq_no), oq->max_count); |
| |
| oq_ctl = octep_vf_read_csr64(oct, CN93_VF_SDP_R_OUT_CONTROL(oq_no)); |
| oq_ctl &= ~GENMASK_ULL(22, 0); //clear the ISIZE and BSIZE (22-0) |
| oq_ctl |= (oq->buffer_size & GENMASK_ULL(15, 0)); //populate the BSIZE (15-0) |
| octep_vf_write_csr64(oct, CN93_VF_SDP_R_OUT_CONTROL(oq_no), oq_ctl); |
| |
| /* Get the mapped address of the pkt_sent and pkts_credit regs */ |
| oq->pkts_sent_reg = oct->mmio.hw_addr + CN93_VF_SDP_R_OUT_CNTS(oq_no); |
| oq->pkts_credit_reg = oct->mmio.hw_addr + CN93_VF_SDP_R_OUT_SLIST_DBELL(oq_no); |
| |
| time_threshold = CFG_GET_OQ_INTR_TIME(oct->conf); |
| reg_val = ((u64)time_threshold << 32) | CFG_GET_OQ_INTR_PKT(oct->conf); |
| octep_vf_write_csr64(oct, CN93_VF_SDP_R_OUT_INT_LEVELS(oq_no), reg_val); |
| } |
| |
| /* Setup registers for a VF mailbox */ |
| static void octep_vf_setup_mbox_regs_cn93(struct octep_vf_device *oct, int q_no) |
| { |
| struct octep_vf_mbox *mbox = oct->mbox; |
| |
| /* PF to VF DATA reg. VF reads from this reg */ |
| mbox->mbox_read_reg = oct->mmio.hw_addr + CN93_VF_SDP_R_MBOX_PF_VF_DATA(q_no); |
| |
| /* VF mbox interrupt reg */ |
| mbox->mbox_int_reg = oct->mmio.hw_addr + CN93_VF_SDP_R_MBOX_PF_VF_INT(q_no); |
| |
| /* VF to PF DATA reg. VF writes into this reg */ |
| mbox->mbox_write_reg = oct->mmio.hw_addr + CN93_VF_SDP_R_MBOX_VF_PF_DATA(q_no); |
| } |
| |
| /* Mailbox Interrupt handler */ |
| static void cn93_handle_vf_mbox_intr(struct octep_vf_device *oct) |
| { |
| if (oct->mbox) |
| schedule_work(&oct->mbox->wk.work); |
| else |
| dev_err(&oct->pdev->dev, "cannot schedule work on invalid mbox\n"); |
| } |
| |
| /* Tx/Rx queue interrupt handler */ |
| static irqreturn_t octep_vf_ioq_intr_handler_cn93(void *data) |
| { |
| struct octep_vf_ioq_vector *vector = data; |
| struct octep_vf_device *oct; |
| struct octep_vf_oq *oq; |
| u64 reg_val; |
| |
| oct = vector->octep_vf_dev; |
| oq = vector->oq; |
| /* Mailbox interrupt arrives along with interrupt of tx/rx ring pair 0 */ |
| if (oq->q_no == 0) { |
| reg_val = octep_vf_read_csr64(oct, CN93_VF_SDP_R_MBOX_PF_VF_INT(0)); |
| if (reg_val & CN93_VF_SDP_R_MBOX_PF_VF_INT_STATUS) { |
| cn93_handle_vf_mbox_intr(oct); |
| octep_vf_write_csr64(oct, CN93_VF_SDP_R_MBOX_PF_VF_INT(0), reg_val); |
| } |
| } |
| napi_schedule_irqoff(oq->napi); |
| return IRQ_HANDLED; |
| } |
| |
| /* Re-initialize Octeon hardware registers */ |
| static void octep_vf_reinit_regs_cn93(struct octep_vf_device *oct) |
| { |
| u32 i; |
| |
| for (i = 0; i < CFG_GET_PORTS_ACTIVE_IO_RINGS(oct->conf); i++) |
| oct->hw_ops.setup_iq_regs(oct, i); |
| |
| for (i = 0; i < CFG_GET_PORTS_ACTIVE_IO_RINGS(oct->conf); i++) |
| oct->hw_ops.setup_oq_regs(oct, i); |
| |
| oct->hw_ops.enable_interrupts(oct); |
| oct->hw_ops.enable_io_queues(oct); |
| |
| for (i = 0; i < CFG_GET_PORTS_ACTIVE_IO_RINGS(oct->conf); i++) |
| writel(oct->oq[i]->max_count, oct->oq[i]->pkts_credit_reg); |
| } |
| |
| /* Enable all interrupts */ |
| static void octep_vf_enable_interrupts_cn93(struct octep_vf_device *oct) |
| { |
| int num_rings, q; |
| u64 reg_val; |
| |
| num_rings = CFG_GET_PORTS_ACTIVE_IO_RINGS(oct->conf); |
| for (q = 0; q < num_rings; q++) { |
| reg_val = octep_vf_read_csr64(oct, CN93_VF_SDP_R_IN_INT_LEVELS(q)); |
| reg_val |= BIT_ULL_MASK(62); |
| octep_vf_write_csr64(oct, CN93_VF_SDP_R_IN_INT_LEVELS(q), reg_val); |
| |
| reg_val = octep_vf_read_csr64(oct, CN93_VF_SDP_R_OUT_INT_LEVELS(q)); |
| reg_val |= BIT_ULL_MASK(62); |
| octep_vf_write_csr64(oct, CN93_VF_SDP_R_OUT_INT_LEVELS(q), reg_val); |
| } |
| /* Enable PF to VF mbox interrupt by setting 2nd bit*/ |
| octep_vf_write_csr64(oct, CN93_VF_SDP_R_MBOX_PF_VF_INT(0), |
| CN93_VF_SDP_R_MBOX_PF_VF_INT_ENAB); |
| } |
| |
| /* Disable all interrupts */ |
| static void octep_vf_disable_interrupts_cn93(struct octep_vf_device *oct) |
| { |
| int num_rings, q; |
| u64 reg_val; |
| |
| /* Disable PF to VF mbox interrupt by setting 2nd bit*/ |
| if (oct->mbox) |
| octep_vf_write_csr64(oct, CN93_VF_SDP_R_MBOX_PF_VF_INT(0), 0x0); |
| |
| num_rings = CFG_GET_PORTS_ACTIVE_IO_RINGS(oct->conf); |
| for (q = 0; q < num_rings; q++) { |
| reg_val = octep_vf_read_csr64(oct, CN93_VF_SDP_R_IN_INT_LEVELS(q)); |
| reg_val &= ~BIT_ULL_MASK(62); |
| octep_vf_write_csr64(oct, CN93_VF_SDP_R_IN_INT_LEVELS(q), reg_val); |
| |
| reg_val = octep_vf_read_csr64(oct, CN93_VF_SDP_R_OUT_INT_LEVELS(q)); |
| reg_val &= ~BIT_ULL_MASK(62); |
| octep_vf_write_csr64(oct, CN93_VF_SDP_R_OUT_INT_LEVELS(q), reg_val); |
| } |
| } |
| |
| /* Get new Octeon Read Index: index of descriptor that Octeon reads next. */ |
| static u32 octep_vf_update_iq_read_index_cn93(struct octep_vf_iq *iq) |
| { |
| u32 pkt_in_done = readl(iq->inst_cnt_reg); |
| u32 last_done, new_idx; |
| |
| last_done = pkt_in_done - iq->pkt_in_done; |
| iq->pkt_in_done = pkt_in_done; |
| |
| new_idx = (iq->octep_vf_read_index + last_done) % iq->max_count; |
| |
| return new_idx; |
| } |
| |
| /* Enable a hardware Tx Queue */ |
| static void octep_vf_enable_iq_cn93(struct octep_vf_device *oct, int iq_no) |
| { |
| u64 loop = HZ; |
| u64 reg_val; |
| |
| octep_vf_write_csr64(oct, CN93_VF_SDP_R_IN_INSTR_DBELL(iq_no), GENMASK_ULL(31, 0)); |
| |
| while (octep_vf_read_csr64(oct, CN93_VF_SDP_R_IN_INSTR_DBELL(iq_no)) && |
| loop--) { |
| schedule_timeout_interruptible(1); |
| } |
| |
| reg_val = octep_vf_read_csr64(oct, CN93_VF_SDP_R_IN_INT_LEVELS(iq_no)); |
| reg_val |= BIT_ULL_MASK(62); |
| octep_vf_write_csr64(oct, CN93_VF_SDP_R_IN_INT_LEVELS(iq_no), reg_val); |
| |
| reg_val = octep_vf_read_csr64(oct, CN93_VF_SDP_R_IN_ENABLE(iq_no)); |
| reg_val |= ULL(1); |
| octep_vf_write_csr64(oct, CN93_VF_SDP_R_IN_ENABLE(iq_no), reg_val); |
| } |
| |
| /* Enable a hardware Rx Queue */ |
| static void octep_vf_enable_oq_cn93(struct octep_vf_device *oct, int oq_no) |
| { |
| u64 reg_val; |
| |
| reg_val = octep_vf_read_csr64(oct, CN93_VF_SDP_R_OUT_INT_LEVELS(oq_no)); |
| reg_val |= BIT_ULL_MASK(62); |
| octep_vf_write_csr64(oct, CN93_VF_SDP_R_OUT_INT_LEVELS(oq_no), reg_val); |
| |
| octep_vf_write_csr64(oct, CN93_VF_SDP_R_OUT_SLIST_DBELL(oq_no), GENMASK_ULL(31, 0)); |
| |
| reg_val = octep_vf_read_csr64(oct, CN93_VF_SDP_R_OUT_ENABLE(oq_no)); |
| reg_val |= ULL(1); |
| octep_vf_write_csr64(oct, CN93_VF_SDP_R_OUT_ENABLE(oq_no), reg_val); |
| } |
| |
| /* Enable all hardware Tx/Rx Queues assigned to VF */ |
| static void octep_vf_enable_io_queues_cn93(struct octep_vf_device *oct) |
| { |
| u8 q; |
| |
| for (q = 0; q < CFG_GET_PORTS_ACTIVE_IO_RINGS(oct->conf); q++) { |
| octep_vf_enable_iq_cn93(oct, q); |
| octep_vf_enable_oq_cn93(oct, q); |
| } |
| } |
| |
| /* Disable a hardware Tx Queue assigned to VF */ |
| static void octep_vf_disable_iq_cn93(struct octep_vf_device *oct, int iq_no) |
| { |
| u64 reg_val; |
| |
| reg_val = octep_vf_read_csr64(oct, CN93_VF_SDP_R_IN_ENABLE(iq_no)); |
| reg_val &= ~ULL(1); |
| octep_vf_write_csr64(oct, CN93_VF_SDP_R_IN_ENABLE(iq_no), reg_val); |
| } |
| |
| /* Disable a hardware Rx Queue assigned to VF */ |
| static void octep_vf_disable_oq_cn93(struct octep_vf_device *oct, int oq_no) |
| { |
| u64 reg_val; |
| |
| reg_val = octep_vf_read_csr64(oct, CN93_VF_SDP_R_OUT_ENABLE(oq_no)); |
| reg_val &= ~ULL(1); |
| octep_vf_write_csr64(oct, CN93_VF_SDP_R_OUT_ENABLE(oq_no), reg_val); |
| } |
| |
| /* Disable all hardware Tx/Rx Queues assigned to VF */ |
| static void octep_vf_disable_io_queues_cn93(struct octep_vf_device *oct) |
| { |
| int q; |
| |
| for (q = 0; q < CFG_GET_PORTS_ACTIVE_IO_RINGS(oct->conf); q++) { |
| octep_vf_disable_iq_cn93(oct, q); |
| octep_vf_disable_oq_cn93(oct, q); |
| } |
| } |
| |
| /* Dump hardware registers (including Tx/Rx queues) for debugging. */ |
| static void octep_vf_dump_registers_cn93(struct octep_vf_device *oct) |
| { |
| u8 num_rings, q; |
| |
| num_rings = CFG_GET_PORTS_ACTIVE_IO_RINGS(oct->conf); |
| for (q = 0; q < num_rings; q++) |
| cn93_vf_dump_q_regs(oct, q); |
| } |
| |
| /** |
| * octep_vf_device_setup_cn93() - Setup Octeon device. |
| * |
| * @oct: Octeon device private data structure. |
| * |
| * - initialize hardware operations. |
| * - get target side pcie port number for the device. |
| * - set initial configuration and max limits. |
| */ |
| void octep_vf_device_setup_cn93(struct octep_vf_device *oct) |
| { |
| oct->hw_ops.setup_iq_regs = octep_vf_setup_iq_regs_cn93; |
| oct->hw_ops.setup_oq_regs = octep_vf_setup_oq_regs_cn93; |
| oct->hw_ops.setup_mbox_regs = octep_vf_setup_mbox_regs_cn93; |
| |
| oct->hw_ops.ioq_intr_handler = octep_vf_ioq_intr_handler_cn93; |
| oct->hw_ops.reinit_regs = octep_vf_reinit_regs_cn93; |
| |
| oct->hw_ops.enable_interrupts = octep_vf_enable_interrupts_cn93; |
| oct->hw_ops.disable_interrupts = octep_vf_disable_interrupts_cn93; |
| |
| oct->hw_ops.update_iq_read_idx = octep_vf_update_iq_read_index_cn93; |
| |
| oct->hw_ops.enable_iq = octep_vf_enable_iq_cn93; |
| oct->hw_ops.enable_oq = octep_vf_enable_oq_cn93; |
| oct->hw_ops.enable_io_queues = octep_vf_enable_io_queues_cn93; |
| |
| oct->hw_ops.disable_iq = octep_vf_disable_iq_cn93; |
| oct->hw_ops.disable_oq = octep_vf_disable_oq_cn93; |
| oct->hw_ops.disable_io_queues = octep_vf_disable_io_queues_cn93; |
| oct->hw_ops.reset_io_queues = octep_vf_reset_io_queues_cn93; |
| |
| oct->hw_ops.dump_registers = octep_vf_dump_registers_cn93; |
| octep_vf_init_config_cn93_vf(oct); |
| } |