| // SPDX-License-Identifier: (GPL-2.0+ OR BSD-3-Clause) |
| /* Copyright 2017-2019 NXP */ |
| |
| #include "enetc.h" |
| |
| int enetc_setup_cbdr(struct device *dev, struct enetc_hw *hw, int bd_count, |
| struct enetc_cbdr *cbdr) |
| { |
| int size = bd_count * sizeof(struct enetc_cbd); |
| |
| cbdr->bd_base = dma_alloc_coherent(dev, size, &cbdr->bd_dma_base, |
| GFP_KERNEL); |
| if (!cbdr->bd_base) |
| return -ENOMEM; |
| |
| /* h/w requires 128B alignment */ |
| if (!IS_ALIGNED(cbdr->bd_dma_base, 128)) { |
| dma_free_coherent(dev, size, cbdr->bd_base, |
| cbdr->bd_dma_base); |
| return -EINVAL; |
| } |
| |
| cbdr->next_to_clean = 0; |
| cbdr->next_to_use = 0; |
| cbdr->dma_dev = dev; |
| cbdr->bd_count = bd_count; |
| |
| cbdr->pir = hw->reg + ENETC_SICBDRPIR; |
| cbdr->cir = hw->reg + ENETC_SICBDRCIR; |
| cbdr->mr = hw->reg + ENETC_SICBDRMR; |
| |
| /* set CBDR cache attributes */ |
| enetc_wr(hw, ENETC_SICAR2, |
| ENETC_SICAR_RD_COHERENT | ENETC_SICAR_WR_COHERENT); |
| |
| enetc_wr(hw, ENETC_SICBDRBAR0, lower_32_bits(cbdr->bd_dma_base)); |
| enetc_wr(hw, ENETC_SICBDRBAR1, upper_32_bits(cbdr->bd_dma_base)); |
| enetc_wr(hw, ENETC_SICBDRLENR, ENETC_RTBLENR_LEN(cbdr->bd_count)); |
| |
| enetc_wr_reg(cbdr->pir, cbdr->next_to_clean); |
| enetc_wr_reg(cbdr->cir, cbdr->next_to_use); |
| /* enable ring */ |
| enetc_wr_reg(cbdr->mr, BIT(31)); |
| |
| return 0; |
| } |
| |
| void enetc_teardown_cbdr(struct enetc_cbdr *cbdr) |
| { |
| int size = cbdr->bd_count * sizeof(struct enetc_cbd); |
| |
| /* disable ring */ |
| enetc_wr_reg(cbdr->mr, 0); |
| |
| dma_free_coherent(cbdr->dma_dev, size, cbdr->bd_base, |
| cbdr->bd_dma_base); |
| cbdr->bd_base = NULL; |
| cbdr->dma_dev = NULL; |
| } |
| |
| static void enetc_clean_cbdr(struct enetc_cbdr *ring) |
| { |
| struct enetc_cbd *dest_cbd; |
| int i, status; |
| |
| i = ring->next_to_clean; |
| |
| while (enetc_rd_reg(ring->cir) != i) { |
| dest_cbd = ENETC_CBD(*ring, i); |
| status = dest_cbd->status_flags & ENETC_CBD_STATUS_MASK; |
| if (status) |
| dev_warn(ring->dma_dev, "CMD err %04x for cmd %04x\n", |
| status, dest_cbd->cmd); |
| |
| memset(dest_cbd, 0, sizeof(*dest_cbd)); |
| |
| i = (i + 1) % ring->bd_count; |
| } |
| |
| ring->next_to_clean = i; |
| } |
| |
| static int enetc_cbd_unused(struct enetc_cbdr *r) |
| { |
| return (r->next_to_clean - r->next_to_use - 1 + r->bd_count) % |
| r->bd_count; |
| } |
| |
| int enetc_send_cmd(struct enetc_si *si, struct enetc_cbd *cbd) |
| { |
| struct enetc_cbdr *ring = &si->cbd_ring; |
| int timeout = ENETC_CBDR_TIMEOUT; |
| struct enetc_cbd *dest_cbd; |
| int i; |
| |
| if (unlikely(!ring->bd_base)) |
| return -EIO; |
| |
| if (unlikely(!enetc_cbd_unused(ring))) |
| enetc_clean_cbdr(ring); |
| |
| i = ring->next_to_use; |
| dest_cbd = ENETC_CBD(*ring, i); |
| |
| /* copy command to the ring */ |
| *dest_cbd = *cbd; |
| i = (i + 1) % ring->bd_count; |
| |
| ring->next_to_use = i; |
| /* let H/W know BD ring has been updated */ |
| enetc_wr_reg(ring->pir, i); |
| |
| do { |
| if (enetc_rd_reg(ring->cir) == i) |
| break; |
| udelay(10); /* cannot sleep, rtnl_lock() */ |
| timeout -= 10; |
| } while (timeout); |
| |
| if (!timeout) |
| return -EBUSY; |
| |
| /* CBD may writeback data, feedback up level */ |
| *cbd = *dest_cbd; |
| |
| enetc_clean_cbdr(ring); |
| |
| return 0; |
| } |
| |
| int enetc_clear_mac_flt_entry(struct enetc_si *si, int index) |
| { |
| struct enetc_cbd cbd; |
| |
| memset(&cbd, 0, sizeof(cbd)); |
| |
| cbd.cls = 1; |
| cbd.status_flags = ENETC_CBD_FLAGS_SF; |
| cbd.index = cpu_to_le16(index); |
| |
| return enetc_send_cmd(si, &cbd); |
| } |
| |
| int enetc_set_mac_flt_entry(struct enetc_si *si, int index, |
| char *mac_addr, int si_map) |
| { |
| struct enetc_cbd cbd; |
| u32 upper; |
| u16 lower; |
| |
| memset(&cbd, 0, sizeof(cbd)); |
| |
| /* fill up the "set" descriptor */ |
| cbd.cls = 1; |
| cbd.status_flags = ENETC_CBD_FLAGS_SF; |
| cbd.index = cpu_to_le16(index); |
| cbd.opt[3] = cpu_to_le32(si_map); |
| /* enable entry */ |
| cbd.opt[0] = cpu_to_le32(BIT(31)); |
| |
| upper = *(const u32 *)mac_addr; |
| lower = *(const u16 *)(mac_addr + 4); |
| cbd.addr[0] = cpu_to_le32(upper); |
| cbd.addr[1] = cpu_to_le32(lower); |
| |
| return enetc_send_cmd(si, &cbd); |
| } |
| |
| #define RFSE_ALIGN 64 |
| /* Set entry in RFS table */ |
| int enetc_set_fs_entry(struct enetc_si *si, struct enetc_cmd_rfse *rfse, |
| int index) |
| { |
| struct enetc_cbdr *ring = &si->cbd_ring; |
| struct enetc_cbd cbd = {.cmd = 0}; |
| dma_addr_t dma, dma_align; |
| void *tmp, *tmp_align; |
| int err; |
| |
| /* fill up the "set" descriptor */ |
| cbd.cmd = 0; |
| cbd.cls = 4; |
| cbd.index = cpu_to_le16(index); |
| cbd.length = cpu_to_le16(sizeof(*rfse)); |
| cbd.opt[3] = cpu_to_le32(0); /* SI */ |
| |
| tmp = dma_alloc_coherent(ring->dma_dev, sizeof(*rfse) + RFSE_ALIGN, |
| &dma, GFP_KERNEL); |
| if (!tmp) { |
| dev_err(ring->dma_dev, "DMA mapping of RFS entry failed!\n"); |
| return -ENOMEM; |
| } |
| |
| dma_align = ALIGN(dma, RFSE_ALIGN); |
| tmp_align = PTR_ALIGN(tmp, RFSE_ALIGN); |
| memcpy(tmp_align, rfse, sizeof(*rfse)); |
| |
| cbd.addr[0] = cpu_to_le32(lower_32_bits(dma_align)); |
| cbd.addr[1] = cpu_to_le32(upper_32_bits(dma_align)); |
| |
| err = enetc_send_cmd(si, &cbd); |
| if (err) |
| dev_err(ring->dma_dev, "FS entry add failed (%d)!", err); |
| |
| dma_free_coherent(ring->dma_dev, sizeof(*rfse) + RFSE_ALIGN, |
| tmp, dma); |
| |
| return err; |
| } |
| |
| #define RSSE_ALIGN 64 |
| static int enetc_cmd_rss_table(struct enetc_si *si, u32 *table, int count, |
| bool read) |
| { |
| struct enetc_cbdr *ring = &si->cbd_ring; |
| struct enetc_cbd cbd = {.cmd = 0}; |
| dma_addr_t dma, dma_align; |
| u8 *tmp, *tmp_align; |
| int err, i; |
| |
| if (count < RSSE_ALIGN) |
| /* HW only takes in a full 64 entry table */ |
| return -EINVAL; |
| |
| tmp = dma_alloc_coherent(ring->dma_dev, count + RSSE_ALIGN, |
| &dma, GFP_KERNEL); |
| if (!tmp) { |
| dev_err(ring->dma_dev, "DMA mapping of RSS table failed!\n"); |
| return -ENOMEM; |
| } |
| dma_align = ALIGN(dma, RSSE_ALIGN); |
| tmp_align = PTR_ALIGN(tmp, RSSE_ALIGN); |
| |
| if (!read) |
| for (i = 0; i < count; i++) |
| tmp_align[i] = (u8)(table[i]); |
| |
| /* fill up the descriptor */ |
| cbd.cmd = read ? 2 : 1; |
| cbd.cls = 3; |
| cbd.length = cpu_to_le16(count); |
| |
| cbd.addr[0] = cpu_to_le32(lower_32_bits(dma_align)); |
| cbd.addr[1] = cpu_to_le32(upper_32_bits(dma_align)); |
| |
| err = enetc_send_cmd(si, &cbd); |
| if (err) |
| dev_err(ring->dma_dev, "RSS cmd failed (%d)!", err); |
| |
| if (read) |
| for (i = 0; i < count; i++) |
| table[i] = tmp_align[i]; |
| |
| dma_free_coherent(ring->dma_dev, count + RSSE_ALIGN, tmp, dma); |
| |
| return err; |
| } |
| |
| /* Get RSS table */ |
| int enetc_get_rss_table(struct enetc_si *si, u32 *table, int count) |
| { |
| return enetc_cmd_rss_table(si, table, count, true); |
| } |
| |
| /* Set RSS table */ |
| int enetc_set_rss_table(struct enetc_si *si, const u32 *table, int count) |
| { |
| return enetc_cmd_rss_table(si, (u32 *)table, count, false); |
| } |