| // SPDX-License-Identifier: GPL-2.0 OR BSD-3-Clause |
| /* |
| * rtase is the Linux device driver released for Realtek Automotive Switch |
| * controllers with PCI-Express interface. |
| * |
| * Copyright(c) 2024 Realtek Semiconductor Corp. |
| * |
| * Below is a simplified block diagram of the chip and its relevant interfaces. |
| * |
| * ************************* |
| * * * |
| * * CPU network device * |
| * * * |
| * * +-------------+ * |
| * * | PCIE Host | * |
| * ***********++************ |
| * || |
| * PCIE |
| * || |
| * ********************++********************** |
| * * | PCIE Endpoint | * |
| * * +---------------+ * |
| * * | GMAC | * |
| * * +--++--+ Realtek * |
| * * || RTL90xx Series * |
| * * || * |
| * * +-------------++----------------+ * |
| * * | | MAC | | * |
| * * | +-----+ | * |
| * * | | * |
| * * | Ethernet Switch Core | * |
| * * | | * |
| * * | +-----+ +-----+ | * |
| * * | | MAC |...........| MAC | | * |
| * * +---+-----+-----------+-----+---+ * |
| * * | PHY |...........| PHY | * |
| * * +--++-+ +--++-+ * |
| * *************||****************||*********** |
| * |
| * The block of the Realtek RTL90xx series is our entire chip architecture, |
| * the GMAC is connected to the switch core, and there is no PHY in between. |
| * In addition, this driver is mainly used to control GMAC, but does not |
| * control the switch core, so it is not the same as DSA. Linux only plays |
| * the role of a normal leaf node in this model. |
| */ |
| |
| #include <linux/crc32.h> |
| #include <linux/dma-mapping.h> |
| #include <linux/etherdevice.h> |
| #include <linux/if_vlan.h> |
| #include <linux/in.h> |
| #include <linux/init.h> |
| #include <linux/interrupt.h> |
| #include <linux/io.h> |
| #include <linux/iopoll.h> |
| #include <linux/ip.h> |
| #include <linux/ipv6.h> |
| #include <linux/mdio.h> |
| #include <linux/module.h> |
| #include <linux/netdevice.h> |
| #include <linux/pci.h> |
| #include <linux/pm_runtime.h> |
| #include <linux/prefetch.h> |
| #include <linux/rtnetlink.h> |
| #include <linux/tcp.h> |
| #include <asm/irq.h> |
| #include <net/ip6_checksum.h> |
| #include <net/netdev_queues.h> |
| #include <net/page_pool/helpers.h> |
| #include <net/pkt_cls.h> |
| |
| #include "rtase.h" |
| |
| #define RTK_OPTS1_DEBUG_VALUE 0x0BADBEEF |
| #define RTK_MAGIC_NUMBER 0x0BADBADBADBADBAD |
| |
| static const struct pci_device_id rtase_pci_tbl[] = { |
| {PCI_VDEVICE(REALTEK, 0x906A)}, |
| {} |
| }; |
| |
| MODULE_DEVICE_TABLE(pci, rtase_pci_tbl); |
| |
| MODULE_AUTHOR("Realtek ARD Software Team"); |
| MODULE_DESCRIPTION("Network Driver for the PCIe interface of Realtek Automotive Ethernet Switch"); |
| MODULE_LICENSE("Dual BSD/GPL"); |
| |
| struct rtase_counters { |
| __le64 tx_packets; |
| __le64 rx_packets; |
| __le64 tx_errors; |
| __le32 rx_errors; |
| __le16 rx_missed; |
| __le16 align_errors; |
| __le32 tx_one_collision; |
| __le32 tx_multi_collision; |
| __le64 rx_unicast; |
| __le64 rx_broadcast; |
| __le32 rx_multicast; |
| __le16 tx_aborted; |
| __le16 tx_underrun; |
| } __packed; |
| |
| static void rtase_w8(const struct rtase_private *tp, u16 reg, u8 val8) |
| { |
| writeb(val8, tp->mmio_addr + reg); |
| } |
| |
| static void rtase_w16(const struct rtase_private *tp, u16 reg, u16 val16) |
| { |
| writew(val16, tp->mmio_addr + reg); |
| } |
| |
| static void rtase_w32(const struct rtase_private *tp, u16 reg, u32 val32) |
| { |
| writel(val32, tp->mmio_addr + reg); |
| } |
| |
| static u8 rtase_r8(const struct rtase_private *tp, u16 reg) |
| { |
| return readb(tp->mmio_addr + reg); |
| } |
| |
| static u16 rtase_r16(const struct rtase_private *tp, u16 reg) |
| { |
| return readw(tp->mmio_addr + reg); |
| } |
| |
| static u32 rtase_r32(const struct rtase_private *tp, u16 reg) |
| { |
| return readl(tp->mmio_addr + reg); |
| } |
| |
| static void rtase_free_desc(struct rtase_private *tp) |
| { |
| struct pci_dev *pdev = tp->pdev; |
| u32 i; |
| |
| for (i = 0; i < tp->func_tx_queue_num; i++) { |
| if (!tp->tx_ring[i].desc) |
| continue; |
| |
| dma_free_coherent(&pdev->dev, RTASE_TX_RING_DESC_SIZE, |
| tp->tx_ring[i].desc, |
| tp->tx_ring[i].phy_addr); |
| tp->tx_ring[i].desc = NULL; |
| } |
| |
| for (i = 0; i < tp->func_rx_queue_num; i++) { |
| if (!tp->rx_ring[i].desc) |
| continue; |
| |
| dma_free_coherent(&pdev->dev, RTASE_RX_RING_DESC_SIZE, |
| tp->rx_ring[i].desc, |
| tp->rx_ring[i].phy_addr); |
| tp->rx_ring[i].desc = NULL; |
| } |
| } |
| |
| static int rtase_alloc_desc(struct rtase_private *tp) |
| { |
| struct pci_dev *pdev = tp->pdev; |
| u32 i; |
| |
| /* rx and tx descriptors needs 256 bytes alignment. |
| * dma_alloc_coherent provides more. |
| */ |
| for (i = 0; i < tp->func_tx_queue_num; i++) { |
| tp->tx_ring[i].desc = |
| dma_alloc_coherent(&pdev->dev, |
| RTASE_TX_RING_DESC_SIZE, |
| &tp->tx_ring[i].phy_addr, |
| GFP_KERNEL); |
| if (!tp->tx_ring[i].desc) |
| goto err_out; |
| } |
| |
| for (i = 0; i < tp->func_rx_queue_num; i++) { |
| tp->rx_ring[i].desc = |
| dma_alloc_coherent(&pdev->dev, |
| RTASE_RX_RING_DESC_SIZE, |
| &tp->rx_ring[i].phy_addr, |
| GFP_KERNEL); |
| if (!tp->rx_ring[i].desc) |
| goto err_out; |
| } |
| |
| return 0; |
| |
| err_out: |
| rtase_free_desc(tp); |
| return -ENOMEM; |
| } |
| |
| static void rtase_unmap_tx_skb(struct pci_dev *pdev, u32 len, |
| struct rtase_tx_desc *desc) |
| { |
| dma_unmap_single(&pdev->dev, le64_to_cpu(desc->addr), len, |
| DMA_TO_DEVICE); |
| desc->opts1 = cpu_to_le32(RTK_OPTS1_DEBUG_VALUE); |
| desc->opts2 = 0x00; |
| desc->addr = cpu_to_le64(RTK_MAGIC_NUMBER); |
| } |
| |
| static void rtase_tx_clear_range(struct rtase_ring *ring, u32 start, u32 n) |
| { |
| struct rtase_tx_desc *desc_base = ring->desc; |
| struct rtase_private *tp = ring->ivec->tp; |
| u32 i; |
| |
| for (i = 0; i < n; i++) { |
| u32 entry = (start + i) % RTASE_NUM_DESC; |
| struct rtase_tx_desc *desc = desc_base + entry; |
| u32 len = ring->mis.len[entry]; |
| struct sk_buff *skb; |
| |
| if (len == 0) |
| continue; |
| |
| rtase_unmap_tx_skb(tp->pdev, len, desc); |
| ring->mis.len[entry] = 0; |
| skb = ring->skbuff[entry]; |
| if (!skb) |
| continue; |
| |
| tp->stats.tx_dropped++; |
| dev_kfree_skb_any(skb); |
| ring->skbuff[entry] = NULL; |
| } |
| } |
| |
| static void rtase_tx_clear(struct rtase_private *tp) |
| { |
| struct rtase_ring *ring; |
| u16 i; |
| |
| for (i = 0; i < tp->func_tx_queue_num; i++) { |
| ring = &tp->tx_ring[i]; |
| rtase_tx_clear_range(ring, ring->dirty_idx, RTASE_NUM_DESC); |
| ring->cur_idx = 0; |
| ring->dirty_idx = 0; |
| } |
| } |
| |
| static void rtase_mark_to_asic(union rtase_rx_desc *desc, u32 rx_buf_sz) |
| { |
| u32 eor = le32_to_cpu(desc->desc_cmd.opts1) & RTASE_RING_END; |
| |
| desc->desc_status.opts2 = 0; |
| /* force memory writes to complete before releasing descriptor */ |
| dma_wmb(); |
| WRITE_ONCE(desc->desc_cmd.opts1, |
| cpu_to_le32(RTASE_DESC_OWN | eor | rx_buf_sz)); |
| } |
| |
| static u32 rtase_tx_avail(struct rtase_ring *ring) |
| { |
| return READ_ONCE(ring->dirty_idx) + RTASE_NUM_DESC - |
| READ_ONCE(ring->cur_idx); |
| } |
| |
| static int tx_handler(struct rtase_ring *ring, int budget) |
| { |
| const struct rtase_private *tp = ring->ivec->tp; |
| struct net_device *dev = tp->dev; |
| u32 dirty_tx, tx_left; |
| u32 bytes_compl = 0; |
| u32 pkts_compl = 0; |
| int workdone = 0; |
| |
| dirty_tx = ring->dirty_idx; |
| tx_left = READ_ONCE(ring->cur_idx) - dirty_tx; |
| |
| while (tx_left > 0) { |
| u32 entry = dirty_tx % RTASE_NUM_DESC; |
| struct rtase_tx_desc *desc = ring->desc + |
| sizeof(struct rtase_tx_desc) * entry; |
| u32 status; |
| |
| status = le32_to_cpu(desc->opts1); |
| |
| if (status & RTASE_DESC_OWN) |
| break; |
| |
| rtase_unmap_tx_skb(tp->pdev, ring->mis.len[entry], desc); |
| ring->mis.len[entry] = 0; |
| if (ring->skbuff[entry]) { |
| pkts_compl++; |
| bytes_compl += ring->skbuff[entry]->len; |
| napi_consume_skb(ring->skbuff[entry], budget); |
| ring->skbuff[entry] = NULL; |
| } |
| |
| dirty_tx++; |
| tx_left--; |
| workdone++; |
| |
| if (workdone == RTASE_TX_BUDGET_DEFAULT) |
| break; |
| } |
| |
| if (ring->dirty_idx != dirty_tx) { |
| dev_sw_netstats_tx_add(dev, pkts_compl, bytes_compl); |
| WRITE_ONCE(ring->dirty_idx, dirty_tx); |
| |
| netif_subqueue_completed_wake(dev, ring->index, pkts_compl, |
| bytes_compl, |
| rtase_tx_avail(ring), |
| RTASE_TX_START_THRS); |
| |
| if (ring->cur_idx != dirty_tx) |
| rtase_w8(tp, RTASE_TPPOLL, BIT(ring->index)); |
| } |
| |
| return 0; |
| } |
| |
| static void rtase_tx_desc_init(struct rtase_private *tp, u16 idx) |
| { |
| struct rtase_ring *ring = &tp->tx_ring[idx]; |
| struct rtase_tx_desc *desc; |
| u32 i; |
| |
| memset(ring->desc, 0x0, RTASE_TX_RING_DESC_SIZE); |
| memset(ring->skbuff, 0x0, sizeof(ring->skbuff)); |
| ring->cur_idx = 0; |
| ring->dirty_idx = 0; |
| ring->index = idx; |
| ring->alloc_fail = 0; |
| |
| for (i = 0; i < RTASE_NUM_DESC; i++) { |
| ring->mis.len[i] = 0; |
| if ((RTASE_NUM_DESC - 1) == i) { |
| desc = ring->desc + sizeof(struct rtase_tx_desc) * i; |
| desc->opts1 = cpu_to_le32(RTASE_RING_END); |
| } |
| } |
| |
| ring->ring_handler = tx_handler; |
| if (idx < 4) { |
| ring->ivec = &tp->int_vector[idx]; |
| list_add_tail(&ring->ring_entry, |
| &tp->int_vector[idx].ring_list); |
| } else { |
| ring->ivec = &tp->int_vector[0]; |
| list_add_tail(&ring->ring_entry, &tp->int_vector[0].ring_list); |
| } |
| } |
| |
| static void rtase_map_to_asic(union rtase_rx_desc *desc, dma_addr_t mapping, |
| u32 rx_buf_sz) |
| { |
| desc->desc_cmd.addr = cpu_to_le64(mapping); |
| |
| rtase_mark_to_asic(desc, rx_buf_sz); |
| } |
| |
| static void rtase_make_unusable_by_asic(union rtase_rx_desc *desc) |
| { |
| desc->desc_cmd.addr = cpu_to_le64(RTK_MAGIC_NUMBER); |
| desc->desc_cmd.opts1 &= ~cpu_to_le32(RTASE_DESC_OWN | RSVD_MASK); |
| } |
| |
| static int rtase_alloc_rx_data_buf(struct rtase_ring *ring, |
| void **p_data_buf, |
| union rtase_rx_desc *desc, |
| dma_addr_t *rx_phy_addr) |
| { |
| struct rtase_int_vector *ivec = ring->ivec; |
| const struct rtase_private *tp = ivec->tp; |
| dma_addr_t mapping; |
| struct page *page; |
| |
| page = page_pool_dev_alloc_pages(tp->page_pool); |
| if (!page) { |
| ring->alloc_fail++; |
| goto err_out; |
| } |
| |
| *p_data_buf = page_address(page); |
| mapping = page_pool_get_dma_addr(page); |
| *rx_phy_addr = mapping; |
| rtase_map_to_asic(desc, mapping, tp->rx_buf_sz); |
| |
| return 0; |
| |
| err_out: |
| rtase_make_unusable_by_asic(desc); |
| |
| return -ENOMEM; |
| } |
| |
| static u32 rtase_rx_ring_fill(struct rtase_ring *ring, u32 ring_start, |
| u32 ring_end) |
| { |
| union rtase_rx_desc *desc_base = ring->desc; |
| u32 cur; |
| |
| for (cur = ring_start; ring_end - cur > 0; cur++) { |
| u32 i = cur % RTASE_NUM_DESC; |
| union rtase_rx_desc *desc = desc_base + i; |
| int ret; |
| |
| if (ring->data_buf[i]) |
| continue; |
| |
| ret = rtase_alloc_rx_data_buf(ring, &ring->data_buf[i], desc, |
| &ring->mis.data_phy_addr[i]); |
| if (ret) |
| break; |
| } |
| |
| return cur - ring_start; |
| } |
| |
| static void rtase_mark_as_last_descriptor(union rtase_rx_desc *desc) |
| { |
| desc->desc_cmd.opts1 |= cpu_to_le32(RTASE_RING_END); |
| } |
| |
| static void rtase_rx_ring_clear(struct page_pool *page_pool, |
| struct rtase_ring *ring) |
| { |
| union rtase_rx_desc *desc; |
| struct page *page; |
| u32 i; |
| |
| for (i = 0; i < RTASE_NUM_DESC; i++) { |
| desc = ring->desc + sizeof(union rtase_rx_desc) * i; |
| page = virt_to_head_page(ring->data_buf[i]); |
| |
| if (ring->data_buf[i]) |
| page_pool_put_full_page(page_pool, page, true); |
| |
| rtase_make_unusable_by_asic(desc); |
| } |
| } |
| |
| static int rtase_fragmented_frame(u32 status) |
| { |
| return (status & (RTASE_RX_FIRST_FRAG | RTASE_RX_LAST_FRAG)) != |
| (RTASE_RX_FIRST_FRAG | RTASE_RX_LAST_FRAG); |
| } |
| |
| static void rtase_rx_csum(const struct rtase_private *tp, struct sk_buff *skb, |
| const union rtase_rx_desc *desc) |
| { |
| u32 opts2 = le32_to_cpu(desc->desc_status.opts2); |
| |
| /* rx csum offload */ |
| if (((opts2 & RTASE_RX_V4F) && !(opts2 & RTASE_RX_IPF)) || |
| (opts2 & RTASE_RX_V6F)) { |
| if (((opts2 & RTASE_RX_TCPT) && !(opts2 & RTASE_RX_TCPF)) || |
| ((opts2 & RTASE_RX_UDPT) && !(opts2 & RTASE_RX_UDPF))) |
| skb->ip_summed = CHECKSUM_UNNECESSARY; |
| else |
| skb->ip_summed = CHECKSUM_NONE; |
| } else { |
| skb->ip_summed = CHECKSUM_NONE; |
| } |
| } |
| |
| static void rtase_rx_vlan_skb(union rtase_rx_desc *desc, struct sk_buff *skb) |
| { |
| u32 opts2 = le32_to_cpu(desc->desc_status.opts2); |
| |
| if (!(opts2 & RTASE_RX_VLAN_TAG)) |
| return; |
| |
| __vlan_hwaccel_put_tag(skb, htons(ETH_P_8021Q), |
| swab16(opts2 & RTASE_VLAN_TAG_MASK)); |
| } |
| |
| static void rtase_rx_skb(const struct rtase_ring *ring, struct sk_buff *skb) |
| { |
| struct rtase_int_vector *ivec = ring->ivec; |
| |
| napi_gro_receive(&ivec->napi, skb); |
| } |
| |
| static int rx_handler(struct rtase_ring *ring, int budget) |
| { |
| union rtase_rx_desc *desc_base = ring->desc; |
| u32 pkt_size, cur_rx, delta, entry, status; |
| struct rtase_private *tp = ring->ivec->tp; |
| struct net_device *dev = tp->dev; |
| union rtase_rx_desc *desc; |
| struct sk_buff *skb; |
| int workdone = 0; |
| |
| cur_rx = ring->cur_idx; |
| entry = cur_rx % RTASE_NUM_DESC; |
| desc = &desc_base[entry]; |
| |
| while (workdone < budget) { |
| status = le32_to_cpu(desc->desc_status.opts1); |
| |
| if (status & RTASE_DESC_OWN) |
| break; |
| |
| /* This barrier is needed to keep us from reading |
| * any other fields out of the rx descriptor until |
| * we know the status of RTASE_DESC_OWN |
| */ |
| dma_rmb(); |
| |
| if (unlikely(status & RTASE_RX_RES)) { |
| if (net_ratelimit()) |
| netdev_warn(dev, "Rx ERROR. status = %08x\n", |
| status); |
| |
| tp->stats.rx_errors++; |
| |
| if (status & (RTASE_RX_RWT | RTASE_RX_RUNT)) |
| tp->stats.rx_length_errors++; |
| |
| if (status & RTASE_RX_CRC) |
| tp->stats.rx_crc_errors++; |
| |
| if (dev->features & NETIF_F_RXALL) |
| goto process_pkt; |
| |
| rtase_mark_to_asic(desc, tp->rx_buf_sz); |
| goto skip_process_pkt; |
| } |
| |
| process_pkt: |
| pkt_size = status & RTASE_RX_PKT_SIZE_MASK; |
| if (likely(!(dev->features & NETIF_F_RXFCS))) |
| pkt_size -= ETH_FCS_LEN; |
| |
| /* The driver does not support incoming fragmented frames. |
| * They are seen as a symptom of over-mtu sized frames. |
| */ |
| if (unlikely(rtase_fragmented_frame(status))) { |
| tp->stats.rx_dropped++; |
| tp->stats.rx_length_errors++; |
| rtase_mark_to_asic(desc, tp->rx_buf_sz); |
| goto skip_process_pkt; |
| } |
| |
| dma_sync_single_for_cpu(&tp->pdev->dev, |
| ring->mis.data_phy_addr[entry], |
| tp->rx_buf_sz, DMA_FROM_DEVICE); |
| |
| skb = build_skb(ring->data_buf[entry], PAGE_SIZE); |
| if (!skb) { |
| tp->stats.rx_dropped++; |
| rtase_mark_to_asic(desc, tp->rx_buf_sz); |
| goto skip_process_pkt; |
| } |
| ring->data_buf[entry] = NULL; |
| |
| if (dev->features & NETIF_F_RXCSUM) |
| rtase_rx_csum(tp, skb, desc); |
| |
| skb_put(skb, pkt_size); |
| skb_mark_for_recycle(skb); |
| skb->protocol = eth_type_trans(skb, dev); |
| |
| if (skb->pkt_type == PACKET_MULTICAST) |
| tp->stats.multicast++; |
| |
| rtase_rx_vlan_skb(desc, skb); |
| rtase_rx_skb(ring, skb); |
| |
| dev_sw_netstats_rx_add(dev, pkt_size); |
| |
| skip_process_pkt: |
| workdone++; |
| cur_rx++; |
| entry = cur_rx % RTASE_NUM_DESC; |
| desc = ring->desc + sizeof(union rtase_rx_desc) * entry; |
| } |
| |
| ring->cur_idx = cur_rx; |
| delta = rtase_rx_ring_fill(ring, ring->dirty_idx, ring->cur_idx); |
| ring->dirty_idx += delta; |
| |
| return workdone; |
| } |
| |
| static void rtase_rx_desc_init(struct rtase_private *tp, u16 idx) |
| { |
| struct rtase_ring *ring = &tp->rx_ring[idx]; |
| u16 i; |
| |
| memset(ring->desc, 0x0, RTASE_RX_RING_DESC_SIZE); |
| memset(ring->data_buf, 0x0, sizeof(ring->data_buf)); |
| ring->cur_idx = 0; |
| ring->dirty_idx = 0; |
| ring->index = idx; |
| ring->alloc_fail = 0; |
| |
| for (i = 0; i < RTASE_NUM_DESC; i++) |
| ring->mis.data_phy_addr[i] = 0; |
| |
| ring->ring_handler = rx_handler; |
| ring->ivec = &tp->int_vector[idx]; |
| list_add_tail(&ring->ring_entry, &tp->int_vector[idx].ring_list); |
| } |
| |
| static void rtase_rx_clear(struct rtase_private *tp) |
| { |
| u32 i; |
| |
| for (i = 0; i < tp->func_rx_queue_num; i++) |
| rtase_rx_ring_clear(tp->page_pool, &tp->rx_ring[i]); |
| |
| page_pool_destroy(tp->page_pool); |
| tp->page_pool = NULL; |
| } |
| |
| static int rtase_init_ring(const struct net_device *dev) |
| { |
| struct rtase_private *tp = netdev_priv(dev); |
| struct page_pool_params pp_params = { 0 }; |
| struct page_pool *page_pool; |
| u32 num; |
| u16 i; |
| |
| pp_params.flags = PP_FLAG_DMA_MAP | PP_FLAG_DMA_SYNC_DEV; |
| pp_params.order = 0; |
| pp_params.pool_size = RTASE_NUM_DESC * tp->func_rx_queue_num; |
| pp_params.nid = dev_to_node(&tp->pdev->dev); |
| pp_params.dev = &tp->pdev->dev; |
| pp_params.dma_dir = DMA_FROM_DEVICE; |
| pp_params.max_len = PAGE_SIZE; |
| pp_params.offset = 0; |
| |
| page_pool = page_pool_create(&pp_params); |
| if (IS_ERR(page_pool)) { |
| netdev_err(tp->dev, "failed to create page pool\n"); |
| return -ENOMEM; |
| } |
| |
| tp->page_pool = page_pool; |
| |
| for (i = 0; i < tp->func_tx_queue_num; i++) |
| rtase_tx_desc_init(tp, i); |
| |
| for (i = 0; i < tp->func_rx_queue_num; i++) { |
| rtase_rx_desc_init(tp, i); |
| |
| num = rtase_rx_ring_fill(&tp->rx_ring[i], 0, RTASE_NUM_DESC); |
| if (num != RTASE_NUM_DESC) |
| goto err_out; |
| |
| rtase_mark_as_last_descriptor(tp->rx_ring[i].desc + |
| sizeof(union rtase_rx_desc) * |
| (RTASE_NUM_DESC - 1)); |
| } |
| |
| return 0; |
| |
| err_out: |
| rtase_rx_clear(tp); |
| return -ENOMEM; |
| } |
| |
| static void rtase_interrupt_mitigation(const struct rtase_private *tp) |
| { |
| u32 i; |
| |
| for (i = 0; i < tp->func_tx_queue_num; i++) |
| rtase_w16(tp, RTASE_INT_MITI_TX + i * 2, tp->tx_int_mit); |
| |
| for (i = 0; i < tp->func_rx_queue_num; i++) |
| rtase_w16(tp, RTASE_INT_MITI_RX + i * 2, tp->rx_int_mit); |
| } |
| |
| static void rtase_tally_counter_addr_fill(const struct rtase_private *tp) |
| { |
| rtase_w32(tp, RTASE_DTCCR4, upper_32_bits(tp->tally_paddr)); |
| rtase_w32(tp, RTASE_DTCCR0, lower_32_bits(tp->tally_paddr)); |
| } |
| |
| static void rtase_tally_counter_clear(const struct rtase_private *tp) |
| { |
| u32 cmd = lower_32_bits(tp->tally_paddr); |
| |
| rtase_w32(tp, RTASE_DTCCR4, upper_32_bits(tp->tally_paddr)); |
| rtase_w32(tp, RTASE_DTCCR0, cmd | RTASE_COUNTER_RESET); |
| } |
| |
| static void rtase_desc_addr_fill(const struct rtase_private *tp) |
| { |
| const struct rtase_ring *ring; |
| u16 i, cmd, val; |
| int err; |
| |
| for (i = 0; i < tp->func_tx_queue_num; i++) { |
| ring = &tp->tx_ring[i]; |
| |
| rtase_w32(tp, RTASE_TX_DESC_ADDR0, |
| lower_32_bits(ring->phy_addr)); |
| rtase_w32(tp, RTASE_TX_DESC_ADDR4, |
| upper_32_bits(ring->phy_addr)); |
| |
| cmd = i | RTASE_TX_DESC_CMD_WE | RTASE_TX_DESC_CMD_CS; |
| rtase_w16(tp, RTASE_TX_DESC_COMMAND, cmd); |
| |
| err = read_poll_timeout(rtase_r16, val, |
| !(val & RTASE_TX_DESC_CMD_CS), 10, |
| 1000, false, tp, |
| RTASE_TX_DESC_COMMAND); |
| |
| if (err == -ETIMEDOUT) |
| netdev_err(tp->dev, |
| "error occurred in fill tx descriptor\n"); |
| } |
| |
| for (i = 0; i < tp->func_rx_queue_num; i++) { |
| ring = &tp->rx_ring[i]; |
| |
| if (i == 0) { |
| rtase_w32(tp, RTASE_Q0_RX_DESC_ADDR0, |
| lower_32_bits(ring->phy_addr)); |
| rtase_w32(tp, RTASE_Q0_RX_DESC_ADDR4, |
| upper_32_bits(ring->phy_addr)); |
| } else { |
| rtase_w32(tp, (RTASE_Q1_RX_DESC_ADDR0 + ((i - 1) * 8)), |
| lower_32_bits(ring->phy_addr)); |
| rtase_w32(tp, (RTASE_Q1_RX_DESC_ADDR4 + ((i - 1) * 8)), |
| upper_32_bits(ring->phy_addr)); |
| } |
| } |
| } |
| |
| static void rtase_hw_set_features(const struct net_device *dev, |
| netdev_features_t features) |
| { |
| const struct rtase_private *tp = netdev_priv(dev); |
| u16 rx_config, val; |
| |
| rx_config = rtase_r16(tp, RTASE_RX_CONFIG_0); |
| if (features & NETIF_F_RXALL) |
| rx_config |= (RTASE_ACCEPT_ERR | RTASE_ACCEPT_RUNT); |
| else |
| rx_config &= ~(RTASE_ACCEPT_ERR | RTASE_ACCEPT_RUNT); |
| |
| rtase_w16(tp, RTASE_RX_CONFIG_0, rx_config); |
| |
| val = rtase_r16(tp, RTASE_CPLUS_CMD); |
| if (features & NETIF_F_RXCSUM) |
| rtase_w16(tp, RTASE_CPLUS_CMD, val | RTASE_RX_CHKSUM); |
| else |
| rtase_w16(tp, RTASE_CPLUS_CMD, val & ~RTASE_RX_CHKSUM); |
| |
| rx_config = rtase_r16(tp, RTASE_RX_CONFIG_1); |
| if (dev->features & NETIF_F_HW_VLAN_CTAG_RX) |
| rx_config |= (RTASE_INNER_VLAN_DETAG_EN | |
| RTASE_OUTER_VLAN_DETAG_EN); |
| else |
| rx_config &= ~(RTASE_INNER_VLAN_DETAG_EN | |
| RTASE_OUTER_VLAN_DETAG_EN); |
| |
| rtase_w16(tp, RTASE_RX_CONFIG_1, rx_config); |
| } |
| |
| static void rtase_hw_set_rx_packet_filter(struct net_device *dev) |
| { |
| u32 mc_filter[2] = { 0xFFFFFFFF, 0xFFFFFFFF }; |
| struct rtase_private *tp = netdev_priv(dev); |
| u16 rx_mode; |
| |
| rx_mode = rtase_r16(tp, RTASE_RX_CONFIG_0) & ~RTASE_ACCEPT_MASK; |
| rx_mode |= RTASE_ACCEPT_BROADCAST | RTASE_ACCEPT_MYPHYS; |
| |
| if (dev->flags & IFF_PROMISC) { |
| rx_mode |= RTASE_ACCEPT_MULTICAST | RTASE_ACCEPT_ALLPHYS; |
| } else if (dev->flags & IFF_ALLMULTI) { |
| rx_mode |= RTASE_ACCEPT_MULTICAST; |
| } else { |
| struct netdev_hw_addr *hw_addr; |
| |
| mc_filter[0] = 0; |
| mc_filter[1] = 0; |
| |
| netdev_for_each_mc_addr(hw_addr, dev) { |
| u32 bit_nr = eth_hw_addr_crc(hw_addr); |
| u32 idx = u32_get_bits(bit_nr, BIT(31)); |
| u32 bit = u32_get_bits(bit_nr, |
| RTASE_MULTICAST_FILTER_MASK); |
| |
| mc_filter[idx] |= BIT(bit); |
| rx_mode |= RTASE_ACCEPT_MULTICAST; |
| } |
| } |
| |
| if (dev->features & NETIF_F_RXALL) |
| rx_mode |= RTASE_ACCEPT_ERR | RTASE_ACCEPT_RUNT; |
| |
| rtase_w32(tp, RTASE_MAR0, swab32(mc_filter[1])); |
| rtase_w32(tp, RTASE_MAR1, swab32(mc_filter[0])); |
| rtase_w16(tp, RTASE_RX_CONFIG_0, rx_mode); |
| } |
| |
| static void rtase_irq_dis_and_clear(const struct rtase_private *tp) |
| { |
| const struct rtase_int_vector *ivec = &tp->int_vector[0]; |
| u32 val1; |
| u16 val2; |
| u8 i; |
| |
| rtase_w32(tp, ivec->imr_addr, 0); |
| val1 = rtase_r32(tp, ivec->isr_addr); |
| rtase_w32(tp, ivec->isr_addr, val1); |
| |
| for (i = 1; i < tp->int_nums; i++) { |
| ivec = &tp->int_vector[i]; |
| rtase_w16(tp, ivec->imr_addr, 0); |
| val2 = rtase_r16(tp, ivec->isr_addr); |
| rtase_w16(tp, ivec->isr_addr, val2); |
| } |
| } |
| |
| static void rtase_poll_timeout(const struct rtase_private *tp, u32 cond, |
| u32 sleep_us, u64 timeout_us, u16 reg) |
| { |
| int err; |
| u8 val; |
| |
| err = read_poll_timeout(rtase_r8, val, val & cond, sleep_us, |
| timeout_us, false, tp, reg); |
| |
| if (err == -ETIMEDOUT) |
| netdev_err(tp->dev, "poll reg 0x00%x timeout\n", reg); |
| } |
| |
| static void rtase_nic_reset(const struct net_device *dev) |
| { |
| const struct rtase_private *tp = netdev_priv(dev); |
| u16 rx_config; |
| u8 val; |
| |
| rx_config = rtase_r16(tp, RTASE_RX_CONFIG_0); |
| rtase_w16(tp, RTASE_RX_CONFIG_0, rx_config & ~RTASE_ACCEPT_MASK); |
| |
| val = rtase_r8(tp, RTASE_MISC); |
| rtase_w8(tp, RTASE_MISC, val | RTASE_RX_DV_GATE_EN); |
| |
| val = rtase_r8(tp, RTASE_CHIP_CMD); |
| rtase_w8(tp, RTASE_CHIP_CMD, val | RTASE_STOP_REQ); |
| mdelay(2); |
| |
| rtase_poll_timeout(tp, RTASE_STOP_REQ_DONE, 100, 150000, |
| RTASE_CHIP_CMD); |
| |
| rtase_poll_timeout(tp, RTASE_TX_FIFO_EMPTY, 100, 100000, |
| RTASE_FIFOR); |
| |
| rtase_poll_timeout(tp, RTASE_RX_FIFO_EMPTY, 100, 100000, |
| RTASE_FIFOR); |
| |
| val = rtase_r8(tp, RTASE_CHIP_CMD); |
| rtase_w8(tp, RTASE_CHIP_CMD, val & ~(RTASE_TE | RTASE_RE)); |
| val = rtase_r8(tp, RTASE_CHIP_CMD); |
| rtase_w8(tp, RTASE_CHIP_CMD, val & ~RTASE_STOP_REQ); |
| |
| rtase_w16(tp, RTASE_RX_CONFIG_0, rx_config); |
| } |
| |
| static void rtase_hw_reset(const struct net_device *dev) |
| { |
| const struct rtase_private *tp = netdev_priv(dev); |
| |
| rtase_irq_dis_and_clear(tp); |
| |
| rtase_nic_reset(dev); |
| } |
| |
| static void rtase_set_rx_queue(const struct rtase_private *tp) |
| { |
| u16 reg_data; |
| |
| reg_data = rtase_r16(tp, RTASE_FCR); |
| switch (tp->func_rx_queue_num) { |
| case 1: |
| u16p_replace_bits(®_data, 0x1, RTASE_FCR_RXQ_MASK); |
| break; |
| case 2: |
| u16p_replace_bits(®_data, 0x2, RTASE_FCR_RXQ_MASK); |
| break; |
| case 4: |
| u16p_replace_bits(®_data, 0x3, RTASE_FCR_RXQ_MASK); |
| break; |
| } |
| rtase_w16(tp, RTASE_FCR, reg_data); |
| } |
| |
| static void rtase_set_tx_queue(const struct rtase_private *tp) |
| { |
| u16 reg_data; |
| |
| reg_data = rtase_r16(tp, RTASE_TX_CONFIG_1); |
| switch (tp->tx_queue_ctrl) { |
| case 1: |
| u16p_replace_bits(®_data, 0x0, RTASE_TC_MODE_MASK); |
| break; |
| case 2: |
| u16p_replace_bits(®_data, 0x1, RTASE_TC_MODE_MASK); |
| break; |
| case 3: |
| case 4: |
| u16p_replace_bits(®_data, 0x2, RTASE_TC_MODE_MASK); |
| break; |
| default: |
| u16p_replace_bits(®_data, 0x3, RTASE_TC_MODE_MASK); |
| break; |
| } |
| rtase_w16(tp, RTASE_TX_CONFIG_1, reg_data); |
| } |
| |
| static void rtase_hw_config(struct net_device *dev) |
| { |
| const struct rtase_private *tp = netdev_priv(dev); |
| u32 reg_data32; |
| u16 reg_data16; |
| |
| rtase_hw_reset(dev); |
| |
| /* set rx dma burst */ |
| reg_data16 = rtase_r16(tp, RTASE_RX_CONFIG_0); |
| reg_data16 &= ~(RTASE_RX_SINGLE_TAG | RTASE_RX_SINGLE_FETCH); |
| u16p_replace_bits(®_data16, RTASE_RX_DMA_BURST_256, |
| RTASE_RX_MX_DMA_MASK); |
| rtase_w16(tp, RTASE_RX_CONFIG_0, reg_data16); |
| |
| /* new rx descritpor */ |
| reg_data16 = rtase_r16(tp, RTASE_RX_CONFIG_1); |
| reg_data16 |= RTASE_RX_NEW_DESC_FORMAT_EN | RTASE_PCIE_NEW_FLOW; |
| u16p_replace_bits(®_data16, 0xF, RTASE_RX_MAX_FETCH_DESC_MASK); |
| rtase_w16(tp, RTASE_RX_CONFIG_1, reg_data16); |
| |
| rtase_set_rx_queue(tp); |
| |
| rtase_interrupt_mitigation(tp); |
| |
| /* set tx dma burst size and interframe gap time */ |
| reg_data32 = rtase_r32(tp, RTASE_TX_CONFIG_0); |
| u32p_replace_bits(®_data32, RTASE_TX_DMA_BURST_UNLIMITED, |
| RTASE_TX_DMA_MASK); |
| u32p_replace_bits(®_data32, RTASE_INTERFRAMEGAP, |
| RTASE_TX_INTER_FRAME_GAP_MASK); |
| rtase_w32(tp, RTASE_TX_CONFIG_0, reg_data32); |
| |
| /* new tx descriptor */ |
| reg_data16 = rtase_r16(tp, RTASE_TFUN_CTRL); |
| rtase_w16(tp, RTASE_TFUN_CTRL, reg_data16 | |
| RTASE_TX_NEW_DESC_FORMAT_EN); |
| |
| /* tx fetch desc number */ |
| rtase_w8(tp, RTASE_TDFNR, 0x10); |
| |
| /* tag num select */ |
| reg_data16 = rtase_r16(tp, RTASE_MTPS); |
| u16p_replace_bits(®_data16, 0x4, RTASE_TAG_NUM_SEL_MASK); |
| rtase_w16(tp, RTASE_MTPS, reg_data16); |
| |
| rtase_set_tx_queue(tp); |
| |
| rtase_w16(tp, RTASE_TOKSEL, 0x5555); |
| |
| rtase_tally_counter_addr_fill(tp); |
| rtase_desc_addr_fill(tp); |
| rtase_hw_set_features(dev, dev->features); |
| |
| /* enable flow control */ |
| reg_data16 = rtase_r16(tp, RTASE_CPLUS_CMD); |
| reg_data16 |= (RTASE_FORCE_TXFLOW_EN | RTASE_FORCE_RXFLOW_EN); |
| rtase_w16(tp, RTASE_CPLUS_CMD, reg_data16); |
| /* set near fifo threshold - rx missed issue. */ |
| rtase_w16(tp, RTASE_RFIFONFULL, 0x190); |
| |
| rtase_w16(tp, RTASE_RMS, tp->rx_buf_sz); |
| |
| rtase_hw_set_rx_packet_filter(dev); |
| } |
| |
| static void rtase_nic_enable(const struct net_device *dev) |
| { |
| const struct rtase_private *tp = netdev_priv(dev); |
| u16 rcr = rtase_r16(tp, RTASE_RX_CONFIG_1); |
| u8 val; |
| |
| rtase_w16(tp, RTASE_RX_CONFIG_1, rcr & ~RTASE_PCIE_RELOAD_EN); |
| rtase_w16(tp, RTASE_RX_CONFIG_1, rcr | RTASE_PCIE_RELOAD_EN); |
| |
| val = rtase_r8(tp, RTASE_CHIP_CMD); |
| rtase_w8(tp, RTASE_CHIP_CMD, val | RTASE_TE | RTASE_RE); |
| |
| val = rtase_r8(tp, RTASE_MISC); |
| rtase_w8(tp, RTASE_MISC, val & ~RTASE_RX_DV_GATE_EN); |
| } |
| |
| static void rtase_enable_hw_interrupt(const struct rtase_private *tp) |
| { |
| const struct rtase_int_vector *ivec = &tp->int_vector[0]; |
| u32 i; |
| |
| rtase_w32(tp, ivec->imr_addr, ivec->imr); |
| |
| for (i = 1; i < tp->int_nums; i++) { |
| ivec = &tp->int_vector[i]; |
| rtase_w16(tp, ivec->imr_addr, ivec->imr); |
| } |
| } |
| |
| static void rtase_hw_start(const struct net_device *dev) |
| { |
| const struct rtase_private *tp = netdev_priv(dev); |
| |
| rtase_nic_enable(dev); |
| rtase_enable_hw_interrupt(tp); |
| } |
| |
| /* the interrupt handler does RXQ0 and TXQ0, TXQ4~7 interrutp status |
| */ |
| static irqreturn_t rtase_interrupt(int irq, void *dev_instance) |
| { |
| const struct rtase_private *tp; |
| struct rtase_int_vector *ivec; |
| u32 status; |
| |
| ivec = dev_instance; |
| tp = ivec->tp; |
| status = rtase_r32(tp, ivec->isr_addr); |
| |
| rtase_w32(tp, ivec->imr_addr, 0x0); |
| rtase_w32(tp, ivec->isr_addr, status & ~RTASE_FOVW); |
| |
| if (napi_schedule_prep(&ivec->napi)) |
| __napi_schedule(&ivec->napi); |
| |
| return IRQ_HANDLED; |
| } |
| |
| /* the interrupt handler does RXQ1&TXQ1 or RXQ2&TXQ2 or RXQ3&TXQ3 interrupt |
| * status according to interrupt vector |
| */ |
| static irqreturn_t rtase_q_interrupt(int irq, void *dev_instance) |
| { |
| const struct rtase_private *tp; |
| struct rtase_int_vector *ivec; |
| u16 status; |
| |
| ivec = dev_instance; |
| tp = ivec->tp; |
| status = rtase_r16(tp, ivec->isr_addr); |
| |
| rtase_w16(tp, ivec->imr_addr, 0x0); |
| rtase_w16(tp, ivec->isr_addr, status); |
| |
| if (napi_schedule_prep(&ivec->napi)) |
| __napi_schedule(&ivec->napi); |
| |
| return IRQ_HANDLED; |
| } |
| |
| static int rtase_poll(struct napi_struct *napi, int budget) |
| { |
| const struct rtase_int_vector *ivec; |
| const struct rtase_private *tp; |
| struct rtase_ring *ring; |
| int total_workdone = 0; |
| |
| ivec = container_of(napi, struct rtase_int_vector, napi); |
| tp = ivec->tp; |
| |
| list_for_each_entry(ring, &ivec->ring_list, ring_entry) |
| total_workdone += ring->ring_handler(ring, budget); |
| |
| if (total_workdone >= budget) |
| return budget; |
| |
| if (napi_complete_done(napi, total_workdone)) { |
| if (!ivec->index) |
| rtase_w32(tp, ivec->imr_addr, ivec->imr); |
| else |
| rtase_w16(tp, ivec->imr_addr, ivec->imr); |
| } |
| |
| return total_workdone; |
| } |
| |
| static int rtase_open(struct net_device *dev) |
| { |
| struct rtase_private *tp = netdev_priv(dev); |
| const struct pci_dev *pdev = tp->pdev; |
| struct rtase_int_vector *ivec; |
| u16 i = 0, j; |
| int ret; |
| |
| ivec = &tp->int_vector[0]; |
| tp->rx_buf_sz = RTASE_RX_BUF_SIZE; |
| |
| ret = rtase_alloc_desc(tp); |
| if (ret) |
| return ret; |
| |
| ret = rtase_init_ring(dev); |
| if (ret) |
| goto err_free_all_allocated_mem; |
| |
| rtase_hw_config(dev); |
| |
| if (tp->sw_flag & RTASE_SWF_MSIX_ENABLED) { |
| ret = request_irq(ivec->irq, rtase_interrupt, 0, |
| dev->name, ivec); |
| if (ret) |
| goto err_free_all_allocated_irq; |
| |
| /* request other interrupts to handle multiqueue */ |
| for (i = 1; i < tp->int_nums; i++) { |
| ivec = &tp->int_vector[i]; |
| snprintf(ivec->name, sizeof(ivec->name), "%s_int%i", |
| tp->dev->name, i); |
| ret = request_irq(ivec->irq, rtase_q_interrupt, 0, |
| ivec->name, ivec); |
| if (ret) |
| goto err_free_all_allocated_irq; |
| } |
| } else { |
| ret = request_irq(pdev->irq, rtase_interrupt, 0, dev->name, |
| ivec); |
| if (ret) |
| goto err_free_all_allocated_mem; |
| } |
| |
| rtase_hw_start(dev); |
| |
| for (i = 0; i < tp->int_nums; i++) { |
| ivec = &tp->int_vector[i]; |
| napi_enable(&ivec->napi); |
| } |
| |
| netif_carrier_on(dev); |
| netif_wake_queue(dev); |
| |
| return 0; |
| |
| err_free_all_allocated_irq: |
| for (j = 0; j < i; j++) |
| free_irq(tp->int_vector[j].irq, &tp->int_vector[j]); |
| |
| err_free_all_allocated_mem: |
| rtase_free_desc(tp); |
| |
| return ret; |
| } |
| |
| static void rtase_down(struct net_device *dev) |
| { |
| struct rtase_private *tp = netdev_priv(dev); |
| struct rtase_int_vector *ivec; |
| struct rtase_ring *ring, *tmp; |
| u32 i; |
| |
| for (i = 0; i < tp->int_nums; i++) { |
| ivec = &tp->int_vector[i]; |
| napi_disable(&ivec->napi); |
| list_for_each_entry_safe(ring, tmp, &ivec->ring_list, |
| ring_entry) |
| list_del(&ring->ring_entry); |
| } |
| |
| netif_tx_disable(dev); |
| |
| netif_carrier_off(dev); |
| |
| rtase_hw_reset(dev); |
| |
| rtase_tx_clear(tp); |
| |
| rtase_rx_clear(tp); |
| } |
| |
| static int rtase_close(struct net_device *dev) |
| { |
| struct rtase_private *tp = netdev_priv(dev); |
| const struct pci_dev *pdev = tp->pdev; |
| u32 i; |
| |
| rtase_down(dev); |
| |
| if (tp->sw_flag & RTASE_SWF_MSIX_ENABLED) { |
| for (i = 0; i < tp->int_nums; i++) |
| free_irq(tp->int_vector[i].irq, &tp->int_vector[i]); |
| |
| } else { |
| free_irq(pdev->irq, &tp->int_vector[0]); |
| } |
| |
| rtase_free_desc(tp); |
| |
| return 0; |
| } |
| |
| static u32 rtase_tx_vlan_tag(const struct rtase_private *tp, |
| const struct sk_buff *skb) |
| { |
| return (skb_vlan_tag_present(skb)) ? |
| (RTASE_TX_VLAN_TAG | swab16(skb_vlan_tag_get(skb))) : 0x00; |
| } |
| |
| static u32 rtase_tx_csum(struct sk_buff *skb, const struct net_device *dev) |
| { |
| u32 csum_cmd = 0; |
| u8 ip_protocol; |
| |
| switch (vlan_get_protocol(skb)) { |
| case htons(ETH_P_IP): |
| csum_cmd = RTASE_TX_IPCS_C; |
| ip_protocol = ip_hdr(skb)->protocol; |
| break; |
| |
| case htons(ETH_P_IPV6): |
| csum_cmd = RTASE_TX_IPV6F_C; |
| ip_protocol = ipv6_hdr(skb)->nexthdr; |
| break; |
| |
| default: |
| ip_protocol = IPPROTO_RAW; |
| break; |
| } |
| |
| if (ip_protocol == IPPROTO_TCP) |
| csum_cmd |= RTASE_TX_TCPCS_C; |
| else if (ip_protocol == IPPROTO_UDP) |
| csum_cmd |= RTASE_TX_UDPCS_C; |
| |
| csum_cmd |= u32_encode_bits(skb_transport_offset(skb), |
| RTASE_TCPHO_MASK); |
| |
| return csum_cmd; |
| } |
| |
| static int rtase_xmit_frags(struct rtase_ring *ring, struct sk_buff *skb, |
| u32 opts1, u32 opts2) |
| { |
| const struct skb_shared_info *info = skb_shinfo(skb); |
| const struct rtase_private *tp = ring->ivec->tp; |
| const u8 nr_frags = info->nr_frags; |
| struct rtase_tx_desc *txd = NULL; |
| u32 cur_frag, entry; |
| |
| entry = ring->cur_idx; |
| for (cur_frag = 0; cur_frag < nr_frags; cur_frag++) { |
| const skb_frag_t *frag = &info->frags[cur_frag]; |
| dma_addr_t mapping; |
| u32 status, len; |
| void *addr; |
| |
| entry = (entry + 1) % RTASE_NUM_DESC; |
| |
| txd = ring->desc + sizeof(struct rtase_tx_desc) * entry; |
| len = skb_frag_size(frag); |
| addr = skb_frag_address(frag); |
| mapping = dma_map_single(&tp->pdev->dev, addr, len, |
| DMA_TO_DEVICE); |
| |
| if (unlikely(dma_mapping_error(&tp->pdev->dev, mapping))) { |
| if (unlikely(net_ratelimit())) |
| netdev_err(tp->dev, |
| "Failed to map TX fragments DMA!\n"); |
| |
| goto err_out; |
| } |
| |
| if (((entry + 1) % RTASE_NUM_DESC) == 0) |
| status = (opts1 | len | RTASE_RING_END); |
| else |
| status = opts1 | len; |
| |
| if (cur_frag == (nr_frags - 1)) { |
| ring->skbuff[entry] = skb; |
| status |= RTASE_TX_LAST_FRAG; |
| } |
| |
| ring->mis.len[entry] = len; |
| txd->addr = cpu_to_le64(mapping); |
| txd->opts2 = cpu_to_le32(opts2); |
| |
| /* make sure the operating fields have been updated */ |
| dma_wmb(); |
| txd->opts1 = cpu_to_le32(status); |
| } |
| |
| return cur_frag; |
| |
| err_out: |
| rtase_tx_clear_range(ring, ring->cur_idx + 1, cur_frag); |
| return -EIO; |
| } |
| |
| static netdev_tx_t rtase_start_xmit(struct sk_buff *skb, |
| struct net_device *dev) |
| { |
| struct skb_shared_info *shinfo = skb_shinfo(skb); |
| struct rtase_private *tp = netdev_priv(dev); |
| u32 q_idx, entry, len, opts1, opts2; |
| struct netdev_queue *tx_queue; |
| bool stop_queue, door_bell; |
| u32 mss = shinfo->gso_size; |
| struct rtase_tx_desc *txd; |
| struct rtase_ring *ring; |
| dma_addr_t mapping; |
| int frags; |
| |
| /* multiqueues */ |
| q_idx = skb_get_queue_mapping(skb); |
| ring = &tp->tx_ring[q_idx]; |
| tx_queue = netdev_get_tx_queue(dev, q_idx); |
| |
| if (unlikely(!rtase_tx_avail(ring))) { |
| if (net_ratelimit()) |
| netdev_err(dev, |
| "BUG! Tx Ring full when queue awake!\n"); |
| |
| netif_stop_queue(dev); |
| return NETDEV_TX_BUSY; |
| } |
| |
| entry = ring->cur_idx % RTASE_NUM_DESC; |
| txd = ring->desc + sizeof(struct rtase_tx_desc) * entry; |
| |
| opts1 = RTASE_DESC_OWN; |
| opts2 = rtase_tx_vlan_tag(tp, skb); |
| |
| /* tcp segmentation offload (or tcp large send) */ |
| if (mss) { |
| if (shinfo->gso_type & SKB_GSO_TCPV4) { |
| opts1 |= RTASE_GIANT_SEND_V4; |
| } else if (shinfo->gso_type & SKB_GSO_TCPV6) { |
| if (skb_cow_head(skb, 0)) |
| goto err_dma_0; |
| |
| tcp_v6_gso_csum_prep(skb); |
| opts1 |= RTASE_GIANT_SEND_V6; |
| } else { |
| WARN_ON_ONCE(1); |
| } |
| |
| opts1 |= u32_encode_bits(skb_transport_offset(skb), |
| RTASE_TCPHO_MASK); |
| opts2 |= u32_encode_bits(mss, RTASE_MSS_MASK); |
| } else if (skb->ip_summed == CHECKSUM_PARTIAL) { |
| opts2 |= rtase_tx_csum(skb, dev); |
| } |
| |
| frags = rtase_xmit_frags(ring, skb, opts1, opts2); |
| if (unlikely(frags < 0)) |
| goto err_dma_0; |
| |
| if (frags) { |
| len = skb_headlen(skb); |
| opts1 |= RTASE_TX_FIRST_FRAG; |
| } else { |
| len = skb->len; |
| ring->skbuff[entry] = skb; |
| opts1 |= RTASE_TX_FIRST_FRAG | RTASE_TX_LAST_FRAG; |
| } |
| |
| if (((entry + 1) % RTASE_NUM_DESC) == 0) |
| opts1 |= (len | RTASE_RING_END); |
| else |
| opts1 |= len; |
| |
| mapping = dma_map_single(&tp->pdev->dev, skb->data, len, |
| DMA_TO_DEVICE); |
| |
| if (unlikely(dma_mapping_error(&tp->pdev->dev, mapping))) { |
| if (unlikely(net_ratelimit())) |
| netdev_err(dev, "Failed to map TX DMA!\n"); |
| |
| goto err_dma_1; |
| } |
| |
| ring->mis.len[entry] = len; |
| txd->addr = cpu_to_le64(mapping); |
| txd->opts2 = cpu_to_le32(opts2); |
| txd->opts1 = cpu_to_le32(opts1 & ~RTASE_DESC_OWN); |
| |
| /* make sure the operating fields have been updated */ |
| dma_wmb(); |
| |
| door_bell = __netdev_tx_sent_queue(tx_queue, skb->len, |
| netdev_xmit_more()); |
| |
| txd->opts1 = cpu_to_le32(opts1); |
| |
| skb_tx_timestamp(skb); |
| |
| /* tx needs to see descriptor changes before updated cur_idx */ |
| smp_wmb(); |
| |
| WRITE_ONCE(ring->cur_idx, ring->cur_idx + frags + 1); |
| |
| stop_queue = !netif_subqueue_maybe_stop(dev, ring->index, |
| rtase_tx_avail(ring), |
| RTASE_TX_STOP_THRS, |
| RTASE_TX_START_THRS); |
| |
| if (door_bell || stop_queue) |
| rtase_w8(tp, RTASE_TPPOLL, BIT(ring->index)); |
| |
| return NETDEV_TX_OK; |
| |
| err_dma_1: |
| ring->skbuff[entry] = NULL; |
| rtase_tx_clear_range(ring, ring->cur_idx + 1, frags); |
| |
| err_dma_0: |
| tp->stats.tx_dropped++; |
| dev_kfree_skb_any(skb); |
| return NETDEV_TX_OK; |
| } |
| |
| static void rtase_set_rx_mode(struct net_device *dev) |
| { |
| rtase_hw_set_rx_packet_filter(dev); |
| } |
| |
| static void rtase_enable_eem_write(const struct rtase_private *tp) |
| { |
| u8 val; |
| |
| val = rtase_r8(tp, RTASE_EEM); |
| rtase_w8(tp, RTASE_EEM, val | RTASE_EEM_UNLOCK); |
| } |
| |
| static void rtase_disable_eem_write(const struct rtase_private *tp) |
| { |
| u8 val; |
| |
| val = rtase_r8(tp, RTASE_EEM); |
| rtase_w8(tp, RTASE_EEM, val & ~RTASE_EEM_UNLOCK); |
| } |
| |
| static void rtase_rar_set(const struct rtase_private *tp, const u8 *addr) |
| { |
| u32 rar_low, rar_high; |
| |
| rar_low = (u32)addr[0] | ((u32)addr[1] << 8) | |
| ((u32)addr[2] << 16) | ((u32)addr[3] << 24); |
| |
| rar_high = (u32)addr[4] | ((u32)addr[5] << 8); |
| |
| rtase_enable_eem_write(tp); |
| rtase_w32(tp, RTASE_MAC0, rar_low); |
| rtase_w32(tp, RTASE_MAC4, rar_high); |
| rtase_disable_eem_write(tp); |
| rtase_w16(tp, RTASE_LBK_CTRL, RTASE_LBK_ATLD | RTASE_LBK_CLR); |
| } |
| |
| static int rtase_set_mac_address(struct net_device *dev, void *p) |
| { |
| struct rtase_private *tp = netdev_priv(dev); |
| int ret; |
| |
| ret = eth_mac_addr(dev, p); |
| if (ret) |
| return ret; |
| |
| rtase_rar_set(tp, dev->dev_addr); |
| |
| return 0; |
| } |
| |
| static int rtase_change_mtu(struct net_device *dev, int new_mtu) |
| { |
| dev->mtu = new_mtu; |
| |
| netdev_update_features(dev); |
| |
| return 0; |
| } |
| |
| static void rtase_wait_for_quiescence(const struct net_device *dev) |
| { |
| struct rtase_private *tp = netdev_priv(dev); |
| struct rtase_int_vector *ivec; |
| u32 i; |
| |
| for (i = 0; i < tp->int_nums; i++) { |
| ivec = &tp->int_vector[i]; |
| synchronize_irq(ivec->irq); |
| /* wait for any pending NAPI task to complete */ |
| napi_disable(&ivec->napi); |
| } |
| |
| rtase_irq_dis_and_clear(tp); |
| |
| for (i = 0; i < tp->int_nums; i++) { |
| ivec = &tp->int_vector[i]; |
| napi_enable(&ivec->napi); |
| } |
| } |
| |
| static void rtase_sw_reset(struct net_device *dev) |
| { |
| struct rtase_private *tp = netdev_priv(dev); |
| int ret; |
| |
| netif_stop_queue(dev); |
| netif_carrier_off(dev); |
| rtase_hw_reset(dev); |
| |
| /* let's wait a bit while any (async) irq lands on */ |
| rtase_wait_for_quiescence(dev); |
| rtase_tx_clear(tp); |
| rtase_rx_clear(tp); |
| |
| ret = rtase_init_ring(dev); |
| if (ret) { |
| netdev_err(dev, "unable to init ring\n"); |
| rtase_free_desc(tp); |
| return; |
| } |
| |
| rtase_hw_config(dev); |
| /* always link, so start to transmit & receive */ |
| rtase_hw_start(dev); |
| |
| netif_carrier_on(dev); |
| netif_wake_queue(dev); |
| } |
| |
| static void rtase_dump_tally_counter(const struct rtase_private *tp) |
| { |
| dma_addr_t paddr = tp->tally_paddr; |
| u32 cmd = lower_32_bits(paddr); |
| u32 val; |
| int err; |
| |
| rtase_w32(tp, RTASE_DTCCR4, upper_32_bits(paddr)); |
| rtase_w32(tp, RTASE_DTCCR0, cmd); |
| rtase_w32(tp, RTASE_DTCCR0, cmd | RTASE_COUNTER_DUMP); |
| |
| err = read_poll_timeout(rtase_r32, val, !(val & RTASE_COUNTER_DUMP), |
| 10, 250, false, tp, RTASE_DTCCR0); |
| |
| if (err == -ETIMEDOUT) |
| netdev_err(tp->dev, "error occurred in dump tally counter\n"); |
| } |
| |
| static void rtase_dump_state(const struct net_device *dev) |
| { |
| const struct rtase_private *tp = netdev_priv(dev); |
| int max_reg_size = RTASE_PCI_REGS_SIZE; |
| const struct rtase_counters *counters; |
| const struct rtase_ring *ring; |
| u32 dword_rd; |
| int n = 0; |
| |
| ring = &tp->tx_ring[0]; |
| netdev_err(dev, "Tx descriptor info:\n"); |
| netdev_err(dev, "Tx curIdx = 0x%x\n", ring->cur_idx); |
| netdev_err(dev, "Tx dirtyIdx = 0x%x\n", ring->dirty_idx); |
| netdev_err(dev, "Tx phyAddr = %pad\n", &ring->phy_addr); |
| |
| ring = &tp->rx_ring[0]; |
| netdev_err(dev, "Rx descriptor info:\n"); |
| netdev_err(dev, "Rx curIdx = 0x%x\n", ring->cur_idx); |
| netdev_err(dev, "Rx dirtyIdx = 0x%x\n", ring->dirty_idx); |
| netdev_err(dev, "Rx phyAddr = %pad\n", &ring->phy_addr); |
| |
| netdev_err(dev, "Device Registers:\n"); |
| netdev_err(dev, "Chip Command = 0x%02x\n", |
| rtase_r8(tp, RTASE_CHIP_CMD)); |
| netdev_err(dev, "IMR = %08x\n", rtase_r32(tp, RTASE_IMR0)); |
| netdev_err(dev, "ISR = %08x\n", rtase_r32(tp, RTASE_ISR0)); |
| netdev_err(dev, "Boot Ctrl Reg(0xE004) = %04x\n", |
| rtase_r16(tp, RTASE_BOOT_CTL)); |
| netdev_err(dev, "EPHY ISR(0xE014) = %04x\n", |
| rtase_r16(tp, RTASE_EPHY_ISR)); |
| netdev_err(dev, "EPHY IMR(0xE016) = %04x\n", |
| rtase_r16(tp, RTASE_EPHY_IMR)); |
| netdev_err(dev, "CLKSW SET REG(0xE018) = %04x\n", |
| rtase_r16(tp, RTASE_CLKSW_SET)); |
| |
| netdev_err(dev, "Dump PCI Registers:\n"); |
| |
| while (n < max_reg_size) { |
| if ((n % RTASE_DWORD_MOD) == 0) |
| netdev_err(tp->dev, "0x%03x:\n", n); |
| |
| pci_read_config_dword(tp->pdev, n, &dword_rd); |
| netdev_err(tp->dev, "%08x\n", dword_rd); |
| n += 4; |
| } |
| |
| netdev_err(dev, "Dump tally counter:\n"); |
| counters = tp->tally_vaddr; |
| rtase_dump_tally_counter(tp); |
| |
| netdev_err(dev, "tx_packets %lld\n", |
| le64_to_cpu(counters->tx_packets)); |
| netdev_err(dev, "rx_packets %lld\n", |
| le64_to_cpu(counters->rx_packets)); |
| netdev_err(dev, "tx_errors %lld\n", |
| le64_to_cpu(counters->tx_errors)); |
| netdev_err(dev, "rx_errors %d\n", |
| le32_to_cpu(counters->rx_errors)); |
| netdev_err(dev, "rx_missed %d\n", |
| le16_to_cpu(counters->rx_missed)); |
| netdev_err(dev, "align_errors %d\n", |
| le16_to_cpu(counters->align_errors)); |
| netdev_err(dev, "tx_one_collision %d\n", |
| le32_to_cpu(counters->tx_one_collision)); |
| netdev_err(dev, "tx_multi_collision %d\n", |
| le32_to_cpu(counters->tx_multi_collision)); |
| netdev_err(dev, "rx_unicast %lld\n", |
| le64_to_cpu(counters->rx_unicast)); |
| netdev_err(dev, "rx_broadcast %lld\n", |
| le64_to_cpu(counters->rx_broadcast)); |
| netdev_err(dev, "rx_multicast %d\n", |
| le32_to_cpu(counters->rx_multicast)); |
| netdev_err(dev, "tx_aborted %d\n", |
| le16_to_cpu(counters->tx_aborted)); |
| netdev_err(dev, "tx_underrun %d\n", |
| le16_to_cpu(counters->tx_underrun)); |
| } |
| |
| static void rtase_tx_timeout(struct net_device *dev, unsigned int txqueue) |
| { |
| rtase_dump_state(dev); |
| rtase_sw_reset(dev); |
| } |
| |
| static void rtase_get_stats64(struct net_device *dev, |
| struct rtnl_link_stats64 *stats) |
| { |
| const struct rtase_private *tp = netdev_priv(dev); |
| const struct rtase_counters *counters; |
| |
| counters = tp->tally_vaddr; |
| |
| dev_fetch_sw_netstats(stats, dev->tstats); |
| |
| /* fetch additional counter values missing in stats collected by driver |
| * from tally counter |
| */ |
| rtase_dump_tally_counter(tp); |
| stats->rx_errors = tp->stats.rx_errors; |
| stats->tx_errors = le64_to_cpu(counters->tx_errors); |
| stats->rx_dropped = tp->stats.rx_dropped; |
| stats->tx_dropped = tp->stats.tx_dropped; |
| stats->multicast = tp->stats.multicast; |
| stats->rx_length_errors = tp->stats.rx_length_errors; |
| } |
| |
| static netdev_features_t rtase_fix_features(struct net_device *dev, |
| netdev_features_t features) |
| { |
| netdev_features_t features_fix = features; |
| |
| /* not support TSO for jumbo frames */ |
| if (dev->mtu > ETH_DATA_LEN) |
| features_fix &= ~NETIF_F_ALL_TSO; |
| |
| return features_fix; |
| } |
| |
| static int rtase_set_features(struct net_device *dev, |
| netdev_features_t features) |
| { |
| netdev_features_t features_set = features; |
| |
| features_set &= NETIF_F_RXALL | NETIF_F_RXCSUM | |
| NETIF_F_HW_VLAN_CTAG_RX; |
| |
| if (features_set ^ dev->features) |
| rtase_hw_set_features(dev, features_set); |
| |
| return 0; |
| } |
| |
| static const struct net_device_ops rtase_netdev_ops = { |
| .ndo_open = rtase_open, |
| .ndo_stop = rtase_close, |
| .ndo_start_xmit = rtase_start_xmit, |
| .ndo_set_rx_mode = rtase_set_rx_mode, |
| .ndo_set_mac_address = rtase_set_mac_address, |
| .ndo_change_mtu = rtase_change_mtu, |
| .ndo_tx_timeout = rtase_tx_timeout, |
| .ndo_get_stats64 = rtase_get_stats64, |
| .ndo_fix_features = rtase_fix_features, |
| .ndo_set_features = rtase_set_features, |
| }; |
| |
| static void rtase_get_mac_address(struct net_device *dev) |
| { |
| struct rtase_private *tp = netdev_priv(dev); |
| u8 mac_addr[ETH_ALEN] __aligned(2) = {}; |
| u32 i; |
| |
| for (i = 0; i < ETH_ALEN; i++) |
| mac_addr[i] = rtase_r8(tp, RTASE_MAC0 + i); |
| |
| if (!is_valid_ether_addr(mac_addr)) { |
| eth_hw_addr_random(dev); |
| netdev_warn(dev, "Random ether addr %pM\n", dev->dev_addr); |
| } else { |
| eth_hw_addr_set(dev, mac_addr); |
| ether_addr_copy(dev->perm_addr, dev->dev_addr); |
| } |
| |
| rtase_rar_set(tp, dev->dev_addr); |
| } |
| |
| static int rtase_get_settings(struct net_device *dev, |
| struct ethtool_link_ksettings *cmd) |
| { |
| u32 supported = SUPPORTED_MII | SUPPORTED_Pause | SUPPORTED_Asym_Pause; |
| |
| ethtool_convert_legacy_u32_to_link_mode(cmd->link_modes.supported, |
| supported); |
| cmd->base.speed = SPEED_5000; |
| cmd->base.duplex = DUPLEX_FULL; |
| cmd->base.port = PORT_MII; |
| cmd->base.autoneg = AUTONEG_DISABLE; |
| |
| return 0; |
| } |
| |
| static void rtase_get_pauseparam(struct net_device *dev, |
| struct ethtool_pauseparam *pause) |
| { |
| const struct rtase_private *tp = netdev_priv(dev); |
| u16 value = rtase_r16(tp, RTASE_CPLUS_CMD); |
| |
| pause->autoneg = AUTONEG_DISABLE; |
| pause->tx_pause = !!(value & RTASE_FORCE_TXFLOW_EN); |
| pause->rx_pause = !!(value & RTASE_FORCE_RXFLOW_EN); |
| } |
| |
| static int rtase_set_pauseparam(struct net_device *dev, |
| struct ethtool_pauseparam *pause) |
| { |
| const struct rtase_private *tp = netdev_priv(dev); |
| u16 value = rtase_r16(tp, RTASE_CPLUS_CMD); |
| |
| if (pause->autoneg) |
| return -EOPNOTSUPP; |
| |
| value &= ~(RTASE_FORCE_TXFLOW_EN | RTASE_FORCE_RXFLOW_EN); |
| |
| if (pause->tx_pause) |
| value |= RTASE_FORCE_TXFLOW_EN; |
| |
| if (pause->rx_pause) |
| value |= RTASE_FORCE_RXFLOW_EN; |
| |
| rtase_w16(tp, RTASE_CPLUS_CMD, value); |
| return 0; |
| } |
| |
| static void rtase_get_eth_mac_stats(struct net_device *dev, |
| struct ethtool_eth_mac_stats *stats) |
| { |
| struct rtase_private *tp = netdev_priv(dev); |
| const struct rtase_counters *counters; |
| |
| counters = tp->tally_vaddr; |
| |
| rtase_dump_tally_counter(tp); |
| |
| stats->FramesTransmittedOK = le64_to_cpu(counters->tx_packets); |
| stats->FramesReceivedOK = le64_to_cpu(counters->rx_packets); |
| stats->FramesLostDueToIntMACXmitError = |
| le64_to_cpu(counters->tx_errors); |
| stats->BroadcastFramesReceivedOK = le64_to_cpu(counters->rx_broadcast); |
| } |
| |
| static const struct ethtool_ops rtase_ethtool_ops = { |
| .get_link = ethtool_op_get_link, |
| .get_link_ksettings = rtase_get_settings, |
| .get_pauseparam = rtase_get_pauseparam, |
| .set_pauseparam = rtase_set_pauseparam, |
| .get_eth_mac_stats = rtase_get_eth_mac_stats, |
| .get_ts_info = ethtool_op_get_ts_info, |
| }; |
| |
| static void rtase_init_netdev_ops(struct net_device *dev) |
| { |
| dev->netdev_ops = &rtase_netdev_ops; |
| dev->ethtool_ops = &rtase_ethtool_ops; |
| } |
| |
| static void rtase_reset_interrupt(struct pci_dev *pdev, |
| const struct rtase_private *tp) |
| { |
| if (tp->sw_flag & RTASE_SWF_MSIX_ENABLED) |
| pci_disable_msix(pdev); |
| else |
| pci_disable_msi(pdev); |
| } |
| |
| static int rtase_alloc_msix(struct pci_dev *pdev, struct rtase_private *tp) |
| { |
| int ret, irq; |
| u16 i; |
| |
| memset(tp->msix_entry, 0x0, RTASE_NUM_MSIX * |
| sizeof(struct msix_entry)); |
| |
| for (i = 0; i < RTASE_NUM_MSIX; i++) |
| tp->msix_entry[i].entry = i; |
| |
| ret = pci_enable_msix_exact(pdev, tp->msix_entry, tp->int_nums); |
| |
| if (ret) |
| return ret; |
| |
| for (i = 0; i < tp->int_nums; i++) { |
| irq = pci_irq_vector(pdev, i); |
| if (!irq) { |
| pci_disable_msix(pdev); |
| return irq; |
| } |
| |
| tp->int_vector[i].irq = irq; |
| } |
| |
| return 0; |
| } |
| |
| static int rtase_alloc_interrupt(struct pci_dev *pdev, |
| struct rtase_private *tp) |
| { |
| int ret; |
| |
| ret = rtase_alloc_msix(pdev, tp); |
| if (ret) { |
| ret = pci_enable_msi(pdev); |
| if (ret) { |
| dev_err(&pdev->dev, |
| "unable to alloc interrupt.(MSI)\n"); |
| return ret; |
| } |
| |
| tp->sw_flag |= RTASE_SWF_MSI_ENABLED; |
| } else { |
| tp->sw_flag |= RTASE_SWF_MSIX_ENABLED; |
| } |
| |
| return 0; |
| } |
| |
| static void rtase_init_hardware(const struct rtase_private *tp) |
| { |
| u16 i; |
| |
| for (i = 0; i < RTASE_VLAN_FILTER_ENTRY_NUM; i++) |
| rtase_w32(tp, RTASE_VLAN_ENTRY_0 + i * 4, 0); |
| } |
| |
| static void rtase_init_int_vector(struct rtase_private *tp) |
| { |
| u16 i; |
| |
| /* interrupt vector 0 */ |
| tp->int_vector[0].tp = tp; |
| tp->int_vector[0].index = 0; |
| tp->int_vector[0].imr_addr = RTASE_IMR0; |
| tp->int_vector[0].isr_addr = RTASE_ISR0; |
| tp->int_vector[0].imr = RTASE_ROK | RTASE_RDU | RTASE_TOK | |
| RTASE_TOK4 | RTASE_TOK5 | RTASE_TOK6 | |
| RTASE_TOK7; |
| tp->int_vector[0].poll = rtase_poll; |
| |
| memset(tp->int_vector[0].name, 0x0, sizeof(tp->int_vector[0].name)); |
| INIT_LIST_HEAD(&tp->int_vector[0].ring_list); |
| |
| netif_napi_add(tp->dev, &tp->int_vector[0].napi, |
| tp->int_vector[0].poll); |
| |
| /* interrupt vector 1 ~ 3 */ |
| for (i = 1; i < tp->int_nums; i++) { |
| tp->int_vector[i].tp = tp; |
| tp->int_vector[i].index = i; |
| tp->int_vector[i].imr_addr = RTASE_IMR1 + (i - 1) * 4; |
| tp->int_vector[i].isr_addr = RTASE_ISR1 + (i - 1) * 4; |
| tp->int_vector[i].imr = RTASE_Q_ROK | RTASE_Q_RDU | |
| RTASE_Q_TOK; |
| tp->int_vector[i].poll = rtase_poll; |
| |
| memset(tp->int_vector[i].name, 0x0, |
| sizeof(tp->int_vector[0].name)); |
| INIT_LIST_HEAD(&tp->int_vector[i].ring_list); |
| |
| netif_napi_add(tp->dev, &tp->int_vector[i].napi, |
| tp->int_vector[i].poll); |
| } |
| } |
| |
| static u16 rtase_calc_time_mitigation(u32 time_us) |
| { |
| u8 msb, time_count, time_unit; |
| u16 int_miti; |
| |
| time_us = min_t(int, time_us, RTASE_MITI_MAX_TIME); |
| |
| msb = fls(time_us); |
| if (msb >= RTASE_MITI_COUNT_BIT_NUM) { |
| time_unit = msb - RTASE_MITI_COUNT_BIT_NUM; |
| time_count = time_us >> (msb - RTASE_MITI_COUNT_BIT_NUM); |
| } else { |
| time_unit = 0; |
| time_count = time_us; |
| } |
| |
| int_miti = u16_encode_bits(time_count, RTASE_MITI_TIME_COUNT_MASK) | |
| u16_encode_bits(time_unit, RTASE_MITI_TIME_UNIT_MASK); |
| |
| return int_miti; |
| } |
| |
| static u16 rtase_calc_packet_num_mitigation(u16 pkt_num) |
| { |
| u8 msb, pkt_num_count, pkt_num_unit; |
| u16 int_miti; |
| |
| pkt_num = min_t(int, pkt_num, RTASE_MITI_MAX_PKT_NUM); |
| |
| if (pkt_num > 60) { |
| pkt_num_unit = RTASE_MITI_MAX_PKT_NUM_IDX; |
| pkt_num_count = pkt_num / RTASE_MITI_MAX_PKT_NUM_UNIT; |
| } else { |
| msb = fls(pkt_num); |
| if (msb >= RTASE_MITI_COUNT_BIT_NUM) { |
| pkt_num_unit = msb - RTASE_MITI_COUNT_BIT_NUM; |
| pkt_num_count = pkt_num >> (msb - |
| RTASE_MITI_COUNT_BIT_NUM); |
| } else { |
| pkt_num_unit = 0; |
| pkt_num_count = pkt_num; |
| } |
| } |
| |
| int_miti = u16_encode_bits(pkt_num_count, |
| RTASE_MITI_PKT_NUM_COUNT_MASK) | |
| u16_encode_bits(pkt_num_unit, |
| RTASE_MITI_PKT_NUM_UNIT_MASK); |
| |
| return int_miti; |
| } |
| |
| static void rtase_init_software_variable(struct pci_dev *pdev, |
| struct rtase_private *tp) |
| { |
| u16 int_miti; |
| |
| tp->tx_queue_ctrl = RTASE_TXQ_CTRL; |
| tp->func_tx_queue_num = RTASE_FUNC_TXQ_NUM; |
| tp->func_rx_queue_num = RTASE_FUNC_RXQ_NUM; |
| tp->int_nums = RTASE_INTERRUPT_NUM; |
| |
| int_miti = rtase_calc_time_mitigation(RTASE_MITI_DEFAULT_TIME) | |
| rtase_calc_packet_num_mitigation(RTASE_MITI_DEFAULT_PKT_NUM); |
| tp->tx_int_mit = int_miti; |
| tp->rx_int_mit = int_miti; |
| |
| tp->sw_flag = 0; |
| |
| rtase_init_int_vector(tp); |
| |
| /* MTU range: 60 - hw-specific max */ |
| tp->dev->min_mtu = ETH_ZLEN; |
| tp->dev->max_mtu = RTASE_MAX_JUMBO_SIZE; |
| } |
| |
| static bool rtase_check_mac_version_valid(struct rtase_private *tp) |
| { |
| u32 hw_ver = rtase_r32(tp, RTASE_TX_CONFIG_0) & RTASE_HW_VER_MASK; |
| bool known_ver = false; |
| |
| switch (hw_ver) { |
| case 0x00800000: |
| case 0x04000000: |
| case 0x04800000: |
| known_ver = true; |
| break; |
| } |
| |
| return known_ver; |
| } |
| |
| static int rtase_init_board(struct pci_dev *pdev, struct net_device **dev_out, |
| void __iomem **ioaddr_out) |
| { |
| struct net_device *dev; |
| void __iomem *ioaddr; |
| int ret = -ENOMEM; |
| |
| /* dev zeroed in alloc_etherdev */ |
| dev = alloc_etherdev_mq(sizeof(struct rtase_private), |
| RTASE_FUNC_TXQ_NUM); |
| if (!dev) |
| goto err_out; |
| |
| SET_NETDEV_DEV(dev, &pdev->dev); |
| |
| ret = pci_enable_device(pdev); |
| if (ret < 0) |
| goto err_out_free_dev; |
| |
| /* make sure PCI base addr 1 is MMIO */ |
| if (!(pci_resource_flags(pdev, 2) & IORESOURCE_MEM)) { |
| ret = -ENODEV; |
| goto err_out_disable; |
| } |
| |
| /* check for weird/broken PCI region reporting */ |
| if (pci_resource_len(pdev, 2) < RTASE_REGS_SIZE) { |
| ret = -ENODEV; |
| goto err_out_disable; |
| } |
| |
| ret = pci_request_regions(pdev, KBUILD_MODNAME); |
| if (ret < 0) |
| goto err_out_disable; |
| |
| ret = dma_set_mask_and_coherent(&pdev->dev, DMA_BIT_MASK(64)); |
| if (ret) { |
| dev_err(&pdev->dev, "no usable dma addressing method\n"); |
| goto err_out_free_res; |
| } |
| |
| pci_set_master(pdev); |
| |
| /* ioremap MMIO region */ |
| ioaddr = ioremap(pci_resource_start(pdev, 2), |
| pci_resource_len(pdev, 2)); |
| if (!ioaddr) { |
| ret = -EIO; |
| goto err_out_free_res; |
| } |
| |
| *ioaddr_out = ioaddr; |
| *dev_out = dev; |
| |
| return ret; |
| |
| err_out_free_res: |
| pci_release_regions(pdev); |
| |
| err_out_disable: |
| pci_disable_device(pdev); |
| |
| err_out_free_dev: |
| free_netdev(dev); |
| |
| err_out: |
| *ioaddr_out = NULL; |
| *dev_out = NULL; |
| |
| return ret; |
| } |
| |
| static void rtase_release_board(struct pci_dev *pdev, struct net_device *dev, |
| void __iomem *ioaddr) |
| { |
| const struct rtase_private *tp = netdev_priv(dev); |
| |
| rtase_rar_set(tp, tp->dev->perm_addr); |
| iounmap(ioaddr); |
| |
| if (tp->sw_flag & RTASE_SWF_MSIX_ENABLED) |
| pci_disable_msix(pdev); |
| else |
| pci_disable_msi(pdev); |
| |
| pci_release_regions(pdev); |
| pci_disable_device(pdev); |
| free_netdev(dev); |
| } |
| |
| static int rtase_init_one(struct pci_dev *pdev, |
| const struct pci_device_id *ent) |
| { |
| struct net_device *dev = NULL; |
| struct rtase_int_vector *ivec; |
| void __iomem *ioaddr = NULL; |
| struct rtase_private *tp; |
| int ret, i; |
| |
| if (!pdev->is_physfn && pdev->is_virtfn) { |
| dev_err(&pdev->dev, |
| "This module does not support a virtual function."); |
| return -EINVAL; |
| } |
| |
| dev_dbg(&pdev->dev, "Automotive Switch Ethernet driver loaded\n"); |
| |
| ret = rtase_init_board(pdev, &dev, &ioaddr); |
| if (ret != 0) |
| return ret; |
| |
| tp = netdev_priv(dev); |
| tp->mmio_addr = ioaddr; |
| tp->dev = dev; |
| tp->pdev = pdev; |
| |
| /* identify chip attached to board */ |
| if (!rtase_check_mac_version_valid(tp)) |
| return dev_err_probe(&pdev->dev, -ENODEV, |
| "unknown chip version, contact rtase maintainers (see MAINTAINERS file)\n"); |
| |
| rtase_init_software_variable(pdev, tp); |
| rtase_init_hardware(tp); |
| |
| ret = rtase_alloc_interrupt(pdev, tp); |
| if (ret < 0) { |
| dev_err(&pdev->dev, "unable to alloc MSIX/MSI\n"); |
| goto err_out_1; |
| } |
| |
| rtase_init_netdev_ops(dev); |
| |
| dev->pcpu_stat_type = NETDEV_PCPU_STAT_TSTATS; |
| |
| dev->features |= NETIF_F_HW_VLAN_CTAG_TX | NETIF_F_HW_VLAN_CTAG_RX | |
| NETIF_F_IP_CSUM | NETIF_F_HIGHDMA | |
| NETIF_F_RXCSUM | NETIF_F_SG | |
| NETIF_F_TSO | NETIF_F_IPV6_CSUM | |
| NETIF_F_TSO6; |
| |
| dev->hw_features = NETIF_F_SG | NETIF_F_IP_CSUM | |
| NETIF_F_TSO | NETIF_F_RXCSUM | |
| NETIF_F_HW_VLAN_CTAG_TX | NETIF_F_HW_VLAN_CTAG_RX | |
| NETIF_F_RXALL | NETIF_F_RXFCS | |
| NETIF_F_IPV6_CSUM | NETIF_F_TSO6; |
| |
| dev->vlan_features = NETIF_F_SG | NETIF_F_IP_CSUM | NETIF_F_TSO | |
| NETIF_F_HIGHDMA; |
| dev->priv_flags |= IFF_LIVE_ADDR_CHANGE; |
| netif_set_tso_max_size(dev, RTASE_LSO_64K); |
| netif_set_tso_max_segs(dev, RTASE_NIC_MAX_PHYS_BUF_COUNT_LSO2); |
| |
| rtase_get_mac_address(dev); |
| |
| tp->tally_vaddr = dma_alloc_coherent(&pdev->dev, |
| sizeof(*tp->tally_vaddr), |
| &tp->tally_paddr, |
| GFP_KERNEL); |
| if (!tp->tally_vaddr) { |
| ret = -ENOMEM; |
| goto err_out; |
| } |
| |
| rtase_tally_counter_clear(tp); |
| |
| pci_set_drvdata(pdev, dev); |
| |
| netif_carrier_off(dev); |
| |
| ret = register_netdev(dev); |
| if (ret != 0) |
| goto err_out; |
| |
| netdev_dbg(dev, "%pM, IRQ %d\n", dev->dev_addr, dev->irq); |
| |
| return 0; |
| |
| err_out: |
| if (tp->tally_vaddr) { |
| dma_free_coherent(&pdev->dev, |
| sizeof(*tp->tally_vaddr), |
| tp->tally_vaddr, |
| tp->tally_paddr); |
| |
| tp->tally_vaddr = NULL; |
| } |
| |
| err_out_1: |
| for (i = 0; i < tp->int_nums; i++) { |
| ivec = &tp->int_vector[i]; |
| netif_napi_del(&ivec->napi); |
| } |
| |
| rtase_release_board(pdev, dev, ioaddr); |
| |
| return ret; |
| } |
| |
| static void rtase_remove_one(struct pci_dev *pdev) |
| { |
| struct net_device *dev = pci_get_drvdata(pdev); |
| struct rtase_private *tp = netdev_priv(dev); |
| struct rtase_int_vector *ivec; |
| u32 i; |
| |
| unregister_netdev(dev); |
| |
| for (i = 0; i < tp->int_nums; i++) { |
| ivec = &tp->int_vector[i]; |
| netif_napi_del(&ivec->napi); |
| } |
| |
| rtase_reset_interrupt(pdev, tp); |
| if (tp->tally_vaddr) { |
| dma_free_coherent(&pdev->dev, |
| sizeof(*tp->tally_vaddr), |
| tp->tally_vaddr, |
| tp->tally_paddr); |
| tp->tally_vaddr = NULL; |
| } |
| |
| rtase_release_board(pdev, dev, tp->mmio_addr); |
| pci_set_drvdata(pdev, NULL); |
| } |
| |
| static void rtase_shutdown(struct pci_dev *pdev) |
| { |
| struct net_device *dev = pci_get_drvdata(pdev); |
| const struct rtase_private *tp; |
| |
| tp = netdev_priv(dev); |
| |
| if (netif_running(dev)) |
| rtase_close(dev); |
| |
| rtase_reset_interrupt(pdev, tp); |
| } |
| |
| static int rtase_suspend(struct device *device) |
| { |
| struct net_device *dev = dev_get_drvdata(device); |
| |
| if (netif_running(dev)) { |
| netif_device_detach(dev); |
| rtase_hw_reset(dev); |
| } |
| |
| return 0; |
| } |
| |
| static int rtase_resume(struct device *device) |
| { |
| struct net_device *dev = dev_get_drvdata(device); |
| struct rtase_private *tp = netdev_priv(dev); |
| int ret; |
| |
| /* restore last modified mac address */ |
| rtase_rar_set(tp, dev->dev_addr); |
| |
| if (!netif_running(dev)) |
| goto out; |
| |
| rtase_wait_for_quiescence(dev); |
| |
| rtase_tx_clear(tp); |
| rtase_rx_clear(tp); |
| |
| ret = rtase_init_ring(dev); |
| if (ret) { |
| netdev_err(dev, "unable to init ring\n"); |
| rtase_free_desc(tp); |
| return -ENOMEM; |
| } |
| |
| rtase_hw_config(dev); |
| /* always link, so start to transmit & receive */ |
| rtase_hw_start(dev); |
| |
| netif_device_attach(dev); |
| out: |
| |
| return 0; |
| } |
| |
| static const struct dev_pm_ops rtase_pm_ops = { |
| SYSTEM_SLEEP_PM_OPS(rtase_suspend, rtase_resume) |
| }; |
| |
| static struct pci_driver rtase_pci_driver = { |
| .name = KBUILD_MODNAME, |
| .id_table = rtase_pci_tbl, |
| .probe = rtase_init_one, |
| .remove = rtase_remove_one, |
| .shutdown = rtase_shutdown, |
| .driver.pm = pm_ptr(&rtase_pm_ops), |
| }; |
| |
| module_pci_driver(rtase_pci_driver); |