| // SPDX-License-Identifier: GPL-2.0 |
| /* Copyright(c) 1999 - 2018 Intel Corporation. */ |
| |
| #include "ixgbe.h" |
| #include <linux/if_ether.h> |
| #include <linux/gfp.h> |
| #include <linux/if_vlan.h> |
| #include <generated/utsrelease.h> |
| #include <scsi/scsi_cmnd.h> |
| #include <scsi/scsi_device.h> |
| #include <scsi/fc/fc_fs.h> |
| #include <scsi/fc/fc_fcoe.h> |
| #include <scsi/libfc.h> |
| #include <scsi/libfcoe.h> |
| |
| /** |
| * ixgbe_fcoe_clear_ddp - clear the given ddp context |
| * @ddp: ptr to the ixgbe_fcoe_ddp |
| * |
| * Returns : none |
| * |
| */ |
| static inline void ixgbe_fcoe_clear_ddp(struct ixgbe_fcoe_ddp *ddp) |
| { |
| ddp->len = 0; |
| ddp->err = 1; |
| ddp->udl = NULL; |
| ddp->udp = 0UL; |
| ddp->sgl = NULL; |
| ddp->sgc = 0; |
| } |
| |
| /** |
| * ixgbe_fcoe_ddp_put - free the ddp context for a given xid |
| * @netdev: the corresponding net_device |
| * @xid: the xid that corresponding ddp will be freed |
| * |
| * This is the implementation of net_device_ops.ndo_fcoe_ddp_done |
| * and it is expected to be called by ULD, i.e., FCP layer of libfc |
| * to release the corresponding ddp context when the I/O is done. |
| * |
| * Returns : data length already ddp-ed in bytes |
| */ |
| int ixgbe_fcoe_ddp_put(struct net_device *netdev, u16 xid) |
| { |
| int len; |
| struct ixgbe_fcoe *fcoe; |
| struct ixgbe_adapter *adapter; |
| struct ixgbe_fcoe_ddp *ddp; |
| struct ixgbe_hw *hw; |
| u32 fcbuff; |
| |
| if (!netdev) |
| return 0; |
| |
| if (xid >= netdev->fcoe_ddp_xid) |
| return 0; |
| |
| adapter = netdev_priv(netdev); |
| fcoe = &adapter->fcoe; |
| ddp = &fcoe->ddp[xid]; |
| if (!ddp->udl) |
| return 0; |
| |
| hw = &adapter->hw; |
| len = ddp->len; |
| /* if no error then skip ddp context invalidation */ |
| if (!ddp->err) |
| goto skip_ddpinv; |
| |
| if (hw->mac.type == ixgbe_mac_X550) { |
| /* X550 does not require DDP FCoE lock */ |
| |
| IXGBE_WRITE_REG(hw, IXGBE_FCDFC(0, xid), 0); |
| IXGBE_WRITE_REG(hw, IXGBE_FCDFC(3, xid), |
| (xid | IXGBE_FCFLTRW_WE)); |
| |
| /* program FCBUFF */ |
| IXGBE_WRITE_REG(hw, IXGBE_FCDDC(2, xid), 0); |
| |
| /* program FCDMARW */ |
| IXGBE_WRITE_REG(hw, IXGBE_FCDDC(3, xid), |
| (xid | IXGBE_FCDMARW_WE)); |
| |
| /* read FCBUFF to check context invalidated */ |
| IXGBE_WRITE_REG(hw, IXGBE_FCDDC(3, xid), |
| (xid | IXGBE_FCDMARW_RE)); |
| fcbuff = IXGBE_READ_REG(hw, IXGBE_FCDDC(2, xid)); |
| } else { |
| /* other hardware requires DDP FCoE lock */ |
| spin_lock_bh(&fcoe->lock); |
| IXGBE_WRITE_REG(hw, IXGBE_FCFLT, 0); |
| IXGBE_WRITE_REG(hw, IXGBE_FCFLTRW, |
| (xid | IXGBE_FCFLTRW_WE)); |
| IXGBE_WRITE_REG(hw, IXGBE_FCBUFF, 0); |
| IXGBE_WRITE_REG(hw, IXGBE_FCDMARW, |
| (xid | IXGBE_FCDMARW_WE)); |
| |
| /* guaranteed to be invalidated after 100us */ |
| IXGBE_WRITE_REG(hw, IXGBE_FCDMARW, |
| (xid | IXGBE_FCDMARW_RE)); |
| fcbuff = IXGBE_READ_REG(hw, IXGBE_FCBUFF); |
| spin_unlock_bh(&fcoe->lock); |
| } |
| |
| if (fcbuff & IXGBE_FCBUFF_VALID) |
| usleep_range(100, 150); |
| |
| skip_ddpinv: |
| if (ddp->sgl) |
| dma_unmap_sg(&adapter->pdev->dev, ddp->sgl, ddp->sgc, |
| DMA_FROM_DEVICE); |
| if (ddp->pool) { |
| dma_pool_free(ddp->pool, ddp->udl, ddp->udp); |
| ddp->pool = NULL; |
| } |
| |
| ixgbe_fcoe_clear_ddp(ddp); |
| |
| return len; |
| } |
| |
| /** |
| * ixgbe_fcoe_ddp_setup - called to set up ddp context |
| * @netdev: the corresponding net_device |
| * @xid: the exchange id requesting ddp |
| * @sgl: the scatter-gather list for this request |
| * @sgc: the number of scatter-gather items |
| * @target_mode: 1 to setup target mode, 0 to setup initiator mode |
| * |
| * Returns : 1 for success and 0 for no ddp |
| */ |
| static int ixgbe_fcoe_ddp_setup(struct net_device *netdev, u16 xid, |
| struct scatterlist *sgl, unsigned int sgc, |
| int target_mode) |
| { |
| struct ixgbe_adapter *adapter; |
| struct ixgbe_hw *hw; |
| struct ixgbe_fcoe *fcoe; |
| struct ixgbe_fcoe_ddp *ddp; |
| struct ixgbe_fcoe_ddp_pool *ddp_pool; |
| struct scatterlist *sg; |
| unsigned int i, j, dmacount; |
| unsigned int len; |
| static const unsigned int bufflen = IXGBE_FCBUFF_MIN; |
| unsigned int firstoff = 0; |
| unsigned int lastsize; |
| unsigned int thisoff = 0; |
| unsigned int thislen = 0; |
| u32 fcbuff, fcdmarw, fcfltrw, fcrxctl; |
| dma_addr_t addr = 0; |
| |
| if (!netdev || !sgl) |
| return 0; |
| |
| adapter = netdev_priv(netdev); |
| if (xid >= netdev->fcoe_ddp_xid) { |
| e_warn(drv, "xid=0x%x out-of-range\n", xid); |
| return 0; |
| } |
| |
| /* no DDP if we are already down or resetting */ |
| if (test_bit(__IXGBE_DOWN, &adapter->state) || |
| test_bit(__IXGBE_RESETTING, &adapter->state)) |
| return 0; |
| |
| fcoe = &adapter->fcoe; |
| ddp = &fcoe->ddp[xid]; |
| if (ddp->sgl) { |
| e_err(drv, "xid 0x%x w/ non-null sgl=%p nents=%d\n", |
| xid, ddp->sgl, ddp->sgc); |
| return 0; |
| } |
| ixgbe_fcoe_clear_ddp(ddp); |
| |
| |
| if (!fcoe->ddp_pool) { |
| e_warn(drv, "No ddp_pool resources allocated\n"); |
| return 0; |
| } |
| |
| ddp_pool = per_cpu_ptr(fcoe->ddp_pool, get_cpu()); |
| if (!ddp_pool->pool) { |
| e_warn(drv, "xid=0x%x no ddp pool for fcoe\n", xid); |
| goto out_noddp; |
| } |
| |
| /* setup dma from scsi command sgl */ |
| dmacount = dma_map_sg(&adapter->pdev->dev, sgl, sgc, DMA_FROM_DEVICE); |
| if (dmacount == 0) { |
| e_err(drv, "xid 0x%x DMA map error\n", xid); |
| goto out_noddp; |
| } |
| |
| /* alloc the udl from per cpu ddp pool */ |
| ddp->udl = dma_pool_alloc(ddp_pool->pool, GFP_ATOMIC, &ddp->udp); |
| if (!ddp->udl) { |
| e_err(drv, "failed allocated ddp context\n"); |
| goto out_noddp_unmap; |
| } |
| ddp->pool = ddp_pool->pool; |
| ddp->sgl = sgl; |
| ddp->sgc = sgc; |
| |
| j = 0; |
| for_each_sg(sgl, sg, dmacount, i) { |
| addr = sg_dma_address(sg); |
| len = sg_dma_len(sg); |
| while (len) { |
| /* max number of buffers allowed in one DDP context */ |
| if (j >= IXGBE_BUFFCNT_MAX) { |
| ddp_pool->noddp++; |
| goto out_noddp_free; |
| } |
| |
| /* get the offset of length of current buffer */ |
| thisoff = addr & ((dma_addr_t)bufflen - 1); |
| thislen = min((bufflen - thisoff), len); |
| /* |
| * all but the 1st buffer (j == 0) |
| * must be aligned on bufflen |
| */ |
| if ((j != 0) && (thisoff)) |
| goto out_noddp_free; |
| /* |
| * all but the last buffer |
| * ((i == (dmacount - 1)) && (thislen == len)) |
| * must end at bufflen |
| */ |
| if (((i != (dmacount - 1)) || (thislen != len)) |
| && ((thislen + thisoff) != bufflen)) |
| goto out_noddp_free; |
| |
| ddp->udl[j] = (u64)(addr - thisoff); |
| /* only the first buffer may have none-zero offset */ |
| if (j == 0) |
| firstoff = thisoff; |
| len -= thislen; |
| addr += thislen; |
| j++; |
| } |
| } |
| /* only the last buffer may have non-full bufflen */ |
| lastsize = thisoff + thislen; |
| |
| /* |
| * lastsize can not be buffer len. |
| * If it is then adding another buffer with lastsize = 1. |
| */ |
| if (lastsize == bufflen) { |
| if (j >= IXGBE_BUFFCNT_MAX) { |
| ddp_pool->noddp_ext_buff++; |
| goto out_noddp_free; |
| } |
| |
| ddp->udl[j] = (u64)(fcoe->extra_ddp_buffer_dma); |
| j++; |
| lastsize = 1; |
| } |
| put_cpu(); |
| |
| fcbuff = (IXGBE_FCBUFF_4KB << IXGBE_FCBUFF_BUFFSIZE_SHIFT); |
| fcbuff |= ((j & 0xff) << IXGBE_FCBUFF_BUFFCNT_SHIFT); |
| fcbuff |= (firstoff << IXGBE_FCBUFF_OFFSET_SHIFT); |
| /* Set WRCONTX bit to allow DDP for target */ |
| if (target_mode) |
| fcbuff |= (IXGBE_FCBUFF_WRCONTX); |
| fcbuff |= (IXGBE_FCBUFF_VALID); |
| |
| fcdmarw = xid; |
| fcdmarw |= IXGBE_FCDMARW_WE; |
| fcdmarw |= (lastsize << IXGBE_FCDMARW_LASTSIZE_SHIFT); |
| |
| fcfltrw = xid; |
| fcfltrw |= IXGBE_FCFLTRW_WE; |
| |
| /* program DMA context */ |
| hw = &adapter->hw; |
| |
| /* turn on last frame indication for target mode as FCP_RSPtarget is |
| * supposed to send FCP_RSP when it is done. */ |
| if (target_mode && !test_bit(__IXGBE_FCOE_TARGET, &fcoe->mode)) { |
| set_bit(__IXGBE_FCOE_TARGET, &fcoe->mode); |
| fcrxctl = IXGBE_READ_REG(hw, IXGBE_FCRXCTRL); |
| fcrxctl |= IXGBE_FCRXCTRL_LASTSEQH; |
| IXGBE_WRITE_REG(hw, IXGBE_FCRXCTRL, fcrxctl); |
| } |
| |
| if (hw->mac.type == ixgbe_mac_X550) { |
| /* X550 does not require DDP lock */ |
| |
| IXGBE_WRITE_REG(hw, IXGBE_FCDDC(0, xid), |
| ddp->udp & DMA_BIT_MASK(32)); |
| IXGBE_WRITE_REG(hw, IXGBE_FCDDC(1, xid), (u64)ddp->udp >> 32); |
| IXGBE_WRITE_REG(hw, IXGBE_FCDDC(2, xid), fcbuff); |
| IXGBE_WRITE_REG(hw, IXGBE_FCDDC(3, xid), fcdmarw); |
| /* program filter context */ |
| IXGBE_WRITE_REG(hw, IXGBE_FCDFC(0, xid), IXGBE_FCFLT_VALID); |
| IXGBE_WRITE_REG(hw, IXGBE_FCDFC(1, xid), 0); |
| IXGBE_WRITE_REG(hw, IXGBE_FCDFC(3, xid), fcfltrw); |
| } else { |
| /* DDP lock for indirect DDP context access */ |
| spin_lock_bh(&fcoe->lock); |
| |
| IXGBE_WRITE_REG(hw, IXGBE_FCPTRL, ddp->udp & DMA_BIT_MASK(32)); |
| IXGBE_WRITE_REG(hw, IXGBE_FCPTRH, (u64)ddp->udp >> 32); |
| IXGBE_WRITE_REG(hw, IXGBE_FCBUFF, fcbuff); |
| IXGBE_WRITE_REG(hw, IXGBE_FCDMARW, fcdmarw); |
| /* program filter context */ |
| IXGBE_WRITE_REG(hw, IXGBE_FCPARAM, 0); |
| IXGBE_WRITE_REG(hw, IXGBE_FCFLT, IXGBE_FCFLT_VALID); |
| IXGBE_WRITE_REG(hw, IXGBE_FCFLTRW, fcfltrw); |
| |
| spin_unlock_bh(&fcoe->lock); |
| } |
| |
| return 1; |
| |
| out_noddp_free: |
| dma_pool_free(ddp->pool, ddp->udl, ddp->udp); |
| ixgbe_fcoe_clear_ddp(ddp); |
| |
| out_noddp_unmap: |
| dma_unmap_sg(&adapter->pdev->dev, sgl, sgc, DMA_FROM_DEVICE); |
| out_noddp: |
| put_cpu(); |
| return 0; |
| } |
| |
| /** |
| * ixgbe_fcoe_ddp_get - called to set up ddp context in initiator mode |
| * @netdev: the corresponding net_device |
| * @xid: the exchange id requesting ddp |
| * @sgl: the scatter-gather list for this request |
| * @sgc: the number of scatter-gather items |
| * |
| * This is the implementation of net_device_ops.ndo_fcoe_ddp_setup |
| * and is expected to be called from ULD, e.g., FCP layer of libfc |
| * to set up ddp for the corresponding xid of the given sglist for |
| * the corresponding I/O. |
| * |
| * Returns : 1 for success and 0 for no ddp |
| */ |
| int ixgbe_fcoe_ddp_get(struct net_device *netdev, u16 xid, |
| struct scatterlist *sgl, unsigned int sgc) |
| { |
| return ixgbe_fcoe_ddp_setup(netdev, xid, sgl, sgc, 0); |
| } |
| |
| /** |
| * ixgbe_fcoe_ddp_target - called to set up ddp context in target mode |
| * @netdev: the corresponding net_device |
| * @xid: the exchange id requesting ddp |
| * @sgl: the scatter-gather list for this request |
| * @sgc: the number of scatter-gather items |
| * |
| * This is the implementation of net_device_ops.ndo_fcoe_ddp_target |
| * and is expected to be called from ULD, e.g., FCP layer of libfc |
| * to set up ddp for the corresponding xid of the given sglist for |
| * the corresponding I/O. The DDP in target mode is a write I/O request |
| * from the initiator. |
| * |
| * Returns : 1 for success and 0 for no ddp |
| */ |
| int ixgbe_fcoe_ddp_target(struct net_device *netdev, u16 xid, |
| struct scatterlist *sgl, unsigned int sgc) |
| { |
| return ixgbe_fcoe_ddp_setup(netdev, xid, sgl, sgc, 1); |
| } |
| |
| /** |
| * ixgbe_fcoe_ddp - check ddp status and mark it done |
| * @adapter: ixgbe adapter |
| * @rx_desc: advanced rx descriptor |
| * @skb: the skb holding the received data |
| * |
| * This checks ddp status. |
| * |
| * Returns : < 0 indicates an error or not a FCiE ddp, 0 indicates |
| * not passing the skb to ULD, > 0 indicates is the length of data |
| * being ddped. |
| */ |
| int ixgbe_fcoe_ddp(struct ixgbe_adapter *adapter, |
| union ixgbe_adv_rx_desc *rx_desc, |
| struct sk_buff *skb) |
| { |
| int rc = -EINVAL; |
| struct ixgbe_fcoe *fcoe; |
| struct ixgbe_fcoe_ddp *ddp; |
| struct fc_frame_header *fh; |
| struct fcoe_crc_eof *crc; |
| __le32 fcerr = ixgbe_test_staterr(rx_desc, IXGBE_RXDADV_ERR_FCERR); |
| __le32 ddp_err; |
| int ddp_max; |
| u32 fctl; |
| u16 xid; |
| |
| if (fcerr == cpu_to_le32(IXGBE_FCERR_BADCRC)) |
| skb->ip_summed = CHECKSUM_NONE; |
| else |
| skb->ip_summed = CHECKSUM_UNNECESSARY; |
| |
| if (eth_hdr(skb)->h_proto == htons(ETH_P_8021Q)) |
| fh = (struct fc_frame_header *)(skb->data + |
| sizeof(struct vlan_hdr) + sizeof(struct fcoe_hdr)); |
| else |
| fh = (struct fc_frame_header *)(skb->data + |
| sizeof(struct fcoe_hdr)); |
| |
| fctl = ntoh24(fh->fh_f_ctl); |
| if (fctl & FC_FC_EX_CTX) |
| xid = be16_to_cpu(fh->fh_ox_id); |
| else |
| xid = be16_to_cpu(fh->fh_rx_id); |
| |
| ddp_max = IXGBE_FCOE_DDP_MAX; |
| /* X550 has different DDP Max limit */ |
| if (adapter->hw.mac.type == ixgbe_mac_X550) |
| ddp_max = IXGBE_FCOE_DDP_MAX_X550; |
| if (xid >= ddp_max) |
| return -EINVAL; |
| |
| fcoe = &adapter->fcoe; |
| ddp = &fcoe->ddp[xid]; |
| if (!ddp->udl) |
| return -EINVAL; |
| |
| ddp_err = ixgbe_test_staterr(rx_desc, IXGBE_RXDADV_ERR_FCEOFE | |
| IXGBE_RXDADV_ERR_FCERR); |
| if (ddp_err) |
| return -EINVAL; |
| |
| switch (ixgbe_test_staterr(rx_desc, IXGBE_RXDADV_STAT_FCSTAT)) { |
| /* return 0 to bypass going to ULD for DDPed data */ |
| case cpu_to_le32(IXGBE_RXDADV_STAT_FCSTAT_DDP): |
| /* update length of DDPed data */ |
| ddp->len = le32_to_cpu(rx_desc->wb.lower.hi_dword.rss); |
| rc = 0; |
| break; |
| /* unmap the sg list when FCPRSP is received */ |
| case cpu_to_le32(IXGBE_RXDADV_STAT_FCSTAT_FCPRSP): |
| dma_unmap_sg(&adapter->pdev->dev, ddp->sgl, |
| ddp->sgc, DMA_FROM_DEVICE); |
| ddp->err = (__force u32)ddp_err; |
| ddp->sgl = NULL; |
| ddp->sgc = 0; |
| fallthrough; |
| /* if DDP length is present pass it through to ULD */ |
| case cpu_to_le32(IXGBE_RXDADV_STAT_FCSTAT_NODDP): |
| /* update length of DDPed data */ |
| ddp->len = le32_to_cpu(rx_desc->wb.lower.hi_dword.rss); |
| if (ddp->len) |
| rc = ddp->len; |
| break; |
| /* no match will return as an error */ |
| case cpu_to_le32(IXGBE_RXDADV_STAT_FCSTAT_NOMTCH): |
| default: |
| break; |
| } |
| |
| /* In target mode, check the last data frame of the sequence. |
| * For DDP in target mode, data is already DDPed but the header |
| * indication of the last data frame ould allow is to tell if we |
| * got all the data and the ULP can send FCP_RSP back, as this is |
| * not a full fcoe frame, we fill the trailer here so it won't be |
| * dropped by the ULP stack. |
| */ |
| if ((fh->fh_r_ctl == FC_RCTL_DD_SOL_DATA) && |
| (fctl & FC_FC_END_SEQ)) { |
| skb_linearize(skb); |
| crc = skb_put(skb, sizeof(*crc)); |
| crc->fcoe_eof = FC_EOF_T; |
| } |
| |
| return rc; |
| } |
| |
| /** |
| * ixgbe_fso - ixgbe FCoE Sequence Offload (FSO) |
| * @tx_ring: tx desc ring |
| * @first: first tx_buffer structure containing skb, tx_flags, and protocol |
| * @hdr_len: hdr_len to be returned |
| * |
| * This sets up large send offload for FCoE |
| * |
| * Returns : 0 indicates success, < 0 for error |
| */ |
| int ixgbe_fso(struct ixgbe_ring *tx_ring, |
| struct ixgbe_tx_buffer *first, |
| u8 *hdr_len) |
| { |
| struct sk_buff *skb = first->skb; |
| struct fc_frame_header *fh; |
| u32 vlan_macip_lens; |
| u32 fcoe_sof_eof = 0; |
| u32 mss_l4len_idx; |
| u32 type_tucmd = IXGBE_ADVTXT_TUCMD_FCOE; |
| u8 sof, eof; |
| |
| if (skb_is_gso(skb) && (skb_shinfo(skb)->gso_type != SKB_GSO_FCOE)) { |
| dev_err(tx_ring->dev, "Wrong gso type %d:expecting SKB_GSO_FCOE\n", |
| skb_shinfo(skb)->gso_type); |
| return -EINVAL; |
| } |
| |
| /* resets the header to point fcoe/fc */ |
| skb_set_network_header(skb, skb->mac_len); |
| skb_set_transport_header(skb, skb->mac_len + |
| sizeof(struct fcoe_hdr)); |
| |
| /* sets up SOF and ORIS */ |
| sof = ((struct fcoe_hdr *)skb_network_header(skb))->fcoe_sof; |
| switch (sof) { |
| case FC_SOF_I2: |
| fcoe_sof_eof = IXGBE_ADVTXD_FCOEF_ORIS; |
| break; |
| case FC_SOF_I3: |
| fcoe_sof_eof = IXGBE_ADVTXD_FCOEF_SOF | |
| IXGBE_ADVTXD_FCOEF_ORIS; |
| break; |
| case FC_SOF_N2: |
| break; |
| case FC_SOF_N3: |
| fcoe_sof_eof = IXGBE_ADVTXD_FCOEF_SOF; |
| break; |
| default: |
| dev_warn(tx_ring->dev, "unknown sof = 0x%x\n", sof); |
| return -EINVAL; |
| } |
| |
| /* the first byte of the last dword is EOF */ |
| skb_copy_bits(skb, skb->len - 4, &eof, 1); |
| /* sets up EOF and ORIE */ |
| switch (eof) { |
| case FC_EOF_N: |
| fcoe_sof_eof |= IXGBE_ADVTXD_FCOEF_EOF_N; |
| break; |
| case FC_EOF_T: |
| /* lso needs ORIE */ |
| if (skb_is_gso(skb)) |
| fcoe_sof_eof |= IXGBE_ADVTXD_FCOEF_EOF_N | |
| IXGBE_ADVTXD_FCOEF_ORIE; |
| else |
| fcoe_sof_eof |= IXGBE_ADVTXD_FCOEF_EOF_T; |
| break; |
| case FC_EOF_NI: |
| fcoe_sof_eof |= IXGBE_ADVTXD_FCOEF_EOF_NI; |
| break; |
| case FC_EOF_A: |
| fcoe_sof_eof |= IXGBE_ADVTXD_FCOEF_EOF_A; |
| break; |
| default: |
| dev_warn(tx_ring->dev, "unknown eof = 0x%x\n", eof); |
| return -EINVAL; |
| } |
| |
| /* sets up PARINC indicating data offset */ |
| fh = (struct fc_frame_header *)skb_transport_header(skb); |
| if (fh->fh_f_ctl[2] & FC_FC_REL_OFF) |
| fcoe_sof_eof |= IXGBE_ADVTXD_FCOEF_PARINC; |
| |
| /* include trailer in headlen as it is replicated per frame */ |
| *hdr_len = sizeof(struct fcoe_crc_eof); |
| |
| /* hdr_len includes fc_hdr if FCoE LSO is enabled */ |
| if (skb_is_gso(skb)) { |
| *hdr_len += skb_transport_offset(skb) + |
| sizeof(struct fc_frame_header); |
| /* update gso_segs and bytecount */ |
| first->gso_segs = DIV_ROUND_UP(skb->len - *hdr_len, |
| skb_shinfo(skb)->gso_size); |
| first->bytecount += (first->gso_segs - 1) * *hdr_len; |
| first->tx_flags |= IXGBE_TX_FLAGS_TSO; |
| /* Hardware expects L4T to be RSV for FCoE TSO */ |
| type_tucmd |= IXGBE_ADVTXD_TUCMD_L4T_RSV; |
| } |
| |
| /* set flag indicating FCOE to ixgbe_tx_map call */ |
| first->tx_flags |= IXGBE_TX_FLAGS_FCOE | IXGBE_TX_FLAGS_CC; |
| |
| /* mss_l4len_id: use 0 for FSO as TSO, no need for L4LEN */ |
| mss_l4len_idx = skb_shinfo(skb)->gso_size << IXGBE_ADVTXD_MSS_SHIFT; |
| |
| /* vlan_macip_lens: HEADLEN, MACLEN, VLAN tag */ |
| vlan_macip_lens = skb_transport_offset(skb) + |
| sizeof(struct fc_frame_header); |
| vlan_macip_lens |= (skb_transport_offset(skb) - 4) |
| << IXGBE_ADVTXD_MACLEN_SHIFT; |
| vlan_macip_lens |= first->tx_flags & IXGBE_TX_FLAGS_VLAN_MASK; |
| |
| /* write context desc */ |
| ixgbe_tx_ctxtdesc(tx_ring, vlan_macip_lens, fcoe_sof_eof, |
| type_tucmd, mss_l4len_idx); |
| |
| return 0; |
| } |
| |
| static void ixgbe_fcoe_dma_pool_free(struct ixgbe_fcoe *fcoe, unsigned int cpu) |
| { |
| struct ixgbe_fcoe_ddp_pool *ddp_pool; |
| |
| ddp_pool = per_cpu_ptr(fcoe->ddp_pool, cpu); |
| dma_pool_destroy(ddp_pool->pool); |
| ddp_pool->pool = NULL; |
| } |
| |
| static int ixgbe_fcoe_dma_pool_alloc(struct ixgbe_fcoe *fcoe, |
| struct device *dev, |
| unsigned int cpu) |
| { |
| struct ixgbe_fcoe_ddp_pool *ddp_pool; |
| struct dma_pool *pool; |
| char pool_name[32]; |
| |
| snprintf(pool_name, 32, "ixgbe_fcoe_ddp_%u", cpu); |
| |
| pool = dma_pool_create(pool_name, dev, IXGBE_FCPTR_MAX, |
| IXGBE_FCPTR_ALIGN, PAGE_SIZE); |
| if (!pool) |
| return -ENOMEM; |
| |
| ddp_pool = per_cpu_ptr(fcoe->ddp_pool, cpu); |
| ddp_pool->pool = pool; |
| ddp_pool->noddp = 0; |
| ddp_pool->noddp_ext_buff = 0; |
| |
| return 0; |
| } |
| |
| /** |
| * ixgbe_configure_fcoe - configures registers for fcoe at start |
| * @adapter: ptr to ixgbe adapter |
| * |
| * This sets up FCoE related registers |
| * |
| * Returns : none |
| */ |
| void ixgbe_configure_fcoe(struct ixgbe_adapter *adapter) |
| { |
| struct ixgbe_ring_feature *fcoe = &adapter->ring_feature[RING_F_FCOE]; |
| struct ixgbe_hw *hw = &adapter->hw; |
| int i, fcoe_q, fcoe_i, fcoe_q_h = 0; |
| int fcreta_size; |
| u32 etqf; |
| |
| /* Minimal functionality for FCoE requires at least CRC offloads */ |
| if (!(adapter->netdev->features & NETIF_F_FCOE_CRC)) |
| return; |
| |
| /* Enable L2 EtherType filter for FCoE, needed for FCoE CRC and DDP */ |
| etqf = ETH_P_FCOE | IXGBE_ETQF_FCOE | IXGBE_ETQF_FILTER_EN; |
| if (adapter->flags & IXGBE_FLAG_SRIOV_ENABLED) { |
| etqf |= IXGBE_ETQF_POOL_ENABLE; |
| etqf |= VMDQ_P(0) << IXGBE_ETQF_POOL_SHIFT; |
| } |
| IXGBE_WRITE_REG(hw, IXGBE_ETQF(IXGBE_ETQF_FILTER_FCOE), etqf); |
| IXGBE_WRITE_REG(hw, IXGBE_ETQS(IXGBE_ETQF_FILTER_FCOE), 0); |
| |
| /* leave registers un-configured if FCoE is disabled */ |
| if (!(adapter->flags & IXGBE_FLAG_FCOE_ENABLED)) |
| return; |
| |
| /* Use one or more Rx queues for FCoE by redirection table */ |
| fcreta_size = IXGBE_FCRETA_SIZE; |
| if (adapter->hw.mac.type == ixgbe_mac_X550) |
| fcreta_size = IXGBE_FCRETA_SIZE_X550; |
| |
| for (i = 0; i < fcreta_size; i++) { |
| if (adapter->hw.mac.type == ixgbe_mac_X550) { |
| int fcoe_i_h = fcoe->offset + ((i + fcreta_size) % |
| fcoe->indices); |
| fcoe_q_h = adapter->rx_ring[fcoe_i_h]->reg_idx; |
| fcoe_q_h = FIELD_PREP(IXGBE_FCRETA_ENTRY_HIGH_MASK, |
| fcoe_q_h); |
| } |
| |
| fcoe_i = fcoe->offset + (i % fcoe->indices); |
| fcoe_i &= IXGBE_FCRETA_ENTRY_MASK; |
| fcoe_q = adapter->rx_ring[fcoe_i]->reg_idx; |
| fcoe_q |= fcoe_q_h; |
| IXGBE_WRITE_REG(hw, IXGBE_FCRETA(i), fcoe_q); |
| } |
| IXGBE_WRITE_REG(hw, IXGBE_FCRECTL, IXGBE_FCRECTL_ENA); |
| |
| /* Enable L2 EtherType filter for FIP */ |
| etqf = ETH_P_FIP | IXGBE_ETQF_FILTER_EN; |
| if (adapter->flags & IXGBE_FLAG_SRIOV_ENABLED) { |
| etqf |= IXGBE_ETQF_POOL_ENABLE; |
| etqf |= VMDQ_P(0) << IXGBE_ETQF_POOL_SHIFT; |
| } |
| IXGBE_WRITE_REG(hw, IXGBE_ETQF(IXGBE_ETQF_FILTER_FIP), etqf); |
| |
| /* Send FIP frames to the first FCoE queue */ |
| fcoe_q = adapter->rx_ring[fcoe->offset]->reg_idx; |
| IXGBE_WRITE_REG(hw, IXGBE_ETQS(IXGBE_ETQF_FILTER_FIP), |
| IXGBE_ETQS_QUEUE_EN | |
| (fcoe_q << IXGBE_ETQS_RX_QUEUE_SHIFT)); |
| |
| /* Configure FCoE Rx control */ |
| IXGBE_WRITE_REG(hw, IXGBE_FCRXCTRL, |
| IXGBE_FCRXCTRL_FCCRCBO | |
| (FC_FCOE_VER << IXGBE_FCRXCTRL_FCOEVER_SHIFT)); |
| } |
| |
| /** |
| * ixgbe_free_fcoe_ddp_resources - release all fcoe ddp context resources |
| * @adapter : ixgbe adapter |
| * |
| * Cleans up outstanding ddp context resources |
| * |
| * Returns : none |
| */ |
| void ixgbe_free_fcoe_ddp_resources(struct ixgbe_adapter *adapter) |
| { |
| struct ixgbe_fcoe *fcoe = &adapter->fcoe; |
| int cpu, i, ddp_max; |
| |
| /* do nothing if no DDP pools were allocated */ |
| if (!fcoe->ddp_pool) |
| return; |
| |
| ddp_max = IXGBE_FCOE_DDP_MAX; |
| /* X550 has different DDP Max limit */ |
| if (adapter->hw.mac.type == ixgbe_mac_X550) |
| ddp_max = IXGBE_FCOE_DDP_MAX_X550; |
| |
| for (i = 0; i < ddp_max; i++) |
| ixgbe_fcoe_ddp_put(adapter->netdev, i); |
| |
| for_each_possible_cpu(cpu) |
| ixgbe_fcoe_dma_pool_free(fcoe, cpu); |
| |
| dma_unmap_single(&adapter->pdev->dev, |
| fcoe->extra_ddp_buffer_dma, |
| IXGBE_FCBUFF_MIN, |
| DMA_FROM_DEVICE); |
| kfree(fcoe->extra_ddp_buffer); |
| |
| fcoe->extra_ddp_buffer = NULL; |
| fcoe->extra_ddp_buffer_dma = 0; |
| } |
| |
| /** |
| * ixgbe_setup_fcoe_ddp_resources - setup all fcoe ddp context resources |
| * @adapter: ixgbe adapter |
| * |
| * Sets up ddp context resouces |
| * |
| * Returns : 0 indicates success or -EINVAL on failure |
| */ |
| int ixgbe_setup_fcoe_ddp_resources(struct ixgbe_adapter *adapter) |
| { |
| struct ixgbe_fcoe *fcoe = &adapter->fcoe; |
| struct device *dev = &adapter->pdev->dev; |
| void *buffer; |
| dma_addr_t dma; |
| unsigned int cpu; |
| |
| /* do nothing if no DDP pools were allocated */ |
| if (!fcoe->ddp_pool) |
| return 0; |
| |
| /* Extra buffer to be shared by all DDPs for HW work around */ |
| buffer = kmalloc(IXGBE_FCBUFF_MIN, GFP_KERNEL); |
| if (!buffer) |
| return -ENOMEM; |
| |
| dma = dma_map_single(dev, buffer, IXGBE_FCBUFF_MIN, DMA_FROM_DEVICE); |
| if (dma_mapping_error(dev, dma)) { |
| e_err(drv, "failed to map extra DDP buffer\n"); |
| kfree(buffer); |
| return -ENOMEM; |
| } |
| |
| fcoe->extra_ddp_buffer = buffer; |
| fcoe->extra_ddp_buffer_dma = dma; |
| |
| /* allocate pci pool for each cpu */ |
| for_each_possible_cpu(cpu) { |
| int err = ixgbe_fcoe_dma_pool_alloc(fcoe, dev, cpu); |
| if (!err) |
| continue; |
| |
| e_err(drv, "failed to alloc DDP pool on cpu:%d\n", cpu); |
| ixgbe_free_fcoe_ddp_resources(adapter); |
| return -ENOMEM; |
| } |
| |
| return 0; |
| } |
| |
| static int ixgbe_fcoe_ddp_enable(struct ixgbe_adapter *adapter) |
| { |
| struct ixgbe_fcoe *fcoe = &adapter->fcoe; |
| |
| if (!(adapter->flags & IXGBE_FLAG_FCOE_CAPABLE)) |
| return -EINVAL; |
| |
| fcoe->ddp_pool = alloc_percpu(struct ixgbe_fcoe_ddp_pool); |
| |
| if (!fcoe->ddp_pool) { |
| e_err(drv, "failed to allocate percpu DDP resources\n"); |
| return -ENOMEM; |
| } |
| |
| adapter->netdev->fcoe_ddp_xid = IXGBE_FCOE_DDP_MAX - 1; |
| /* X550 has different DDP Max limit */ |
| if (adapter->hw.mac.type == ixgbe_mac_X550) |
| adapter->netdev->fcoe_ddp_xid = IXGBE_FCOE_DDP_MAX_X550 - 1; |
| |
| return 0; |
| } |
| |
| static void ixgbe_fcoe_ddp_disable(struct ixgbe_adapter *adapter) |
| { |
| struct ixgbe_fcoe *fcoe = &adapter->fcoe; |
| |
| adapter->netdev->fcoe_ddp_xid = 0; |
| |
| if (!fcoe->ddp_pool) |
| return; |
| |
| free_percpu(fcoe->ddp_pool); |
| fcoe->ddp_pool = NULL; |
| } |
| |
| /** |
| * ixgbe_fcoe_enable - turn on FCoE offload feature |
| * @netdev: the corresponding netdev |
| * |
| * Turns on FCoE offload feature in 82599. |
| * |
| * Returns : 0 indicates success or -EINVAL on failure |
| */ |
| int ixgbe_fcoe_enable(struct net_device *netdev) |
| { |
| struct ixgbe_adapter *adapter = netdev_priv(netdev); |
| struct ixgbe_fcoe *fcoe = &adapter->fcoe; |
| |
| atomic_inc(&fcoe->refcnt); |
| |
| if (!(adapter->flags & IXGBE_FLAG_FCOE_CAPABLE)) |
| return -EINVAL; |
| |
| if (adapter->flags & IXGBE_FLAG_FCOE_ENABLED) |
| return -EINVAL; |
| |
| e_info(drv, "Enabling FCoE offload features.\n"); |
| |
| if (adapter->flags & IXGBE_FLAG_SRIOV_ENABLED) |
| e_warn(probe, "Enabling FCoE on PF will disable legacy VFs\n"); |
| |
| if (netif_running(netdev)) |
| netdev->netdev_ops->ndo_stop(netdev); |
| |
| /* Allocate per CPU memory to track DDP pools */ |
| ixgbe_fcoe_ddp_enable(adapter); |
| |
| /* enable FCoE and notify stack */ |
| adapter->flags |= IXGBE_FLAG_FCOE_ENABLED; |
| netdev->fcoe_mtu = true; |
| netdev_features_change(netdev); |
| |
| /* release existing queues and reallocate them */ |
| ixgbe_clear_interrupt_scheme(adapter); |
| ixgbe_init_interrupt_scheme(adapter); |
| |
| if (netif_running(netdev)) |
| netdev->netdev_ops->ndo_open(netdev); |
| |
| return 0; |
| } |
| |
| /** |
| * ixgbe_fcoe_disable - turn off FCoE offload feature |
| * @netdev: the corresponding netdev |
| * |
| * Turns off FCoE offload feature in 82599. |
| * |
| * Returns : 0 indicates success or -EINVAL on failure |
| */ |
| int ixgbe_fcoe_disable(struct net_device *netdev) |
| { |
| struct ixgbe_adapter *adapter = netdev_priv(netdev); |
| |
| if (!atomic_dec_and_test(&adapter->fcoe.refcnt)) |
| return -EINVAL; |
| |
| if (!(adapter->flags & IXGBE_FLAG_FCOE_ENABLED)) |
| return -EINVAL; |
| |
| e_info(drv, "Disabling FCoE offload features.\n"); |
| if (netif_running(netdev)) |
| netdev->netdev_ops->ndo_stop(netdev); |
| |
| /* Free per CPU memory to track DDP pools */ |
| ixgbe_fcoe_ddp_disable(adapter); |
| |
| /* disable FCoE and notify stack */ |
| adapter->flags &= ~IXGBE_FLAG_FCOE_ENABLED; |
| netdev->fcoe_mtu = false; |
| |
| netdev_features_change(netdev); |
| |
| /* release existing queues and reallocate them */ |
| ixgbe_clear_interrupt_scheme(adapter); |
| ixgbe_init_interrupt_scheme(adapter); |
| |
| if (netif_running(netdev)) |
| netdev->netdev_ops->ndo_open(netdev); |
| |
| return 0; |
| } |
| |
| /** |
| * ixgbe_fcoe_get_wwn - get world wide name for the node or the port |
| * @netdev : ixgbe adapter |
| * @wwn : the world wide name |
| * @type: the type of world wide name |
| * |
| * Returns the node or port world wide name if both the prefix and the san |
| * mac address are valid, then the wwn is formed based on the NAA-2 for |
| * IEEE Extended name identifier (ref. to T10 FC-LS Spec., Sec. 15.3). |
| * |
| * Returns : 0 on success |
| */ |
| int ixgbe_fcoe_get_wwn(struct net_device *netdev, u64 *wwn, int type) |
| { |
| u16 prefix = 0xffff; |
| struct ixgbe_adapter *adapter = netdev_priv(netdev); |
| struct ixgbe_mac_info *mac = &adapter->hw.mac; |
| |
| switch (type) { |
| case NETDEV_FCOE_WWNN: |
| prefix = mac->wwnn_prefix; |
| break; |
| case NETDEV_FCOE_WWPN: |
| prefix = mac->wwpn_prefix; |
| break; |
| default: |
| break; |
| } |
| |
| if ((prefix != 0xffff) && |
| is_valid_ether_addr(mac->san_addr)) { |
| *wwn = ((u64) prefix << 48) | |
| ((u64) mac->san_addr[0] << 40) | |
| ((u64) mac->san_addr[1] << 32) | |
| ((u64) mac->san_addr[2] << 24) | |
| ((u64) mac->san_addr[3] << 16) | |
| ((u64) mac->san_addr[4] << 8) | |
| ((u64) mac->san_addr[5]); |
| return 0; |
| } |
| return -EINVAL; |
| } |
| |
| /** |
| * ixgbe_fcoe_get_hbainfo - get FCoE HBA information |
| * @netdev : ixgbe adapter |
| * @info : HBA information |
| * |
| * Returns ixgbe HBA information |
| * |
| * Returns : 0 on success |
| */ |
| int ixgbe_fcoe_get_hbainfo(struct net_device *netdev, |
| struct netdev_fcoe_hbainfo *info) |
| { |
| struct ixgbe_adapter *adapter = netdev_priv(netdev); |
| struct ixgbe_hw *hw = &adapter->hw; |
| u64 dsn; |
| |
| if (!info) |
| return -EINVAL; |
| |
| /* Don't return information on unsupported devices */ |
| if (!(adapter->flags & IXGBE_FLAG_FCOE_ENABLED)) |
| return -EINVAL; |
| |
| /* Manufacturer */ |
| snprintf(info->manufacturer, sizeof(info->manufacturer), |
| "Intel Corporation"); |
| |
| /* Serial Number */ |
| |
| /* Get the PCI-e Device Serial Number Capability */ |
| dsn = pci_get_dsn(adapter->pdev); |
| if (dsn) |
| snprintf(info->serial_number, sizeof(info->serial_number), |
| "%016llX", dsn); |
| else |
| snprintf(info->serial_number, sizeof(info->serial_number), |
| "Unknown"); |
| |
| /* Hardware Version */ |
| snprintf(info->hardware_version, |
| sizeof(info->hardware_version), |
| "Rev %d", hw->revision_id); |
| /* Driver Name/Version */ |
| snprintf(info->driver_version, |
| sizeof(info->driver_version), |
| "%s v%s", |
| ixgbe_driver_name, |
| UTS_RELEASE); |
| /* Firmware Version */ |
| strscpy(info->firmware_version, adapter->eeprom_id, |
| sizeof(info->firmware_version)); |
| |
| /* Model */ |
| if (hw->mac.type == ixgbe_mac_82599EB) { |
| snprintf(info->model, |
| sizeof(info->model), |
| "Intel 82599"); |
| } else if (hw->mac.type == ixgbe_mac_X550) { |
| snprintf(info->model, |
| sizeof(info->model), |
| "Intel X550"); |
| } else { |
| snprintf(info->model, |
| sizeof(info->model), |
| "Intel X540"); |
| } |
| |
| /* Model Description */ |
| snprintf(info->model_description, |
| sizeof(info->model_description), |
| "%s", |
| ixgbe_default_device_descr); |
| |
| return 0; |
| } |
| |
| /** |
| * ixgbe_fcoe_get_tc - get the current TC that fcoe is mapped to |
| * @adapter: pointer to the device adapter structure |
| * |
| * Return : TC that FCoE is mapped to |
| */ |
| u8 ixgbe_fcoe_get_tc(struct ixgbe_adapter *adapter) |
| { |
| #ifdef CONFIG_IXGBE_DCB |
| return netdev_get_prio_tc_map(adapter->netdev, adapter->fcoe.up); |
| #else |
| return 0; |
| #endif |
| } |