| /* SPDX-License-Identifier: GPL-2.0 */ |
| /* Copyright (c) 2018, Intel Corporation. */ |
| |
| #ifndef _ICE_LAN_TX_RX_H_ |
| #define _ICE_LAN_TX_RX_H_ |
| |
| union ice_32byte_rx_desc { |
| struct { |
| __le64 pkt_addr; /* Packet buffer address */ |
| __le64 hdr_addr; /* Header buffer address */ |
| /* bit 0 of hdr_addr is DD bit */ |
| __le64 rsvd1; |
| __le64 rsvd2; |
| } read; |
| struct { |
| struct { |
| struct { |
| __le16 mirroring_status; |
| __le16 l2tag1; |
| } lo_dword; |
| union { |
| __le32 rss; /* RSS Hash */ |
| __le32 fd_id; /* Flow Director filter id */ |
| } hi_dword; |
| } qword0; |
| struct { |
| /* status/error/PTYPE/length */ |
| __le64 status_error_len; |
| } qword1; |
| struct { |
| __le16 ext_status; /* extended status */ |
| __le16 rsvd; |
| __le16 l2tag2_1; |
| __le16 l2tag2_2; |
| } qword2; |
| struct { |
| __le32 reserved; |
| __le32 fd_id; |
| } qword3; |
| } wb; /* writeback */ |
| }; |
| |
| struct ice_rx_ptype_decoded { |
| u32 ptype:10; |
| u32 known:1; |
| u32 outer_ip:1; |
| u32 outer_ip_ver:2; |
| u32 outer_frag:1; |
| u32 tunnel_type:3; |
| u32 tunnel_end_prot:2; |
| u32 tunnel_end_frag:1; |
| u32 inner_prot:4; |
| u32 payload_layer:3; |
| }; |
| |
| enum ice_rx_ptype_outer_ip { |
| ICE_RX_PTYPE_OUTER_L2 = 0, |
| ICE_RX_PTYPE_OUTER_IP = 1, |
| }; |
| |
| enum ice_rx_ptype_outer_ip_ver { |
| ICE_RX_PTYPE_OUTER_NONE = 0, |
| ICE_RX_PTYPE_OUTER_IPV4 = 1, |
| ICE_RX_PTYPE_OUTER_IPV6 = 2, |
| }; |
| |
| enum ice_rx_ptype_outer_fragmented { |
| ICE_RX_PTYPE_NOT_FRAG = 0, |
| ICE_RX_PTYPE_FRAG = 1, |
| }; |
| |
| enum ice_rx_ptype_tunnel_type { |
| ICE_RX_PTYPE_TUNNEL_NONE = 0, |
| ICE_RX_PTYPE_TUNNEL_IP_IP = 1, |
| ICE_RX_PTYPE_TUNNEL_IP_GRENAT = 2, |
| ICE_RX_PTYPE_TUNNEL_IP_GRENAT_MAC = 3, |
| ICE_RX_PTYPE_TUNNEL_IP_GRENAT_MAC_VLAN = 4, |
| }; |
| |
| enum ice_rx_ptype_tunnel_end_prot { |
| ICE_RX_PTYPE_TUNNEL_END_NONE = 0, |
| ICE_RX_PTYPE_TUNNEL_END_IPV4 = 1, |
| ICE_RX_PTYPE_TUNNEL_END_IPV6 = 2, |
| }; |
| |
| enum ice_rx_ptype_inner_prot { |
| ICE_RX_PTYPE_INNER_PROT_NONE = 0, |
| ICE_RX_PTYPE_INNER_PROT_UDP = 1, |
| ICE_RX_PTYPE_INNER_PROT_TCP = 2, |
| ICE_RX_PTYPE_INNER_PROT_SCTP = 3, |
| ICE_RX_PTYPE_INNER_PROT_ICMP = 4, |
| ICE_RX_PTYPE_INNER_PROT_TIMESYNC = 5, |
| }; |
| |
| enum ice_rx_ptype_payload_layer { |
| ICE_RX_PTYPE_PAYLOAD_LAYER_NONE = 0, |
| ICE_RX_PTYPE_PAYLOAD_LAYER_PAY2 = 1, |
| ICE_RX_PTYPE_PAYLOAD_LAYER_PAY3 = 2, |
| ICE_RX_PTYPE_PAYLOAD_LAYER_PAY4 = 3, |
| }; |
| |
| /* RX Flex Descriptor |
| * This descriptor is used instead of the legacy version descriptor when |
| * ice_rlan_ctx.adv_desc is set |
| */ |
| union ice_32b_rx_flex_desc { |
| struct { |
| __le64 pkt_addr; /* Packet buffer address */ |
| __le64 hdr_addr; /* Header buffer address */ |
| /* bit 0 of hdr_addr is DD bit */ |
| __le64 rsvd1; |
| __le64 rsvd2; |
| } read; |
| struct { |
| /* Qword 0 */ |
| u8 rxdid; /* descriptor builder profile id */ |
| u8 mir_id_umb_cast; /* mirror=[5:0], umb=[7:6] */ |
| __le16 ptype_flex_flags0; /* ptype=[9:0], ff0=[15:10] */ |
| __le16 pkt_len; /* [15:14] are reserved */ |
| __le16 hdr_len_sph_flex_flags1; /* header=[10:0] */ |
| /* sph=[11:11] */ |
| /* ff1/ext=[15:12] */ |
| |
| /* Qword 1 */ |
| __le16 status_error0; |
| __le16 l2tag1; |
| __le16 flex_meta0; |
| __le16 flex_meta1; |
| |
| /* Qword 2 */ |
| __le16 status_error1; |
| u8 flex_flags2; |
| u8 time_stamp_low; |
| __le16 l2tag2_1st; |
| __le16 l2tag2_2nd; |
| |
| /* Qword 3 */ |
| __le16 flex_meta2; |
| __le16 flex_meta3; |
| union { |
| struct { |
| __le16 flex_meta4; |
| __le16 flex_meta5; |
| } flex; |
| __le32 ts_high; |
| } flex_ts; |
| } wb; /* writeback */ |
| }; |
| |
| /* Rx Flex Descriptor NIC Profile |
| * This descriptor corresponds to RxDID 2 which contains |
| * metadata fields for RSS, flow id and timestamp info |
| */ |
| struct ice_32b_rx_flex_desc_nic { |
| /* Qword 0 */ |
| u8 rxdid; |
| u8 mir_id_umb_cast; |
| __le16 ptype_flexi_flags0; |
| __le16 pkt_len; |
| __le16 hdr_len_sph_flex_flags1; |
| |
| /* Qword 1 */ |
| __le16 status_error0; |
| __le16 l2tag1; |
| __le32 rss_hash; |
| |
| /* Qword 2 */ |
| __le16 status_error1; |
| u8 flexi_flags2; |
| u8 ts_low; |
| __le16 l2tag2_1st; |
| __le16 l2tag2_2nd; |
| |
| /* Qword 3 */ |
| __le32 flow_id; |
| union { |
| struct { |
| __le16 vlan_id; |
| __le16 flow_id_ipv6; |
| } flex; |
| __le32 ts_high; |
| } flex_ts; |
| }; |
| |
| /* Receive Flex Descriptor profile IDs: There are a total |
| * of 64 profiles where profile IDs 0/1 are for legacy; and |
| * profiles 2-63 are flex profiles that can be programmed |
| * with a specific metadata (profile 7 reserved for HW) |
| */ |
| enum ice_rxdid { |
| ICE_RXDID_LEGACY_0 = 0, |
| ICE_RXDID_LEGACY_1 = 1, |
| ICE_RXDID_FLEX_NIC = 2, |
| ICE_RXDID_FLEX_NIC_2 = 6, |
| ICE_RXDID_HW = 7, |
| ICE_RXDID_LAST = 63, |
| }; |
| |
| /* Receive Flex Descriptor Rx opcode values */ |
| #define ICE_RX_OPC_MDID 0x01 |
| |
| /* Receive Descriptor MDID values */ |
| enum ice_flex_rx_mdid { |
| ICE_RX_MDID_FLOW_ID_LOWER = 5, |
| ICE_RX_MDID_FLOW_ID_HIGH, |
| ICE_RX_MDID_SRC_VSI = 19, |
| ICE_RX_MDID_HASH_LOW = 56, |
| ICE_RX_MDID_HASH_HIGH, |
| }; |
| |
| /* Rx Flag64 packet flag bits */ |
| enum ice_rx_flg64_bits { |
| ICE_RXFLG_PKT_DSI = 0, |
| ICE_RXFLG_EVLAN_x8100 = 15, |
| ICE_RXFLG_EVLAN_x9100, |
| ICE_RXFLG_VLAN_x8100, |
| ICE_RXFLG_TNL_MAC = 22, |
| ICE_RXFLG_TNL_VLAN, |
| ICE_RXFLG_PKT_FRG, |
| ICE_RXFLG_FIN = 32, |
| ICE_RXFLG_SYN, |
| ICE_RXFLG_RST, |
| ICE_RXFLG_TNL0 = 38, |
| ICE_RXFLG_TNL1, |
| ICE_RXFLG_TNL2, |
| ICE_RXFLG_UDP_GRE, |
| ICE_RXFLG_RSVD = 63 |
| }; |
| |
| /* for ice_32byte_rx_flex_desc.ptype_flexi_flags0 member */ |
| #define ICE_RX_FLEX_DESC_PTYPE_M (0x3FF) /* 10-bits */ |
| |
| /* for ice_32byte_rx_flex_desc.pkt_length member */ |
| #define ICE_RX_FLX_DESC_PKT_LEN_M (0x3FFF) /* 14-bits */ |
| |
| enum ice_rx_flex_desc_status_error_0_bits { |
| /* Note: These are predefined bit offsets */ |
| ICE_RX_FLEX_DESC_STATUS0_DD_S = 0, |
| ICE_RX_FLEX_DESC_STATUS0_EOF_S, |
| ICE_RX_FLEX_DESC_STATUS0_HBO_S, |
| ICE_RX_FLEX_DESC_STATUS0_L3L4P_S, |
| ICE_RX_FLEX_DESC_STATUS0_XSUM_IPE_S, |
| ICE_RX_FLEX_DESC_STATUS0_XSUM_L4E_S, |
| ICE_RX_FLEX_DESC_STATUS0_XSUM_EIPE_S, |
| ICE_RX_FLEX_DESC_STATUS0_XSUM_EUDPE_S, |
| ICE_RX_FLEX_DESC_STATUS0_LPBK_S, |
| ICE_RX_FLEX_DESC_STATUS0_IPV6EXADD_S, |
| ICE_RX_FLEX_DESC_STATUS0_RXE_S, |
| ICE_RX_FLEX_DESC_STATUS0_CRCP_S, |
| ICE_RX_FLEX_DESC_STATUS0_RSS_VALID_S, |
| ICE_RX_FLEX_DESC_STATUS0_L2TAG1P_S, |
| ICE_RX_FLEX_DESC_STATUS0_XTRMD0_VALID_S, |
| ICE_RX_FLEX_DESC_STATUS0_XTRMD1_VALID_S, |
| ICE_RX_FLEX_DESC_STATUS0_LAST /* this entry must be last!!! */ |
| }; |
| |
| #define ICE_RXQ_CTX_SIZE_DWORDS 8 |
| #define ICE_RXQ_CTX_SZ (ICE_RXQ_CTX_SIZE_DWORDS * sizeof(u32)) |
| |
| /* RLAN Rx queue context data |
| * |
| * The sizes of the variables may be larger than needed due to crossing byte |
| * boundaries. If we do not have the width of the variable set to the correct |
| * size then we could end up shifting bits off the top of the variable when the |
| * variable is at the top of a byte and crosses over into the next byte. |
| */ |
| struct ice_rlan_ctx { |
| u16 head; |
| u16 cpuid; /* bigger than needed, see above for reason */ |
| #define ICE_RLAN_BASE_S 7 |
| u64 base; |
| u16 qlen; |
| #define ICE_RLAN_CTX_DBUF_S 7 |
| u16 dbuf; /* bigger than needed, see above for reason */ |
| #define ICE_RLAN_CTX_HBUF_S 6 |
| u16 hbuf; /* bigger than needed, see above for reason */ |
| u8 dtype; |
| u8 dsize; |
| u8 crcstrip; |
| u8 l2tsel; |
| u8 hsplit_0; |
| u8 hsplit_1; |
| u8 showiv; |
| u32 rxmax; /* bigger than needed, see above for reason */ |
| u8 tphrdesc_ena; |
| u8 tphwdesc_ena; |
| u8 tphdata_ena; |
| u8 tphhead_ena; |
| u16 lrxqthresh; /* bigger than needed, see above for reason */ |
| }; |
| |
| struct ice_ctx_ele { |
| u16 offset; |
| u16 size_of; |
| u16 width; |
| u16 lsb; |
| }; |
| |
| #define ICE_CTX_STORE(_struct, _ele, _width, _lsb) { \ |
| .offset = offsetof(struct _struct, _ele), \ |
| .size_of = FIELD_SIZEOF(struct _struct, _ele), \ |
| .width = _width, \ |
| .lsb = _lsb, \ |
| } |
| |
| /* for hsplit_0 field of Rx RLAN context */ |
| enum ice_rlan_ctx_rx_hsplit_0 { |
| ICE_RLAN_RX_HSPLIT_0_NO_SPLIT = 0, |
| ICE_RLAN_RX_HSPLIT_0_SPLIT_L2 = 1, |
| ICE_RLAN_RX_HSPLIT_0_SPLIT_IP = 2, |
| ICE_RLAN_RX_HSPLIT_0_SPLIT_TCP_UDP = 4, |
| ICE_RLAN_RX_HSPLIT_0_SPLIT_SCTP = 8, |
| }; |
| |
| /* for hsplit_1 field of Rx RLAN context */ |
| enum ice_rlan_ctx_rx_hsplit_1 { |
| ICE_RLAN_RX_HSPLIT_1_NO_SPLIT = 0, |
| ICE_RLAN_RX_HSPLIT_1_SPLIT_L2 = 1, |
| ICE_RLAN_RX_HSPLIT_1_SPLIT_ALWAYS = 2, |
| }; |
| |
| /* TX Descriptor */ |
| struct ice_tx_desc { |
| __le64 buf_addr; /* Address of descriptor's data buf */ |
| __le64 cmd_type_offset_bsz; |
| }; |
| |
| enum ice_tx_desc_dtype_value { |
| ICE_TX_DESC_DTYPE_DATA = 0x0, |
| ICE_TX_DESC_DTYPE_CTX = 0x1, |
| /* DESC_DONE - HW has completed write-back of descriptor */ |
| ICE_TX_DESC_DTYPE_DESC_DONE = 0xF, |
| }; |
| |
| #define ICE_TXD_QW1_CMD_S 4 |
| #define ICE_TXD_QW1_CMD_M (0xFFFUL << ICE_TXD_QW1_CMD_S) |
| |
| enum ice_tx_desc_cmd_bits { |
| ICE_TX_DESC_CMD_EOP = 0x0001, |
| ICE_TX_DESC_CMD_RS = 0x0002, |
| ICE_TX_DESC_CMD_IL2TAG1 = 0x0008, |
| ICE_TX_DESC_CMD_IIPT_IPV6 = 0x0020, /* 2 BITS */ |
| ICE_TX_DESC_CMD_IIPT_IPV4 = 0x0040, /* 2 BITS */ |
| ICE_TX_DESC_CMD_IIPT_IPV4_CSUM = 0x0060, /* 2 BITS */ |
| ICE_TX_DESC_CMD_L4T_EOFT_TCP = 0x0100, /* 2 BITS */ |
| ICE_TX_DESC_CMD_L4T_EOFT_UDP = 0x0300, /* 2 BITS */ |
| }; |
| |
| #define ICE_TXD_QW1_OFFSET_S 16 |
| #define ICE_TXD_QW1_OFFSET_M (0x3FFFFULL << ICE_TXD_QW1_OFFSET_S) |
| |
| enum ice_tx_desc_len_fields { |
| /* Note: These are predefined bit offsets */ |
| ICE_TX_DESC_LEN_MACLEN_S = 0, /* 7 BITS */ |
| ICE_TX_DESC_LEN_IPLEN_S = 7, /* 7 BITS */ |
| ICE_TX_DESC_LEN_L4_LEN_S = 14 /* 4 BITS */ |
| }; |
| |
| #define ICE_TXD_QW1_MACLEN_M (0x7FUL << ICE_TX_DESC_LEN_MACLEN_S) |
| #define ICE_TXD_QW1_IPLEN_M (0x7FUL << ICE_TX_DESC_LEN_IPLEN_S) |
| #define ICE_TXD_QW1_L4LEN_M (0xFUL << ICE_TX_DESC_LEN_L4_LEN_S) |
| |
| /* Tx descriptor field limits in bytes */ |
| #define ICE_TXD_MACLEN_MAX ((ICE_TXD_QW1_MACLEN_M >> \ |
| ICE_TX_DESC_LEN_MACLEN_S) * ICE_BYTES_PER_WORD) |
| #define ICE_TXD_IPLEN_MAX ((ICE_TXD_QW1_IPLEN_M >> \ |
| ICE_TX_DESC_LEN_IPLEN_S) * ICE_BYTES_PER_DWORD) |
| #define ICE_TXD_L4LEN_MAX ((ICE_TXD_QW1_L4LEN_M >> \ |
| ICE_TX_DESC_LEN_L4_LEN_S) * ICE_BYTES_PER_DWORD) |
| |
| #define ICE_TXD_QW1_TX_BUF_SZ_S 34 |
| #define ICE_TXD_QW1_L2TAG1_S 48 |
| |
| /* Context descriptors */ |
| struct ice_tx_ctx_desc { |
| __le32 tunneling_params; |
| __le16 l2tag2; |
| __le16 rsvd; |
| __le64 qw1; |
| }; |
| |
| #define ICE_TXD_CTX_QW1_CMD_S 4 |
| #define ICE_TXD_CTX_QW1_CMD_M (0x7FUL << ICE_TXD_CTX_QW1_CMD_S) |
| |
| #define ICE_TXD_CTX_QW1_TSO_LEN_S 30 |
| #define ICE_TXD_CTX_QW1_TSO_LEN_M \ |
| (0x3FFFFULL << ICE_TXD_CTX_QW1_TSO_LEN_S) |
| |
| #define ICE_TXD_CTX_QW1_MSS_S 50 |
| |
| enum ice_tx_ctx_desc_cmd_bits { |
| ICE_TX_CTX_DESC_TSO = 0x01, |
| ICE_TX_CTX_DESC_TSYN = 0x02, |
| ICE_TX_CTX_DESC_IL2TAG2 = 0x04, |
| ICE_TX_CTX_DESC_IL2TAG2_IL2H = 0x08, |
| ICE_TX_CTX_DESC_SWTCH_NOTAG = 0x00, |
| ICE_TX_CTX_DESC_SWTCH_UPLINK = 0x10, |
| ICE_TX_CTX_DESC_SWTCH_LOCAL = 0x20, |
| ICE_TX_CTX_DESC_SWTCH_VSI = 0x30, |
| ICE_TX_CTX_DESC_RESERVED = 0x40 |
| }; |
| |
| #define ICE_LAN_TXQ_MAX_QGRPS 127 |
| #define ICE_LAN_TXQ_MAX_QDIS 1023 |
| |
| /* Tx queue context data |
| * |
| * The sizes of the variables may be larger than needed due to crossing byte |
| * boundaries. If we do not have the width of the variable set to the correct |
| * size then we could end up shifting bits off the top of the variable when the |
| * variable is at the top of a byte and crosses over into the next byte. |
| */ |
| struct ice_tlan_ctx { |
| #define ICE_TLAN_CTX_BASE_S 7 |
| u64 base; /* base is defined in 128-byte units */ |
| u8 port_num; |
| u16 cgd_num; /* bigger than needed, see above for reason */ |
| u8 pf_num; |
| u16 vmvf_num; |
| u8 vmvf_type; |
| #define ICE_TLAN_CTX_VMVF_TYPE_VF 0 |
| #define ICE_TLAN_CTX_VMVF_TYPE_VMQ 1 |
| #define ICE_TLAN_CTX_VMVF_TYPE_PF 2 |
| u16 src_vsi; |
| u8 tsyn_ena; |
| u8 alt_vlan; |
| u16 cpuid; /* bigger than needed, see above for reason */ |
| u8 wb_mode; |
| u8 tphrd_desc; |
| u8 tphrd; |
| u8 tphwr_desc; |
| u16 cmpq_id; |
| u16 qnum_in_func; |
| u8 itr_notification_mode; |
| u8 adjust_prof_id; |
| u32 qlen; /* bigger than needed, see above for reason */ |
| u8 quanta_prof_idx; |
| u8 tso_ena; |
| u16 tso_qnum; |
| u8 legacy_int; |
| u8 drop_ena; |
| u8 cache_prof_idx; |
| u8 pkt_shaper_prof_idx; |
| u8 int_q_state; /* width not needed - internal do not write */ |
| }; |
| |
| /* macro to make the table lines short */ |
| #define ICE_PTT(PTYPE, OUTER_IP, OUTER_IP_VER, OUTER_FRAG, T, TE, TEF, I, PL)\ |
| { PTYPE, \ |
| 1, \ |
| ICE_RX_PTYPE_OUTER_##OUTER_IP, \ |
| ICE_RX_PTYPE_OUTER_##OUTER_IP_VER, \ |
| ICE_RX_PTYPE_##OUTER_FRAG, \ |
| ICE_RX_PTYPE_TUNNEL_##T, \ |
| ICE_RX_PTYPE_TUNNEL_END_##TE, \ |
| ICE_RX_PTYPE_##TEF, \ |
| ICE_RX_PTYPE_INNER_PROT_##I, \ |
| ICE_RX_PTYPE_PAYLOAD_LAYER_##PL } |
| |
| #define ICE_PTT_UNUSED_ENTRY(PTYPE) { PTYPE, 0, 0, 0, 0, 0, 0, 0, 0, 0 } |
| |
| /* shorter macros makes the table fit but are terse */ |
| #define ICE_RX_PTYPE_NOF ICE_RX_PTYPE_NOT_FRAG |
| |
| /* Lookup table mapping the HW PTYPE to the bit field for decoding */ |
| static const struct ice_rx_ptype_decoded ice_ptype_lkup[] = { |
| /* L2 Packet types */ |
| ICE_PTT_UNUSED_ENTRY(0), |
| ICE_PTT(1, L2, NONE, NOF, NONE, NONE, NOF, NONE, PAY2), |
| ICE_PTT(2, L2, NONE, NOF, NONE, NONE, NOF, NONE, NONE), |
| }; |
| |
| static inline struct ice_rx_ptype_decoded ice_decode_rx_desc_ptype(u16 ptype) |
| { |
| return ice_ptype_lkup[ptype]; |
| } |
| |
| #define ICE_LINK_SPEED_UNKNOWN 0 |
| #define ICE_LINK_SPEED_10MBPS 10 |
| #define ICE_LINK_SPEED_100MBPS 100 |
| #define ICE_LINK_SPEED_1000MBPS 1000 |
| #define ICE_LINK_SPEED_2500MBPS 2500 |
| #define ICE_LINK_SPEED_5000MBPS 5000 |
| #define ICE_LINK_SPEED_10000MBPS 10000 |
| #define ICE_LINK_SPEED_20000MBPS 20000 |
| #define ICE_LINK_SPEED_25000MBPS 25000 |
| #define ICE_LINK_SPEED_40000MBPS 40000 |
| |
| #endif /* _ICE_LAN_TX_RX_H_ */ |