| /* SPDX-License-Identifier: (GPL-2.0-only OR BSD-3-Clause) */ |
| /* QLogic qed NIC Driver |
| * Copyright (c) 2015-2017 QLogic Corporation |
| * Copyright (c) 2019-2020 Marvell International Ltd. |
| */ |
| |
| #ifndef _QED_SRIOV_H |
| #define _QED_SRIOV_H |
| #include <linux/types.h> |
| #include "qed_vf.h" |
| |
| #define QED_ETH_VF_NUM_MAC_FILTERS 1 |
| #define QED_ETH_VF_NUM_VLAN_FILTERS 2 |
| #define QED_VF_ARRAY_LENGTH (3) |
| |
| #ifdef CONFIG_QED_SRIOV |
| #define IS_VF(cdev) ((cdev)->b_is_vf) |
| #define IS_PF(cdev) (!((cdev)->b_is_vf)) |
| #define IS_PF_SRIOV(p_hwfn) (!!((p_hwfn)->cdev->p_iov_info)) |
| #else |
| #define IS_VF(cdev) (0) |
| #define IS_PF(cdev) (1) |
| #define IS_PF_SRIOV(p_hwfn) (0) |
| #endif |
| #define IS_PF_SRIOV_ALLOC(p_hwfn) (!!((p_hwfn)->pf_iov_info)) |
| |
| #define QED_MAX_VF_CHAINS_PER_PF 16 |
| |
| #define QED_ETH_MAX_VF_NUM_VLAN_FILTERS \ |
| (MAX_NUM_VFS * QED_ETH_VF_NUM_VLAN_FILTERS) |
| |
| enum qed_iov_vport_update_flag { |
| QED_IOV_VP_UPDATE_ACTIVATE, |
| QED_IOV_VP_UPDATE_VLAN_STRIP, |
| QED_IOV_VP_UPDATE_TX_SWITCH, |
| QED_IOV_VP_UPDATE_MCAST, |
| QED_IOV_VP_UPDATE_ACCEPT_PARAM, |
| QED_IOV_VP_UPDATE_RSS, |
| QED_IOV_VP_UPDATE_ACCEPT_ANY_VLAN, |
| QED_IOV_VP_UPDATE_SGE_TPA, |
| QED_IOV_VP_UPDATE_MAX, |
| }; |
| |
| struct qed_public_vf_info { |
| /* These copies will later be reflected in the bulletin board, |
| * but this copy should be newer. |
| */ |
| u8 forced_mac[ETH_ALEN]; |
| u16 forced_vlan; |
| u8 mac[ETH_ALEN]; |
| |
| /* IFLA_VF_LINK_STATE_<X> */ |
| int link_state; |
| |
| /* Currently configured Tx rate in MB/sec. 0 if unconfigured */ |
| int tx_rate; |
| |
| /* Trusted VFs can configure promiscuous mode. |
| * Also store shadow promisc configuration if needed. |
| */ |
| bool is_trusted_configured; |
| bool is_trusted_request; |
| u8 rx_accept_mode; |
| u8 tx_accept_mode; |
| }; |
| |
| struct qed_iov_vf_init_params { |
| u16 rel_vf_id; |
| |
| /* Number of requested Queues; Currently, don't support different |
| * number of Rx/Tx queues. |
| */ |
| |
| u16 num_queues; |
| |
| /* Allow the client to choose which qzones to use for Rx/Tx, |
| * and which queue_base to use for Tx queues on a per-queue basis. |
| * Notice values should be relative to the PF resources. |
| */ |
| u16 req_rx_queue[QED_MAX_VF_CHAINS_PER_PF]; |
| u16 req_tx_queue[QED_MAX_VF_CHAINS_PER_PF]; |
| }; |
| |
| /* This struct is part of qed_dev and contains data relevant to all hwfns; |
| * Initialized only if SR-IOV cpabability is exposed in PCIe config space. |
| */ |
| struct qed_hw_sriov_info { |
| int pos; /* capability position */ |
| int nres; /* number of resources */ |
| u32 cap; /* SR-IOV Capabilities */ |
| u16 ctrl; /* SR-IOV Control */ |
| u16 total_vfs; /* total VFs associated with the PF */ |
| u16 num_vfs; /* number of vfs that have been started */ |
| u16 initial_vfs; /* initial VFs associated with the PF */ |
| u16 nr_virtfn; /* number of VFs available */ |
| u16 offset; /* first VF Routing ID offset */ |
| u16 stride; /* following VF stride */ |
| u16 vf_device_id; /* VF device id */ |
| u32 pgsz; /* page size for BAR alignment */ |
| u8 link; /* Function Dependency Link */ |
| |
| u32 first_vf_in_pf; |
| }; |
| |
| /* This mailbox is maintained per VF in its PF contains all information |
| * required for sending / receiving a message. |
| */ |
| struct qed_iov_vf_mbx { |
| union vfpf_tlvs *req_virt; |
| dma_addr_t req_phys; |
| union pfvf_tlvs *reply_virt; |
| dma_addr_t reply_phys; |
| |
| /* Address in VF where a pending message is located */ |
| dma_addr_t pending_req; |
| |
| /* Message from VF awaits handling */ |
| bool b_pending_msg; |
| |
| u8 *offset; |
| |
| /* saved VF request header */ |
| struct vfpf_first_tlv first_tlv; |
| }; |
| |
| #define QED_IOV_LEGACY_QID_RX (0) |
| #define QED_IOV_LEGACY_QID_TX (1) |
| #define QED_IOV_QID_INVALID (0xFE) |
| |
| struct qed_vf_queue_cid { |
| bool b_is_tx; |
| struct qed_queue_cid *p_cid; |
| }; |
| |
| /* Describes a qzone associated with the VF */ |
| struct qed_vf_queue { |
| u16 fw_rx_qid; |
| u16 fw_tx_qid; |
| |
| struct qed_vf_queue_cid cids[MAX_QUEUES_PER_QZONE]; |
| }; |
| |
| enum vf_state { |
| VF_FREE = 0, /* VF ready to be acquired holds no resc */ |
| VF_ACQUIRED, /* VF, acquired, but not initialized */ |
| VF_ENABLED, /* VF, Enabled */ |
| VF_RESET, /* VF, FLR'd, pending cleanup */ |
| VF_STOPPED /* VF, Stopped */ |
| }; |
| |
| struct qed_vf_vlan_shadow { |
| bool used; |
| u16 vid; |
| }; |
| |
| struct qed_vf_shadow_config { |
| /* Shadow copy of all guest vlans */ |
| struct qed_vf_vlan_shadow vlans[QED_ETH_VF_NUM_VLAN_FILTERS + 1]; |
| |
| /* Shadow copy of all configured MACs; Empty if forcing MACs */ |
| u8 macs[QED_ETH_VF_NUM_MAC_FILTERS][ETH_ALEN]; |
| u8 inner_vlan_removal; |
| }; |
| |
| /* PFs maintain an array of this structure, per VF */ |
| struct qed_vf_info { |
| struct qed_iov_vf_mbx vf_mbx; |
| enum vf_state state; |
| bool b_init; |
| bool b_malicious; |
| u8 to_disable; |
| |
| struct qed_bulletin bulletin; |
| dma_addr_t vf_bulletin; |
| |
| /* PF saves a copy of the last VF acquire message */ |
| struct vfpf_acquire_tlv acquire; |
| |
| u32 concrete_fid; |
| u16 opaque_fid; |
| u16 mtu; |
| |
| u8 vport_id; |
| u8 relative_vf_id; |
| u8 abs_vf_id; |
| #define QED_VF_ABS_ID(p_hwfn, p_vf) (QED_PATH_ID(p_hwfn) ? \ |
| (p_vf)->abs_vf_id + MAX_NUM_VFS_BB : \ |
| (p_vf)->abs_vf_id) |
| |
| u8 vport_instance; |
| u8 num_rxqs; |
| u8 num_txqs; |
| |
| u16 rx_coal; |
| u16 tx_coal; |
| |
| u8 num_sbs; |
| |
| u8 num_mac_filters; |
| u8 num_vlan_filters; |
| |
| struct qed_vf_queue vf_queues[QED_MAX_VF_CHAINS_PER_PF]; |
| u16 igu_sbs[QED_MAX_VF_CHAINS_PER_PF]; |
| u8 num_active_rxqs; |
| struct qed_public_vf_info p_vf_info; |
| bool spoof_chk; |
| bool req_spoofchk_val; |
| |
| /* Stores the configuration requested by VF */ |
| struct qed_vf_shadow_config shadow_config; |
| |
| /* A bitfield using bulletin's valid-map bits, used to indicate |
| * which of the bulletin board features have been configured. |
| */ |
| u64 configured_features; |
| #define QED_IOV_CONFIGURED_FEATURES_MASK ((1 << MAC_ADDR_FORCED) | \ |
| (1 << VLAN_ADDR_FORCED)) |
| }; |
| |
| /* This structure is part of qed_hwfn and used only for PFs that have sriov |
| * capability enabled. |
| */ |
| struct qed_pf_iov { |
| struct qed_vf_info vfs_array[MAX_NUM_VFS]; |
| u64 pending_flr[QED_VF_ARRAY_LENGTH]; |
| |
| /* Allocate message address continuosuly and split to each VF */ |
| void *mbx_msg_virt_addr; |
| dma_addr_t mbx_msg_phys_addr; |
| u32 mbx_msg_size; |
| void *mbx_reply_virt_addr; |
| dma_addr_t mbx_reply_phys_addr; |
| u32 mbx_reply_size; |
| void *p_bulletins; |
| dma_addr_t bulletins_phys; |
| u32 bulletins_size; |
| }; |
| |
| enum qed_iov_wq_flag { |
| QED_IOV_WQ_MSG_FLAG, |
| QED_IOV_WQ_SET_UNICAST_FILTER_FLAG, |
| QED_IOV_WQ_BULLETIN_UPDATE_FLAG, |
| QED_IOV_WQ_STOP_WQ_FLAG, |
| QED_IOV_WQ_FLR_FLAG, |
| QED_IOV_WQ_TRUST_FLAG, |
| QED_IOV_WQ_VF_FORCE_LINK_QUERY_FLAG, |
| }; |
| |
| extern const struct qed_iov_hv_ops qed_iov_ops_pass; |
| |
| #ifdef CONFIG_QED_SRIOV |
| /** |
| * qed_iov_is_valid_vfid(): Check if given VF ID @vfid is valid |
| * w.r.t. @b_enabled_only value |
| * if b_enabled_only = true - only enabled |
| * VF id is valid. |
| * else any VF id less than max_vfs is valid. |
| * |
| * @p_hwfn: HW device data. |
| * @rel_vf_id: Relative VF ID. |
| * @b_enabled_only: consider only enabled VF. |
| * @b_non_malicious: true iff we want to validate vf isn't malicious. |
| * |
| * Return: bool - true for valid VF ID |
| */ |
| bool qed_iov_is_valid_vfid(struct qed_hwfn *p_hwfn, |
| int rel_vf_id, |
| bool b_enabled_only, bool b_non_malicious); |
| |
| /** |
| * qed_iov_get_next_active_vf(): Given a VF index, return index of |
| * next [including that] active VF. |
| * |
| * @p_hwfn: HW device data. |
| * @rel_vf_id: VF ID. |
| * |
| * Return: MAX_NUM_VFS in case no further active VFs, otherwise index. |
| */ |
| u16 qed_iov_get_next_active_vf(struct qed_hwfn *p_hwfn, u16 rel_vf_id); |
| |
| void qed_iov_bulletin_set_udp_ports(struct qed_hwfn *p_hwfn, |
| int vfid, u16 vxlan_port, u16 geneve_port); |
| |
| /** |
| * qed_iov_hw_info(): Read sriov related information and allocated resources |
| * reads from configuration space, shmem, etc. |
| * |
| * @p_hwfn: HW device data. |
| * |
| * Return: Int. |
| */ |
| int qed_iov_hw_info(struct qed_hwfn *p_hwfn); |
| |
| /** |
| * qed_add_tlv(): place a given tlv on the tlv buffer at next offset |
| * |
| * @p_hwfn: HW device data. |
| * @offset: offset. |
| * @type: Type |
| * @length: Length. |
| * |
| * Return: pointer to the newly placed tlv |
| */ |
| void *qed_add_tlv(struct qed_hwfn *p_hwfn, u8 **offset, u16 type, u16 length); |
| |
| /** |
| * qed_dp_tlv_list(): list the types and lengths of the tlvs on the buffer |
| * |
| * @p_hwfn: HW device data. |
| * @tlvs_list: Tlvs_list. |
| * |
| * Return: Void. |
| */ |
| void qed_dp_tlv_list(struct qed_hwfn *p_hwfn, void *tlvs_list); |
| |
| /** |
| * qed_sriov_vfpf_malicious(): Handle malicious VF/PF. |
| * |
| * @p_hwfn: HW device data. |
| * @p_data: Pointer to data. |
| * |
| * Return: Void. |
| */ |
| void qed_sriov_vfpf_malicious(struct qed_hwfn *p_hwfn, |
| struct fw_err_data *p_data); |
| |
| /** |
| * qed_sriov_eqe_event(): Callback for SRIOV events. |
| * |
| * @p_hwfn: HW device data. |
| * @opcode: Opcode. |
| * @echo: Echo. |
| * @data: data |
| * @fw_return_code: FW return code. |
| * |
| * Return: Int. |
| */ |
| int qed_sriov_eqe_event(struct qed_hwfn *p_hwfn, u8 opcode, __le16 echo, |
| union event_ring_data *data, u8 fw_return_code); |
| |
| /** |
| * qed_iov_alloc(): allocate sriov related resources |
| * |
| * @p_hwfn: HW device data. |
| * |
| * Return: Int. |
| */ |
| int qed_iov_alloc(struct qed_hwfn *p_hwfn); |
| |
| /** |
| * qed_iov_setup(): setup sriov related resources |
| * |
| * @p_hwfn: HW device data. |
| * |
| * Return: Void. |
| */ |
| void qed_iov_setup(struct qed_hwfn *p_hwfn); |
| |
| /** |
| * qed_iov_free(): free sriov related resources |
| * |
| * @p_hwfn: HW device data. |
| * |
| * Return: Void. |
| */ |
| void qed_iov_free(struct qed_hwfn *p_hwfn); |
| |
| /** |
| * qed_iov_free_hw_info(): free sriov related memory that was |
| * allocated during hw_prepare |
| * |
| * @cdev: Qed dev pointer. |
| * |
| * Return: Void. |
| */ |
| void qed_iov_free_hw_info(struct qed_dev *cdev); |
| |
| /** |
| * qed_iov_mark_vf_flr(): Mark structs of vfs that have been FLR-ed. |
| * |
| * @p_hwfn: HW device data. |
| * @disabled_vfs: bitmask of all VFs on path that were FLRed |
| * |
| * Return: true iff one of the PF's vfs got FLRed. false otherwise. |
| */ |
| bool qed_iov_mark_vf_flr(struct qed_hwfn *p_hwfn, u32 *disabled_vfs); |
| |
| /** |
| * qed_iov_search_list_tlvs(): Search extended TLVs in request/reply buffer. |
| * |
| * @p_hwfn: HW device data. |
| * @p_tlvs_list: Pointer to tlvs list |
| * @req_type: Type of TLV |
| * |
| * Return: pointer to tlv type if found, otherwise returns NULL. |
| */ |
| void *qed_iov_search_list_tlvs(struct qed_hwfn *p_hwfn, |
| void *p_tlvs_list, u16 req_type); |
| |
| void qed_iov_wq_stop(struct qed_dev *cdev, bool schedule_first); |
| int qed_iov_wq_start(struct qed_dev *cdev); |
| |
| void qed_schedule_iov(struct qed_hwfn *hwfn, enum qed_iov_wq_flag flag); |
| void qed_vf_start_iov_wq(struct qed_dev *cdev); |
| int qed_sriov_disable(struct qed_dev *cdev, bool pci_enabled); |
| void qed_inform_vf_link_state(struct qed_hwfn *hwfn); |
| #else |
| static inline bool |
| qed_iov_is_valid_vfid(struct qed_hwfn *p_hwfn, |
| int rel_vf_id, bool b_enabled_only, bool b_non_malicious) |
| { |
| return false; |
| } |
| |
| static inline u16 qed_iov_get_next_active_vf(struct qed_hwfn *p_hwfn, |
| u16 rel_vf_id) |
| { |
| return MAX_NUM_VFS; |
| } |
| |
| static inline void |
| qed_iov_bulletin_set_udp_ports(struct qed_hwfn *p_hwfn, int vfid, |
| u16 vxlan_port, u16 geneve_port) |
| { |
| } |
| |
| static inline int qed_iov_hw_info(struct qed_hwfn *p_hwfn) |
| { |
| return 0; |
| } |
| |
| static inline int qed_iov_alloc(struct qed_hwfn *p_hwfn) |
| { |
| return 0; |
| } |
| |
| static inline void qed_iov_setup(struct qed_hwfn *p_hwfn) |
| { |
| } |
| |
| static inline void qed_iov_free(struct qed_hwfn *p_hwfn) |
| { |
| } |
| |
| static inline void qed_iov_free_hw_info(struct qed_dev *cdev) |
| { |
| } |
| |
| static inline bool qed_iov_mark_vf_flr(struct qed_hwfn *p_hwfn, |
| u32 *disabled_vfs) |
| { |
| return false; |
| } |
| |
| static inline void qed_iov_wq_stop(struct qed_dev *cdev, bool schedule_first) |
| { |
| } |
| |
| static inline int qed_iov_wq_start(struct qed_dev *cdev) |
| { |
| return 0; |
| } |
| |
| static inline void qed_schedule_iov(struct qed_hwfn *hwfn, |
| enum qed_iov_wq_flag flag) |
| { |
| } |
| |
| static inline void qed_vf_start_iov_wq(struct qed_dev *cdev) |
| { |
| } |
| |
| static inline int qed_sriov_disable(struct qed_dev *cdev, bool pci_enabled) |
| { |
| return 0; |
| } |
| |
| static inline void qed_inform_vf_link_state(struct qed_hwfn *hwfn) |
| { |
| } |
| |
| static inline void qed_sriov_vfpf_malicious(struct qed_hwfn *p_hwfn, |
| struct fw_err_data *p_data) |
| { |
| } |
| |
| static inline int qed_sriov_eqe_event(struct qed_hwfn *p_hwfn, u8 opcode, |
| __le16 echo, union event_ring_data *data, |
| u8 fw_return_code) |
| { |
| return 0; |
| } |
| #endif |
| |
| #define qed_for_each_vf(_p_hwfn, _i) \ |
| for (_i = qed_iov_get_next_active_vf(_p_hwfn, 0); \ |
| _i < MAX_NUM_VFS; \ |
| _i = qed_iov_get_next_active_vf(_p_hwfn, _i + 1)) |
| |
| #endif |