blob: 63deb120359cfaa753f4edb1b6501724f588f0c0 [file] [log] [blame]
/* SPDX-License-Identifier: GPL-2.0-only */
/* Copyright (C) 2023 Intel Corporation */
#ifndef _VIRTCHNL2_H_
#define _VIRTCHNL2_H_
#include <linux/if_ether.h>
/* All opcodes associated with virtchnl2 are prefixed with virtchnl2 or
* VIRTCHNL2. Any future opcodes, offloads/capabilities, structures,
* and defines must be prefixed with virtchnl2 or VIRTCHNL2 to avoid confusion.
*
* PF/VF uses the virtchnl2 interface defined in this header file to communicate
* with device Control Plane (CP). Driver and the CP may run on different
* platforms with different endianness. To avoid byte order discrepancies,
* all the structures in this header follow little-endian format.
*
* This is an interface definition file where existing enums and their values
* must remain unchanged over time, so we specify explicit values for all enums.
*/
/* This macro is used to generate compilation errors if a structure
* is not exactly the correct length.
*/
#define VIRTCHNL2_CHECK_STRUCT_LEN(n, X) \
static_assert((n) == sizeof(struct X))
/* New major set of opcodes introduced and so leaving room for
* old misc opcodes to be added in future. Also these opcodes may only
* be used if both the PF and VF have successfully negotiated the
* VIRTCHNL version as 2.0 during VIRTCHNL2_OP_VERSION exchange.
*/
enum virtchnl2_op {
VIRTCHNL2_OP_UNKNOWN = 0,
VIRTCHNL2_OP_VERSION = 1,
VIRTCHNL2_OP_GET_CAPS = 500,
VIRTCHNL2_OP_CREATE_VPORT = 501,
VIRTCHNL2_OP_DESTROY_VPORT = 502,
VIRTCHNL2_OP_ENABLE_VPORT = 503,
VIRTCHNL2_OP_DISABLE_VPORT = 504,
VIRTCHNL2_OP_CONFIG_TX_QUEUES = 505,
VIRTCHNL2_OP_CONFIG_RX_QUEUES = 506,
VIRTCHNL2_OP_ENABLE_QUEUES = 507,
VIRTCHNL2_OP_DISABLE_QUEUES = 508,
VIRTCHNL2_OP_ADD_QUEUES = 509,
VIRTCHNL2_OP_DEL_QUEUES = 510,
VIRTCHNL2_OP_MAP_QUEUE_VECTOR = 511,
VIRTCHNL2_OP_UNMAP_QUEUE_VECTOR = 512,
VIRTCHNL2_OP_GET_RSS_KEY = 513,
VIRTCHNL2_OP_SET_RSS_KEY = 514,
VIRTCHNL2_OP_GET_RSS_LUT = 515,
VIRTCHNL2_OP_SET_RSS_LUT = 516,
VIRTCHNL2_OP_GET_RSS_HASH = 517,
VIRTCHNL2_OP_SET_RSS_HASH = 518,
VIRTCHNL2_OP_SET_SRIOV_VFS = 519,
VIRTCHNL2_OP_ALLOC_VECTORS = 520,
VIRTCHNL2_OP_DEALLOC_VECTORS = 521,
VIRTCHNL2_OP_EVENT = 522,
VIRTCHNL2_OP_GET_STATS = 523,
VIRTCHNL2_OP_RESET_VF = 524,
VIRTCHNL2_OP_GET_EDT_CAPS = 525,
VIRTCHNL2_OP_GET_PTYPE_INFO = 526,
/* Opcode 527 and 528 are reserved for VIRTCHNL2_OP_GET_PTYPE_ID and
* VIRTCHNL2_OP_GET_PTYPE_INFO_RAW.
* Opcodes 529, 530, 531, 532 and 533 are reserved.
*/
VIRTCHNL2_OP_LOOPBACK = 534,
VIRTCHNL2_OP_ADD_MAC_ADDR = 535,
VIRTCHNL2_OP_DEL_MAC_ADDR = 536,
VIRTCHNL2_OP_CONFIG_PROMISCUOUS_MODE = 537,
};
/**
* enum virtchnl2_vport_type - Type of virtual port.
* @VIRTCHNL2_VPORT_TYPE_DEFAULT: Default virtual port type.
*/
enum virtchnl2_vport_type {
VIRTCHNL2_VPORT_TYPE_DEFAULT = 0,
};
/**
* enum virtchnl2_queue_model - Type of queue model.
* @VIRTCHNL2_QUEUE_MODEL_SINGLE: Single queue model.
* @VIRTCHNL2_QUEUE_MODEL_SPLIT: Split queue model.
*
* In the single queue model, the same transmit descriptor queue is used by
* software to post descriptors to hardware and by hardware to post completed
* descriptors to software.
* Likewise, the same receive descriptor queue is used by hardware to post
* completions to software and by software to post buffers to hardware.
*
* In the split queue model, hardware uses transmit completion queues to post
* descriptor/buffer completions to software, while software uses transmit
* descriptor queues to post descriptors to hardware.
* Likewise, hardware posts descriptor completions to the receive descriptor
* queue, while software uses receive buffer queues to post buffers to hardware.
*/
enum virtchnl2_queue_model {
VIRTCHNL2_QUEUE_MODEL_SINGLE = 0,
VIRTCHNL2_QUEUE_MODEL_SPLIT = 1,
};
/* Checksum offload capability flags */
enum virtchnl2_cap_txrx_csum {
VIRTCHNL2_CAP_TX_CSUM_L3_IPV4 = BIT(0),
VIRTCHNL2_CAP_TX_CSUM_L4_IPV4_TCP = BIT(1),
VIRTCHNL2_CAP_TX_CSUM_L4_IPV4_UDP = BIT(2),
VIRTCHNL2_CAP_TX_CSUM_L4_IPV4_SCTP = BIT(3),
VIRTCHNL2_CAP_TX_CSUM_L4_IPV6_TCP = BIT(4),
VIRTCHNL2_CAP_TX_CSUM_L4_IPV6_UDP = BIT(5),
VIRTCHNL2_CAP_TX_CSUM_L4_IPV6_SCTP = BIT(6),
VIRTCHNL2_CAP_TX_CSUM_GENERIC = BIT(7),
VIRTCHNL2_CAP_RX_CSUM_L3_IPV4 = BIT(8),
VIRTCHNL2_CAP_RX_CSUM_L4_IPV4_TCP = BIT(9),
VIRTCHNL2_CAP_RX_CSUM_L4_IPV4_UDP = BIT(10),
VIRTCHNL2_CAP_RX_CSUM_L4_IPV4_SCTP = BIT(11),
VIRTCHNL2_CAP_RX_CSUM_L4_IPV6_TCP = BIT(12),
VIRTCHNL2_CAP_RX_CSUM_L4_IPV6_UDP = BIT(13),
VIRTCHNL2_CAP_RX_CSUM_L4_IPV6_SCTP = BIT(14),
VIRTCHNL2_CAP_RX_CSUM_GENERIC = BIT(15),
VIRTCHNL2_CAP_TX_CSUM_L3_SINGLE_TUNNEL = BIT(16),
VIRTCHNL2_CAP_TX_CSUM_L3_DOUBLE_TUNNEL = BIT(17),
VIRTCHNL2_CAP_RX_CSUM_L3_SINGLE_TUNNEL = BIT(18),
VIRTCHNL2_CAP_RX_CSUM_L3_DOUBLE_TUNNEL = BIT(19),
VIRTCHNL2_CAP_TX_CSUM_L4_SINGLE_TUNNEL = BIT(20),
VIRTCHNL2_CAP_TX_CSUM_L4_DOUBLE_TUNNEL = BIT(21),
VIRTCHNL2_CAP_RX_CSUM_L4_SINGLE_TUNNEL = BIT(22),
VIRTCHNL2_CAP_RX_CSUM_L4_DOUBLE_TUNNEL = BIT(23),
};
/* Segmentation offload capability flags */
enum virtchnl2_cap_seg {
VIRTCHNL2_CAP_SEG_IPV4_TCP = BIT(0),
VIRTCHNL2_CAP_SEG_IPV4_UDP = BIT(1),
VIRTCHNL2_CAP_SEG_IPV4_SCTP = BIT(2),
VIRTCHNL2_CAP_SEG_IPV6_TCP = BIT(3),
VIRTCHNL2_CAP_SEG_IPV6_UDP = BIT(4),
VIRTCHNL2_CAP_SEG_IPV6_SCTP = BIT(5),
VIRTCHNL2_CAP_SEG_GENERIC = BIT(6),
VIRTCHNL2_CAP_SEG_TX_SINGLE_TUNNEL = BIT(7),
VIRTCHNL2_CAP_SEG_TX_DOUBLE_TUNNEL = BIT(8),
};
/* Receive Side Scaling Flow type capability flags */
enum virtchnl2_cap_rss {
VIRTCHNL2_CAP_RSS_IPV4_TCP = BIT(0),
VIRTCHNL2_CAP_RSS_IPV4_UDP = BIT(1),
VIRTCHNL2_CAP_RSS_IPV4_SCTP = BIT(2),
VIRTCHNL2_CAP_RSS_IPV4_OTHER = BIT(3),
VIRTCHNL2_CAP_RSS_IPV6_TCP = BIT(4),
VIRTCHNL2_CAP_RSS_IPV6_UDP = BIT(5),
VIRTCHNL2_CAP_RSS_IPV6_SCTP = BIT(6),
VIRTCHNL2_CAP_RSS_IPV6_OTHER = BIT(7),
VIRTCHNL2_CAP_RSS_IPV4_AH = BIT(8),
VIRTCHNL2_CAP_RSS_IPV4_ESP = BIT(9),
VIRTCHNL2_CAP_RSS_IPV4_AH_ESP = BIT(10),
VIRTCHNL2_CAP_RSS_IPV6_AH = BIT(11),
VIRTCHNL2_CAP_RSS_IPV6_ESP = BIT(12),
VIRTCHNL2_CAP_RSS_IPV6_AH_ESP = BIT(13),
};
/* Header split capability flags */
enum virtchnl2_cap_rx_hsplit_at {
/* for prepended metadata */
VIRTCHNL2_CAP_RX_HSPLIT_AT_L2 = BIT(0),
/* all VLANs go into header buffer */
VIRTCHNL2_CAP_RX_HSPLIT_AT_L3 = BIT(1),
VIRTCHNL2_CAP_RX_HSPLIT_AT_L4V4 = BIT(2),
VIRTCHNL2_CAP_RX_HSPLIT_AT_L4V6 = BIT(3),
};
/* Receive Side Coalescing offload capability flags */
enum virtchnl2_cap_rsc {
VIRTCHNL2_CAP_RSC_IPV4_TCP = BIT(0),
VIRTCHNL2_CAP_RSC_IPV4_SCTP = BIT(1),
VIRTCHNL2_CAP_RSC_IPV6_TCP = BIT(2),
VIRTCHNL2_CAP_RSC_IPV6_SCTP = BIT(3),
};
/* Other capability flags */
enum virtchnl2_cap_other {
VIRTCHNL2_CAP_RDMA = BIT_ULL(0),
VIRTCHNL2_CAP_SRIOV = BIT_ULL(1),
VIRTCHNL2_CAP_MACFILTER = BIT_ULL(2),
VIRTCHNL2_CAP_FLOW_DIRECTOR = BIT_ULL(3),
/* Queue based scheduling using split queue model */
VIRTCHNL2_CAP_SPLITQ_QSCHED = BIT_ULL(4),
VIRTCHNL2_CAP_CRC = BIT_ULL(5),
VIRTCHNL2_CAP_ADQ = BIT_ULL(6),
VIRTCHNL2_CAP_WB_ON_ITR = BIT_ULL(7),
VIRTCHNL2_CAP_PROMISC = BIT_ULL(8),
VIRTCHNL2_CAP_LINK_SPEED = BIT_ULL(9),
VIRTCHNL2_CAP_INLINE_IPSEC = BIT_ULL(10),
VIRTCHNL2_CAP_LARGE_NUM_QUEUES = BIT_ULL(11),
VIRTCHNL2_CAP_VLAN = BIT_ULL(12),
VIRTCHNL2_CAP_PTP = BIT_ULL(13),
/* EDT: Earliest Departure Time capability used for Timing Wheel */
VIRTCHNL2_CAP_EDT = BIT_ULL(14),
VIRTCHNL2_CAP_ADV_RSS = BIT_ULL(15),
VIRTCHNL2_CAP_FDIR = BIT_ULL(16),
VIRTCHNL2_CAP_RX_FLEX_DESC = BIT_ULL(17),
VIRTCHNL2_CAP_PTYPE = BIT_ULL(18),
VIRTCHNL2_CAP_LOOPBACK = BIT_ULL(19),
/* Other capability 20 is reserved */
/* this must be the last capability */
VIRTCHNL2_CAP_OEM = BIT_ULL(63),
};
/* underlying device type */
enum virtchl2_device_type {
VIRTCHNL2_MEV_DEVICE = 0,
};
/**
* enum virtchnl2_txq_sched_mode - Transmit Queue Scheduling Modes.
* @VIRTCHNL2_TXQ_SCHED_MODE_QUEUE: Queue mode is the legacy mode i.e. inorder
* completions where descriptors and buffers
* are completed at the same time.
* @VIRTCHNL2_TXQ_SCHED_MODE_FLOW: Flow scheduling mode allows for out of order
* packet processing where descriptors are
* cleaned in order, but buffers can be
* completed out of order.
*/
enum virtchnl2_txq_sched_mode {
VIRTCHNL2_TXQ_SCHED_MODE_QUEUE = 0,
VIRTCHNL2_TXQ_SCHED_MODE_FLOW = 1,
};
/**
* enum virtchnl2_rxq_flags - Receive Queue Feature flags.
* @VIRTCHNL2_RXQ_RSC: Rx queue RSC flag.
* @VIRTCHNL2_RXQ_HDR_SPLIT: Rx queue header split flag.
* @VIRTCHNL2_RXQ_IMMEDIATE_WRITE_BACK: When set, packet descriptors are flushed
* by hardware immediately after processing
* each packet.
* @VIRTCHNL2_RX_DESC_SIZE_16BYTE: Rx queue 16 byte descriptor size.
* @VIRTCHNL2_RX_DESC_SIZE_32BYTE: Rx queue 32 byte descriptor size.
*/
enum virtchnl2_rxq_flags {
VIRTCHNL2_RXQ_RSC = BIT(0),
VIRTCHNL2_RXQ_HDR_SPLIT = BIT(1),
VIRTCHNL2_RXQ_IMMEDIATE_WRITE_BACK = BIT(2),
VIRTCHNL2_RX_DESC_SIZE_16BYTE = BIT(3),
VIRTCHNL2_RX_DESC_SIZE_32BYTE = BIT(4),
};
/* Type of RSS algorithm */
enum virtchnl2_rss_alg {
VIRTCHNL2_RSS_ALG_TOEPLITZ_ASYMMETRIC = 0,
VIRTCHNL2_RSS_ALG_R_ASYMMETRIC = 1,
VIRTCHNL2_RSS_ALG_TOEPLITZ_SYMMETRIC = 2,
VIRTCHNL2_RSS_ALG_XOR_SYMMETRIC = 3,
};
/* Type of event */
enum virtchnl2_event_codes {
VIRTCHNL2_EVENT_UNKNOWN = 0,
VIRTCHNL2_EVENT_LINK_CHANGE = 1,
/* Event type 2, 3 are reserved */
};
/* Transmit and Receive queue types are valid in legacy as well as split queue
* models. With Split Queue model, 2 additional types are introduced -
* TX_COMPLETION and RX_BUFFER. In split queue model, receive corresponds to
* the queue where hardware posts completions.
*/
enum virtchnl2_queue_type {
VIRTCHNL2_QUEUE_TYPE_TX = 0,
VIRTCHNL2_QUEUE_TYPE_RX = 1,
VIRTCHNL2_QUEUE_TYPE_TX_COMPLETION = 2,
VIRTCHNL2_QUEUE_TYPE_RX_BUFFER = 3,
VIRTCHNL2_QUEUE_TYPE_CONFIG_TX = 4,
VIRTCHNL2_QUEUE_TYPE_CONFIG_RX = 5,
/* Queue types 6, 7, 8, 9 are reserved */
VIRTCHNL2_QUEUE_TYPE_MBX_TX = 10,
VIRTCHNL2_QUEUE_TYPE_MBX_RX = 11,
};
/* Interrupt throttling rate index */
enum virtchnl2_itr_idx {
VIRTCHNL2_ITR_IDX_0 = 0,
VIRTCHNL2_ITR_IDX_1 = 1,
};
/**
* enum virtchnl2_mac_addr_type - MAC address types.
* @VIRTCHNL2_MAC_ADDR_PRIMARY: PF/VF driver should set this type for the
* primary/device unicast MAC address filter for
* VIRTCHNL2_OP_ADD_MAC_ADDR and
* VIRTCHNL2_OP_DEL_MAC_ADDR. This allows for the
* underlying control plane function to accurately
* track the MAC address and for VM/function reset.
*
* @VIRTCHNL2_MAC_ADDR_EXTRA: PF/VF driver should set this type for any extra
* unicast and/or multicast filters that are being
* added/deleted via VIRTCHNL2_OP_ADD_MAC_ADDR or
* VIRTCHNL2_OP_DEL_MAC_ADDR.
*/
enum virtchnl2_mac_addr_type {
VIRTCHNL2_MAC_ADDR_PRIMARY = 1,
VIRTCHNL2_MAC_ADDR_EXTRA = 2,
};
/* Flags used for promiscuous mode */
enum virtchnl2_promisc_flags {
VIRTCHNL2_UNICAST_PROMISC = BIT(0),
VIRTCHNL2_MULTICAST_PROMISC = BIT(1),
};
/* Protocol header type within a packet segment. A segment consists of one or
* more protocol headers that make up a logical group of protocol headers. Each
* logical group of protocol headers encapsulates or is encapsulated using/by
* tunneling or encapsulation protocols for network virtualization.
*/
enum virtchnl2_proto_hdr_type {
/* VIRTCHNL2_PROTO_HDR_ANY is a mandatory protocol id */
VIRTCHNL2_PROTO_HDR_ANY = 0,
VIRTCHNL2_PROTO_HDR_PRE_MAC = 1,
/* VIRTCHNL2_PROTO_HDR_MAC is a mandatory protocol id */
VIRTCHNL2_PROTO_HDR_MAC = 2,
VIRTCHNL2_PROTO_HDR_POST_MAC = 3,
VIRTCHNL2_PROTO_HDR_ETHERTYPE = 4,
VIRTCHNL2_PROTO_HDR_VLAN = 5,
VIRTCHNL2_PROTO_HDR_SVLAN = 6,
VIRTCHNL2_PROTO_HDR_CVLAN = 7,
VIRTCHNL2_PROTO_HDR_MPLS = 8,
VIRTCHNL2_PROTO_HDR_UMPLS = 9,
VIRTCHNL2_PROTO_HDR_MMPLS = 10,
VIRTCHNL2_PROTO_HDR_PTP = 11,
VIRTCHNL2_PROTO_HDR_CTRL = 12,
VIRTCHNL2_PROTO_HDR_LLDP = 13,
VIRTCHNL2_PROTO_HDR_ARP = 14,
VIRTCHNL2_PROTO_HDR_ECP = 15,
VIRTCHNL2_PROTO_HDR_EAPOL = 16,
VIRTCHNL2_PROTO_HDR_PPPOD = 17,
VIRTCHNL2_PROTO_HDR_PPPOE = 18,
/* VIRTCHNL2_PROTO_HDR_IPV4 is a mandatory protocol id */
VIRTCHNL2_PROTO_HDR_IPV4 = 19,
/* IPv4 and IPv6 Fragment header types are only associated to
* VIRTCHNL2_PROTO_HDR_IPV4 and VIRTCHNL2_PROTO_HDR_IPV6 respectively,
* cannot be used independently.
*/
/* VIRTCHNL2_PROTO_HDR_IPV4_FRAG is a mandatory protocol id */
VIRTCHNL2_PROTO_HDR_IPV4_FRAG = 20,
/* VIRTCHNL2_PROTO_HDR_IPV6 is a mandatory protocol id */
VIRTCHNL2_PROTO_HDR_IPV6 = 21,
/* VIRTCHNL2_PROTO_HDR_IPV6_FRAG is a mandatory protocol id */
VIRTCHNL2_PROTO_HDR_IPV6_FRAG = 22,
VIRTCHNL2_PROTO_HDR_IPV6_EH = 23,
/* VIRTCHNL2_PROTO_HDR_UDP is a mandatory protocol id */
VIRTCHNL2_PROTO_HDR_UDP = 24,
/* VIRTCHNL2_PROTO_HDR_TCP is a mandatory protocol id */
VIRTCHNL2_PROTO_HDR_TCP = 25,
/* VIRTCHNL2_PROTO_HDR_SCTP is a mandatory protocol id */
VIRTCHNL2_PROTO_HDR_SCTP = 26,
/* VIRTCHNL2_PROTO_HDR_ICMP is a mandatory protocol id */
VIRTCHNL2_PROTO_HDR_ICMP = 27,
/* VIRTCHNL2_PROTO_HDR_ICMPV6 is a mandatory protocol id */
VIRTCHNL2_PROTO_HDR_ICMPV6 = 28,
VIRTCHNL2_PROTO_HDR_IGMP = 29,
VIRTCHNL2_PROTO_HDR_AH = 30,
VIRTCHNL2_PROTO_HDR_ESP = 31,
VIRTCHNL2_PROTO_HDR_IKE = 32,
VIRTCHNL2_PROTO_HDR_NATT_KEEP = 33,
/* VIRTCHNL2_PROTO_HDR_PAY is a mandatory protocol id */
VIRTCHNL2_PROTO_HDR_PAY = 34,
VIRTCHNL2_PROTO_HDR_L2TPV2 = 35,
VIRTCHNL2_PROTO_HDR_L2TPV2_CONTROL = 36,
VIRTCHNL2_PROTO_HDR_L2TPV3 = 37,
VIRTCHNL2_PROTO_HDR_GTP = 38,
VIRTCHNL2_PROTO_HDR_GTP_EH = 39,
VIRTCHNL2_PROTO_HDR_GTPCV2 = 40,
VIRTCHNL2_PROTO_HDR_GTPC_TEID = 41,
VIRTCHNL2_PROTO_HDR_GTPU = 42,
VIRTCHNL2_PROTO_HDR_GTPU_UL = 43,
VIRTCHNL2_PROTO_HDR_GTPU_DL = 44,
VIRTCHNL2_PROTO_HDR_ECPRI = 45,
VIRTCHNL2_PROTO_HDR_VRRP = 46,
VIRTCHNL2_PROTO_HDR_OSPF = 47,
/* VIRTCHNL2_PROTO_HDR_TUN is a mandatory protocol id */
VIRTCHNL2_PROTO_HDR_TUN = 48,
VIRTCHNL2_PROTO_HDR_GRE = 49,
VIRTCHNL2_PROTO_HDR_NVGRE = 50,
VIRTCHNL2_PROTO_HDR_VXLAN = 51,
VIRTCHNL2_PROTO_HDR_VXLAN_GPE = 52,
VIRTCHNL2_PROTO_HDR_GENEVE = 53,
VIRTCHNL2_PROTO_HDR_NSH = 54,
VIRTCHNL2_PROTO_HDR_QUIC = 55,
VIRTCHNL2_PROTO_HDR_PFCP = 56,
VIRTCHNL2_PROTO_HDR_PFCP_NODE = 57,
VIRTCHNL2_PROTO_HDR_PFCP_SESSION = 58,
VIRTCHNL2_PROTO_HDR_RTP = 59,
VIRTCHNL2_PROTO_HDR_ROCE = 60,
VIRTCHNL2_PROTO_HDR_ROCEV1 = 61,
VIRTCHNL2_PROTO_HDR_ROCEV2 = 62,
/* Protocol ids up to 32767 are reserved.
* 32768 - 65534 are used for user defined protocol ids.
* VIRTCHNL2_PROTO_HDR_NO_PROTO is a mandatory protocol id.
*/
VIRTCHNL2_PROTO_HDR_NO_PROTO = 65535,
};
enum virtchl2_version {
VIRTCHNL2_VERSION_MINOR_0 = 0,
VIRTCHNL2_VERSION_MAJOR_2 = 2,
};
/**
* struct virtchnl2_edt_caps - Get EDT granularity and time horizon.
* @tstamp_granularity_ns: Timestamp granularity in nanoseconds.
* @time_horizon_ns: Total time window in nanoseconds.
*
* Associated with VIRTCHNL2_OP_GET_EDT_CAPS.
*/
struct virtchnl2_edt_caps {
__le64 tstamp_granularity_ns;
__le64 time_horizon_ns;
};
VIRTCHNL2_CHECK_STRUCT_LEN(16, virtchnl2_edt_caps);
/**
* struct virtchnl2_version_info - Version information.
* @major: Major version.
* @minor: Minor version.
*
* PF/VF posts its version number to the CP. CP responds with its version number
* in the same format, along with a return code.
* If there is a major version mismatch, then the PF/VF cannot operate.
* If there is a minor version mismatch, then the PF/VF can operate but should
* add a warning to the system log.
*
* This version opcode MUST always be specified as == 1, regardless of other
* changes in the API. The CP must always respond to this message without
* error regardless of version mismatch.
*
* Associated with VIRTCHNL2_OP_VERSION.
*/
struct virtchnl2_version_info {
__le32 major;
__le32 minor;
};
VIRTCHNL2_CHECK_STRUCT_LEN(8, virtchnl2_version_info);
/**
* struct virtchnl2_get_capabilities - Capabilities info.
* @csum_caps: See enum virtchnl2_cap_txrx_csum.
* @seg_caps: See enum virtchnl2_cap_seg.
* @hsplit_caps: See enum virtchnl2_cap_rx_hsplit_at.
* @rsc_caps: See enum virtchnl2_cap_rsc.
* @rss_caps: See enum virtchnl2_cap_rss.
* @other_caps: See enum virtchnl2_cap_other.
* @mailbox_dyn_ctl: DYN_CTL register offset and vector id for mailbox
* provided by CP.
* @mailbox_vector_id: Mailbox vector id.
* @num_allocated_vectors: Maximum number of allocated vectors for the device.
* @max_rx_q: Maximum number of supported Rx queues.
* @max_tx_q: Maximum number of supported Tx queues.
* @max_rx_bufq: Maximum number of supported buffer queues.
* @max_tx_complq: Maximum number of supported completion queues.
* @max_sriov_vfs: The PF sends the maximum VFs it is requesting. The CP
* responds with the maximum VFs granted.
* @max_vports: Maximum number of vports that can be supported.
* @default_num_vports: Default number of vports driver should allocate on load.
* @max_tx_hdr_size: Max header length hardware can parse/checksum, in bytes.
* @max_sg_bufs_per_tx_pkt: Max number of scatter gather buffers that can be
* sent per transmit packet without needing to be
* linearized.
* @pad: Padding.
* @reserved: Reserved.
* @device_type: See enum virtchl2_device_type.
* @min_sso_packet_len: Min packet length supported by device for single
* segment offload.
* @max_hdr_buf_per_lso: Max number of header buffers that can be used for
* an LSO.
* @pad1: Padding for future extensions.
*
* Dataplane driver sends this message to CP to negotiate capabilities and
* provides a virtchnl2_get_capabilities structure with its desired
* capabilities, max_sriov_vfs and num_allocated_vectors.
* CP responds with a virtchnl2_get_capabilities structure updated
* with allowed capabilities and the other fields as below.
* If PF sets max_sriov_vfs as 0, CP will respond with max number of VFs
* that can be created by this PF. For any other value 'n', CP responds
* with max_sriov_vfs set to min(n, x) where x is the max number of VFs
* allowed by CP's policy. max_sriov_vfs is not applicable for VFs.
* If dataplane driver sets num_allocated_vectors as 0, CP will respond with 1
* which is default vector associated with the default mailbox. For any other
* value 'n', CP responds with a value <= n based on the CP's policy of
* max number of vectors for a PF.
* CP will respond with the vector ID of mailbox allocated to the PF in
* mailbox_vector_id and the number of itr index registers in itr_idx_map.
* It also responds with default number of vports that the dataplane driver
* should comeup with in default_num_vports and maximum number of vports that
* can be supported in max_vports.
*
* Associated with VIRTCHNL2_OP_GET_CAPS.
*/
struct virtchnl2_get_capabilities {
__le32 csum_caps;
__le32 seg_caps;
__le32 hsplit_caps;
__le32 rsc_caps;
__le64 rss_caps;
__le64 other_caps;
__le32 mailbox_dyn_ctl;
__le16 mailbox_vector_id;
__le16 num_allocated_vectors;
__le16 max_rx_q;
__le16 max_tx_q;
__le16 max_rx_bufq;
__le16 max_tx_complq;
__le16 max_sriov_vfs;
__le16 max_vports;
__le16 default_num_vports;
__le16 max_tx_hdr_size;
u8 max_sg_bufs_per_tx_pkt;
u8 pad[3];
u8 reserved[4];
__le32 device_type;
u8 min_sso_packet_len;
u8 max_hdr_buf_per_lso;
u8 pad1[10];
};
VIRTCHNL2_CHECK_STRUCT_LEN(80, virtchnl2_get_capabilities);
/**
* struct virtchnl2_queue_reg_chunk - Single queue chunk.
* @type: See enum virtchnl2_queue_type.
* @start_queue_id: Start Queue ID.
* @num_queues: Number of queues in the chunk.
* @pad: Padding.
* @qtail_reg_start: Queue tail register offset.
* @qtail_reg_spacing: Queue tail register spacing.
* @pad1: Padding for future extensions.
*/
struct virtchnl2_queue_reg_chunk {
__le32 type;
__le32 start_queue_id;
__le32 num_queues;
__le32 pad;
__le64 qtail_reg_start;
__le32 qtail_reg_spacing;
u8 pad1[4];
};
VIRTCHNL2_CHECK_STRUCT_LEN(32, virtchnl2_queue_reg_chunk);
/**
* struct virtchnl2_queue_reg_chunks - Specify several chunks of contiguous
* queues.
* @num_chunks: Number of chunks.
* @pad: Padding.
* @chunks: Chunks of queue info.
*/
struct virtchnl2_queue_reg_chunks {
__le16 num_chunks;
u8 pad[6];
struct virtchnl2_queue_reg_chunk chunks[] __counted_by_le(num_chunks);
};
VIRTCHNL2_CHECK_STRUCT_LEN(8, virtchnl2_queue_reg_chunks);
/**
* struct virtchnl2_create_vport - Create vport config info.
* @vport_type: See enum virtchnl2_vport_type.
* @txq_model: See virtchnl2_queue_model.
* @rxq_model: See virtchnl2_queue_model.
* @num_tx_q: Number of Tx queues.
* @num_tx_complq: Valid only if txq_model is split queue.
* @num_rx_q: Number of Rx queues.
* @num_rx_bufq: Valid only if rxq_model is split queue.
* @default_rx_q: Relative receive queue index to be used as default.
* @vport_index: Used to align PF and CP in case of default multiple vports,
* it is filled by the PF and CP returns the same value, to
* enable the driver to support multiple asynchronous parallel
* CREATE_VPORT requests and associate a response to a specific
* request.
* @max_mtu: Max MTU. CP populates this field on response.
* @vport_id: Vport id. CP populates this field on response.
* @default_mac_addr: Default MAC address.
* @pad: Padding.
* @rx_desc_ids: See VIRTCHNL2_RX_DESC_IDS definitions.
* @tx_desc_ids: See VIRTCHNL2_TX_DESC_IDS definitions.
* @pad1: Padding.
* @rss_algorithm: RSS algorithm.
* @rss_key_size: RSS key size.
* @rss_lut_size: RSS LUT size.
* @rx_split_pos: See enum virtchnl2_cap_rx_hsplit_at.
* @pad2: Padding.
* @chunks: Chunks of contiguous queues.
*
* PF sends this message to CP to create a vport by filling in required
* fields of virtchnl2_create_vport structure.
* CP responds with the updated virtchnl2_create_vport structure containing the
* necessary fields followed by chunks which in turn will have an array of
* num_chunks entries of virtchnl2_queue_chunk structures.
*
* Associated with VIRTCHNL2_OP_CREATE_VPORT.
*/
struct virtchnl2_create_vport {
__le16 vport_type;
__le16 txq_model;
__le16 rxq_model;
__le16 num_tx_q;
__le16 num_tx_complq;
__le16 num_rx_q;
__le16 num_rx_bufq;
__le16 default_rx_q;
__le16 vport_index;
/* CP populates the following fields on response */
__le16 max_mtu;
__le32 vport_id;
u8 default_mac_addr[ETH_ALEN];
__le16 pad;
__le64 rx_desc_ids;
__le64 tx_desc_ids;
u8 pad1[72];
__le32 rss_algorithm;
__le16 rss_key_size;
__le16 rss_lut_size;
__le32 rx_split_pos;
u8 pad2[20];
struct virtchnl2_queue_reg_chunks chunks;
};
VIRTCHNL2_CHECK_STRUCT_LEN(160, virtchnl2_create_vport);
/**
* struct virtchnl2_vport - Vport ID info.
* @vport_id: Vport id.
* @pad: Padding for future extensions.
*
* PF sends this message to CP to destroy, enable or disable a vport by filling
* in the vport_id in virtchnl2_vport structure.
* CP responds with the status of the requested operation.
*
* Associated with VIRTCHNL2_OP_DESTROY_VPORT, VIRTCHNL2_OP_ENABLE_VPORT,
* VIRTCHNL2_OP_DISABLE_VPORT.
*/
struct virtchnl2_vport {
__le32 vport_id;
u8 pad[4];
};
VIRTCHNL2_CHECK_STRUCT_LEN(8, virtchnl2_vport);
/**
* struct virtchnl2_txq_info - Transmit queue config info
* @dma_ring_addr: DMA address.
* @type: See enum virtchnl2_queue_type.
* @queue_id: Queue ID.
* @relative_queue_id: Valid only if queue model is split and type is transmit
* queue. Used in many to one mapping of transmit queues to
* completion queue.
* @model: See enum virtchnl2_queue_model.
* @sched_mode: See enum virtchnl2_txq_sched_mode.
* @qflags: TX queue feature flags.
* @ring_len: Ring length.
* @tx_compl_queue_id: Valid only if queue model is split and type is transmit
* queue.
* @peer_type: Valid only if queue type is VIRTCHNL2_QUEUE_TYPE_MAILBOX_TX
* @peer_rx_queue_id: Valid only if queue type is CONFIG_TX and used to deliver
* messages for the respective CONFIG_TX queue.
* @pad: Padding.
* @egress_pasid: Egress PASID info.
* @egress_hdr_pasid: Egress HDR passid.
* @egress_buf_pasid: Egress buf passid.
* @pad1: Padding for future extensions.
*/
struct virtchnl2_txq_info {
__le64 dma_ring_addr;
__le32 type;
__le32 queue_id;
__le16 relative_queue_id;
__le16 model;
__le16 sched_mode;
__le16 qflags;
__le16 ring_len;
__le16 tx_compl_queue_id;
__le16 peer_type;
__le16 peer_rx_queue_id;
u8 pad[4];
__le32 egress_pasid;
__le32 egress_hdr_pasid;
__le32 egress_buf_pasid;
u8 pad1[8];
};
VIRTCHNL2_CHECK_STRUCT_LEN(56, virtchnl2_txq_info);
/**
* struct virtchnl2_config_tx_queues - TX queue config.
* @vport_id: Vport id.
* @num_qinfo: Number of virtchnl2_txq_info structs.
* @pad: Padding.
* @qinfo: Tx queues config info.
*
* PF sends this message to set up parameters for one or more transmit queues.
* This message contains an array of num_qinfo instances of virtchnl2_txq_info
* structures. CP configures requested queues and returns a status code. If
* num_qinfo specified is greater than the number of queues associated with the
* vport, an error is returned and no queues are configured.
*
* Associated with VIRTCHNL2_OP_CONFIG_TX_QUEUES.
*/
struct virtchnl2_config_tx_queues {
__le32 vport_id;
__le16 num_qinfo;
u8 pad[10];
struct virtchnl2_txq_info qinfo[] __counted_by_le(num_qinfo);
};
VIRTCHNL2_CHECK_STRUCT_LEN(16, virtchnl2_config_tx_queues);
/**
* struct virtchnl2_rxq_info - Receive queue config info.
* @desc_ids: See VIRTCHNL2_RX_DESC_IDS definitions.
* @dma_ring_addr: See VIRTCHNL2_RX_DESC_IDS definitions.
* @type: See enum virtchnl2_queue_type.
* @queue_id: Queue id.
* @model: See enum virtchnl2_queue_model.
* @hdr_buffer_size: Header buffer size.
* @data_buffer_size: Data buffer size.
* @max_pkt_size: Max packet size.
* @ring_len: Ring length.
* @buffer_notif_stride: Buffer notification stride in units of 32-descriptors.
* This field must be a power of 2.
* @pad: Padding.
* @dma_head_wb_addr: Applicable only for receive buffer queues.
* @qflags: Applicable only for receive completion queues.
* See enum virtchnl2_rxq_flags.
* @rx_buffer_low_watermark: Rx buffer low watermark.
* @rx_bufq1_id: Buffer queue index of the first buffer queue associated with
* the Rx queue. Valid only in split queue model.
* @rx_bufq2_id: Buffer queue index of the second buffer queue associated with
* the Rx queue. Valid only in split queue model.
* @bufq2_ena: It indicates if there is a second buffer, rx_bufq2_id is valid
* only if this field is set.
* @pad1: Padding.
* @ingress_pasid: Ingress PASID.
* @ingress_hdr_pasid: Ingress PASID header.
* @ingress_buf_pasid: Ingress PASID buffer.
* @pad2: Padding for future extensions.
*/
struct virtchnl2_rxq_info {
__le64 desc_ids;
__le64 dma_ring_addr;
__le32 type;
__le32 queue_id;
__le16 model;
__le16 hdr_buffer_size;
__le32 data_buffer_size;
__le32 max_pkt_size;
__le16 ring_len;
u8 buffer_notif_stride;
u8 pad;
__le64 dma_head_wb_addr;
__le16 qflags;
__le16 rx_buffer_low_watermark;
__le16 rx_bufq1_id;
__le16 rx_bufq2_id;
u8 bufq2_ena;
u8 pad1[3];
__le32 ingress_pasid;
__le32 ingress_hdr_pasid;
__le32 ingress_buf_pasid;
u8 pad2[16];
};
VIRTCHNL2_CHECK_STRUCT_LEN(88, virtchnl2_rxq_info);
/**
* struct virtchnl2_config_rx_queues - Rx queues config.
* @vport_id: Vport id.
* @num_qinfo: Number of instances.
* @pad: Padding.
* @qinfo: Rx queues config info.
*
* PF sends this message to set up parameters for one or more receive queues.
* This message contains an array of num_qinfo instances of virtchnl2_rxq_info
* structures. CP configures requested queues and returns a status code.
* If the number of queues specified is greater than the number of queues
* associated with the vport, an error is returned and no queues are configured.
*
* Associated with VIRTCHNL2_OP_CONFIG_RX_QUEUES.
*/
struct virtchnl2_config_rx_queues {
__le32 vport_id;
__le16 num_qinfo;
u8 pad[18];
struct virtchnl2_rxq_info qinfo[] __counted_by_le(num_qinfo);
};
VIRTCHNL2_CHECK_STRUCT_LEN(24, virtchnl2_config_rx_queues);
/**
* struct virtchnl2_add_queues - data for VIRTCHNL2_OP_ADD_QUEUES.
* @vport_id: Vport id.
* @num_tx_q: Number of Tx qieues.
* @num_tx_complq: Number of Tx completion queues.
* @num_rx_q: Number of Rx queues.
* @num_rx_bufq: Number of Rx buffer queues.
* @pad: Padding.
* @chunks: Chunks of contiguous queues.
*
* PF sends this message to request additional transmit/receive queues beyond
* the ones that were assigned via CREATE_VPORT request. virtchnl2_add_queues
* structure is used to specify the number of each type of queues.
* CP responds with the same structure with the actual number of queues assigned
* followed by num_chunks of virtchnl2_queue_chunk structures.
*
* Associated with VIRTCHNL2_OP_ADD_QUEUES.
*/
struct virtchnl2_add_queues {
__le32 vport_id;
__le16 num_tx_q;
__le16 num_tx_complq;
__le16 num_rx_q;
__le16 num_rx_bufq;
u8 pad[4];
struct virtchnl2_queue_reg_chunks chunks;
};
VIRTCHNL2_CHECK_STRUCT_LEN(24, virtchnl2_add_queues);
/**
* struct virtchnl2_vector_chunk - Structure to specify a chunk of contiguous
* interrupt vectors.
* @start_vector_id: Start vector id.
* @start_evv_id: Start EVV id.
* @num_vectors: Number of vectors.
* @pad: Padding.
* @dynctl_reg_start: DYN_CTL register offset.
* @dynctl_reg_spacing: register spacing between DYN_CTL registers of 2
* consecutive vectors.
* @itrn_reg_start: ITRN register offset.
* @itrn_reg_spacing: Register spacing between dynctl registers of 2
* consecutive vectors.
* @itrn_index_spacing: Register spacing between itrn registers of the same
* vector where n=0..2.
* @pad1: Padding for future extensions.
*
* Register offsets and spacing provided by CP.
* Dynamic control registers are used for enabling/disabling/re-enabling
* interrupts and updating interrupt rates in the hotpath. Any changes
* to interrupt rates in the dynamic control registers will be reflected
* in the interrupt throttling rate registers.
* itrn registers are used to update interrupt rates for specific
* interrupt indices without modifying the state of the interrupt.
*/
struct virtchnl2_vector_chunk {
__le16 start_vector_id;
__le16 start_evv_id;
__le16 num_vectors;
__le16 pad;
__le32 dynctl_reg_start;
__le32 dynctl_reg_spacing;
__le32 itrn_reg_start;
__le32 itrn_reg_spacing;
__le32 itrn_index_spacing;
u8 pad1[4];
};
VIRTCHNL2_CHECK_STRUCT_LEN(32, virtchnl2_vector_chunk);
/**
* struct virtchnl2_vector_chunks - chunks of contiguous interrupt vectors.
* @num_vchunks: number of vector chunks.
* @pad: Padding.
* @vchunks: Chunks of contiguous vector info.
*
* PF sends virtchnl2_vector_chunks struct to specify the vectors it is giving
* away. CP performs requested action and returns status.
*
* Associated with VIRTCHNL2_OP_DEALLOC_VECTORS.
*/
struct virtchnl2_vector_chunks {
__le16 num_vchunks;
u8 pad[14];
struct virtchnl2_vector_chunk vchunks[] __counted_by_le(num_vchunks);
};
VIRTCHNL2_CHECK_STRUCT_LEN(16, virtchnl2_vector_chunks);
/**
* struct virtchnl2_alloc_vectors - vector allocation info.
* @num_vectors: Number of vectors.
* @pad: Padding.
* @vchunks: Chunks of contiguous vector info.
*
* PF sends this message to request additional interrupt vectors beyond the
* ones that were assigned via GET_CAPS request. virtchnl2_alloc_vectors
* structure is used to specify the number of vectors requested. CP responds
* with the same structure with the actual number of vectors assigned followed
* by virtchnl2_vector_chunks structure identifying the vector ids.
*
* Associated with VIRTCHNL2_OP_ALLOC_VECTORS.
*/
struct virtchnl2_alloc_vectors {
__le16 num_vectors;
u8 pad[14];
struct virtchnl2_vector_chunks vchunks;
};
VIRTCHNL2_CHECK_STRUCT_LEN(32, virtchnl2_alloc_vectors);
/**
* struct virtchnl2_rss_lut - RSS LUT info.
* @vport_id: Vport id.
* @lut_entries_start: Start of LUT entries.
* @lut_entries: Number of LUT entrties.
* @pad: Padding.
* @lut: RSS lookup table.
*
* PF sends this message to get or set RSS lookup table. Only supported if
* both PF and CP drivers set the VIRTCHNL2_CAP_RSS bit during configuration
* negotiation.
*
* Associated with VIRTCHNL2_OP_GET_RSS_LUT and VIRTCHNL2_OP_SET_RSS_LUT.
*/
struct virtchnl2_rss_lut {
__le32 vport_id;
__le16 lut_entries_start;
__le16 lut_entries;
u8 pad[4];
__le32 lut[] __counted_by_le(lut_entries);
};
VIRTCHNL2_CHECK_STRUCT_LEN(12, virtchnl2_rss_lut);
/**
* struct virtchnl2_rss_hash - RSS hash info.
* @ptype_groups: Packet type groups bitmap.
* @vport_id: Vport id.
* @pad: Padding for future extensions.
*
* PF sends these messages to get and set the hash filter enable bits for RSS.
* By default, the CP sets these to all possible traffic types that the
* hardware supports. The PF can query this value if it wants to change the
* traffic types that are hashed by the hardware.
* Only supported if both PF and CP drivers set the VIRTCHNL2_CAP_RSS bit
* during configuration negotiation.
*
* Associated with VIRTCHNL2_OP_GET_RSS_HASH and VIRTCHNL2_OP_SET_RSS_HASH
*/
struct virtchnl2_rss_hash {
__le64 ptype_groups;
__le32 vport_id;
u8 pad[4];
};
VIRTCHNL2_CHECK_STRUCT_LEN(16, virtchnl2_rss_hash);
/**
* struct virtchnl2_sriov_vfs_info - VFs info.
* @num_vfs: Number of VFs.
* @pad: Padding for future extensions.
*
* This message is used to set number of SRIOV VFs to be created. The actual
* allocation of resources for the VFs in terms of vport, queues and interrupts
* is done by CP. When this call completes, the IDPF driver calls
* pci_enable_sriov to let the OS instantiate the SRIOV PCIE devices.
* The number of VFs set to 0 will destroy all the VFs of this function.
*
* Associated with VIRTCHNL2_OP_SET_SRIOV_VFS.
*/
struct virtchnl2_sriov_vfs_info {
__le16 num_vfs;
__le16 pad;
};
VIRTCHNL2_CHECK_STRUCT_LEN(4, virtchnl2_sriov_vfs_info);
/**
* struct virtchnl2_ptype - Packet type info.
* @ptype_id_10: 10-bit packet type.
* @ptype_id_8: 8-bit packet type.
* @proto_id_count: Number of protocol ids the packet supports, maximum of 32
* protocol ids are supported.
* @pad: Padding.
* @proto_id: proto_id_count decides the allocation of protocol id array.
* See enum virtchnl2_proto_hdr_type.
*
* Based on the descriptor type the PF supports, CP fills ptype_id_10 or
* ptype_id_8 for flex and base descriptor respectively. If ptype_id_10 value
* is set to 0xFFFF, PF should consider this ptype as dummy one and it is the
* last ptype.
*/
struct virtchnl2_ptype {
__le16 ptype_id_10;
u8 ptype_id_8;
u8 proto_id_count;
__le16 pad;
__le16 proto_id[] __counted_by(proto_id_count);
} __packed __aligned(2);
VIRTCHNL2_CHECK_STRUCT_LEN(6, virtchnl2_ptype);
/**
* struct virtchnl2_get_ptype_info - Packet type info.
* @start_ptype_id: Starting ptype ID.
* @num_ptypes: Number of packet types from start_ptype_id.
* @pad: Padding for future extensions.
*
* The total number of supported packet types is based on the descriptor type.
* For the flex descriptor, it is 1024 (10-bit ptype), and for the base
* descriptor, it is 256 (8-bit ptype). Send this message to the CP by
* populating the 'start_ptype_id' and the 'num_ptypes'. CP responds with the
* 'start_ptype_id', 'num_ptypes', and the array of ptype (virtchnl2_ptype) that
* are added at the end of the 'virtchnl2_get_ptype_info' message (Note: There
* is no specific field for the ptypes but are added at the end of the
* ptype info message. PF/VF is expected to extract the ptypes accordingly.
* Reason for doing this is because compiler doesn't allow nested flexible
* array fields).
*
* If all the ptypes don't fit into one mailbox buffer, CP splits the
* ptype info into multiple messages, where each message will have its own
* 'start_ptype_id', 'num_ptypes', and the ptype array itself. When CP is done
* updating all the ptype information extracted from the package (the number of
* ptypes extracted might be less than what PF/VF expects), it will append a
* dummy ptype (which has 'ptype_id_10' of 'struct virtchnl2_ptype' as 0xFFFF)
* to the ptype array.
*
* PF/VF is expected to receive multiple VIRTCHNL2_OP_GET_PTYPE_INFO messages.
*
* Associated with VIRTCHNL2_OP_GET_PTYPE_INFO.
*/
struct virtchnl2_get_ptype_info {
__le16 start_ptype_id;
__le16 num_ptypes;
__le32 pad;
};
VIRTCHNL2_CHECK_STRUCT_LEN(8, virtchnl2_get_ptype_info);
/**
* struct virtchnl2_vport_stats - Vport statistics.
* @vport_id: Vport id.
* @pad: Padding.
* @rx_bytes: Received bytes.
* @rx_unicast: Received unicast packets.
* @rx_multicast: Received multicast packets.
* @rx_broadcast: Received broadcast packets.
* @rx_discards: Discarded packets on receive.
* @rx_errors: Receive errors.
* @rx_unknown_protocol: Unlnown protocol.
* @tx_bytes: Transmitted bytes.
* @tx_unicast: Transmitted unicast packets.
* @tx_multicast: Transmitted multicast packets.
* @tx_broadcast: Transmitted broadcast packets.
* @tx_discards: Discarded packets on transmit.
* @tx_errors: Transmit errors.
* @rx_invalid_frame_length: Packets with invalid frame length.
* @rx_overflow_drop: Packets dropped on buffer overflow.
*
* PF/VF sends this message to CP to get the update stats by specifying the
* vport_id. CP responds with stats in struct virtchnl2_vport_stats.
*
* Associated with VIRTCHNL2_OP_GET_STATS.
*/
struct virtchnl2_vport_stats {
__le32 vport_id;
u8 pad[4];
__le64 rx_bytes;
__le64 rx_unicast;
__le64 rx_multicast;
__le64 rx_broadcast;
__le64 rx_discards;
__le64 rx_errors;
__le64 rx_unknown_protocol;
__le64 tx_bytes;
__le64 tx_unicast;
__le64 tx_multicast;
__le64 tx_broadcast;
__le64 tx_discards;
__le64 tx_errors;
__le64 rx_invalid_frame_length;
__le64 rx_overflow_drop;
};
VIRTCHNL2_CHECK_STRUCT_LEN(128, virtchnl2_vport_stats);
/**
* struct virtchnl2_event - Event info.
* @event: Event opcode. See enum virtchnl2_event_codes.
* @link_speed: Link_speed provided in Mbps.
* @vport_id: Vport ID.
* @link_status: Link status.
* @pad: Padding.
* @reserved: Reserved.
*
* CP sends this message to inform the PF/VF driver of events that may affect
* it. No direct response is expected from the driver, though it may generate
* other messages in response to this one.
*
* Associated with VIRTCHNL2_OP_EVENT.
*/
struct virtchnl2_event {
__le32 event;
__le32 link_speed;
__le32 vport_id;
u8 link_status;
u8 pad;
__le16 reserved;
};
VIRTCHNL2_CHECK_STRUCT_LEN(16, virtchnl2_event);
/**
* struct virtchnl2_rss_key - RSS key info.
* @vport_id: Vport id.
* @key_len: Length of RSS key.
* @pad: Padding.
* @key_flex: RSS hash key, packed bytes.
* PF/VF sends this message to get or set RSS key. Only supported if both
* PF/VF and CP drivers set the VIRTCHNL2_CAP_RSS bit during configuration
* negotiation.
*
* Associated with VIRTCHNL2_OP_GET_RSS_KEY and VIRTCHNL2_OP_SET_RSS_KEY.
*/
struct virtchnl2_rss_key {
__le32 vport_id;
__le16 key_len;
u8 pad;
u8 key_flex[] __counted_by_le(key_len);
} __packed;
VIRTCHNL2_CHECK_STRUCT_LEN(7, virtchnl2_rss_key);
/**
* struct virtchnl2_queue_chunk - chunk of contiguous queues
* @type: See enum virtchnl2_queue_type.
* @start_queue_id: Starting queue id.
* @num_queues: Number of queues.
* @pad: Padding for future extensions.
*/
struct virtchnl2_queue_chunk {
__le32 type;
__le32 start_queue_id;
__le32 num_queues;
u8 pad[4];
};
VIRTCHNL2_CHECK_STRUCT_LEN(16, virtchnl2_queue_chunk);
/* struct virtchnl2_queue_chunks - chunks of contiguous queues
* @num_chunks: Number of chunks.
* @pad: Padding.
* @chunks: Chunks of contiguous queues info.
*/
struct virtchnl2_queue_chunks {
__le16 num_chunks;
u8 pad[6];
struct virtchnl2_queue_chunk chunks[] __counted_by_le(num_chunks);
};
VIRTCHNL2_CHECK_STRUCT_LEN(8, virtchnl2_queue_chunks);
/**
* struct virtchnl2_del_ena_dis_queues - Enable/disable queues info.
* @vport_id: Vport id.
* @pad: Padding.
* @chunks: Chunks of contiguous queues info.
*
* PF sends these messages to enable, disable or delete queues specified in
* chunks. PF sends virtchnl2_del_ena_dis_queues struct to specify the queues
* to be enabled/disabled/deleted. Also applicable to single queue receive or
* transmit. CP performs requested action and returns status.
*
* Associated with VIRTCHNL2_OP_ENABLE_QUEUES, VIRTCHNL2_OP_DISABLE_QUEUES and
* VIRTCHNL2_OP_DISABLE_QUEUES.
*/
struct virtchnl2_del_ena_dis_queues {
__le32 vport_id;
u8 pad[4];
struct virtchnl2_queue_chunks chunks;
};
VIRTCHNL2_CHECK_STRUCT_LEN(16, virtchnl2_del_ena_dis_queues);
/**
* struct virtchnl2_queue_vector - Queue to vector mapping.
* @queue_id: Queue id.
* @vector_id: Vector id.
* @pad: Padding.
* @itr_idx: See enum virtchnl2_itr_idx.
* @queue_type: See enum virtchnl2_queue_type.
* @pad1: Padding for future extensions.
*/
struct virtchnl2_queue_vector {
__le32 queue_id;
__le16 vector_id;
u8 pad[2];
__le32 itr_idx;
__le32 queue_type;
u8 pad1[8];
};
VIRTCHNL2_CHECK_STRUCT_LEN(24, virtchnl2_queue_vector);
/**
* struct virtchnl2_queue_vector_maps - Map/unmap queues info.
* @vport_id: Vport id.
* @num_qv_maps: Number of queue vector maps.
* @pad: Padding.
* @qv_maps: Queue to vector maps.
*
* PF sends this message to map or unmap queues to vectors and interrupt
* throttling rate index registers. External data buffer contains
* virtchnl2_queue_vector_maps structure that contains num_qv_maps of
* virtchnl2_queue_vector structures. CP maps the requested queue vector maps
* after validating the queue and vector ids and returns a status code.
*
* Associated with VIRTCHNL2_OP_MAP_QUEUE_VECTOR and
* VIRTCHNL2_OP_UNMAP_QUEUE_VECTOR.
*/
struct virtchnl2_queue_vector_maps {
__le32 vport_id;
__le16 num_qv_maps;
u8 pad[10];
struct virtchnl2_queue_vector qv_maps[] __counted_by_le(num_qv_maps);
};
VIRTCHNL2_CHECK_STRUCT_LEN(16, virtchnl2_queue_vector_maps);
/**
* struct virtchnl2_loopback - Loopback info.
* @vport_id: Vport id.
* @enable: Enable/disable.
* @pad: Padding for future extensions.
*
* PF/VF sends this message to transition to/from the loopback state. Setting
* the 'enable' to 1 enables the loopback state and setting 'enable' to 0
* disables it. CP configures the state to loopback and returns status.
*
* Associated with VIRTCHNL2_OP_LOOPBACK.
*/
struct virtchnl2_loopback {
__le32 vport_id;
u8 enable;
u8 pad[3];
};
VIRTCHNL2_CHECK_STRUCT_LEN(8, virtchnl2_loopback);
/* struct virtchnl2_mac_addr - MAC address info.
* @addr: MAC address.
* @type: MAC type. See enum virtchnl2_mac_addr_type.
* @pad: Padding for future extensions.
*/
struct virtchnl2_mac_addr {
u8 addr[ETH_ALEN];
u8 type;
u8 pad;
};
VIRTCHNL2_CHECK_STRUCT_LEN(8, virtchnl2_mac_addr);
/**
* struct virtchnl2_mac_addr_list - List of MAC addresses.
* @vport_id: Vport id.
* @num_mac_addr: Number of MAC addresses.
* @pad: Padding.
* @mac_addr_list: List with MAC address info.
*
* PF/VF driver uses this structure to send list of MAC addresses to be
* added/deleted to the CP where as CP performs the action and returns the
* status.
*
* Associated with VIRTCHNL2_OP_ADD_MAC_ADDR and VIRTCHNL2_OP_DEL_MAC_ADDR.
*/
struct virtchnl2_mac_addr_list {
__le32 vport_id;
__le16 num_mac_addr;
u8 pad[2];
struct virtchnl2_mac_addr mac_addr_list[] __counted_by_le(num_mac_addr);
};
VIRTCHNL2_CHECK_STRUCT_LEN(8, virtchnl2_mac_addr_list);
/**
* struct virtchnl2_promisc_info - Promisc type info.
* @vport_id: Vport id.
* @flags: See enum virtchnl2_promisc_flags.
* @pad: Padding for future extensions.
*
* PF/VF sends vport id and flags to the CP where as CP performs the action
* and returns the status.
*
* Associated with VIRTCHNL2_OP_CONFIG_PROMISCUOUS_MODE.
*/
struct virtchnl2_promisc_info {
__le32 vport_id;
/* See VIRTCHNL2_PROMISC_FLAGS definitions */
__le16 flags;
u8 pad[2];
};
VIRTCHNL2_CHECK_STRUCT_LEN(8, virtchnl2_promisc_info);
#endif /* _VIRTCHNL_2_H_ */