| /******************************************************************* |
| * Copyright © 1997-2007 Alacritech, Inc. All rights reserved |
| * |
| * $Id: sxghif.h,v 1.5 2008/07/24 19:18:22 chris Exp $ |
| * |
| * sxghif.h: |
| * |
| * This file contains structures and definitions for the |
| * Alacritech Sahara host interface |
| ******************************************************************/ |
| |
| #define DBG 1 |
| |
| /* UCODE Registers */ |
| struct sxg_ucode_regs { |
| /* Address 0 - 0x3F = Command codes 0-15 for TCB 0. Excode 0 */ |
| u32 Icr; /* Code = 0 (extended), ExCode = 0 - Int control */ |
| u32 RsvdReg1; /* Code = 1 - TOE -NA */ |
| u32 RsvdReg2; /* Code = 2 - TOE -NA */ |
| u32 RsvdReg3; /* Code = 3 - TOE -NA */ |
| u32 RsvdReg4; /* Code = 4 - TOE -NA */ |
| u32 RsvdReg5; /* Code = 5 - TOE -NA */ |
| u32 CardUp; /* Code = 6 - Microcode initialized when 1 */ |
| u32 RsvdReg7; /* Code = 7 - TOE -NA */ |
| u32 ConfigStat; /* Code = 8 - Configuration data load status */ |
| u32 RsvdReg9; /* Code = 9 - TOE -NA */ |
| u32 CodeNotUsed[6]; /* Codes 10-15 not used. ExCode = 0 */ |
| /* This brings us to ExCode 1 at address 0x40 = Interrupt status pointer */ |
| u32 Isp; /* Code = 0 (extended), ExCode = 1 */ |
| u32 PadEx1[15]; /* Codes 1-15 not used with extended codes */ |
| /* ExCode 2 = Interrupt Status Register */ |
| u32 Isr; /* Code = 0 (extended), ExCode = 2 */ |
| u32 PadEx2[15]; |
| /* ExCode 3 = Event base register. Location of event rings */ |
| u32 EventBase; /* Code = 0 (extended), ExCode = 3 */ |
| u32 PadEx3[15]; |
| /* ExCode 4 = Event ring size */ |
| u32 EventSize; /* Code = 0 (extended), ExCode = 4 */ |
| u32 PadEx4[15]; |
| /* ExCode 5 = TCB Buffers base address */ |
| u32 TcbBase; /* Code = 0 (extended), ExCode = 5 */ |
| u32 PadEx5[15]; |
| /* ExCode 6 = TCB Composite Buffers base address */ |
| u32 TcbCompBase; /* Code = 0 (extended), ExCode = 6 */ |
| u32 PadEx6[15]; |
| /* ExCode 7 = Transmit ring base address */ |
| u32 XmtBase; /* Code = 0 (extended), ExCode = 7 */ |
| u32 PadEx7[15]; |
| /* ExCode 8 = Transmit ring size */ |
| u32 XmtSize; /* Code = 0 (extended), ExCode = 8 */ |
| u32 PadEx8[15]; |
| /* ExCode 9 = Receive ring base address */ |
| u32 RcvBase; /* Code = 0 (extended), ExCode = 9 */ |
| u32 PadEx9[15]; |
| /* ExCode 10 = Receive ring size */ |
| u32 RcvSize; /* Code = 0 (extended), ExCode = 10 */ |
| u32 PadEx10[15]; |
| /* ExCode 11 = Read EEPROM/Flash Config */ |
| u32 Config; /* Code = 0 (extended), ExCode = 11 */ |
| u32 PadEx11[15]; |
| /* ExCode 12 = Multicast bits 31:0 */ |
| u32 McastLow; /* Code = 0 (extended), ExCode = 12 */ |
| u32 PadEx12[15]; |
| /* ExCode 13 = Multicast bits 63:32 */ |
| u32 McastHigh; /* Code = 0 (extended), ExCode = 13 */ |
| u32 PadEx13[15]; |
| /* ExCode 14 = Ping */ |
| u32 Ping; /* Code = 0 (extended), ExCode = 14 */ |
| u32 PadEx14[15]; |
| /* ExCode 15 = Link MTU */ |
| u32 LinkMtu; /* Code = 0 (extended), ExCode = 15 */ |
| u32 PadEx15[15]; |
| /* ExCode 16 = Download synchronization */ |
| u32 LoadSync; /* Code = 0 (extended), ExCode = 16 */ |
| u32 PadEx16[15]; |
| /* ExCode 17 = Upper DRAM address bits on 32-bit systems */ |
| u32 Upper; /* Code = 0 (extended), ExCode = 17 */ |
| u32 PadEx17[15]; |
| /* ExCode 18 = Slowpath Send Index Address */ |
| u32 SPSendIndex; /* Code = 0 (extended), ExCode = 18 */ |
| u32 PadEx18[15]; |
| /* ExCode 19 = Get ucode statistics */ |
| u32 GetUcodeStats; /* Code = 0 (extended), ExCode = 19 */ |
| u32 PadEx19[15]; |
| /* ExCode 20 = Aggregation - See sxgmisc.c:SxgSetInterruptAggregation */ |
| u32 Aggregation; /* Code = 0 (extended), ExCode = 20 */ |
| u32 PadEx20[15]; |
| /* ExCode 21 = Receive MDL push timer */ |
| u32 PushTicks; /* Code = 0 (extended), ExCode = 21 */ |
| u32 PadEx21[15]; |
| /* ExCode 22 = ACK Frequency */ |
| u32 AckFrequency; /* Code = 0 (extended), ExCode = 22 */ |
| u32 PadEx22[15]; |
| /* ExCode 23 = TOE NA */ |
| u32 RsvdReg23; |
| u32 PadEx23[15]; |
| /* ExCode 24 = TOE NA */ |
| u32 RsvdReg24; |
| u32 PadEx24[15]; |
| /* ExCode 25 = TOE NA */ |
| u32 RsvdReg25; /* Code = 0 (extended), ExCode = 25 */ |
| u32 PadEx25[15]; |
| /* ExCode 26 = Receive checksum requirements */ |
| u32 ReceiveChecksum; /* Code = 0 (extended), ExCode = 26 */ |
| u32 PadEx26[15]; |
| /* ExCode 27 = RSS Requirements */ |
| u32 Rss; /* Code = 0 (extended), ExCode = 27 */ |
| u32 PadEx27[15]; |
| /* ExCode 28 = RSS Table */ |
| u32 RssTable; /* Code = 0 (extended), ExCode = 28 */ |
| u32 PadEx28[15]; |
| /* ExCode 29 = Event ring release entries */ |
| u32 EventRelease; /* Code = 0 (extended), ExCode = 29 */ |
| u32 PadEx29[15]; |
| /* ExCode 30 = Number of receive bufferlist commands on ring 0 */ |
| u32 RcvCmd; /* Code = 0 (extended), ExCode = 30 */ |
| u32 PadEx30[15]; |
| /* ExCode 31 = slowpath transmit command - Data[31:0] = 1 */ |
| u32 XmtCmd; /* Code = 0 (extended), ExCode = 31 */ |
| u32 PadEx31[15]; |
| /* ExCode 32 = Dump command */ |
| u32 DumpCmd; /* Code = 0 (extended), ExCode = 32 */ |
| u32 PadEx32[15]; |
| /* ExCode 33 = Debug command */ |
| u32 DebugCmd; /* Code = 0 (extended), ExCode = 33 */ |
| u32 PadEx33[15]; |
| /* |
| * There are 128 possible extended commands - each of account for 16 |
| * words (including the non-relevent base command codes 1-15). |
| * Pad for the remainder of these here to bring us to the next CPU |
| * base. As extended codes are added, reduce the first array value in |
| * the following field |
| */ |
| u32 PadToNextCpu[94][16]; /* 94 = 128 - 34 (34 = Excodes 0 - 33)*/ |
| }; |
| |
| /* Interrupt control register (0) values */ |
| #define SXG_ICR_DISABLE 0x00000000 |
| #define SXG_ICR_ENABLE 0x00000001 |
| #define SXG_ICR_MASK 0x00000002 |
| #define SXG_ICR_MSGID_MASK 0xFFFF0000 |
| #define SXG_ICR_MSGID_SHIFT 16 |
| #define SXG_ICR(_MessageId, _Data) \ |
| ((((_MessageId) << SXG_ICR_MSGID_SHIFT) & \ |
| SXG_ICR_MSGID_MASK) | (_Data)) |
| |
| #define SXG_MIN_AGG_DEFAULT 0x0010 /* Minimum aggregation default */ |
| #define SXG_MAX_AGG_DEFAULT 0x0040 /* Maximum aggregation default */ |
| #define SXG_MAX_AGG_SHIFT 16 /* Maximum in top 16 bits of register */ |
| /* Disable interrupt aggregation on xmt */ |
| #define SXG_AGG_XMT_DISABLE 0x80000000 |
| |
| /* The Microcode supports up to 16 RSS queues (RevB) */ |
| #define SXG_MAX_RSS 16 |
| #define SXG_MAX_RSS_REVA 8 |
| |
| #define SXG_MAX_RSS_TABLE_SIZE 256 /* 256-byte max */ |
| |
| #define SXG_RSS_REVA_TCP6 0x00000001 /* RSS TCP over IPv6 */ |
| #define SXG_RSS_REVA_TCP4 0x00000002 /* RSS TCP over IPv4 */ |
| #define SXG_RSS_IP 0x00000001 /* RSS TCP over IPv6 */ |
| #define SXG_RSS_TCP 0x00000002 /* RSS TCP over IPv4 */ |
| #define SXG_RSS_LEGACY 0x00000004 /* Line-base interrupts */ |
| #define SXG_RSS_TABLE_SIZE 0x0000FF00 /* Table size mask */ |
| |
| #define SXG_RSS_TABLE_SHIFT 8 |
| #define SXG_RSS_BASE_CPU 0x00FF0000 /* Base CPU (not used) */ |
| #define SXG_RSS_BASE_SHIFT 16 |
| |
| #define SXG_RCV_IP_CSUM_ENABLED 0x00000001 /* ExCode 26 (ReceiveChecksum) */ |
| #define SXG_RCV_TCP_CSUM_ENABLED 0x00000002 /* ExCode 26 (ReceiveChecksum) */ |
| |
| #define SXG_XMT_CPUID_SHIFT 16 |
| |
| /* |
| * Status returned by ucode in the ConfigStat reg (see above) when attempted |
| * to load configuration data from the EEPROM/Flash. |
| */ |
| #define SXG_CFG_TIMEOUT 1 /* init value - timeout if unchanged */ |
| #define SXG_CFG_LOAD_EEPROM 2 /* config data loaded from EEPROM */ |
| #define SXG_CFG_LOAD_FLASH 3 /* config data loaded from flash */ |
| #define SXG_CFG_LOAD_INVALID 4 /* no valid config data found */ |
| #define SXG_CFG_LOAD_ERROR 5 /* hardware error */ |
| |
| #define SXG_CHECK_FOR_HANG_TIME 5 |
| |
| /* |
| * TCB registers - This is really the same register memory area as UCODE_REGS |
| * above, but defined differently. Bits 17:06 of the address define the TCB, |
| * which means each TCB area occupies 0x40 (64) bytes, or 16 u32S. What really |
| * is happening is that these registers occupy the "PadEx[15]" areas in the |
| * struct sxg_ucode_regs definition above |
| */ |
| struct sxg_tcb_regs { |
| u32 ExCode; /* Extended codes - see SXG_UCODE_REGS */ |
| u32 Xmt; /* Code = 1 - # of Xmt descriptors added to ring */ |
| u32 Rcv; /* Code = 2 - # of Rcv descriptors added to ring */ |
| u32 Rsvd1; /* Code = 3 - TOE NA */ |
| u32 Rsvd2; /* Code = 4 - TOE NA */ |
| u32 Rsvd3; /* Code = 5 - TOE NA */ |
| u32 Invalid1; /* Code = 6 - Reserved for "CardUp" see above */ |
| u32 Rsvd4; /* Code = 7 - TOE NA */ |
| u32 Invalid2; /* Code = 8 - Reserved for "ConfigStat" see above */ |
| u32 Rsvd5; /* Code = 9 - TOE NA */ |
| u32 Pad[6]; /* Codes 10-15 - Not used. */ |
| }; |
| |
| /*************************************************************************** |
| * ISR Format |
| * 31 0 |
| * _______________________________________ |
| * | | | | | | | | | |
| * |____|____|____|____|____|____|____|____| |
| * ^^^^ ^^^^ ^^^^ ^^^^ \ / |
| * ERR --|||| |||| |||| |||| ----------------- |
| * EVENT ---||| |||| |||| |||| | |
| * ----|| |||| |||| |||| |-- Crash Address |
| * UPC -----| |||| |||| |||| |
| * LEVENT -------|||| |||| |||| |
| * PDQF --------||| |||| |||| |
| * RMISS ---------|| |||| |||| |
| * BREAK ----------| |||| |||| |
| * HBEATOK ------------|||| |||| |
| * NOHBEAT -------------||| |||| |
| * ERFULL --------------|| |||| |
| * XDROP ---------------| |||| |
| * -----------------|||| |
| * -----------------||||--\ |
| * ||---|-CpuId of crash |
| * |----/ |
| ***************************************************************************/ |
| #define SXG_ISR_ERR 0x80000000 /* Error */ |
| #define SXG_ISR_EVENT 0x40000000 /* Event ring event */ |
| #define SXG_ISR_NONE1 0x20000000 /* Not used */ |
| #define SXG_ISR_UPC 0x10000000 /* Dump/debug command complete*/ |
| #define SXG_ISR_LINK 0x08000000 /* Link event */ |
| #define SXG_ISR_PDQF 0x04000000 /* Processed data queue full */ |
| #define SXG_ISR_RMISS 0x02000000 /* Drop - no host buf */ |
| #define SXG_ISR_BREAK 0x01000000 /* Breakpoint hit */ |
| #define SXG_ISR_PING 0x00800000 /* Heartbeat response */ |
| #define SXG_ISR_DEAD 0x00400000 /* Card crash */ |
| #define SXG_ISR_ERFULL 0x00200000 /* Event ring full */ |
| #define SXG_ISR_XDROP 0x00100000 /* XMT Drop - no DRAM bufs or XMT err */ |
| #define SXG_ISR_SPSEND 0x00080000 /* Slow send complete */ |
| #define SXG_ISR_CPU 0x00070000 /* Dead CPU mask */ |
| #define SXG_ISR_CPU_SHIFT 16 /* Dead CPU shift */ |
| #define SXG_ISR_CRASH 0x0000FFFF /* Crash address mask */ |
| |
| /*************************************************************************** |
| * Event Ring entry |
| * |
| * 31 15 0 |
| * .___________________.___________________. |
| * |<------------ Pad 0 ------------>| |
| * |_________|_________|_________|_________|0 0x00 |
| * |<------------ Pad 1 ------------>| |
| * |_________|_________|_________|_________|4 0x04 |
| * |<------------ Pad 2 ------------>| |
| * |_________|_________|_________|_________|8 0x08 |
| * |<----------- Event Word 0 ------------>| |
| * |_________|_________|_________|_________|12 0x0c |
| * |<----------- Event Word 1 ------------>| |
| * |_________|_________|_________|_________|16 0x10 |
| * |<------------- Toeplitz ------------>| |
| * |_________|_________|_________|_________|20 0x14 |
| * |<----- Length ---->|<------ TCB Id --->| |
| * |_________|_________|_________|_________|24 0x18 |
| * |<----- Status ---->|Evnt Code|Flsh Code| |
| * |_________|_________|_________|_________|28 0x1c |
| * ^ ^^^^ ^^^^ |
| * |- VALID |||| ||||- RBUFC |
| * |||| |||-- SLOWR |
| * |||| ||--- UNUSED |
| * |||| |---- FASTC |
| * ||||------ FASTR |
| * |||------- |
| * ||-------- |
| * |--------- |
| * |
| * Slowpath status: |
| * _______________________________________ |
| * |<----- Status ---->|Evnt Code|Flsh Code| |
| * |_________|Cmd Index|_________|_________|28 0x1c |
| * ^^^ ^^^^ |
| * ||| ||||- ISTCPIP6 |
| * ||| |||-- IPONLY |
| * ||| ||--- RCVERR |
| * ||| |---- IPCBAD |
| * |||------ TCPCBAD |
| * ||------- ISTCPIP |
| * |-------- SCERR |
| * |
| ************************************************************************/ |
| #pragma pack(push, 1) |
| struct sxg_event { |
| u32 Pad[1]; /* not used */ |
| u32 SndUna; /* SndUna value */ |
| u32 Resid; /* receive MDL resid */ |
| union { |
| void * HostHandle; /* Receive host handle */ |
| u32 Rsvd1; /* TOE NA */ |
| struct { |
| u32 NotUsed; |
| u32 Rsvd2; /* TOE NA */ |
| } Flush; |
| }; |
| u32 Toeplitz; /* RSS Toeplitz hash */ |
| union { |
| ushort Rsvd3; /* TOE NA */ |
| ushort HdrOffset; /* Slowpath */ |
| }; |
| ushort Length; |
| unsigned char Rsvd4; /* TOE NA */ |
| unsigned char Code; /* Event code */ |
| unsigned char CommandIndex; /* New ring index */ |
| unsigned char Status; /* Event status */ |
| }; |
| #pragma pack(pop) |
| |
| /* Event code definitions */ |
| #define EVENT_CODE_BUFFERS 0x01 /* Receive buffer list command (ring 0) */ |
| #define EVENT_CODE_SLOWRCV 0x02 /* Slowpath receive */ |
| #define EVENT_CODE_UNUSED 0x04 /* Was slowpath commands complete */ |
| |
| /* Status values */ |
| #define EVENT_STATUS_VALID 0x80 /* Entry valid */ |
| |
| /* Slowpath status */ |
| #define EVENT_STATUS_ERROR 0x40 /* Completed with error. Index in next byte */ |
| #define EVENT_STATUS_TCPIP4 0x20 /* TCPIPv4 frame */ |
| #define EVENT_STATUS_TCPBAD 0x10 /* Bad TCP checksum */ |
| #define EVENT_STATUS_IPBAD 0x08 /* Bad IP checksum */ |
| #define EVENT_STATUS_RCVERR 0x04 /* Slowpath receive error */ |
| #define EVENT_STATUS_IPONLY 0x02 /* IP frame */ |
| #define EVENT_STATUS_TCPIP6 0x01 /* TCPIPv6 frame */ |
| #define EVENT_STATUS_TCPIP 0x21 /* Combination of v4 and v6 */ |
| |
| /* |
| * Event ring |
| * Size must be power of 2, between 128 and 16k |
| */ |
| #define EVENT_RING_SIZE 4096 |
| #define EVENT_RING_BATCH 16 /* Hand entries back 16 at a time. */ |
| /* Stop processing events after 4096 (256 * 16) */ |
| #define EVENT_BATCH_LIMIT 256 |
| |
| struct sxg_event_ring { |
| struct sxg_event Ring[EVENT_RING_SIZE]; |
| }; |
| |
| /* TCB Buffers */ |
| /* Maximum number of TCBS supported by hardware/microcode */ |
| #define SXG_MAX_TCB 4096 |
| /* Minimum TCBs before we fail initialization */ |
| #define SXG_MIN_TCB 512 |
| /* |
| * TCB Hash |
| * The bucket is determined by bits 11:4 of the toeplitz if we support 4k |
| * offloaded connections, 10:4 if we support 2k and so on. |
| */ |
| #define SXG_TCB_BUCKET_SHIFT 4 |
| #define SXG_TCB_PER_BUCKET 16 |
| #define SXG_TCB_BUCKET_MASK 0xFF0 /* Bucket portion of TCB ID */ |
| #define SXG_TCB_ELEMENT_MASK 0x00F /* Element within bucket */ |
| #define SXG_TCB_BUCKETS 256 /* 256 * 16 = 4k */ |
| |
| #define SXG_TCB_BUFFER_SIZE 512 /* ASSERT format is correct */ |
| |
| #define SXG_TCB_RCVQ_SIZE 736 |
| |
| #define SXG_TCB_COMPOSITE_BUFFER_SIZE 1024 |
| |
| #define SXG_LOCATE_TCP_FRAME_HDR(_TcpObject, _IPv6) \ |
| (((_TcpObject)->VlanId) ? \ |
| ((_IPv6) ? /* Vlan frame header = yes */ \ |
| &(_TcpObject)->CompBuffer->Frame.HasVlan.TcpIp6.SxgTcp: \ |
| &(_TcpObject)->CompBuffer->Frame.HasVlan.TcpIp.SxgTcp): \ |
| ((_IPv6) ? /* Vlan frame header = No */ \ |
| &(_TcpObject)->CompBuffer->Frame.NoVlan.TcpIp6.SxgTcp : \ |
| &(_TcpObject)->CompBuffer->Frame.NoVlan.TcpIp.SxgTcp)) |
| |
| #define SXG_LOCATE_IP_FRAME_HDR(_TcpObject) \ |
| (_TcpObject)->VlanId ? \ |
| &(_TcpObject)->CompBuffer->Frame.HasVlan.TcpIp.Ip: \ |
| &(_TcpObject)->CompBuffer->Frame.NoVlan.TcpIp.Ip |
| |
| #define SXG_LOCATE_IP6_FRAME_HDR(TcpObject) \ |
| (_TcpObject)->VlanId ? \ |
| &(_TcpObject)->CompBuffer->Frame.HasVlan.TcpIp6.Ip: \ |
| &(_TcpObject)->CompBuffer->Frame.NoVlan.TcpIp6.Ip |
| |
| #if DBG |
| /* |
| * Horrible kludge to distinguish dumb-nic, slowpath, and |
| * fastpath traffic. Decrement the HopLimit by one |
| * for slowpath, two for fastpath. This assumes the limit is measurably |
| * greater than two, which I think is reasonable. |
| * Obviously this is DBG only. Maybe remove later, or #if 0 so we |
| * can set it when needed |
| */ |
| #define SXG_DBG_HOP_LIMIT(_TcpObject, _FastPath) { \ |
| PIPV6_HDR _Ip6FrameHdr; \ |
| if ((_TcpObject)->IPv6) { \ |
| _Ip6FrameHdr = SXG_LOCATE_IP6_FRAME_HDR((_TcpObject)); \ |
| if (_FastPath) { \ |
| _Ip6FrameHdr->HopLimit = \ |
| (_TcpObject)->Cached.TtlOrHopLimit - 2; \ |
| } else { \ |
| _Ip6FrameHdr->HopLimit = \ |
| (_TcpObject)->Cached.TtlOrHopLimit - 1; \ |
| } \ |
| } \ |
| } |
| #else |
| /* Do nothing with free build */ |
| #define SXG_DBG_HOP_LIMIT(_TcpObject, _FastPath) |
| #endif |
| |
| /* Receive and transmit rings */ |
| #define SXG_MAX_RING_SIZE 256 |
| #define SXG_XMT_RING_SIZE 128 /* Start with 128 */ |
| #define SXG_RCV_RING_SIZE 128 /* Start with 128 */ |
| #define SXG_MAX_ENTRIES 4096 |
| #define SXG_JUMBO_RCV_RING_SIZE 32 |
| |
| /* Structure and macros to manage a ring */ |
| struct sxg_ring_info { |
| /* Where we add entries - Note unsigned char:RING_SIZE */ |
| unsigned char Head; |
| unsigned char Tail; /* Where we pull off completed entries */ |
| ushort Size; /* Ring size - Must be multiple of 2 */ |
| void * Context[SXG_MAX_RING_SIZE]; /* Shadow ring */ |
| }; |
| |
| #define SXG_INITIALIZE_RING(_ring, _size) { \ |
| (_ring).Head = 0; \ |
| (_ring).Tail = 0; \ |
| (_ring).Size = (_size); \ |
| } |
| |
| #define SXG_ADVANCE_INDEX(_index, _size) \ |
| ((_index) = ((_index) + 1) & ((_size) - 1)) |
| #define SXG_PREVIOUS_INDEX(_index, _size) \ |
| (((_index) - 1) &((_size) - 1)) |
| #define SXG_RING_EMPTY(_ring) ((_ring)->Head == (_ring)->Tail) |
| #define SXG_RING_FULL(_ring) \ |
| ((((_ring)->Head + 1) & ((_ring)->Size - 1)) == (_ring)->Tail) |
| #define SXG_RING_ADVANCE_HEAD(_ring) \ |
| SXG_ADVANCE_INDEX((_ring)->Head, ((_ring)->Size)) |
| #define SXG_RING_RETREAT_HEAD(_ring) ((_ring)->Head = \ |
| SXG_PREVIOUS_INDEX((_ring)->Head, (_ring)->Size)) |
| #define SXG_RING_ADVANCE_TAIL(_ring) { \ |
| ASSERT((_ring)->Tail != (_ring)->Head); \ |
| SXG_ADVANCE_INDEX((_ring)->Tail, ((_ring)->Size)); \ |
| } |
| /* |
| * Set cmd to the next available ring entry, set the shadow context |
| * entry and advance the ring. |
| * The appropriate lock must be held when calling this macro |
| */ |
| #define SXG_GET_CMD(_ring, _ringinfo, _cmd, _context) { \ |
| if(SXG_RING_FULL(_ringinfo)) { \ |
| (_cmd) = NULL; \ |
| } else { \ |
| (_cmd) = &(_ring)->Descriptors[(_ringinfo)->Head]; \ |
| (_ringinfo)->Context[(_ringinfo)->Head] = (void *)(_context);\ |
| SXG_RING_ADVANCE_HEAD(_ringinfo); \ |
| } \ |
| } |
| |
| /* |
| * Abort the previously allocated command by retreating the head. |
| * NOTE - The appopriate lock MUST NOT BE DROPPED between the SXG_GET_CMD |
| * and SXG_ABORT_CMD calls. |
| */ |
| #define SXG_ABORT_CMD(_ringinfo) { \ |
| ASSERT(!(SXG_RING_EMPTY(_ringinfo))); \ |
| SXG_RING_RETREAT_HEAD(_ringinfo); \ |
| (_ringinfo)->Context[(_ringinfo)->Head] = NULL; \ |
| } |
| |
| /* |
| * For the given ring, return a pointer to the tail cmd and context, |
| * clear the context and advance the tail |
| */ |
| #define SXG_RETURN_CMD(_ring, _ringinfo, _cmd, _context) { \ |
| (_cmd) = &(_ring)->Descriptors[(_ringinfo)->Tail]; \ |
| (_context) = (_ringinfo)->Context[(_ringinfo)->Tail]; \ |
| (_ringinfo)->Context[(_ringinfo)->Tail] = NULL; \ |
| SXG_RING_ADVANCE_TAIL(_ringinfo); \ |
| } |
| |
| /* |
| * For a given ring find out how much the first pointer is ahead of |
| * the second pointer. "ahead" recognises the fact that the ring can wrap |
| */ |
| static inline int sxg_ring_get_forward_diff (struct sxg_ring_info *ringinfo, |
| int a, int b) { |
| if ((a < 0 || a > ringinfo->Size ) || (b < 0 || b > ringinfo->Size)) |
| return -1; |
| if (a > b) /* _a is lagging _b and _b has not wrapped around */ |
| return (a - b); |
| else |
| return ((ringinfo->Size - (b - a))); |
| } |
| |
| /*************************************************************** |
| * Host Command Buffer - commands to INIC via the Cmd Rings |
| * |
| * 31 15 0 |
| * .___________________.___________________. |
| * |<-------------- Sgl Low -------------->| |
| * |_________|_________|_________|_________|0 0x00 |
| * |<-------------- Sgl High ------------->| |
| * |_________|_________|_________|_________|4 0x04 |
| * |<------------- Sge 0 Low ----------->| |
| * |_________|_________|_________|_________|8 0x08 |
| * |<------------- Sge 0 High ----------->| |
| * |_________|_________|_________|_________|12 0x0c |
| * |<------------ Sge 0 Length ---------->| |
| * |_________|_________|_________|_________|16 0x10 |
| * |<----------- Window Update ----------->| |
| * |<-------- SP 1st SGE offset ---------->| |
| * |_________|_________|_________|_________|20 0x14 |
| * |<----------- Total Length ------------>| |
| * |_________|_________|_________|_________|24 0x18 |
| * |<----- LCnt ------>|<----- Flags ----->| |
| * |_________|_________|_________|_________|28 0x1c |
| ****************************************************************/ |
| #pragma pack(push, 1) |
| struct sxg_cmd { |
| dma64_addr_t Sgl; /* Physical address of SGL */ |
| union { |
| struct { |
| dma64_addr_t FirstSgeAddress; /* Address of first SGE */ |
| u32 FirstSgeLength; /* Length of first SGE */ |
| union { |
| u32 Rsvd1; /* TOE NA */ |
| u32 SgeOffset; /* Slowpath - 2nd SGE offset */ |
| /* MDL completion - clobbers update */ |
| u32 Resid; |
| }; |
| union { |
| u32 TotalLength; /* Total transfer length */ |
| u32 Mss; /* LSO MSS */ |
| }; |
| } Buffer; |
| }; |
| union { |
| struct { |
| unsigned char Flags:4; /* slowpath flags */ |
| unsigned char IpHl:4; /* Ip header length (>>2) */ |
| unsigned char MacLen; /* Mac header len */ |
| } CsumFlags; |
| struct { |
| ushort Flags:4; /* slowpath flags */ |
| ushort TcpHdrOff:7; /* TCP */ |
| ushort MacLen:5; /* Mac header len */ |
| } LsoFlags; |
| ushort Flags; /* flags */ |
| }; |
| union { |
| ushort SgEntries; /* SG entry count including first sge */ |
| struct { |
| unsigned char Status; /* Copied from event status */ |
| unsigned char NotUsed; |
| } Status; |
| }; |
| }; |
| #pragma pack(pop) |
| |
| #pragma pack(push, 1) |
| struct vlan_hdr { |
| ushort VlanTci; |
| ushort VlanTpid; |
| }; |
| #pragma pack(pop) |
| |
| /******************************************************************** |
| * Slowpath Flags: |
| * |
| * |
| * LSS Flags: |
| * .--- |
| * /.--- TCP Large segment send |
| * //.--- |
| * ///.--- |
| * 3 1 1 //// |
| * 1 5 0 |||| |
| * .___________________.____________vvvv. |
| * | |MAC | TCP | | |
| * | LCnt |hlen|hdroff|Flgs| |
| * |___________________|||||||||||||____| |
| * |
| * |
| * Checksum Flags |
| * |
| * .--- |
| * /.--- |
| * //.--- Checksum TCP |
| * ///.--- Checksum IP |
| * 3 1 //// No bits - normal send |
| * 1 5 7 |||| |
| * .___________________._______________vvvv. |
| * | | Offload | IP | | |
| * | LCnt |MAC hlen |Hlen|Flgs| |
| * |___________________|____|____|____|____| |
| * |
| *****************************************************************/ |
| /* Slowpath CMD flags */ |
| #define SXG_SLOWCMD_CSUM_IP 0x01 /* Checksum IP */ |
| #define SXG_SLOWCMD_CSUM_TCP 0x02 /* Checksum TCP */ |
| #define SXG_SLOWCMD_LSO 0x04 /* Large segment send */ |
| |
| struct sxg_xmt_ring { |
| struct sxg_cmd Descriptors[SXG_XMT_RING_SIZE]; |
| }; |
| |
| struct sxg_rcv_ring { |
| struct sxg_cmd Descriptors[SXG_RCV_RING_SIZE]; |
| }; |
| |
| /* |
| * Share memory buffer types - Used to identify asynchronous |
| * shared memory allocation |
| */ |
| enum sxg_buffer_type { |
| SXG_BUFFER_TYPE_RCV, /* Receive buffer */ |
| SXG_BUFFER_TYPE_SGL /* SGL buffer */ |
| }; |
| |
| /* State for SXG buffers */ |
| #define SXG_BUFFER_FREE 0x01 |
| #define SXG_BUFFER_BUSY 0x02 |
| #define SXG_BUFFER_ONCARD 0x04 |
| #define SXG_BUFFER_UPSTREAM 0x08 |
| |
| /* |
| * Receive data buffers |
| * |
| * Receive data buffers are given to the Sahara card 128 at a time. |
| * This is accomplished by filling in a "receive descriptor block" |
| * with 128 "receive descriptors". Each descriptor consists of |
| * a physical address, which the card uses as the address to |
| * DMA data into, and a virtual address, which is given back |
| * to the host in the "HostHandle" portion of an event. |
| * The receive descriptor data structure is defined below |
| * as sxg_rcv_data_descriptor, and the corresponding block |
| * is defined as sxg_rcv_descriptor_block. |
| * |
| * This receive descriptor block is given to the card by filling |
| * in the Sgl field of a sxg_cmd entry from pAdapt->RcvRings[0] |
| * with the physical address of the receive descriptor block. |
| * |
| * Both the receive buffers and the receive descriptor blocks |
| * require additional data structures to maintain them |
| * on a free queue and contain other information associated with them. |
| * Those data structures are defined as the sxg_rcv_data_buffer_hdr |
| * and sxg_rcv_descriptor_block_hdr respectively. |
| * |
| * Since both the receive buffers and the receive descriptor block |
| * must be accessible by the card, both must be allocated out of |
| * shared memory. To ensure that we always have a descriptor |
| * block available for every 128 buffers, we allocate all of |
| * these resources together in a single block. This entire |
| * block is managed by a struct sxg_rcv_block_hdr, who's sole purpose |
| * is to maintain address information so that the entire block |
| * can be free later. |
| * |
| * Further complicating matters is the fact that the receive |
| * buffers must be variable in length in order to accomodate |
| * jumbo frame configurations. We configure the buffer |
| * length so that the buffer and it's corresponding struct |
| * sxg_rcv_data_buffer_hdr structure add up to an even |
| * boundary. Then we place the remaining data structures after 128 |
| * of them as shown in the following diagram: |
| * |
| * _________________________________________ |
| * | | |
| * | Variable length receive buffer #1 | |
| * |_________________________________________| |
| * | | |
| * | sxg_rcv_data_buffer_hdr #1 | |
| * |_________________________________________| <== Even 2k or 10k boundary |
| * | | |
| * | ... repeat 2-128 .. | |
| * |_________________________________________| |
| * | | |
| * | struct sxg_rcv_descriptor_block | |
| * | Contains sxg_rcv_data_descriptor * 128 | |
| * |_________________________________________| |
| * | | |
| * | struct sxg_rcv_descriptor_block_hdr | |
| * |_________________________________________| |
| * | | |
| * | struct sxg_rcv_block_hdr | |
| * |_________________________________________| |
| * |
| * Memory consumption: |
| * Non-jumbo: |
| * Buffers and sxg_rcv_data_buffer_hdr = 2k * 128 = 256k |
| * + struct sxg_rcv_descriptor_block = 2k |
| * + struct sxg_rcv_descriptor_block_hdr = ~32 |
| * + struct sxg_rcv_block_hdr = ~32 |
| * => Total = ~258k/block |
| * |
| * Jumbo: |
| * Buffers and sxg_rcv_data_buffer_hdr = 10k * 128 = 1280k |
| * + struct sxg_rcv_descriptor_block = 2k |
| * + struct sxg_rcv_descriptor_block_hdr = ~32 |
| * + struct sxg_rcv_block_hdr = ~32 |
| * => Total = ~1282k/block |
| * |
| */ |
| #define SXG_RCV_DATA_BUFFERS 8192 /* Amount to give to the card */ |
| #define SXG_INITIAL_RCV_DATA_BUFFERS 16384 /* Initial pool of buffers */ |
| /* Minimum amount and when to get more */ |
| #define SXG_MIN_RCV_DATA_BUFFERS 4096 |
| #define SXG_MAX_RCV_BLOCKS 256 /* = 32k receive buffers */ |
| /* Amount to give to the card in case of jumbo frames */ |
| #define SXG_JUMBO_RCV_DATA_BUFFERS 2048 |
| /* Initial pool of buffers in case of jumbo buffers */ |
| #define SXG_INITIAL_JUMBO_RCV_DATA_BUFFERS 4096 |
| #define SXG_MIN_JUMBO_RCV_DATA_BUFFERS 1024 |
| |
| /* Receive buffer header */ |
| struct sxg_rcv_data_buffer_hdr { |
| dma64_addr_t PhysicalAddress; /* Buffer physical address */ |
| /* |
| * Note - DO NOT USE the VirtualAddress field to locate data. |
| * Use the sxg.h:SXG_RECEIVE_DATA_LOCATION macro instead. |
| */ |
| struct list_entry FreeList; /* Free queue of buffers */ |
| unsigned char State; /* See SXG_BUFFER state above */ |
| struct sk_buff * skb; /* Double mapped (nbl and pkt)*/ |
| }; |
| |
| /* |
| * SxgSlowReceive uses the PACKET (skb) contained |
| * in the struct sxg_rcv_data_buffer_hdr when indicating dumb-nic data |
| */ |
| #define SxgDumbRcvPacket skb |
| |
| /* Space for struct sxg_rcv_data_buffer_hdr */ |
| #define SXG_RCV_DATA_HDR_SIZE sizeof(struct sxg_rcv_data_buffer_hdr) |
| /* Non jumbo = 2k including HDR */ |
| #define SXG_RCV_DATA_BUFFER_SIZE 2048 |
| /* jumbo = 10k including HDR */ |
| #define SXG_RCV_JUMBO_BUFFER_SIZE 10240 |
| |
| /* Receive data descriptor */ |
| struct sxg_rcv_data_descriptor { |
| union { |
| struct sk_buff *VirtualAddress; /* Host handle */ |
| u64 ForceTo8Bytes; /*Force x86 to 8-byte boundary*/ |
| }; |
| dma64_addr_t PhysicalAddress; |
| }; |
| |
| /* Receive descriptor block */ |
| #define SXG_RCV_DESCRIPTORS_PER_BLOCK 128 |
| #define SXG_RCV_DESCRIPTOR_BLOCK_SIZE 2048 /* For sanity check */ |
| |
| struct sxg_rcv_descriptor_block { |
| struct sxg_rcv_data_descriptor Descriptors[SXG_RCV_DESCRIPTORS_PER_BLOCK]; |
| }; |
| |
| /* Receive descriptor block header */ |
| struct sxg_rcv_descriptor_block_hdr { |
| void *VirtualAddress; /* start of 2k buffer */ |
| dma64_addr_t PhysicalAddress;/* and it's physical address */ |
| struct list_entry FreeList;/* free queue of descriptor blocks */ |
| unsigned char State; /* see sxg_buffer state above */ |
| }; |
| |
| /* Receive block header */ |
| struct sxg_rcv_block_hdr { |
| void *VirtualAddress; /* Start of virtual memory */ |
| dma64_addr_t PhysicalAddress;/* ..and it's physical address*/ |
| struct list_entry AllList; /* Queue of all SXG_RCV_BLOCKS*/ |
| }; |
| |
| /* Macros to determine data structure offsets into receive block */ |
| #define SXG_RCV_BLOCK_SIZE(_Buffersize) \ |
| (((_Buffersize) * SXG_RCV_DESCRIPTORS_PER_BLOCK) + \ |
| (sizeof(struct sxg_rcv_descriptor_block)) + \ |
| (sizeof(struct sxg_rcv_descriptor_block_hdr)) + \ |
| (sizeof(struct sxg_rcv_block_hdr))) |
| #define SXG_RCV_BUFFER_DATA_SIZE(_Buffersize) \ |
| ((_Buffersize) - SXG_RCV_DATA_HDR_SIZE) |
| #define SXG_RCV_DATA_BUFFER_HDR_OFFSET(_Buffersize) \ |
| ((_Buffersize) - SXG_RCV_DATA_HDR_SIZE) |
| #define SXG_RCV_DESCRIPTOR_BLOCK_OFFSET(_Buffersize) \ |
| ((_Buffersize) * SXG_RCV_DESCRIPTORS_PER_BLOCK) |
| #define SXG_RCV_DESCRIPTOR_BLOCK_HDR_OFFSET(_Buffersize) \ |
| (((_Buffersize) * SXG_RCV_DESCRIPTORS_PER_BLOCK) + \ |
| (sizeof(struct sxg_rcv_descriptor_block))) |
| #define SXG_RCV_BLOCK_HDR_OFFSET(_Buffersize) \ |
| (((_Buffersize) * SXG_RCV_DESCRIPTORS_PER_BLOCK) + \ |
| (sizeof(struct sxg_rcv_descriptor_block)) + \ |
| (sizeof(struct sxg_rcv_descriptor_block_hdr))) |
| |
| /* Scatter gather list buffer */ |
| #define SXG_INITIAL_SGL_BUFFERS 8192 /* Initial pool of SGL buffers */ |
| #define SXG_MIN_SGL_BUFFERS 2048 /* Minimum amount and when to get more*/ |
| /* Maximum to allocate (note ADAPT:ushort) */ |
| #define SXG_MAX_SGL_BUFFERS 16384 |
| |
| /* |
| * SXG_SGL_POOL_PROPERTIES - This structure is used to define a pool of SGL |
| * buffers. These buffers are allocated out of shared memory and used to |
| * contain a physical scatter gather list structure that is shared |
| * with the card. |
| * |
| * We split our SGL buffers into multiple pools based on size. The motivation |
| * is that some applications perform very large I/Os (1MB for example), so |
| * we need to be able to allocate an SGL to accommodate such a request. |
| * But such an SGL would require 256 24-byte SG entries - ~6k. |
| * Given that the vast majority of I/Os are much smaller than 1M, allocating |
| * a single pool of SGL buffers would be a horribly inefficient use of |
| * memory. |
| * |
| * The following structure includes two fields relating to its size. |
| * The NBSize field specifies the largest NET_BUFFER that can be handled |
| * by the particular pool. The SGEntries field defines the size, in |
| * entries, of the SGL for that pool. The SGEntries is determined by |
| * dividing the NBSize by the expected page size (4k), and then padding |
| * it by some appropriate amount as insurance (20% or so..??). |
| */ |
| struct sxg_sgl_pool_properties { |
| u32 NBSize; /* Largest NET_BUFFER size for this pool */ |
| ushort SGEntries; /* Number of entries in SGL */ |
| ushort InitialBuffers; /* Number to allocate at initializationtime */ |
| ushort MinBuffers; /* When to get more */ |
| ushort MaxBuffers; /* When to stop */ |
| ushort PerCpuThreshold;/* See sxgh.h:SXG_RESOURCES */ |
| }; |
| |
| /* |
| * At the moment I'm going to statically initialize 4 pools: |
| * 100k buffer pool: The vast majority of the expected buffers are expected |
| * to be less than or equal to 100k. At 30 entries per and |
| * 8k initial buffers amounts to ~4MB of memory |
| * NOTE - This used to be 64K with 20 entries, but during |
| * WHQL NDIS 6.0 Testing (2c_mini6stress) MS does their |
| * best to send absurd NBL's with ridiculous SGLs, we |
| * have received 400byte sends contained in SGL's that |
| * have 28 entries |
| * 1M buffer pool: Buffers between 64k and 1M. Allocate 256 initial |
| * buffers with 300 entries each => ~2MB of memory |
| * 5M buffer pool: Not expected often, if at all. 32 initial buffers |
| * at 1500 entries each => ~1MB of memory |
| * 10M buffer pool: Not expected at all, except under pathelogical conditions. |
| * Allocate one at initialization time. |
| * Note - 10M is the current limit of what we can realistically |
| * support due to the sahara SGL bug described in the |
| * SAHARA SGL WORKAROUND below. We will likely adjust the |
| * number of pools and/or pool properties over time. |
| */ |
| #define SXG_NUM_SGL_POOLS 4 |
| #define INITIALIZE_SGL_POOL_PROPERTIES \ |
| struct sxg_sgl_pool_properties SxgSglPoolProperties[SXG_NUM_SGL_POOLS] =\ |
| { \ |
| { 102400, 30, 8192, 2048, 16384, 256}, \ |
| { 1048576, 300, 256, 128, 1024, 16}, \ |
| { 5252880, 1500, 32, 16, 512, 0}, \ |
| {10485760, 2700, 2, 4, 32, 0}, \ |
| }; |
| |
| extern struct sxg_sgl_pool_properties SxgSglPoolProperties[]; |
| |
| #define SXG_MAX_SGL_BUFFER_SIZE \ |
| SxgSglPoolProperties[SXG_NUM_SGL_POOLS - 1].NBSize |
| |
| /* |
| * SAHARA SGL WORKAROUND!! |
| * The current Sahara card uses a 16-bit counter when advancing |
| * SGL address locations. This means that if an SGL crosses |
| * a 64k boundary, the hardware will actually skip back to |
| * the start of the previous 64k boundary, with obviously |
| * undesirable results. |
| * |
| * We currently workaround this issue by allocating SGL buffers |
| * in 64k blocks and skipping over buffers that straddle the boundary. |
| */ |
| #define SXG_INVALID_SGL(phys_addr,len) \ |
| (((phys_addr >> 16) != ( (phys_addr + len) >> 16 ))) |
| |
| /* |
| * Allocate SGLs in blocks so we can skip over invalid entries. |
| * We allocation 64k worth of SGL buffers, including the |
| * struct sxg_sgl_block_hdr, plus one for padding |
| */ |
| #define SXG_SGL_BLOCK_SIZE 65536 |
| #define SXG_SGL_ALLOCATION_SIZE(_Pool) \ |
| SXG_SGL_BLOCK_SIZE + SXG_SGL_SIZE(_Pool) |
| |
| struct sxg_sgl_block_hdr { |
| ushort Pool; /* Associated SGL pool */ |
| /* struct sxg_scatter_gather blocks */ |
| struct list_entry List; |
| dma64_addr_t PhysicalAddress;/* physical address */ |
| }; |
| |
| /* |
| * The following definition denotes the maximum block of memory that the |
| * card can DMA to.It is specified in the call to NdisMRegisterScatterGatherDma. |
| * For now, use the same value as used in the Slic/Oasis driver, which |
| * is 128M. That should cover any expected MDL that I can think of. |
| */ |
| #define SXG_MAX_PHYS_MAP (1024 * 1024 * 128) |
| |
| /* Self identifying structure type */ |
| enum SXG_SGL_TYPE { |
| SXG_SGL_DUMB, /* Dumb NIC SGL */ |
| SXG_SGL_SLOW, /* Slowpath protocol header - see below */ |
| SXG_SGL_CHIMNEY /* Chimney offload SGL */ |
| }; |
| |
| /* |
| * The ucode expects an NDIS SGL structure that |
| * is formatted for an x64 system. When running |
| * on an x64 system, we can simply hand the NDIS SGL |
| * to the card directly. For x86 systems we must reconstruct |
| * the SGL. The following structure defines an x64 |
| * formatted SGL entry |
| */ |
| struct sxg_x64_sge { |
| dma64_addr_t Address; /* same as wdm.h */ |
| u32 Length; /* same as wdm.h */ |
| u32 CompilerPad; /* The compiler pads to 8-bytes */ |
| u64 Reserved; /* u32 * in wdm.h. Force to 8 bytes */ |
| }; |
| |
| /* |
| * Our SGL structure - Essentially the same as |
| * wdm.h:SCATTER_GATHER_LIST. Note the variable number of |
| * elements based on the pool specified above |
| */ |
| struct sxg_x64_sgl { |
| u32 NumberOfElements; |
| u32 *Reserved; |
| struct sxg_x64_sge Elements[1]; /* Variable */ |
| }; |
| |
| struct sxg_scatter_gather { |
| enum SXG_SGL_TYPE Type; /* FIRST! Dumb-nic or offload */ |
| ushort Pool; /* Associated SGL pool */ |
| ushort Entries; /* SGL total entries */ |
| void * adapter; /* Back pointer to adapter */ |
| /* Free struct sxg_scatter_gather blocks */ |
| struct list_entry FreeList; |
| /* All struct sxg_scatter_gather blocks */ |
| struct list_entry AllList; |
| dma64_addr_t PhysicalAddress;/* physical address */ |
| unsigned char State; /* See SXG_BUFFER state above */ |
| unsigned char CmdIndex; /* Command ring index */ |
| struct sk_buff *DumbPacket; /* Associated Packet */ |
| /* For asynchronous completions */ |
| u32 Direction; |
| u32 CurOffset; /* Current SGL offset */ |
| u32 SglRef; /* SGL reference count */ |
| struct vlan_hdr VlanTag; /* VLAN tag to be inserted into SGL */ |
| struct sxg_x64_sgl *pSgl; /* SGL Addr. Possibly &Sgl */ |
| struct sxg_x64_sgl Sgl; /* SGL handed to card */ |
| }; |
| |
| /* |
| * Note - the "- 1" is because struct sxg_scatter_gather=>struct sxg_x64_sgl |
| * includes 1 SGE.. |
| */ |
| #define SXG_SGL_SIZE(_Pool) \ |
| (sizeof(struct sxg_scatter_gather) + \ |
| ((SxgSglPoolProperties[_Pool].SGEntries - 1) * \ |
| sizeof(struct sxg_x64_sge))) |
| |
| /* Force NDIS to give us it's own buffer so we can reformat to our own */ |
| #define SXG_SGL_BUFFER(_SxgSgl) NULL |
| #define SXG_SGL_BUFFER_LENGTH(_SxgSgl) 0 |
| #define SXG_SGL_BUF_SIZE 0 |
| |
| /* |
| #if defined(CONFIG_X86_64) |
| #define SXG_SGL_BUFFER(_SxgSgl) (&_SxgSgl->Sgl) |
| #define SXG_SGL_BUFFER_LENGTH(_SxgSgl) ((_SxgSgl)->Entries * \ |
| sizeof(struct sxg_x64_sge)) |
| #define SXG_SGL_BUF_SIZE sizeof(struct sxg_x64_sgl) |
| #elif defined(CONFIG_X86) |
| // Force NDIS to give us it's own buffer so we can reformat to our own |
| #define SXG_SGL_BUFFER(_SxgSgl) NULL |
| #define SXG_SGL_BUFFER_LENGTH(_SxgSgl) 0 |
| #define SXG_SGL_BUF_SIZE 0 |
| #else |
| #error staging: sxg: driver is for X86 only! |
| #endif |
| */ |
| /* Microcode statistics */ |
| struct sxg_ucode_stats { |
| u32 RPDQOflow; /* PDQ overflow (unframed ie dq & drop 1st) */ |
| u32 XDrops; /* Xmt drops due to no xmt buffer */ |
| u32 ERDrops; /* Rcv drops due to ER full */ |
| u32 NBDrops; /* Rcv drops due to out of host buffers */ |
| u32 PQDrops; /* Rcv drops due to PDQ full */ |
| /* Rcv drops due to bad frame: no link addr match, frlen > max */ |
| u32 BFDrops; |
| u32 UPDrops; /* Rcv drops due to UPFq full */ |
| u32 XNoBufs; /* Xmt drop due to no DRAM Xmit buffer or PxyBuf */ |
| }; |
| |
| /* |
| * Macros for handling the Offload engine values |
| */ |
| /* Number of positions to shift Network Header Length before passing to card */ |
| #define SXG_NW_HDR_LEN_SHIFT 2 |