1 /**************************************************************************
3 * Copyright (C) 2000-2008 Alacritech, Inc. All rights reserved.
5 * Redistribution and use in source and binary forms, with or without
6 * modification, are permitted provided that the following conditions
9 * 1. Redistributions of source code must retain the above copyright
10 * notice, this list of conditions and the following disclaimer.
11 * 2. Redistributions in binary form must reproduce the above
12 * copyright notice, this list of conditions and the following
13 * disclaimer in the documentation and/or other materials provided
14 * with the distribution.
16 * THIS SOFTWARE IS PROVIDED BY ALACRITECH, INC. ``AS IS'' AND ANY
17 * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
18 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
19 * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL ALACRITECH, INC. OR
20 * CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
21 * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
22 * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF
23 * USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
24 * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,
25 * OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT
26 * OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
29 * The views and conclusions contained in the software and documentation
30 * are those of the authors and should not be interpreted as representing
31 * official policies, either expressed or implied, of Alacritech, Inc.
33 **************************************************************************/
38 * The SXG driver for Alacritech's 10Gbe products.
40 * NOTE: This is the standard, non-accelerated version of Alacritech's
44 #include <linux/kernel.h>
45 #include <linux/string.h>
46 #include <linux/errno.h>
47 #include <linux/module.h>
48 #include <linux/moduleparam.h>
49 #include <linux/ioport.h>
50 #include <linux/slab.h>
51 #include <linux/interrupt.h>
52 #include <linux/timer.h>
53 #include <linux/pci.h>
54 #include <linux/spinlock.h>
55 #include <linux/init.h>
56 #include <linux/netdevice.h>
57 #include <linux/etherdevice.h>
58 #include <linux/ethtool.h>
59 #include <linux/skbuff.h>
60 #include <linux/delay.h>
61 #include <linux/types.h>
62 #include <linux/dma-mapping.h>
63 #include <linux/mii.h>
65 #define SLIC_GET_STATS_ENABLED 0
66 #define LINUX_FREES_ADAPTER_RESOURCES 1
67 #define SXG_OFFLOAD_IP_CHECKSUM 0
68 #define SXG_POWER_MANAGEMENT_ENABLED 0
78 #include "sxgphycode.h"
79 #include "saharadbgdownload.h"
81 static int sxg_allocate_buffer_memory(struct adapter_t *adapter, u32 Size,
82 enum sxg_buffer_type BufferType);
83 static void sxg_allocate_rcvblock_complete(struct adapter_t *adapter,
85 dma_addr_t PhysicalAddress,
87 static void sxg_allocate_sgl_buffer_complete(struct adapter_t *adapter,
88 struct sxg_scatter_gather *SxgSgl,
89 dma_addr_t PhysicalAddress,
92 static void sxg_mcast_init_crc32(void);
93 static int sxg_entry_open(struct net_device *dev);
94 static int sxg_entry_halt(struct net_device *dev);
95 static int sxg_ioctl(struct net_device *dev, struct ifreq *rq, int cmd);
96 static int sxg_send_packets(struct sk_buff *skb, struct net_device *dev);
97 static int sxg_transmit_packet(struct adapter_t *adapter, struct sk_buff *skb);
98 static int sxg_dumb_sgl(struct sxg_x64_sgl *pSgl,
99 struct sxg_scatter_gather *SxgSgl);
101 static void sxg_handle_interrupt(struct adapter_t *adapter);
102 static int sxg_process_isr(struct adapter_t *adapter, u32 MessageId);
103 static u32 sxg_process_event_queue(struct adapter_t *adapter, u32 RssId);
104 static void sxg_complete_slow_send(struct adapter_t *adapter, int irq_context);
105 static struct sk_buff *sxg_slow_receive(struct adapter_t *adapter,
106 struct sxg_event *Event);
107 static void sxg_process_rcv_error(struct adapter_t *adapter, u32 ErrorStatus);
108 static bool sxg_mac_filter(struct adapter_t *adapter,
109 struct ether_header *EtherHdr, ushort length);
111 static struct net_device_stats *sxg_get_stats(struct net_device * dev);
112 void sxg_free_resources(struct adapter_t *adapter);
113 void sxg_free_rcvblocks(struct adapter_t *adapter);
114 void sxg_free_sgl_buffers(struct adapter_t *adapter);
115 void sxg_unmap_resources(struct adapter_t *adapter);
116 void sxg_free_mcast_addrs(struct adapter_t *adapter);
117 void sxg_collect_statistics(struct adapter_t *adapter);
121 static int sxg_mac_set_address(struct net_device *dev, void *ptr);
122 static void sxg_mcast_set_list(struct net_device *dev);
124 static void sxg_adapter_set_hwaddr(struct adapter_t *adapter);
126 static void sxg_unmap_mmio_space(struct adapter_t *adapter);
128 static int sxg_initialize_adapter(struct adapter_t *adapter);
129 static void sxg_stock_rcv_buffers(struct adapter_t *adapter);
130 static void sxg_complete_descriptor_blocks(struct adapter_t *adapter,
131 unsigned char Index);
132 static int sxg_initialize_link(struct adapter_t *adapter);
133 static int sxg_phy_init(struct adapter_t *adapter);
134 static void sxg_link_event(struct adapter_t *adapter);
135 static enum SXG_LINK_STATE sxg_get_link_state(struct adapter_t *adapter);
136 static void sxg_link_state(struct adapter_t *adapter,
137 enum SXG_LINK_STATE LinkState);
138 static int sxg_write_mdio_reg(struct adapter_t *adapter,
139 u32 DevAddr, u32 RegAddr, u32 Value);
140 static int sxg_read_mdio_reg(struct adapter_t *adapter,
141 u32 DevAddr, u32 RegAddr, u32 *pValue);
143 static unsigned int sxg_first_init = 1;
144 static char *sxg_banner =
145 "Alacritech SLIC Technology(tm) Server and Storage \
146 10Gbe Accelerator (Non-Accelerated)\n";
148 static int sxg_debug = 1;
149 static int debug = -1;
150 static struct net_device *head_netdevice = NULL;
152 static struct sxgbase_driver sxg_global = {
155 static int intagg_delay = 100;
156 static u32 dynamic_intagg = 0;
158 char sxg_driver_name[] = "sxg";
159 #define DRV_AUTHOR "Alacritech, Inc. Engineering"
160 #define DRV_DESCRIPTION \
161 "Alacritech SLIC Techonology(tm) Non-Accelerated 10Gbe Driver"
162 #define DRV_COPYRIGHT \
163 "Copyright 2000-2008 Alacritech, Inc. All rights reserved."
165 MODULE_AUTHOR(DRV_AUTHOR);
166 MODULE_DESCRIPTION(DRV_DESCRIPTION);
167 MODULE_LICENSE("GPL");
169 module_param(dynamic_intagg, int, 0);
170 MODULE_PARM_DESC(dynamic_intagg, "Dynamic Interrupt Aggregation Setting");
171 module_param(intagg_delay, int, 0);
172 MODULE_PARM_DESC(intagg_delay, "uSec Interrupt Aggregation Delay");
174 static struct pci_device_id sxg_pci_tbl[] __devinitdata = {
175 {PCI_DEVICE(SXG_VENDOR_ID, SXG_DEVICE_ID)},
179 MODULE_DEVICE_TABLE(pci, sxg_pci_tbl);
181 static inline void sxg_reg32_write(void __iomem *reg, u32 value, bool flush)
188 static inline void sxg_reg64_write(struct adapter_t *adapter, void __iomem *reg,
191 u32 value_high = (u32) (value >> 32);
192 u32 value_low = (u32) (value & 0x00000000FFFFFFFF);
195 spin_lock_irqsave(&adapter->Bit64RegLock, flags);
196 writel(value_high, (void __iomem *)(&adapter->UcodeRegs[cpu].Upper));
197 writel(value_low, reg);
198 spin_unlock_irqrestore(&adapter->Bit64RegLock, flags);
201 static void sxg_init_driver(void)
203 if (sxg_first_init) {
204 DBG_ERROR("sxg: %s sxg_first_init set jiffies[%lx]\n",
207 spin_lock_init(&sxg_global.driver_lock);
211 static void sxg_dbg_macaddrs(struct adapter_t *adapter)
213 DBG_ERROR(" (%s) curr %2.2X:%2.2X:%2.2X:%2.2X:%2.2X:%2.2X\n",
214 adapter->netdev->name, adapter->currmacaddr[0],
215 adapter->currmacaddr[1], adapter->currmacaddr[2],
216 adapter->currmacaddr[3], adapter->currmacaddr[4],
217 adapter->currmacaddr[5]);
218 DBG_ERROR(" (%s) mac %2.2X:%2.2X:%2.2X:%2.2X:%2.2X:%2.2X\n",
219 adapter->netdev->name, adapter->macaddr[0],
220 adapter->macaddr[1], adapter->macaddr[2],
221 adapter->macaddr[3], adapter->macaddr[4],
222 adapter->macaddr[5]);
227 static struct sxg_driver SxgDriver;
230 static struct sxg_trace_buffer LSxgTraceBuffer;
232 static struct sxg_trace_buffer *SxgTraceBuffer = NULL;
235 * sxg_download_microcode
237 * Download Microcode to Sahara adapter
240 * adapter - A pointer to our adapter structure
241 * UcodeSel - microcode file selection
246 static bool sxg_download_microcode(struct adapter_t *adapter,
247 enum SXG_UCODE_SEL UcodeSel)
249 struct sxg_hw_regs *HwRegs = adapter->HwRegs;
252 u32 *Instruction = NULL;
253 u32 BaseAddress, AddressOffset, Address;
259 u32 sectionStart[16];
261 SXG_TRACE(TRACE_SXG, SxgTraceBuffer, TRACE_NOISY, "DnldUcod",
263 DBG_ERROR("sxg: %s ENTER\n", __func__);
266 case SXG_UCODE_SAHARA: /* Sahara operational ucode */
267 numSections = SNumSections;
268 for (i = 0; i < numSections; i++) {
269 sectionSize[i] = SSectionSize[i];
270 sectionStart[i] = SSectionStart[i];
274 printk(KERN_ERR KBUILD_MODNAME
275 ": Woah, big error with the microcode!\n");
279 DBG_ERROR("sxg: RESET THE CARD\n");
280 /* First, reset the card */
281 WRITE_REG(HwRegs->Reset, 0xDEAD, FLUSH);
284 * Download each section of the microcode as specified in
285 * its download file. The *download.c file is generated using
286 * the saharaobjtoc facility which converts the metastep .obj
287 * file to a .c file which contains a two dimentional array.
289 for (Section = 0; Section < numSections; Section++) {
290 DBG_ERROR("sxg: SECTION # %d\n", Section);
292 case SXG_UCODE_SAHARA:
293 Instruction = (u32 *) & SaharaUCode[Section][0];
299 BaseAddress = sectionStart[Section];
300 /* Size in instructions */
301 ThisSectionSize = sectionSize[Section] / 12;
302 for (AddressOffset = 0; AddressOffset < ThisSectionSize;
304 Address = BaseAddress + AddressOffset;
305 ASSERT((Address & ~MICROCODE_ADDRESS_MASK) == 0);
306 /* Write instruction bits 31 - 0 */
307 WRITE_REG(HwRegs->UcodeDataLow, *Instruction, FLUSH);
308 /* Write instruction bits 63-32 */
309 WRITE_REG(HwRegs->UcodeDataMiddle, *(Instruction + 1),
311 /* Write instruction bits 95-64 */
312 WRITE_REG(HwRegs->UcodeDataHigh, *(Instruction + 2),
314 /* Write instruction address with the WRITE bit set */
315 WRITE_REG(HwRegs->UcodeAddr,
316 (Address | MICROCODE_ADDRESS_WRITE), FLUSH);
318 * Sahara bug in the ucode download logic - the write to DataLow
319 * for the next instruction could get corrupted. To avoid this,
320 * write to DataLow again for this instruction (which may get
321 * corrupted, but it doesn't matter), then increment the address
322 * and write the data for the next instruction to DataLow. That
323 * write should succeed.
325 WRITE_REG(HwRegs->UcodeDataLow, *Instruction, TRUE);
326 /* Advance 3 u32S to start of next instruction */
331 * Now repeat the entire operation reading the instruction back and
332 * checking for parity errors
334 for (Section = 0; Section < numSections; Section++) {
335 DBG_ERROR("sxg: check SECTION # %d\n", Section);
337 case SXG_UCODE_SAHARA:
338 Instruction = (u32 *) & SaharaUCode[Section][0];
344 BaseAddress = sectionStart[Section];
345 /* Size in instructions */
346 ThisSectionSize = sectionSize[Section] / 12;
347 for (AddressOffset = 0; AddressOffset < ThisSectionSize;
349 Address = BaseAddress + AddressOffset;
350 /* Write the address with the READ bit set */
351 WRITE_REG(HwRegs->UcodeAddr,
352 (Address | MICROCODE_ADDRESS_READ), FLUSH);
353 /* Read it back and check parity bit. */
354 READ_REG(HwRegs->UcodeAddr, ValueRead);
355 if (ValueRead & MICROCODE_ADDRESS_PARITY) {
356 DBG_ERROR("sxg: %s PARITY ERROR\n",
359 return FALSE; /* Parity error */
361 ASSERT((ValueRead & MICROCODE_ADDRESS_MASK) == Address);
362 /* Read the instruction back and compare */
363 READ_REG(HwRegs->UcodeDataLow, ValueRead);
364 if (ValueRead != *Instruction) {
365 DBG_ERROR("sxg: %s MISCOMPARE LOW\n",
367 return FALSE; /* Miscompare */
369 READ_REG(HwRegs->UcodeDataMiddle, ValueRead);
370 if (ValueRead != *(Instruction + 1)) {
371 DBG_ERROR("sxg: %s MISCOMPARE MIDDLE\n",
373 return FALSE; /* Miscompare */
375 READ_REG(HwRegs->UcodeDataHigh, ValueRead);
376 if (ValueRead != *(Instruction + 2)) {
377 DBG_ERROR("sxg: %s MISCOMPARE HIGH\n",
379 return FALSE; /* Miscompare */
381 /* Advance 3 u32S to start of next instruction */
386 /* Everything OK, Go. */
387 WRITE_REG(HwRegs->UcodeAddr, MICROCODE_ADDRESS_GO, FLUSH);
390 * Poll the CardUp register to wait for microcode to initialize
391 * Give up after 10,000 attemps (500ms).
393 for (i = 0; i < 10000; i++) {
395 READ_REG(adapter->UcodeRegs[0].CardUp, ValueRead);
396 if (ValueRead == 0xCAFE) {
397 DBG_ERROR("sxg: %s BOO YA 0xCAFE\n", __func__);
402 DBG_ERROR("sxg: %s TIMEOUT\n", __func__);
404 return FALSE; /* Timeout */
407 * Now write the LoadSync register. This is used to
408 * synchronize with the card so it can scribble on the memory
409 * that contained 0xCAFE from the "CardUp" step above
411 if (UcodeSel == SXG_UCODE_SAHARA) {
412 WRITE_REG(adapter->UcodeRegs[0].LoadSync, 0, FLUSH);
415 SXG_TRACE(TRACE_SXG, SxgTraceBuffer, TRACE_NOISY, "XDnldUcd",
417 DBG_ERROR("sxg: %s EXIT\n", __func__);
423 * sxg_allocate_resources - Allocate memory and locks
426 * adapter - A pointer to our adapter structure
430 static int sxg_allocate_resources(struct adapter_t *adapter)
434 u32 RssIds, IsrCount;
435 /* struct sxg_xmt_ring *XmtRing; */
436 /* struct sxg_rcv_ring *RcvRing; */
438 DBG_ERROR("%s ENTER\n", __func__);
440 SXG_TRACE(TRACE_SXG, SxgTraceBuffer, TRACE_NOISY, "AllocRes",
443 /* Windows tells us how many CPUs it plans to use for */
445 RssIds = SXG_RSS_CPU_COUNT(adapter);
446 IsrCount = adapter->MsiEnabled ? RssIds : 1;
448 DBG_ERROR("%s Setup the spinlocks\n", __func__);
450 /* Allocate spinlocks and initialize listheads first. */
451 spin_lock_init(&adapter->RcvQLock);
452 spin_lock_init(&adapter->SglQLock);
453 spin_lock_init(&adapter->XmtZeroLock);
454 spin_lock_init(&adapter->Bit64RegLock);
455 spin_lock_init(&adapter->AdapterLock);
456 atomic_set(&adapter->pending_allocations, 0);
458 DBG_ERROR("%s Setup the lists\n", __func__);
460 InitializeListHead(&adapter->FreeRcvBuffers);
461 InitializeListHead(&adapter->FreeRcvBlocks);
462 InitializeListHead(&adapter->AllRcvBlocks);
463 InitializeListHead(&adapter->FreeSglBuffers);
464 InitializeListHead(&adapter->AllSglBuffers);
467 * Mark these basic allocations done. This flags essentially
468 * tells the SxgFreeResources routine that it can grab spinlocks
469 * and reference listheads.
471 adapter->BasicAllocations = TRUE;
473 * Main allocation loop. Start with the maximum supported by
474 * the microcode and back off if memory allocation
475 * fails. If we hit a minimum, fail.
479 DBG_ERROR("%s Allocate XmtRings size[%x]\n", __func__,
480 (unsigned int)(sizeof(struct sxg_xmt_ring) * 1));
483 * Start with big items first - receive and transmit rings.
484 * At the moment I'm going to keep the ring size fixed and
485 * adjust the TCBs if we fail. Later we might
486 * consider reducing the ring size as well..
488 adapter->XmtRings = pci_alloc_consistent(adapter->pcidev,
489 sizeof(struct sxg_xmt_ring) *
491 &adapter->PXmtRings);
492 DBG_ERROR("%s XmtRings[%p]\n", __func__, adapter->XmtRings);
494 if (!adapter->XmtRings) {
495 goto per_tcb_allocation_failed;
497 memset(adapter->XmtRings, 0, sizeof(struct sxg_xmt_ring) * 1);
499 DBG_ERROR("%s Allocate RcvRings size[%x]\n", __func__,
500 (unsigned int)(sizeof(struct sxg_rcv_ring) * 1));
502 pci_alloc_consistent(adapter->pcidev,
503 sizeof(struct sxg_rcv_ring) * 1,
504 &adapter->PRcvRings);
505 DBG_ERROR("%s RcvRings[%p]\n", __func__, adapter->RcvRings);
506 if (!adapter->RcvRings) {
507 goto per_tcb_allocation_failed;
509 memset(adapter->RcvRings, 0, sizeof(struct sxg_rcv_ring) * 1);
510 adapter->ucode_stats = kzalloc(sizeof(struct sxg_ucode_stats), GFP_ATOMIC);
511 adapter->pucode_stats = pci_map_single(adapter->pcidev,
512 adapter->ucode_stats,
513 sizeof(struct sxg_ucode_stats),
515 // memset(adapter->ucode_stats, 0, sizeof(struct sxg_ucode_stats));
518 per_tcb_allocation_failed:
519 /* an allocation failed. Free any successful allocations. */
520 if (adapter->XmtRings) {
521 pci_free_consistent(adapter->pcidev,
522 sizeof(struct sxg_xmt_ring) * 1,
525 adapter->XmtRings = NULL;
527 if (adapter->RcvRings) {
528 pci_free_consistent(adapter->pcidev,
529 sizeof(struct sxg_rcv_ring) * 1,
532 adapter->RcvRings = NULL;
534 /* Loop around and try again.... */
535 if (adapter->ucode_stats) {
536 pci_unmap_single(adapter->pcidev,
537 sizeof(struct sxg_ucode_stats),
538 adapter->pucode_stats, PCI_DMA_FROMDEVICE);
539 adapter->ucode_stats = NULL;
544 DBG_ERROR("%s Initialize RCV ZERO and XMT ZERO rings\n", __func__);
545 /* Initialize rcv zero and xmt zero rings */
546 SXG_INITIALIZE_RING(adapter->RcvRingZeroInfo, SXG_RCV_RING_SIZE);
547 SXG_INITIALIZE_RING(adapter->XmtRingZeroInfo, SXG_XMT_RING_SIZE);
549 /* Sanity check receive data structure format */
550 /* ASSERT((adapter->ReceiveBufferSize == SXG_RCV_DATA_BUFFER_SIZE) ||
551 (adapter->ReceiveBufferSize == SXG_RCV_JUMBO_BUFFER_SIZE)); */
552 ASSERT(sizeof(struct sxg_rcv_descriptor_block) ==
553 SXG_RCV_DESCRIPTOR_BLOCK_SIZE);
556 * Allocate receive data buffers. We allocate a block of buffers and
557 * a corresponding descriptor block at once. See sxghw.h:SXG_RCV_BLOCK
559 for (i = 0; i < SXG_INITIAL_RCV_DATA_BUFFERS;
560 i += SXG_RCV_DESCRIPTORS_PER_BLOCK) {
561 sxg_allocate_buffer_memory(adapter,
562 SXG_RCV_BLOCK_SIZE(SXG_RCV_DATA_HDR_SIZE),
563 SXG_BUFFER_TYPE_RCV);
566 * NBL resource allocation can fail in the 'AllocateComplete' routine,
567 * which doesn't return status. Make sure we got the number of buffers
570 if (adapter->FreeRcvBufferCount < SXG_INITIAL_RCV_DATA_BUFFERS) {
571 SXG_TRACE(TRACE_SXG, SxgTraceBuffer, TRACE_NOISY, "XAResF6",
572 adapter, adapter->FreeRcvBufferCount, SXG_MAX_ENTRIES,
574 return (STATUS_RESOURCES);
577 DBG_ERROR("%s Allocate EventRings size[%x]\n", __func__,
578 (unsigned int)(sizeof(struct sxg_event_ring) * RssIds));
580 /* Allocate event queues. */
581 adapter->EventRings = pci_alloc_consistent(adapter->pcidev,
582 sizeof(struct sxg_event_ring) *
584 &adapter->PEventRings);
586 if (!adapter->EventRings) {
587 /* Caller will call SxgFreeAdapter to clean up above
589 SXG_TRACE(TRACE_SXG, SxgTraceBuffer, TRACE_NOISY, "XAResF8",
590 adapter, SXG_MAX_ENTRIES, 0, 0);
591 status = STATUS_RESOURCES;
592 goto per_tcb_allocation_failed;
594 memset(adapter->EventRings, 0, sizeof(struct sxg_event_ring) * RssIds);
596 DBG_ERROR("%s Allocate ISR size[%x]\n", __func__, IsrCount);
598 adapter->Isr = pci_alloc_consistent(adapter->pcidev,
599 IsrCount, &adapter->PIsr);
601 /* Caller will call SxgFreeAdapter to clean up above
603 SXG_TRACE(TRACE_SXG, SxgTraceBuffer, TRACE_NOISY, "XAResF9",
604 adapter, SXG_MAX_ENTRIES, 0, 0);
605 status = STATUS_RESOURCES;
606 goto per_tcb_allocation_failed;
608 memset(adapter->Isr, 0, sizeof(u32) * IsrCount);
610 DBG_ERROR("%s Allocate shared XMT ring zero index location size[%x]\n",
611 __func__, (unsigned int)sizeof(u32));
613 /* Allocate shared XMT ring zero index location */
614 adapter->XmtRingZeroIndex = pci_alloc_consistent(adapter->pcidev,
618 if (!adapter->XmtRingZeroIndex) {
619 SXG_TRACE(TRACE_SXG, SxgTraceBuffer, TRACE_NOISY, "XAResF10",
620 adapter, SXG_MAX_ENTRIES, 0, 0);
621 status = STATUS_RESOURCES;
622 goto per_tcb_allocation_failed;
624 memset(adapter->XmtRingZeroIndex, 0, sizeof(u32));
626 SXG_TRACE(TRACE_SXG, SxgTraceBuffer, TRACE_NOISY, "XAlcResS",
627 adapter, SXG_MAX_ENTRIES, 0, 0);
629 DBG_ERROR("%s EXIT\n", __func__);
630 return (STATUS_SUCCESS);
636 * Set up PCI Configuration space
639 * pcidev - A pointer to our adapter structure
641 static void sxg_config_pci(struct pci_dev *pcidev)
646 pci_read_config_word(pcidev, PCI_COMMAND, &pci_command);
647 DBG_ERROR("sxg: %s PCI command[%4.4x]\n", __func__, pci_command);
648 /* Set the command register */
649 new_command = pci_command | (
650 /* Memory Space Enable */
652 /* Bus master enable */
654 /* Memory write and invalidate */
655 PCI_COMMAND_INVALIDATE |
656 /* Parity error response */
660 /* Fast back-to-back */
661 PCI_COMMAND_FAST_BACK);
662 if (pci_command != new_command) {
663 DBG_ERROR("%s -- Updating PCI COMMAND register %4.4x->%4.4x.\n",
664 __func__, pci_command, new_command);
665 pci_write_config_word(pcidev, PCI_COMMAND, new_command);
671 * @adapter : Pointer to the adapter structure for the card
672 * This function will read the configuration data from EEPROM/FLASH
674 static inline int sxg_read_config(struct adapter_t *adapter)
676 /* struct sxg_config data; */
677 struct sw_cfg_data *data;
679 unsigned long status;
682 data = pci_alloc_consistent(adapter->pcidev,
683 sizeof(struct sw_cfg_data), &p_addr);
686 * We cant get even this much memory. Raise a hell
689 printk(KERN_ERR"%s : Could not allocate memory for reading \
690 EEPROM\n", __FUNCTION__);
694 WRITE_REG(adapter->UcodeRegs[0].ConfigStat, SXG_CFG_TIMEOUT, TRUE);
696 WRITE_REG64(adapter, adapter->UcodeRegs[0].Config, p_addr, 0);
697 for(i=0; i<1000; i++) {
698 READ_REG(adapter->UcodeRegs[0].ConfigStat, status);
699 if (status != SXG_CFG_TIMEOUT)
701 mdelay(1); /* Do we really need this */
705 /* Config read from EEPROM succeeded */
706 case SXG_CFG_LOAD_EEPROM:
707 /* Config read from Flash succeeded */
708 case SXG_CFG_LOAD_FLASH:
709 /* Copy the MAC address to adapter structure */
710 /* TODO: We are not doing the remaining part : FRU,
713 memcpy(adapter->macaddr, data->MacAddr[0].MacAddr,
714 sizeof(struct sxg_config_mac));
716 case SXG_CFG_TIMEOUT:
717 case SXG_CFG_LOAD_INVALID:
718 case SXG_CFG_LOAD_ERROR:
719 default: /* Fix default handler later */
720 printk(KERN_WARNING"%s : We could not read the config \
721 word. Status = %ld\n", __FUNCTION__, status);
724 pci_free_consistent(adapter->pcidev, sizeof(struct sw_cfg_data), data,
726 if (adapter->netdev) {
727 memcpy(adapter->netdev->dev_addr, adapter->currmacaddr, 6);
728 memcpy(adapter->netdev->perm_addr, adapter->currmacaddr, 6);
730 printk("LINSYS : These are the new MAC address\n");
731 sxg_dbg_macaddrs(adapter);
736 static int sxg_entry_probe(struct pci_dev *pcidev,
737 const struct pci_device_id *pci_tbl_entry)
739 static int did_version = 0;
741 struct net_device *netdev;
742 struct adapter_t *adapter;
743 void __iomem *memmapped_ioaddr;
745 ulong mmio_start = 0;
748 DBG_ERROR("sxg: %s 2.6 VERSION ENTER jiffies[%lx] cpu %d\n",
749 __func__, jiffies, smp_processor_id());
751 /* Initialize trace buffer */
753 SxgTraceBuffer = &LSxgTraceBuffer;
754 SXG_TRACE_INIT(SxgTraceBuffer, TRACE_NOISY);
757 sxg_global.dynamic_intagg = dynamic_intagg;
759 err = pci_enable_device(pcidev);
761 DBG_ERROR("Call pci_enable_device(%p) status[%x]\n", pcidev, err);
766 if (sxg_debug > 0 && did_version++ == 0) {
767 printk(KERN_INFO "%s\n", sxg_banner);
768 printk(KERN_INFO "%s\n", SXG_DRV_VERSION);
771 if (!(err = pci_set_dma_mask(pcidev, DMA_64BIT_MASK))) {
772 DBG_ERROR("pci_set_dma_mask(DMA_64BIT_MASK) successful\n");
774 if ((err = pci_set_dma_mask(pcidev, DMA_32BIT_MASK))) {
776 ("No usable DMA configuration, aborting err[%x]\n",
780 DBG_ERROR("pci_set_dma_mask(DMA_32BIT_MASK) successful\n");
783 DBG_ERROR("Call pci_request_regions\n");
785 err = pci_request_regions(pcidev, sxg_driver_name);
787 DBG_ERROR("pci_request_regions FAILED err[%x]\n", err);
791 DBG_ERROR("call pci_set_master\n");
792 pci_set_master(pcidev);
794 DBG_ERROR("call alloc_etherdev\n");
795 netdev = alloc_etherdev(sizeof(struct adapter_t));
798 goto err_out_exit_sxg_probe;
800 DBG_ERROR("alloc_etherdev for slic netdev[%p]\n", netdev);
802 SET_NETDEV_DEV(netdev, &pcidev->dev);
804 pci_set_drvdata(pcidev, netdev);
805 adapter = netdev_priv(netdev);
806 adapter->netdev = netdev;
807 adapter->pcidev = pcidev;
809 mmio_start = pci_resource_start(pcidev, 0);
810 mmio_len = pci_resource_len(pcidev, 0);
812 DBG_ERROR("sxg: call ioremap(mmio_start[%lx], mmio_len[%lx])\n",
813 mmio_start, mmio_len);
815 memmapped_ioaddr = ioremap(mmio_start, mmio_len);
816 DBG_ERROR("sxg: %s MEMMAPPED_IOADDR [%p]\n", __func__,
818 if (!memmapped_ioaddr) {
819 DBG_ERROR("%s cannot remap MMIO region %lx @ %lx\n",
820 __func__, mmio_len, mmio_start);
821 goto err_out_free_mmio_region;
824 DBG_ERROR("sxg: %s found Alacritech SXG PCI, MMIO at %p, start[%lx] \
825 len[%lx], IRQ %d.\n", __func__, memmapped_ioaddr, mmio_start,
826 mmio_len, pcidev->irq);
828 adapter->HwRegs = (void *)memmapped_ioaddr;
829 adapter->base_addr = memmapped_ioaddr;
831 mmio_start = pci_resource_start(pcidev, 2);
832 mmio_len = pci_resource_len(pcidev, 2);
834 DBG_ERROR("sxg: call ioremap(mmio_start[%lx], mmio_len[%lx])\n",
835 mmio_start, mmio_len);
837 memmapped_ioaddr = ioremap(mmio_start, mmio_len);
838 DBG_ERROR("sxg: %s MEMMAPPED_IOADDR [%p]\n", __func__,
840 if (!memmapped_ioaddr) {
841 DBG_ERROR("%s cannot remap MMIO region %lx @ %lx\n",
842 __func__, mmio_len, mmio_start);
843 goto err_out_free_mmio_region;
846 DBG_ERROR("sxg: %s found Alacritech SXG PCI, MMIO at %p, "
847 "start[%lx] len[%lx], IRQ %d.\n", __func__,
848 memmapped_ioaddr, mmio_start, mmio_len, pcidev->irq);
850 adapter->UcodeRegs = (void *)memmapped_ioaddr;
852 adapter->State = SXG_STATE_INITIALIZING;
854 * Maintain a list of all adapters anchored by
855 * the global SxgDriver structure.
857 adapter->Next = SxgDriver.Adapters;
858 SxgDriver.Adapters = adapter;
859 adapter->AdapterID = ++SxgDriver.AdapterID;
861 /* Initialize CRC table used to determine multicast hash */
862 sxg_mcast_init_crc32();
864 adapter->JumboEnabled = FALSE;
865 adapter->RssEnabled = FALSE;
866 if (adapter->JumboEnabled) {
867 adapter->FrameSize = JUMBOMAXFRAME;
868 adapter->ReceiveBufferSize = SXG_RCV_JUMBO_BUFFER_SIZE;
870 adapter->FrameSize = ETHERMAXFRAME;
871 adapter->ReceiveBufferSize = SXG_RCV_DATA_BUFFER_SIZE;
875 * status = SXG_READ_EEPROM(adapter);
881 DBG_ERROR("sxg: %s ENTER sxg_config_pci\n", __func__);
882 sxg_config_pci(pcidev);
883 DBG_ERROR("sxg: %s EXIT sxg_config_pci\n", __func__);
885 DBG_ERROR("sxg: %s ENTER sxg_init_driver\n", __func__);
887 DBG_ERROR("sxg: %s EXIT sxg_init_driver\n", __func__);
889 adapter->vendid = pci_tbl_entry->vendor;
890 adapter->devid = pci_tbl_entry->device;
891 adapter->subsysid = pci_tbl_entry->subdevice;
892 adapter->slotnumber = ((pcidev->devfn >> 3) & 0x1F);
893 adapter->functionnumber = (pcidev->devfn & 0x7);
894 adapter->memorylength = pci_resource_len(pcidev, 0);
895 adapter->irq = pcidev->irq;
896 adapter->next_netdevice = head_netdevice;
897 head_netdevice = netdev;
898 adapter->port = 0; /*adapter->functionnumber; */
900 /* Allocate memory and other resources */
901 DBG_ERROR("sxg: %s ENTER sxg_allocate_resources\n", __func__);
902 status = sxg_allocate_resources(adapter);
903 DBG_ERROR("sxg: %s EXIT sxg_allocate_resources status %x\n",
905 if (status != STATUS_SUCCESS) {
909 DBG_ERROR("sxg: %s ENTER sxg_download_microcode\n", __func__);
910 if (sxg_download_microcode(adapter, SXG_UCODE_SAHARA)) {
911 DBG_ERROR("sxg: %s ENTER sxg_adapter_set_hwaddr\n",
913 sxg_read_config(adapter);
914 sxg_adapter_set_hwaddr(adapter);
916 adapter->state = ADAPT_FAIL;
917 adapter->linkstate = LINK_DOWN;
918 DBG_ERROR("sxg_download_microcode FAILED status[%x]\n", status);
921 netdev->base_addr = (unsigned long)adapter->base_addr;
922 netdev->irq = adapter->irq;
923 netdev->open = sxg_entry_open;
924 netdev->stop = sxg_entry_halt;
925 netdev->hard_start_xmit = sxg_send_packets;
926 netdev->do_ioctl = sxg_ioctl;
928 netdev->set_mac_address = sxg_mac_set_address;
930 netdev->get_stats = sxg_get_stats;
931 netdev->set_multicast_list = sxg_mcast_set_list;
932 SET_ETHTOOL_OPS(netdev, &sxg_nic_ethtool_ops);
934 strcpy(netdev->name, "eth%d");
935 /* strcpy(netdev->name, pci_name(pcidev)); */
936 if ((err = register_netdev(netdev))) {
937 DBG_ERROR("Cannot register net device, aborting. %s\n",
943 ("sxg: %s addr 0x%lx, irq %d, MAC addr \
944 %02X:%02X:%02X:%02X:%02X:%02X\n",
945 netdev->name, netdev->base_addr, pcidev->irq, netdev->dev_addr[0],
946 netdev->dev_addr[1], netdev->dev_addr[2], netdev->dev_addr[3],
947 netdev->dev_addr[4], netdev->dev_addr[5]);
950 ASSERT(status == FALSE);
951 /* sxg_free_adapter(adapter); */
953 DBG_ERROR("sxg: %s EXIT status[%x] jiffies[%lx] cpu %d\n", __func__,
954 status, jiffies, smp_processor_id());
958 iounmap((void *)memmapped_ioaddr);
960 err_out_free_mmio_region:
961 release_mem_region(mmio_start, mmio_len);
963 err_out_exit_sxg_probe:
965 DBG_ERROR("%s EXIT jiffies[%lx] cpu %d\n", __func__, jiffies,
972 * LINE BASE Interrupt routines..
974 * sxg_disable_interrupt
976 * DisableInterrupt Handler
980 * adapter: Our adapter structure
985 static void sxg_disable_interrupt(struct adapter_t *adapter)
987 SXG_TRACE(TRACE_SXG, SxgTraceBuffer, TRACE_NOISY, "DisIntr",
988 adapter, adapter->InterruptsEnabled, 0, 0);
989 /* For now, RSS is disabled with line based interrupts */
990 ASSERT(adapter->RssEnabled == FALSE);
991 ASSERT(adapter->MsiEnabled == FALSE);
992 /* Turn off interrupts by writing to the icr register. */
993 WRITE_REG(adapter->UcodeRegs[0].Icr, SXG_ICR(0, SXG_ICR_DISABLE), TRUE);
995 adapter->InterruptsEnabled = 0;
997 SXG_TRACE(TRACE_SXG, SxgTraceBuffer, TRACE_NOISY, "XDisIntr",
998 adapter, adapter->InterruptsEnabled, 0, 0);
1002 * sxg_enable_interrupt
1004 * EnableInterrupt Handler
1008 * adapter: Our adapter structure
1013 static void sxg_enable_interrupt(struct adapter_t *adapter)
1015 SXG_TRACE(TRACE_SXG, SxgTraceBuffer, TRACE_NOISY, "EnIntr",
1016 adapter, adapter->InterruptsEnabled, 0, 0);
1017 /* For now, RSS is disabled with line based interrupts */
1018 ASSERT(adapter->RssEnabled == FALSE);
1019 ASSERT(adapter->MsiEnabled == FALSE);
1020 /* Turn on interrupts by writing to the icr register. */
1021 WRITE_REG(adapter->UcodeRegs[0].Icr, SXG_ICR(0, SXG_ICR_ENABLE), TRUE);
1023 adapter->InterruptsEnabled = 1;
1025 SXG_TRACE(TRACE_SXG, SxgTraceBuffer, TRACE_NOISY, "XEnIntr",
1030 * sxg_isr - Process an line-based interrupt
1033 * Context - Our adapter structure
1034 * QueueDefault - Output parameter to queue to default CPU
1035 * TargetCpus - Output bitmap to schedule DPC's
1037 * Return Value: TRUE if our interrupt
1039 static irqreturn_t sxg_isr(int irq, void *dev_id)
1041 struct net_device *dev = (struct net_device *) dev_id;
1042 struct adapter_t *adapter = (struct adapter_t *) netdev_priv(dev);
1044 if(adapter->state != ADAPT_UP)
1046 adapter->Stats.NumInts++;
1047 if (adapter->Isr[0] == 0) {
1049 * The SLIC driver used to experience a number of spurious
1050 * interrupts due to the delay associated with the masking of
1051 * the interrupt (we'd bounce back in here). If we see that
1052 * again with Sahara,add a READ_REG of the Icr register after
1053 * the WRITE_REG below.
1055 adapter->Stats.FalseInts++;
1059 * Move the Isr contents and clear the value in
1060 * shared memory, and mask interrupts
1062 adapter->IsrCopy[0] = adapter->Isr[0];
1063 adapter->Isr[0] = 0;
1064 WRITE_REG(adapter->UcodeRegs[0].Icr, SXG_ICR(0, SXG_ICR_MASK), TRUE);
1065 /* ASSERT(adapter->IsrDpcsPending == 0); */
1066 #if XXXTODO /* RSS Stuff */
1068 * If RSS is enabled and the ISR specifies SXG_ISR_EVENT, then
1069 * schedule DPC's based on event queues.
1071 if (adapter->RssEnabled && (adapter->IsrCopy[0] & SXG_ISR_EVENT)) {
1073 i < adapter->RssSystemInfo->ProcessorInfo.RssCpuCount;
1075 struct sxg_event_ring *EventRing =
1076 &adapter->EventRings[i];
1077 struct sxg_event *Event =
1078 &EventRing->Ring[adapter->NextEvent[i]];
1080 adapter->RssSystemInfo->RssIdToCpu[i];
1081 if (Event->Status & EVENT_STATUS_VALID) {
1082 adapter->IsrDpcsPending++;
1083 CpuMask |= (1 << Cpu);
1088 * Now, either schedule the CPUs specified by the CpuMask,
1092 *QueueDefault = FALSE;
1094 adapter->IsrDpcsPending = 1;
1095 *QueueDefault = TRUE;
1097 *TargetCpus = CpuMask;
1099 /* There are no DPCs in Linux, so call the handler now */
1100 sxg_handle_interrupt(adapter);
1105 int debug_inthandler = 0;
1107 static void sxg_handle_interrupt(struct adapter_t *adapter)
1109 /* unsigned char RssId = 0; */
1112 if (++debug_inthandler < 20) {
1113 DBG_ERROR("Enter sxg_handle_interrupt ISR[%x]\n",
1114 adapter->IsrCopy[0]);
1116 SXG_TRACE(TRACE_SXG, SxgTraceBuffer, TRACE_NOISY, "HndlIntr",
1117 adapter, adapter->IsrCopy[0], 0, 0);
1118 /* For now, RSS is disabled with line based interrupts */
1119 ASSERT(adapter->RssEnabled == FALSE);
1120 ASSERT(adapter->MsiEnabled == FALSE);
1121 ASSERT(adapter->IsrCopy[0]);
1123 /* Always process the event queue. */
1124 sxg_process_event_queue(adapter,
1125 (adapter->RssEnabled ? /*RssId */ 0 : 0));
1127 #if XXXTODO /* RSS stuff */
1128 if (--adapter->IsrDpcsPending) {
1130 ASSERT(adapter->RssEnabled);
1131 SXG_TRACE(TRACE_SXG, SxgTraceBuffer, TRACE_NOISY, "DPCsPend",
1136 /* Last (or only) DPC processes the ISR and clears the interrupt. */
1137 NewIsr = sxg_process_isr(adapter, 0);
1138 /* Reenable interrupts */
1139 adapter->IsrCopy[0] = 0;
1140 SXG_TRACE(TRACE_SXG, SxgTraceBuffer, TRACE_NOISY, "ClearIsr",
1141 adapter, NewIsr, 0, 0);
1143 if (debug_inthandler < 20) {
1145 ("Exit sxg_handle_interrupt2 after enabling interrupt\n");
1148 WRITE_REG(adapter->UcodeRegs[0].Isr, NewIsr, TRUE);
1150 SXG_TRACE(TRACE_SXG, SxgTraceBuffer, TRACE_NOISY, "XHndlInt",
1155 * sxg_process_isr - Process an interrupt. Called from the line-based and
1156 * message based interrupt DPC routines
1159 * adapter - Our adapter structure
1160 * Queue - The ISR that needs processing
1165 static int sxg_process_isr(struct adapter_t *adapter, u32 MessageId)
1167 u32 Isr = adapter->IsrCopy[MessageId];
1170 SXG_TRACE(TRACE_SXG, SxgTraceBuffer, TRACE_NOISY, "ProcIsr",
1171 adapter, Isr, 0, 0);
1173 DBG_ERROR("%s: Entering with %d ISR value\n", __FUNCTION__, Isr);
1175 if (Isr & SXG_ISR_ERR) {
1176 if (Isr & SXG_ISR_PDQF) {
1177 adapter->Stats.PdqFull++;
1178 DBG_ERROR("%s: SXG_ISR_ERR PDQF!!\n", __func__);
1180 /* No host buffer */
1181 if (Isr & SXG_ISR_RMISS) {
1183 * There is a bunch of code in the SLIC driver which
1184 * attempts to process more receive events per DPC
1185 * if we start to fall behind. We'll probablyd
1186 * need to do something similar here, but hold
1187 * off for now. I don't want to make the code more
1188 * complicated than strictly needed.
1190 adapter->Stats.RcvNoBuffer++;
1191 adapter->stats.rx_missed_errors++;
1192 if (adapter->Stats.RcvNoBuffer < 5) {
1193 DBG_ERROR("%s: SXG_ISR_ERR RMISS!!\n",
1198 if (Isr & SXG_ISR_DEAD) {
1200 * Set aside the crash info and set the adapter state
1203 adapter->CrashCpu = (unsigned char)
1204 ((Isr & SXG_ISR_CPU) >> SXG_ISR_CPU_SHIFT);
1205 adapter->CrashLocation = (ushort) (Isr & SXG_ISR_CRASH);
1206 adapter->Dead = TRUE;
1207 DBG_ERROR("%s: ISR_DEAD %x, CPU: %d\n", __func__,
1208 adapter->CrashLocation, adapter->CrashCpu);
1210 /* Event ring full */
1211 if (Isr & SXG_ISR_ERFULL) {
1213 * Same issue as RMISS, really. This means the
1214 * host is falling behind the card. Need to increase
1215 * event ring size, process more events per interrupt,
1216 * and/or reduce/remove interrupt aggregation.
1218 adapter->Stats.EventRingFull++;
1219 DBG_ERROR("%s: SXG_ISR_ERR EVENT RING FULL!!\n",
1222 /* Transmit drop - no DRAM buffers or XMT error */
1223 if (Isr & SXG_ISR_XDROP) {
1224 adapter->Stats.XmtDrops++;
1225 adapter->Stats.XmtErrors++;
1226 DBG_ERROR("%s: SXG_ISR_ERR XDROP!!\n", __func__);
1229 /* Slowpath send completions */
1230 if (Isr & SXG_ISR_SPSEND) {
1231 sxg_complete_slow_send(adapter, 1);
1234 if (Isr & SXG_ISR_UPC) {
1235 /* Maybe change when debug is added.. */
1236 ASSERT(adapter->DumpCmdRunning);
1237 adapter->DumpCmdRunning = FALSE;
1240 if (Isr & SXG_ISR_LINK) {
1241 sxg_link_event(adapter);
1243 /* Debug - breakpoint hit */
1244 if (Isr & SXG_ISR_BREAK) {
1246 * At the moment AGDB isn't written to support interactive
1247 * debug sessions. When it is, this interrupt will be used to
1248 * signal AGDB that it has hit a breakpoint. For now, ASSERT.
1252 /* Heartbeat response */
1253 if (Isr & SXG_ISR_PING) {
1254 adapter->PingOutstanding = FALSE;
1256 SXG_TRACE(TRACE_SXG, SxgTraceBuffer, TRACE_NOISY, "XProcIsr",
1257 adapter, Isr, NewIsr, 0);
1263 * sxg_process_event_queue - Process our event queue
1266 * - adapter - Adapter structure
1267 * - RssId - The event queue requiring processing
1272 static u32 sxg_process_event_queue(struct adapter_t *adapter, u32 RssId)
1274 struct sxg_event_ring *EventRing = &adapter->EventRings[RssId];
1275 struct sxg_event *Event = &EventRing->Ring[adapter->NextEvent[RssId]];
1276 u32 EventsProcessed = 0, Batches = 0;
1278 struct sk_buff *skb;
1279 #ifdef LINUX_HANDLES_RCV_INDICATION_LISTS
1280 struct sk_buff *prev_skb = NULL;
1281 struct sk_buff *IndicationList[SXG_RCV_ARRAYSIZE];
1283 struct sxg_rcv_data_buffer_hdr *RcvDataBufferHdr;
1285 u32 ReturnStatus = 0;
1287 ASSERT((adapter->State == SXG_STATE_RUNNING) ||
1288 (adapter->State == SXG_STATE_PAUSING) ||
1289 (adapter->State == SXG_STATE_PAUSED) ||
1290 (adapter->State == SXG_STATE_HALTING));
1292 * We may still have unprocessed events on the queue if
1293 * the card crashed. Don't process them.
1295 if (adapter->Dead) {
1299 * In theory there should only be a single processor that
1300 * accesses this queue, and only at interrupt-DPC time. So/
1301 * we shouldn't need a lock for any of this.
1303 while (Event->Status & EVENT_STATUS_VALID) {
1304 SXG_TRACE(TRACE_SXG, SxgTraceBuffer, TRACE_NOISY, "Event",
1305 Event, Event->Code, Event->Status,
1306 adapter->NextEvent);
1307 switch (Event->Code) {
1308 case EVENT_CODE_BUFFERS:
1309 /* struct sxg_ring_info Head & Tail == unsigned char */
1310 ASSERT(!(Event->CommandIndex & 0xFF00));
1311 sxg_complete_descriptor_blocks(adapter,
1312 Event->CommandIndex);
1314 case EVENT_CODE_SLOWRCV:
1315 --adapter->RcvBuffersOnCard;
1316 if ((skb = sxg_slow_receive(adapter, Event))) {
1318 #ifdef LINUX_HANDLES_RCV_INDICATION_LISTS
1319 /* Add it to our indication list */
1320 SXG_ADD_RCV_PACKET(adapter, skb, prev_skb,
1321 IndicationList, num_skbs);
1323 * Linux, we just pass up each skb to the
1324 * protocol above at this point, there is no
1325 * capability of an indication list.
1328 /* CHECK skb_pull(skb, INIC_RCVBUF_HEADSIZE); */
1329 /* (rcvbuf->length & IRHDDR_FLEN_MSK); */
1330 rx_bytes = Event->Length;
1331 adapter->stats.rx_packets++;
1332 adapter->stats.rx_bytes += rx_bytes;
1333 #if SXG_OFFLOAD_IP_CHECKSUM
1334 skb->ip_summed = CHECKSUM_UNNECESSARY;
1336 skb->dev = adapter->netdev;
1342 DBG_ERROR("%s: ERROR Invalid EventCode %d\n",
1343 __func__, Event->Code);
1347 * See if we need to restock card receive buffers.
1348 * There are two things to note here:
1349 * First - This test is not SMP safe. The
1350 * adapter->BuffersOnCard field is protected via atomic
1351 * interlocked calls, but we do not protect it with respect
1352 * to these tests. The only way to do that is with a lock,
1353 * and I don't want to grab a lock every time we adjust the
1354 * BuffersOnCard count. Instead, we allow the buffer
1355 * replenishment to be off once in a while. The worst that
1356 * can happen is the card is given on more-or-less descriptor
1357 * block than the arbitrary value we've chosen. No big deal
1358 * In short DO NOT ADD A LOCK HERE, OR WHERE RcvBuffersOnCard
1360 * Second - We expect this test to rarely
1361 * evaluate to true. We attempt to refill descriptor blocks
1362 * as they are returned to us (sxg_complete_descriptor_blocks)
1363 * so The only time this should evaluate to true is when
1364 * sxg_complete_descriptor_blocks failed to allocate
1367 if (adapter->RcvBuffersOnCard < SXG_RCV_DATA_BUFFERS) {
1368 sxg_stock_rcv_buffers(adapter);
1371 * It's more efficient to just set this to zero.
1372 * But clearing the top bit saves potential debug info...
1374 Event->Status &= ~EVENT_STATUS_VALID;
1375 /* Advance to the next event */
1376 SXG_ADVANCE_INDEX(adapter->NextEvent[RssId], EVENT_RING_SIZE);
1377 Event = &EventRing->Ring[adapter->NextEvent[RssId]];
1379 if (EventsProcessed == EVENT_RING_BATCH) {
1380 /* Release a batch of events back to the card */
1381 WRITE_REG(adapter->UcodeRegs[RssId].EventRelease,
1382 EVENT_RING_BATCH, FALSE);
1383 EventsProcessed = 0;
1385 * If we've processed our batch limit, break out of the
1386 * loop and return SXG_ISR_EVENT to arrange for us to
1389 if (Batches++ == EVENT_BATCH_LIMIT) {
1390 SXG_TRACE(TRACE_SXG, SxgTraceBuffer,
1391 TRACE_NOISY, "EvtLimit", Batches,
1392 adapter->NextEvent, 0, 0);
1393 ReturnStatus = SXG_ISR_EVENT;
1398 #ifdef LINUX_HANDLES_RCV_INDICATION_LISTS
1399 /* Indicate any received dumb-nic frames */
1400 SXG_INDICATE_PACKETS(adapter, IndicationList, num_skbs);
1402 /* Release events back to the card. */
1403 if (EventsProcessed) {
1404 WRITE_REG(adapter->UcodeRegs[RssId].EventRelease,
1405 EventsProcessed, FALSE);
1407 SXG_TRACE(TRACE_SXG, SxgTraceBuffer, TRACE_NOISY, "XPrcEvnt",
1408 Batches, EventsProcessed, adapter->NextEvent, num_skbs);
1410 return (ReturnStatus);
1414 * sxg_complete_slow_send - Complete slowpath or dumb-nic sends
1417 * adapter - A pointer to our adapter structure
1418 * irq_context - An integer to denote if we are in interrupt context
1422 static void sxg_complete_slow_send(struct adapter_t *adapter, int irq_context)
1424 struct sxg_xmt_ring *XmtRing = &adapter->XmtRings[0];
1425 struct sxg_ring_info *XmtRingInfo = &adapter->XmtRingZeroInfo;
1427 struct sxg_cmd *XmtCmd;
1428 unsigned long flags;
1429 unsigned long sgl_flags;
1430 unsigned int processed_count = 0;
1433 * NOTE - This lock is dropped and regrabbed in this loop.
1434 * This means two different processors can both be running/
1435 * through this loop. Be *very* careful.
1438 if(!spin_trylock(&adapter->XmtZeroLock))
1442 spin_lock_irqsave(&adapter->XmtZeroLock, flags);
1444 SXG_TRACE(TRACE_SXG, SxgTraceBuffer, TRACE_NOISY, "CmpSnds",
1445 adapter, XmtRingInfo->Head, XmtRingInfo->Tail, 0);
1447 while ((XmtRingInfo->Tail != *adapter->XmtRingZeroIndex)
1448 && processed_count++ < SXG_COMPLETE_SLOW_SEND_LIMIT) {
1450 * Locate the current Cmd (ring descriptor entry), and
1451 * associated SGL, and advance the tail
1453 SXG_RETURN_CMD(XmtRing, XmtRingInfo, XmtCmd, ContextType);
1454 ASSERT(ContextType);
1455 SXG_TRACE(TRACE_SXG, SxgTraceBuffer, TRACE_NOISY, "CmpSnd",
1456 XmtRingInfo->Head, XmtRingInfo->Tail, XmtCmd, 0);
1457 /* Clear the SGL field. */
1460 switch (*ContextType) {
1463 struct sk_buff *skb;
1464 struct sxg_scatter_gather *SxgSgl =
1465 (struct sxg_scatter_gather *)ContextType;
1466 dma64_addr_t FirstSgeAddress;
1469 /* Dumb-nic send. Command context is the dumb-nic SGL */
1470 skb = (struct sk_buff *)ContextType;
1471 skb = SxgSgl->DumbPacket;
1472 FirstSgeAddress = XmtCmd->Buffer.FirstSgeAddress;
1473 FirstSgeLength = XmtCmd->Buffer.FirstSgeLength;
1474 /* Complete the send */
1475 SXG_TRACE(TRACE_SXG, SxgTraceBuffer,
1476 TRACE_IMPORTANT, "DmSndCmp", skb, 0,
1478 ASSERT(adapter->Stats.XmtQLen);
1479 adapter->Stats.XmtQLen--;/* within XmtZeroLock */
1480 adapter->Stats.XmtOk++;
1482 * Now drop the lock and complete the send
1483 * back to Microsoft. We need to drop the lock
1484 * because Microsoft can come back with a
1485 * chimney send, which results in a double trip
1489 spin_unlock(&adapter->XmtZeroLock);
1491 spin_unlock_irqrestore(
1492 &adapter->XmtZeroLock, flags);
1494 SxgSgl->DumbPacket = NULL;
1495 SXG_COMPLETE_DUMB_SEND(adapter, skb,
1498 SXG_FREE_SGL_BUFFER(adapter, SxgSgl, NULL,
1500 /* and reacquire.. */
1502 if(!spin_trylock(&adapter->XmtZeroLock))
1506 spin_lock_irqsave(&adapter->XmtZeroLock, flags);
1514 spin_unlock(&adapter->XmtZeroLock);
1516 spin_unlock_irqrestore(&adapter->XmtZeroLock, flags);
1518 SXG_TRACE(TRACE_SXG, SxgTraceBuffer, TRACE_NOISY, "CmpSnd",
1519 adapter, XmtRingInfo->Head, XmtRingInfo->Tail, 0);
1526 * adapter - A pointer to our adapter structure
1527 * Event - Receive event
1531 static struct sk_buff *sxg_slow_receive(struct adapter_t *adapter,
1532 struct sxg_event *Event)
1534 u32 BufferSize = adapter->ReceiveBufferSize;
1535 struct sxg_rcv_data_buffer_hdr *RcvDataBufferHdr;
1536 struct sk_buff *Packet;
1537 static int read_counter = 0;
1539 RcvDataBufferHdr = (struct sxg_rcv_data_buffer_hdr *) Event->HostHandle;
1540 if(read_counter++ & 0x100)
1542 sxg_collect_statistics(adapter);
1545 ASSERT(RcvDataBufferHdr);
1546 ASSERT(RcvDataBufferHdr->State == SXG_BUFFER_ONCARD);
1547 SXG_TRACE(TRACE_SXG, SxgTraceBuffer, TRACE_IMPORTANT, "SlowRcv", Event,
1548 RcvDataBufferHdr, RcvDataBufferHdr->State,
1549 /*RcvDataBufferHdr->VirtualAddress*/ 0);
1550 /* Drop rcv frames in non-running state */
1551 switch (adapter->State) {
1552 case SXG_STATE_RUNNING:
1554 case SXG_STATE_PAUSING:
1555 case SXG_STATE_PAUSED:
1556 case SXG_STATE_HALTING:
1564 * memcpy(SXG_RECEIVE_DATA_LOCATION(RcvDataBufferHdr),
1565 * RcvDataBufferHdr->VirtualAddress, Event->Length);
1568 /* Change buffer state to UPSTREAM */
1569 RcvDataBufferHdr->State = SXG_BUFFER_UPSTREAM;
1570 if (Event->Status & EVENT_STATUS_RCVERR) {
1571 SXG_TRACE(TRACE_SXG, SxgTraceBuffer, TRACE_NOISY, "RcvError",
1572 Event, Event->Status, Event->HostHandle, 0);
1573 /* XXXTODO - Remove this print later */
1574 DBG_ERROR("SXG: Receive error %x\n", *(u32 *)
1575 SXG_RECEIVE_DATA_LOCATION(RcvDataBufferHdr));
1576 sxg_process_rcv_error(adapter, *(u32 *)
1577 SXG_RECEIVE_DATA_LOCATION
1578 (RcvDataBufferHdr));
1581 #if XXXTODO /* VLAN stuff */
1582 /* If there's a VLAN tag, extract it and validate it */
1583 if (((struct ether_header *)
1584 (SXG_RECEIVE_DATA_LOCATION(RcvDataBufferHdr)))->EtherType
1585 == ETHERTYPE_VLAN) {
1586 if (SxgExtractVlanHeader(adapter, RcvDataBufferHdr, Event) !=
1588 SXG_TRACE(TRACE_SXG, SxgTraceBuffer, TRACE_NOISY,
1590 SXG_RECEIVE_DATA_LOCATION(RcvDataBufferHdr),
1596 /* Dumb-nic frame. See if it passes our mac filter and update stats */
1599 * ASK if (!sxg_mac_filter(adapter,
1600 * SXG_RECEIVE_DATA_LOCATION(RcvDataBufferHdr),
1602 * SXG_TRACE(TRACE_SXG, SxgTraceBuffer, TRACE_NOISY, "RcvFiltr",
1603 * Event, SXG_RECEIVE_DATA_LOCATION(RcvDataBufferHdr),
1604 * Event->Length, 0);
1609 Packet = RcvDataBufferHdr->SxgDumbRcvPacket;
1610 SXG_ADJUST_RCV_PACKET(Packet, RcvDataBufferHdr, Event);
1611 Packet->protocol = eth_type_trans(Packet, adapter->netdev);
1613 SXG_TRACE(TRACE_SXG, SxgTraceBuffer, TRACE_IMPORTANT, "DumbRcv",
1614 RcvDataBufferHdr, Packet, Event->Length, 0);
1615 /* Lastly adjust the receive packet length. */
1616 RcvDataBufferHdr->SxgDumbRcvPacket = NULL;
1617 RcvDataBufferHdr->PhysicalAddress = NULL;
1618 SXG_ALLOCATE_RCV_PACKET(adapter, RcvDataBufferHdr, BufferSize);
1619 if (RcvDataBufferHdr->skb)
1621 spin_lock(&adapter->RcvQLock);
1622 SXG_FREE_RCV_DATA_BUFFER(adapter, RcvDataBufferHdr);
1623 // adapter->RcvBuffersOnCard ++;
1624 spin_unlock(&adapter->RcvQLock);
1629 SXG_TRACE(TRACE_SXG, SxgTraceBuffer, TRACE_NOISY, "DropRcv",
1630 RcvDataBufferHdr, Event->Length, 0, 0);
1631 adapter->Stats.RcvDiscards++;
1632 spin_lock(&adapter->RcvQLock);
1633 SXG_FREE_RCV_DATA_BUFFER(adapter, RcvDataBufferHdr);
1634 spin_unlock(&adapter->RcvQLock);
1639 * sxg_process_rcv_error - process receive error and update
1643 * adapter - Adapter structure
1644 * ErrorStatus - 4-byte receive error status
1646 * Return Value : None
1648 static void sxg_process_rcv_error(struct adapter_t *adapter, u32 ErrorStatus)
1652 adapter->Stats.RcvErrors++;
1654 if (ErrorStatus & SXG_RCV_STATUS_TRANSPORT_ERROR) {
1655 Error = ErrorStatus & SXG_RCV_STATUS_TRANSPORT_MASK;
1657 case SXG_RCV_STATUS_TRANSPORT_CSUM:
1658 adapter->Stats.TransportCsum++;
1660 case SXG_RCV_STATUS_TRANSPORT_UFLOW:
1661 adapter->Stats.TransportUflow++;
1663 case SXG_RCV_STATUS_TRANSPORT_HDRLEN:
1664 adapter->Stats.TransportHdrLen++;
1668 if (ErrorStatus & SXG_RCV_STATUS_NETWORK_ERROR) {
1669 Error = ErrorStatus & SXG_RCV_STATUS_NETWORK_MASK;
1671 case SXG_RCV_STATUS_NETWORK_CSUM:
1672 adapter->Stats.NetworkCsum++;
1674 case SXG_RCV_STATUS_NETWORK_UFLOW:
1675 adapter->Stats.NetworkUflow++;
1677 case SXG_RCV_STATUS_NETWORK_HDRLEN:
1678 adapter->Stats.NetworkHdrLen++;
1682 if (ErrorStatus & SXG_RCV_STATUS_PARITY) {
1683 adapter->Stats.Parity++;
1685 if (ErrorStatus & SXG_RCV_STATUS_LINK_ERROR) {
1686 Error = ErrorStatus & SXG_RCV_STATUS_LINK_MASK;
1688 case SXG_RCV_STATUS_LINK_PARITY:
1689 adapter->Stats.LinkParity++;
1691 case SXG_RCV_STATUS_LINK_EARLY:
1692 adapter->Stats.LinkEarly++;
1694 case SXG_RCV_STATUS_LINK_BUFOFLOW:
1695 adapter->Stats.LinkBufOflow++;
1697 case SXG_RCV_STATUS_LINK_CODE:
1698 adapter->Stats.LinkCode++;
1700 case SXG_RCV_STATUS_LINK_DRIBBLE:
1701 adapter->Stats.LinkDribble++;
1703 case SXG_RCV_STATUS_LINK_CRC:
1704 adapter->Stats.LinkCrc++;
1706 case SXG_RCV_STATUS_LINK_OFLOW:
1707 adapter->Stats.LinkOflow++;
1709 case SXG_RCV_STATUS_LINK_UFLOW:
1710 adapter->Stats.LinkUflow++;
1720 * adapter - Adapter structure
1721 * pether - Ethernet header
1722 * length - Frame length
1724 * Return Value : TRUE if the frame is to be allowed
1726 static bool sxg_mac_filter(struct adapter_t *adapter,
1727 struct ether_header *EtherHdr, ushort length)
1731 if (SXG_MULTICAST_PACKET(EtherHdr)) {
1732 if (SXG_BROADCAST_PACKET(EtherHdr)) {
1734 if (adapter->MacFilter & MAC_BCAST) {
1735 adapter->Stats.DumbRcvBcastPkts++;
1736 adapter->Stats.DumbRcvBcastBytes += length;
1737 adapter->Stats.DumbRcvPkts++;
1738 adapter->Stats.DumbRcvBytes += length;
1743 if (adapter->MacFilter & MAC_ALLMCAST) {
1744 adapter->Stats.DumbRcvMcastPkts++;
1745 adapter->Stats.DumbRcvMcastBytes += length;
1746 adapter->Stats.DumbRcvPkts++;
1747 adapter->Stats.DumbRcvBytes += length;
1750 if (adapter->MacFilter & MAC_MCAST) {
1751 struct sxg_multicast_address *MulticastAddrs =
1752 adapter->MulticastAddrs;
1753 while (MulticastAddrs) {
1754 ETHER_EQ_ADDR(MulticastAddrs->Address,
1755 EtherHdr->ether_dhost,
1761 DumbRcvMcastBytes += length;
1762 adapter->Stats.DumbRcvPkts++;
1763 adapter->Stats.DumbRcvBytes +=
1767 MulticastAddrs = MulticastAddrs->Next;
1771 } else if (adapter->MacFilter & MAC_DIRECTED) {
1773 * Not broadcast or multicast. Must be directed at us or
1774 * the card is in promiscuous mode. Either way, consider it
1775 * ours if MAC_DIRECTED is set
1777 adapter->Stats.DumbRcvUcastPkts++;
1778 adapter->Stats.DumbRcvUcastBytes += length;
1779 adapter->Stats.DumbRcvPkts++;
1780 adapter->Stats.DumbRcvBytes += length;
1783 if (adapter->MacFilter & MAC_PROMISC) {
1784 /* Whatever it is, keep it. */
1785 adapter->Stats.DumbRcvPkts++;
1786 adapter->Stats.DumbRcvBytes += length;
1789 adapter->Stats.RcvDiscards++;
1793 static int sxg_register_interrupt(struct adapter_t *adapter)
1795 if (!adapter->intrregistered) {
1799 ("sxg: %s AllocAdaptRsrcs adapter[%p] dev->irq[%x] %x\n",
1800 __func__, adapter, adapter->netdev->irq, NR_IRQS);
1802 spin_unlock_irqrestore(&sxg_global.driver_lock,
1805 retval = request_irq(adapter->netdev->irq,
1808 adapter->netdev->name, adapter->netdev);
1810 spin_lock_irqsave(&sxg_global.driver_lock, sxg_global.flags);
1813 DBG_ERROR("sxg: request_irq (%s) FAILED [%x]\n",
1814 adapter->netdev->name, retval);
1817 adapter->intrregistered = 1;
1818 adapter->IntRegistered = TRUE;
1819 /* Disable RSS with line-based interrupts */
1820 adapter->MsiEnabled = FALSE;
1821 adapter->RssEnabled = FALSE;
1822 DBG_ERROR("sxg: %s AllocAdaptRsrcs adapter[%p] dev->irq[%x]\n",
1823 __func__, adapter, adapter->netdev->irq);
1825 return (STATUS_SUCCESS);
1828 static void sxg_deregister_interrupt(struct adapter_t *adapter)
1830 DBG_ERROR("sxg: %s ENTER adapter[%p]\n", __func__, adapter);
1832 slic_init_cleanup(adapter);
1834 memset(&adapter->stats, 0, sizeof(struct net_device_stats));
1835 adapter->error_interrupts = 0;
1836 adapter->rcv_interrupts = 0;
1837 adapter->xmit_interrupts = 0;
1838 adapter->linkevent_interrupts = 0;
1839 adapter->upr_interrupts = 0;
1840 adapter->num_isrs = 0;
1841 adapter->xmit_completes = 0;
1842 adapter->rcv_broadcasts = 0;
1843 adapter->rcv_multicasts = 0;
1844 adapter->rcv_unicasts = 0;
1845 DBG_ERROR("sxg: %s EXIT\n", __func__);
1851 * Perform initialization of our slic interface.
1854 static int sxg_if_init(struct adapter_t *adapter)
1856 struct net_device *dev = adapter->netdev;
1859 DBG_ERROR("sxg: %s (%s) ENTER states[%d:%d] flags[%x]\n",
1860 __func__, adapter->netdev->name,
1862 adapter->linkstate, dev->flags);
1864 /* adapter should be down at this point */
1865 if (adapter->state != ADAPT_DOWN) {
1866 DBG_ERROR("sxg_if_init adapter->state != ADAPT_DOWN\n");
1869 ASSERT(adapter->linkstate == LINK_DOWN);
1871 adapter->devflags_prev = dev->flags;
1872 adapter->macopts = MAC_DIRECTED;
1874 DBG_ERROR("sxg: %s (%s) Set MAC options: ", __func__,
1875 adapter->netdev->name);
1876 if (dev->flags & IFF_BROADCAST) {
1877 adapter->macopts |= MAC_BCAST;
1878 DBG_ERROR("BCAST ");
1880 if (dev->flags & IFF_PROMISC) {
1881 adapter->macopts |= MAC_PROMISC;
1882 DBG_ERROR("PROMISC ");
1884 if (dev->flags & IFF_ALLMULTI) {
1885 adapter->macopts |= MAC_ALLMCAST;
1886 DBG_ERROR("ALL_MCAST ");
1888 if (dev->flags & IFF_MULTICAST) {
1889 adapter->macopts |= MAC_MCAST;
1890 DBG_ERROR("MCAST ");
1894 status = sxg_register_interrupt(adapter);
1895 if (status != STATUS_SUCCESS) {
1896 DBG_ERROR("sxg_if_init: sxg_register_interrupt FAILED %x\n",
1898 sxg_deregister_interrupt(adapter);
1902 adapter->state = ADAPT_UP;
1904 /* clear any pending events, then enable interrupts */
1905 DBG_ERROR("sxg: %s ENABLE interrupts(slic)\n", __func__);
1907 return (STATUS_SUCCESS);
1910 static int sxg_entry_open(struct net_device *dev)
1912 struct adapter_t *adapter = (struct adapter_t *) netdev_priv(dev);
1916 DBG_ERROR("sxg: %s adapter->activated[%d]\n", __func__,
1917 adapter->activated);
1919 ("sxg: %s (%s): [jiffies[%lx] cpu %d] dev[%p] adapt[%p] port[%d]\n",
1920 __func__, adapter->netdev->name, jiffies, smp_processor_id(),
1921 adapter->netdev, adapter, adapter->port);
1923 netif_stop_queue(adapter->netdev);
1925 spin_lock_irqsave(&sxg_global.driver_lock, sxg_global.flags);
1926 if (!adapter->activated) {
1927 sxg_global.num_sxg_ports_active++;
1928 adapter->activated = 1;
1930 /* Initialize the adapter */
1931 DBG_ERROR("sxg: %s ENTER sxg_initialize_adapter\n", __func__);
1932 status = sxg_initialize_adapter(adapter);
1933 DBG_ERROR("sxg: %s EXIT sxg_initialize_adapter status[%x]\n",
1936 if (status == STATUS_SUCCESS) {
1937 DBG_ERROR("sxg: %s ENTER sxg_if_init\n", __func__);
1938 status = sxg_if_init(adapter);
1939 DBG_ERROR("sxg: %s EXIT sxg_if_init status[%x]\n", __func__,
1943 if (status != STATUS_SUCCESS) {
1944 if (adapter->activated) {
1945 sxg_global.num_sxg_ports_active--;
1946 adapter->activated = 0;
1948 spin_unlock_irqrestore(&sxg_global.driver_lock,
1952 DBG_ERROR("sxg: %s ENABLE ALL INTERRUPTS\n", __func__);
1954 /* Enable interrupts */
1955 SXG_ENABLE_ALL_INTERRUPTS(adapter);
1957 DBG_ERROR("sxg: %s EXIT\n", __func__);
1959 spin_unlock_irqrestore(&sxg_global.driver_lock, sxg_global.flags);
1960 return STATUS_SUCCESS;
1963 static void __devexit sxg_entry_remove(struct pci_dev *pcidev)
1965 struct net_device *dev = pci_get_drvdata(pcidev);
1967 unsigned int mmio_len = 0;
1968 struct adapter_t *adapter = (struct adapter_t *) netdev_priv(dev);
1970 flush_scheduled_work();
1972 /* Deallocate Resources */
1973 unregister_netdev(dev);
1974 sxg_free_resources(adapter);
1977 DBG_ERROR("sxg: %s ENTER dev[%p] adapter[%p]\n", __func__, dev,
1980 mmio_start = pci_resource_start(pcidev, 0);
1981 mmio_len = pci_resource_len(pcidev, 0);
1983 DBG_ERROR("sxg: %s rel_region(0) start[%x] len[%x]\n", __func__,
1984 mmio_start, mmio_len);
1985 release_mem_region(mmio_start, mmio_len);
1987 mmio_start = pci_resource_start(pcidev, 2);
1988 mmio_len = pci_resource_len(pcidev, 2);
1990 DBG_ERROR("sxg: %s rel_region(2) start[%x] len[%x]\n", __FUNCTION__,
1991 mmio_start, mmio_len);
1992 release_mem_region(mmio_start, mmio_len);
1994 pci_disable_device(pcidev);
1996 DBG_ERROR("sxg: %s deallocate device\n", __func__);
1998 DBG_ERROR("sxg: %s EXIT\n", __func__);
2001 static int sxg_entry_halt(struct net_device *dev)
2003 struct adapter_t *adapter = (struct adapter_t *) netdev_priv(dev);
2005 spin_lock_irqsave(&sxg_global.driver_lock, sxg_global.flags);
2006 DBG_ERROR("sxg: %s (%s) ENTER\n", __func__, dev->name);
2008 netif_stop_queue(adapter->netdev);
2009 adapter->state = ADAPT_DOWN;
2010 adapter->linkstate = LINK_DOWN;
2011 adapter->devflags_prev = 0;
2012 DBG_ERROR("sxg: %s (%s) set adapter[%p] state to ADAPT_DOWN(%d)\n",
2013 __func__, dev->name, adapter, adapter->state);
2015 DBG_ERROR("sxg: %s (%s) EXIT\n", __func__, dev->name);
2016 DBG_ERROR("sxg: %s EXIT\n", __func__);
2018 /* Disable interrupts */
2019 SXG_DISABLE_ALL_INTERRUPTS(adapter);
2021 spin_unlock_irqrestore(&sxg_global.driver_lock, sxg_global.flags);
2023 sxg_deregister_interrupt(adapter);
2024 return (STATUS_SUCCESS);
2027 static int sxg_ioctl(struct net_device *dev, struct ifreq *rq, int cmd)
2030 /* DBG_ERROR("sxg: %s cmd[%x] rq[%p] dev[%p]\n", __func__, cmd, rq, dev);*/
2032 case SIOCSLICSETINTAGG:
2034 /* struct adapter_t *adapter = (struct adapter_t *)
2040 if (copy_from_user(data, rq->ifr_data, 28)) {
2041 DBG_ERROR("copy_from_user FAILED getting \
2047 "%s: set interrupt aggregation to %d\n",
2053 /* DBG_ERROR("sxg: %s UNSUPPORTED[%x]\n", __func__, cmd); */
2059 #define NORMAL_ETHFRAME 0
2062 * sxg_send_packets - Send a skb packet
2065 * skb - The packet to send
2066 * dev - Our linux net device that refs our adapter
2069 * 0 regardless of outcome XXXTODO refer to e1000 driver
2071 static int sxg_send_packets(struct sk_buff *skb, struct net_device *dev)
2073 struct adapter_t *adapter = (struct adapter_t *) netdev_priv(dev);
2074 u32 status = STATUS_SUCCESS;
2077 * DBG_ERROR("sxg: %s ENTER sxg_send_packets skb[%p]\n", __FUNCTION__,
2080 printk("ASK:sxg_send_packets: skb[%p]\n", skb);
2082 /* Check the adapter state */
2083 switch (adapter->State) {
2084 case SXG_STATE_INITIALIZING:
2085 case SXG_STATE_HALTED:
2086 case SXG_STATE_SHUTDOWN:
2087 ASSERT(0); /* unexpected */
2089 case SXG_STATE_RESETTING:
2090 case SXG_STATE_SLEEP:
2091 case SXG_STATE_BOOTDIAG:
2092 case SXG_STATE_DIAG:
2093 case SXG_STATE_HALTING:
2094 status = STATUS_FAILURE;
2096 case SXG_STATE_RUNNING:
2097 if (adapter->LinkState != SXG_LINK_UP) {
2098 status = STATUS_FAILURE;
2103 status = STATUS_FAILURE;
2105 if (status != STATUS_SUCCESS) {
2109 status = sxg_transmit_packet(adapter, skb);
2110 if (status == STATUS_SUCCESS) {
2115 /* reject & complete all the packets if they cant be sent */
2116 if (status != STATUS_SUCCESS) {
2118 /* sxg_send_packets_fail(adapter, skb, status); */
2120 SXG_DROP_DUMB_SEND(adapter, skb);
2121 adapter->stats.tx_dropped++;
2122 return NETDEV_TX_BUSY;
2125 DBG_ERROR("sxg: %s EXIT sxg_send_packets status[%x]\n", __func__,
2129 return NETDEV_TX_OK;
2133 * sxg_transmit_packet
2135 * This function transmits a single packet.
2138 * adapter - Pointer to our adapter structure
2139 * skb - The packet to be sent
2141 * Return - STATUS of send
2143 static int sxg_transmit_packet(struct adapter_t *adapter, struct sk_buff *skb)
2145 struct sxg_x64_sgl *pSgl;
2146 struct sxg_scatter_gather *SxgSgl;
2147 unsigned long sgl_flags;
2148 /* void *SglBuffer; */
2149 /* u32 SglBufferLength; */
2152 * The vast majority of work is done in the shared
2153 * sxg_dumb_sgl routine.
2155 SXG_TRACE(TRACE_SXG, SxgTraceBuffer, TRACE_NOISY, "DumbSend",
2156 adapter, skb, 0, 0);
2158 /* Allocate a SGL buffer */
2159 SXG_GET_SGL_BUFFER(adapter, SxgSgl, 0);
2161 adapter->Stats.NoSglBuf++;
2162 adapter->Stats.XmtErrors++;
2163 SXG_TRACE(TRACE_SXG, SxgTraceBuffer, TRACE_NOISY, "SndPktF1",
2164 adapter, skb, 0, 0);
2165 return (STATUS_RESOURCES);
2167 ASSERT(SxgSgl->adapter == adapter);
2168 /*SglBuffer = SXG_SGL_BUFFER(SxgSgl);
2169 SglBufferLength = SXG_SGL_BUF_SIZE; */
2170 SxgSgl->VlanTag.VlanTci = 0;
2171 SxgSgl->VlanTag.VlanTpid = 0;
2172 SxgSgl->Type = SXG_SGL_DUMB;
2173 SxgSgl->DumbPacket = skb;
2176 /* Call the common sxg_dumb_sgl routine to complete the send. */
2177 return (sxg_dumb_sgl(pSgl, SxgSgl));
2185 * SxgSgl - struct sxg_scatter_gather
2188 * Status of send operation.
2190 static int sxg_dumb_sgl(struct sxg_x64_sgl *pSgl,
2191 struct sxg_scatter_gather *SxgSgl)
2193 struct adapter_t *adapter = SxgSgl->adapter;
2194 struct sk_buff *skb = SxgSgl->DumbPacket;
2195 /* For now, all dumb-nic sends go on RSS queue zero */
2196 struct sxg_xmt_ring *XmtRing = &adapter->XmtRings[0];
2197 struct sxg_ring_info *XmtRingInfo = &adapter->XmtRingZeroInfo;
2198 struct sxg_cmd *XmtCmd = NULL;
2199 /* u32 Index = 0; */
2200 u32 DataLength = skb->len;
2201 /* unsigned int BufLen; */
2202 /* u32 SglOffset; */
2204 unsigned long flags;
2206 SXG_TRACE(TRACE_SXG, SxgTraceBuffer, TRACE_NOISY, "DumbSgl",
2207 pSgl, SxgSgl, 0, 0);
2209 /* Set aside a pointer to the sgl */
2210 SxgSgl->pSgl = pSgl;
2212 /* Sanity check that our SGL format is as we expect. */
2213 ASSERT(sizeof(struct sxg_x64_sge) == sizeof(struct sxg_x64_sge));
2214 /* Shouldn't be a vlan tag on this frame */
2215 ASSERT(SxgSgl->VlanTag.VlanTci == 0);
2216 ASSERT(SxgSgl->VlanTag.VlanTpid == 0);
2219 * From here below we work with the SGL placed in our
2223 SxgSgl->Sgl.NumberOfElements = 1;
2225 /* Grab the spinlock and acquire a command */
2226 spin_lock_irqsave(&adapter->XmtZeroLock, flags);
2227 SXG_GET_CMD(XmtRing, XmtRingInfo, XmtCmd, SxgSgl);
2228 if (XmtCmd == NULL) {
2230 * Call sxg_complete_slow_send to see if we can
2231 * free up any XmtRingZero entries and then try again
2234 spin_unlock_irqrestore(&adapter->XmtZeroLock, flags);
2235 sxg_complete_slow_send(adapter, 0);
2236 spin_lock_irqsave(&adapter->XmtZeroLock, flags);
2237 SXG_GET_CMD(XmtRing, XmtRingInfo, XmtCmd, SxgSgl);
2238 if (XmtCmd == NULL) {
2239 adapter->Stats.XmtZeroFull++;
2243 SXG_TRACE(TRACE_SXG, SxgTraceBuffer, TRACE_NOISY, "DumbCmd",
2244 XmtCmd, XmtRingInfo->Head, XmtRingInfo->Tail, 0);
2246 adapter->Stats.DumbXmtPkts++;
2247 adapter->stats.tx_packets++;
2248 adapter->Stats.DumbXmtBytes += DataLength;
2249 adapter->stats.tx_bytes += DataLength;
2250 #if XXXTODO /* Stats stuff */
2251 if (SXG_MULTICAST_PACKET(EtherHdr)) {
2252 if (SXG_BROADCAST_PACKET(EtherHdr)) {
2253 adapter->Stats.DumbXmtBcastPkts++;
2254 adapter->Stats.DumbXmtBcastBytes += DataLength;
2256 adapter->Stats.DumbXmtMcastPkts++;
2257 adapter->Stats.DumbXmtMcastBytes += DataLength;
2260 adapter->Stats.DumbXmtUcastPkts++;
2261 adapter->Stats.DumbXmtUcastBytes += DataLength;
2265 * Fill in the command
2266 * Copy out the first SGE to the command and adjust for offset
2268 phys_addr = pci_map_single(adapter->pcidev, skb->data, skb->len,
2270 memset(XmtCmd, '\0', sizeof(*XmtCmd));
2271 XmtCmd->Buffer.FirstSgeAddress = phys_addr;
2272 XmtCmd->Buffer.FirstSgeLength = DataLength;
2273 XmtCmd->Buffer.SgeOffset = 0;
2274 XmtCmd->Buffer.TotalLength = DataLength;
2275 XmtCmd->SgEntries = 1;
2278 * Advance transmit cmd descripter by 1.
2279 * NOTE - See comments in SxgTcpOutput where we write
2280 * to the XmtCmd register regarding CPU ID values and/or
2281 * multiple commands.
2283 WRITE_REG(adapter->UcodeRegs[0].XmtCmd, 1, TRUE);
2284 adapter->Stats.XmtQLen++; /* Stats within lock */
2285 spin_unlock_irqrestore(&adapter->XmtZeroLock, flags);
2286 SXG_TRACE(TRACE_SXG, SxgTraceBuffer, TRACE_NOISY, "XDumSgl2",
2287 XmtCmd, pSgl, SxgSgl, 0);
2288 return STATUS_SUCCESS;
2292 * NOTE - Only jump to this label AFTER grabbing the
2293 * XmtZeroLock, and DO NOT DROP IT between the
2294 * command allocation and the following abort.
2297 SXG_ABORT_CMD(XmtRingInfo);
2299 spin_unlock_irqrestore(&adapter->XmtZeroLock, flags);
2300 return STATUS_FAILURE;
2304 * Jump to this label if failure occurs before the
2305 * XmtZeroLock is grabbed
2307 adapter->Stats.XmtErrors++;
2308 adapter->stats.tx_errors++;
2309 SXG_TRACE(TRACE_SXG, SxgTraceBuffer, TRACE_IMPORTANT, "DumSGFal",
2310 pSgl, SxgSgl, XmtRingInfo->Head, XmtRingInfo->Tail);
2311 /* SxgSgl->DumbPacket is the skb */
2312 // SXG_COMPLETE_DUMB_SEND(adapter, SxgSgl->DumbPacket);
2316 * Link management functions
2318 * sxg_initialize_link - Initialize the link stuff
2321 * adapter - A pointer to our adapter structure
2326 static int sxg_initialize_link(struct adapter_t *adapter)
2328 struct sxg_hw_regs *HwRegs = adapter->HwRegs;
2334 SXG_TRACE(TRACE_SXG, SxgTraceBuffer, TRACE_NOISY, "InitLink",
2337 /* Reset PHY and XGXS module */
2338 WRITE_REG(HwRegs->LinkStatus, LS_SERDES_POWER_DOWN, TRUE);
2340 /* Reset transmit configuration register */
2341 WRITE_REG(HwRegs->XmtConfig, XMT_CONFIG_RESET, TRUE);
2343 /* Reset receive configuration register */
2344 WRITE_REG(HwRegs->RcvConfig, RCV_CONFIG_RESET, TRUE);
2346 /* Reset all MAC modules */
2347 WRITE_REG(HwRegs->MacConfig0, AXGMAC_CFG0_SUB_RESET, TRUE);
2351 * XXXTODO - This assumes the MAC address (0a:0b:0c:0d:0e:0f)
2352 * is stored with the first nibble (0a) in the byte 0
2353 * of the Mac address. Possibly reverse?
2355 Value = *(u32 *) adapter->macaddr;
2356 WRITE_REG(HwRegs->LinkAddress0Low, Value, TRUE);
2357 /* also write the MAC address to the MAC. Endian is reversed. */
2358 WRITE_REG(HwRegs->MacAddressLow, ntohl(Value), TRUE);
2359 Value = (*(u16 *) & adapter->macaddr[4] & 0x0000FFFF);
2360 WRITE_REG(HwRegs->LinkAddress0High, Value | LINK_ADDRESS_ENABLE, TRUE);
2361 /* endian swap for the MAC (put high bytes in bits [31:16], swapped) */
2362 Value = ntohl(Value);
2363 WRITE_REG(HwRegs->MacAddressHigh, Value, TRUE);
2364 /* Link address 1 */
2365 WRITE_REG(HwRegs->LinkAddress1Low, 0, TRUE);
2366 WRITE_REG(HwRegs->LinkAddress1High, 0, TRUE);
2367 /* Link address 2 */
2368 WRITE_REG(HwRegs->LinkAddress2Low, 0, TRUE);
2369 WRITE_REG(HwRegs->LinkAddress2High, 0, TRUE);
2370 /* Link address 3 */
2371 WRITE_REG(HwRegs->LinkAddress3Low, 0, TRUE);
2372 WRITE_REG(HwRegs->LinkAddress3High, 0, TRUE);
2374 /* Enable MAC modules */
2375 WRITE_REG(HwRegs->MacConfig0, 0, TRUE);
2378 WRITE_REG(HwRegs->MacConfig1, (
2379 /* Allow sending of pause */
2380 AXGMAC_CFG1_XMT_PAUSE |
2382 AXGMAC_CFG1_XMT_EN |
2383 /* Enable detection of pause */
2384 AXGMAC_CFG1_RCV_PAUSE |
2385 /* Enable receive */
2386 AXGMAC_CFG1_RCV_EN |
2387 /* short frame detection */
2388 AXGMAC_CFG1_SHORT_ASSERT |
2389 /* Verify frame length */
2390 AXGMAC_CFG1_CHECK_LEN |
2392 AXGMAC_CFG1_GEN_FCS |
2393 /* Pad frames to 64 bytes */
2394 AXGMAC_CFG1_PAD_64),
2397 /* Set AXGMAC max frame length if jumbo. Not needed for standard MTU */
2398 if (adapter->JumboEnabled) {
2399 WRITE_REG(HwRegs->MacMaxFrameLen, AXGMAC_MAXFRAME_JUMBO, TRUE);
2402 * AMIIM Configuration Register -
2403 * The value placed in the AXGMAC_AMIIM_CFG_HALF_CLOCK portion
2404 * (bottom bits) of this register is used to determine the MDC frequency
2405 * as specified in the A-XGMAC Design Document. This value must not be
2406 * zero. The following value (62 or 0x3E) is based on our MAC transmit
2407 * clock frequency (MTCLK) of 312.5 MHz. Given a maximum MDIO clock
2408 * frequency of 2.5 MHz (see the PHY spec), we get:
2409 * 312.5/(2*(X+1)) < 2.5 ==> X = 62.
2410 * This value happens to be the default value for this register, so we
2411 * really don't have to do this.
2413 WRITE_REG(HwRegs->MacAmiimConfig, 0x0000003E, TRUE);
2415 /* Power up and enable PHY and XAUI/XGXS/Serdes logic */
2416 WRITE_REG(HwRegs->LinkStatus,
2419 LS_XGXS_CTL | LS_PHY_CLK_EN | LS_ATTN_ALARM), TRUE);
2420 DBG_ERROR("After Power Up and enable PHY in sxg_initialize_link\n");
2423 * Per information given by Aeluros, wait 100 ms after removing reset.
2424 * It's not enough to wait for the self-clearing reset bit in reg 0 to
2429 /* Verify the PHY has come up by checking that the Reset bit has
2432 status = sxg_read_mdio_reg(adapter,
2433 MIIM_DEV_PHY_PMA, /* PHY PMA/PMD module */
2434 PHY_PMA_CONTROL1, /* PMA/PMD control register */
2436 DBG_ERROR("After sxg_read_mdio_reg Value[%x] fail=%x\n", Value,
2437 (Value & PMA_CONTROL1_RESET));
2438 if (status != STATUS_SUCCESS)
2439 return (STATUS_FAILURE);
2440 if (Value & PMA_CONTROL1_RESET) /* reset complete if bit is 0 */
2441 return (STATUS_FAILURE);
2443 /* The SERDES should be initialized by now - confirm */
2444 READ_REG(HwRegs->LinkStatus, Value);
2445 if (Value & LS_SERDES_DOWN) /* verify SERDES is initialized */
2446 return (STATUS_FAILURE);
2448 /* The XAUI link should also be up - confirm */
2449 if (!(Value & LS_XAUI_LINK_UP)) /* verify XAUI link is up */
2450 return (STATUS_FAILURE);
2452 /* Initialize the PHY */
2453 status = sxg_phy_init(adapter);
2454 if (status != STATUS_SUCCESS)
2455 return (STATUS_FAILURE);
2457 /* Enable the Link Alarm */
2459 /* MIIM_DEV_PHY_PMA - PHY PMA/PMD module
2460 * LASI_CONTROL - LASI control register
2461 * LASI_CTL_LS_ALARM_ENABLE - enable link alarm bit
2463 status = sxg_write_mdio_reg(adapter, MIIM_DEV_PHY_PMA,
2465 LASI_CTL_LS_ALARM_ENABLE);
2466 if (status != STATUS_SUCCESS)
2467 return (STATUS_FAILURE);
2469 /* XXXTODO - temporary - verify bit is set */
2471 /* MIIM_DEV_PHY_PMA - PHY PMA/PMD module
2472 * LASI_CONTROL - LASI control register
2474 status = sxg_read_mdio_reg(adapter, MIIM_DEV_PHY_PMA,
2478 if (status != STATUS_SUCCESS)
2479 return (STATUS_FAILURE);
2480 if (!(Value & LASI_CTL_LS_ALARM_ENABLE)) {
2481 DBG_ERROR("Error! LASI Control Alarm Enable bit not set!\n");
2483 /* Enable receive */
2484 MaxFrame = adapter->JumboEnabled ? JUMBOMAXFRAME : ETHERMAXFRAME;
2485 ConfigData = (RCV_CONFIG_ENABLE |
2486 RCV_CONFIG_ENPARSE |
2488 RCV_CONFIG_RCVPAUSE |
2491 RCV_CONFIG_HASH_16 |
2492 RCV_CONFIG_SOCKET | RCV_CONFIG_BUFSIZE(MaxFrame));
2493 WRITE_REG(HwRegs->RcvConfig, ConfigData, TRUE);
2495 WRITE_REG(HwRegs->XmtConfig, XMT_CONFIG_ENABLE, TRUE);
2497 /* Mark the link as down. We'll get a link event when it comes up. */
2498 sxg_link_state(adapter, SXG_LINK_DOWN);
2500 SXG_TRACE(TRACE_SXG, SxgTraceBuffer, TRACE_NOISY, "XInitLnk",
2502 return (STATUS_SUCCESS);
2506 * sxg_phy_init - Initialize the PHY
2509 * adapter - A pointer to our adapter structure
2514 static int sxg_phy_init(struct adapter_t *adapter)
2517 struct phy_ucode *p;
2520 DBG_ERROR("ENTER %s\n", __func__);
2522 /* MIIM_DEV_PHY_PMA - PHY PMA/PMD module
2523 * 0xC205 - PHY ID register (?)
2524 * &Value - XXXTODO - add def
2526 status = sxg_read_mdio_reg(adapter, MIIM_DEV_PHY_PMA,
2529 if (status != STATUS_SUCCESS)
2530 return (STATUS_FAILURE);
2532 if (Value == 0x0012) {
2533 /* 0x0012 == AEL2005C PHY(?) - XXXTODO - add def */
2534 DBG_ERROR("AEL2005C PHY detected. Downloading PHY \
2537 /* Initialize AEL2005C PHY and download PHY microcode */
2538 for (p = PhyUcode; p->Addr != 0xFFFF; p++) {
2540 /* if address == 0, data == sleep time in ms */
2543 /* write the given data to the specified address */
2544 status = sxg_write_mdio_reg(adapter,
2550 if (status != STATUS_SUCCESS)
2551 return (STATUS_FAILURE);
2555 DBG_ERROR("EXIT %s\n", __func__);
2557 return (STATUS_SUCCESS);
2561 * sxg_link_event - Process a link event notification from the card
2564 * adapter - A pointer to our adapter structure
2569 static void sxg_link_event(struct adapter_t *adapter)
2571 struct sxg_hw_regs *HwRegs = adapter->HwRegs;
2572 enum SXG_LINK_STATE LinkState;
2576 SXG_TRACE(TRACE_SXG, SxgTraceBuffer, TRACE_NOISY, "LinkEvnt",
2578 DBG_ERROR("ENTER %s\n", __func__);
2580 /* Check the Link Status register. We should have a Link Alarm. */
2581 READ_REG(HwRegs->LinkStatus, Value);
2582 if (Value & LS_LINK_ALARM) {
2584 * We got a Link Status alarm. First, pause to let the
2585 * link state settle (it can bounce a number of times)
2589 /* Now clear the alarm by reading the LASI status register. */
2590 /* MIIM_DEV_PHY_PMA - PHY PMA/PMD module */
2591 status = sxg_read_mdio_reg(adapter, MIIM_DEV_PHY_PMA,
2592 /* LASI status register */
2595 if (status != STATUS_SUCCESS) {
2596 DBG_ERROR("Error reading LASI Status MDIO register!\n");
2597 sxg_link_state(adapter, SXG_LINK_DOWN);
2600 ASSERT(Value & LASI_STATUS_LS_ALARM);
2602 /* Now get and set the link state */
2603 LinkState = sxg_get_link_state(adapter);
2604 sxg_link_state(adapter, LinkState);
2605 DBG_ERROR("SXG: Link Alarm occurred. Link is %s\n",
2606 ((LinkState == SXG_LINK_UP) ? "UP" : "DOWN"));
2609 * XXXTODO - Assuming Link Attention is only being generated
2610 * for the Link Alarm pin (and not for a XAUI Link Status change)
2611 * , then it's impossible to get here. Yet we've gotten here
2612 * twice (under extreme conditions - bouncing the link up and
2613 * down many times a second). Needs further investigation.
2615 DBG_ERROR("SXG: sxg_link_event: Can't get here!\n");
2616 DBG_ERROR("SXG: Link Status == 0x%08X.\n", Value);
2619 DBG_ERROR("EXIT %s\n", __func__);
2624 * sxg_get_link_state - Determine if the link is up or down
2627 * adapter - A pointer to our adapter structure
2632 static enum SXG_LINK_STATE sxg_get_link_state(struct adapter_t *adapter)
2637 DBG_ERROR("ENTER %s\n", __func__);
2639 SXG_TRACE(TRACE_SXG, SxgTraceBuffer, TRACE_NOISY, "GetLink",
2643 * Per the Xenpak spec (and the IEEE 10Gb spec?), the link is up if
2644 * the following 3 bits (from 3 different MDIO registers) are all true.
2647 /* MIIM_DEV_PHY_PMA - PHY PMA/PMD module */
2648 status = sxg_read_mdio_reg(adapter, MIIM_DEV_PHY_PMA,
2649 /* PMA/PMD Receive Signal Detect register */
2652 if (status != STATUS_SUCCESS)
2655 /* If PMA/PMD receive signal detect is 0, then the link is down */
2656 if (!(Value & PMA_RCV_DETECT))
2657 return (SXG_LINK_DOWN);
2659 /* MIIM_DEV_PHY_PCS - PHY PCS module */
2660 status = sxg_read_mdio_reg(adapter, MIIM_DEV_PHY_PCS,
2661 /* PCS 10GBASE-R Status 1 register */
2662 PHY_PCS_10G_STATUS1,
2664 if (status != STATUS_SUCCESS)
2667 /* If PCS is not locked to receive blocks, then the link is down */
2668 if (!(Value & PCS_10B_BLOCK_LOCK))
2669 return (SXG_LINK_DOWN);
2671 status = sxg_read_mdio_reg(adapter, MIIM_DEV_PHY_XS,/* PHY XS module */
2672 /* XS Lane Status register */
2675 if (status != STATUS_SUCCESS)
2678 /* If XS transmit lanes are not aligned, then the link is down */
2679 if (!(Value & XS_LANE_ALIGN))
2680 return (SXG_LINK_DOWN);
2682 /* All 3 bits are true, so the link is up */
2683 DBG_ERROR("EXIT %s\n", __func__);
2685 return (SXG_LINK_UP);
2688 /* An error occurred reading an MDIO register. This shouldn't happen. */
2689 DBG_ERROR("Error reading an MDIO register!\n");
2691 return (SXG_LINK_DOWN);
2694 static void sxg_indicate_link_state(struct adapter_t *adapter,
2695 enum SXG_LINK_STATE LinkState)
2697 if (adapter->LinkState == SXG_LINK_UP) {
2698 DBG_ERROR("%s: LINK now UP, call netif_start_queue\n",
2700 netif_start_queue(adapter->netdev);
2702 DBG_ERROR("%s: LINK now DOWN, call netif_stop_queue\n",
2704 netif_stop_queue(adapter->netdev);
2709 * sxg_link_state - Set the link state and if necessary, indicate.
2710 * This routine the central point of processing for all link state changes.
2711 * Nothing else in the driver should alter the link state or perform
2712 * link state indications
2715 * adapter - A pointer to our adapter structure
2716 * LinkState - The link state
2721 static void sxg_link_state(struct adapter_t *adapter,
2722 enum SXG_LINK_STATE LinkState)
2724 SXG_TRACE(TRACE_SXG, SxgTraceBuffer, TRACE_IMPORTANT, "LnkINDCT",
2725 adapter, LinkState, adapter->LinkState, adapter->State);
2727 DBG_ERROR("ENTER %s\n", __func__);
2730 * Hold the adapter lock during this routine. Maybe move
2731 * the lock to the caller.
2733 /* IMP TODO : Check if we can survive without taking this lock */
2734 // spin_lock(&adapter->AdapterLock);
2735 if (LinkState == adapter->LinkState) {
2736 /* Nothing changed.. */
2737 // spin_unlock(&adapter->AdapterLock);
2738 DBG_ERROR("EXIT #0 %s. Link status = %d\n",
2739 __func__, LinkState);
2742 /* Save the adapter state */
2743 adapter->LinkState = LinkState;
2745 /* Drop the lock and indicate link state */
2746 // spin_unlock(&adapter->AdapterLock);
2747 DBG_ERROR("EXIT #1 %s\n", __func__);
2749 sxg_indicate_link_state(adapter, LinkState);
2753 * sxg_write_mdio_reg - Write to a register on the MDIO bus
2756 * adapter - A pointer to our adapter structure
2757 * DevAddr - MDIO device number being addressed
2758 * RegAddr - register address for the specified MDIO device
2759 * Value - value to write to the MDIO register
2764 static int sxg_write_mdio_reg(struct adapter_t *adapter,
2765 u32 DevAddr, u32 RegAddr, u32 Value)
2767 struct sxg_hw_regs *HwRegs = adapter->HwRegs;
2768 /* Address operation (written to MIIM field reg) */
2770 /* Write operation (written to MIIM field reg) */
2772 u32 Cmd;/* Command (written to MIIM command reg) */
2776 /* DBG_ERROR("ENTER %s\n", __func__); */
2778 SXG_TRACE(TRACE_SXG, SxgTraceBuffer, TRACE_NOISY, "WrtMDIO",
2781 /* Ensure values don't exceed field width */
2782 DevAddr &= 0x001F; /* 5-bit field */
2783 RegAddr &= 0xFFFF; /* 16-bit field */
2784 Value &= 0xFFFF; /* 16-bit field */
2786 /* Set MIIM field register bits for an MIIM address operation */
2787 AddrOp = (MIIM_PORT_NUM << AXGMAC_AMIIM_FIELD_PORT_SHIFT) |
2788 (DevAddr << AXGMAC_AMIIM_FIELD_DEV_SHIFT) |
2789 (MIIM_TA_10GB << AXGMAC_AMIIM_FIELD_TA_SHIFT) |
2790 (MIIM_OP_ADDR << AXGMAC_AMIIM_FIELD_OP_SHIFT) | RegAddr;
2792 /* Set MIIM field register bits for an MIIM write operation */
2793 WriteOp = (MIIM_PORT_NUM << AXGMAC_AMIIM_FIELD_PORT_SHIFT) |
2794 (DevAddr << AXGMAC_AMIIM_FIELD_DEV_SHIFT) |
2795 (MIIM_TA_10GB << AXGMAC_AMIIM_FIELD_TA_SHIFT) |
2796 (MIIM_OP_WRITE << AXGMAC_AMIIM_FIELD_OP_SHIFT) | Value;
2798 /* Set MIIM command register bits to execute an MIIM command */
2799 Cmd = AXGMAC_AMIIM_CMD_START | AXGMAC_AMIIM_CMD_10G_OPERATION;
2801 /* Reset the command register command bit (in case it's not 0) */
2802 WRITE_REG(HwRegs->MacAmiimCmd, 0, TRUE);
2804 /* MIIM write to set the address of the specified MDIO register */
2805 WRITE_REG(HwRegs->MacAmiimField, AddrOp, TRUE);
2807 /* Write to MIIM Command Register to execute to address operation */
2808 WRITE_REG(HwRegs->MacAmiimCmd, Cmd, TRUE);
2810 /* Poll AMIIM Indicator register to wait for completion */
2811 Timeout = SXG_LINK_TIMEOUT;
2813 udelay(100); /* Timeout in 100us units */
2814 READ_REG(HwRegs->MacAmiimIndicator, ValueRead);
2815 if (--Timeout == 0) {
2816 return (STATUS_FAILURE);
2818 } while (ValueRead & AXGMAC_AMIIM_INDC_BUSY);
2820 /* Reset the command register command bit */
2821 WRITE_REG(HwRegs->MacAmiimCmd, 0, TRUE);
2823 /* MIIM write to set up an MDIO write operation */
2824 WRITE_REG(HwRegs->MacAmiimField, WriteOp, TRUE);
2826 /* Write to MIIM Command Register to execute the write operation */
2827 WRITE_REG(HwRegs->MacAmiimCmd, Cmd, TRUE);
2829 /* Poll AMIIM Indicator register to wait for completion */
2830 Timeout = SXG_LINK_TIMEOUT;
2832 udelay(100); /* Timeout in 100us units */
2833 READ_REG(HwRegs->MacAmiimIndicator, ValueRead);
2834 if (--Timeout == 0) {
2835 return (STATUS_FAILURE);
2837 } while (ValueRead & AXGMAC_AMIIM_INDC_BUSY);
2839 /* DBG_ERROR("EXIT %s\n", __func__); */
2841 return (STATUS_SUCCESS);
2845 * sxg_read_mdio_reg - Read a register on the MDIO bus
2848 * adapter - A pointer to our adapter structure
2849 * DevAddr - MDIO device number being addressed
2850 * RegAddr - register address for the specified MDIO device
2851 * pValue - pointer to where to put data read from the MDIO register
2856 static int sxg_read_mdio_reg(struct adapter_t *adapter,
2857 u32 DevAddr, u32 RegAddr, u32 *pValue)
2859 struct sxg_hw_regs *HwRegs = adapter->HwRegs;
2860 u32 AddrOp; /* Address operation (written to MIIM field reg) */
2861 u32 ReadOp; /* Read operation (written to MIIM field reg) */
2862 u32 Cmd; /* Command (written to MIIM command reg) */
2866 SXG_TRACE(TRACE_SXG, SxgTraceBuffer, TRACE_NOISY, "WrtMDIO",
2868 DBG_ERROR("ENTER %s\n", __FUNCTION__);
2870 /* Ensure values don't exceed field width */
2871 DevAddr &= 0x001F; /* 5-bit field */
2872 RegAddr &= 0xFFFF; /* 16-bit field */
2874 /* Set MIIM field register bits for an MIIM address operation */
2875 AddrOp = (MIIM_PORT_NUM << AXGMAC_AMIIM_FIELD_PORT_SHIFT) |
2876 (DevAddr << AXGMAC_AMIIM_FIELD_DEV_SHIFT) |
2877 (MIIM_TA_10GB << AXGMAC_AMIIM_FIELD_TA_SHIFT) |
2878 (MIIM_OP_ADDR << AXGMAC_AMIIM_FIELD_OP_SHIFT) | RegAddr;
2880 /* Set MIIM field register bits for an MIIM read operation */
2881 ReadOp = (MIIM_PORT_NUM << AXGMAC_AMIIM_FIELD_PORT_SHIFT) |
2882 (DevAddr << AXGMAC_AMIIM_FIELD_DEV_SHIFT) |
2883 (MIIM_TA_10GB << AXGMAC_AMIIM_FIELD_TA_SHIFT) |
2884 (MIIM_OP_READ << AXGMAC_AMIIM_FIELD_OP_SHIFT);
2886 /* Set MIIM command register bits to execute an MIIM command */
2887 Cmd = AXGMAC_AMIIM_CMD_START | AXGMAC_AMIIM_CMD_10G_OPERATION;
2889 /* Reset the command register command bit (in case it's not 0) */
2890 WRITE_REG(HwRegs->MacAmiimCmd, 0, TRUE);
2892 /* MIIM write to set the address of the specified MDIO register */
2893 WRITE_REG(HwRegs->MacAmiimField, AddrOp, TRUE);
2895 /* Write to MIIM Command Register to execute to address operation */
2896 WRITE_REG(HwRegs->MacAmiimCmd, Cmd, TRUE);
2898 /* Poll AMIIM Indicator register to wait for completion */
2899 Timeout = SXG_LINK_TIMEOUT;
2901 udelay(100); /* Timeout in 100us units */
2902 READ_REG(HwRegs->MacAmiimIndicator, ValueRead);
2903 if (--Timeout == 0) {
2904 DBG_ERROR("EXIT %s with STATUS_FAILURE 1\n", __FUNCTION__);
2906 return (STATUS_FAILURE);
2908 } while (ValueRead & AXGMAC_AMIIM_INDC_BUSY);
2910 /* Reset the command register command bit */
2911 WRITE_REG(HwRegs->MacAmiimCmd, 0, TRUE);
2913 /* MIIM write to set up an MDIO register read operation */
2914 WRITE_REG(HwRegs->MacAmiimField, ReadOp, TRUE);
2916 /* Write to MIIM Command Register to execute the read operation */
2917 WRITE_REG(HwRegs->MacAmiimCmd, Cmd, TRUE);
2919 /* Poll AMIIM Indicator register to wait for completion */
2920 Timeout = SXG_LINK_TIMEOUT;
2922 udelay(100); /* Timeout in 100us units */
2923 READ_REG(HwRegs->MacAmiimIndicator, ValueRead);
2924 if (--Timeout == 0) {
2925 DBG_ERROR("EXIT %s with STATUS_FAILURE 2\n", __FUNCTION__);
2927 return (STATUS_FAILURE);
2929 } while (ValueRead & AXGMAC_AMIIM_INDC_BUSY);
2931 /* Read the MDIO register data back from the field register */
2932 READ_REG(HwRegs->MacAmiimField, *pValue);
2933 *pValue &= 0xFFFF; /* data is in the lower 16 bits */
2935 DBG_ERROR("EXIT %s\n", __FUNCTION__);
2937 return (STATUS_SUCCESS);
2941 * Functions to obtain the CRC corresponding to the destination mac address.
2942 * This is a standard ethernet CRC in that it is a 32-bit, reflected CRC using
2944 * x^32 + x^26 + x^23 + x^22 + x^16 + x^12 + x^11 + x^10 + x^8 + x^7 + x^5
2945 * + x^4 + x^2 + x^1.
2947 * After the CRC for the 6 bytes is generated (but before the value is
2948 * complemented), we must then transpose the value and return bits 30-23.
2950 static u32 sxg_crc_table[256];/* Table of CRC's for all possible byte values */
2951 static u32 sxg_crc_init; /* Is table initialized */
2953 /* Contruct the CRC32 table */
2954 static void sxg_mcast_init_crc32(void)
2956 u32 c; /* CRC shit reg */
2957 u32 e = 0; /* Poly X-or pattern */
2958 int i; /* counter */
2959 int k; /* byte being shifted into crc */
2961 static int p[] = { 0, 1, 2, 4, 5, 7, 8, 10, 11, 12, 16, 22, 23, 26 };
2963 for (i = 0; i < sizeof(p) / sizeof(int); i++) {
2964 e |= 1L << (31 - p[i]);
2967 for (i = 1; i < 256; i++) {
2969 for (k = 8; k; k--) {
2970 c = c & 1 ? (c >> 1) ^ e : c >> 1;
2972 sxg_crc_table[i] = c;
2977 * Return the MAC hast as described above.
2979 static unsigned char sxg_mcast_get_mac_hash(char *macaddr)
2984 unsigned char machash = 0;
2986 if (!sxg_crc_init) {
2987 sxg_mcast_init_crc32();
2991 crc = 0xFFFFFFFF; /* Preload shift register, per crc-32 spec */
2992 for (i = 0, p = macaddr; i < 6; ++p, ++i) {
2993 crc = (crc >> 8) ^ sxg_crc_table[(crc ^ *p) & 0xFF];
2996 /* Return bits 1-8, transposed */
2997 for (i = 1; i < 9; i++) {
2998 machash |= (((crc >> i) & 1) << (8 - i));
3004 static void sxg_mcast_set_mask(struct adapter_t *adapter)
3006 struct sxg_ucode_regs *sxg_regs = adapter->UcodeRegs;
3008 DBG_ERROR("%s ENTER (%s) macopts[%x] mask[%llx]\n", __func__,
3009 adapter->netdev->name, (unsigned int)adapter->MacFilter,
3010 adapter->MulticastMask);
3012 if (adapter->MacFilter & (MAC_ALLMCAST | MAC_PROMISC)) {
3014 * Turn on all multicast addresses. We have to do this for
3015 * promiscuous mode as well as ALLMCAST mode. It saves the
3016 * Microcode from having keep state about the MAC configuration
3018 /* DBG_ERROR("sxg: %s macopts = MAC_ALLMCAST | MAC_PROMISC\n
3019 * SLUT MODE!!!\n",__func__);
3021 WRITE_REG(sxg_regs->McastLow, 0xFFFFFFFF, FLUSH);
3022 WRITE_REG(sxg_regs->McastHigh, 0xFFFFFFFF, FLUSH);
3023 /* DBG_ERROR("%s (%s) WRITE to slic_regs slic_mcastlow&high \
3024 * 0xFFFFFFFF\n",__func__, adapter->netdev->name);
3029 * Commit our multicast mast to the SLIC by writing to the
3030 * multicast address mask registers
3032 DBG_ERROR("%s (%s) WRITE mcastlow[%lx] mcasthigh[%lx]\n",
3033 __func__, adapter->netdev->name,
3034 ((ulong) (adapter->MulticastMask & 0xFFFFFFFF)),
3036 ((adapter->MulticastMask >> 32) & 0xFFFFFFFF)));
3038 WRITE_REG(sxg_regs->McastLow,
3039 (u32) (adapter->MulticastMask & 0xFFFFFFFF), FLUSH);
3040 WRITE_REG(sxg_regs->McastHigh,
3042 MulticastMask >> 32) & 0xFFFFFFFF), FLUSH);
3047 * Allocate a mcast_address structure to hold the multicast address.
3050 static int sxg_mcast_add_list(struct adapter_t *adapter, char *address)
3052 struct mcast_address *mcaddr, *mlist;
3055 /* Check to see if it already exists */
3056 mlist = adapter->mcastaddrs;
3058 ETHER_EQ_ADDR(mlist->address, address, equaladdr);
3060 return (STATUS_SUCCESS);
3062 mlist = mlist->next;
3065 /* Doesn't already exist. Allocate a structure to hold it */
3066 mcaddr = kmalloc(sizeof(struct mcast_address), GFP_ATOMIC);
3070 memcpy(mcaddr->address, address, 6);
3072 mcaddr->next = adapter->mcastaddrs;
3073 adapter->mcastaddrs = mcaddr;
3075 return (STATUS_SUCCESS);
3078 static void sxg_mcast_set_bit(struct adapter_t *adapter, char *address)
3080 unsigned char crcpoly;
3082 /* Get the CRC polynomial for the mac address */
3083 crcpoly = sxg_mcast_get_mac_hash(address);
3086 * We only have space on the SLIC for 64 entries. Lop
3087 * off the top two bits. (2^6 = 64)
3091 /* OR in the new bit into our 64 bit mask. */
3092 adapter->MulticastMask |= (u64) 1 << crcpoly;
3095 static void sxg_mcast_set_list(struct net_device *dev)
3097 struct adapter_t *adapter = (struct adapter_t *) netdev_priv(dev);
3100 if (dev->flags & IFF_PROMISC) {
3101 adapter->MacFilter |= MAC_PROMISC;
3103 //XXX handle other flags as well
3104 sxg_mcast_set_mask(adapter);
3107 static void sxg_unmap_mmio_space(struct adapter_t *adapter)
3109 #if LINUX_FREES_ADAPTER_RESOURCES
3111 * if (adapter->Regs) {
3112 * iounmap(adapter->Regs);
3114 * adapter->slic_regs = NULL;
3119 void sxg_free_sgl_buffers(struct adapter_t *adapter)
3121 struct list_entry *ple;
3122 struct sxg_scatter_gather *Sgl;
3124 while(!(IsListEmpty(&adapter->AllSglBuffers))) {
3125 ple = RemoveHeadList(&adapter->AllSglBuffers);
3126 Sgl = container_of(ple, struct sxg_scatter_gather, AllList);
3128 adapter->AllSglBufferCount--;
3132 void sxg_free_rcvblocks(struct adapter_t *adapter)
3135 void *temp_RcvBlock;
3136 struct list_entry *ple;
3137 struct sxg_rcv_block_hdr *RcvBlockHdr;
3138 struct sxg_rcv_data_buffer_hdr *RcvDataBufferHdr;
3139 ASSERT((adapter->state == SXG_STATE_INITIALIZING) ||
3140 (adapter->state == SXG_STATE_HALTING));
3141 while(!(IsListEmpty(&adapter->AllRcvBlocks))) {
3143 ple = RemoveHeadList(&adapter->AllRcvBlocks);
3144 RcvBlockHdr = container_of(ple, struct sxg_rcv_block_hdr, AllList);
3146 if(RcvBlockHdr->VirtualAddress) {
3147 temp_RcvBlock = RcvBlockHdr->VirtualAddress;
3149 for(i=0; i< SXG_RCV_DESCRIPTORS_PER_BLOCK;
3150 i++, temp_RcvBlock += SXG_RCV_DATA_HDR_SIZE) {
3152 (struct sxg_rcv_data_buffer_hdr *)temp_RcvBlock;
3153 SXG_FREE_RCV_PACKET(RcvDataBufferHdr);
3157 pci_free_consistent(adapter->pcidev,
3158 SXG_RCV_BLOCK_SIZE(SXG_RCV_DATA_HDR_SIZE),
3159 RcvBlockHdr->VirtualAddress,
3160 RcvBlockHdr->PhysicalAddress);
3161 adapter->AllRcvBlockCount--;
3163 ASSERT(adapter->AllRcvBlockCount == 0);
3164 SXG_TRACE(TRACE_SXG, SxgTraceBuffer, TRACE_NOISY, "XFrRBlk",
3167 void sxg_free_mcast_addrs(struct adapter_t *adapter)
3169 struct sxg_multicast_address *address;
3170 while(adapter->MulticastAddrs) {
3171 address = adapter->MulticastAddrs;
3172 adapter->MulticastAddrs = address->Next;
3176 adapter->MulticastMask= 0;
3179 void sxg_unmap_resources(struct adapter_t *adapter)
3181 if(adapter->HwRegs) {
3182 iounmap((void *)adapter->HwRegs);
3184 if(adapter->UcodeRegs) {
3185 iounmap((void *)adapter->UcodeRegs);
3188 ASSERT(adapter->AllRcvBlockCount == 0);
3189 SXG_TRACE(TRACE_SXG, SxgTraceBuffer, TRACE_NOISY, "XFrRBlk",
3196 * sxg_free_resources - Free everything allocated in SxgAllocateResources
3199 * adapter - A pointer to our adapter structure
3204 void sxg_free_resources(struct adapter_t *adapter)
3206 u32 RssIds, IsrCount;
3208 struct net_device *netdev = adapter->netdev;
3209 RssIds = SXG_RSS_CPU_COUNT(adapter);
3210 IsrCount = adapter->MsiEnabled ? RssIds : 1;
3212 if (adapter->BasicAllocations == FALSE) {
3214 * No allocations have been made, including spinlocks,
3215 * or listhead initializations. Return.
3221 free_irq(adapter->netdev->irq, netdev);
3224 if (!(IsListEmpty(&adapter->AllRcvBlocks))) {
3225 sxg_free_rcvblocks(adapter);
3227 if (!(IsListEmpty(&adapter->AllSglBuffers))) {
3228 sxg_free_sgl_buffers(adapter);
3231 if (adapter->XmtRingZeroIndex) {
3232 pci_free_consistent(adapter->pcidev,
3234 adapter->XmtRingZeroIndex,
3235 adapter->PXmtRingZeroIndex);
3238 pci_free_consistent(adapter->pcidev,
3239 sizeof(u32) * IsrCount,
3240 adapter->Isr, adapter->PIsr);
3243 if (adapter->EventRings) {
3244 pci_free_consistent(adapter->pcidev,
3245 sizeof(struct sxg_event_ring) * RssIds,
3246 adapter->EventRings, adapter->PEventRings);
3248 if (adapter->RcvRings) {
3249 pci_free_consistent(adapter->pcidev,
3250 sizeof(struct sxg_rcv_ring) * 1,
3252 adapter->PRcvRings);
3253 adapter->RcvRings = NULL;
3256 if(adapter->XmtRings) {
3257 pci_free_consistent(adapter->pcidev,
3258 sizeof(struct sxg_xmt_ring) * 1,
3260 adapter->PXmtRings);
3261 adapter->XmtRings = NULL;
3264 if (adapter->ucode_stats) {
3265 pci_unmap_single(adapter->pcidev,
3266 sizeof(struct sxg_ucode_stats),
3267 adapter->pucode_stats, PCI_DMA_FROMDEVICE);
3268 adapter->ucode_stats = NULL;
3272 /* Unmap register spaces */
3273 sxg_unmap_resources(adapter);
3275 sxg_free_mcast_addrs(adapter);
3277 adapter->BasicAllocations = FALSE;
3282 * sxg_allocate_complete -
3284 * This routine is called when a memory allocation has completed.
3287 * struct adapter_t * - Our adapter structure
3288 * VirtualAddress - Memory virtual address
3289 * PhysicalAddress - Memory physical address
3290 * Length - Length of memory allocated (or 0)
3291 * Context - The type of buffer allocated
3296 static void sxg_allocate_complete(struct adapter_t *adapter,
3297 void *VirtualAddress,
3298 dma_addr_t PhysicalAddress,
3299 u32 Length, enum sxg_buffer_type Context)
3301 SXG_TRACE(TRACE_SXG, SxgTraceBuffer, TRACE_NOISY, "AllocCmp",
3302 adapter, VirtualAddress, Length, Context);
3303 ASSERT(atomic_read(&adapter->pending_allocations));
3304 atomic_dec(&adapter->pending_allocations);
3308 case SXG_BUFFER_TYPE_RCV:
3309 sxg_allocate_rcvblock_complete(adapter,
3311 PhysicalAddress, Length);
3313 case SXG_BUFFER_TYPE_SGL:
3314 sxg_allocate_sgl_buffer_complete(adapter, (struct sxg_scatter_gather *)
3316 PhysicalAddress, Length);
3319 SXG_TRACE(TRACE_SXG, SxgTraceBuffer, TRACE_NOISY, "XAlocCmp",
3320 adapter, VirtualAddress, Length, Context);
3324 * sxg_allocate_buffer_memory - Shared memory allocation routine used for
3325 * synchronous and asynchronous buffer allocations
3328 * adapter - A pointer to our adapter structure
3329 * Size - block size to allocate
3330 * BufferType - Type of buffer to allocate
3335 static int sxg_allocate_buffer_memory(struct adapter_t *adapter,
3336 u32 Size, enum sxg_buffer_type BufferType)
3342 SXG_TRACE(TRACE_SXG, SxgTraceBuffer, TRACE_NOISY, "AllocMem",
3343 adapter, Size, BufferType, 0);
3345 * Grab the adapter lock and check the state. If we're in anything other
3346 * than INITIALIZING or RUNNING state, fail. This is to prevent
3347 * allocations in an improper driver state
3350 atomic_inc(&adapter->pending_allocations);
3352 if(BufferType != SXG_BUFFER_TYPE_SGL)
3353 Buffer = pci_alloc_consistent(adapter->pcidev, Size, &pBuffer);
3355 Buffer = kzalloc(Size, GFP_ATOMIC);
3358 if (Buffer == NULL) {
3360 * Decrement the AllocationsPending count while holding
3361 * the lock. Pause processing relies on this
3363 atomic_dec(&adapter->pending_allocations);
3364 SXG_TRACE(TRACE_SXG, SxgTraceBuffer, TRACE_NOISY, "AlcMemF1",
3365 adapter, Size, BufferType, 0);
3366 return (STATUS_RESOURCES);
3368 sxg_allocate_complete(adapter, Buffer, pBuffer, Size, BufferType);
3369 status = STATUS_SUCCESS;
3371 SXG_TRACE(TRACE_SXG, SxgTraceBuffer, TRACE_NOISY, "XAlocMem",
3372 adapter, Size, BufferType, status);
3377 * sxg_allocate_rcvblock_complete - Complete a receive descriptor
3381 * adapter - A pointer to our adapter structure
3382 * RcvBlock - receive block virtual address
3383 * PhysicalAddress - Physical address
3384 * Length - Memory length
3388 static void sxg_allocate_rcvblock_complete(struct adapter_t *adapter,
3390 dma_addr_t PhysicalAddress,
3394 u32 BufferSize = adapter->ReceiveBufferSize;
3396 void *temp_RcvBlock;
3397 struct sxg_rcv_block_hdr *RcvBlockHdr;
3398 struct sxg_rcv_data_buffer_hdr *RcvDataBufferHdr;
3399 struct sxg_rcv_descriptor_block *RcvDescriptorBlock;
3400 struct sxg_rcv_descriptor_block_hdr *RcvDescriptorBlockHdr;
3402 SXG_TRACE(TRACE_SXG, SxgTraceBuffer, TRACE_NOISY, "AlRcvBlk",
3403 adapter, RcvBlock, Length, 0);
3404 if (RcvBlock == NULL) {
3407 memset(RcvBlock, 0, Length);
3408 ASSERT((BufferSize == SXG_RCV_DATA_BUFFER_SIZE) ||
3409 (BufferSize == SXG_RCV_JUMBO_BUFFER_SIZE));
3410 ASSERT(Length == SXG_RCV_BLOCK_SIZE(SXG_RCV_DATA_HDR_SIZE));
3412 * First, initialize the contained pool of receive data buffers.
3413 * This initialization requires NBL/NB/MDL allocations, if any of them
3414 * fail, free the block and return without queueing the shared memory
3416 //RcvDataBuffer = RcvBlock;
3417 temp_RcvBlock = RcvBlock;
3418 for (i = 0; i < SXG_RCV_DESCRIPTORS_PER_BLOCK;
3419 i++, temp_RcvBlock += SXG_RCV_DATA_HDR_SIZE) {
3420 RcvDataBufferHdr = (struct sxg_rcv_data_buffer_hdr *)
3422 /* For FREE macro assertion */
3423 RcvDataBufferHdr->State = SXG_BUFFER_UPSTREAM;
3424 SXG_ALLOCATE_RCV_PACKET(adapter, RcvDataBufferHdr, BufferSize);
3425 if (RcvDataBufferHdr->SxgDumbRcvPacket == NULL)
3431 * Place this entire block of memory on the AllRcvBlocks queue so it
3435 RcvBlockHdr = (struct sxg_rcv_block_hdr *) ((unsigned char *)RcvBlock +
3436 SXG_RCV_BLOCK_HDR_OFFSET(SXG_RCV_DATA_HDR_SIZE));
3437 RcvBlockHdr->VirtualAddress = RcvBlock;
3438 RcvBlockHdr->PhysicalAddress = PhysicalAddress;
3439 spin_lock(&adapter->RcvQLock);
3440 adapter->AllRcvBlockCount++;
3441 InsertTailList(&adapter->AllRcvBlocks, &RcvBlockHdr->AllList);
3442 spin_unlock(&adapter->RcvQLock);
3444 /* Now free the contained receive data buffers that we
3445 * initialized above */
3446 temp_RcvBlock = RcvBlock;
3447 for (i = 0, Paddr = PhysicalAddress;
3448 i < SXG_RCV_DESCRIPTORS_PER_BLOCK;
3449 i++, Paddr += SXG_RCV_DATA_HDR_SIZE,
3450 temp_RcvBlock += SXG_RCV_DATA_HDR_SIZE) {
3452 (struct sxg_rcv_data_buffer_hdr *)temp_RcvBlock;
3453 spin_lock(&adapter->RcvQLock);
3454 SXG_FREE_RCV_DATA_BUFFER(adapter, RcvDataBufferHdr);
3455 spin_unlock(&adapter->RcvQLock);
3458 /* Locate the descriptor block and put it on a separate free queue */
3459 RcvDescriptorBlock =
3460 (struct sxg_rcv_descriptor_block *) ((unsigned char *)RcvBlock +
3461 SXG_RCV_DESCRIPTOR_BLOCK_OFFSET
3462 (SXG_RCV_DATA_HDR_SIZE));
3463 RcvDescriptorBlockHdr =
3464 (struct sxg_rcv_descriptor_block_hdr *) ((unsigned char *)RcvBlock +
3465 SXG_RCV_DESCRIPTOR_BLOCK_HDR_OFFSET
3466 (SXG_RCV_DATA_HDR_SIZE));
3467 RcvDescriptorBlockHdr->VirtualAddress = RcvDescriptorBlock;
3468 RcvDescriptorBlockHdr->PhysicalAddress = Paddr;
3469 spin_lock(&adapter->RcvQLock);
3470 SXG_FREE_RCV_DESCRIPTOR_BLOCK(adapter, RcvDescriptorBlockHdr);
3471 spin_unlock(&adapter->RcvQLock);
3472 SXG_TRACE(TRACE_SXG, SxgTraceBuffer, TRACE_NOISY, "XAlRBlk",
3473 adapter, RcvBlock, Length, 0);
3476 /* Free any allocated resources */
3478 temp_RcvBlock = RcvBlock;
3479 for (i = 0; i < SXG_RCV_DESCRIPTORS_PER_BLOCK;
3480 i++, temp_RcvBlock += SXG_RCV_DATA_HDR_SIZE) {
3482 (struct sxg_rcv_data_buffer_hdr *)temp_RcvBlock;
3483 SXG_FREE_RCV_PACKET(RcvDataBufferHdr);
3485 pci_free_consistent(adapter->pcidev,
3486 Length, RcvBlock, PhysicalAddress);
3488 DBG_ERROR("%s: OUT OF RESOURCES\n", __func__);
3489 SXG_TRACE(TRACE_SXG, SxgTraceBuffer, TRACE_IMPORTANT, "RcvAFail",
3490 adapter, adapter->FreeRcvBufferCount,
3491 adapter->FreeRcvBlockCount, adapter->AllRcvBlockCount);
3492 adapter->Stats.NoMem++;
3496 * sxg_allocate_sgl_buffer_complete - Complete a SGL buffer allocation
3499 * adapter - A pointer to our adapter structure
3500 * SxgSgl - struct sxg_scatter_gather buffer
3501 * PhysicalAddress - Physical address
3502 * Length - Memory length
3506 static void sxg_allocate_sgl_buffer_complete(struct adapter_t *adapter,
3507 struct sxg_scatter_gather *SxgSgl,
3508 dma_addr_t PhysicalAddress,
3511 unsigned long sgl_flags;
3512 SXG_TRACE(TRACE_SXG, SxgTraceBuffer, TRACE_NOISY, "AlSglCmp",
3513 adapter, SxgSgl, Length, 0);
3515 spin_lock_irqsave(&adapter->SglQLock, sgl_flags);
3517 spin_lock(&adapter->SglQLock);
3518 adapter->AllSglBufferCount++;
3519 /* PhysicalAddress; */
3520 SxgSgl->PhysicalAddress = PhysicalAddress;
3521 /* Initialize backpointer once */
3522 SxgSgl->adapter = adapter;
3523 InsertTailList(&adapter->AllSglBuffers, &SxgSgl->AllList);
3525 spin_unlock_irqrestore(&adapter->SglQLock, sgl_flags);
3527 spin_unlock(&adapter->SglQLock);
3528 SxgSgl->State = SXG_BUFFER_BUSY;
3529 SXG_FREE_SGL_BUFFER(adapter, SxgSgl, NULL, in_irq());
3530 SXG_TRACE(TRACE_SXG, SxgTraceBuffer, TRACE_NOISY, "XAlSgl",
3531 adapter, SxgSgl, Length, 0);
3535 static void sxg_adapter_set_hwaddr(struct adapter_t *adapter)
3538 * DBG_ERROR ("%s ENTER card->config_set[%x] port[%d] physport[%d] \
3539 * funct#[%d]\n", __func__, card->config_set,
3540 * adapter->port, adapter->physport, adapter->functionnumber);
3542 * sxg_dbg_macaddrs(adapter);
3544 /* DBG_ERROR ("%s AFTER copying from config.macinfo into currmacaddr\n",
3548 /* sxg_dbg_macaddrs(adapter); */
3550 struct net_device * dev = adapter->netdev;
3553 printk("sxg: Dev is Null\n");
3556 DBG_ERROR("%s ENTER (%s)\n", __FUNCTION__, adapter->netdev->name);
3558 if (netif_running(dev)) {
3565 if (!(adapter->currmacaddr[0] ||
3566 adapter->currmacaddr[1] ||
3567 adapter->currmacaddr[2] ||
3568 adapter->currmacaddr[3] ||
3569 adapter->currmacaddr[4] || adapter->currmacaddr[5])) {
3570 memcpy(adapter->currmacaddr, adapter->macaddr, 6);
3572 if (adapter->netdev) {
3573 memcpy(adapter->netdev->dev_addr, adapter->currmacaddr, 6);
3574 memcpy(adapter->netdev->perm_addr, adapter->currmacaddr, 6);
3576 /* DBG_ERROR ("%s EXIT port %d\n", __func__, adapter->port); */
3577 sxg_dbg_macaddrs(adapter);
3582 static int sxg_mac_set_address(struct net_device *dev, void *ptr)
3584 struct adapter_t *adapter = (struct adapter_t *) netdev_priv(dev);
3585 struct sockaddr *addr = ptr;
3587 DBG_ERROR("%s ENTER (%s)\n", __func__, adapter->netdev->name);
3589 if (netif_running(dev)) {
3595 DBG_ERROR("sxg: %s (%s) curr %2.2X:%2.2X:%2.2X:%2.2X:%2.2X:%2.2X\n",
3596 __func__, adapter->netdev->name, adapter->currmacaddr[0],
3597 adapter->currmacaddr[1], adapter->currmacaddr[2],
3598 adapter->currmacaddr[3], adapter->currmacaddr[4],
3599 adapter->currmacaddr[5]);
3600 memcpy(dev->dev_addr, addr->sa_data, dev->addr_len);
3601 memcpy(adapter->currmacaddr, addr->sa_data, dev->addr_len);
3602 DBG_ERROR("sxg: %s (%s) new %2.2X:%2.2X:%2.2X:%2.2X:%2.2X:%2.2X\n",
3603 __func__, adapter->netdev->name, adapter->currmacaddr[0],
3604 adapter->currmacaddr[1], adapter->currmacaddr[2],
3605 adapter->currmacaddr[3], adapter->currmacaddr[4],
3606 adapter->currmacaddr[5]);
3608 sxg_config_set(adapter, TRUE);
3614 * SXG DRIVER FUNCTIONS (below)
3616 * sxg_initialize_adapter - Initialize adapter
3619 * adapter - A pointer to our adapter structure
3623 static int sxg_initialize_adapter(struct adapter_t *adapter)
3625 u32 RssIds, IsrCount;
3629 SXG_TRACE(TRACE_SXG, SxgTraceBuffer, TRACE_NOISY, "InitAdpt",
3632 RssIds = 1; /* XXXTODO SXG_RSS_CPU_COUNT(adapter); */
3633 IsrCount = adapter->MsiEnabled ? RssIds : 1;
3636 * Sanity check SXG_UCODE_REGS structure definition to
3637 * make sure the length is correct
3639 ASSERT(sizeof(struct sxg_ucode_regs) == SXG_REGISTER_SIZE_PER_CPU);
3641 /* Disable interrupts */
3642 SXG_DISABLE_ALL_INTERRUPTS(adapter);
3645 ASSERT((adapter->FrameSize == ETHERMAXFRAME) ||
3646 (adapter->FrameSize == JUMBOMAXFRAME));
3647 WRITE_REG(adapter->UcodeRegs[0].LinkMtu, adapter->FrameSize, TRUE);
3649 /* Set event ring base address and size */
3650 WRITE_REG64(adapter,
3651 adapter->UcodeRegs[0].EventBase, adapter->PEventRings, 0);
3652 WRITE_REG(adapter->UcodeRegs[0].EventSize, EVENT_RING_SIZE, TRUE);
3654 /* Per-ISR initialization */
3655 for (i = 0; i < IsrCount; i++) {
3657 /* Set interrupt status pointer */
3658 Addr = adapter->PIsr + (i * sizeof(u32));
3659 WRITE_REG64(adapter, adapter->UcodeRegs[i].Isp, Addr, i);
3662 /* XMT ring zero index */
3663 WRITE_REG64(adapter,
3664 adapter->UcodeRegs[0].SPSendIndex,
3665 adapter->PXmtRingZeroIndex, 0);
3667 /* Per-RSS initialization */
3668 for (i = 0; i < RssIds; i++) {
3669 /* Release all event ring entries to the Microcode */
3670 WRITE_REG(adapter->UcodeRegs[i].EventRelease, EVENT_RING_SIZE,
3674 /* Transmit ring base and size */
3675 WRITE_REG64(adapter,
3676 adapter->UcodeRegs[0].XmtBase, adapter->PXmtRings, 0);
3677 WRITE_REG(adapter->UcodeRegs[0].XmtSize, SXG_XMT_RING_SIZE, TRUE);
3679 /* Receive ring base and size */
3680 WRITE_REG64(adapter,
3681 adapter->UcodeRegs[0].RcvBase, adapter->PRcvRings, 0);
3682 WRITE_REG(adapter->UcodeRegs[0].RcvSize, SXG_RCV_RING_SIZE, TRUE);
3684 /* Populate the card with receive buffers */
3685 sxg_stock_rcv_buffers(adapter);
3688 * Initialize checksum offload capabilities. At the moment we always
3689 * enable IP and TCP receive checksums on the card. Depending on the
3690 * checksum configuration specified by the user, we can choose to
3691 * report or ignore the checksum information provided by the card.
3693 WRITE_REG(adapter->UcodeRegs[0].ReceiveChecksum,
3694 SXG_RCV_TCP_CSUM_ENABLED | SXG_RCV_IP_CSUM_ENABLED, TRUE);
3696 /* Initialize the MAC, XAUI */
3697 DBG_ERROR("sxg: %s ENTER sxg_initialize_link\n", __func__);
3698 status = sxg_initialize_link(adapter);
3699 DBG_ERROR("sxg: %s EXIT sxg_initialize_link status[%x]\n", __func__,
3701 if (status != STATUS_SUCCESS) {
3705 * Initialize Dead to FALSE.
3706 * SlicCheckForHang or SlicDumpThread will take it from here.
3708 adapter->Dead = FALSE;
3709 adapter->PingOutstanding = FALSE;
3710 adapter->State = SXG_STATE_RUNNING;
3712 SXG_TRACE(TRACE_SXG, SxgTraceBuffer, TRACE_NOISY, "XInit",
3714 return (STATUS_SUCCESS);
3718 * sxg_fill_descriptor_block - Populate a descriptor block and give it to
3719 * the card. The caller should hold the RcvQLock
3722 * adapter - A pointer to our adapter structure
3723 * RcvDescriptorBlockHdr - Descriptor block to fill
3728 static int sxg_fill_descriptor_block(struct adapter_t *adapter,
3729 struct sxg_rcv_descriptor_block_hdr *RcvDescriptorBlockHdr)
3732 struct sxg_ring_info *RcvRingInfo = &adapter->RcvRingZeroInfo;
3733 struct sxg_rcv_data_buffer_hdr *RcvDataBufferHdr;
3734 struct sxg_rcv_descriptor_block *RcvDescriptorBlock;
3735 struct sxg_cmd *RingDescriptorCmd;
3736 struct sxg_rcv_ring *RingZero = &adapter->RcvRings[0];
3738 SXG_TRACE(TRACE_SXG, SxgTraceBuffer, TRACE_NOISY, "FilBlk",
3739 adapter, adapter->RcvBuffersOnCard,
3740 adapter->FreeRcvBufferCount, adapter->AllRcvBlockCount);
3742 ASSERT(RcvDescriptorBlockHdr);
3745 * If we don't have the resources to fill the descriptor block,
3748 if ((adapter->FreeRcvBufferCount < SXG_RCV_DESCRIPTORS_PER_BLOCK) ||
3749 SXG_RING_FULL(RcvRingInfo)) {
3750 adapter->Stats.NoMem++;
3751 return (STATUS_FAILURE);
3753 /* Get a ring descriptor command */
3754 SXG_GET_CMD(RingZero,
3755 RcvRingInfo, RingDescriptorCmd, RcvDescriptorBlockHdr);
3756 ASSERT(RingDescriptorCmd);
3757 RcvDescriptorBlockHdr->State = SXG_BUFFER_ONCARD;
3758 RcvDescriptorBlock = (struct sxg_rcv_descriptor_block *)
3759 RcvDescriptorBlockHdr->VirtualAddress;
3761 /* Fill in the descriptor block */
3762 for (i = 0; i < SXG_RCV_DESCRIPTORS_PER_BLOCK; i++) {
3763 SXG_GET_RCV_DATA_BUFFER(adapter, RcvDataBufferHdr);
3764 ASSERT(RcvDataBufferHdr);
3765 // ASSERT(RcvDataBufferHdr->SxgDumbRcvPacket);
3766 if (!RcvDataBufferHdr->SxgDumbRcvPacket) {
3767 SXG_ALLOCATE_RCV_PACKET(adapter, RcvDataBufferHdr,
3768 adapter->ReceiveBufferSize);
3769 if(RcvDataBufferHdr->skb)
3770 RcvDataBufferHdr->SxgDumbRcvPacket =
3771 RcvDataBufferHdr->skb;
3775 SXG_REINIATIALIZE_PACKET(RcvDataBufferHdr->SxgDumbRcvPacket);
3776 RcvDataBufferHdr->State = SXG_BUFFER_ONCARD;
3777 RcvDescriptorBlock->Descriptors[i].VirtualAddress =
3778 (void *)RcvDataBufferHdr;
3780 RcvDescriptorBlock->Descriptors[i].PhysicalAddress =
3781 RcvDataBufferHdr->PhysicalAddress;
3783 /* Add the descriptor block to receive descriptor ring 0 */
3784 RingDescriptorCmd->Sgl = RcvDescriptorBlockHdr->PhysicalAddress;
3787 * RcvBuffersOnCard is not protected via the receive lock (see
3788 * sxg_process_event_queue) We don't want to grap a lock every time a
3789 * buffer is returned to us, so we use atomic interlocked functions
3792 adapter->RcvBuffersOnCard += SXG_RCV_DESCRIPTORS_PER_BLOCK;
3794 SXG_TRACE(TRACE_SXG, SxgTraceBuffer, TRACE_NOISY, "DscBlk",
3795 RcvDescriptorBlockHdr,
3796 RingDescriptorCmd, RcvRingInfo->Head, RcvRingInfo->Tail);
3798 WRITE_REG(adapter->UcodeRegs[0].RcvCmd, 1, true);
3799 SXG_TRACE(TRACE_SXG, SxgTraceBuffer, TRACE_NOISY, "XFilBlk",
3800 adapter, adapter->RcvBuffersOnCard,
3801 adapter->FreeRcvBufferCount, adapter->AllRcvBlockCount);
3802 return (STATUS_SUCCESS);
3808 * sxg_stock_rcv_buffers - Stock the card with receive buffers
3811 * adapter - A pointer to our adapter structure
3816 static void sxg_stock_rcv_buffers(struct adapter_t *adapter)
3818 struct sxg_rcv_descriptor_block_hdr *RcvDescriptorBlockHdr;
3820 SXG_TRACE(TRACE_SXG, SxgTraceBuffer, TRACE_NOISY, "StockBuf",
3821 adapter, adapter->RcvBuffersOnCard,
3822 adapter->FreeRcvBufferCount, adapter->AllRcvBlockCount);
3824 * First, see if we've got less than our minimum threshold of
3825 * receive buffers, there isn't an allocation in progress, and
3826 * we haven't exceeded our maximum.. get another block of buffers
3827 * None of this needs to be SMP safe. It's round numbers.
3829 if ((adapter->FreeRcvBufferCount < SXG_MIN_RCV_DATA_BUFFERS) &&
3830 (adapter->AllRcvBlockCount < SXG_MAX_RCV_BLOCKS) &&
3831 (atomic_read(&adapter->pending_allocations) == 0)) {
3832 sxg_allocate_buffer_memory(adapter,
3834 (SXG_RCV_DATA_HDR_SIZE),
3835 SXG_BUFFER_TYPE_RCV);
3837 /* Now grab the RcvQLock lock and proceed */
3838 spin_lock(&adapter->RcvQLock);
3839 while (adapter->RcvBuffersOnCard < SXG_RCV_DATA_BUFFERS) {
3840 struct list_entry *_ple;
3842 /* Get a descriptor block */
3843 RcvDescriptorBlockHdr = NULL;
3844 if (adapter->FreeRcvBlockCount) {
3845 _ple = RemoveHeadList(&adapter->FreeRcvBlocks);
3846 RcvDescriptorBlockHdr =
3847 container_of(_ple, struct sxg_rcv_descriptor_block_hdr,
3849 adapter->FreeRcvBlockCount--;
3850 RcvDescriptorBlockHdr->State = SXG_BUFFER_BUSY;
3853 if (RcvDescriptorBlockHdr == NULL) {
3855 adapter->Stats.NoMem++;
3858 /* Fill in the descriptor block and give it to the card */
3859 if (sxg_fill_descriptor_block(adapter, RcvDescriptorBlockHdr) ==
3861 /* Free the descriptor block */
3862 SXG_FREE_RCV_DESCRIPTOR_BLOCK(adapter,
3863 RcvDescriptorBlockHdr);
3867 spin_unlock(&adapter->RcvQLock);
3868 SXG_TRACE(TRACE_SXG, SxgTraceBuffer, TRACE_NOISY, "XFilBlks",
3869 adapter, adapter->RcvBuffersOnCard,
3870 adapter->FreeRcvBufferCount, adapter->AllRcvBlockCount);
3874 * sxg_complete_descriptor_blocks - Return descriptor blocks that have been
3875 * completed by the microcode
3878 * adapter - A pointer to our adapter structure
3879 * Index - Where the microcode is up to
3884 static void sxg_complete_descriptor_blocks(struct adapter_t *adapter,
3885 unsigned char Index)
3887 struct sxg_rcv_ring *RingZero = &adapter->RcvRings[0];
3888 struct sxg_ring_info *RcvRingInfo = &adapter->RcvRingZeroInfo;
3889 struct sxg_rcv_descriptor_block_hdr *RcvDescriptorBlockHdr;
3890 struct sxg_cmd *RingDescriptorCmd;
3892 SXG_TRACE(TRACE_SXG, SxgTraceBuffer, TRACE_NOISY, "CmpRBlks",
3893 adapter, Index, RcvRingInfo->Head, RcvRingInfo->Tail);
3895 /* Now grab the RcvQLock lock and proceed */
3896 spin_lock(&adapter->RcvQLock);
3897 ASSERT(Index != RcvRingInfo->Tail);
3898 while (sxg_ring_get_forward_diff(RcvRingInfo, Index,
3899 RcvRingInfo->Tail) > 3) {
3901 * Locate the current Cmd (ring descriptor entry), and
3902 * associated receive descriptor block, and advance
3905 SXG_RETURN_CMD(RingZero,
3907 RingDescriptorCmd, RcvDescriptorBlockHdr);
3908 SXG_TRACE(TRACE_SXG, SxgTraceBuffer, TRACE_NOISY, "CmpRBlk",
3909 RcvRingInfo->Head, RcvRingInfo->Tail,
3910 RingDescriptorCmd, RcvDescriptorBlockHdr);
3912 /* Clear the SGL field */
3913 RingDescriptorCmd->Sgl = 0;
3915 * Attempt to refill it and hand it right back to the
3916 * card. If we fail to refill it, free the descriptor block
3917 * header. The card will be restocked later via the
3918 * RcvBuffersOnCard test
3920 if (sxg_fill_descriptor_block(adapter,
3921 RcvDescriptorBlockHdr) == STATUS_FAILURE)
3922 SXG_FREE_RCV_DESCRIPTOR_BLOCK(adapter,
3923 RcvDescriptorBlockHdr);
3925 spin_unlock(&adapter->RcvQLock);
3926 SXG_TRACE(TRACE_SXG, SxgTraceBuffer, TRACE_NOISY, "XCRBlks",
3927 adapter, Index, RcvRingInfo->Head, RcvRingInfo->Tail);
3931 * Read the statistics which the card has been maintaining.
3933 void sxg_collect_statistics(struct adapter_t *adapter)
3935 if(adapter->ucode_stats)
3936 WRITE_REG64(adapter, adapter->UcodeRegs[0].GetUcodeStats, adapter->pucode_stats, 0);
3937 adapter->stats.rx_fifo_errors = adapter->ucode_stats->ERDrops;
3938 adapter->stats.rx_over_errors = adapter->ucode_stats->NBDrops;
3939 adapter->stats.tx_fifo_errors = adapter->ucode_stats->XDrops;
3942 static struct net_device_stats *sxg_get_stats(struct net_device * dev)
3944 struct adapter_t *adapter = netdev_priv(dev);
3946 sxg_collect_statistics(adapter);
3947 return (&adapter->stats);
3950 static struct pci_driver sxg_driver = {
3951 .name = sxg_driver_name,
3952 .id_table = sxg_pci_tbl,
3953 .probe = sxg_entry_probe,
3954 .remove = sxg_entry_remove,
3955 #if SXG_POWER_MANAGEMENT_ENABLED
3956 .suspend = sxgpm_suspend,
3957 .resume = sxgpm_resume,
3959 /* .shutdown = slic_shutdown, MOOK_INVESTIGATE */
3962 static int __init sxg_module_init(void)
3969 return pci_register_driver(&sxg_driver);
3972 static void __exit sxg_module_cleanup(void)
3974 pci_unregister_driver(&sxg_driver);
3977 module_init(sxg_module_init);
3978 module_exit(sxg_module_cleanup);