Merge master.kernel.org:/pub/scm/linux/kernel/git/davem/net-2.6
[deliverable/linux.git] / drivers / net / s2io.c
CommitLineData
1da177e4 1/************************************************************************
776bd20f 2 * s2io.c: A Linux PCI-X Ethernet driver for Neterion 10GbE Server NIC
1da177e4
LT
3 * Copyright(c) 2002-2005 Neterion Inc.
4
5 * This software may be used and distributed according to the terms of
6 * the GNU General Public License (GPL), incorporated herein by reference.
7 * Drivers based on or derived from this code fall under the GPL and must
8 * retain the authorship, copyright and license notice. This file is not
9 * a complete program and may only be used when the entire operating
10 * system is licensed under the GPL.
11 * See the file COPYING in this distribution for more information.
12 *
13 * Credits:
20346722 14 * Jeff Garzik : For pointing out the improper error condition
15 * check in the s2io_xmit routine and also some
16 * issues in the Tx watch dog function. Also for
17 * patiently answering all those innumerable
1da177e4
LT
18 * questions regaring the 2.6 porting issues.
19 * Stephen Hemminger : Providing proper 2.6 porting mechanism for some
20 * macros available only in 2.6 Kernel.
20346722 21 * Francois Romieu : For pointing out all code part that were
1da177e4 22 * deprecated and also styling related comments.
20346722 23 * Grant Grundler : For helping me get rid of some Architecture
1da177e4
LT
24 * dependent code.
25 * Christopher Hellwig : Some more 2.6 specific issues in the driver.
20346722 26 *
1da177e4
LT
27 * The module loadable parameters that are supported by the driver and a brief
28 * explaination of all the variables.
20346722 29 * rx_ring_num : This can be used to program the number of receive rings used
30 * in the driver.
776bd20f 31 * rx_ring_sz: This defines the number of descriptors each ring can have. This
1da177e4 32 * is also an array of size 8.
da6971d8
AR
33 * rx_ring_mode: This defines the operation mode of all 8 rings. The valid
34 * values are 1, 2 and 3.
1da177e4 35 * tx_fifo_num: This defines the number of Tx FIFOs thats used int the driver.
20346722 36 * tx_fifo_len: This too is an array of 8. Each element defines the number of
1da177e4 37 * Tx descriptors that can be associated with each corresponding FIFO.
1da177e4
LT
38 ************************************************************************/
39
40#include <linux/config.h>
41#include <linux/module.h>
42#include <linux/types.h>
43#include <linux/errno.h>
44#include <linux/ioport.h>
45#include <linux/pci.h>
1e7f0bd8 46#include <linux/dma-mapping.h>
1da177e4
LT
47#include <linux/kernel.h>
48#include <linux/netdevice.h>
49#include <linux/etherdevice.h>
50#include <linux/skbuff.h>
51#include <linux/init.h>
52#include <linux/delay.h>
53#include <linux/stddef.h>
54#include <linux/ioctl.h>
55#include <linux/timex.h>
56#include <linux/sched.h>
57#include <linux/ethtool.h>
1da177e4 58#include <linux/workqueue.h>
be3a6b02 59#include <linux/if_vlan.h>
7d3d0439
RA
60#include <linux/ip.h>
61#include <linux/tcp.h>
62#include <net/tcp.h>
1da177e4 63
1da177e4
LT
64#include <asm/system.h>
65#include <asm/uaccess.h>
20346722 66#include <asm/io.h>
fe931395 67#include <asm/div64.h>
1da177e4
LT
68
69/* local include */
70#include "s2io.h"
71#include "s2io-regs.h"
72
7d3d0439 73#define DRV_VERSION "2.0.11.2"
6c1792f4 74
1da177e4 75/* S2io Driver name & version. */
20346722 76static char s2io_driver_name[] = "Neterion";
6c1792f4 77static char s2io_driver_version[] = DRV_VERSION;
1da177e4 78
26df54bf
AB
79static int rxd_size[4] = {32,48,48,64};
80static int rxd_count[4] = {127,85,85,63};
da6971d8 81
5e25b9dd 82static inline int RXD_IS_UP2DT(RxD_t *rxdp)
83{
84 int ret;
85
86 ret = ((!(rxdp->Control_1 & RXD_OWN_XENA)) &&
87 (GET_RXD_MARKER(rxdp->Control_2) != THE_RXD_MARK));
88
89 return ret;
90}
91
20346722 92/*
1da177e4
LT
93 * Cards with following subsystem_id have a link state indication
94 * problem, 600B, 600C, 600D, 640B, 640C and 640D.
95 * macro below identifies these cards given the subsystem_id.
96 */
541ae68f 97#define CARDS_WITH_FAULTY_LINK_INDICATORS(dev_type, subid) \
98 (dev_type == XFRAME_I_DEVICE) ? \
99 ((((subid >= 0x600B) && (subid <= 0x600D)) || \
100 ((subid >= 0x640B) && (subid <= 0x640D))) ? 1 : 0) : 0
1da177e4
LT
101
102#define LINK_IS_UP(val64) (!(val64 & (ADAPTER_STATUS_RMAC_REMOTE_FAULT | \
103 ADAPTER_STATUS_RMAC_LOCAL_FAULT)))
104#define TASKLET_IN_USE test_and_set_bit(0, (&sp->tasklet_status))
105#define PANIC 1
106#define LOW 2
107static inline int rx_buffer_level(nic_t * sp, int rxb_size, int ring)
108{
109 int level = 0;
20346722 110 mac_info_t *mac_control;
111
112 mac_control = &sp->mac_control;
113 if ((mac_control->rings[ring].pkt_cnt - rxb_size) > 16) {
1da177e4 114 level = LOW;
da6971d8 115 if (rxb_size <= rxd_count[sp->rxd_mode]) {
1da177e4
LT
116 level = PANIC;
117 }
118 }
119
120 return level;
121}
122
123/* Ethtool related variables and Macros. */
124static char s2io_gstrings[][ETH_GSTRING_LEN] = {
125 "Register test\t(offline)",
126 "Eeprom test\t(offline)",
127 "Link test\t(online)",
128 "RLDRAM test\t(offline)",
129 "BIST Test\t(offline)"
130};
131
132static char ethtool_stats_keys[][ETH_GSTRING_LEN] = {
133 {"tmac_frms"},
134 {"tmac_data_octets"},
135 {"tmac_drop_frms"},
136 {"tmac_mcst_frms"},
137 {"tmac_bcst_frms"},
138 {"tmac_pause_ctrl_frms"},
139 {"tmac_any_err_frms"},
140 {"tmac_vld_ip_octets"},
141 {"tmac_vld_ip"},
142 {"tmac_drop_ip"},
143 {"tmac_icmp"},
144 {"tmac_rst_tcp"},
145 {"tmac_tcp"},
146 {"tmac_udp"},
147 {"rmac_vld_frms"},
148 {"rmac_data_octets"},
149 {"rmac_fcs_err_frms"},
150 {"rmac_drop_frms"},
151 {"rmac_vld_mcst_frms"},
152 {"rmac_vld_bcst_frms"},
153 {"rmac_in_rng_len_err_frms"},
154 {"rmac_long_frms"},
155 {"rmac_pause_ctrl_frms"},
156 {"rmac_discarded_frms"},
157 {"rmac_usized_frms"},
158 {"rmac_osized_frms"},
159 {"rmac_frag_frms"},
160 {"rmac_jabber_frms"},
161 {"rmac_ip"},
162 {"rmac_ip_octets"},
163 {"rmac_hdr_err_ip"},
164 {"rmac_drop_ip"},
165 {"rmac_icmp"},
166 {"rmac_tcp"},
167 {"rmac_udp"},
168 {"rmac_err_drp_udp"},
169 {"rmac_pause_cnt"},
170 {"rmac_accepted_ip"},
171 {"rmac_err_tcp"},
7ba013ac 172 {"\n DRIVER STATISTICS"},
173 {"single_bit_ecc_errs"},
174 {"double_bit_ecc_errs"},
7d3d0439
RA
175 ("lro_aggregated_pkts"),
176 ("lro_flush_both_count"),
177 ("lro_out_of_sequence_pkts"),
178 ("lro_flush_due_to_max_pkts"),
179 ("lro_avg_aggr_pkts"),
1da177e4
LT
180};
181
182#define S2IO_STAT_LEN sizeof(ethtool_stats_keys)/ ETH_GSTRING_LEN
183#define S2IO_STAT_STRINGS_LEN S2IO_STAT_LEN * ETH_GSTRING_LEN
184
185#define S2IO_TEST_LEN sizeof(s2io_gstrings) / ETH_GSTRING_LEN
186#define S2IO_STRINGS_LEN S2IO_TEST_LEN * ETH_GSTRING_LEN
187
25fff88e 188#define S2IO_TIMER_CONF(timer, handle, arg, exp) \
189 init_timer(&timer); \
190 timer.function = handle; \
191 timer.data = (unsigned long) arg; \
192 mod_timer(&timer, (jiffies + exp)) \
193
be3a6b02 194/* Add the vlan */
195static void s2io_vlan_rx_register(struct net_device *dev,
196 struct vlan_group *grp)
197{
198 nic_t *nic = dev->priv;
199 unsigned long flags;
200
201 spin_lock_irqsave(&nic->tx_lock, flags);
202 nic->vlgrp = grp;
203 spin_unlock_irqrestore(&nic->tx_lock, flags);
204}
205
206/* Unregister the vlan */
207static void s2io_vlan_rx_kill_vid(struct net_device *dev, unsigned long vid)
208{
209 nic_t *nic = dev->priv;
210 unsigned long flags;
211
212 spin_lock_irqsave(&nic->tx_lock, flags);
213 if (nic->vlgrp)
214 nic->vlgrp->vlan_devices[vid] = NULL;
215 spin_unlock_irqrestore(&nic->tx_lock, flags);
216}
217
20346722 218/*
1da177e4
LT
219 * Constants to be programmed into the Xena's registers, to configure
220 * the XAUI.
221 */
222
223#define SWITCH_SIGN 0xA5A5A5A5A5A5A5A5ULL
224#define END_SIGN 0x0
225
f71e1309 226static const u64 herc_act_dtx_cfg[] = {
541ae68f 227 /* Set address */
e960fc5c 228 0x8000051536750000ULL, 0x80000515367500E0ULL,
541ae68f 229 /* Write data */
e960fc5c 230 0x8000051536750004ULL, 0x80000515367500E4ULL,
541ae68f 231 /* Set address */
232 0x80010515003F0000ULL, 0x80010515003F00E0ULL,
233 /* Write data */
234 0x80010515003F0004ULL, 0x80010515003F00E4ULL,
235 /* Set address */
e960fc5c 236 0x801205150D440000ULL, 0x801205150D4400E0ULL,
237 /* Write data */
238 0x801205150D440004ULL, 0x801205150D4400E4ULL,
239 /* Set address */
541ae68f 240 0x80020515F2100000ULL, 0x80020515F21000E0ULL,
241 /* Write data */
242 0x80020515F2100004ULL, 0x80020515F21000E4ULL,
243 /* Done */
244 END_SIGN
245};
246
f71e1309 247static const u64 xena_mdio_cfg[] = {
1da177e4
LT
248 /* Reset PMA PLL */
249 0xC001010000000000ULL, 0xC0010100000000E0ULL,
250 0xC0010100008000E4ULL,
251 /* Remove Reset from PMA PLL */
252 0xC001010000000000ULL, 0xC0010100000000E0ULL,
253 0xC0010100000000E4ULL,
254 END_SIGN
255};
256
f71e1309 257static const u64 xena_dtx_cfg[] = {
1da177e4
LT
258 0x8000051500000000ULL, 0x80000515000000E0ULL,
259 0x80000515D93500E4ULL, 0x8001051500000000ULL,
260 0x80010515000000E0ULL, 0x80010515001E00E4ULL,
261 0x8002051500000000ULL, 0x80020515000000E0ULL,
262 0x80020515F21000E4ULL,
263 /* Set PADLOOPBACKN */
264 0x8002051500000000ULL, 0x80020515000000E0ULL,
265 0x80020515B20000E4ULL, 0x8003051500000000ULL,
266 0x80030515000000E0ULL, 0x80030515B20000E4ULL,
267 0x8004051500000000ULL, 0x80040515000000E0ULL,
268 0x80040515B20000E4ULL, 0x8005051500000000ULL,
269 0x80050515000000E0ULL, 0x80050515B20000E4ULL,
270 SWITCH_SIGN,
271 /* Remove PADLOOPBACKN */
272 0x8002051500000000ULL, 0x80020515000000E0ULL,
273 0x80020515F20000E4ULL, 0x8003051500000000ULL,
274 0x80030515000000E0ULL, 0x80030515F20000E4ULL,
275 0x8004051500000000ULL, 0x80040515000000E0ULL,
276 0x80040515F20000E4ULL, 0x8005051500000000ULL,
277 0x80050515000000E0ULL, 0x80050515F20000E4ULL,
278 END_SIGN
279};
280
20346722 281/*
1da177e4
LT
282 * Constants for Fixing the MacAddress problem seen mostly on
283 * Alpha machines.
284 */
f71e1309 285static const u64 fix_mac[] = {
1da177e4
LT
286 0x0060000000000000ULL, 0x0060600000000000ULL,
287 0x0040600000000000ULL, 0x0000600000000000ULL,
288 0x0020600000000000ULL, 0x0060600000000000ULL,
289 0x0020600000000000ULL, 0x0060600000000000ULL,
290 0x0020600000000000ULL, 0x0060600000000000ULL,
291 0x0020600000000000ULL, 0x0060600000000000ULL,
292 0x0020600000000000ULL, 0x0060600000000000ULL,
293 0x0020600000000000ULL, 0x0060600000000000ULL,
294 0x0020600000000000ULL, 0x0060600000000000ULL,
295 0x0020600000000000ULL, 0x0060600000000000ULL,
296 0x0020600000000000ULL, 0x0060600000000000ULL,
297 0x0020600000000000ULL, 0x0060600000000000ULL,
298 0x0020600000000000ULL, 0x0000600000000000ULL,
299 0x0040600000000000ULL, 0x0060600000000000ULL,
300 END_SIGN
301};
302
303/* Module Loadable parameters. */
304static unsigned int tx_fifo_num = 1;
305static unsigned int tx_fifo_len[MAX_TX_FIFOS] =
306 {[0 ...(MAX_TX_FIFOS - 1)] = 0 };
307static unsigned int rx_ring_num = 1;
308static unsigned int rx_ring_sz[MAX_RX_RINGS] =
309 {[0 ...(MAX_RX_RINGS - 1)] = 0 };
20346722 310static unsigned int rts_frm_len[MAX_RX_RINGS] =
311 {[0 ...(MAX_RX_RINGS - 1)] = 0 };
da6971d8 312static unsigned int rx_ring_mode = 1;
5e25b9dd 313static unsigned int use_continuous_tx_intrs = 1;
1da177e4
LT
314static unsigned int rmac_pause_time = 65535;
315static unsigned int mc_pause_threshold_q0q3 = 187;
316static unsigned int mc_pause_threshold_q4q7 = 187;
317static unsigned int shared_splits;
318static unsigned int tmac_util_period = 5;
319static unsigned int rmac_util_period = 5;
b6e3f982 320static unsigned int bimodal = 0;
da6971d8 321static unsigned int l3l4hdr_size = 128;
1da177e4
LT
322#ifndef CONFIG_S2IO_NAPI
323static unsigned int indicate_max_pkts;
324#endif
303bcb4b 325/* Frequency of Rx desc syncs expressed as power of 2 */
326static unsigned int rxsync_frequency = 3;
cc6e7c44
RA
327/* Interrupt type. Values can be 0(INTA), 1(MSI), 2(MSI_X) */
328static unsigned int intr_type = 0;
7d3d0439
RA
329/* Large receive offload feature */
330static unsigned int lro = 0;
331/* Max pkts to be aggregated by LRO at one time. If not specified,
332 * aggregation happens until we hit max IP pkt size(64K)
333 */
334static unsigned int lro_max_pkts = 0xFFFF;
1da177e4 335
20346722 336/*
1da177e4 337 * S2IO device table.
20346722 338 * This table lists all the devices that this driver supports.
1da177e4
LT
339 */
340static struct pci_device_id s2io_tbl[] __devinitdata = {
341 {PCI_VENDOR_ID_S2IO, PCI_DEVICE_ID_S2IO_WIN,
342 PCI_ANY_ID, PCI_ANY_ID},
343 {PCI_VENDOR_ID_S2IO, PCI_DEVICE_ID_S2IO_UNI,
344 PCI_ANY_ID, PCI_ANY_ID},
345 {PCI_VENDOR_ID_S2IO, PCI_DEVICE_ID_HERC_WIN,
20346722 346 PCI_ANY_ID, PCI_ANY_ID},
347 {PCI_VENDOR_ID_S2IO, PCI_DEVICE_ID_HERC_UNI,
348 PCI_ANY_ID, PCI_ANY_ID},
1da177e4
LT
349 {0,}
350};
351
352MODULE_DEVICE_TABLE(pci, s2io_tbl);
353
354static struct pci_driver s2io_driver = {
355 .name = "S2IO",
356 .id_table = s2io_tbl,
357 .probe = s2io_init_nic,
358 .remove = __devexit_p(s2io_rem_nic),
359};
360
361/* A simplifier macro used both by init and free shared_mem Fns(). */
362#define TXD_MEM_PAGE_CNT(len, per_each) ((len+per_each - 1) / per_each)
363
364/**
365 * init_shared_mem - Allocation and Initialization of Memory
366 * @nic: Device private variable.
20346722 367 * Description: The function allocates all the memory areas shared
368 * between the NIC and the driver. This includes Tx descriptors,
1da177e4
LT
369 * Rx descriptors and the statistics block.
370 */
371
372static int init_shared_mem(struct s2io_nic *nic)
373{
374 u32 size;
375 void *tmp_v_addr, *tmp_v_addr_next;
376 dma_addr_t tmp_p_addr, tmp_p_addr_next;
377 RxD_block_t *pre_rxd_blk = NULL;
20346722 378 int i, j, blk_cnt, rx_sz, tx_sz;
1da177e4
LT
379 int lst_size, lst_per_page;
380 struct net_device *dev = nic->dev;
8ae418cf 381 unsigned long tmp;
1da177e4 382 buffAdd_t *ba;
1da177e4
LT
383
384 mac_info_t *mac_control;
385 struct config_param *config;
386
387 mac_control = &nic->mac_control;
388 config = &nic->config;
389
390
391 /* Allocation and initialization of TXDLs in FIOFs */
392 size = 0;
393 for (i = 0; i < config->tx_fifo_num; i++) {
394 size += config->tx_cfg[i].fifo_len;
395 }
396 if (size > MAX_AVAILABLE_TXDS) {
0b1f7ebe 397 DBG_PRINT(ERR_DBG, "%s: Requested TxDs too high, ",
398 __FUNCTION__);
399 DBG_PRINT(ERR_DBG, "Requested: %d, max supported: 8192\n", size);
1da177e4
LT
400 return FAILURE;
401 }
402
403 lst_size = (sizeof(TxD_t) * config->max_txds);
20346722 404 tx_sz = lst_size * size;
1da177e4
LT
405 lst_per_page = PAGE_SIZE / lst_size;
406
407 for (i = 0; i < config->tx_fifo_num; i++) {
408 int fifo_len = config->tx_cfg[i].fifo_len;
409 int list_holder_size = fifo_len * sizeof(list_info_hold_t);
20346722 410 mac_control->fifos[i].list_info = kmalloc(list_holder_size,
411 GFP_KERNEL);
412 if (!mac_control->fifos[i].list_info) {
1da177e4
LT
413 DBG_PRINT(ERR_DBG,
414 "Malloc failed for list_info\n");
415 return -ENOMEM;
416 }
20346722 417 memset(mac_control->fifos[i].list_info, 0, list_holder_size);
1da177e4
LT
418 }
419 for (i = 0; i < config->tx_fifo_num; i++) {
420 int page_num = TXD_MEM_PAGE_CNT(config->tx_cfg[i].fifo_len,
421 lst_per_page);
20346722 422 mac_control->fifos[i].tx_curr_put_info.offset = 0;
423 mac_control->fifos[i].tx_curr_put_info.fifo_len =
1da177e4 424 config->tx_cfg[i].fifo_len - 1;
20346722 425 mac_control->fifos[i].tx_curr_get_info.offset = 0;
426 mac_control->fifos[i].tx_curr_get_info.fifo_len =
1da177e4 427 config->tx_cfg[i].fifo_len - 1;
20346722 428 mac_control->fifos[i].fifo_no = i;
429 mac_control->fifos[i].nic = nic;
fed5eccd 430 mac_control->fifos[i].max_txds = MAX_SKB_FRAGS + 2;
20346722 431
1da177e4
LT
432 for (j = 0; j < page_num; j++) {
433 int k = 0;
434 dma_addr_t tmp_p;
435 void *tmp_v;
436 tmp_v = pci_alloc_consistent(nic->pdev,
437 PAGE_SIZE, &tmp_p);
438 if (!tmp_v) {
439 DBG_PRINT(ERR_DBG,
440 "pci_alloc_consistent ");
441 DBG_PRINT(ERR_DBG, "failed for TxDL\n");
442 return -ENOMEM;
443 }
776bd20f 444 /* If we got a zero DMA address(can happen on
445 * certain platforms like PPC), reallocate.
446 * Store virtual address of page we don't want,
447 * to be freed later.
448 */
449 if (!tmp_p) {
450 mac_control->zerodma_virt_addr = tmp_v;
451 DBG_PRINT(INIT_DBG,
452 "%s: Zero DMA address for TxDL. ", dev->name);
453 DBG_PRINT(INIT_DBG,
6b4d617d 454 "Virtual address %p\n", tmp_v);
776bd20f 455 tmp_v = pci_alloc_consistent(nic->pdev,
456 PAGE_SIZE, &tmp_p);
457 if (!tmp_v) {
458 DBG_PRINT(ERR_DBG,
459 "pci_alloc_consistent ");
460 DBG_PRINT(ERR_DBG, "failed for TxDL\n");
461 return -ENOMEM;
462 }
463 }
1da177e4
LT
464 while (k < lst_per_page) {
465 int l = (j * lst_per_page) + k;
466 if (l == config->tx_cfg[i].fifo_len)
20346722 467 break;
468 mac_control->fifos[i].list_info[l].list_virt_addr =
1da177e4 469 tmp_v + (k * lst_size);
20346722 470 mac_control->fifos[i].list_info[l].list_phy_addr =
1da177e4
LT
471 tmp_p + (k * lst_size);
472 k++;
473 }
474 }
475 }
1da177e4 476
fed5eccd
AR
477 nic->ufo_in_band_v = kmalloc((sizeof(u64) * size), GFP_KERNEL);
478 if (!nic->ufo_in_band_v)
479 return -ENOMEM;
480
1da177e4
LT
481 /* Allocation and initialization of RXDs in Rings */
482 size = 0;
483 for (i = 0; i < config->rx_ring_num; i++) {
da6971d8
AR
484 if (config->rx_cfg[i].num_rxd %
485 (rxd_count[nic->rxd_mode] + 1)) {
1da177e4
LT
486 DBG_PRINT(ERR_DBG, "%s: RxD count of ", dev->name);
487 DBG_PRINT(ERR_DBG, "Ring%d is not a multiple of ",
488 i);
489 DBG_PRINT(ERR_DBG, "RxDs per Block");
490 return FAILURE;
491 }
492 size += config->rx_cfg[i].num_rxd;
20346722 493 mac_control->rings[i].block_count =
da6971d8
AR
494 config->rx_cfg[i].num_rxd /
495 (rxd_count[nic->rxd_mode] + 1 );
496 mac_control->rings[i].pkt_cnt = config->rx_cfg[i].num_rxd -
497 mac_control->rings[i].block_count;
1da177e4 498 }
da6971d8
AR
499 if (nic->rxd_mode == RXD_MODE_1)
500 size = (size * (sizeof(RxD1_t)));
501 else
502 size = (size * (sizeof(RxD3_t)));
20346722 503 rx_sz = size;
1da177e4
LT
504
505 for (i = 0; i < config->rx_ring_num; i++) {
20346722 506 mac_control->rings[i].rx_curr_get_info.block_index = 0;
507 mac_control->rings[i].rx_curr_get_info.offset = 0;
508 mac_control->rings[i].rx_curr_get_info.ring_len =
1da177e4 509 config->rx_cfg[i].num_rxd - 1;
20346722 510 mac_control->rings[i].rx_curr_put_info.block_index = 0;
511 mac_control->rings[i].rx_curr_put_info.offset = 0;
512 mac_control->rings[i].rx_curr_put_info.ring_len =
1da177e4 513 config->rx_cfg[i].num_rxd - 1;
20346722 514 mac_control->rings[i].nic = nic;
515 mac_control->rings[i].ring_no = i;
516
da6971d8
AR
517 blk_cnt = config->rx_cfg[i].num_rxd /
518 (rxd_count[nic->rxd_mode] + 1);
1da177e4
LT
519 /* Allocating all the Rx blocks */
520 for (j = 0; j < blk_cnt; j++) {
da6971d8
AR
521 rx_block_info_t *rx_blocks;
522 int l;
523
524 rx_blocks = &mac_control->rings[i].rx_blocks[j];
525 size = SIZE_OF_BLOCK; //size is always page size
1da177e4
LT
526 tmp_v_addr = pci_alloc_consistent(nic->pdev, size,
527 &tmp_p_addr);
528 if (tmp_v_addr == NULL) {
529 /*
20346722 530 * In case of failure, free_shared_mem()
531 * is called, which should free any
532 * memory that was alloced till the
1da177e4
LT
533 * failure happened.
534 */
da6971d8 535 rx_blocks->block_virt_addr = tmp_v_addr;
1da177e4
LT
536 return -ENOMEM;
537 }
538 memset(tmp_v_addr, 0, size);
da6971d8
AR
539 rx_blocks->block_virt_addr = tmp_v_addr;
540 rx_blocks->block_dma_addr = tmp_p_addr;
541 rx_blocks->rxds = kmalloc(sizeof(rxd_info_t)*
542 rxd_count[nic->rxd_mode],
543 GFP_KERNEL);
544 for (l=0; l<rxd_count[nic->rxd_mode];l++) {
545 rx_blocks->rxds[l].virt_addr =
546 rx_blocks->block_virt_addr +
547 (rxd_size[nic->rxd_mode] * l);
548 rx_blocks->rxds[l].dma_addr =
549 rx_blocks->block_dma_addr +
550 (rxd_size[nic->rxd_mode] * l);
551 }
552
20346722 553 mac_control->rings[i].rx_blocks[j].block_virt_addr =
554 tmp_v_addr;
555 mac_control->rings[i].rx_blocks[j].block_dma_addr =
556 tmp_p_addr;
1da177e4
LT
557 }
558 /* Interlinking all Rx Blocks */
559 for (j = 0; j < blk_cnt; j++) {
20346722 560 tmp_v_addr =
561 mac_control->rings[i].rx_blocks[j].block_virt_addr;
1da177e4 562 tmp_v_addr_next =
20346722 563 mac_control->rings[i].rx_blocks[(j + 1) %
1da177e4 564 blk_cnt].block_virt_addr;
20346722 565 tmp_p_addr =
566 mac_control->rings[i].rx_blocks[j].block_dma_addr;
1da177e4 567 tmp_p_addr_next =
20346722 568 mac_control->rings[i].rx_blocks[(j + 1) %
1da177e4
LT
569 blk_cnt].block_dma_addr;
570
571 pre_rxd_blk = (RxD_block_t *) tmp_v_addr;
1da177e4
LT
572 pre_rxd_blk->reserved_2_pNext_RxD_block =
573 (unsigned long) tmp_v_addr_next;
1da177e4
LT
574 pre_rxd_blk->pNext_RxD_Blk_physical =
575 (u64) tmp_p_addr_next;
576 }
577 }
da6971d8
AR
578 if (nic->rxd_mode >= RXD_MODE_3A) {
579 /*
580 * Allocation of Storages for buffer addresses in 2BUFF mode
581 * and the buffers as well.
582 */
583 for (i = 0; i < config->rx_ring_num; i++) {
584 blk_cnt = config->rx_cfg[i].num_rxd /
585 (rxd_count[nic->rxd_mode]+ 1);
586 mac_control->rings[i].ba =
587 kmalloc((sizeof(buffAdd_t *) * blk_cnt),
1da177e4 588 GFP_KERNEL);
da6971d8 589 if (!mac_control->rings[i].ba)
1da177e4 590 return -ENOMEM;
da6971d8
AR
591 for (j = 0; j < blk_cnt; j++) {
592 int k = 0;
593 mac_control->rings[i].ba[j] =
594 kmalloc((sizeof(buffAdd_t) *
595 (rxd_count[nic->rxd_mode] + 1)),
596 GFP_KERNEL);
597 if (!mac_control->rings[i].ba[j])
1da177e4 598 return -ENOMEM;
da6971d8
AR
599 while (k != rxd_count[nic->rxd_mode]) {
600 ba = &mac_control->rings[i].ba[j][k];
601
602 ba->ba_0_org = (void *) kmalloc
603 (BUF0_LEN + ALIGN_SIZE, GFP_KERNEL);
604 if (!ba->ba_0_org)
605 return -ENOMEM;
606 tmp = (unsigned long)ba->ba_0_org;
607 tmp += ALIGN_SIZE;
608 tmp &= ~((unsigned long) ALIGN_SIZE);
609 ba->ba_0 = (void *) tmp;
610
611 ba->ba_1_org = (void *) kmalloc
612 (BUF1_LEN + ALIGN_SIZE, GFP_KERNEL);
613 if (!ba->ba_1_org)
614 return -ENOMEM;
615 tmp = (unsigned long) ba->ba_1_org;
616 tmp += ALIGN_SIZE;
617 tmp &= ~((unsigned long) ALIGN_SIZE);
618 ba->ba_1 = (void *) tmp;
619 k++;
620 }
1da177e4
LT
621 }
622 }
623 }
1da177e4
LT
624
625 /* Allocation and initialization of Statistics block */
626 size = sizeof(StatInfo_t);
627 mac_control->stats_mem = pci_alloc_consistent
628 (nic->pdev, size, &mac_control->stats_mem_phy);
629
630 if (!mac_control->stats_mem) {
20346722 631 /*
632 * In case of failure, free_shared_mem() is called, which
633 * should free any memory that was alloced till the
1da177e4
LT
634 * failure happened.
635 */
636 return -ENOMEM;
637 }
638 mac_control->stats_mem_sz = size;
639
640 tmp_v_addr = mac_control->stats_mem;
641 mac_control->stats_info = (StatInfo_t *) tmp_v_addr;
642 memset(tmp_v_addr, 0, size);
1da177e4
LT
643 DBG_PRINT(INIT_DBG, "%s:Ring Mem PHY: 0x%llx\n", dev->name,
644 (unsigned long long) tmp_p_addr);
645
646 return SUCCESS;
647}
648
20346722 649/**
650 * free_shared_mem - Free the allocated Memory
1da177e4
LT
651 * @nic: Device private variable.
652 * Description: This function is to free all memory locations allocated by
653 * the init_shared_mem() function and return it to the kernel.
654 */
655
656static void free_shared_mem(struct s2io_nic *nic)
657{
658 int i, j, blk_cnt, size;
659 void *tmp_v_addr;
660 dma_addr_t tmp_p_addr;
661 mac_info_t *mac_control;
662 struct config_param *config;
663 int lst_size, lst_per_page;
776bd20f 664 struct net_device *dev = nic->dev;
1da177e4
LT
665
666 if (!nic)
667 return;
668
669 mac_control = &nic->mac_control;
670 config = &nic->config;
671
672 lst_size = (sizeof(TxD_t) * config->max_txds);
673 lst_per_page = PAGE_SIZE / lst_size;
674
675 for (i = 0; i < config->tx_fifo_num; i++) {
676 int page_num = TXD_MEM_PAGE_CNT(config->tx_cfg[i].fifo_len,
677 lst_per_page);
678 for (j = 0; j < page_num; j++) {
679 int mem_blks = (j * lst_per_page);
776bd20f 680 if (!mac_control->fifos[i].list_info)
681 return;
682 if (!mac_control->fifos[i].list_info[mem_blks].
683 list_virt_addr)
1da177e4
LT
684 break;
685 pci_free_consistent(nic->pdev, PAGE_SIZE,
20346722 686 mac_control->fifos[i].
687 list_info[mem_blks].
1da177e4 688 list_virt_addr,
20346722 689 mac_control->fifos[i].
690 list_info[mem_blks].
1da177e4
LT
691 list_phy_addr);
692 }
776bd20f 693 /* If we got a zero DMA address during allocation,
694 * free the page now
695 */
696 if (mac_control->zerodma_virt_addr) {
697 pci_free_consistent(nic->pdev, PAGE_SIZE,
698 mac_control->zerodma_virt_addr,
699 (dma_addr_t)0);
700 DBG_PRINT(INIT_DBG,
6b4d617d
AM
701 "%s: Freeing TxDL with zero DMA addr. ",
702 dev->name);
703 DBG_PRINT(INIT_DBG, "Virtual address %p\n",
704 mac_control->zerodma_virt_addr);
776bd20f 705 }
20346722 706 kfree(mac_control->fifos[i].list_info);
1da177e4
LT
707 }
708
1da177e4 709 size = SIZE_OF_BLOCK;
1da177e4 710 for (i = 0; i < config->rx_ring_num; i++) {
20346722 711 blk_cnt = mac_control->rings[i].block_count;
1da177e4 712 for (j = 0; j < blk_cnt; j++) {
20346722 713 tmp_v_addr = mac_control->rings[i].rx_blocks[j].
714 block_virt_addr;
715 tmp_p_addr = mac_control->rings[i].rx_blocks[j].
716 block_dma_addr;
1da177e4
LT
717 if (tmp_v_addr == NULL)
718 break;
719 pci_free_consistent(nic->pdev, size,
720 tmp_v_addr, tmp_p_addr);
da6971d8 721 kfree(mac_control->rings[i].rx_blocks[j].rxds);
1da177e4
LT
722 }
723 }
724
da6971d8
AR
725 if (nic->rxd_mode >= RXD_MODE_3A) {
726 /* Freeing buffer storage addresses in 2BUFF mode. */
727 for (i = 0; i < config->rx_ring_num; i++) {
728 blk_cnt = config->rx_cfg[i].num_rxd /
729 (rxd_count[nic->rxd_mode] + 1);
730 for (j = 0; j < blk_cnt; j++) {
731 int k = 0;
732 if (!mac_control->rings[i].ba[j])
733 continue;
734 while (k != rxd_count[nic->rxd_mode]) {
735 buffAdd_t *ba =
736 &mac_control->rings[i].ba[j][k];
737 kfree(ba->ba_0_org);
738 kfree(ba->ba_1_org);
739 k++;
740 }
741 kfree(mac_control->rings[i].ba[j]);
1da177e4 742 }
da6971d8 743 kfree(mac_control->rings[i].ba);
1da177e4 744 }
1da177e4 745 }
1da177e4
LT
746
747 if (mac_control->stats_mem) {
748 pci_free_consistent(nic->pdev,
749 mac_control->stats_mem_sz,
750 mac_control->stats_mem,
751 mac_control->stats_mem_phy);
752 }
fed5eccd
AR
753 if (nic->ufo_in_band_v)
754 kfree(nic->ufo_in_band_v);
1da177e4
LT
755}
756
541ae68f 757/**
758 * s2io_verify_pci_mode -
759 */
760
761static int s2io_verify_pci_mode(nic_t *nic)
762{
509a2671 763 XENA_dev_config_t __iomem *bar0 = nic->bar0;
541ae68f 764 register u64 val64 = 0;
765 int mode;
766
767 val64 = readq(&bar0->pci_mode);
768 mode = (u8)GET_PCI_MODE(val64);
769
770 if ( val64 & PCI_MODE_UNKNOWN_MODE)
771 return -1; /* Unknown PCI mode */
772 return mode;
773}
774
775
776/**
777 * s2io_print_pci_mode -
778 */
779static int s2io_print_pci_mode(nic_t *nic)
780{
509a2671 781 XENA_dev_config_t __iomem *bar0 = nic->bar0;
541ae68f 782 register u64 val64 = 0;
783 int mode;
784 struct config_param *config = &nic->config;
785
786 val64 = readq(&bar0->pci_mode);
787 mode = (u8)GET_PCI_MODE(val64);
788
789 if ( val64 & PCI_MODE_UNKNOWN_MODE)
790 return -1; /* Unknown PCI mode */
791
792 if (val64 & PCI_MODE_32_BITS) {
793 DBG_PRINT(ERR_DBG, "%s: Device is on 32 bit ", nic->dev->name);
794 } else {
795 DBG_PRINT(ERR_DBG, "%s: Device is on 64 bit ", nic->dev->name);
796 }
797
798 switch(mode) {
799 case PCI_MODE_PCI_33:
800 DBG_PRINT(ERR_DBG, "33MHz PCI bus\n");
801 config->bus_speed = 33;
802 break;
803 case PCI_MODE_PCI_66:
804 DBG_PRINT(ERR_DBG, "66MHz PCI bus\n");
805 config->bus_speed = 133;
806 break;
807 case PCI_MODE_PCIX_M1_66:
808 DBG_PRINT(ERR_DBG, "66MHz PCIX(M1) bus\n");
809 config->bus_speed = 133; /* Herc doubles the clock rate */
810 break;
811 case PCI_MODE_PCIX_M1_100:
812 DBG_PRINT(ERR_DBG, "100MHz PCIX(M1) bus\n");
813 config->bus_speed = 200;
814 break;
815 case PCI_MODE_PCIX_M1_133:
816 DBG_PRINT(ERR_DBG, "133MHz PCIX(M1) bus\n");
817 config->bus_speed = 266;
818 break;
819 case PCI_MODE_PCIX_M2_66:
820 DBG_PRINT(ERR_DBG, "133MHz PCIX(M2) bus\n");
821 config->bus_speed = 133;
822 break;
823 case PCI_MODE_PCIX_M2_100:
824 DBG_PRINT(ERR_DBG, "200MHz PCIX(M2) bus\n");
825 config->bus_speed = 200;
826 break;
827 case PCI_MODE_PCIX_M2_133:
828 DBG_PRINT(ERR_DBG, "266MHz PCIX(M2) bus\n");
829 config->bus_speed = 266;
830 break;
831 default:
832 return -1; /* Unsupported bus speed */
833 }
834
835 return mode;
836}
837
20346722 838/**
839 * init_nic - Initialization of hardware
1da177e4 840 * @nic: device peivate variable
20346722 841 * Description: The function sequentially configures every block
842 * of the H/W from their reset values.
843 * Return Value: SUCCESS on success and
1da177e4
LT
844 * '-1' on failure (endian settings incorrect).
845 */
846
847static int init_nic(struct s2io_nic *nic)
848{
849 XENA_dev_config_t __iomem *bar0 = nic->bar0;
850 struct net_device *dev = nic->dev;
851 register u64 val64 = 0;
852 void __iomem *add;
853 u32 time;
854 int i, j;
855 mac_info_t *mac_control;
856 struct config_param *config;
857 int mdio_cnt = 0, dtx_cnt = 0;
858 unsigned long long mem_share;
20346722 859 int mem_size;
1da177e4
LT
860
861 mac_control = &nic->mac_control;
862 config = &nic->config;
863
5e25b9dd 864 /* to set the swapper controle on the card */
20346722 865 if(s2io_set_swapper(nic)) {
1da177e4
LT
866 DBG_PRINT(ERR_DBG,"ERROR: Setting Swapper failed\n");
867 return -1;
868 }
869
541ae68f 870 /*
871 * Herc requires EOI to be removed from reset before XGXS, so..
872 */
873 if (nic->device_type & XFRAME_II_DEVICE) {
874 val64 = 0xA500000000ULL;
875 writeq(val64, &bar0->sw_reset);
876 msleep(500);
877 val64 = readq(&bar0->sw_reset);
878 }
879
1da177e4
LT
880 /* Remove XGXS from reset state */
881 val64 = 0;
882 writeq(val64, &bar0->sw_reset);
1da177e4 883 msleep(500);
20346722 884 val64 = readq(&bar0->sw_reset);
1da177e4
LT
885
886 /* Enable Receiving broadcasts */
887 add = &bar0->mac_cfg;
888 val64 = readq(&bar0->mac_cfg);
889 val64 |= MAC_RMAC_BCAST_ENABLE;
890 writeq(RMAC_CFG_KEY(0x4C0D), &bar0->rmac_cfg_key);
891 writel((u32) val64, add);
892 writeq(RMAC_CFG_KEY(0x4C0D), &bar0->rmac_cfg_key);
893 writel((u32) (val64 >> 32), (add + 4));
894
895 /* Read registers in all blocks */
896 val64 = readq(&bar0->mac_int_mask);
897 val64 = readq(&bar0->mc_int_mask);
898 val64 = readq(&bar0->xgxs_int_mask);
899
900 /* Set MTU */
901 val64 = dev->mtu;
902 writeq(vBIT(val64, 2, 14), &bar0->rmac_max_pyld_len);
903
20346722 904 /*
905 * Configuring the XAUI Interface of Xena.
1da177e4 906 * ***************************************
20346722 907 * To Configure the Xena's XAUI, one has to write a series
908 * of 64 bit values into two registers in a particular
909 * sequence. Hence a macro 'SWITCH_SIGN' has been defined
910 * which will be defined in the array of configuration values
541ae68f 911 * (xena_dtx_cfg & xena_mdio_cfg) at appropriate places
20346722 912 * to switch writing from one regsiter to another. We continue
1da177e4 913 * writing these values until we encounter the 'END_SIGN' macro.
20346722 914 * For example, After making a series of 21 writes into
915 * dtx_control register the 'SWITCH_SIGN' appears and hence we
1da177e4
LT
916 * start writing into mdio_control until we encounter END_SIGN.
917 */
541ae68f 918 if (nic->device_type & XFRAME_II_DEVICE) {
919 while (herc_act_dtx_cfg[dtx_cnt] != END_SIGN) {
303bcb4b 920 SPECIAL_REG_WRITE(herc_act_dtx_cfg[dtx_cnt],
1da177e4 921 &bar0->dtx_control, UF);
541ae68f 922 if (dtx_cnt & 0x1)
923 msleep(1); /* Necessary!! */
1da177e4
LT
924 dtx_cnt++;
925 }
541ae68f 926 } else {
927 while (1) {
928 dtx_cfg:
929 while (xena_dtx_cfg[dtx_cnt] != END_SIGN) {
930 if (xena_dtx_cfg[dtx_cnt] == SWITCH_SIGN) {
931 dtx_cnt++;
932 goto mdio_cfg;
933 }
934 SPECIAL_REG_WRITE(xena_dtx_cfg[dtx_cnt],
935 &bar0->dtx_control, UF);
936 val64 = readq(&bar0->dtx_control);
937 dtx_cnt++;
938 }
939 mdio_cfg:
940 while (xena_mdio_cfg[mdio_cnt] != END_SIGN) {
941 if (xena_mdio_cfg[mdio_cnt] == SWITCH_SIGN) {
942 mdio_cnt++;
943 goto dtx_cfg;
944 }
945 SPECIAL_REG_WRITE(xena_mdio_cfg[mdio_cnt],
946 &bar0->mdio_control, UF);
947 val64 = readq(&bar0->mdio_control);
1da177e4 948 mdio_cnt++;
541ae68f 949 }
950 if ((xena_dtx_cfg[dtx_cnt] == END_SIGN) &&
951 (xena_mdio_cfg[mdio_cnt] == END_SIGN)) {
952 break;
953 } else {
1da177e4
LT
954 goto dtx_cfg;
955 }
1da177e4
LT
956 }
957 }
958
959 /* Tx DMA Initialization */
960 val64 = 0;
961 writeq(val64, &bar0->tx_fifo_partition_0);
962 writeq(val64, &bar0->tx_fifo_partition_1);
963 writeq(val64, &bar0->tx_fifo_partition_2);
964 writeq(val64, &bar0->tx_fifo_partition_3);
965
966
967 for (i = 0, j = 0; i < config->tx_fifo_num; i++) {
968 val64 |=
969 vBIT(config->tx_cfg[i].fifo_len - 1, ((i * 32) + 19),
970 13) | vBIT(config->tx_cfg[i].fifo_priority,
971 ((i * 32) + 5), 3);
972
973 if (i == (config->tx_fifo_num - 1)) {
974 if (i % 2 == 0)
975 i++;
976 }
977
978 switch (i) {
979 case 1:
980 writeq(val64, &bar0->tx_fifo_partition_0);
981 val64 = 0;
982 break;
983 case 3:
984 writeq(val64, &bar0->tx_fifo_partition_1);
985 val64 = 0;
986 break;
987 case 5:
988 writeq(val64, &bar0->tx_fifo_partition_2);
989 val64 = 0;
990 break;
991 case 7:
992 writeq(val64, &bar0->tx_fifo_partition_3);
993 break;
994 }
995 }
996
997 /* Enable Tx FIFO partition 0. */
998 val64 = readq(&bar0->tx_fifo_partition_0);
999 val64 |= BIT(0); /* To enable the FIFO partition. */
1000 writeq(val64, &bar0->tx_fifo_partition_0);
1001
5e25b9dd 1002 /*
1003 * Disable 4 PCCs for Xena1, 2 and 3 as per H/W bug
1004 * SXE-008 TRANSMIT DMA ARBITRATION ISSUE.
1005 */
541ae68f 1006 if ((nic->device_type == XFRAME_I_DEVICE) &&
1007 (get_xena_rev_id(nic->pdev) < 4))
5e25b9dd 1008 writeq(PCC_ENABLE_FOUR, &bar0->pcc_enable);
1009
1da177e4
LT
1010 val64 = readq(&bar0->tx_fifo_partition_0);
1011 DBG_PRINT(INIT_DBG, "Fifo partition at: 0x%p is: 0x%llx\n",
1012 &bar0->tx_fifo_partition_0, (unsigned long long) val64);
1013
20346722 1014 /*
1015 * Initialization of Tx_PA_CONFIG register to ignore packet
1da177e4
LT
1016 * integrity checking.
1017 */
1018 val64 = readq(&bar0->tx_pa_cfg);
1019 val64 |= TX_PA_CFG_IGNORE_FRM_ERR | TX_PA_CFG_IGNORE_SNAP_OUI |
1020 TX_PA_CFG_IGNORE_LLC_CTRL | TX_PA_CFG_IGNORE_L2_ERR;
1021 writeq(val64, &bar0->tx_pa_cfg);
1022
1023 /* Rx DMA intialization. */
1024 val64 = 0;
1025 for (i = 0; i < config->rx_ring_num; i++) {
1026 val64 |=
1027 vBIT(config->rx_cfg[i].ring_priority, (5 + (i * 8)),
1028 3);
1029 }
1030 writeq(val64, &bar0->rx_queue_priority);
1031
20346722 1032 /*
1033 * Allocating equal share of memory to all the
1da177e4
LT
1034 * configured Rings.
1035 */
1036 val64 = 0;
541ae68f 1037 if (nic->device_type & XFRAME_II_DEVICE)
1038 mem_size = 32;
1039 else
1040 mem_size = 64;
1041
1da177e4
LT
1042 for (i = 0; i < config->rx_ring_num; i++) {
1043 switch (i) {
1044 case 0:
20346722 1045 mem_share = (mem_size / config->rx_ring_num +
1046 mem_size % config->rx_ring_num);
1da177e4
LT
1047 val64 |= RX_QUEUE_CFG_Q0_SZ(mem_share);
1048 continue;
1049 case 1:
20346722 1050 mem_share = (mem_size / config->rx_ring_num);
1da177e4
LT
1051 val64 |= RX_QUEUE_CFG_Q1_SZ(mem_share);
1052 continue;
1053 case 2:
20346722 1054 mem_share = (mem_size / config->rx_ring_num);
1da177e4
LT
1055 val64 |= RX_QUEUE_CFG_Q2_SZ(mem_share);
1056 continue;
1057 case 3:
20346722 1058 mem_share = (mem_size / config->rx_ring_num);
1da177e4
LT
1059 val64 |= RX_QUEUE_CFG_Q3_SZ(mem_share);
1060 continue;
1061 case 4:
20346722 1062 mem_share = (mem_size / config->rx_ring_num);
1da177e4
LT
1063 val64 |= RX_QUEUE_CFG_Q4_SZ(mem_share);
1064 continue;
1065 case 5:
20346722 1066 mem_share = (mem_size / config->rx_ring_num);
1da177e4
LT
1067 val64 |= RX_QUEUE_CFG_Q5_SZ(mem_share);
1068 continue;
1069 case 6:
20346722 1070 mem_share = (mem_size / config->rx_ring_num);
1da177e4
LT
1071 val64 |= RX_QUEUE_CFG_Q6_SZ(mem_share);
1072 continue;
1073 case 7:
20346722 1074 mem_share = (mem_size / config->rx_ring_num);
1da177e4
LT
1075 val64 |= RX_QUEUE_CFG_Q7_SZ(mem_share);
1076 continue;
1077 }
1078 }
1079 writeq(val64, &bar0->rx_queue_cfg);
1080
20346722 1081 /*
5e25b9dd 1082 * Filling Tx round robin registers
1083 * as per the number of FIFOs
1da177e4 1084 */
5e25b9dd 1085 switch (config->tx_fifo_num) {
1086 case 1:
1087 val64 = 0x0000000000000000ULL;
1088 writeq(val64, &bar0->tx_w_round_robin_0);
1089 writeq(val64, &bar0->tx_w_round_robin_1);
1090 writeq(val64, &bar0->tx_w_round_robin_2);
1091 writeq(val64, &bar0->tx_w_round_robin_3);
1092 writeq(val64, &bar0->tx_w_round_robin_4);
1093 break;
1094 case 2:
1095 val64 = 0x0000010000010000ULL;
1096 writeq(val64, &bar0->tx_w_round_robin_0);
1097 val64 = 0x0100000100000100ULL;
1098 writeq(val64, &bar0->tx_w_round_robin_1);
1099 val64 = 0x0001000001000001ULL;
1100 writeq(val64, &bar0->tx_w_round_robin_2);
1101 val64 = 0x0000010000010000ULL;
1102 writeq(val64, &bar0->tx_w_round_robin_3);
1103 val64 = 0x0100000000000000ULL;
1104 writeq(val64, &bar0->tx_w_round_robin_4);
1105 break;
1106 case 3:
1107 val64 = 0x0001000102000001ULL;
1108 writeq(val64, &bar0->tx_w_round_robin_0);
1109 val64 = 0x0001020000010001ULL;
1110 writeq(val64, &bar0->tx_w_round_robin_1);
1111 val64 = 0x0200000100010200ULL;
1112 writeq(val64, &bar0->tx_w_round_robin_2);
1113 val64 = 0x0001000102000001ULL;
1114 writeq(val64, &bar0->tx_w_round_robin_3);
1115 val64 = 0x0001020000000000ULL;
1116 writeq(val64, &bar0->tx_w_round_robin_4);
1117 break;
1118 case 4:
1119 val64 = 0x0001020300010200ULL;
1120 writeq(val64, &bar0->tx_w_round_robin_0);
1121 val64 = 0x0100000102030001ULL;
1122 writeq(val64, &bar0->tx_w_round_robin_1);
1123 val64 = 0x0200010000010203ULL;
1124 writeq(val64, &bar0->tx_w_round_robin_2);
1125 val64 = 0x0001020001000001ULL;
1126 writeq(val64, &bar0->tx_w_round_robin_3);
1127 val64 = 0x0203000100000000ULL;
1128 writeq(val64, &bar0->tx_w_round_robin_4);
1129 break;
1130 case 5:
1131 val64 = 0x0001000203000102ULL;
1132 writeq(val64, &bar0->tx_w_round_robin_0);
1133 val64 = 0x0001020001030004ULL;
1134 writeq(val64, &bar0->tx_w_round_robin_1);
1135 val64 = 0x0001000203000102ULL;
1136 writeq(val64, &bar0->tx_w_round_robin_2);
1137 val64 = 0x0001020001030004ULL;
1138 writeq(val64, &bar0->tx_w_round_robin_3);
1139 val64 = 0x0001000000000000ULL;
1140 writeq(val64, &bar0->tx_w_round_robin_4);
1141 break;
1142 case 6:
1143 val64 = 0x0001020304000102ULL;
1144 writeq(val64, &bar0->tx_w_round_robin_0);
1145 val64 = 0x0304050001020001ULL;
1146 writeq(val64, &bar0->tx_w_round_robin_1);
1147 val64 = 0x0203000100000102ULL;
1148 writeq(val64, &bar0->tx_w_round_robin_2);
1149 val64 = 0x0304000102030405ULL;
1150 writeq(val64, &bar0->tx_w_round_robin_3);
1151 val64 = 0x0001000200000000ULL;
1152 writeq(val64, &bar0->tx_w_round_robin_4);
1153 break;
1154 case 7:
1155 val64 = 0x0001020001020300ULL;
1156 writeq(val64, &bar0->tx_w_round_robin_0);
1157 val64 = 0x0102030400010203ULL;
1158 writeq(val64, &bar0->tx_w_round_robin_1);
1159 val64 = 0x0405060001020001ULL;
1160 writeq(val64, &bar0->tx_w_round_robin_2);
1161 val64 = 0x0304050000010200ULL;
1162 writeq(val64, &bar0->tx_w_round_robin_3);
1163 val64 = 0x0102030000000000ULL;
1164 writeq(val64, &bar0->tx_w_round_robin_4);
1165 break;
1166 case 8:
1167 val64 = 0x0001020300040105ULL;
1168 writeq(val64, &bar0->tx_w_round_robin_0);
1169 val64 = 0x0200030106000204ULL;
1170 writeq(val64, &bar0->tx_w_round_robin_1);
1171 val64 = 0x0103000502010007ULL;
1172 writeq(val64, &bar0->tx_w_round_robin_2);
1173 val64 = 0x0304010002060500ULL;
1174 writeq(val64, &bar0->tx_w_round_robin_3);
1175 val64 = 0x0103020400000000ULL;
1176 writeq(val64, &bar0->tx_w_round_robin_4);
1177 break;
1178 }
1179
1180 /* Filling the Rx round robin registers as per the
1181 * number of Rings and steering based on QoS.
1182 */
1183 switch (config->rx_ring_num) {
1184 case 1:
1185 val64 = 0x8080808080808080ULL;
1186 writeq(val64, &bar0->rts_qos_steering);
1187 break;
1188 case 2:
1189 val64 = 0x0000010000010000ULL;
1190 writeq(val64, &bar0->rx_w_round_robin_0);
1191 val64 = 0x0100000100000100ULL;
1192 writeq(val64, &bar0->rx_w_round_robin_1);
1193 val64 = 0x0001000001000001ULL;
1194 writeq(val64, &bar0->rx_w_round_robin_2);
1195 val64 = 0x0000010000010000ULL;
1196 writeq(val64, &bar0->rx_w_round_robin_3);
1197 val64 = 0x0100000000000000ULL;
1198 writeq(val64, &bar0->rx_w_round_robin_4);
1199
1200 val64 = 0x8080808040404040ULL;
1201 writeq(val64, &bar0->rts_qos_steering);
1202 break;
1203 case 3:
1204 val64 = 0x0001000102000001ULL;
1205 writeq(val64, &bar0->rx_w_round_robin_0);
1206 val64 = 0x0001020000010001ULL;
1207 writeq(val64, &bar0->rx_w_round_robin_1);
1208 val64 = 0x0200000100010200ULL;
1209 writeq(val64, &bar0->rx_w_round_robin_2);
1210 val64 = 0x0001000102000001ULL;
1211 writeq(val64, &bar0->rx_w_round_robin_3);
1212 val64 = 0x0001020000000000ULL;
1213 writeq(val64, &bar0->rx_w_round_robin_4);
1214
1215 val64 = 0x8080804040402020ULL;
1216 writeq(val64, &bar0->rts_qos_steering);
1217 break;
1218 case 4:
1219 val64 = 0x0001020300010200ULL;
1220 writeq(val64, &bar0->rx_w_round_robin_0);
1221 val64 = 0x0100000102030001ULL;
1222 writeq(val64, &bar0->rx_w_round_robin_1);
1223 val64 = 0x0200010000010203ULL;
1224 writeq(val64, &bar0->rx_w_round_robin_2);
1225 val64 = 0x0001020001000001ULL;
1226 writeq(val64, &bar0->rx_w_round_robin_3);
1227 val64 = 0x0203000100000000ULL;
1228 writeq(val64, &bar0->rx_w_round_robin_4);
1229
1230 val64 = 0x8080404020201010ULL;
1231 writeq(val64, &bar0->rts_qos_steering);
1232 break;
1233 case 5:
1234 val64 = 0x0001000203000102ULL;
1235 writeq(val64, &bar0->rx_w_round_robin_0);
1236 val64 = 0x0001020001030004ULL;
1237 writeq(val64, &bar0->rx_w_round_robin_1);
1238 val64 = 0x0001000203000102ULL;
1239 writeq(val64, &bar0->rx_w_round_robin_2);
1240 val64 = 0x0001020001030004ULL;
1241 writeq(val64, &bar0->rx_w_round_robin_3);
1242 val64 = 0x0001000000000000ULL;
1243 writeq(val64, &bar0->rx_w_round_robin_4);
1244
1245 val64 = 0x8080404020201008ULL;
1246 writeq(val64, &bar0->rts_qos_steering);
1247 break;
1248 case 6:
1249 val64 = 0x0001020304000102ULL;
1250 writeq(val64, &bar0->rx_w_round_robin_0);
1251 val64 = 0x0304050001020001ULL;
1252 writeq(val64, &bar0->rx_w_round_robin_1);
1253 val64 = 0x0203000100000102ULL;
1254 writeq(val64, &bar0->rx_w_round_robin_2);
1255 val64 = 0x0304000102030405ULL;
1256 writeq(val64, &bar0->rx_w_round_robin_3);
1257 val64 = 0x0001000200000000ULL;
1258 writeq(val64, &bar0->rx_w_round_robin_4);
1259
1260 val64 = 0x8080404020100804ULL;
1261 writeq(val64, &bar0->rts_qos_steering);
1262 break;
1263 case 7:
1264 val64 = 0x0001020001020300ULL;
1265 writeq(val64, &bar0->rx_w_round_robin_0);
1266 val64 = 0x0102030400010203ULL;
1267 writeq(val64, &bar0->rx_w_round_robin_1);
1268 val64 = 0x0405060001020001ULL;
1269 writeq(val64, &bar0->rx_w_round_robin_2);
1270 val64 = 0x0304050000010200ULL;
1271 writeq(val64, &bar0->rx_w_round_robin_3);
1272 val64 = 0x0102030000000000ULL;
1273 writeq(val64, &bar0->rx_w_round_robin_4);
1274
1275 val64 = 0x8080402010080402ULL;
1276 writeq(val64, &bar0->rts_qos_steering);
1277 break;
1278 case 8:
1279 val64 = 0x0001020300040105ULL;
1280 writeq(val64, &bar0->rx_w_round_robin_0);
1281 val64 = 0x0200030106000204ULL;
1282 writeq(val64, &bar0->rx_w_round_robin_1);
1283 val64 = 0x0103000502010007ULL;
1284 writeq(val64, &bar0->rx_w_round_robin_2);
1285 val64 = 0x0304010002060500ULL;
1286 writeq(val64, &bar0->rx_w_round_robin_3);
1287 val64 = 0x0103020400000000ULL;
1288 writeq(val64, &bar0->rx_w_round_robin_4);
1289
1290 val64 = 0x8040201008040201ULL;
1291 writeq(val64, &bar0->rts_qos_steering);
1292 break;
1293 }
1da177e4
LT
1294
1295 /* UDP Fix */
1296 val64 = 0;
20346722 1297 for (i = 0; i < 8; i++)
1da177e4
LT
1298 writeq(val64, &bar0->rts_frm_len_n[i]);
1299
5e25b9dd 1300 /* Set the default rts frame length for the rings configured */
1301 val64 = MAC_RTS_FRM_LEN_SET(dev->mtu+22);
1302 for (i = 0 ; i < config->rx_ring_num ; i++)
1303 writeq(val64, &bar0->rts_frm_len_n[i]);
1304
1305 /* Set the frame length for the configured rings
1306 * desired by the user
1307 */
1308 for (i = 0; i < config->rx_ring_num; i++) {
1309 /* If rts_frm_len[i] == 0 then it is assumed that user not
1310 * specified frame length steering.
1311 * If the user provides the frame length then program
1312 * the rts_frm_len register for those values or else
1313 * leave it as it is.
1314 */
1315 if (rts_frm_len[i] != 0) {
1316 writeq(MAC_RTS_FRM_LEN_SET(rts_frm_len[i]),
1317 &bar0->rts_frm_len_n[i]);
1318 }
1319 }
1da177e4 1320
20346722 1321 /* Program statistics memory */
1da177e4 1322 writeq(mac_control->stats_mem_phy, &bar0->stat_addr);
1da177e4 1323
541ae68f 1324 if (nic->device_type == XFRAME_II_DEVICE) {
1325 val64 = STAT_BC(0x320);
1326 writeq(val64, &bar0->stat_byte_cnt);
1327 }
1328
20346722 1329 /*
1da177e4
LT
1330 * Initializing the sampling rate for the device to calculate the
1331 * bandwidth utilization.
1332 */
1333 val64 = MAC_TX_LINK_UTIL_VAL(tmac_util_period) |
1334 MAC_RX_LINK_UTIL_VAL(rmac_util_period);
1335 writeq(val64, &bar0->mac_link_util);
1336
1337
20346722 1338 /*
1339 * Initializing the Transmit and Receive Traffic Interrupt
1da177e4
LT
1340 * Scheme.
1341 */
20346722 1342 /*
1343 * TTI Initialization. Default Tx timer gets us about
1da177e4
LT
1344 * 250 interrupts per sec. Continuous interrupts are enabled
1345 * by default.
1346 */
541ae68f 1347 if (nic->device_type == XFRAME_II_DEVICE) {
1348 int count = (nic->config.bus_speed * 125)/2;
1349 val64 = TTI_DATA1_MEM_TX_TIMER_VAL(count);
1350 } else {
1351
1352 val64 = TTI_DATA1_MEM_TX_TIMER_VAL(0x2078);
1353 }
1354 val64 |= TTI_DATA1_MEM_TX_URNG_A(0xA) |
1da177e4 1355 TTI_DATA1_MEM_TX_URNG_B(0x10) |
5e25b9dd 1356 TTI_DATA1_MEM_TX_URNG_C(0x30) | TTI_DATA1_MEM_TX_TIMER_AC_EN;
541ae68f 1357 if (use_continuous_tx_intrs)
1358 val64 |= TTI_DATA1_MEM_TX_TIMER_CI_EN;
1da177e4
LT
1359 writeq(val64, &bar0->tti_data1_mem);
1360
1361 val64 = TTI_DATA2_MEM_TX_UFC_A(0x10) |
1362 TTI_DATA2_MEM_TX_UFC_B(0x20) |
5e25b9dd 1363 TTI_DATA2_MEM_TX_UFC_C(0x70) | TTI_DATA2_MEM_TX_UFC_D(0x80);
1da177e4
LT
1364 writeq(val64, &bar0->tti_data2_mem);
1365
1366 val64 = TTI_CMD_MEM_WE | TTI_CMD_MEM_STROBE_NEW_CMD;
1367 writeq(val64, &bar0->tti_command_mem);
1368
20346722 1369 /*
1da177e4
LT
1370 * Once the operation completes, the Strobe bit of the command
1371 * register will be reset. We poll for this particular condition
1372 * We wait for a maximum of 500ms for the operation to complete,
1373 * if it's not complete by then we return error.
1374 */
1375 time = 0;
1376 while (TRUE) {
1377 val64 = readq(&bar0->tti_command_mem);
1378 if (!(val64 & TTI_CMD_MEM_STROBE_NEW_CMD)) {
1379 break;
1380 }
1381 if (time > 10) {
1382 DBG_PRINT(ERR_DBG, "%s: TTI init Failed\n",
1383 dev->name);
1384 return -1;
1385 }
1386 msleep(50);
1387 time++;
1388 }
1389
b6e3f982 1390 if (nic->config.bimodal) {
1391 int k = 0;
1392 for (k = 0; k < config->rx_ring_num; k++) {
1393 val64 = TTI_CMD_MEM_WE | TTI_CMD_MEM_STROBE_NEW_CMD;
1394 val64 |= TTI_CMD_MEM_OFFSET(0x38+k);
1395 writeq(val64, &bar0->tti_command_mem);
541ae68f 1396
541ae68f 1397 /*
b6e3f982 1398 * Once the operation completes, the Strobe bit of the command
1399 * register will be reset. We poll for this particular condition
1400 * We wait for a maximum of 500ms for the operation to complete,
1401 * if it's not complete by then we return error.
1402 */
1403 time = 0;
1404 while (TRUE) {
1405 val64 = readq(&bar0->tti_command_mem);
1406 if (!(val64 & TTI_CMD_MEM_STROBE_NEW_CMD)) {
1407 break;
1408 }
1409 if (time > 10) {
1410 DBG_PRINT(ERR_DBG,
1411 "%s: TTI init Failed\n",
1412 dev->name);
1413 return -1;
1414 }
1415 time++;
1416 msleep(50);
1417 }
1418 }
541ae68f 1419 } else {
1da177e4 1420
b6e3f982 1421 /* RTI Initialization */
1422 if (nic->device_type == XFRAME_II_DEVICE) {
1423 /*
1424 * Programmed to generate Apprx 500 Intrs per
1425 * second
1426 */
1427 int count = (nic->config.bus_speed * 125)/4;
1428 val64 = RTI_DATA1_MEM_RX_TIMER_VAL(count);
1429 } else {
1430 val64 = RTI_DATA1_MEM_RX_TIMER_VAL(0xFFF);
1431 }
1432 val64 |= RTI_DATA1_MEM_RX_URNG_A(0xA) |
1433 RTI_DATA1_MEM_RX_URNG_B(0x10) |
1434 RTI_DATA1_MEM_RX_URNG_C(0x30) | RTI_DATA1_MEM_RX_TIMER_AC_EN;
1da177e4 1435
b6e3f982 1436 writeq(val64, &bar0->rti_data1_mem);
1da177e4 1437
b6e3f982 1438 val64 = RTI_DATA2_MEM_RX_UFC_A(0x1) |
cc6e7c44
RA
1439 RTI_DATA2_MEM_RX_UFC_B(0x2) ;
1440 if (nic->intr_type == MSI_X)
1441 val64 |= (RTI_DATA2_MEM_RX_UFC_C(0x20) | \
1442 RTI_DATA2_MEM_RX_UFC_D(0x40));
1443 else
1444 val64 |= (RTI_DATA2_MEM_RX_UFC_C(0x40) | \
1445 RTI_DATA2_MEM_RX_UFC_D(0x80));
b6e3f982 1446 writeq(val64, &bar0->rti_data2_mem);
1da177e4 1447
b6e3f982 1448 for (i = 0; i < config->rx_ring_num; i++) {
1449 val64 = RTI_CMD_MEM_WE | RTI_CMD_MEM_STROBE_NEW_CMD
1450 | RTI_CMD_MEM_OFFSET(i);
1451 writeq(val64, &bar0->rti_command_mem);
1452
1453 /*
1454 * Once the operation completes, the Strobe bit of the
1455 * command register will be reset. We poll for this
1456 * particular condition. We wait for a maximum of 500ms
1457 * for the operation to complete, if it's not complete
1458 * by then we return error.
1459 */
1460 time = 0;
1461 while (TRUE) {
1462 val64 = readq(&bar0->rti_command_mem);
1463 if (!(val64 & RTI_CMD_MEM_STROBE_NEW_CMD)) {
1464 break;
1465 }
1466 if (time > 10) {
1467 DBG_PRINT(ERR_DBG, "%s: RTI init Failed\n",
1468 dev->name);
1469 return -1;
1470 }
1471 time++;
1472 msleep(50);
1473 }
1da177e4 1474 }
1da177e4
LT
1475 }
1476
20346722 1477 /*
1478 * Initializing proper values as Pause threshold into all
1da177e4
LT
1479 * the 8 Queues on Rx side.
1480 */
1481 writeq(0xffbbffbbffbbffbbULL, &bar0->mc_pause_thresh_q0q3);
1482 writeq(0xffbbffbbffbbffbbULL, &bar0->mc_pause_thresh_q4q7);
1483
1484 /* Disable RMAC PAD STRIPPING */
509a2671 1485 add = &bar0->mac_cfg;
1da177e4
LT
1486 val64 = readq(&bar0->mac_cfg);
1487 val64 &= ~(MAC_CFG_RMAC_STRIP_PAD);
1488 writeq(RMAC_CFG_KEY(0x4C0D), &bar0->rmac_cfg_key);
1489 writel((u32) (val64), add);
1490 writeq(RMAC_CFG_KEY(0x4C0D), &bar0->rmac_cfg_key);
1491 writel((u32) (val64 >> 32), (add + 4));
1492 val64 = readq(&bar0->mac_cfg);
1493
7d3d0439
RA
1494 /* Enable FCS stripping by adapter */
1495 add = &bar0->mac_cfg;
1496 val64 = readq(&bar0->mac_cfg);
1497 val64 |= MAC_CFG_RMAC_STRIP_FCS;
1498 if (nic->device_type == XFRAME_II_DEVICE)
1499 writeq(val64, &bar0->mac_cfg);
1500 else {
1501 writeq(RMAC_CFG_KEY(0x4C0D), &bar0->rmac_cfg_key);
1502 writel((u32) (val64), add);
1503 writeq(RMAC_CFG_KEY(0x4C0D), &bar0->rmac_cfg_key);
1504 writel((u32) (val64 >> 32), (add + 4));
1505 }
1506
20346722 1507 /*
1508 * Set the time value to be inserted in the pause frame
1da177e4
LT
1509 * generated by xena.
1510 */
1511 val64 = readq(&bar0->rmac_pause_cfg);
1512 val64 &= ~(RMAC_PAUSE_HG_PTIME(0xffff));
1513 val64 |= RMAC_PAUSE_HG_PTIME(nic->mac_control.rmac_pause_time);
1514 writeq(val64, &bar0->rmac_pause_cfg);
1515
20346722 1516 /*
1da177e4
LT
1517 * Set the Threshold Limit for Generating the pause frame
1518 * If the amount of data in any Queue exceeds ratio of
1519 * (mac_control.mc_pause_threshold_q0q3 or q4q7)/256
1520 * pause frame is generated
1521 */
1522 val64 = 0;
1523 for (i = 0; i < 4; i++) {
1524 val64 |=
1525 (((u64) 0xFF00 | nic->mac_control.
1526 mc_pause_threshold_q0q3)
1527 << (i * 2 * 8));
1528 }
1529 writeq(val64, &bar0->mc_pause_thresh_q0q3);
1530
1531 val64 = 0;
1532 for (i = 0; i < 4; i++) {
1533 val64 |=
1534 (((u64) 0xFF00 | nic->mac_control.
1535 mc_pause_threshold_q4q7)
1536 << (i * 2 * 8));
1537 }
1538 writeq(val64, &bar0->mc_pause_thresh_q4q7);
1539
20346722 1540 /*
1541 * TxDMA will stop Read request if the number of read split has
1da177e4
LT
1542 * exceeded the limit pointed by shared_splits
1543 */
1544 val64 = readq(&bar0->pic_control);
1545 val64 |= PIC_CNTL_SHARED_SPLITS(shared_splits);
1546 writeq(val64, &bar0->pic_control);
1547
541ae68f 1548 /*
1549 * Programming the Herc to split every write transaction
1550 * that does not start on an ADB to reduce disconnects.
1551 */
1552 if (nic->device_type == XFRAME_II_DEVICE) {
1553 val64 = WREQ_SPLIT_MASK_SET_MASK(255);
1554 writeq(val64, &bar0->wreq_split_mask);
1555 }
1556
a371a07d 1557 /* Setting Link stability period to 64 ms */
1558 if (nic->device_type == XFRAME_II_DEVICE) {
1559 val64 = MISC_LINK_STABILITY_PRD(3);
1560 writeq(val64, &bar0->misc_control);
1561 }
1562
1da177e4
LT
1563 return SUCCESS;
1564}
a371a07d 1565#define LINK_UP_DOWN_INTERRUPT 1
1566#define MAC_RMAC_ERR_TIMER 2
1567
ac1f60db 1568static int s2io_link_fault_indication(nic_t *nic)
a371a07d 1569{
cc6e7c44
RA
1570 if (nic->intr_type != INTA)
1571 return MAC_RMAC_ERR_TIMER;
a371a07d 1572 if (nic->device_type == XFRAME_II_DEVICE)
1573 return LINK_UP_DOWN_INTERRUPT;
1574 else
1575 return MAC_RMAC_ERR_TIMER;
1576}
1da177e4 1577
20346722 1578/**
1579 * en_dis_able_nic_intrs - Enable or Disable the interrupts
1da177e4
LT
1580 * @nic: device private variable,
1581 * @mask: A mask indicating which Intr block must be modified and,
1582 * @flag: A flag indicating whether to enable or disable the Intrs.
1583 * Description: This function will either disable or enable the interrupts
20346722 1584 * depending on the flag argument. The mask argument can be used to
1585 * enable/disable any Intr block.
1da177e4
LT
1586 * Return Value: NONE.
1587 */
1588
1589static void en_dis_able_nic_intrs(struct s2io_nic *nic, u16 mask, int flag)
1590{
1591 XENA_dev_config_t __iomem *bar0 = nic->bar0;
1592 register u64 val64 = 0, temp64 = 0;
1593
1594 /* Top level interrupt classification */
1595 /* PIC Interrupts */
1596 if ((mask & (TX_PIC_INTR | RX_PIC_INTR))) {
1597 /* Enable PIC Intrs in the general intr mask register */
1598 val64 = TXPIC_INT_M | PIC_RX_INT_M;
1599 if (flag == ENABLE_INTRS) {
1600 temp64 = readq(&bar0->general_int_mask);
1601 temp64 &= ~((u64) val64);
1602 writeq(temp64, &bar0->general_int_mask);
20346722 1603 /*
a371a07d 1604 * If Hercules adapter enable GPIO otherwise
1605 * disabled all PCIX, Flash, MDIO, IIC and GPIO
20346722 1606 * interrupts for now.
1607 * TODO
1da177e4 1608 */
a371a07d 1609 if (s2io_link_fault_indication(nic) ==
1610 LINK_UP_DOWN_INTERRUPT ) {
1611 temp64 = readq(&bar0->pic_int_mask);
1612 temp64 &= ~((u64) PIC_INT_GPIO);
1613 writeq(temp64, &bar0->pic_int_mask);
1614 temp64 = readq(&bar0->gpio_int_mask);
1615 temp64 &= ~((u64) GPIO_INT_MASK_LINK_UP);
1616 writeq(temp64, &bar0->gpio_int_mask);
1617 } else {
1618 writeq(DISABLE_ALL_INTRS, &bar0->pic_int_mask);
1619 }
20346722 1620 /*
1da177e4
LT
1621 * No MSI Support is available presently, so TTI and
1622 * RTI interrupts are also disabled.
1623 */
1624 } else if (flag == DISABLE_INTRS) {
20346722 1625 /*
1626 * Disable PIC Intrs in the general
1627 * intr mask register
1da177e4
LT
1628 */
1629 writeq(DISABLE_ALL_INTRS, &bar0->pic_int_mask);
1630 temp64 = readq(&bar0->general_int_mask);
1631 val64 |= temp64;
1632 writeq(val64, &bar0->general_int_mask);
1633 }
1634 }
1635
1636 /* DMA Interrupts */
1637 /* Enabling/Disabling Tx DMA interrupts */
1638 if (mask & TX_DMA_INTR) {
1639 /* Enable TxDMA Intrs in the general intr mask register */
1640 val64 = TXDMA_INT_M;
1641 if (flag == ENABLE_INTRS) {
1642 temp64 = readq(&bar0->general_int_mask);
1643 temp64 &= ~((u64) val64);
1644 writeq(temp64, &bar0->general_int_mask);
20346722 1645 /*
1646 * Keep all interrupts other than PFC interrupt
1da177e4
LT
1647 * and PCC interrupt disabled in DMA level.
1648 */
1649 val64 = DISABLE_ALL_INTRS & ~(TXDMA_PFC_INT_M |
1650 TXDMA_PCC_INT_M);
1651 writeq(val64, &bar0->txdma_int_mask);
20346722 1652 /*
1653 * Enable only the MISC error 1 interrupt in PFC block
1da177e4
LT
1654 */
1655 val64 = DISABLE_ALL_INTRS & (~PFC_MISC_ERR_1);
1656 writeq(val64, &bar0->pfc_err_mask);
20346722 1657 /*
1658 * Enable only the FB_ECC error interrupt in PCC block
1da177e4
LT
1659 */
1660 val64 = DISABLE_ALL_INTRS & (~PCC_FB_ECC_ERR);
1661 writeq(val64, &bar0->pcc_err_mask);
1662 } else if (flag == DISABLE_INTRS) {
20346722 1663 /*
1664 * Disable TxDMA Intrs in the general intr mask
1665 * register
1da177e4
LT
1666 */
1667 writeq(DISABLE_ALL_INTRS, &bar0->txdma_int_mask);
1668 writeq(DISABLE_ALL_INTRS, &bar0->pfc_err_mask);
1669 temp64 = readq(&bar0->general_int_mask);
1670 val64 |= temp64;
1671 writeq(val64, &bar0->general_int_mask);
1672 }
1673 }
1674
1675 /* Enabling/Disabling Rx DMA interrupts */
1676 if (mask & RX_DMA_INTR) {
1677 /* Enable RxDMA Intrs in the general intr mask register */
1678 val64 = RXDMA_INT_M;
1679 if (flag == ENABLE_INTRS) {
1680 temp64 = readq(&bar0->general_int_mask);
1681 temp64 &= ~((u64) val64);
1682 writeq(temp64, &bar0->general_int_mask);
20346722 1683 /*
1684 * All RxDMA block interrupts are disabled for now
1685 * TODO
1da177e4
LT
1686 */
1687 writeq(DISABLE_ALL_INTRS, &bar0->rxdma_int_mask);
1688 } else if (flag == DISABLE_INTRS) {
20346722 1689 /*
1690 * Disable RxDMA Intrs in the general intr mask
1691 * register
1da177e4
LT
1692 */
1693 writeq(DISABLE_ALL_INTRS, &bar0->rxdma_int_mask);
1694 temp64 = readq(&bar0->general_int_mask);
1695 val64 |= temp64;
1696 writeq(val64, &bar0->general_int_mask);
1697 }
1698 }
1699
1700 /* MAC Interrupts */
1701 /* Enabling/Disabling MAC interrupts */
1702 if (mask & (TX_MAC_INTR | RX_MAC_INTR)) {
1703 val64 = TXMAC_INT_M | RXMAC_INT_M;
1704 if (flag == ENABLE_INTRS) {
1705 temp64 = readq(&bar0->general_int_mask);
1706 temp64 &= ~((u64) val64);
1707 writeq(temp64, &bar0->general_int_mask);
20346722 1708 /*
1709 * All MAC block error interrupts are disabled for now
1da177e4
LT
1710 * TODO
1711 */
1da177e4 1712 } else if (flag == DISABLE_INTRS) {
20346722 1713 /*
1714 * Disable MAC Intrs in the general intr mask register
1da177e4
LT
1715 */
1716 writeq(DISABLE_ALL_INTRS, &bar0->mac_int_mask);
1717 writeq(DISABLE_ALL_INTRS,
1718 &bar0->mac_rmac_err_mask);
1719
1720 temp64 = readq(&bar0->general_int_mask);
1721 val64 |= temp64;
1722 writeq(val64, &bar0->general_int_mask);
1723 }
1724 }
1725
1726 /* XGXS Interrupts */
1727 if (mask & (TX_XGXS_INTR | RX_XGXS_INTR)) {
1728 val64 = TXXGXS_INT_M | RXXGXS_INT_M;
1729 if (flag == ENABLE_INTRS) {
1730 temp64 = readq(&bar0->general_int_mask);
1731 temp64 &= ~((u64) val64);
1732 writeq(temp64, &bar0->general_int_mask);
20346722 1733 /*
1da177e4 1734 * All XGXS block error interrupts are disabled for now
20346722 1735 * TODO
1da177e4
LT
1736 */
1737 writeq(DISABLE_ALL_INTRS, &bar0->xgxs_int_mask);
1738 } else if (flag == DISABLE_INTRS) {
20346722 1739 /*
1740 * Disable MC Intrs in the general intr mask register
1da177e4
LT
1741 */
1742 writeq(DISABLE_ALL_INTRS, &bar0->xgxs_int_mask);
1743 temp64 = readq(&bar0->general_int_mask);
1744 val64 |= temp64;
1745 writeq(val64, &bar0->general_int_mask);
1746 }
1747 }
1748
1749 /* Memory Controller(MC) interrupts */
1750 if (mask & MC_INTR) {
1751 val64 = MC_INT_M;
1752 if (flag == ENABLE_INTRS) {
1753 temp64 = readq(&bar0->general_int_mask);
1754 temp64 &= ~((u64) val64);
1755 writeq(temp64, &bar0->general_int_mask);
20346722 1756 /*
5e25b9dd 1757 * Enable all MC Intrs.
1da177e4 1758 */
5e25b9dd 1759 writeq(0x0, &bar0->mc_int_mask);
1760 writeq(0x0, &bar0->mc_err_mask);
1da177e4
LT
1761 } else if (flag == DISABLE_INTRS) {
1762 /*
1763 * Disable MC Intrs in the general intr mask register
1764 */
1765 writeq(DISABLE_ALL_INTRS, &bar0->mc_int_mask);
1766 temp64 = readq(&bar0->general_int_mask);
1767 val64 |= temp64;
1768 writeq(val64, &bar0->general_int_mask);
1769 }
1770 }
1771
1772
1773 /* Tx traffic interrupts */
1774 if (mask & TX_TRAFFIC_INTR) {
1775 val64 = TXTRAFFIC_INT_M;
1776 if (flag == ENABLE_INTRS) {
1777 temp64 = readq(&bar0->general_int_mask);
1778 temp64 &= ~((u64) val64);
1779 writeq(temp64, &bar0->general_int_mask);
20346722 1780 /*
1da177e4 1781 * Enable all the Tx side interrupts
20346722 1782 * writing 0 Enables all 64 TX interrupt levels
1da177e4
LT
1783 */
1784 writeq(0x0, &bar0->tx_traffic_mask);
1785 } else if (flag == DISABLE_INTRS) {
20346722 1786 /*
1787 * Disable Tx Traffic Intrs in the general intr mask
1da177e4
LT
1788 * register.
1789 */
1790 writeq(DISABLE_ALL_INTRS, &bar0->tx_traffic_mask);
1791 temp64 = readq(&bar0->general_int_mask);
1792 val64 |= temp64;
1793 writeq(val64, &bar0->general_int_mask);
1794 }
1795 }
1796
1797 /* Rx traffic interrupts */
1798 if (mask & RX_TRAFFIC_INTR) {
1799 val64 = RXTRAFFIC_INT_M;
1800 if (flag == ENABLE_INTRS) {
1801 temp64 = readq(&bar0->general_int_mask);
1802 temp64 &= ~((u64) val64);
1803 writeq(temp64, &bar0->general_int_mask);
1804 /* writing 0 Enables all 8 RX interrupt levels */
1805 writeq(0x0, &bar0->rx_traffic_mask);
1806 } else if (flag == DISABLE_INTRS) {
20346722 1807 /*
1808 * Disable Rx Traffic Intrs in the general intr mask
1da177e4
LT
1809 * register.
1810 */
1811 writeq(DISABLE_ALL_INTRS, &bar0->rx_traffic_mask);
1812 temp64 = readq(&bar0->general_int_mask);
1813 val64 |= temp64;
1814 writeq(val64, &bar0->general_int_mask);
1815 }
1816 }
1817}
1818
541ae68f 1819static int check_prc_pcc_state(u64 val64, int flag, int rev_id, int herc)
20346722 1820{
1821 int ret = 0;
1822
1823 if (flag == FALSE) {
541ae68f 1824 if ((!herc && (rev_id >= 4)) || herc) {
5e25b9dd 1825 if (!(val64 & ADAPTER_STATUS_RMAC_PCC_IDLE) &&
1826 ((val64 & ADAPTER_STATUS_RC_PRC_QUIESCENT) ==
1827 ADAPTER_STATUS_RC_PRC_QUIESCENT)) {
1828 ret = 1;
1829 }
541ae68f 1830 }else {
5e25b9dd 1831 if (!(val64 & ADAPTER_STATUS_RMAC_PCC_FOUR_IDLE) &&
1832 ((val64 & ADAPTER_STATUS_RC_PRC_QUIESCENT) ==
1833 ADAPTER_STATUS_RC_PRC_QUIESCENT)) {
1834 ret = 1;
1835 }
20346722 1836 }
1837 } else {
541ae68f 1838 if ((!herc && (rev_id >= 4)) || herc) {
5e25b9dd 1839 if (((val64 & ADAPTER_STATUS_RMAC_PCC_IDLE) ==
1840 ADAPTER_STATUS_RMAC_PCC_IDLE) &&
1841 (!(val64 & ADAPTER_STATUS_RC_PRC_QUIESCENT) ||
1842 ((val64 & ADAPTER_STATUS_RC_PRC_QUIESCENT) ==
1843 ADAPTER_STATUS_RC_PRC_QUIESCENT))) {
1844 ret = 1;
1845 }
1846 } else {
1847 if (((val64 & ADAPTER_STATUS_RMAC_PCC_FOUR_IDLE) ==
1848 ADAPTER_STATUS_RMAC_PCC_FOUR_IDLE) &&
1849 (!(val64 & ADAPTER_STATUS_RC_PRC_QUIESCENT) ||
1850 ((val64 & ADAPTER_STATUS_RC_PRC_QUIESCENT) ==
1851 ADAPTER_STATUS_RC_PRC_QUIESCENT))) {
1852 ret = 1;
1853 }
20346722 1854 }
1855 }
1856
1857 return ret;
1858}
1859/**
1860 * verify_xena_quiescence - Checks whether the H/W is ready
1da177e4
LT
1861 * @val64 : Value read from adapter status register.
1862 * @flag : indicates if the adapter enable bit was ever written once
1863 * before.
1864 * Description: Returns whether the H/W is ready to go or not. Depending
20346722 1865 * on whether adapter enable bit was written or not the comparison
1da177e4
LT
1866 * differs and the calling function passes the input argument flag to
1867 * indicate this.
20346722 1868 * Return: 1 If xena is quiescence
1da177e4
LT
1869 * 0 If Xena is not quiescence
1870 */
1871
20346722 1872static int verify_xena_quiescence(nic_t *sp, u64 val64, int flag)
1da177e4 1873{
541ae68f 1874 int ret = 0, herc;
1da177e4 1875 u64 tmp64 = ~((u64) val64);
5e25b9dd 1876 int rev_id = get_xena_rev_id(sp->pdev);
1da177e4 1877
541ae68f 1878 herc = (sp->device_type == XFRAME_II_DEVICE);
1da177e4
LT
1879 if (!
1880 (tmp64 &
1881 (ADAPTER_STATUS_TDMA_READY | ADAPTER_STATUS_RDMA_READY |
1882 ADAPTER_STATUS_PFC_READY | ADAPTER_STATUS_TMAC_BUF_EMPTY |
1883 ADAPTER_STATUS_PIC_QUIESCENT | ADAPTER_STATUS_MC_DRAM_READY |
1884 ADAPTER_STATUS_MC_QUEUES_READY | ADAPTER_STATUS_M_PLL_LOCK |
1885 ADAPTER_STATUS_P_PLL_LOCK))) {
541ae68f 1886 ret = check_prc_pcc_state(val64, flag, rev_id, herc);
1da177e4
LT
1887 }
1888
1889 return ret;
1890}
1891
1892/**
1893 * fix_mac_address - Fix for Mac addr problem on Alpha platforms
1894 * @sp: Pointer to device specifc structure
20346722 1895 * Description :
1da177e4
LT
1896 * New procedure to clear mac address reading problems on Alpha platforms
1897 *
1898 */
1899
ac1f60db 1900static void fix_mac_address(nic_t * sp)
1da177e4
LT
1901{
1902 XENA_dev_config_t __iomem *bar0 = sp->bar0;
1903 u64 val64;
1904 int i = 0;
1905
1906 while (fix_mac[i] != END_SIGN) {
1907 writeq(fix_mac[i++], &bar0->gpio_control);
20346722 1908 udelay(10);
1da177e4
LT
1909 val64 = readq(&bar0->gpio_control);
1910 }
1911}
1912
1913/**
20346722 1914 * start_nic - Turns the device on
1da177e4 1915 * @nic : device private variable.
20346722 1916 * Description:
1917 * This function actually turns the device on. Before this function is
1918 * called,all Registers are configured from their reset states
1919 * and shared memory is allocated but the NIC is still quiescent. On
1da177e4
LT
1920 * calling this function, the device interrupts are cleared and the NIC is
1921 * literally switched on by writing into the adapter control register.
20346722 1922 * Return Value:
1da177e4
LT
1923 * SUCCESS on success and -1 on failure.
1924 */
1925
1926static int start_nic(struct s2io_nic *nic)
1927{
1928 XENA_dev_config_t __iomem *bar0 = nic->bar0;
1929 struct net_device *dev = nic->dev;
1930 register u64 val64 = 0;
20346722 1931 u16 interruptible;
1932 u16 subid, i;
1da177e4
LT
1933 mac_info_t *mac_control;
1934 struct config_param *config;
1935
1936 mac_control = &nic->mac_control;
1937 config = &nic->config;
1938
1939 /* PRC Initialization and configuration */
1940 for (i = 0; i < config->rx_ring_num; i++) {
20346722 1941 writeq((u64) mac_control->rings[i].rx_blocks[0].block_dma_addr,
1da177e4
LT
1942 &bar0->prc_rxd0_n[i]);
1943
1944 val64 = readq(&bar0->prc_ctrl_n[i]);
b6e3f982 1945 if (nic->config.bimodal)
1946 val64 |= PRC_CTRL_BIMODAL_INTERRUPT;
da6971d8
AR
1947 if (nic->rxd_mode == RXD_MODE_1)
1948 val64 |= PRC_CTRL_RC_ENABLED;
1949 else
1950 val64 |= PRC_CTRL_RC_ENABLED | PRC_CTRL_RING_MODE_3;
1da177e4
LT
1951 writeq(val64, &bar0->prc_ctrl_n[i]);
1952 }
1953
da6971d8
AR
1954 if (nic->rxd_mode == RXD_MODE_3B) {
1955 /* Enabling 2 buffer mode by writing into Rx_pa_cfg reg. */
1956 val64 = readq(&bar0->rx_pa_cfg);
1957 val64 |= RX_PA_CFG_IGNORE_L2_ERR;
1958 writeq(val64, &bar0->rx_pa_cfg);
1959 }
1da177e4 1960
20346722 1961 /*
1da177e4
LT
1962 * Enabling MC-RLDRAM. After enabling the device, we timeout
1963 * for around 100ms, which is approximately the time required
1964 * for the device to be ready for operation.
1965 */
1966 val64 = readq(&bar0->mc_rldram_mrs);
1967 val64 |= MC_RLDRAM_QUEUE_SIZE_ENABLE | MC_RLDRAM_MRS_ENABLE;
1968 SPECIAL_REG_WRITE(val64, &bar0->mc_rldram_mrs, UF);
1969 val64 = readq(&bar0->mc_rldram_mrs);
1970
20346722 1971 msleep(100); /* Delay by around 100 ms. */
1da177e4
LT
1972
1973 /* Enabling ECC Protection. */
1974 val64 = readq(&bar0->adapter_control);
1975 val64 &= ~ADAPTER_ECC_EN;
1976 writeq(val64, &bar0->adapter_control);
1977
20346722 1978 /*
1979 * Clearing any possible Link state change interrupts that
1da177e4
LT
1980 * could have popped up just before Enabling the card.
1981 */
1982 val64 = readq(&bar0->mac_rmac_err_reg);
1983 if (val64)
1984 writeq(val64, &bar0->mac_rmac_err_reg);
1985
20346722 1986 /*
1987 * Verify if the device is ready to be enabled, if so enable
1da177e4
LT
1988 * it.
1989 */
1990 val64 = readq(&bar0->adapter_status);
20346722 1991 if (!verify_xena_quiescence(nic, val64, nic->device_enabled_once)) {
1da177e4
LT
1992 DBG_PRINT(ERR_DBG, "%s: device is not ready, ", dev->name);
1993 DBG_PRINT(ERR_DBG, "Adapter status reads: 0x%llx\n",
1994 (unsigned long long) val64);
1995 return FAILURE;
1996 }
1997
1998 /* Enable select interrupts */
cc6e7c44
RA
1999 if (nic->intr_type != INTA)
2000 en_dis_able_nic_intrs(nic, ENA_ALL_INTRS, DISABLE_INTRS);
2001 else {
2002 interruptible = TX_TRAFFIC_INTR | RX_TRAFFIC_INTR;
2003 interruptible |= TX_PIC_INTR | RX_PIC_INTR;
2004 interruptible |= TX_MAC_INTR | RX_MAC_INTR;
2005 en_dis_able_nic_intrs(nic, interruptible, ENABLE_INTRS);
2006 }
1da177e4 2007
20346722 2008 /*
1da177e4 2009 * With some switches, link might be already up at this point.
20346722 2010 * Because of this weird behavior, when we enable laser,
2011 * we may not get link. We need to handle this. We cannot
2012 * figure out which switch is misbehaving. So we are forced to
2013 * make a global change.
1da177e4
LT
2014 */
2015
2016 /* Enabling Laser. */
2017 val64 = readq(&bar0->adapter_control);
2018 val64 |= ADAPTER_EOI_TX_ON;
2019 writeq(val64, &bar0->adapter_control);
2020
2021 /* SXE-002: Initialize link and activity LED */
2022 subid = nic->pdev->subsystem_device;
541ae68f 2023 if (((subid & 0xFF) >= 0x07) &&
2024 (nic->device_type == XFRAME_I_DEVICE)) {
1da177e4
LT
2025 val64 = readq(&bar0->gpio_control);
2026 val64 |= 0x0000800000000000ULL;
2027 writeq(val64, &bar0->gpio_control);
2028 val64 = 0x0411040400000000ULL;
509a2671 2029 writeq(val64, (void __iomem *)bar0 + 0x2700);
1da177e4
LT
2030 }
2031
20346722 2032 /*
2033 * Don't see link state interrupts on certain switches, so
1da177e4
LT
2034 * directly scheduling a link state task from here.
2035 */
2036 schedule_work(&nic->set_link_task);
2037
1da177e4
LT
2038 return SUCCESS;
2039}
fed5eccd
AR
2040/**
2041 * s2io_txdl_getskb - Get the skb from txdl, unmap and return skb
2042 */
2043static struct sk_buff *s2io_txdl_getskb(fifo_info_t *fifo_data, TxD_t *txdlp, int get_off)
2044{
2045 nic_t *nic = fifo_data->nic;
2046 struct sk_buff *skb;
2047 TxD_t *txds;
2048 u16 j, frg_cnt;
2049
2050 txds = txdlp;
26b7625c 2051 if (txds->Host_Control == (u64)(long)nic->ufo_in_band_v) {
fed5eccd
AR
2052 pci_unmap_single(nic->pdev, (dma_addr_t)
2053 txds->Buffer_Pointer, sizeof(u64),
2054 PCI_DMA_TODEVICE);
2055 txds++;
2056 }
2057
2058 skb = (struct sk_buff *) ((unsigned long)
2059 txds->Host_Control);
2060 if (!skb) {
2061 memset(txdlp, 0, (sizeof(TxD_t) * fifo_data->max_txds));
2062 return NULL;
2063 }
2064 pci_unmap_single(nic->pdev, (dma_addr_t)
2065 txds->Buffer_Pointer,
2066 skb->len - skb->data_len,
2067 PCI_DMA_TODEVICE);
2068 frg_cnt = skb_shinfo(skb)->nr_frags;
2069 if (frg_cnt) {
2070 txds++;
2071 for (j = 0; j < frg_cnt; j++, txds++) {
2072 skb_frag_t *frag = &skb_shinfo(skb)->frags[j];
2073 if (!txds->Buffer_Pointer)
2074 break;
2075 pci_unmap_page(nic->pdev, (dma_addr_t)
2076 txds->Buffer_Pointer,
2077 frag->size, PCI_DMA_TODEVICE);
2078 }
2079 }
2080 txdlp->Host_Control = 0;
2081 return(skb);
2082}
1da177e4 2083
20346722 2084/**
2085 * free_tx_buffers - Free all queued Tx buffers
1da177e4 2086 * @nic : device private variable.
20346722 2087 * Description:
1da177e4 2088 * Free all queued Tx buffers.
20346722 2089 * Return Value: void
1da177e4
LT
2090*/
2091
2092static void free_tx_buffers(struct s2io_nic *nic)
2093{
2094 struct net_device *dev = nic->dev;
2095 struct sk_buff *skb;
2096 TxD_t *txdp;
2097 int i, j;
2098 mac_info_t *mac_control;
2099 struct config_param *config;
fed5eccd 2100 int cnt = 0;
1da177e4
LT
2101
2102 mac_control = &nic->mac_control;
2103 config = &nic->config;
2104
2105 for (i = 0; i < config->tx_fifo_num; i++) {
2106 for (j = 0; j < config->tx_cfg[i].fifo_len - 1; j++) {
20346722 2107 txdp = (TxD_t *) mac_control->fifos[i].list_info[j].
1da177e4 2108 list_virt_addr;
fed5eccd
AR
2109 skb = s2io_txdl_getskb(&mac_control->fifos[i], txdp, j);
2110 if (skb) {
2111 dev_kfree_skb(skb);
2112 cnt++;
1da177e4 2113 }
1da177e4
LT
2114 }
2115 DBG_PRINT(INTR_DBG,
2116 "%s:forcibly freeing %d skbs on FIFO%d\n",
2117 dev->name, cnt, i);
20346722 2118 mac_control->fifos[i].tx_curr_get_info.offset = 0;
2119 mac_control->fifos[i].tx_curr_put_info.offset = 0;
1da177e4
LT
2120 }
2121}
2122
20346722 2123/**
2124 * stop_nic - To stop the nic
1da177e4 2125 * @nic ; device private variable.
20346722 2126 * Description:
2127 * This function does exactly the opposite of what the start_nic()
1da177e4
LT
2128 * function does. This function is called to stop the device.
2129 * Return Value:
2130 * void.
2131 */
2132
2133static void stop_nic(struct s2io_nic *nic)
2134{
2135 XENA_dev_config_t __iomem *bar0 = nic->bar0;
2136 register u64 val64 = 0;
2137 u16 interruptible, i;
2138 mac_info_t *mac_control;
2139 struct config_param *config;
2140
2141 mac_control = &nic->mac_control;
2142 config = &nic->config;
2143
2144 /* Disable all interrupts */
e960fc5c 2145 interruptible = TX_TRAFFIC_INTR | RX_TRAFFIC_INTR;
a371a07d 2146 interruptible |= TX_PIC_INTR | RX_PIC_INTR;
2147 interruptible |= TX_MAC_INTR | RX_MAC_INTR;
1da177e4
LT
2148 en_dis_able_nic_intrs(nic, interruptible, DISABLE_INTRS);
2149
2150 /* Disable PRCs */
2151 for (i = 0; i < config->rx_ring_num; i++) {
2152 val64 = readq(&bar0->prc_ctrl_n[i]);
2153 val64 &= ~((u64) PRC_CTRL_RC_ENABLED);
2154 writeq(val64, &bar0->prc_ctrl_n[i]);
2155 }
2156}
2157
26df54bf 2158static int fill_rxd_3buf(nic_t *nic, RxD_t *rxdp, struct sk_buff *skb)
da6971d8
AR
2159{
2160 struct net_device *dev = nic->dev;
2161 struct sk_buff *frag_list;
50eb8006 2162 void *tmp;
da6971d8
AR
2163
2164 /* Buffer-1 receives L3/L4 headers */
2165 ((RxD3_t*)rxdp)->Buffer1_ptr = pci_map_single
2166 (nic->pdev, skb->data, l3l4hdr_size + 4,
2167 PCI_DMA_FROMDEVICE);
2168
2169 /* skb_shinfo(skb)->frag_list will have L4 data payload */
2170 skb_shinfo(skb)->frag_list = dev_alloc_skb(dev->mtu + ALIGN_SIZE);
2171 if (skb_shinfo(skb)->frag_list == NULL) {
2172 DBG_PRINT(ERR_DBG, "%s: dev_alloc_skb failed\n ", dev->name);
2173 return -ENOMEM ;
2174 }
2175 frag_list = skb_shinfo(skb)->frag_list;
2176 frag_list->next = NULL;
50eb8006
JG
2177 tmp = (void *)ALIGN((long)frag_list->data, ALIGN_SIZE + 1);
2178 frag_list->data = tmp;
2179 frag_list->tail = tmp;
da6971d8
AR
2180
2181 /* Buffer-2 receives L4 data payload */
2182 ((RxD3_t*)rxdp)->Buffer2_ptr = pci_map_single(nic->pdev,
2183 frag_list->data, dev->mtu,
2184 PCI_DMA_FROMDEVICE);
2185 rxdp->Control_2 |= SET_BUFFER1_SIZE_3(l3l4hdr_size + 4);
2186 rxdp->Control_2 |= SET_BUFFER2_SIZE_3(dev->mtu);
2187
2188 return SUCCESS;
2189}
2190
20346722 2191/**
2192 * fill_rx_buffers - Allocates the Rx side skbs
1da177e4 2193 * @nic: device private variable
20346722 2194 * @ring_no: ring number
2195 * Description:
1da177e4
LT
2196 * The function allocates Rx side skbs and puts the physical
2197 * address of these buffers into the RxD buffer pointers, so that the NIC
2198 * can DMA the received frame into these locations.
2199 * The NIC supports 3 receive modes, viz
2200 * 1. single buffer,
2201 * 2. three buffer and
2202 * 3. Five buffer modes.
20346722 2203 * Each mode defines how many fragments the received frame will be split
2204 * up into by the NIC. The frame is split into L3 header, L4 Header,
1da177e4
LT
2205 * L4 payload in three buffer mode and in 5 buffer mode, L4 payload itself
2206 * is split into 3 fragments. As of now only single buffer mode is
2207 * supported.
2208 * Return Value:
2209 * SUCCESS on success or an appropriate -ve value on failure.
2210 */
2211
ac1f60db 2212static int fill_rx_buffers(struct s2io_nic *nic, int ring_no)
1da177e4
LT
2213{
2214 struct net_device *dev = nic->dev;
2215 struct sk_buff *skb;
2216 RxD_t *rxdp;
2217 int off, off1, size, block_no, block_no1;
1da177e4 2218 u32 alloc_tab = 0;
20346722 2219 u32 alloc_cnt;
1da177e4
LT
2220 mac_info_t *mac_control;
2221 struct config_param *config;
20346722 2222 u64 tmp;
1da177e4 2223 buffAdd_t *ba;
1da177e4
LT
2224#ifndef CONFIG_S2IO_NAPI
2225 unsigned long flags;
2226#endif
303bcb4b 2227 RxD_t *first_rxdp = NULL;
1da177e4
LT
2228
2229 mac_control = &nic->mac_control;
2230 config = &nic->config;
20346722 2231 alloc_cnt = mac_control->rings[ring_no].pkt_cnt -
2232 atomic_read(&nic->rx_bufs_left[ring_no]);
1da177e4
LT
2233
2234 while (alloc_tab < alloc_cnt) {
20346722 2235 block_no = mac_control->rings[ring_no].rx_curr_put_info.
1da177e4 2236 block_index;
20346722 2237 block_no1 = mac_control->rings[ring_no].rx_curr_get_info.
1da177e4 2238 block_index;
20346722 2239 off = mac_control->rings[ring_no].rx_curr_put_info.offset;
2240 off1 = mac_control->rings[ring_no].rx_curr_get_info.offset;
1da177e4 2241
da6971d8
AR
2242 rxdp = mac_control->rings[ring_no].
2243 rx_blocks[block_no].rxds[off].virt_addr;
2244
2245 if ((block_no == block_no1) && (off == off1) &&
2246 (rxdp->Host_Control)) {
2247 DBG_PRINT(INTR_DBG, "%s: Get and Put",
2248 dev->name);
1da177e4
LT
2249 DBG_PRINT(INTR_DBG, " info equated\n");
2250 goto end;
2251 }
da6971d8 2252 if (off && (off == rxd_count[nic->rxd_mode])) {
20346722 2253 mac_control->rings[ring_no].rx_curr_put_info.
1da177e4 2254 block_index++;
da6971d8
AR
2255 if (mac_control->rings[ring_no].rx_curr_put_info.
2256 block_index == mac_control->rings[ring_no].
2257 block_count)
2258 mac_control->rings[ring_no].rx_curr_put_info.
2259 block_index = 0;
2260 block_no = mac_control->rings[ring_no].
2261 rx_curr_put_info.block_index;
2262 if (off == rxd_count[nic->rxd_mode])
2263 off = 0;
20346722 2264 mac_control->rings[ring_no].rx_curr_put_info.
da6971d8
AR
2265 offset = off;
2266 rxdp = mac_control->rings[ring_no].
2267 rx_blocks[block_no].block_virt_addr;
1da177e4
LT
2268 DBG_PRINT(INTR_DBG, "%s: Next block at: %p\n",
2269 dev->name, rxdp);
2270 }
2271#ifndef CONFIG_S2IO_NAPI
2272 spin_lock_irqsave(&nic->put_lock, flags);
20346722 2273 mac_control->rings[ring_no].put_pos =
da6971d8 2274 (block_no * (rxd_count[nic->rxd_mode] + 1)) + off;
1da177e4
LT
2275 spin_unlock_irqrestore(&nic->put_lock, flags);
2276#endif
da6971d8
AR
2277 if ((rxdp->Control_1 & RXD_OWN_XENA) &&
2278 ((nic->rxd_mode >= RXD_MODE_3A) &&
2279 (rxdp->Control_2 & BIT(0)))) {
20346722 2280 mac_control->rings[ring_no].rx_curr_put_info.
da6971d8 2281 offset = off;
1da177e4
LT
2282 goto end;
2283 }
da6971d8
AR
2284 /* calculate size of skb based on ring mode */
2285 size = dev->mtu + HEADER_ETHERNET_II_802_3_SIZE +
2286 HEADER_802_2_SIZE + HEADER_SNAP_SIZE;
2287 if (nic->rxd_mode == RXD_MODE_1)
2288 size += NET_IP_ALIGN;
2289 else if (nic->rxd_mode == RXD_MODE_3B)
2290 size = dev->mtu + ALIGN_SIZE + BUF0_LEN + 4;
2291 else
2292 size = l3l4hdr_size + ALIGN_SIZE + BUF0_LEN + 4;
1da177e4 2293
da6971d8
AR
2294 /* allocate skb */
2295 skb = dev_alloc_skb(size);
2296 if(!skb) {
1da177e4
LT
2297 DBG_PRINT(ERR_DBG, "%s: Out of ", dev->name);
2298 DBG_PRINT(ERR_DBG, "memory to allocate SKBs\n");
303bcb4b 2299 if (first_rxdp) {
2300 wmb();
2301 first_rxdp->Control_1 |= RXD_OWN_XENA;
2302 }
da6971d8
AR
2303 return -ENOMEM ;
2304 }
2305 if (nic->rxd_mode == RXD_MODE_1) {
2306 /* 1 buffer mode - normal operation mode */
2307 memset(rxdp, 0, sizeof(RxD1_t));
2308 skb_reserve(skb, NET_IP_ALIGN);
2309 ((RxD1_t*)rxdp)->Buffer0_ptr = pci_map_single
2310 (nic->pdev, skb->data, size, PCI_DMA_FROMDEVICE);
2311 rxdp->Control_2 &= (~MASK_BUFFER0_SIZE_1);
2312 rxdp->Control_2 |= SET_BUFFER0_SIZE_1(size);
2313
2314 } else if (nic->rxd_mode >= RXD_MODE_3A) {
2315 /*
2316 * 2 or 3 buffer mode -
2317 * Both 2 buffer mode and 3 buffer mode provides 128
2318 * byte aligned receive buffers.
2319 *
2320 * 3 buffer mode provides header separation where in
2321 * skb->data will have L3/L4 headers where as
2322 * skb_shinfo(skb)->frag_list will have the L4 data
2323 * payload
2324 */
2325
2326 memset(rxdp, 0, sizeof(RxD3_t));
2327 ba = &mac_control->rings[ring_no].ba[block_no][off];
2328 skb_reserve(skb, BUF0_LEN);
2329 tmp = (u64)(unsigned long) skb->data;
2330 tmp += ALIGN_SIZE;
2331 tmp &= ~ALIGN_SIZE;
2332 skb->data = (void *) (unsigned long)tmp;
2333 skb->tail = (void *) (unsigned long)tmp;
2334
2335 ((RxD3_t*)rxdp)->Buffer0_ptr =
2336 pci_map_single(nic->pdev, ba->ba_0, BUF0_LEN,
2337 PCI_DMA_FROMDEVICE);
2338 rxdp->Control_2 = SET_BUFFER0_SIZE_3(BUF0_LEN);
2339 if (nic->rxd_mode == RXD_MODE_3B) {
2340 /* Two buffer mode */
2341
2342 /*
2343 * Buffer2 will have L3/L4 header plus
2344 * L4 payload
2345 */
2346 ((RxD3_t*)rxdp)->Buffer2_ptr = pci_map_single
2347 (nic->pdev, skb->data, dev->mtu + 4,
2348 PCI_DMA_FROMDEVICE);
2349
2350 /* Buffer-1 will be dummy buffer not used */
2351 ((RxD3_t*)rxdp)->Buffer1_ptr =
2352 pci_map_single(nic->pdev, ba->ba_1, BUF1_LEN,
2353 PCI_DMA_FROMDEVICE);
2354 rxdp->Control_2 |= SET_BUFFER1_SIZE_3(1);
2355 rxdp->Control_2 |= SET_BUFFER2_SIZE_3
2356 (dev->mtu + 4);
2357 } else {
2358 /* 3 buffer mode */
2359 if (fill_rxd_3buf(nic, rxdp, skb) == -ENOMEM) {
2360 dev_kfree_skb_irq(skb);
2361 if (first_rxdp) {
2362 wmb();
2363 first_rxdp->Control_1 |=
2364 RXD_OWN_XENA;
2365 }
2366 return -ENOMEM ;
2367 }
2368 }
2369 rxdp->Control_2 |= BIT(0);
1da177e4 2370 }
1da177e4 2371 rxdp->Host_Control = (unsigned long) (skb);
303bcb4b 2372 if (alloc_tab & ((1 << rxsync_frequency) - 1))
2373 rxdp->Control_1 |= RXD_OWN_XENA;
1da177e4 2374 off++;
da6971d8
AR
2375 if (off == (rxd_count[nic->rxd_mode] + 1))
2376 off = 0;
20346722 2377 mac_control->rings[ring_no].rx_curr_put_info.offset = off;
20346722 2378
da6971d8 2379 rxdp->Control_2 |= SET_RXD_MARKER;
303bcb4b 2380 if (!(alloc_tab & ((1 << rxsync_frequency) - 1))) {
2381 if (first_rxdp) {
2382 wmb();
2383 first_rxdp->Control_1 |= RXD_OWN_XENA;
2384 }
2385 first_rxdp = rxdp;
2386 }
1da177e4
LT
2387 atomic_inc(&nic->rx_bufs_left[ring_no]);
2388 alloc_tab++;
2389 }
2390
2391 end:
303bcb4b 2392 /* Transfer ownership of first descriptor to adapter just before
2393 * exiting. Before that, use memory barrier so that ownership
2394 * and other fields are seen by adapter correctly.
2395 */
2396 if (first_rxdp) {
2397 wmb();
2398 first_rxdp->Control_1 |= RXD_OWN_XENA;
2399 }
2400
1da177e4
LT
2401 return SUCCESS;
2402}
2403
da6971d8
AR
2404static void free_rxd_blk(struct s2io_nic *sp, int ring_no, int blk)
2405{
2406 struct net_device *dev = sp->dev;
2407 int j;
2408 struct sk_buff *skb;
2409 RxD_t *rxdp;
2410 mac_info_t *mac_control;
2411 buffAdd_t *ba;
2412
2413 mac_control = &sp->mac_control;
2414 for (j = 0 ; j < rxd_count[sp->rxd_mode]; j++) {
2415 rxdp = mac_control->rings[ring_no].
2416 rx_blocks[blk].rxds[j].virt_addr;
2417 skb = (struct sk_buff *)
2418 ((unsigned long) rxdp->Host_Control);
2419 if (!skb) {
2420 continue;
2421 }
2422 if (sp->rxd_mode == RXD_MODE_1) {
2423 pci_unmap_single(sp->pdev, (dma_addr_t)
2424 ((RxD1_t*)rxdp)->Buffer0_ptr,
2425 dev->mtu +
2426 HEADER_ETHERNET_II_802_3_SIZE
2427 + HEADER_802_2_SIZE +
2428 HEADER_SNAP_SIZE,
2429 PCI_DMA_FROMDEVICE);
2430 memset(rxdp, 0, sizeof(RxD1_t));
2431 } else if(sp->rxd_mode == RXD_MODE_3B) {
2432 ba = &mac_control->rings[ring_no].
2433 ba[blk][j];
2434 pci_unmap_single(sp->pdev, (dma_addr_t)
2435 ((RxD3_t*)rxdp)->Buffer0_ptr,
2436 BUF0_LEN,
2437 PCI_DMA_FROMDEVICE);
2438 pci_unmap_single(sp->pdev, (dma_addr_t)
2439 ((RxD3_t*)rxdp)->Buffer1_ptr,
2440 BUF1_LEN,
2441 PCI_DMA_FROMDEVICE);
2442 pci_unmap_single(sp->pdev, (dma_addr_t)
2443 ((RxD3_t*)rxdp)->Buffer2_ptr,
2444 dev->mtu + 4,
2445 PCI_DMA_FROMDEVICE);
2446 memset(rxdp, 0, sizeof(RxD3_t));
2447 } else {
2448 pci_unmap_single(sp->pdev, (dma_addr_t)
2449 ((RxD3_t*)rxdp)->Buffer0_ptr, BUF0_LEN,
2450 PCI_DMA_FROMDEVICE);
2451 pci_unmap_single(sp->pdev, (dma_addr_t)
2452 ((RxD3_t*)rxdp)->Buffer1_ptr,
2453 l3l4hdr_size + 4,
2454 PCI_DMA_FROMDEVICE);
2455 pci_unmap_single(sp->pdev, (dma_addr_t)
2456 ((RxD3_t*)rxdp)->Buffer2_ptr, dev->mtu,
2457 PCI_DMA_FROMDEVICE);
2458 memset(rxdp, 0, sizeof(RxD3_t));
2459 }
2460 dev_kfree_skb(skb);
2461 atomic_dec(&sp->rx_bufs_left[ring_no]);
2462 }
2463}
2464
1da177e4 2465/**
20346722 2466 * free_rx_buffers - Frees all Rx buffers
1da177e4 2467 * @sp: device private variable.
20346722 2468 * Description:
1da177e4
LT
2469 * This function will free all Rx buffers allocated by host.
2470 * Return Value:
2471 * NONE.
2472 */
2473
2474static void free_rx_buffers(struct s2io_nic *sp)
2475{
2476 struct net_device *dev = sp->dev;
da6971d8 2477 int i, blk = 0, buf_cnt = 0;
1da177e4
LT
2478 mac_info_t *mac_control;
2479 struct config_param *config;
1da177e4
LT
2480
2481 mac_control = &sp->mac_control;
2482 config = &sp->config;
2483
2484 for (i = 0; i < config->rx_ring_num; i++) {
da6971d8
AR
2485 for (blk = 0; blk < rx_ring_sz[i]; blk++)
2486 free_rxd_blk(sp,i,blk);
1da177e4 2487
20346722 2488 mac_control->rings[i].rx_curr_put_info.block_index = 0;
2489 mac_control->rings[i].rx_curr_get_info.block_index = 0;
2490 mac_control->rings[i].rx_curr_put_info.offset = 0;
2491 mac_control->rings[i].rx_curr_get_info.offset = 0;
1da177e4
LT
2492 atomic_set(&sp->rx_bufs_left[i], 0);
2493 DBG_PRINT(INIT_DBG, "%s:Freed 0x%x Rx Buffers on ring%d\n",
2494 dev->name, buf_cnt, i);
2495 }
2496}
2497
2498/**
2499 * s2io_poll - Rx interrupt handler for NAPI support
2500 * @dev : pointer to the device structure.
20346722 2501 * @budget : The number of packets that were budgeted to be processed
1da177e4
LT
2502 * during one pass through the 'Poll" function.
2503 * Description:
2504 * Comes into picture only if NAPI support has been incorporated. It does
2505 * the same thing that rx_intr_handler does, but not in a interrupt context
2506 * also It will process only a given number of packets.
2507 * Return value:
2508 * 0 on success and 1 if there are No Rx packets to be processed.
2509 */
2510
20346722 2511#if defined(CONFIG_S2IO_NAPI)
1da177e4
LT
2512static int s2io_poll(struct net_device *dev, int *budget)
2513{
2514 nic_t *nic = dev->priv;
20346722 2515 int pkt_cnt = 0, org_pkts_to_process;
1da177e4
LT
2516 mac_info_t *mac_control;
2517 struct config_param *config;
509a2671 2518 XENA_dev_config_t __iomem *bar0 = nic->bar0;
20346722 2519 u64 val64;
2520 int i;
1da177e4 2521
7ba013ac 2522 atomic_inc(&nic->isr_cnt);
1da177e4
LT
2523 mac_control = &nic->mac_control;
2524 config = &nic->config;
2525
20346722 2526 nic->pkts_to_process = *budget;
2527 if (nic->pkts_to_process > dev->quota)
2528 nic->pkts_to_process = dev->quota;
2529 org_pkts_to_process = nic->pkts_to_process;
1da177e4
LT
2530
2531 val64 = readq(&bar0->rx_traffic_int);
2532 writeq(val64, &bar0->rx_traffic_int);
2533
2534 for (i = 0; i < config->rx_ring_num; i++) {
20346722 2535 rx_intr_handler(&mac_control->rings[i]);
2536 pkt_cnt = org_pkts_to_process - nic->pkts_to_process;
2537 if (!nic->pkts_to_process) {
2538 /* Quota for the current iteration has been met */
2539 goto no_rx;
1da177e4 2540 }
1da177e4
LT
2541 }
2542 if (!pkt_cnt)
2543 pkt_cnt = 1;
2544
2545 dev->quota -= pkt_cnt;
2546 *budget -= pkt_cnt;
2547 netif_rx_complete(dev);
2548
2549 for (i = 0; i < config->rx_ring_num; i++) {
2550 if (fill_rx_buffers(nic, i) == -ENOMEM) {
2551 DBG_PRINT(ERR_DBG, "%s:Out of memory", dev->name);
2552 DBG_PRINT(ERR_DBG, " in Rx Poll!!\n");
2553 break;
2554 }
2555 }
2556 /* Re enable the Rx interrupts. */
2557 en_dis_able_nic_intrs(nic, RX_TRAFFIC_INTR, ENABLE_INTRS);
7ba013ac 2558 atomic_dec(&nic->isr_cnt);
1da177e4
LT
2559 return 0;
2560
20346722 2561no_rx:
1da177e4
LT
2562 dev->quota -= pkt_cnt;
2563 *budget -= pkt_cnt;
2564
2565 for (i = 0; i < config->rx_ring_num; i++) {
2566 if (fill_rx_buffers(nic, i) == -ENOMEM) {
2567 DBG_PRINT(ERR_DBG, "%s:Out of memory", dev->name);
2568 DBG_PRINT(ERR_DBG, " in Rx Poll!!\n");
2569 break;
2570 }
2571 }
7ba013ac 2572 atomic_dec(&nic->isr_cnt);
1da177e4
LT
2573 return 1;
2574}
20346722 2575#endif
2576
2577/**
1da177e4
LT
2578 * rx_intr_handler - Rx interrupt handler
2579 * @nic: device private variable.
20346722 2580 * Description:
2581 * If the interrupt is because of a received frame or if the
1da177e4 2582 * receive ring contains fresh as yet un-processed frames,this function is
20346722 2583 * called. It picks out the RxD at which place the last Rx processing had
2584 * stopped and sends the skb to the OSM's Rx handler and then increments
1da177e4
LT
2585 * the offset.
2586 * Return Value:
2587 * NONE.
2588 */
20346722 2589static void rx_intr_handler(ring_info_t *ring_data)
1da177e4 2590{
20346722 2591 nic_t *nic = ring_data->nic;
1da177e4 2592 struct net_device *dev = (struct net_device *) nic->dev;
da6971d8 2593 int get_block, put_block, put_offset;
1da177e4
LT
2594 rx_curr_get_info_t get_info, put_info;
2595 RxD_t *rxdp;
2596 struct sk_buff *skb;
20346722 2597#ifndef CONFIG_S2IO_NAPI
2598 int pkt_cnt = 0;
1da177e4 2599#endif
7d3d0439
RA
2600 int i;
2601
7ba013ac 2602 spin_lock(&nic->rx_lock);
2603 if (atomic_read(&nic->card_state) == CARD_DOWN) {
776bd20f 2604 DBG_PRINT(INTR_DBG, "%s: %s going down for reset\n",
7ba013ac 2605 __FUNCTION__, dev->name);
2606 spin_unlock(&nic->rx_lock);
776bd20f 2607 return;
7ba013ac 2608 }
2609
20346722 2610 get_info = ring_data->rx_curr_get_info;
2611 get_block = get_info.block_index;
2612 put_info = ring_data->rx_curr_put_info;
2613 put_block = put_info.block_index;
da6971d8 2614 rxdp = ring_data->rx_blocks[get_block].rxds[get_info.offset].virt_addr;
20346722 2615#ifndef CONFIG_S2IO_NAPI
2616 spin_lock(&nic->put_lock);
2617 put_offset = ring_data->put_pos;
2618 spin_unlock(&nic->put_lock);
2619#else
da6971d8 2620 put_offset = (put_block * (rxd_count[nic->rxd_mode] + 1)) +
20346722 2621 put_info.offset;
2622#endif
da6971d8
AR
2623 while (RXD_IS_UP2DT(rxdp)) {
2624 /* If your are next to put index then it's FIFO full condition */
2625 if ((get_block == put_block) &&
2626 (get_info.offset + 1) == put_info.offset) {
2627 DBG_PRINT(ERR_DBG, "%s: Ring Full\n",dev->name);
2628 break;
2629 }
20346722 2630 skb = (struct sk_buff *) ((unsigned long)rxdp->Host_Control);
2631 if (skb == NULL) {
2632 DBG_PRINT(ERR_DBG, "%s: The skb is ",
2633 dev->name);
2634 DBG_PRINT(ERR_DBG, "Null in Rx Intr\n");
7ba013ac 2635 spin_unlock(&nic->rx_lock);
20346722 2636 return;
1da177e4 2637 }
da6971d8
AR
2638 if (nic->rxd_mode == RXD_MODE_1) {
2639 pci_unmap_single(nic->pdev, (dma_addr_t)
2640 ((RxD1_t*)rxdp)->Buffer0_ptr,
20346722 2641 dev->mtu +
2642 HEADER_ETHERNET_II_802_3_SIZE +
2643 HEADER_802_2_SIZE +
2644 HEADER_SNAP_SIZE,
2645 PCI_DMA_FROMDEVICE);
da6971d8
AR
2646 } else if (nic->rxd_mode == RXD_MODE_3B) {
2647 pci_unmap_single(nic->pdev, (dma_addr_t)
2648 ((RxD3_t*)rxdp)->Buffer0_ptr,
20346722 2649 BUF0_LEN, PCI_DMA_FROMDEVICE);
da6971d8
AR
2650 pci_unmap_single(nic->pdev, (dma_addr_t)
2651 ((RxD3_t*)rxdp)->Buffer1_ptr,
20346722 2652 BUF1_LEN, PCI_DMA_FROMDEVICE);
da6971d8
AR
2653 pci_unmap_single(nic->pdev, (dma_addr_t)
2654 ((RxD3_t*)rxdp)->Buffer2_ptr,
2655 dev->mtu + 4,
20346722 2656 PCI_DMA_FROMDEVICE);
da6971d8
AR
2657 } else {
2658 pci_unmap_single(nic->pdev, (dma_addr_t)
2659 ((RxD3_t*)rxdp)->Buffer0_ptr, BUF0_LEN,
2660 PCI_DMA_FROMDEVICE);
2661 pci_unmap_single(nic->pdev, (dma_addr_t)
2662 ((RxD3_t*)rxdp)->Buffer1_ptr,
2663 l3l4hdr_size + 4,
2664 PCI_DMA_FROMDEVICE);
2665 pci_unmap_single(nic->pdev, (dma_addr_t)
2666 ((RxD3_t*)rxdp)->Buffer2_ptr,
2667 dev->mtu, PCI_DMA_FROMDEVICE);
2668 }
20346722 2669 rx_osm_handler(ring_data, rxdp);
2670 get_info.offset++;
da6971d8
AR
2671 ring_data->rx_curr_get_info.offset = get_info.offset;
2672 rxdp = ring_data->rx_blocks[get_block].
2673 rxds[get_info.offset].virt_addr;
2674 if (get_info.offset == rxd_count[nic->rxd_mode]) {
20346722 2675 get_info.offset = 0;
da6971d8 2676 ring_data->rx_curr_get_info.offset = get_info.offset;
20346722 2677 get_block++;
da6971d8
AR
2678 if (get_block == ring_data->block_count)
2679 get_block = 0;
2680 ring_data->rx_curr_get_info.block_index = get_block;
20346722 2681 rxdp = ring_data->rx_blocks[get_block].block_virt_addr;
2682 }
1da177e4 2683
20346722 2684#ifdef CONFIG_S2IO_NAPI
2685 nic->pkts_to_process -= 1;
2686 if (!nic->pkts_to_process)
2687 break;
2688#else
2689 pkt_cnt++;
1da177e4
LT
2690 if ((indicate_max_pkts) && (pkt_cnt > indicate_max_pkts))
2691 break;
20346722 2692#endif
1da177e4 2693 }
7d3d0439
RA
2694 if (nic->lro) {
2695 /* Clear all LRO sessions before exiting */
2696 for (i=0; i<MAX_LRO_SESSIONS; i++) {
2697 lro_t *lro = &nic->lro0_n[i];
2698 if (lro->in_use) {
2699 update_L3L4_header(nic, lro);
2700 queue_rx_frame(lro->parent);
2701 clear_lro_session(lro);
2702 }
2703 }
2704 }
2705
7ba013ac 2706 spin_unlock(&nic->rx_lock);
1da177e4 2707}
20346722 2708
2709/**
1da177e4
LT
2710 * tx_intr_handler - Transmit interrupt handler
2711 * @nic : device private variable
20346722 2712 * Description:
2713 * If an interrupt was raised to indicate DMA complete of the
2714 * Tx packet, this function is called. It identifies the last TxD
2715 * whose buffer was freed and frees all skbs whose data have already
1da177e4
LT
2716 * DMA'ed into the NICs internal memory.
2717 * Return Value:
2718 * NONE
2719 */
2720
20346722 2721static void tx_intr_handler(fifo_info_t *fifo_data)
1da177e4 2722{
20346722 2723 nic_t *nic = fifo_data->nic;
1da177e4
LT
2724 struct net_device *dev = (struct net_device *) nic->dev;
2725 tx_curr_get_info_t get_info, put_info;
2726 struct sk_buff *skb;
2727 TxD_t *txdlp;
1da177e4 2728
20346722 2729 get_info = fifo_data->tx_curr_get_info;
2730 put_info = fifo_data->tx_curr_put_info;
2731 txdlp = (TxD_t *) fifo_data->list_info[get_info.offset].
2732 list_virt_addr;
2733 while ((!(txdlp->Control_1 & TXD_LIST_OWN_XENA)) &&
2734 (get_info.offset != put_info.offset) &&
2735 (txdlp->Host_Control)) {
2736 /* Check for TxD errors */
2737 if (txdlp->Control_1 & TXD_T_CODE) {
2738 unsigned long long err;
2739 err = txdlp->Control_1 & TXD_T_CODE;
776bd20f 2740 if ((err >> 48) == 0xA) {
2741 DBG_PRINT(TX_DBG, "TxD returned due \
cc6e7c44 2742to loss of link\n");
776bd20f 2743 }
2744 else {
2745 DBG_PRINT(ERR_DBG, "***TxD error \
cc6e7c44 2746%llx\n", err);
776bd20f 2747 }
20346722 2748 }
1da177e4 2749
fed5eccd 2750 skb = s2io_txdl_getskb(fifo_data, txdlp, get_info.offset);
20346722 2751 if (skb == NULL) {
2752 DBG_PRINT(ERR_DBG, "%s: Null skb ",
2753 __FUNCTION__);
2754 DBG_PRINT(ERR_DBG, "in Tx Free Intr\n");
2755 return;
2756 }
2757
20346722 2758 /* Updating the statistics block */
20346722 2759 nic->stats.tx_bytes += skb->len;
2760 dev_kfree_skb_irq(skb);
2761
2762 get_info.offset++;
2763 get_info.offset %= get_info.fifo_len + 1;
2764 txdlp = (TxD_t *) fifo_data->list_info
2765 [get_info.offset].list_virt_addr;
2766 fifo_data->tx_curr_get_info.offset =
2767 get_info.offset;
1da177e4
LT
2768 }
2769
2770 spin_lock(&nic->tx_lock);
2771 if (netif_queue_stopped(dev))
2772 netif_wake_queue(dev);
2773 spin_unlock(&nic->tx_lock);
2774}
2775
20346722 2776/**
1da177e4
LT
2777 * alarm_intr_handler - Alarm Interrrupt handler
2778 * @nic: device private variable
20346722 2779 * Description: If the interrupt was neither because of Rx packet or Tx
1da177e4 2780 * complete, this function is called. If the interrupt was to indicate
20346722 2781 * a loss of link, the OSM link status handler is invoked for any other
2782 * alarm interrupt the block that raised the interrupt is displayed
1da177e4
LT
2783 * and a H/W reset is issued.
2784 * Return Value:
2785 * NONE
2786*/
2787
2788static void alarm_intr_handler(struct s2io_nic *nic)
2789{
2790 struct net_device *dev = (struct net_device *) nic->dev;
2791 XENA_dev_config_t __iomem *bar0 = nic->bar0;
2792 register u64 val64 = 0, err_reg = 0;
2793
2794 /* Handling link status change error Intr */
a371a07d 2795 if (s2io_link_fault_indication(nic) == MAC_RMAC_ERR_TIMER) {
2796 err_reg = readq(&bar0->mac_rmac_err_reg);
2797 writeq(err_reg, &bar0->mac_rmac_err_reg);
2798 if (err_reg & RMAC_LINK_STATE_CHANGE_INT) {
2799 schedule_work(&nic->set_link_task);
2800 }
1da177e4
LT
2801 }
2802
5e25b9dd 2803 /* Handling Ecc errors */
2804 val64 = readq(&bar0->mc_err_reg);
2805 writeq(val64, &bar0->mc_err_reg);
2806 if (val64 & (MC_ERR_REG_ECC_ALL_SNG | MC_ERR_REG_ECC_ALL_DBL)) {
2807 if (val64 & MC_ERR_REG_ECC_ALL_DBL) {
7ba013ac 2808 nic->mac_control.stats_info->sw_stat.
2809 double_ecc_errs++;
776bd20f 2810 DBG_PRINT(INIT_DBG, "%s: Device indicates ",
5e25b9dd 2811 dev->name);
776bd20f 2812 DBG_PRINT(INIT_DBG, "double ECC error!!\n");
e960fc5c 2813 if (nic->device_type != XFRAME_II_DEVICE) {
776bd20f 2814 /* Reset XframeI only if critical error */
2815 if (val64 & (MC_ERR_REG_MIRI_ECC_DB_ERR_0 |
2816 MC_ERR_REG_MIRI_ECC_DB_ERR_1)) {
2817 netif_stop_queue(dev);
2818 schedule_work(&nic->rst_timer_task);
2819 }
e960fc5c 2820 }
5e25b9dd 2821 } else {
7ba013ac 2822 nic->mac_control.stats_info->sw_stat.
2823 single_ecc_errs++;
5e25b9dd 2824 }
2825 }
2826
1da177e4
LT
2827 /* In case of a serious error, the device will be Reset. */
2828 val64 = readq(&bar0->serr_source);
2829 if (val64 & SERR_SOURCE_ANY) {
2830 DBG_PRINT(ERR_DBG, "%s: Device indicates ", dev->name);
776bd20f 2831 DBG_PRINT(ERR_DBG, "serious error %llx!!\n",
2832 (unsigned long long)val64);
1da177e4
LT
2833 netif_stop_queue(dev);
2834 schedule_work(&nic->rst_timer_task);
2835 }
2836
2837 /*
2838 * Also as mentioned in the latest Errata sheets if the PCC_FB_ECC
2839 * Error occurs, the adapter will be recycled by disabling the
20346722 2840 * adapter enable bit and enabling it again after the device
1da177e4
LT
2841 * becomes Quiescent.
2842 */
2843 val64 = readq(&bar0->pcc_err_reg);
2844 writeq(val64, &bar0->pcc_err_reg);
2845 if (val64 & PCC_FB_ECC_DB_ERR) {
2846 u64 ac = readq(&bar0->adapter_control);
2847 ac &= ~(ADAPTER_CNTL_EN);
2848 writeq(ac, &bar0->adapter_control);
2849 ac = readq(&bar0->adapter_control);
2850 schedule_work(&nic->set_link_task);
2851 }
2852
2853 /* Other type of interrupts are not being handled now, TODO */
2854}
2855
20346722 2856/**
1da177e4 2857 * wait_for_cmd_complete - waits for a command to complete.
20346722 2858 * @sp : private member of the device structure, which is a pointer to the
1da177e4 2859 * s2io_nic structure.
20346722 2860 * Description: Function that waits for a command to Write into RMAC
2861 * ADDR DATA registers to be completed and returns either success or
2862 * error depending on whether the command was complete or not.
1da177e4
LT
2863 * Return value:
2864 * SUCCESS on success and FAILURE on failure.
2865 */
2866
ac1f60db 2867static int wait_for_cmd_complete(nic_t * sp)
1da177e4
LT
2868{
2869 XENA_dev_config_t __iomem *bar0 = sp->bar0;
2870 int ret = FAILURE, cnt = 0;
2871 u64 val64;
2872
2873 while (TRUE) {
2874 val64 = readq(&bar0->rmac_addr_cmd_mem);
2875 if (!(val64 & RMAC_ADDR_CMD_MEM_STROBE_CMD_EXECUTING)) {
2876 ret = SUCCESS;
2877 break;
2878 }
2879 msleep(50);
2880 if (cnt++ > 10)
2881 break;
2882 }
2883
2884 return ret;
2885}
2886
20346722 2887/**
2888 * s2io_reset - Resets the card.
1da177e4
LT
2889 * @sp : private member of the device structure.
2890 * Description: Function to Reset the card. This function then also
20346722 2891 * restores the previously saved PCI configuration space registers as
1da177e4
LT
2892 * the card reset also resets the configuration space.
2893 * Return value:
2894 * void.
2895 */
2896
26df54bf 2897static void s2io_reset(nic_t * sp)
1da177e4
LT
2898{
2899 XENA_dev_config_t __iomem *bar0 = sp->bar0;
2900 u64 val64;
5e25b9dd 2901 u16 subid, pci_cmd;
1da177e4 2902
0b1f7ebe 2903 /* Back up the PCI-X CMD reg, dont want to lose MMRBC, OST settings */
e960fc5c 2904 pci_read_config_word(sp->pdev, PCIX_COMMAND_REGISTER, &(pci_cmd));
0b1f7ebe 2905
1da177e4
LT
2906 val64 = SW_RESET_ALL;
2907 writeq(val64, &bar0->sw_reset);
2908
20346722 2909 /*
2910 * At this stage, if the PCI write is indeed completed, the
2911 * card is reset and so is the PCI Config space of the device.
2912 * So a read cannot be issued at this stage on any of the
1da177e4
LT
2913 * registers to ensure the write into "sw_reset" register
2914 * has gone through.
2915 * Question: Is there any system call that will explicitly force
2916 * all the write commands still pending on the bus to be pushed
2917 * through?
2918 * As of now I'am just giving a 250ms delay and hoping that the
2919 * PCI write to sw_reset register is done by this time.
2920 */
2921 msleep(250);
2922
e960fc5c 2923 /* Restore the PCI state saved during initialization. */
2924 pci_restore_state(sp->pdev);
2925 pci_write_config_word(sp->pdev, PCIX_COMMAND_REGISTER,
0b1f7ebe 2926 pci_cmd);
1da177e4
LT
2927 s2io_init_pci(sp);
2928
2929 msleep(250);
2930
20346722 2931 /* Set swapper to enable I/O register access */
2932 s2io_set_swapper(sp);
2933
cc6e7c44
RA
2934 /* Restore the MSIX table entries from local variables */
2935 restore_xmsi_data(sp);
2936
5e25b9dd 2937 /* Clear certain PCI/PCI-X fields after reset */
303bcb4b 2938 if (sp->device_type == XFRAME_II_DEVICE) {
2939 /* Clear parity err detect bit */
2940 pci_write_config_word(sp->pdev, PCI_STATUS, 0x8000);
5e25b9dd 2941
303bcb4b 2942 /* Clearing PCIX Ecc status register */
2943 pci_write_config_dword(sp->pdev, 0x68, 0x7C);
5e25b9dd 2944
303bcb4b 2945 /* Clearing PCI_STATUS error reflected here */
2946 writeq(BIT(62), &bar0->txpic_int_reg);
2947 }
5e25b9dd 2948
20346722 2949 /* Reset device statistics maintained by OS */
2950 memset(&sp->stats, 0, sizeof (struct net_device_stats));
2951
1da177e4
LT
2952 /* SXE-002: Configure link and activity LED to turn it off */
2953 subid = sp->pdev->subsystem_device;
541ae68f 2954 if (((subid & 0xFF) >= 0x07) &&
2955 (sp->device_type == XFRAME_I_DEVICE)) {
1da177e4
LT
2956 val64 = readq(&bar0->gpio_control);
2957 val64 |= 0x0000800000000000ULL;
2958 writeq(val64, &bar0->gpio_control);
2959 val64 = 0x0411040400000000ULL;
509a2671 2960 writeq(val64, (void __iomem *)bar0 + 0x2700);
1da177e4
LT
2961 }
2962
541ae68f 2963 /*
2964 * Clear spurious ECC interrupts that would have occured on
2965 * XFRAME II cards after reset.
2966 */
2967 if (sp->device_type == XFRAME_II_DEVICE) {
2968 val64 = readq(&bar0->pcc_err_reg);
2969 writeq(val64, &bar0->pcc_err_reg);
2970 }
2971
1da177e4
LT
2972 sp->device_enabled_once = FALSE;
2973}
2974
2975/**
20346722 2976 * s2io_set_swapper - to set the swapper controle on the card
2977 * @sp : private member of the device structure,
1da177e4 2978 * pointer to the s2io_nic structure.
20346722 2979 * Description: Function to set the swapper control on the card
1da177e4
LT
2980 * correctly depending on the 'endianness' of the system.
2981 * Return value:
2982 * SUCCESS on success and FAILURE on failure.
2983 */
2984
26df54bf 2985static int s2io_set_swapper(nic_t * sp)
1da177e4
LT
2986{
2987 struct net_device *dev = sp->dev;
2988 XENA_dev_config_t __iomem *bar0 = sp->bar0;
2989 u64 val64, valt, valr;
2990
20346722 2991 /*
1da177e4
LT
2992 * Set proper endian settings and verify the same by reading
2993 * the PIF Feed-back register.
2994 */
2995
2996 val64 = readq(&bar0->pif_rd_swapper_fb);
2997 if (val64 != 0x0123456789ABCDEFULL) {
2998 int i = 0;
2999 u64 value[] = { 0xC30000C3C30000C3ULL, /* FE=1, SE=1 */
3000 0x8100008181000081ULL, /* FE=1, SE=0 */
3001 0x4200004242000042ULL, /* FE=0, SE=1 */
3002 0}; /* FE=0, SE=0 */
3003
3004 while(i<4) {
3005 writeq(value[i], &bar0->swapper_ctrl);
3006 val64 = readq(&bar0->pif_rd_swapper_fb);
3007 if (val64 == 0x0123456789ABCDEFULL)
3008 break;
3009 i++;
3010 }
3011 if (i == 4) {
3012 DBG_PRINT(ERR_DBG, "%s: Endian settings are wrong, ",
3013 dev->name);
3014 DBG_PRINT(ERR_DBG, "feedback read %llx\n",
3015 (unsigned long long) val64);
3016 return FAILURE;
3017 }
3018 valr = value[i];
3019 } else {
3020 valr = readq(&bar0->swapper_ctrl);
3021 }
3022
3023 valt = 0x0123456789ABCDEFULL;
3024 writeq(valt, &bar0->xmsi_address);
3025 val64 = readq(&bar0->xmsi_address);
3026
3027 if(val64 != valt) {
3028 int i = 0;
3029 u64 value[] = { 0x00C3C30000C3C300ULL, /* FE=1, SE=1 */
3030 0x0081810000818100ULL, /* FE=1, SE=0 */
3031 0x0042420000424200ULL, /* FE=0, SE=1 */
3032 0}; /* FE=0, SE=0 */
3033
3034 while(i<4) {
3035 writeq((value[i] | valr), &bar0->swapper_ctrl);
3036 writeq(valt, &bar0->xmsi_address);
3037 val64 = readq(&bar0->xmsi_address);
3038 if(val64 == valt)
3039 break;
3040 i++;
3041 }
3042 if(i == 4) {
20346722 3043 unsigned long long x = val64;
1da177e4 3044 DBG_PRINT(ERR_DBG, "Write failed, Xmsi_addr ");
20346722 3045 DBG_PRINT(ERR_DBG, "reads:0x%llx\n", x);
1da177e4
LT
3046 return FAILURE;
3047 }
3048 }
3049 val64 = readq(&bar0->swapper_ctrl);
3050 val64 &= 0xFFFF000000000000ULL;
3051
3052#ifdef __BIG_ENDIAN
20346722 3053 /*
3054 * The device by default set to a big endian format, so a
1da177e4
LT
3055 * big endian driver need not set anything.
3056 */
3057 val64 |= (SWAPPER_CTRL_TXP_FE |
3058 SWAPPER_CTRL_TXP_SE |
3059 SWAPPER_CTRL_TXD_R_FE |
3060 SWAPPER_CTRL_TXD_W_FE |
3061 SWAPPER_CTRL_TXF_R_FE |
3062 SWAPPER_CTRL_RXD_R_FE |
3063 SWAPPER_CTRL_RXD_W_FE |
3064 SWAPPER_CTRL_RXF_W_FE |
3065 SWAPPER_CTRL_XMSI_FE |
1da177e4 3066 SWAPPER_CTRL_STATS_FE | SWAPPER_CTRL_STATS_SE);
92383340 3067 if (sp->intr_type == INTA)
cc6e7c44 3068 val64 |= SWAPPER_CTRL_XMSI_SE;
1da177e4
LT
3069 writeq(val64, &bar0->swapper_ctrl);
3070#else
20346722 3071 /*
1da177e4 3072 * Initially we enable all bits to make it accessible by the
20346722 3073 * driver, then we selectively enable only those bits that
1da177e4
LT
3074 * we want to set.
3075 */
3076 val64 |= (SWAPPER_CTRL_TXP_FE |
3077 SWAPPER_CTRL_TXP_SE |
3078 SWAPPER_CTRL_TXD_R_FE |
3079 SWAPPER_CTRL_TXD_R_SE |
3080 SWAPPER_CTRL_TXD_W_FE |
3081 SWAPPER_CTRL_TXD_W_SE |
3082 SWAPPER_CTRL_TXF_R_FE |
3083 SWAPPER_CTRL_RXD_R_FE |
3084 SWAPPER_CTRL_RXD_R_SE |
3085 SWAPPER_CTRL_RXD_W_FE |
3086 SWAPPER_CTRL_RXD_W_SE |
3087 SWAPPER_CTRL_RXF_W_FE |
3088 SWAPPER_CTRL_XMSI_FE |
1da177e4 3089 SWAPPER_CTRL_STATS_FE | SWAPPER_CTRL_STATS_SE);
cc6e7c44
RA
3090 if (sp->intr_type == INTA)
3091 val64 |= SWAPPER_CTRL_XMSI_SE;
1da177e4
LT
3092 writeq(val64, &bar0->swapper_ctrl);
3093#endif
3094 val64 = readq(&bar0->swapper_ctrl);
3095
20346722 3096 /*
3097 * Verifying if endian settings are accurate by reading a
1da177e4
LT
3098 * feedback register.
3099 */
3100 val64 = readq(&bar0->pif_rd_swapper_fb);
3101 if (val64 != 0x0123456789ABCDEFULL) {
3102 /* Endian settings are incorrect, calls for another dekko. */
3103 DBG_PRINT(ERR_DBG, "%s: Endian settings are wrong, ",
3104 dev->name);
3105 DBG_PRINT(ERR_DBG, "feedback read %llx\n",
3106 (unsigned long long) val64);
3107 return FAILURE;
3108 }
3109
3110 return SUCCESS;
3111}
3112
ac1f60db 3113static int wait_for_msix_trans(nic_t *nic, int i)
cc6e7c44 3114{
37eb47ed 3115 XENA_dev_config_t __iomem *bar0 = nic->bar0;
cc6e7c44
RA
3116 u64 val64;
3117 int ret = 0, cnt = 0;
3118
3119 do {
3120 val64 = readq(&bar0->xmsi_access);
3121 if (!(val64 & BIT(15)))
3122 break;
3123 mdelay(1);
3124 cnt++;
3125 } while(cnt < 5);
3126 if (cnt == 5) {
3127 DBG_PRINT(ERR_DBG, "XMSI # %d Access failed\n", i);
3128 ret = 1;
3129 }
3130
3131 return ret;
3132}
3133
26df54bf 3134static void restore_xmsi_data(nic_t *nic)
cc6e7c44 3135{
37eb47ed 3136 XENA_dev_config_t __iomem *bar0 = nic->bar0;
cc6e7c44
RA
3137 u64 val64;
3138 int i;
3139
3140 for (i=0; i< MAX_REQUESTED_MSI_X; i++) {
3141 writeq(nic->msix_info[i].addr, &bar0->xmsi_address);
3142 writeq(nic->msix_info[i].data, &bar0->xmsi_data);
3143 val64 = (BIT(7) | BIT(15) | vBIT(i, 26, 6));
3144 writeq(val64, &bar0->xmsi_access);
3145 if (wait_for_msix_trans(nic, i)) {
3146 DBG_PRINT(ERR_DBG, "failed in %s\n", __FUNCTION__);
3147 continue;
3148 }
3149 }
3150}
3151
ac1f60db 3152static void store_xmsi_data(nic_t *nic)
cc6e7c44 3153{
37eb47ed 3154 XENA_dev_config_t __iomem *bar0 = nic->bar0;
cc6e7c44
RA
3155 u64 val64, addr, data;
3156 int i;
3157
3158 /* Store and display */
3159 for (i=0; i< MAX_REQUESTED_MSI_X; i++) {
3160 val64 = (BIT(15) | vBIT(i, 26, 6));
3161 writeq(val64, &bar0->xmsi_access);
3162 if (wait_for_msix_trans(nic, i)) {
3163 DBG_PRINT(ERR_DBG, "failed in %s\n", __FUNCTION__);
3164 continue;
3165 }
3166 addr = readq(&bar0->xmsi_address);
3167 data = readq(&bar0->xmsi_data);
3168 if (addr && data) {
3169 nic->msix_info[i].addr = addr;
3170 nic->msix_info[i].data = data;
3171 }
3172 }
3173}
3174
3175int s2io_enable_msi(nic_t *nic)
3176{
37eb47ed 3177 XENA_dev_config_t __iomem *bar0 = nic->bar0;
cc6e7c44
RA
3178 u16 msi_ctrl, msg_val;
3179 struct config_param *config = &nic->config;
3180 struct net_device *dev = nic->dev;
3181 u64 val64, tx_mat, rx_mat;
3182 int i, err;
3183
3184 val64 = readq(&bar0->pic_control);
3185 val64 &= ~BIT(1);
3186 writeq(val64, &bar0->pic_control);
3187
3188 err = pci_enable_msi(nic->pdev);
3189 if (err) {
3190 DBG_PRINT(ERR_DBG, "%s: enabling MSI failed\n",
3191 nic->dev->name);
3192 return err;
3193 }
3194
3195 /*
3196 * Enable MSI and use MSI-1 in stead of the standard MSI-0
3197 * for interrupt handling.
3198 */
3199 pci_read_config_word(nic->pdev, 0x4c, &msg_val);
3200 msg_val ^= 0x1;
3201 pci_write_config_word(nic->pdev, 0x4c, msg_val);
3202 pci_read_config_word(nic->pdev, 0x4c, &msg_val);
3203
3204 pci_read_config_word(nic->pdev, 0x42, &msi_ctrl);
3205 msi_ctrl |= 0x10;
3206 pci_write_config_word(nic->pdev, 0x42, msi_ctrl);
3207
3208 /* program MSI-1 into all usable Tx_Mat and Rx_Mat fields */
3209 tx_mat = readq(&bar0->tx_mat0_n[0]);
3210 for (i=0; i<config->tx_fifo_num; i++) {
3211 tx_mat |= TX_MAT_SET(i, 1);
3212 }
3213 writeq(tx_mat, &bar0->tx_mat0_n[0]);
3214
3215 rx_mat = readq(&bar0->rx_mat);
3216 for (i=0; i<config->rx_ring_num; i++) {
3217 rx_mat |= RX_MAT_SET(i, 1);
3218 }
3219 writeq(rx_mat, &bar0->rx_mat);
3220
3221 dev->irq = nic->pdev->irq;
3222 return 0;
3223}
3224
26df54bf 3225static int s2io_enable_msi_x(nic_t *nic)
cc6e7c44 3226{
37eb47ed 3227 XENA_dev_config_t __iomem *bar0 = nic->bar0;
cc6e7c44
RA
3228 u64 tx_mat, rx_mat;
3229 u16 msi_control; /* Temp variable */
3230 int ret, i, j, msix_indx = 1;
3231
3232 nic->entries = kmalloc(MAX_REQUESTED_MSI_X * sizeof(struct msix_entry),
3233 GFP_KERNEL);
3234 if (nic->entries == NULL) {
3235 DBG_PRINT(ERR_DBG, "%s: Memory allocation failed\n", __FUNCTION__);
3236 return -ENOMEM;
3237 }
3238 memset(nic->entries, 0, MAX_REQUESTED_MSI_X * sizeof(struct msix_entry));
3239
3240 nic->s2io_entries =
3241 kmalloc(MAX_REQUESTED_MSI_X * sizeof(struct s2io_msix_entry),
3242 GFP_KERNEL);
3243 if (nic->s2io_entries == NULL) {
3244 DBG_PRINT(ERR_DBG, "%s: Memory allocation failed\n", __FUNCTION__);
3245 kfree(nic->entries);
3246 return -ENOMEM;
3247 }
3248 memset(nic->s2io_entries, 0,
3249 MAX_REQUESTED_MSI_X * sizeof(struct s2io_msix_entry));
3250
3251 for (i=0; i< MAX_REQUESTED_MSI_X; i++) {
3252 nic->entries[i].entry = i;
3253 nic->s2io_entries[i].entry = i;
3254 nic->s2io_entries[i].arg = NULL;
3255 nic->s2io_entries[i].in_use = 0;
3256 }
3257
3258 tx_mat = readq(&bar0->tx_mat0_n[0]);
3259 for (i=0; i<nic->config.tx_fifo_num; i++, msix_indx++) {
3260 tx_mat |= TX_MAT_SET(i, msix_indx);
3261 nic->s2io_entries[msix_indx].arg = &nic->mac_control.fifos[i];
3262 nic->s2io_entries[msix_indx].type = MSIX_FIFO_TYPE;
3263 nic->s2io_entries[msix_indx].in_use = MSIX_FLG;
3264 }
3265 writeq(tx_mat, &bar0->tx_mat0_n[0]);
3266
3267 if (!nic->config.bimodal) {
3268 rx_mat = readq(&bar0->rx_mat);
3269 for (j=0; j<nic->config.rx_ring_num; j++, msix_indx++) {
3270 rx_mat |= RX_MAT_SET(j, msix_indx);
3271 nic->s2io_entries[msix_indx].arg = &nic->mac_control.rings[j];
3272 nic->s2io_entries[msix_indx].type = MSIX_RING_TYPE;
3273 nic->s2io_entries[msix_indx].in_use = MSIX_FLG;
3274 }
3275 writeq(rx_mat, &bar0->rx_mat);
3276 } else {
3277 tx_mat = readq(&bar0->tx_mat0_n[7]);
3278 for (j=0; j<nic->config.rx_ring_num; j++, msix_indx++) {
3279 tx_mat |= TX_MAT_SET(i, msix_indx);
3280 nic->s2io_entries[msix_indx].arg = &nic->mac_control.rings[j];
3281 nic->s2io_entries[msix_indx].type = MSIX_RING_TYPE;
3282 nic->s2io_entries[msix_indx].in_use = MSIX_FLG;
3283 }
3284 writeq(tx_mat, &bar0->tx_mat0_n[7]);
3285 }
3286
3287 ret = pci_enable_msix(nic->pdev, nic->entries, MAX_REQUESTED_MSI_X);
3288 if (ret) {
3289 DBG_PRINT(ERR_DBG, "%s: Enabling MSIX failed\n", nic->dev->name);
3290 kfree(nic->entries);
3291 kfree(nic->s2io_entries);
3292 nic->entries = NULL;
3293 nic->s2io_entries = NULL;
3294 return -ENOMEM;
3295 }
3296
3297 /*
3298 * To enable MSI-X, MSI also needs to be enabled, due to a bug
3299 * in the herc NIC. (Temp change, needs to be removed later)
3300 */
3301 pci_read_config_word(nic->pdev, 0x42, &msi_control);
3302 msi_control |= 0x1; /* Enable MSI */
3303 pci_write_config_word(nic->pdev, 0x42, msi_control);
3304
3305 return 0;
3306}
3307
1da177e4
LT
3308/* ********************************************************* *
3309 * Functions defined below concern the OS part of the driver *
3310 * ********************************************************* */
3311
20346722 3312/**
1da177e4
LT
3313 * s2io_open - open entry point of the driver
3314 * @dev : pointer to the device structure.
3315 * Description:
3316 * This function is the open entry point of the driver. It mainly calls a
3317 * function to allocate Rx buffers and inserts them into the buffer
20346722 3318 * descriptors and then enables the Rx part of the NIC.
1da177e4
LT
3319 * Return value:
3320 * 0 on success and an appropriate (-)ve integer as defined in errno.h
3321 * file on failure.
3322 */
3323
ac1f60db 3324static int s2io_open(struct net_device *dev)
1da177e4
LT
3325{
3326 nic_t *sp = dev->priv;
3327 int err = 0;
cc6e7c44
RA
3328 int i;
3329 u16 msi_control; /* Temp variable */
1da177e4 3330
20346722 3331 /*
3332 * Make sure you have link off by default every time
1da177e4
LT
3333 * Nic is initialized
3334 */
3335 netif_carrier_off(dev);
0b1f7ebe 3336 sp->last_link_state = 0;
1da177e4
LT
3337
3338 /* Initialize H/W and enable interrupts */
3339 if (s2io_card_up(sp)) {
3340 DBG_PRINT(ERR_DBG, "%s: H/W initialization failed\n",
3341 dev->name);
20346722 3342 err = -ENODEV;
3343 goto hw_init_failed;
1da177e4
LT
3344 }
3345
cc6e7c44
RA
3346 /* Store the values of the MSIX table in the nic_t structure */
3347 store_xmsi_data(sp);
3348
1da177e4 3349 /* After proper initialization of H/W, register ISR */
cc6e7c44
RA
3350 if (sp->intr_type == MSI) {
3351 err = request_irq((int) sp->pdev->irq, s2io_msi_handle,
3352 SA_SHIRQ, sp->name, dev);
3353 if (err) {
3354 DBG_PRINT(ERR_DBG, "%s: MSI registration \
3355failed\n", dev->name);
3356 goto isr_registration_failed;
3357 }
3358 }
3359 if (sp->intr_type == MSI_X) {
3360 for (i=1; (sp->s2io_entries[i].in_use == MSIX_FLG); i++) {
3361 if (sp->s2io_entries[i].type == MSIX_FIFO_TYPE) {
3362 sprintf(sp->desc1, "%s:MSI-X-%d-TX",
3363 dev->name, i);
3364 err = request_irq(sp->entries[i].vector,
3365 s2io_msix_fifo_handle, 0, sp->desc1,
3366 sp->s2io_entries[i].arg);
3367 DBG_PRINT(ERR_DBG, "%s @ 0x%llx\n", sp->desc1,
26b7625c 3368 (unsigned long long)sp->msix_info[i].addr);
cc6e7c44
RA
3369 } else {
3370 sprintf(sp->desc2, "%s:MSI-X-%d-RX",
3371 dev->name, i);
3372 err = request_irq(sp->entries[i].vector,
3373 s2io_msix_ring_handle, 0, sp->desc2,
3374 sp->s2io_entries[i].arg);
3375 DBG_PRINT(ERR_DBG, "%s @ 0x%llx\n", sp->desc2,
26b7625c 3376 (unsigned long long)sp->msix_info[i].addr);
cc6e7c44
RA
3377 }
3378 if (err) {
3379 DBG_PRINT(ERR_DBG, "%s: MSI-X-%d registration \
3380failed\n", dev->name, i);
3381 DBG_PRINT(ERR_DBG, "Returned: %d\n", err);
3382 goto isr_registration_failed;
3383 }
3384 sp->s2io_entries[i].in_use = MSIX_REGISTERED_SUCCESS;
3385 }
3386 }
3387 if (sp->intr_type == INTA) {
3388 err = request_irq((int) sp->pdev->irq, s2io_isr, SA_SHIRQ,
3389 sp->name, dev);
3390 if (err) {
3391 DBG_PRINT(ERR_DBG, "%s: ISR registration failed\n",
3392 dev->name);
3393 goto isr_registration_failed;
3394 }
1da177e4
LT
3395 }
3396
3397 if (s2io_set_mac_addr(dev, dev->dev_addr) == FAILURE) {
3398 DBG_PRINT(ERR_DBG, "Set Mac Address Failed\n");
20346722 3399 err = -ENODEV;
3400 goto setting_mac_address_failed;
1da177e4
LT
3401 }
3402
3403 netif_start_queue(dev);
3404 return 0;
20346722 3405
3406setting_mac_address_failed:
cc6e7c44
RA
3407 if (sp->intr_type != MSI_X)
3408 free_irq(sp->pdev->irq, dev);
20346722 3409isr_registration_failed:
25fff88e 3410 del_timer_sync(&sp->alarm_timer);
cc6e7c44
RA
3411 if (sp->intr_type == MSI_X) {
3412 if (sp->device_type == XFRAME_II_DEVICE) {
3413 for (i=1; (sp->s2io_entries[i].in_use ==
3414 MSIX_REGISTERED_SUCCESS); i++) {
3415 int vector = sp->entries[i].vector;
3416 void *arg = sp->s2io_entries[i].arg;
3417
3418 free_irq(vector, arg);
3419 }
3420 pci_disable_msix(sp->pdev);
3421
3422 /* Temp */
3423 pci_read_config_word(sp->pdev, 0x42, &msi_control);
3424 msi_control &= 0xFFFE; /* Disable MSI */
3425 pci_write_config_word(sp->pdev, 0x42, msi_control);
3426 }
3427 }
3428 else if (sp->intr_type == MSI)
3429 pci_disable_msi(sp->pdev);
20346722 3430 s2io_reset(sp);
3431hw_init_failed:
cc6e7c44
RA
3432 if (sp->intr_type == MSI_X) {
3433 if (sp->entries)
3434 kfree(sp->entries);
3435 if (sp->s2io_entries)
3436 kfree(sp->s2io_entries);
3437 }
20346722 3438 return err;
1da177e4
LT
3439}
3440
3441/**
3442 * s2io_close -close entry point of the driver
3443 * @dev : device pointer.
3444 * Description:
3445 * This is the stop entry point of the driver. It needs to undo exactly
3446 * whatever was done by the open entry point,thus it's usually referred to
3447 * as the close function.Among other things this function mainly stops the
3448 * Rx side of the NIC and frees all the Rx buffers in the Rx rings.
3449 * Return value:
3450 * 0 on success and an appropriate (-)ve integer as defined in errno.h
3451 * file on failure.
3452 */
3453
ac1f60db 3454static int s2io_close(struct net_device *dev)
1da177e4
LT
3455{
3456 nic_t *sp = dev->priv;
cc6e7c44
RA
3457 int i;
3458 u16 msi_control;
3459
1da177e4
LT
3460 flush_scheduled_work();
3461 netif_stop_queue(dev);
3462 /* Reset card, kill tasklet and free Tx and Rx buffers. */
3463 s2io_card_down(sp);
3464
cc6e7c44
RA
3465 if (sp->intr_type == MSI_X) {
3466 if (sp->device_type == XFRAME_II_DEVICE) {
3467 for (i=1; (sp->s2io_entries[i].in_use ==
3468 MSIX_REGISTERED_SUCCESS); i++) {
3469 int vector = sp->entries[i].vector;
3470 void *arg = sp->s2io_entries[i].arg;
3471
3472 free_irq(vector, arg);
3473 }
3474 pci_read_config_word(sp->pdev, 0x42, &msi_control);
3475 msi_control &= 0xFFFE; /* Disable MSI */
3476 pci_write_config_word(sp->pdev, 0x42, msi_control);
3477
3478 pci_disable_msix(sp->pdev);
3479 }
3480 }
3481 else {
3482 free_irq(sp->pdev->irq, dev);
3483 if (sp->intr_type == MSI)
3484 pci_disable_msi(sp->pdev);
3485 }
1da177e4
LT
3486 sp->device_close_flag = TRUE; /* Device is shut down. */
3487 return 0;
3488}
3489
3490/**
3491 * s2io_xmit - Tx entry point of te driver
3492 * @skb : the socket buffer containing the Tx data.
3493 * @dev : device pointer.
3494 * Description :
3495 * This function is the Tx entry point of the driver. S2IO NIC supports
3496 * certain protocol assist features on Tx side, namely CSO, S/G, LSO.
3497 * NOTE: when device cant queue the pkt,just the trans_start variable will
3498 * not be upadted.
3499 * Return value:
3500 * 0 on success & 1 on failure.
3501 */
3502
ac1f60db 3503static int s2io_xmit(struct sk_buff *skb, struct net_device *dev)
1da177e4
LT
3504{
3505 nic_t *sp = dev->priv;
3506 u16 frg_cnt, frg_len, i, queue, queue_len, put_off, get_off;
3507 register u64 val64;
3508 TxD_t *txdp;
3509 TxFIFO_element_t __iomem *tx_fifo;
3510 unsigned long flags;
3511#ifdef NETIF_F_TSO
3512 int mss;
3513#endif
be3a6b02 3514 u16 vlan_tag = 0;
3515 int vlan_priority = 0;
1da177e4
LT
3516 mac_info_t *mac_control;
3517 struct config_param *config;
1da177e4
LT
3518
3519 mac_control = &sp->mac_control;
3520 config = &sp->config;
3521
20346722 3522 DBG_PRINT(TX_DBG, "%s: In Neterion Tx routine\n", dev->name);
1da177e4 3523 spin_lock_irqsave(&sp->tx_lock, flags);
1da177e4 3524 if (atomic_read(&sp->card_state) == CARD_DOWN) {
20346722 3525 DBG_PRINT(TX_DBG, "%s: Card going down for reset\n",
1da177e4
LT
3526 dev->name);
3527 spin_unlock_irqrestore(&sp->tx_lock, flags);
20346722 3528 dev_kfree_skb(skb);
3529 return 0;
1da177e4
LT
3530 }
3531
3532 queue = 0;
1da177e4 3533
be3a6b02 3534 /* Get Fifo number to Transmit based on vlan priority */
3535 if (sp->vlgrp && vlan_tx_tag_present(skb)) {
3536 vlan_tag = vlan_tx_tag_get(skb);
3537 vlan_priority = vlan_tag >> 13;
3538 queue = config->fifo_mapping[vlan_priority];
3539 }
3540
20346722 3541 put_off = (u16) mac_control->fifos[queue].tx_curr_put_info.offset;
3542 get_off = (u16) mac_control->fifos[queue].tx_curr_get_info.offset;
3543 txdp = (TxD_t *) mac_control->fifos[queue].list_info[put_off].
3544 list_virt_addr;
3545
3546 queue_len = mac_control->fifos[queue].tx_curr_put_info.fifo_len + 1;
1da177e4
LT
3547 /* Avoid "put" pointer going beyond "get" pointer */
3548 if (txdp->Host_Control || (((put_off + 1) % queue_len) == get_off)) {
776bd20f 3549 DBG_PRINT(TX_DBG, "Error in xmit, No free TXDs.\n");
1da177e4
LT
3550 netif_stop_queue(dev);
3551 dev_kfree_skb(skb);
3552 spin_unlock_irqrestore(&sp->tx_lock, flags);
3553 return 0;
3554 }
0b1f7ebe 3555
3556 /* A buffer with no data will be dropped */
3557 if (!skb->len) {
3558 DBG_PRINT(TX_DBG, "%s:Buffer has no data..\n", dev->name);
3559 dev_kfree_skb(skb);
3560 spin_unlock_irqrestore(&sp->tx_lock, flags);
3561 return 0;
3562 }
3563
fed5eccd
AR
3564 txdp->Control_1 = 0;
3565 txdp->Control_2 = 0;
1da177e4
LT
3566#ifdef NETIF_F_TSO
3567 mss = skb_shinfo(skb)->tso_size;
3568 if (mss) {
3569 txdp->Control_1 |= TXD_TCP_LSO_EN;
3570 txdp->Control_1 |= TXD_TCP_LSO_MSS(mss);
3571 }
3572#endif
1da177e4
LT
3573 if (skb->ip_summed == CHECKSUM_HW) {
3574 txdp->Control_2 |=
3575 (TXD_TX_CKO_IPV4_EN | TXD_TX_CKO_TCP_EN |
3576 TXD_TX_CKO_UDP_EN);
3577 }
fed5eccd
AR
3578 txdp->Control_1 |= TXD_GATHER_CODE_FIRST;
3579 txdp->Control_1 |= TXD_LIST_OWN_XENA;
1da177e4 3580 txdp->Control_2 |= config->tx_intr_type;
d8892c6e 3581
be3a6b02 3582 if (sp->vlgrp && vlan_tx_tag_present(skb)) {
3583 txdp->Control_2 |= TXD_VLAN_ENABLE;
3584 txdp->Control_2 |= TXD_VLAN_TAG(vlan_tag);
3585 }
3586
fed5eccd
AR
3587 frg_len = skb->len - skb->data_len;
3588 if (skb_shinfo(skb)->ufo_size) {
3589 int ufo_size;
3590
3591 ufo_size = skb_shinfo(skb)->ufo_size;
3592 ufo_size &= ~7;
3593 txdp->Control_1 |= TXD_UFO_EN;
3594 txdp->Control_1 |= TXD_UFO_MSS(ufo_size);
3595 txdp->Control_1 |= TXD_BUFFER0_SIZE(8);
3596#ifdef __BIG_ENDIAN
3597 sp->ufo_in_band_v[put_off] =
3598 (u64)skb_shinfo(skb)->ip6_frag_id;
3599#else
3600 sp->ufo_in_band_v[put_off] =
3601 (u64)skb_shinfo(skb)->ip6_frag_id << 32;
3602#endif
3603 txdp->Host_Control = (unsigned long)sp->ufo_in_band_v;
3604 txdp->Buffer_Pointer = pci_map_single(sp->pdev,
3605 sp->ufo_in_band_v,
3606 sizeof(u64), PCI_DMA_TODEVICE);
3607 txdp++;
3608 txdp->Control_1 = 0;
3609 txdp->Control_2 = 0;
3610 }
1da177e4 3611
fed5eccd
AR
3612 txdp->Buffer_Pointer = pci_map_single
3613 (sp->pdev, skb->data, frg_len, PCI_DMA_TODEVICE);
3614 txdp->Host_Control = (unsigned long) skb;
3615 txdp->Control_1 |= TXD_BUFFER0_SIZE(frg_len);
3616
3617 if (skb_shinfo(skb)->ufo_size)
3618 txdp->Control_1 |= TXD_UFO_EN;
3619
3620 frg_cnt = skb_shinfo(skb)->nr_frags;
1da177e4
LT
3621 /* For fragmented SKB. */
3622 for (i = 0; i < frg_cnt; i++) {
3623 skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
0b1f7ebe 3624 /* A '0' length fragment will be ignored */
3625 if (!frag->size)
3626 continue;
1da177e4
LT
3627 txdp++;
3628 txdp->Buffer_Pointer = (u64) pci_map_page
3629 (sp->pdev, frag->page, frag->page_offset,
3630 frag->size, PCI_DMA_TODEVICE);
efd51b5c 3631 txdp->Control_1 = TXD_BUFFER0_SIZE(frag->size);
fed5eccd
AR
3632 if (skb_shinfo(skb)->ufo_size)
3633 txdp->Control_1 |= TXD_UFO_EN;
1da177e4
LT
3634 }
3635 txdp->Control_1 |= TXD_GATHER_CODE_LAST;
3636
fed5eccd
AR
3637 if (skb_shinfo(skb)->ufo_size)
3638 frg_cnt++; /* as Txd0 was used for inband header */
3639
1da177e4 3640 tx_fifo = mac_control->tx_FIFO_start[queue];
20346722 3641 val64 = mac_control->fifos[queue].list_info[put_off].list_phy_addr;
1da177e4
LT
3642 writeq(val64, &tx_fifo->TxDL_Pointer);
3643
3644 val64 = (TX_FIFO_LAST_TXD_NUM(frg_cnt) | TX_FIFO_FIRST_LIST |
3645 TX_FIFO_LAST_LIST);
20346722 3646
1da177e4
LT
3647#ifdef NETIF_F_TSO
3648 if (mss)
3649 val64 |= TX_FIFO_SPECIAL_FUNC;
3650#endif
fed5eccd
AR
3651 if (skb_shinfo(skb)->ufo_size)
3652 val64 |= TX_FIFO_SPECIAL_FUNC;
1da177e4
LT
3653 writeq(val64, &tx_fifo->List_Control);
3654
303bcb4b 3655 mmiowb();
3656
1da177e4 3657 put_off++;
20346722 3658 put_off %= mac_control->fifos[queue].tx_curr_put_info.fifo_len + 1;
3659 mac_control->fifos[queue].tx_curr_put_info.offset = put_off;
1da177e4
LT
3660
3661 /* Avoid "put" pointer going beyond "get" pointer */
3662 if (((put_off + 1) % queue_len) == get_off) {
3663 DBG_PRINT(TX_DBG,
3664 "No free TxDs for xmit, Put: 0x%x Get:0x%x\n",
3665 put_off, get_off);
3666 netif_stop_queue(dev);
3667 }
3668
3669 dev->trans_start = jiffies;
3670 spin_unlock_irqrestore(&sp->tx_lock, flags);
3671
3672 return 0;
3673}
3674
25fff88e 3675static void
3676s2io_alarm_handle(unsigned long data)
3677{
3678 nic_t *sp = (nic_t *)data;
3679
3680 alarm_intr_handler(sp);
3681 mod_timer(&sp->alarm_timer, jiffies + HZ / 2);
3682}
3683
cc6e7c44
RA
3684static irqreturn_t
3685s2io_msi_handle(int irq, void *dev_id, struct pt_regs *regs)
3686{
3687 struct net_device *dev = (struct net_device *) dev_id;
3688 nic_t *sp = dev->priv;
3689 int i;
3690 int ret;
3691 mac_info_t *mac_control;
3692 struct config_param *config;
3693
3694 atomic_inc(&sp->isr_cnt);
3695 mac_control = &sp->mac_control;
3696 config = &sp->config;
3697 DBG_PRINT(INTR_DBG, "%s: MSI handler\n", __FUNCTION__);
3698
3699 /* If Intr is because of Rx Traffic */
3700 for (i = 0; i < config->rx_ring_num; i++)
3701 rx_intr_handler(&mac_control->rings[i]);
3702
3703 /* If Intr is because of Tx Traffic */
3704 for (i = 0; i < config->tx_fifo_num; i++)
3705 tx_intr_handler(&mac_control->fifos[i]);
3706
3707 /*
3708 * If the Rx buffer count is below the panic threshold then
3709 * reallocate the buffers from the interrupt handler itself,
3710 * else schedule a tasklet to reallocate the buffers.
3711 */
3712 for (i = 0; i < config->rx_ring_num; i++) {
7d3d0439
RA
3713 if (!sp->lro) {
3714 int rxb_size = atomic_read(&sp->rx_bufs_left[i]);
3715 int level = rx_buffer_level(sp, rxb_size, i);
3716
3717 if ((level == PANIC) && (!TASKLET_IN_USE)) {
3718 DBG_PRINT(INTR_DBG, "%s: Rx BD hit ",
3719 dev->name);
3720 DBG_PRINT(INTR_DBG, "PANIC levels\n");
3721 if ((ret = fill_rx_buffers(sp, i)) == -ENOMEM) {
3722 DBG_PRINT(ERR_DBG, "%s:Out of memory",
3723 dev->name);
3724 DBG_PRINT(ERR_DBG, " in ISR!!\n");
3725 clear_bit(0, (&sp->tasklet_status));
3726 atomic_dec(&sp->isr_cnt);
3727 return IRQ_HANDLED;
3728 }
cc6e7c44 3729 clear_bit(0, (&sp->tasklet_status));
7d3d0439
RA
3730 } else if (level == LOW) {
3731 tasklet_schedule(&sp->task);
cc6e7c44 3732 }
7d3d0439
RA
3733 }
3734 else if (fill_rx_buffers(sp, i) == -ENOMEM) {
3735 DBG_PRINT(ERR_DBG, "%s:Out of memory",
3736 dev->name);
3737 DBG_PRINT(ERR_DBG, " in Rx Intr!!\n");
3738 break;
cc6e7c44
RA
3739 }
3740 }
3741
3742 atomic_dec(&sp->isr_cnt);
3743 return IRQ_HANDLED;
3744}
3745
3746static irqreturn_t
3747s2io_msix_ring_handle(int irq, void *dev_id, struct pt_regs *regs)
3748{
3749 ring_info_t *ring = (ring_info_t *)dev_id;
3750 nic_t *sp = ring->nic;
7d3d0439 3751 struct net_device *dev = (struct net_device *) dev_id;
cc6e7c44
RA
3752 int rxb_size, level, rng_n;
3753
3754 atomic_inc(&sp->isr_cnt);
3755 rx_intr_handler(ring);
3756
3757 rng_n = ring->ring_no;
7d3d0439
RA
3758 if (!sp->lro) {
3759 rxb_size = atomic_read(&sp->rx_bufs_left[rng_n]);
3760 level = rx_buffer_level(sp, rxb_size, rng_n);
3761
3762 if ((level == PANIC) && (!TASKLET_IN_USE)) {
3763 int ret;
3764 DBG_PRINT(INTR_DBG, "%s: Rx BD hit ", __FUNCTION__);
3765 DBG_PRINT(INTR_DBG, "PANIC levels\n");
3766 if ((ret = fill_rx_buffers(sp, rng_n)) == -ENOMEM) {
3767 DBG_PRINT(ERR_DBG, "Out of memory in %s",
3768 __FUNCTION__);
3769 clear_bit(0, (&sp->tasklet_status));
3770 return IRQ_HANDLED;
3771 }
cc6e7c44 3772 clear_bit(0, (&sp->tasklet_status));
7d3d0439
RA
3773 } else if (level == LOW) {
3774 tasklet_schedule(&sp->task);
cc6e7c44 3775 }
cc6e7c44 3776 }
7d3d0439
RA
3777 else if (fill_rx_buffers(sp, rng_n) == -ENOMEM) {
3778 DBG_PRINT(ERR_DBG, "%s:Out of memory", dev->name);
3779 DBG_PRINT(ERR_DBG, " in Rx Intr!!\n");
3780 }
3781
cc6e7c44
RA
3782 atomic_dec(&sp->isr_cnt);
3783
3784 return IRQ_HANDLED;
3785}
3786
3787static irqreturn_t
3788s2io_msix_fifo_handle(int irq, void *dev_id, struct pt_regs *regs)
3789{
3790 fifo_info_t *fifo = (fifo_info_t *)dev_id;
3791 nic_t *sp = fifo->nic;
3792
3793 atomic_inc(&sp->isr_cnt);
3794 tx_intr_handler(fifo);
3795 atomic_dec(&sp->isr_cnt);
3796 return IRQ_HANDLED;
3797}
3798
a371a07d 3799static void s2io_txpic_intr_handle(nic_t *sp)
3800{
509a2671 3801 XENA_dev_config_t __iomem *bar0 = sp->bar0;
a371a07d 3802 u64 val64;
3803
3804 val64 = readq(&bar0->pic_int_status);
3805 if (val64 & PIC_INT_GPIO) {
3806 val64 = readq(&bar0->gpio_int_reg);
3807 if ((val64 & GPIO_INT_REG_LINK_DOWN) &&
3808 (val64 & GPIO_INT_REG_LINK_UP)) {
3809 val64 |= GPIO_INT_REG_LINK_DOWN;
3810 val64 |= GPIO_INT_REG_LINK_UP;
3811 writeq(val64, &bar0->gpio_int_reg);
3812 goto masking;
3813 }
3814
3815 if (((sp->last_link_state == LINK_UP) &&
3816 (val64 & GPIO_INT_REG_LINK_DOWN)) ||
3817 ((sp->last_link_state == LINK_DOWN) &&
3818 (val64 & GPIO_INT_REG_LINK_UP))) {
3819 val64 = readq(&bar0->gpio_int_mask);
3820 val64 |= GPIO_INT_MASK_LINK_DOWN;
3821 val64 |= GPIO_INT_MASK_LINK_UP;
3822 writeq(val64, &bar0->gpio_int_mask);
3823 s2io_set_link((unsigned long)sp);
3824 }
3825masking:
3826 if (sp->last_link_state == LINK_UP) {
3827 /*enable down interrupt */
3828 val64 = readq(&bar0->gpio_int_mask);
3829 /* unmasks link down intr */
3830 val64 &= ~GPIO_INT_MASK_LINK_DOWN;
3831 /* masks link up intr */
3832 val64 |= GPIO_INT_MASK_LINK_UP;
3833 writeq(val64, &bar0->gpio_int_mask);
3834 } else {
3835 /*enable UP Interrupt */
3836 val64 = readq(&bar0->gpio_int_mask);
3837 /* unmasks link up interrupt */
3838 val64 &= ~GPIO_INT_MASK_LINK_UP;
3839 /* masks link down interrupt */
3840 val64 |= GPIO_INT_MASK_LINK_DOWN;
3841 writeq(val64, &bar0->gpio_int_mask);
3842 }
3843 }
3844}
3845
1da177e4
LT
3846/**
3847 * s2io_isr - ISR handler of the device .
3848 * @irq: the irq of the device.
3849 * @dev_id: a void pointer to the dev structure of the NIC.
3850 * @pt_regs: pointer to the registers pushed on the stack.
20346722 3851 * Description: This function is the ISR handler of the device. It
3852 * identifies the reason for the interrupt and calls the relevant
3853 * service routines. As a contongency measure, this ISR allocates the
1da177e4
LT
3854 * recv buffers, if their numbers are below the panic value which is
3855 * presently set to 25% of the original number of rcv buffers allocated.
3856 * Return value:
20346722 3857 * IRQ_HANDLED: will be returned if IRQ was handled by this routine
1da177e4
LT
3858 * IRQ_NONE: will be returned if interrupt is not from our device
3859 */
3860static irqreturn_t s2io_isr(int irq, void *dev_id, struct pt_regs *regs)
3861{
3862 struct net_device *dev = (struct net_device *) dev_id;
3863 nic_t *sp = dev->priv;
3864 XENA_dev_config_t __iomem *bar0 = sp->bar0;
20346722 3865 int i;
fe113638 3866 u64 reason = 0, val64;
1da177e4
LT
3867 mac_info_t *mac_control;
3868 struct config_param *config;
3869
7ba013ac 3870 atomic_inc(&sp->isr_cnt);
1da177e4
LT
3871 mac_control = &sp->mac_control;
3872 config = &sp->config;
3873
20346722 3874 /*
1da177e4
LT
3875 * Identify the cause for interrupt and call the appropriate
3876 * interrupt handler. Causes for the interrupt could be;
3877 * 1. Rx of packet.
3878 * 2. Tx complete.
3879 * 3. Link down.
20346722 3880 * 4. Error in any functional blocks of the NIC.
1da177e4
LT
3881 */
3882 reason = readq(&bar0->general_int_status);
3883
3884 if (!reason) {
3885 /* The interrupt was not raised by Xena. */
7ba013ac 3886 atomic_dec(&sp->isr_cnt);
1da177e4
LT
3887 return IRQ_NONE;
3888 }
3889
1da177e4
LT
3890#ifdef CONFIG_S2IO_NAPI
3891 if (reason & GEN_INTR_RXTRAFFIC) {
3892 if (netif_rx_schedule_prep(dev)) {
3893 en_dis_able_nic_intrs(sp, RX_TRAFFIC_INTR,
3894 DISABLE_INTRS);
3895 __netif_rx_schedule(dev);
3896 }
3897 }
3898#else
3899 /* If Intr is because of Rx Traffic */
3900 if (reason & GEN_INTR_RXTRAFFIC) {
fe113638 3901 /*
3902 * rx_traffic_int reg is an R1 register, writing all 1's
3903 * will ensure that the actual interrupt causing bit get's
3904 * cleared and hence a read can be avoided.
3905 */
3906 val64 = 0xFFFFFFFFFFFFFFFFULL;
3907 writeq(val64, &bar0->rx_traffic_int);
20346722 3908 for (i = 0; i < config->rx_ring_num; i++) {
3909 rx_intr_handler(&mac_control->rings[i]);
3910 }
1da177e4
LT
3911 }
3912#endif
3913
20346722 3914 /* If Intr is because of Tx Traffic */
3915 if (reason & GEN_INTR_TXTRAFFIC) {
fe113638 3916 /*
3917 * tx_traffic_int reg is an R1 register, writing all 1's
3918 * will ensure that the actual interrupt causing bit get's
3919 * cleared and hence a read can be avoided.
3920 */
3921 val64 = 0xFFFFFFFFFFFFFFFFULL;
3922 writeq(val64, &bar0->tx_traffic_int);
3923
20346722 3924 for (i = 0; i < config->tx_fifo_num; i++)
3925 tx_intr_handler(&mac_control->fifos[i]);
3926 }
3927
a371a07d 3928 if (reason & GEN_INTR_TXPIC)
3929 s2io_txpic_intr_handle(sp);
20346722 3930 /*
3931 * If the Rx buffer count is below the panic threshold then
3932 * reallocate the buffers from the interrupt handler itself,
1da177e4
LT
3933 * else schedule a tasklet to reallocate the buffers.
3934 */
3935#ifndef CONFIG_S2IO_NAPI
3936 for (i = 0; i < config->rx_ring_num; i++) {
7d3d0439
RA
3937 if (!sp->lro) {
3938 int ret;
3939 int rxb_size = atomic_read(&sp->rx_bufs_left[i]);
3940 int level = rx_buffer_level(sp, rxb_size, i);
3941
3942 if ((level == PANIC) && (!TASKLET_IN_USE)) {
3943 DBG_PRINT(INTR_DBG, "%s: Rx BD hit ",
3944 dev->name);
3945 DBG_PRINT(INTR_DBG, "PANIC levels\n");
3946 if ((ret = fill_rx_buffers(sp, i)) == -ENOMEM) {
3947 DBG_PRINT(ERR_DBG, "%s:Out of memory",
3948 dev->name);
3949 DBG_PRINT(ERR_DBG, " in ISR!!\n");
3950 clear_bit(0, (&sp->tasklet_status));
3951 atomic_dec(&sp->isr_cnt);
3952 return IRQ_HANDLED;
3953 }
1da177e4 3954 clear_bit(0, (&sp->tasklet_status));
7d3d0439
RA
3955 } else if (level == LOW) {
3956 tasklet_schedule(&sp->task);
1da177e4 3957 }
7d3d0439
RA
3958 }
3959 else if (fill_rx_buffers(sp, i) == -ENOMEM) {
3960 DBG_PRINT(ERR_DBG, "%s:Out of memory",
3961 dev->name);
3962 DBG_PRINT(ERR_DBG, " in Rx intr!!\n");
3963 break;
1da177e4
LT
3964 }
3965 }
3966#endif
3967
7ba013ac 3968 atomic_dec(&sp->isr_cnt);
1da177e4
LT
3969 return IRQ_HANDLED;
3970}
3971
7ba013ac 3972/**
3973 * s2io_updt_stats -
3974 */
3975static void s2io_updt_stats(nic_t *sp)
3976{
3977 XENA_dev_config_t __iomem *bar0 = sp->bar0;
3978 u64 val64;
3979 int cnt = 0;
3980
3981 if (atomic_read(&sp->card_state) == CARD_UP) {
3982 /* Apprx 30us on a 133 MHz bus */
3983 val64 = SET_UPDT_CLICKS(10) |
3984 STAT_CFG_ONE_SHOT_EN | STAT_CFG_STAT_EN;
3985 writeq(val64, &bar0->stat_cfg);
3986 do {
3987 udelay(100);
3988 val64 = readq(&bar0->stat_cfg);
3989 if (!(val64 & BIT(0)))
3990 break;
3991 cnt++;
3992 if (cnt == 5)
3993 break; /* Updt failed */
3994 } while(1);
3995 }
3996}
3997
1da177e4 3998/**
20346722 3999 * s2io_get_stats - Updates the device statistics structure.
1da177e4
LT
4000 * @dev : pointer to the device structure.
4001 * Description:
20346722 4002 * This function updates the device statistics structure in the s2io_nic
1da177e4
LT
4003 * structure and returns a pointer to the same.
4004 * Return value:
4005 * pointer to the updated net_device_stats structure.
4006 */
4007
ac1f60db 4008static struct net_device_stats *s2io_get_stats(struct net_device *dev)
1da177e4
LT
4009{
4010 nic_t *sp = dev->priv;
4011 mac_info_t *mac_control;
4012 struct config_param *config;
4013
20346722 4014
1da177e4
LT
4015 mac_control = &sp->mac_control;
4016 config = &sp->config;
4017
7ba013ac 4018 /* Configure Stats for immediate updt */
4019 s2io_updt_stats(sp);
4020
4021 sp->stats.tx_packets =
4022 le32_to_cpu(mac_control->stats_info->tmac_frms);
20346722 4023 sp->stats.tx_errors =
4024 le32_to_cpu(mac_control->stats_info->tmac_any_err_frms);
4025 sp->stats.rx_errors =
4026 le32_to_cpu(mac_control->stats_info->rmac_drop_frms);
4027 sp->stats.multicast =
4028 le32_to_cpu(mac_control->stats_info->rmac_vld_mcst_frms);
1da177e4 4029 sp->stats.rx_length_errors =
20346722 4030 le32_to_cpu(mac_control->stats_info->rmac_long_frms);
1da177e4
LT
4031
4032 return (&sp->stats);
4033}
4034
4035/**
4036 * s2io_set_multicast - entry point for multicast address enable/disable.
4037 * @dev : pointer to the device structure
4038 * Description:
20346722 4039 * This function is a driver entry point which gets called by the kernel
4040 * whenever multicast addresses must be enabled/disabled. This also gets
1da177e4
LT
4041 * called to set/reset promiscuous mode. Depending on the deivce flag, we
4042 * determine, if multicast address must be enabled or if promiscuous mode
4043 * is to be disabled etc.
4044 * Return value:
4045 * void.
4046 */
4047
4048static void s2io_set_multicast(struct net_device *dev)
4049{
4050 int i, j, prev_cnt;
4051 struct dev_mc_list *mclist;
4052 nic_t *sp = dev->priv;
4053 XENA_dev_config_t __iomem *bar0 = sp->bar0;
4054 u64 val64 = 0, multi_mac = 0x010203040506ULL, mask =
4055 0xfeffffffffffULL;
4056 u64 dis_addr = 0xffffffffffffULL, mac_addr = 0;
4057 void __iomem *add;
4058
4059 if ((dev->flags & IFF_ALLMULTI) && (!sp->m_cast_flg)) {
4060 /* Enable all Multicast addresses */
4061 writeq(RMAC_ADDR_DATA0_MEM_ADDR(multi_mac),
4062 &bar0->rmac_addr_data0_mem);
4063 writeq(RMAC_ADDR_DATA1_MEM_MASK(mask),
4064 &bar0->rmac_addr_data1_mem);
4065 val64 = RMAC_ADDR_CMD_MEM_WE |
4066 RMAC_ADDR_CMD_MEM_STROBE_NEW_CMD |
4067 RMAC_ADDR_CMD_MEM_OFFSET(MAC_MC_ALL_MC_ADDR_OFFSET);
4068 writeq(val64, &bar0->rmac_addr_cmd_mem);
4069 /* Wait till command completes */
4070 wait_for_cmd_complete(sp);
4071
4072 sp->m_cast_flg = 1;
4073 sp->all_multi_pos = MAC_MC_ALL_MC_ADDR_OFFSET;
4074 } else if ((dev->flags & IFF_ALLMULTI) && (sp->m_cast_flg)) {
4075 /* Disable all Multicast addresses */
4076 writeq(RMAC_ADDR_DATA0_MEM_ADDR(dis_addr),
4077 &bar0->rmac_addr_data0_mem);
5e25b9dd 4078 writeq(RMAC_ADDR_DATA1_MEM_MASK(0x0),
4079 &bar0->rmac_addr_data1_mem);
1da177e4
LT
4080 val64 = RMAC_ADDR_CMD_MEM_WE |
4081 RMAC_ADDR_CMD_MEM_STROBE_NEW_CMD |
4082 RMAC_ADDR_CMD_MEM_OFFSET(sp->all_multi_pos);
4083 writeq(val64, &bar0->rmac_addr_cmd_mem);
4084 /* Wait till command completes */
4085 wait_for_cmd_complete(sp);
4086
4087 sp->m_cast_flg = 0;
4088 sp->all_multi_pos = 0;
4089 }
4090
4091 if ((dev->flags & IFF_PROMISC) && (!sp->promisc_flg)) {
4092 /* Put the NIC into promiscuous mode */
4093 add = &bar0->mac_cfg;
4094 val64 = readq(&bar0->mac_cfg);
4095 val64 |= MAC_CFG_RMAC_PROM_ENABLE;
4096
4097 writeq(RMAC_CFG_KEY(0x4C0D), &bar0->rmac_cfg_key);
4098 writel((u32) val64, add);
4099 writeq(RMAC_CFG_KEY(0x4C0D), &bar0->rmac_cfg_key);
4100 writel((u32) (val64 >> 32), (add + 4));
4101
4102 val64 = readq(&bar0->mac_cfg);
4103 sp->promisc_flg = 1;
776bd20f 4104 DBG_PRINT(INFO_DBG, "%s: entered promiscuous mode\n",
1da177e4
LT
4105 dev->name);
4106 } else if (!(dev->flags & IFF_PROMISC) && (sp->promisc_flg)) {
4107 /* Remove the NIC from promiscuous mode */
4108 add = &bar0->mac_cfg;
4109 val64 = readq(&bar0->mac_cfg);
4110 val64 &= ~MAC_CFG_RMAC_PROM_ENABLE;
4111
4112 writeq(RMAC_CFG_KEY(0x4C0D), &bar0->rmac_cfg_key);
4113 writel((u32) val64, add);
4114 writeq(RMAC_CFG_KEY(0x4C0D), &bar0->rmac_cfg_key);
4115 writel((u32) (val64 >> 32), (add + 4));
4116
4117 val64 = readq(&bar0->mac_cfg);
4118 sp->promisc_flg = 0;
776bd20f 4119 DBG_PRINT(INFO_DBG, "%s: left promiscuous mode\n",
1da177e4
LT
4120 dev->name);
4121 }
4122
4123 /* Update individual M_CAST address list */
4124 if ((!sp->m_cast_flg) && dev->mc_count) {
4125 if (dev->mc_count >
4126 (MAX_ADDRS_SUPPORTED - MAC_MC_ADDR_START_OFFSET - 1)) {
4127 DBG_PRINT(ERR_DBG, "%s: No more Rx filters ",
4128 dev->name);
4129 DBG_PRINT(ERR_DBG, "can be added, please enable ");
4130 DBG_PRINT(ERR_DBG, "ALL_MULTI instead\n");
4131 return;
4132 }
4133
4134 prev_cnt = sp->mc_addr_count;
4135 sp->mc_addr_count = dev->mc_count;
4136
4137 /* Clear out the previous list of Mc in the H/W. */
4138 for (i = 0; i < prev_cnt; i++) {
4139 writeq(RMAC_ADDR_DATA0_MEM_ADDR(dis_addr),
4140 &bar0->rmac_addr_data0_mem);
4141 writeq(RMAC_ADDR_DATA1_MEM_MASK(0ULL),
20346722 4142 &bar0->rmac_addr_data1_mem);
1da177e4
LT
4143 val64 = RMAC_ADDR_CMD_MEM_WE |
4144 RMAC_ADDR_CMD_MEM_STROBE_NEW_CMD |
4145 RMAC_ADDR_CMD_MEM_OFFSET
4146 (MAC_MC_ADDR_START_OFFSET + i);
4147 writeq(val64, &bar0->rmac_addr_cmd_mem);
4148
4149 /* Wait for command completes */
4150 if (wait_for_cmd_complete(sp)) {
4151 DBG_PRINT(ERR_DBG, "%s: Adding ",
4152 dev->name);
4153 DBG_PRINT(ERR_DBG, "Multicasts failed\n");
4154 return;
4155 }
4156 }
4157
4158 /* Create the new Rx filter list and update the same in H/W. */
4159 for (i = 0, mclist = dev->mc_list; i < dev->mc_count;
4160 i++, mclist = mclist->next) {
4161 memcpy(sp->usr_addrs[i].addr, mclist->dmi_addr,
4162 ETH_ALEN);
a7a80d5a 4163 mac_addr = 0;
1da177e4
LT
4164 for (j = 0; j < ETH_ALEN; j++) {
4165 mac_addr |= mclist->dmi_addr[j];
4166 mac_addr <<= 8;
4167 }
4168 mac_addr >>= 8;
4169 writeq(RMAC_ADDR_DATA0_MEM_ADDR(mac_addr),
4170 &bar0->rmac_addr_data0_mem);
4171 writeq(RMAC_ADDR_DATA1_MEM_MASK(0ULL),
20346722 4172 &bar0->rmac_addr_data1_mem);
1da177e4
LT
4173 val64 = RMAC_ADDR_CMD_MEM_WE |
4174 RMAC_ADDR_CMD_MEM_STROBE_NEW_CMD |
4175 RMAC_ADDR_CMD_MEM_OFFSET
4176 (i + MAC_MC_ADDR_START_OFFSET);
4177 writeq(val64, &bar0->rmac_addr_cmd_mem);
4178
4179 /* Wait for command completes */
4180 if (wait_for_cmd_complete(sp)) {
4181 DBG_PRINT(ERR_DBG, "%s: Adding ",
4182 dev->name);
4183 DBG_PRINT(ERR_DBG, "Multicasts failed\n");
4184 return;
4185 }
4186 }
4187 }
4188}
4189
4190/**
20346722 4191 * s2io_set_mac_addr - Programs the Xframe mac address
1da177e4
LT
4192 * @dev : pointer to the device structure.
4193 * @addr: a uchar pointer to the new mac address which is to be set.
20346722 4194 * Description : This procedure will program the Xframe to receive
1da177e4 4195 * frames with new Mac Address
20346722 4196 * Return value: SUCCESS on success and an appropriate (-)ve integer
1da177e4
LT
4197 * as defined in errno.h file on failure.
4198 */
4199
26df54bf 4200static int s2io_set_mac_addr(struct net_device *dev, u8 * addr)
1da177e4
LT
4201{
4202 nic_t *sp = dev->priv;
4203 XENA_dev_config_t __iomem *bar0 = sp->bar0;
4204 register u64 val64, mac_addr = 0;
4205 int i;
4206
20346722 4207 /*
1da177e4
LT
4208 * Set the new MAC address as the new unicast filter and reflect this
4209 * change on the device address registered with the OS. It will be
20346722 4210 * at offset 0.
1da177e4
LT
4211 */
4212 for (i = 0; i < ETH_ALEN; i++) {
4213 mac_addr <<= 8;
4214 mac_addr |= addr[i];
4215 }
4216
4217 writeq(RMAC_ADDR_DATA0_MEM_ADDR(mac_addr),
4218 &bar0->rmac_addr_data0_mem);
4219
4220 val64 =
4221 RMAC_ADDR_CMD_MEM_WE | RMAC_ADDR_CMD_MEM_STROBE_NEW_CMD |
4222 RMAC_ADDR_CMD_MEM_OFFSET(0);
4223 writeq(val64, &bar0->rmac_addr_cmd_mem);
4224 /* Wait till command completes */
4225 if (wait_for_cmd_complete(sp)) {
4226 DBG_PRINT(ERR_DBG, "%s: set_mac_addr failed\n", dev->name);
4227 return FAILURE;
4228 }
4229
4230 return SUCCESS;
4231}
4232
4233/**
20346722 4234 * s2io_ethtool_sset - Sets different link parameters.
1da177e4
LT
4235 * @sp : private member of the device structure, which is a pointer to the * s2io_nic structure.
4236 * @info: pointer to the structure with parameters given by ethtool to set
4237 * link information.
4238 * Description:
20346722 4239 * The function sets different link parameters provided by the user onto
1da177e4
LT
4240 * the NIC.
4241 * Return value:
4242 * 0 on success.
4243*/
4244
4245static int s2io_ethtool_sset(struct net_device *dev,
4246 struct ethtool_cmd *info)
4247{
4248 nic_t *sp = dev->priv;
4249 if ((info->autoneg == AUTONEG_ENABLE) ||
4250 (info->speed != SPEED_10000) || (info->duplex != DUPLEX_FULL))
4251 return -EINVAL;
4252 else {
4253 s2io_close(sp->dev);
4254 s2io_open(sp->dev);
4255 }
4256
4257 return 0;
4258}
4259
4260/**
20346722 4261 * s2io_ethtol_gset - Return link specific information.
1da177e4
LT
4262 * @sp : private member of the device structure, pointer to the
4263 * s2io_nic structure.
4264 * @info : pointer to the structure with parameters given by ethtool
4265 * to return link information.
4266 * Description:
4267 * Returns link specific information like speed, duplex etc.. to ethtool.
4268 * Return value :
4269 * return 0 on success.
4270 */
4271
4272static int s2io_ethtool_gset(struct net_device *dev, struct ethtool_cmd *info)
4273{
4274 nic_t *sp = dev->priv;
4275 info->supported = (SUPPORTED_10000baseT_Full | SUPPORTED_FIBRE);
4276 info->advertising = (SUPPORTED_10000baseT_Full | SUPPORTED_FIBRE);
4277 info->port = PORT_FIBRE;
4278 /* info->transceiver?? TODO */
4279
4280 if (netif_carrier_ok(sp->dev)) {
4281 info->speed = 10000;
4282 info->duplex = DUPLEX_FULL;
4283 } else {
4284 info->speed = -1;
4285 info->duplex = -1;
4286 }
4287
4288 info->autoneg = AUTONEG_DISABLE;
4289 return 0;
4290}
4291
4292/**
20346722 4293 * s2io_ethtool_gdrvinfo - Returns driver specific information.
4294 * @sp : private member of the device structure, which is a pointer to the
1da177e4
LT
4295 * s2io_nic structure.
4296 * @info : pointer to the structure with parameters given by ethtool to
4297 * return driver information.
4298 * Description:
4299 * Returns driver specefic information like name, version etc.. to ethtool.
4300 * Return value:
4301 * void
4302 */
4303
4304static void s2io_ethtool_gdrvinfo(struct net_device *dev,
4305 struct ethtool_drvinfo *info)
4306{
4307 nic_t *sp = dev->priv;
4308
dbc2309d
JL
4309 strncpy(info->driver, s2io_driver_name, sizeof(info->driver));
4310 strncpy(info->version, s2io_driver_version, sizeof(info->version));
4311 strncpy(info->fw_version, "", sizeof(info->fw_version));
4312 strncpy(info->bus_info, pci_name(sp->pdev), sizeof(info->bus_info));
1da177e4
LT
4313 info->regdump_len = XENA_REG_SPACE;
4314 info->eedump_len = XENA_EEPROM_SPACE;
4315 info->testinfo_len = S2IO_TEST_LEN;
4316 info->n_stats = S2IO_STAT_LEN;
4317}
4318
4319/**
4320 * s2io_ethtool_gregs - dumps the entire space of Xfame into the buffer.
20346722 4321 * @sp: private member of the device structure, which is a pointer to the
1da177e4 4322 * s2io_nic structure.
20346722 4323 * @regs : pointer to the structure with parameters given by ethtool for
1da177e4
LT
4324 * dumping the registers.
4325 * @reg_space: The input argumnet into which all the registers are dumped.
4326 * Description:
4327 * Dumps the entire register space of xFrame NIC into the user given
4328 * buffer area.
4329 * Return value :
4330 * void .
4331*/
4332
4333static void s2io_ethtool_gregs(struct net_device *dev,
4334 struct ethtool_regs *regs, void *space)
4335{
4336 int i;
4337 u64 reg;
4338 u8 *reg_space = (u8 *) space;
4339 nic_t *sp = dev->priv;
4340
4341 regs->len = XENA_REG_SPACE;
4342 regs->version = sp->pdev->subsystem_device;
4343
4344 for (i = 0; i < regs->len; i += 8) {
4345 reg = readq(sp->bar0 + i);
4346 memcpy((reg_space + i), &reg, 8);
4347 }
4348}
4349
4350/**
4351 * s2io_phy_id - timer function that alternates adapter LED.
20346722 4352 * @data : address of the private member of the device structure, which
1da177e4 4353 * is a pointer to the s2io_nic structure, provided as an u32.
20346722 4354 * Description: This is actually the timer function that alternates the
4355 * adapter LED bit of the adapter control bit to set/reset every time on
4356 * invocation. The timer is set for 1/2 a second, hence tha NIC blinks
1da177e4
LT
4357 * once every second.
4358*/
4359static void s2io_phy_id(unsigned long data)
4360{
4361 nic_t *sp = (nic_t *) data;
4362 XENA_dev_config_t __iomem *bar0 = sp->bar0;
4363 u64 val64 = 0;
4364 u16 subid;
4365
4366 subid = sp->pdev->subsystem_device;
541ae68f 4367 if ((sp->device_type == XFRAME_II_DEVICE) ||
4368 ((subid & 0xFF) >= 0x07)) {
1da177e4
LT
4369 val64 = readq(&bar0->gpio_control);
4370 val64 ^= GPIO_CTRL_GPIO_0;
4371 writeq(val64, &bar0->gpio_control);
4372 } else {
4373 val64 = readq(&bar0->adapter_control);
4374 val64 ^= ADAPTER_LED_ON;
4375 writeq(val64, &bar0->adapter_control);
4376 }
4377
4378 mod_timer(&sp->id_timer, jiffies + HZ / 2);
4379}
4380
4381/**
4382 * s2io_ethtool_idnic - To physically identify the nic on the system.
4383 * @sp : private member of the device structure, which is a pointer to the
4384 * s2io_nic structure.
20346722 4385 * @id : pointer to the structure with identification parameters given by
1da177e4
LT
4386 * ethtool.
4387 * Description: Used to physically identify the NIC on the system.
20346722 4388 * The Link LED will blink for a time specified by the user for
1da177e4 4389 * identification.
20346722 4390 * NOTE: The Link has to be Up to be able to blink the LED. Hence
1da177e4
LT
4391 * identification is possible only if it's link is up.
4392 * Return value:
4393 * int , returns 0 on success
4394 */
4395
4396static int s2io_ethtool_idnic(struct net_device *dev, u32 data)
4397{
4398 u64 val64 = 0, last_gpio_ctrl_val;
4399 nic_t *sp = dev->priv;
4400 XENA_dev_config_t __iomem *bar0 = sp->bar0;
4401 u16 subid;
4402
4403 subid = sp->pdev->subsystem_device;
4404 last_gpio_ctrl_val = readq(&bar0->gpio_control);
541ae68f 4405 if ((sp->device_type == XFRAME_I_DEVICE) &&
4406 ((subid & 0xFF) < 0x07)) {
1da177e4
LT
4407 val64 = readq(&bar0->adapter_control);
4408 if (!(val64 & ADAPTER_CNTL_EN)) {
4409 printk(KERN_ERR
4410 "Adapter Link down, cannot blink LED\n");
4411 return -EFAULT;
4412 }
4413 }
4414 if (sp->id_timer.function == NULL) {
4415 init_timer(&sp->id_timer);
4416 sp->id_timer.function = s2io_phy_id;
4417 sp->id_timer.data = (unsigned long) sp;
4418 }
4419 mod_timer(&sp->id_timer, jiffies);
4420 if (data)
20346722 4421 msleep_interruptible(data * HZ);
1da177e4 4422 else
20346722 4423 msleep_interruptible(MAX_FLICKER_TIME);
1da177e4
LT
4424 del_timer_sync(&sp->id_timer);
4425
541ae68f 4426 if (CARDS_WITH_FAULTY_LINK_INDICATORS(sp->device_type, subid)) {
1da177e4
LT
4427 writeq(last_gpio_ctrl_val, &bar0->gpio_control);
4428 last_gpio_ctrl_val = readq(&bar0->gpio_control);
4429 }
4430
4431 return 0;
4432}
4433
4434/**
4435 * s2io_ethtool_getpause_data -Pause frame frame generation and reception.
20346722 4436 * @sp : private member of the device structure, which is a pointer to the
4437 * s2io_nic structure.
1da177e4
LT
4438 * @ep : pointer to the structure with pause parameters given by ethtool.
4439 * Description:
4440 * Returns the Pause frame generation and reception capability of the NIC.
4441 * Return value:
4442 * void
4443 */
4444static void s2io_ethtool_getpause_data(struct net_device *dev,
4445 struct ethtool_pauseparam *ep)
4446{
4447 u64 val64;
4448 nic_t *sp = dev->priv;
4449 XENA_dev_config_t __iomem *bar0 = sp->bar0;
4450
4451 val64 = readq(&bar0->rmac_pause_cfg);
4452 if (val64 & RMAC_PAUSE_GEN_ENABLE)
4453 ep->tx_pause = TRUE;
4454 if (val64 & RMAC_PAUSE_RX_ENABLE)
4455 ep->rx_pause = TRUE;
4456 ep->autoneg = FALSE;
4457}
4458
4459/**
4460 * s2io_ethtool_setpause_data - set/reset pause frame generation.
20346722 4461 * @sp : private member of the device structure, which is a pointer to the
1da177e4
LT
4462 * s2io_nic structure.
4463 * @ep : pointer to the structure with pause parameters given by ethtool.
4464 * Description:
4465 * It can be used to set or reset Pause frame generation or reception
4466 * support of the NIC.
4467 * Return value:
4468 * int, returns 0 on Success
4469 */
4470
4471static int s2io_ethtool_setpause_data(struct net_device *dev,
20346722 4472 struct ethtool_pauseparam *ep)
1da177e4
LT
4473{
4474 u64 val64;
4475 nic_t *sp = dev->priv;
4476 XENA_dev_config_t __iomem *bar0 = sp->bar0;
4477
4478 val64 = readq(&bar0->rmac_pause_cfg);
4479 if (ep->tx_pause)
4480 val64 |= RMAC_PAUSE_GEN_ENABLE;
4481 else
4482 val64 &= ~RMAC_PAUSE_GEN_ENABLE;
4483 if (ep->rx_pause)
4484 val64 |= RMAC_PAUSE_RX_ENABLE;
4485 else
4486 val64 &= ~RMAC_PAUSE_RX_ENABLE;
4487 writeq(val64, &bar0->rmac_pause_cfg);
4488 return 0;
4489}
4490
4491/**
4492 * read_eeprom - reads 4 bytes of data from user given offset.
20346722 4493 * @sp : private member of the device structure, which is a pointer to the
1da177e4
LT
4494 * s2io_nic structure.
4495 * @off : offset at which the data must be written
4496 * @data : Its an output parameter where the data read at the given
20346722 4497 * offset is stored.
1da177e4 4498 * Description:
20346722 4499 * Will read 4 bytes of data from the user given offset and return the
1da177e4
LT
4500 * read data.
4501 * NOTE: Will allow to read only part of the EEPROM visible through the
4502 * I2C bus.
4503 * Return value:
4504 * -1 on failure and 0 on success.
4505 */
4506
4507#define S2IO_DEV_ID 5
ad4ebed0 4508static int read_eeprom(nic_t * sp, int off, u64 * data)
1da177e4
LT
4509{
4510 int ret = -1;
4511 u32 exit_cnt = 0;
4512 u64 val64;
4513 XENA_dev_config_t __iomem *bar0 = sp->bar0;
4514
ad4ebed0 4515 if (sp->device_type == XFRAME_I_DEVICE) {
4516 val64 = I2C_CONTROL_DEV_ID(S2IO_DEV_ID) | I2C_CONTROL_ADDR(off) |
4517 I2C_CONTROL_BYTE_CNT(0x3) | I2C_CONTROL_READ |
4518 I2C_CONTROL_CNTL_START;
4519 SPECIAL_REG_WRITE(val64, &bar0->i2c_control, LF);
1da177e4 4520
ad4ebed0 4521 while (exit_cnt < 5) {
4522 val64 = readq(&bar0->i2c_control);
4523 if (I2C_CONTROL_CNTL_END(val64)) {
4524 *data = I2C_CONTROL_GET_DATA(val64);
4525 ret = 0;
4526 break;
4527 }
4528 msleep(50);
4529 exit_cnt++;
1da177e4 4530 }
1da177e4
LT
4531 }
4532
ad4ebed0 4533 if (sp->device_type == XFRAME_II_DEVICE) {
4534 val64 = SPI_CONTROL_KEY(0x9) | SPI_CONTROL_SEL1 |
4535 SPI_CONTROL_BYTECNT(0x3) |
4536 SPI_CONTROL_CMD(0x3) | SPI_CONTROL_ADDR(off);
4537 SPECIAL_REG_WRITE(val64, &bar0->spi_control, LF);
4538 val64 |= SPI_CONTROL_REQ;
4539 SPECIAL_REG_WRITE(val64, &bar0->spi_control, LF);
4540 while (exit_cnt < 5) {
4541 val64 = readq(&bar0->spi_control);
4542 if (val64 & SPI_CONTROL_NACK) {
4543 ret = 1;
4544 break;
4545 } else if (val64 & SPI_CONTROL_DONE) {
4546 *data = readq(&bar0->spi_data);
4547 *data &= 0xffffff;
4548 ret = 0;
4549 break;
4550 }
4551 msleep(50);
4552 exit_cnt++;
4553 }
4554 }
1da177e4
LT
4555 return ret;
4556}
4557
4558/**
4559 * write_eeprom - actually writes the relevant part of the data value.
4560 * @sp : private member of the device structure, which is a pointer to the
4561 * s2io_nic structure.
4562 * @off : offset at which the data must be written
4563 * @data : The data that is to be written
20346722 4564 * @cnt : Number of bytes of the data that are actually to be written into
1da177e4
LT
4565 * the Eeprom. (max of 3)
4566 * Description:
4567 * Actually writes the relevant part of the data value into the Eeprom
4568 * through the I2C bus.
4569 * Return value:
4570 * 0 on success, -1 on failure.
4571 */
4572
ad4ebed0 4573static int write_eeprom(nic_t * sp, int off, u64 data, int cnt)
1da177e4
LT
4574{
4575 int exit_cnt = 0, ret = -1;
4576 u64 val64;
4577 XENA_dev_config_t __iomem *bar0 = sp->bar0;
4578
ad4ebed0 4579 if (sp->device_type == XFRAME_I_DEVICE) {
4580 val64 = I2C_CONTROL_DEV_ID(S2IO_DEV_ID) | I2C_CONTROL_ADDR(off) |
4581 I2C_CONTROL_BYTE_CNT(cnt) | I2C_CONTROL_SET_DATA((u32)data) |
4582 I2C_CONTROL_CNTL_START;
4583 SPECIAL_REG_WRITE(val64, &bar0->i2c_control, LF);
4584
4585 while (exit_cnt < 5) {
4586 val64 = readq(&bar0->i2c_control);
4587 if (I2C_CONTROL_CNTL_END(val64)) {
4588 if (!(val64 & I2C_CONTROL_NACK))
4589 ret = 0;
4590 break;
4591 }
4592 msleep(50);
4593 exit_cnt++;
4594 }
4595 }
1da177e4 4596
ad4ebed0 4597 if (sp->device_type == XFRAME_II_DEVICE) {
4598 int write_cnt = (cnt == 8) ? 0 : cnt;
4599 writeq(SPI_DATA_WRITE(data,(cnt<<3)), &bar0->spi_data);
4600
4601 val64 = SPI_CONTROL_KEY(0x9) | SPI_CONTROL_SEL1 |
4602 SPI_CONTROL_BYTECNT(write_cnt) |
4603 SPI_CONTROL_CMD(0x2) | SPI_CONTROL_ADDR(off);
4604 SPECIAL_REG_WRITE(val64, &bar0->spi_control, LF);
4605 val64 |= SPI_CONTROL_REQ;
4606 SPECIAL_REG_WRITE(val64, &bar0->spi_control, LF);
4607 while (exit_cnt < 5) {
4608 val64 = readq(&bar0->spi_control);
4609 if (val64 & SPI_CONTROL_NACK) {
4610 ret = 1;
4611 break;
4612 } else if (val64 & SPI_CONTROL_DONE) {
1da177e4 4613 ret = 0;
ad4ebed0 4614 break;
4615 }
4616 msleep(50);
4617 exit_cnt++;
1da177e4 4618 }
1da177e4 4619 }
1da177e4
LT
4620 return ret;
4621}
4622
4623/**
4624 * s2io_ethtool_geeprom - reads the value stored in the Eeprom.
4625 * @sp : private member of the device structure, which is a pointer to the * s2io_nic structure.
20346722 4626 * @eeprom : pointer to the user level structure provided by ethtool,
1da177e4
LT
4627 * containing all relevant information.
4628 * @data_buf : user defined value to be written into Eeprom.
4629 * Description: Reads the values stored in the Eeprom at given offset
4630 * for a given length. Stores these values int the input argument data
4631 * buffer 'data_buf' and returns these to the caller (ethtool.)
4632 * Return value:
4633 * int 0 on success
4634 */
4635
4636static int s2io_ethtool_geeprom(struct net_device *dev,
20346722 4637 struct ethtool_eeprom *eeprom, u8 * data_buf)
1da177e4 4638{
ad4ebed0 4639 u32 i, valid;
4640 u64 data;
1da177e4
LT
4641 nic_t *sp = dev->priv;
4642
4643 eeprom->magic = sp->pdev->vendor | (sp->pdev->device << 16);
4644
4645 if ((eeprom->offset + eeprom->len) > (XENA_EEPROM_SPACE))
4646 eeprom->len = XENA_EEPROM_SPACE - eeprom->offset;
4647
4648 for (i = 0; i < eeprom->len; i += 4) {
4649 if (read_eeprom(sp, (eeprom->offset + i), &data)) {
4650 DBG_PRINT(ERR_DBG, "Read of EEPROM failed\n");
4651 return -EFAULT;
4652 }
4653 valid = INV(data);
4654 memcpy((data_buf + i), &valid, 4);
4655 }
4656 return 0;
4657}
4658
4659/**
4660 * s2io_ethtool_seeprom - tries to write the user provided value in Eeprom
4661 * @sp : private member of the device structure, which is a pointer to the
4662 * s2io_nic structure.
20346722 4663 * @eeprom : pointer to the user level structure provided by ethtool,
1da177e4
LT
4664 * containing all relevant information.
4665 * @data_buf ; user defined value to be written into Eeprom.
4666 * Description:
4667 * Tries to write the user provided value in the Eeprom, at the offset
4668 * given by the user.
4669 * Return value:
4670 * 0 on success, -EFAULT on failure.
4671 */
4672
4673static int s2io_ethtool_seeprom(struct net_device *dev,
4674 struct ethtool_eeprom *eeprom,
4675 u8 * data_buf)
4676{
4677 int len = eeprom->len, cnt = 0;
ad4ebed0 4678 u64 valid = 0, data;
1da177e4
LT
4679 nic_t *sp = dev->priv;
4680
4681 if (eeprom->magic != (sp->pdev->vendor | (sp->pdev->device << 16))) {
4682 DBG_PRINT(ERR_DBG,
4683 "ETHTOOL_WRITE_EEPROM Err: Magic value ");
4684 DBG_PRINT(ERR_DBG, "is wrong, Its not 0x%x\n",
4685 eeprom->magic);
4686 return -EFAULT;
4687 }
4688
4689 while (len) {
4690 data = (u32) data_buf[cnt] & 0x000000FF;
4691 if (data) {
4692 valid = (u32) (data << 24);
4693 } else
4694 valid = data;
4695
4696 if (write_eeprom(sp, (eeprom->offset + cnt), valid, 0)) {
4697 DBG_PRINT(ERR_DBG,
4698 "ETHTOOL_WRITE_EEPROM Err: Cannot ");
4699 DBG_PRINT(ERR_DBG,
4700 "write into the specified offset\n");
4701 return -EFAULT;
4702 }
4703 cnt++;
4704 len--;
4705 }
4706
4707 return 0;
4708}
4709
4710/**
20346722 4711 * s2io_register_test - reads and writes into all clock domains.
4712 * @sp : private member of the device structure, which is a pointer to the
1da177e4
LT
4713 * s2io_nic structure.
4714 * @data : variable that returns the result of each of the test conducted b
4715 * by the driver.
4716 * Description:
4717 * Read and write into all clock domains. The NIC has 3 clock domains,
4718 * see that registers in all the three regions are accessible.
4719 * Return value:
4720 * 0 on success.
4721 */
4722
4723static int s2io_register_test(nic_t * sp, uint64_t * data)
4724{
4725 XENA_dev_config_t __iomem *bar0 = sp->bar0;
ad4ebed0 4726 u64 val64 = 0, exp_val;
1da177e4
LT
4727 int fail = 0;
4728
20346722 4729 val64 = readq(&bar0->pif_rd_swapper_fb);
4730 if (val64 != 0x123456789abcdefULL) {
1da177e4
LT
4731 fail = 1;
4732 DBG_PRINT(INFO_DBG, "Read Test level 1 fails\n");
4733 }
4734
4735 val64 = readq(&bar0->rmac_pause_cfg);
4736 if (val64 != 0xc000ffff00000000ULL) {
4737 fail = 1;
4738 DBG_PRINT(INFO_DBG, "Read Test level 2 fails\n");
4739 }
4740
4741 val64 = readq(&bar0->rx_queue_cfg);
ad4ebed0 4742 if (sp->device_type == XFRAME_II_DEVICE)
4743 exp_val = 0x0404040404040404ULL;
4744 else
4745 exp_val = 0x0808080808080808ULL;
4746 if (val64 != exp_val) {
1da177e4
LT
4747 fail = 1;
4748 DBG_PRINT(INFO_DBG, "Read Test level 3 fails\n");
4749 }
4750
4751 val64 = readq(&bar0->xgxs_efifo_cfg);
4752 if (val64 != 0x000000001923141EULL) {
4753 fail = 1;
4754 DBG_PRINT(INFO_DBG, "Read Test level 4 fails\n");
4755 }
4756
4757 val64 = 0x5A5A5A5A5A5A5A5AULL;
4758 writeq(val64, &bar0->xmsi_data);
4759 val64 = readq(&bar0->xmsi_data);
4760 if (val64 != 0x5A5A5A5A5A5A5A5AULL) {
4761 fail = 1;
4762 DBG_PRINT(ERR_DBG, "Write Test level 1 fails\n");
4763 }
4764
4765 val64 = 0xA5A5A5A5A5A5A5A5ULL;
4766 writeq(val64, &bar0->xmsi_data);
4767 val64 = readq(&bar0->xmsi_data);
4768 if (val64 != 0xA5A5A5A5A5A5A5A5ULL) {
4769 fail = 1;
4770 DBG_PRINT(ERR_DBG, "Write Test level 2 fails\n");
4771 }
4772
4773 *data = fail;
ad4ebed0 4774 return fail;
1da177e4
LT
4775}
4776
4777/**
20346722 4778 * s2io_eeprom_test - to verify that EEprom in the xena can be programmed.
1da177e4
LT
4779 * @sp : private member of the device structure, which is a pointer to the
4780 * s2io_nic structure.
4781 * @data:variable that returns the result of each of the test conducted by
4782 * the driver.
4783 * Description:
20346722 4784 * Verify that EEPROM in the xena can be programmed using I2C_CONTROL
1da177e4
LT
4785 * register.
4786 * Return value:
4787 * 0 on success.
4788 */
4789
4790static int s2io_eeprom_test(nic_t * sp, uint64_t * data)
4791{
4792 int fail = 0;
ad4ebed0 4793 u64 ret_data, org_4F0, org_7F0;
4794 u8 saved_4F0 = 0, saved_7F0 = 0;
4795 struct net_device *dev = sp->dev;
1da177e4
LT
4796
4797 /* Test Write Error at offset 0 */
ad4ebed0 4798 /* Note that SPI interface allows write access to all areas
4799 * of EEPROM. Hence doing all negative testing only for Xframe I.
4800 */
4801 if (sp->device_type == XFRAME_I_DEVICE)
4802 if (!write_eeprom(sp, 0, 0, 3))
4803 fail = 1;
4804
4805 /* Save current values at offsets 0x4F0 and 0x7F0 */
4806 if (!read_eeprom(sp, 0x4F0, &org_4F0))
4807 saved_4F0 = 1;
4808 if (!read_eeprom(sp, 0x7F0, &org_7F0))
4809 saved_7F0 = 1;
1da177e4
LT
4810
4811 /* Test Write at offset 4f0 */
ad4ebed0 4812 if (write_eeprom(sp, 0x4F0, 0x012345, 3))
1da177e4
LT
4813 fail = 1;
4814 if (read_eeprom(sp, 0x4F0, &ret_data))
4815 fail = 1;
4816
ad4ebed0 4817 if (ret_data != 0x012345) {
26b7625c
AM
4818 DBG_PRINT(ERR_DBG, "%s: eeprom test error at offset 0x4F0. "
4819 "Data written %llx Data read %llx\n",
4820 dev->name, (unsigned long long)0x12345,
4821 (unsigned long long)ret_data);
1da177e4 4822 fail = 1;
ad4ebed0 4823 }
1da177e4
LT
4824
4825 /* Reset the EEPROM data go FFFF */
ad4ebed0 4826 write_eeprom(sp, 0x4F0, 0xFFFFFF, 3);
1da177e4
LT
4827
4828 /* Test Write Request Error at offset 0x7c */
ad4ebed0 4829 if (sp->device_type == XFRAME_I_DEVICE)
4830 if (!write_eeprom(sp, 0x07C, 0, 3))
4831 fail = 1;
1da177e4 4832
ad4ebed0 4833 /* Test Write Request at offset 0x7f0 */
4834 if (write_eeprom(sp, 0x7F0, 0x012345, 3))
1da177e4 4835 fail = 1;
ad4ebed0 4836 if (read_eeprom(sp, 0x7F0, &ret_data))
1da177e4
LT
4837 fail = 1;
4838
ad4ebed0 4839 if (ret_data != 0x012345) {
26b7625c
AM
4840 DBG_PRINT(ERR_DBG, "%s: eeprom test error at offset 0x7F0. "
4841 "Data written %llx Data read %llx\n",
4842 dev->name, (unsigned long long)0x12345,
4843 (unsigned long long)ret_data);
1da177e4 4844 fail = 1;
ad4ebed0 4845 }
1da177e4
LT
4846
4847 /* Reset the EEPROM data go FFFF */
ad4ebed0 4848 write_eeprom(sp, 0x7F0, 0xFFFFFF, 3);
1da177e4 4849
ad4ebed0 4850 if (sp->device_type == XFRAME_I_DEVICE) {
4851 /* Test Write Error at offset 0x80 */
4852 if (!write_eeprom(sp, 0x080, 0, 3))
4853 fail = 1;
1da177e4 4854
ad4ebed0 4855 /* Test Write Error at offset 0xfc */
4856 if (!write_eeprom(sp, 0x0FC, 0, 3))
4857 fail = 1;
1da177e4 4858
ad4ebed0 4859 /* Test Write Error at offset 0x100 */
4860 if (!write_eeprom(sp, 0x100, 0, 3))
4861 fail = 1;
1da177e4 4862
ad4ebed0 4863 /* Test Write Error at offset 4ec */
4864 if (!write_eeprom(sp, 0x4EC, 0, 3))
4865 fail = 1;
4866 }
4867
4868 /* Restore values at offsets 0x4F0 and 0x7F0 */
4869 if (saved_4F0)
4870 write_eeprom(sp, 0x4F0, org_4F0, 3);
4871 if (saved_7F0)
4872 write_eeprom(sp, 0x7F0, org_7F0, 3);
1da177e4
LT
4873
4874 *data = fail;
ad4ebed0 4875 return fail;
1da177e4
LT
4876}
4877
4878/**
4879 * s2io_bist_test - invokes the MemBist test of the card .
20346722 4880 * @sp : private member of the device structure, which is a pointer to the
1da177e4 4881 * s2io_nic structure.
20346722 4882 * @data:variable that returns the result of each of the test conducted by
1da177e4
LT
4883 * the driver.
4884 * Description:
4885 * This invokes the MemBist test of the card. We give around
4886 * 2 secs time for the Test to complete. If it's still not complete
20346722 4887 * within this peiod, we consider that the test failed.
1da177e4
LT
4888 * Return value:
4889 * 0 on success and -1 on failure.
4890 */
4891
4892static int s2io_bist_test(nic_t * sp, uint64_t * data)
4893{
4894 u8 bist = 0;
4895 int cnt = 0, ret = -1;
4896
4897 pci_read_config_byte(sp->pdev, PCI_BIST, &bist);
4898 bist |= PCI_BIST_START;
4899 pci_write_config_word(sp->pdev, PCI_BIST, bist);
4900
4901 while (cnt < 20) {
4902 pci_read_config_byte(sp->pdev, PCI_BIST, &bist);
4903 if (!(bist & PCI_BIST_START)) {
4904 *data = (bist & PCI_BIST_CODE_MASK);
4905 ret = 0;
4906 break;
4907 }
4908 msleep(100);
4909 cnt++;
4910 }
4911
4912 return ret;
4913}
4914
4915/**
20346722 4916 * s2io-link_test - verifies the link state of the nic
4917 * @sp ; private member of the device structure, which is a pointer to the
1da177e4
LT
4918 * s2io_nic structure.
4919 * @data: variable that returns the result of each of the test conducted by
4920 * the driver.
4921 * Description:
20346722 4922 * The function verifies the link state of the NIC and updates the input
1da177e4
LT
4923 * argument 'data' appropriately.
4924 * Return value:
4925 * 0 on success.
4926 */
4927
4928static int s2io_link_test(nic_t * sp, uint64_t * data)
4929{
4930 XENA_dev_config_t __iomem *bar0 = sp->bar0;
4931 u64 val64;
4932
4933 val64 = readq(&bar0->adapter_status);
4934 if (val64 & ADAPTER_STATUS_RMAC_LOCAL_FAULT)
4935 *data = 1;
4936
4937 return 0;
4938}
4939
4940/**
20346722 4941 * s2io_rldram_test - offline test for access to the RldRam chip on the NIC
4942 * @sp - private member of the device structure, which is a pointer to the
1da177e4 4943 * s2io_nic structure.
20346722 4944 * @data - variable that returns the result of each of the test
1da177e4
LT
4945 * conducted by the driver.
4946 * Description:
20346722 4947 * This is one of the offline test that tests the read and write
1da177e4
LT
4948 * access to the RldRam chip on the NIC.
4949 * Return value:
4950 * 0 on success.
4951 */
4952
4953static int s2io_rldram_test(nic_t * sp, uint64_t * data)
4954{
4955 XENA_dev_config_t __iomem *bar0 = sp->bar0;
4956 u64 val64;
ad4ebed0 4957 int cnt, iteration = 0, test_fail = 0;
1da177e4
LT
4958
4959 val64 = readq(&bar0->adapter_control);
4960 val64 &= ~ADAPTER_ECC_EN;
4961 writeq(val64, &bar0->adapter_control);
4962
4963 val64 = readq(&bar0->mc_rldram_test_ctrl);
4964 val64 |= MC_RLDRAM_TEST_MODE;
ad4ebed0 4965 SPECIAL_REG_WRITE(val64, &bar0->mc_rldram_test_ctrl, LF);
1da177e4
LT
4966
4967 val64 = readq(&bar0->mc_rldram_mrs);
4968 val64 |= MC_RLDRAM_QUEUE_SIZE_ENABLE;
4969 SPECIAL_REG_WRITE(val64, &bar0->mc_rldram_mrs, UF);
4970
4971 val64 |= MC_RLDRAM_MRS_ENABLE;
4972 SPECIAL_REG_WRITE(val64, &bar0->mc_rldram_mrs, UF);
4973
4974 while (iteration < 2) {
4975 val64 = 0x55555555aaaa0000ULL;
4976 if (iteration == 1) {
4977 val64 ^= 0xFFFFFFFFFFFF0000ULL;
4978 }
4979 writeq(val64, &bar0->mc_rldram_test_d0);
4980
4981 val64 = 0xaaaa5a5555550000ULL;
4982 if (iteration == 1) {
4983 val64 ^= 0xFFFFFFFFFFFF0000ULL;
4984 }
4985 writeq(val64, &bar0->mc_rldram_test_d1);
4986
4987 val64 = 0x55aaaaaaaa5a0000ULL;
4988 if (iteration == 1) {
4989 val64 ^= 0xFFFFFFFFFFFF0000ULL;
4990 }
4991 writeq(val64, &bar0->mc_rldram_test_d2);
4992
ad4ebed0 4993 val64 = (u64) (0x0000003ffffe0100ULL);
1da177e4
LT
4994 writeq(val64, &bar0->mc_rldram_test_add);
4995
ad4ebed0 4996 val64 = MC_RLDRAM_TEST_MODE | MC_RLDRAM_TEST_WRITE |
4997 MC_RLDRAM_TEST_GO;
4998 SPECIAL_REG_WRITE(val64, &bar0->mc_rldram_test_ctrl, LF);
1da177e4
LT
4999
5000 for (cnt = 0; cnt < 5; cnt++) {
5001 val64 = readq(&bar0->mc_rldram_test_ctrl);
5002 if (val64 & MC_RLDRAM_TEST_DONE)
5003 break;
5004 msleep(200);
5005 }
5006
5007 if (cnt == 5)
5008 break;
5009
ad4ebed0 5010 val64 = MC_RLDRAM_TEST_MODE | MC_RLDRAM_TEST_GO;
5011 SPECIAL_REG_WRITE(val64, &bar0->mc_rldram_test_ctrl, LF);
1da177e4
LT
5012
5013 for (cnt = 0; cnt < 5; cnt++) {
5014 val64 = readq(&bar0->mc_rldram_test_ctrl);
5015 if (val64 & MC_RLDRAM_TEST_DONE)
5016 break;
5017 msleep(500);
5018 }
5019
5020 if (cnt == 5)
5021 break;
5022
5023 val64 = readq(&bar0->mc_rldram_test_ctrl);
ad4ebed0 5024 if (!(val64 & MC_RLDRAM_TEST_PASS))
5025 test_fail = 1;
1da177e4
LT
5026
5027 iteration++;
5028 }
5029
ad4ebed0 5030 *data = test_fail;
1da177e4 5031
ad4ebed0 5032 /* Bring the adapter out of test mode */
5033 SPECIAL_REG_WRITE(0, &bar0->mc_rldram_test_ctrl, LF);
5034
5035 return test_fail;
1da177e4
LT
5036}
5037
5038/**
5039 * s2io_ethtool_test - conducts 6 tsets to determine the health of card.
5040 * @sp : private member of the device structure, which is a pointer to the
5041 * s2io_nic structure.
5042 * @ethtest : pointer to a ethtool command specific structure that will be
5043 * returned to the user.
20346722 5044 * @data : variable that returns the result of each of the test
1da177e4
LT
5045 * conducted by the driver.
5046 * Description:
5047 * This function conducts 6 tests ( 4 offline and 2 online) to determine
5048 * the health of the card.
5049 * Return value:
5050 * void
5051 */
5052
5053static void s2io_ethtool_test(struct net_device *dev,
5054 struct ethtool_test *ethtest,
5055 uint64_t * data)
5056{
5057 nic_t *sp = dev->priv;
5058 int orig_state = netif_running(sp->dev);
5059
5060 if (ethtest->flags == ETH_TEST_FL_OFFLINE) {
5061 /* Offline Tests. */
20346722 5062 if (orig_state)
1da177e4 5063 s2io_close(sp->dev);
1da177e4
LT
5064
5065 if (s2io_register_test(sp, &data[0]))
5066 ethtest->flags |= ETH_TEST_FL_FAILED;
5067
5068 s2io_reset(sp);
1da177e4
LT
5069
5070 if (s2io_rldram_test(sp, &data[3]))
5071 ethtest->flags |= ETH_TEST_FL_FAILED;
5072
5073 s2io_reset(sp);
1da177e4
LT
5074
5075 if (s2io_eeprom_test(sp, &data[1]))
5076 ethtest->flags |= ETH_TEST_FL_FAILED;
5077
5078 if (s2io_bist_test(sp, &data[4]))
5079 ethtest->flags |= ETH_TEST_FL_FAILED;
5080
5081 if (orig_state)
5082 s2io_open(sp->dev);
5083
5084 data[2] = 0;
5085 } else {
5086 /* Online Tests. */
5087 if (!orig_state) {
5088 DBG_PRINT(ERR_DBG,
5089 "%s: is not up, cannot run test\n",
5090 dev->name);
5091 data[0] = -1;
5092 data[1] = -1;
5093 data[2] = -1;
5094 data[3] = -1;
5095 data[4] = -1;
5096 }
5097
5098 if (s2io_link_test(sp, &data[2]))
5099 ethtest->flags |= ETH_TEST_FL_FAILED;
5100
5101 data[0] = 0;
5102 data[1] = 0;
5103 data[3] = 0;
5104 data[4] = 0;
5105 }
5106}
5107
5108static void s2io_get_ethtool_stats(struct net_device *dev,
5109 struct ethtool_stats *estats,
5110 u64 * tmp_stats)
5111{
5112 int i = 0;
5113 nic_t *sp = dev->priv;
5114 StatInfo_t *stat_info = sp->mac_control.stats_info;
fe931395 5115 u64 tmp;
1da177e4 5116
7ba013ac 5117 s2io_updt_stats(sp);
541ae68f 5118 tmp_stats[i++] =
5119 (u64)le32_to_cpu(stat_info->tmac_frms_oflow) << 32 |
5120 le32_to_cpu(stat_info->tmac_frms);
5121 tmp_stats[i++] =
5122 (u64)le32_to_cpu(stat_info->tmac_data_octets_oflow) << 32 |
5123 le32_to_cpu(stat_info->tmac_data_octets);
1da177e4 5124 tmp_stats[i++] = le64_to_cpu(stat_info->tmac_drop_frms);
541ae68f 5125 tmp_stats[i++] =
5126 (u64)le32_to_cpu(stat_info->tmac_mcst_frms_oflow) << 32 |
5127 le32_to_cpu(stat_info->tmac_mcst_frms);
5128 tmp_stats[i++] =
5129 (u64)le32_to_cpu(stat_info->tmac_bcst_frms_oflow) << 32 |
5130 le32_to_cpu(stat_info->tmac_bcst_frms);
1da177e4 5131 tmp_stats[i++] = le64_to_cpu(stat_info->tmac_pause_ctrl_frms);
541ae68f 5132 tmp_stats[i++] =
5133 (u64)le32_to_cpu(stat_info->tmac_any_err_frms_oflow) << 32 |
5134 le32_to_cpu(stat_info->tmac_any_err_frms);
1da177e4 5135 tmp_stats[i++] = le64_to_cpu(stat_info->tmac_vld_ip_octets);
541ae68f 5136 tmp_stats[i++] =
5137 (u64)le32_to_cpu(stat_info->tmac_vld_ip_oflow) << 32 |
5138 le32_to_cpu(stat_info->tmac_vld_ip);
5139 tmp_stats[i++] =
5140 (u64)le32_to_cpu(stat_info->tmac_drop_ip_oflow) << 32 |
5141 le32_to_cpu(stat_info->tmac_drop_ip);
5142 tmp_stats[i++] =
5143 (u64)le32_to_cpu(stat_info->tmac_icmp_oflow) << 32 |
5144 le32_to_cpu(stat_info->tmac_icmp);
5145 tmp_stats[i++] =
5146 (u64)le32_to_cpu(stat_info->tmac_rst_tcp_oflow) << 32 |
5147 le32_to_cpu(stat_info->tmac_rst_tcp);
1da177e4 5148 tmp_stats[i++] = le64_to_cpu(stat_info->tmac_tcp);
541ae68f 5149 tmp_stats[i++] = (u64)le32_to_cpu(stat_info->tmac_udp_oflow) << 32 |
5150 le32_to_cpu(stat_info->tmac_udp);
5151 tmp_stats[i++] =
5152 (u64)le32_to_cpu(stat_info->rmac_vld_frms_oflow) << 32 |
5153 le32_to_cpu(stat_info->rmac_vld_frms);
5154 tmp_stats[i++] =
5155 (u64)le32_to_cpu(stat_info->rmac_data_octets_oflow) << 32 |
5156 le32_to_cpu(stat_info->rmac_data_octets);
1da177e4
LT
5157 tmp_stats[i++] = le64_to_cpu(stat_info->rmac_fcs_err_frms);
5158 tmp_stats[i++] = le64_to_cpu(stat_info->rmac_drop_frms);
541ae68f 5159 tmp_stats[i++] =
5160 (u64)le32_to_cpu(stat_info->rmac_vld_mcst_frms_oflow) << 32 |
5161 le32_to_cpu(stat_info->rmac_vld_mcst_frms);
5162 tmp_stats[i++] =
5163 (u64)le32_to_cpu(stat_info->rmac_vld_bcst_frms_oflow) << 32 |
5164 le32_to_cpu(stat_info->rmac_vld_bcst_frms);
1da177e4
LT
5165 tmp_stats[i++] = le32_to_cpu(stat_info->rmac_in_rng_len_err_frms);
5166 tmp_stats[i++] = le64_to_cpu(stat_info->rmac_long_frms);
5167 tmp_stats[i++] = le64_to_cpu(stat_info->rmac_pause_ctrl_frms);
541ae68f 5168 tmp_stats[i++] =
5169 (u64)le32_to_cpu(stat_info->rmac_discarded_frms_oflow) << 32 |
5170 le32_to_cpu(stat_info->rmac_discarded_frms);
5171 tmp_stats[i++] =
5172 (u64)le32_to_cpu(stat_info->rmac_usized_frms_oflow) << 32 |
5173 le32_to_cpu(stat_info->rmac_usized_frms);
5174 tmp_stats[i++] =
5175 (u64)le32_to_cpu(stat_info->rmac_osized_frms_oflow) << 32 |
5176 le32_to_cpu(stat_info->rmac_osized_frms);
5177 tmp_stats[i++] =
5178 (u64)le32_to_cpu(stat_info->rmac_frag_frms_oflow) << 32 |
5179 le32_to_cpu(stat_info->rmac_frag_frms);
5180 tmp_stats[i++] =
5181 (u64)le32_to_cpu(stat_info->rmac_jabber_frms_oflow) << 32 |
5182 le32_to_cpu(stat_info->rmac_jabber_frms);
5183 tmp_stats[i++] = (u64)le32_to_cpu(stat_info->rmac_ip_oflow) << 32 |
5184 le32_to_cpu(stat_info->rmac_ip);
1da177e4
LT
5185 tmp_stats[i++] = le64_to_cpu(stat_info->rmac_ip_octets);
5186 tmp_stats[i++] = le32_to_cpu(stat_info->rmac_hdr_err_ip);
541ae68f 5187 tmp_stats[i++] = (u64)le32_to_cpu(stat_info->rmac_drop_ip_oflow) << 32 |
5188 le32_to_cpu(stat_info->rmac_drop_ip);
5189 tmp_stats[i++] = (u64)le32_to_cpu(stat_info->rmac_icmp_oflow) << 32 |
5190 le32_to_cpu(stat_info->rmac_icmp);
1da177e4 5191 tmp_stats[i++] = le64_to_cpu(stat_info->rmac_tcp);
541ae68f 5192 tmp_stats[i++] = (u64)le32_to_cpu(stat_info->rmac_udp_oflow) << 32 |
5193 le32_to_cpu(stat_info->rmac_udp);
5194 tmp_stats[i++] =
5195 (u64)le32_to_cpu(stat_info->rmac_err_drp_udp_oflow) << 32 |
5196 le32_to_cpu(stat_info->rmac_err_drp_udp);
5197 tmp_stats[i++] =
5198 (u64)le32_to_cpu(stat_info->rmac_pause_cnt_oflow) << 32 |
5199 le32_to_cpu(stat_info->rmac_pause_cnt);
5200 tmp_stats[i++] =
5201 (u64)le32_to_cpu(stat_info->rmac_accepted_ip_oflow) << 32 |
5202 le32_to_cpu(stat_info->rmac_accepted_ip);
1da177e4 5203 tmp_stats[i++] = le32_to_cpu(stat_info->rmac_err_tcp);
7ba013ac 5204 tmp_stats[i++] = 0;
5205 tmp_stats[i++] = stat_info->sw_stat.single_ecc_errs;
5206 tmp_stats[i++] = stat_info->sw_stat.double_ecc_errs;
7d3d0439
RA
5207 tmp_stats[i++] = stat_info->sw_stat.clubbed_frms_cnt;
5208 tmp_stats[i++] = stat_info->sw_stat.sending_both;
5209 tmp_stats[i++] = stat_info->sw_stat.outof_sequence_pkts;
5210 tmp_stats[i++] = stat_info->sw_stat.flush_max_pkts;
fe931395
AM
5211 tmp = 0;
5212 if (stat_info->sw_stat.num_aggregations) {
5213 tmp = stat_info->sw_stat.sum_avg_pkts_aggregated;
5214 do_div(tmp, stat_info->sw_stat.num_aggregations);
5215 }
5216 tmp_stats[i++] = tmp;
1da177e4
LT
5217}
5218
ac1f60db 5219static int s2io_ethtool_get_regs_len(struct net_device *dev)
1da177e4
LT
5220{
5221 return (XENA_REG_SPACE);
5222}
5223
5224
ac1f60db 5225static u32 s2io_ethtool_get_rx_csum(struct net_device * dev)
1da177e4
LT
5226{
5227 nic_t *sp = dev->priv;
5228
5229 return (sp->rx_csum);
5230}
ac1f60db
AB
5231
5232static int s2io_ethtool_set_rx_csum(struct net_device *dev, u32 data)
1da177e4
LT
5233{
5234 nic_t *sp = dev->priv;
5235
5236 if (data)
5237 sp->rx_csum = 1;
5238 else
5239 sp->rx_csum = 0;
5240
5241 return 0;
5242}
ac1f60db
AB
5243
5244static int s2io_get_eeprom_len(struct net_device *dev)
1da177e4
LT
5245{
5246 return (XENA_EEPROM_SPACE);
5247}
5248
ac1f60db 5249static int s2io_ethtool_self_test_count(struct net_device *dev)
1da177e4
LT
5250{
5251 return (S2IO_TEST_LEN);
5252}
ac1f60db
AB
5253
5254static void s2io_ethtool_get_strings(struct net_device *dev,
5255 u32 stringset, u8 * data)
1da177e4
LT
5256{
5257 switch (stringset) {
5258 case ETH_SS_TEST:
5259 memcpy(data, s2io_gstrings, S2IO_STRINGS_LEN);
5260 break;
5261 case ETH_SS_STATS:
5262 memcpy(data, &ethtool_stats_keys,
5263 sizeof(ethtool_stats_keys));
5264 }
5265}
1da177e4
LT
5266static int s2io_ethtool_get_stats_count(struct net_device *dev)
5267{
5268 return (S2IO_STAT_LEN);
5269}
5270
ac1f60db 5271static int s2io_ethtool_op_set_tx_csum(struct net_device *dev, u32 data)
1da177e4
LT
5272{
5273 if (data)
5274 dev->features |= NETIF_F_IP_CSUM;
5275 else
5276 dev->features &= ~NETIF_F_IP_CSUM;
5277
5278 return 0;
5279}
5280
5281
5282static struct ethtool_ops netdev_ethtool_ops = {
5283 .get_settings = s2io_ethtool_gset,
5284 .set_settings = s2io_ethtool_sset,
5285 .get_drvinfo = s2io_ethtool_gdrvinfo,
5286 .get_regs_len = s2io_ethtool_get_regs_len,
5287 .get_regs = s2io_ethtool_gregs,
5288 .get_link = ethtool_op_get_link,
5289 .get_eeprom_len = s2io_get_eeprom_len,
5290 .get_eeprom = s2io_ethtool_geeprom,
5291 .set_eeprom = s2io_ethtool_seeprom,
5292 .get_pauseparam = s2io_ethtool_getpause_data,
5293 .set_pauseparam = s2io_ethtool_setpause_data,
5294 .get_rx_csum = s2io_ethtool_get_rx_csum,
5295 .set_rx_csum = s2io_ethtool_set_rx_csum,
5296 .get_tx_csum = ethtool_op_get_tx_csum,
5297 .set_tx_csum = s2io_ethtool_op_set_tx_csum,
5298 .get_sg = ethtool_op_get_sg,
5299 .set_sg = ethtool_op_set_sg,
5300#ifdef NETIF_F_TSO
5301 .get_tso = ethtool_op_get_tso,
5302 .set_tso = ethtool_op_set_tso,
5303#endif
fed5eccd
AR
5304 .get_ufo = ethtool_op_get_ufo,
5305 .set_ufo = ethtool_op_set_ufo,
1da177e4
LT
5306 .self_test_count = s2io_ethtool_self_test_count,
5307 .self_test = s2io_ethtool_test,
5308 .get_strings = s2io_ethtool_get_strings,
5309 .phys_id = s2io_ethtool_idnic,
5310 .get_stats_count = s2io_ethtool_get_stats_count,
5311 .get_ethtool_stats = s2io_get_ethtool_stats
5312};
5313
5314/**
20346722 5315 * s2io_ioctl - Entry point for the Ioctl
1da177e4
LT
5316 * @dev : Device pointer.
5317 * @ifr : An IOCTL specefic structure, that can contain a pointer to
5318 * a proprietary structure used to pass information to the driver.
5319 * @cmd : This is used to distinguish between the different commands that
5320 * can be passed to the IOCTL functions.
5321 * Description:
20346722 5322 * Currently there are no special functionality supported in IOCTL, hence
5323 * function always return EOPNOTSUPPORTED
1da177e4
LT
5324 */
5325
ac1f60db 5326static int s2io_ioctl(struct net_device *dev, struct ifreq *rq, int cmd)
1da177e4
LT
5327{
5328 return -EOPNOTSUPP;
5329}
5330
5331/**
5332 * s2io_change_mtu - entry point to change MTU size for the device.
5333 * @dev : device pointer.
5334 * @new_mtu : the new MTU size for the device.
5335 * Description: A driver entry point to change MTU size for the device.
5336 * Before changing the MTU the device must be stopped.
5337 * Return value:
5338 * 0 on success and an appropriate (-)ve integer as defined in errno.h
5339 * file on failure.
5340 */
5341
ac1f60db 5342static int s2io_change_mtu(struct net_device *dev, int new_mtu)
1da177e4
LT
5343{
5344 nic_t *sp = dev->priv;
1da177e4
LT
5345
5346 if ((new_mtu < MIN_MTU) || (new_mtu > S2IO_JUMBO_SIZE)) {
5347 DBG_PRINT(ERR_DBG, "%s: MTU size is invalid.\n",
5348 dev->name);
5349 return -EPERM;
5350 }
5351
1da177e4 5352 dev->mtu = new_mtu;
d8892c6e 5353 if (netif_running(dev)) {
5354 s2io_card_down(sp);
5355 netif_stop_queue(dev);
5356 if (s2io_card_up(sp)) {
5357 DBG_PRINT(ERR_DBG, "%s: Device bring up failed\n",
5358 __FUNCTION__);
5359 }
5360 if (netif_queue_stopped(dev))
5361 netif_wake_queue(dev);
5362 } else { /* Device is down */
5363 XENA_dev_config_t __iomem *bar0 = sp->bar0;
5364 u64 val64 = new_mtu;
5365
5366 writeq(vBIT(val64, 2, 14), &bar0->rmac_max_pyld_len);
5367 }
1da177e4
LT
5368
5369 return 0;
5370}
5371
5372/**
5373 * s2io_tasklet - Bottom half of the ISR.
5374 * @dev_adr : address of the device structure in dma_addr_t format.
5375 * Description:
5376 * This is the tasklet or the bottom half of the ISR. This is
20346722 5377 * an extension of the ISR which is scheduled by the scheduler to be run
1da177e4 5378 * when the load on the CPU is low. All low priority tasks of the ISR can
20346722 5379 * be pushed into the tasklet. For now the tasklet is used only to
1da177e4
LT
5380 * replenish the Rx buffers in the Rx buffer descriptors.
5381 * Return value:
5382 * void.
5383 */
5384
5385static void s2io_tasklet(unsigned long dev_addr)
5386{
5387 struct net_device *dev = (struct net_device *) dev_addr;
5388 nic_t *sp = dev->priv;
5389 int i, ret;
5390 mac_info_t *mac_control;
5391 struct config_param *config;
5392
5393 mac_control = &sp->mac_control;
5394 config = &sp->config;
5395
5396 if (!TASKLET_IN_USE) {
5397 for (i = 0; i < config->rx_ring_num; i++) {
5398 ret = fill_rx_buffers(sp, i);
5399 if (ret == -ENOMEM) {
5400 DBG_PRINT(ERR_DBG, "%s: Out of ",
5401 dev->name);
5402 DBG_PRINT(ERR_DBG, "memory in tasklet\n");
5403 break;
5404 } else if (ret == -EFILL) {
5405 DBG_PRINT(ERR_DBG,
5406 "%s: Rx Ring %d is full\n",
5407 dev->name, i);
5408 break;
5409 }
5410 }
5411 clear_bit(0, (&sp->tasklet_status));
5412 }
5413}
5414
5415/**
5416 * s2io_set_link - Set the LInk status
5417 * @data: long pointer to device private structue
5418 * Description: Sets the link status for the adapter
5419 */
5420
5421static void s2io_set_link(unsigned long data)
5422{
5423 nic_t *nic = (nic_t *) data;
5424 struct net_device *dev = nic->dev;
5425 XENA_dev_config_t __iomem *bar0 = nic->bar0;
5426 register u64 val64;
5427 u16 subid;
5428
5429 if (test_and_set_bit(0, &(nic->link_state))) {
5430 /* The card is being reset, no point doing anything */
5431 return;
5432 }
5433
5434 subid = nic->pdev->subsystem_device;
a371a07d 5435 if (s2io_link_fault_indication(nic) == MAC_RMAC_ERR_TIMER) {
5436 /*
5437 * Allow a small delay for the NICs self initiated
5438 * cleanup to complete.
5439 */
5440 msleep(100);
5441 }
1da177e4
LT
5442
5443 val64 = readq(&bar0->adapter_status);
20346722 5444 if (verify_xena_quiescence(nic, val64, nic->device_enabled_once)) {
1da177e4
LT
5445 if (LINK_IS_UP(val64)) {
5446 val64 = readq(&bar0->adapter_control);
5447 val64 |= ADAPTER_CNTL_EN;
5448 writeq(val64, &bar0->adapter_control);
541ae68f 5449 if (CARDS_WITH_FAULTY_LINK_INDICATORS(nic->device_type,
5450 subid)) {
1da177e4
LT
5451 val64 = readq(&bar0->gpio_control);
5452 val64 |= GPIO_CTRL_GPIO_0;
5453 writeq(val64, &bar0->gpio_control);
5454 val64 = readq(&bar0->gpio_control);
5455 } else {
5456 val64 |= ADAPTER_LED_ON;
5457 writeq(val64, &bar0->adapter_control);
5458 }
a371a07d 5459 if (s2io_link_fault_indication(nic) ==
5460 MAC_RMAC_ERR_TIMER) {
5461 val64 = readq(&bar0->adapter_status);
5462 if (!LINK_IS_UP(val64)) {
5463 DBG_PRINT(ERR_DBG, "%s:", dev->name);
5464 DBG_PRINT(ERR_DBG, " Link down");
5465 DBG_PRINT(ERR_DBG, "after ");
5466 DBG_PRINT(ERR_DBG, "enabling ");
5467 DBG_PRINT(ERR_DBG, "device \n");
5468 }
1da177e4
LT
5469 }
5470 if (nic->device_enabled_once == FALSE) {
5471 nic->device_enabled_once = TRUE;
5472 }
5473 s2io_link(nic, LINK_UP);
5474 } else {
541ae68f 5475 if (CARDS_WITH_FAULTY_LINK_INDICATORS(nic->device_type,
5476 subid)) {
1da177e4
LT
5477 val64 = readq(&bar0->gpio_control);
5478 val64 &= ~GPIO_CTRL_GPIO_0;
5479 writeq(val64, &bar0->gpio_control);
5480 val64 = readq(&bar0->gpio_control);
5481 }
5482 s2io_link(nic, LINK_DOWN);
5483 }
5484 } else { /* NIC is not Quiescent. */
5485 DBG_PRINT(ERR_DBG, "%s: Error: ", dev->name);
5486 DBG_PRINT(ERR_DBG, "device is not Quiescent\n");
5487 netif_stop_queue(dev);
5488 }
5489 clear_bit(0, &(nic->link_state));
5490}
5491
5492static void s2io_card_down(nic_t * sp)
5493{
5494 int cnt = 0;
5495 XENA_dev_config_t __iomem *bar0 = sp->bar0;
5496 unsigned long flags;
5497 register u64 val64 = 0;
5498
25fff88e 5499 del_timer_sync(&sp->alarm_timer);
1da177e4 5500 /* If s2io_set_link task is executing, wait till it completes. */
20346722 5501 while (test_and_set_bit(0, &(sp->link_state))) {
1da177e4 5502 msleep(50);
20346722 5503 }
1da177e4
LT
5504 atomic_set(&sp->card_state, CARD_DOWN);
5505
5506 /* disable Tx and Rx traffic on the NIC */
5507 stop_nic(sp);
5508
5509 /* Kill tasklet. */
5510 tasklet_kill(&sp->task);
5511
5512 /* Check if the device is Quiescent and then Reset the NIC */
5513 do {
5514 val64 = readq(&bar0->adapter_status);
20346722 5515 if (verify_xena_quiescence(sp, val64, sp->device_enabled_once)) {
1da177e4
LT
5516 break;
5517 }
5518
5519 msleep(50);
5520 cnt++;
5521 if (cnt == 10) {
5522 DBG_PRINT(ERR_DBG,
5523 "s2io_close:Device not Quiescent ");
5524 DBG_PRINT(ERR_DBG, "adaper status reads 0x%llx\n",
5525 (unsigned long long) val64);
5526 break;
5527 }
5528 } while (1);
1da177e4
LT
5529 s2io_reset(sp);
5530
7ba013ac 5531 /* Waiting till all Interrupt handlers are complete */
5532 cnt = 0;
5533 do {
5534 msleep(10);
5535 if (!atomic_read(&sp->isr_cnt))
5536 break;
5537 cnt++;
5538 } while(cnt < 5);
5539
5540 spin_lock_irqsave(&sp->tx_lock, flags);
5541 /* Free all Tx buffers */
1da177e4 5542 free_tx_buffers(sp);
7ba013ac 5543 spin_unlock_irqrestore(&sp->tx_lock, flags);
5544
5545 /* Free all Rx buffers */
5546 spin_lock_irqsave(&sp->rx_lock, flags);
1da177e4 5547 free_rx_buffers(sp);
7ba013ac 5548 spin_unlock_irqrestore(&sp->rx_lock, flags);
1da177e4 5549
1da177e4
LT
5550 clear_bit(0, &(sp->link_state));
5551}
5552
5553static int s2io_card_up(nic_t * sp)
5554{
cc6e7c44 5555 int i, ret = 0;
1da177e4
LT
5556 mac_info_t *mac_control;
5557 struct config_param *config;
5558 struct net_device *dev = (struct net_device *) sp->dev;
5559
5560 /* Initialize the H/W I/O registers */
5561 if (init_nic(sp) != 0) {
5562 DBG_PRINT(ERR_DBG, "%s: H/W initialization failed\n",
5563 dev->name);
5564 return -ENODEV;
5565 }
5566
cc6e7c44
RA
5567 if (sp->intr_type == MSI)
5568 ret = s2io_enable_msi(sp);
5569 else if (sp->intr_type == MSI_X)
5570 ret = s2io_enable_msi_x(sp);
5571 if (ret) {
5572 DBG_PRINT(ERR_DBG, "%s: Defaulting to INTA\n", dev->name);
5573 sp->intr_type = INTA;
5574 }
5575
20346722 5576 /*
5577 * Initializing the Rx buffers. For now we are considering only 1
1da177e4
LT
5578 * Rx ring and initializing buffers into 30 Rx blocks
5579 */
5580 mac_control = &sp->mac_control;
5581 config = &sp->config;
5582
5583 for (i = 0; i < config->rx_ring_num; i++) {
5584 if ((ret = fill_rx_buffers(sp, i))) {
5585 DBG_PRINT(ERR_DBG, "%s: Out of memory in Open\n",
5586 dev->name);
5587 s2io_reset(sp);
5588 free_rx_buffers(sp);
5589 return -ENOMEM;
5590 }
5591 DBG_PRINT(INFO_DBG, "Buf in ring:%d is %d:\n", i,
5592 atomic_read(&sp->rx_bufs_left[i]));
5593 }
5594
5595 /* Setting its receive mode */
5596 s2io_set_multicast(dev);
5597
7d3d0439
RA
5598 if (sp->lro) {
5599 /* Initialize max aggregatable pkts based on MTU */
5600 sp->lro_max_aggr_per_sess = ((1<<16) - 1) / dev->mtu;
5601 /* Check if we can use(if specified) user provided value */
5602 if (lro_max_pkts < sp->lro_max_aggr_per_sess)
5603 sp->lro_max_aggr_per_sess = lro_max_pkts;
5604 }
5605
1da177e4
LT
5606 /* Enable tasklet for the device */
5607 tasklet_init(&sp->task, s2io_tasklet, (unsigned long) dev);
5608
5609 /* Enable Rx Traffic and interrupts on the NIC */
5610 if (start_nic(sp)) {
5611 DBG_PRINT(ERR_DBG, "%s: Starting NIC failed\n", dev->name);
5612 tasklet_kill(&sp->task);
5613 s2io_reset(sp);
5614 free_irq(dev->irq, dev);
5615 free_rx_buffers(sp);
5616 return -ENODEV;
5617 }
5618
25fff88e 5619 S2IO_TIMER_CONF(sp->alarm_timer, s2io_alarm_handle, sp, (HZ/2));
5620
1da177e4
LT
5621 atomic_set(&sp->card_state, CARD_UP);
5622 return 0;
5623}
5624
20346722 5625/**
1da177e4
LT
5626 * s2io_restart_nic - Resets the NIC.
5627 * @data : long pointer to the device private structure
5628 * Description:
5629 * This function is scheduled to be run by the s2io_tx_watchdog
20346722 5630 * function after 0.5 secs to reset the NIC. The idea is to reduce
1da177e4
LT
5631 * the run time of the watch dog routine which is run holding a
5632 * spin lock.
5633 */
5634
5635static void s2io_restart_nic(unsigned long data)
5636{
5637 struct net_device *dev = (struct net_device *) data;
5638 nic_t *sp = dev->priv;
5639
5640 s2io_card_down(sp);
5641 if (s2io_card_up(sp)) {
5642 DBG_PRINT(ERR_DBG, "%s: Device bring up failed\n",
5643 dev->name);
5644 }
5645 netif_wake_queue(dev);
5646 DBG_PRINT(ERR_DBG, "%s: was reset by Tx watchdog timer\n",
5647 dev->name);
20346722 5648
1da177e4
LT
5649}
5650
20346722 5651/**
5652 * s2io_tx_watchdog - Watchdog for transmit side.
1da177e4
LT
5653 * @dev : Pointer to net device structure
5654 * Description:
5655 * This function is triggered if the Tx Queue is stopped
5656 * for a pre-defined amount of time when the Interface is still up.
5657 * If the Interface is jammed in such a situation, the hardware is
5658 * reset (by s2io_close) and restarted again (by s2io_open) to
5659 * overcome any problem that might have been caused in the hardware.
5660 * Return value:
5661 * void
5662 */
5663
5664static void s2io_tx_watchdog(struct net_device *dev)
5665{
5666 nic_t *sp = dev->priv;
5667
5668 if (netif_carrier_ok(dev)) {
5669 schedule_work(&sp->rst_timer_task);
5670 }
5671}
5672
5673/**
5674 * rx_osm_handler - To perform some OS related operations on SKB.
5675 * @sp: private member of the device structure,pointer to s2io_nic structure.
5676 * @skb : the socket buffer pointer.
5677 * @len : length of the packet
5678 * @cksum : FCS checksum of the frame.
5679 * @ring_no : the ring from which this RxD was extracted.
20346722 5680 * Description:
1da177e4
LT
5681 * This function is called by the Tx interrupt serivce routine to perform
5682 * some OS related operations on the SKB before passing it to the upper
5683 * layers. It mainly checks if the checksum is OK, if so adds it to the
5684 * SKBs cksum variable, increments the Rx packet count and passes the SKB
5685 * to the upper layer. If the checksum is wrong, it increments the Rx
5686 * packet error count, frees the SKB and returns error.
5687 * Return value:
5688 * SUCCESS on success and -1 on failure.
5689 */
20346722 5690static int rx_osm_handler(ring_info_t *ring_data, RxD_t * rxdp)
1da177e4 5691{
20346722 5692 nic_t *sp = ring_data->nic;
1da177e4 5693 struct net_device *dev = (struct net_device *) sp->dev;
20346722 5694 struct sk_buff *skb = (struct sk_buff *)
5695 ((unsigned long) rxdp->Host_Control);
5696 int ring_no = ring_data->ring_no;
1da177e4 5697 u16 l3_csum, l4_csum;
7d3d0439 5698 lro_t *lro;
da6971d8 5699
20346722 5700 skb->dev = dev;
5701 if (rxdp->Control_1 & RXD_T_CODE) {
5702 unsigned long long err = rxdp->Control_1 & RXD_T_CODE;
5703 DBG_PRINT(ERR_DBG, "%s: Rx error Value: 0x%llx\n",
5704 dev->name, err);
1ddc50d4 5705 dev_kfree_skb(skb);
5706 sp->stats.rx_crc_errors++;
5707 atomic_dec(&sp->rx_bufs_left[ring_no]);
5708 rxdp->Host_Control = 0;
5709 return 0;
20346722 5710 }
1da177e4 5711
20346722 5712 /* Updating statistics */
5713 rxdp->Host_Control = 0;
5714 sp->rx_pkt_count++;
5715 sp->stats.rx_packets++;
da6971d8
AR
5716 if (sp->rxd_mode == RXD_MODE_1) {
5717 int len = RXD_GET_BUFFER0_SIZE_1(rxdp->Control_2);
20346722 5718
da6971d8
AR
5719 sp->stats.rx_bytes += len;
5720 skb_put(skb, len);
5721
5722 } else if (sp->rxd_mode >= RXD_MODE_3A) {
5723 int get_block = ring_data->rx_curr_get_info.block_index;
5724 int get_off = ring_data->rx_curr_get_info.offset;
5725 int buf0_len = RXD_GET_BUFFER0_SIZE_3(rxdp->Control_2);
5726 int buf2_len = RXD_GET_BUFFER2_SIZE_3(rxdp->Control_2);
5727 unsigned char *buff = skb_push(skb, buf0_len);
5728
5729 buffAdd_t *ba = &ring_data->ba[get_block][get_off];
5730 sp->stats.rx_bytes += buf0_len + buf2_len;
5731 memcpy(buff, ba->ba_0, buf0_len);
5732
5733 if (sp->rxd_mode == RXD_MODE_3A) {
5734 int buf1_len = RXD_GET_BUFFER1_SIZE_3(rxdp->Control_2);
5735
5736 skb_put(skb, buf1_len);
5737 skb->len += buf2_len;
5738 skb->data_len += buf2_len;
5739 skb->truesize += buf2_len;
5740 skb_put(skb_shinfo(skb)->frag_list, buf2_len);
5741 sp->stats.rx_bytes += buf1_len;
5742
5743 } else
5744 skb_put(skb, buf2_len);
5745 }
20346722 5746
7d3d0439
RA
5747 if ((rxdp->Control_1 & TCP_OR_UDP_FRAME) && ((!sp->lro) ||
5748 (sp->lro && (!(rxdp->Control_1 & RXD_FRAME_IP_FRAG)))) &&
20346722 5749 (sp->rx_csum)) {
5750 l3_csum = RXD_GET_L3_CKSUM(rxdp->Control_1);
1da177e4
LT
5751 l4_csum = RXD_GET_L4_CKSUM(rxdp->Control_1);
5752 if ((l3_csum == L3_CKSUM_OK) && (l4_csum == L4_CKSUM_OK)) {
20346722 5753 /*
1da177e4
LT
5754 * NIC verifies if the Checksum of the received
5755 * frame is Ok or not and accordingly returns
5756 * a flag in the RxD.
5757 */
5758 skb->ip_summed = CHECKSUM_UNNECESSARY;
7d3d0439
RA
5759 if (sp->lro) {
5760 u32 tcp_len;
5761 u8 *tcp;
5762 int ret = 0;
5763
5764 ret = s2io_club_tcp_session(skb->data, &tcp,
5765 &tcp_len, &lro, rxdp, sp);
5766 switch (ret) {
5767 case 3: /* Begin anew */
5768 lro->parent = skb;
5769 goto aggregate;
5770 case 1: /* Aggregate */
5771 {
5772 lro_append_pkt(sp, lro,
5773 skb, tcp_len);
5774 goto aggregate;
5775 }
5776 case 4: /* Flush session */
5777 {
5778 lro_append_pkt(sp, lro,
5779 skb, tcp_len);
5780 queue_rx_frame(lro->parent);
5781 clear_lro_session(lro);
5782 sp->mac_control.stats_info->
5783 sw_stat.flush_max_pkts++;
5784 goto aggregate;
5785 }
5786 case 2: /* Flush both */
5787 lro->parent->data_len =
5788 lro->frags_len;
5789 sp->mac_control.stats_info->
5790 sw_stat.sending_both++;
5791 queue_rx_frame(lro->parent);
5792 clear_lro_session(lro);
5793 goto send_up;
5794 case 0: /* sessions exceeded */
5795 case 5: /*
5796 * First pkt in session not
5797 * L3/L4 aggregatable
5798 */
5799 break;
5800 default:
5801 DBG_PRINT(ERR_DBG,
5802 "%s: Samadhana!!\n",
5803 __FUNCTION__);
5804 BUG();
5805 }
5806 }
1da177e4 5807 } else {
20346722 5808 /*
5809 * Packet with erroneous checksum, let the
1da177e4
LT
5810 * upper layers deal with it.
5811 */
5812 skb->ip_summed = CHECKSUM_NONE;
5813 }
5814 } else {
5815 skb->ip_summed = CHECKSUM_NONE;
5816 }
5817
7d3d0439
RA
5818 if (!sp->lro) {
5819 skb->protocol = eth_type_trans(skb, dev);
1da177e4 5820#ifdef CONFIG_S2IO_NAPI
7d3d0439
RA
5821 if (sp->vlgrp && RXD_GET_VLAN_TAG(rxdp->Control_2)) {
5822 /* Queueing the vlan frame to the upper layer */
5823 vlan_hwaccel_receive_skb(skb, sp->vlgrp,
5824 RXD_GET_VLAN_TAG(rxdp->Control_2));
5825 } else {
5826 netif_receive_skb(skb);
5827 }
1da177e4 5828#else
7d3d0439
RA
5829 if (sp->vlgrp && RXD_GET_VLAN_TAG(rxdp->Control_2)) {
5830 /* Queueing the vlan frame to the upper layer */
5831 vlan_hwaccel_rx(skb, sp->vlgrp,
5832 RXD_GET_VLAN_TAG(rxdp->Control_2));
5833 } else {
5834 netif_rx(skb);
5835 }
1da177e4 5836#endif
7d3d0439
RA
5837 } else {
5838send_up:
5839 queue_rx_frame(skb);
5840 }
1da177e4 5841 dev->last_rx = jiffies;
7d3d0439 5842aggregate:
1da177e4 5843 atomic_dec(&sp->rx_bufs_left[ring_no]);
1da177e4
LT
5844 return SUCCESS;
5845}
5846
5847/**
5848 * s2io_link - stops/starts the Tx queue.
5849 * @sp : private member of the device structure, which is a pointer to the
5850 * s2io_nic structure.
5851 * @link : inidicates whether link is UP/DOWN.
5852 * Description:
5853 * This function stops/starts the Tx queue depending on whether the link
20346722 5854 * status of the NIC is is down or up. This is called by the Alarm
5855 * interrupt handler whenever a link change interrupt comes up.
1da177e4
LT
5856 * Return value:
5857 * void.
5858 */
5859
26df54bf 5860static void s2io_link(nic_t * sp, int link)
1da177e4
LT
5861{
5862 struct net_device *dev = (struct net_device *) sp->dev;
5863
5864 if (link != sp->last_link_state) {
5865 if (link == LINK_DOWN) {
5866 DBG_PRINT(ERR_DBG, "%s: Link down\n", dev->name);
5867 netif_carrier_off(dev);
5868 } else {
5869 DBG_PRINT(ERR_DBG, "%s: Link Up\n", dev->name);
5870 netif_carrier_on(dev);
5871 }
5872 }
5873 sp->last_link_state = link;
5874}
5875
5876/**
20346722 5877 * get_xena_rev_id - to identify revision ID of xena.
5878 * @pdev : PCI Dev structure
5879 * Description:
5880 * Function to identify the Revision ID of xena.
5881 * Return value:
5882 * returns the revision ID of the device.
5883 */
5884
26df54bf 5885static int get_xena_rev_id(struct pci_dev *pdev)
20346722 5886{
5887 u8 id = 0;
5888 int ret;
5889 ret = pci_read_config_byte(pdev, PCI_REVISION_ID, (u8 *) & id);
5890 return id;
5891}
5892
5893/**
5894 * s2io_init_pci -Initialization of PCI and PCI-X configuration registers .
5895 * @sp : private member of the device structure, which is a pointer to the
1da177e4
LT
5896 * s2io_nic structure.
5897 * Description:
5898 * This function initializes a few of the PCI and PCI-X configuration registers
5899 * with recommended values.
5900 * Return value:
5901 * void
5902 */
5903
5904static void s2io_init_pci(nic_t * sp)
5905{
20346722 5906 u16 pci_cmd = 0, pcix_cmd = 0;
1da177e4
LT
5907
5908 /* Enable Data Parity Error Recovery in PCI-X command register. */
5909 pci_read_config_word(sp->pdev, PCIX_COMMAND_REGISTER,
20346722 5910 &(pcix_cmd));
1da177e4 5911 pci_write_config_word(sp->pdev, PCIX_COMMAND_REGISTER,
20346722 5912 (pcix_cmd | 1));
1da177e4 5913 pci_read_config_word(sp->pdev, PCIX_COMMAND_REGISTER,
20346722 5914 &(pcix_cmd));
1da177e4
LT
5915
5916 /* Set the PErr Response bit in PCI command register. */
5917 pci_read_config_word(sp->pdev, PCI_COMMAND, &pci_cmd);
5918 pci_write_config_word(sp->pdev, PCI_COMMAND,
5919 (pci_cmd | PCI_COMMAND_PARITY));
5920 pci_read_config_word(sp->pdev, PCI_COMMAND, &pci_cmd);
5921
1da177e4 5922 /* Forcibly disabling relaxed ordering capability of the card. */
20346722 5923 pcix_cmd &= 0xfffd;
1da177e4 5924 pci_write_config_word(sp->pdev, PCIX_COMMAND_REGISTER,
20346722 5925 pcix_cmd);
1da177e4 5926 pci_read_config_word(sp->pdev, PCIX_COMMAND_REGISTER,
20346722 5927 &(pcix_cmd));
1da177e4
LT
5928}
5929
5930MODULE_AUTHOR("Raghavendra Koushik <raghavendra.koushik@neterion.com>");
5931MODULE_LICENSE("GPL");
6c1792f4
JL
5932MODULE_VERSION(DRV_VERSION);
5933
1da177e4 5934module_param(tx_fifo_num, int, 0);
1da177e4 5935module_param(rx_ring_num, int, 0);
da6971d8 5936module_param(rx_ring_mode, int, 0);
20346722 5937module_param_array(tx_fifo_len, uint, NULL, 0);
5938module_param_array(rx_ring_sz, uint, NULL, 0);
20346722 5939module_param_array(rts_frm_len, uint, NULL, 0);
5e25b9dd 5940module_param(use_continuous_tx_intrs, int, 1);
1da177e4
LT
5941module_param(rmac_pause_time, int, 0);
5942module_param(mc_pause_threshold_q0q3, int, 0);
5943module_param(mc_pause_threshold_q4q7, int, 0);
5944module_param(shared_splits, int, 0);
5945module_param(tmac_util_period, int, 0);
5946module_param(rmac_util_period, int, 0);
b6e3f982 5947module_param(bimodal, bool, 0);
da6971d8 5948module_param(l3l4hdr_size, int , 0);
1da177e4
LT
5949#ifndef CONFIG_S2IO_NAPI
5950module_param(indicate_max_pkts, int, 0);
5951#endif
303bcb4b 5952module_param(rxsync_frequency, int, 0);
cc6e7c44 5953module_param(intr_type, int, 0);
7d3d0439
RA
5954module_param(lro, int, 0);
5955module_param(lro_max_pkts, int, 0);
20346722 5956
1da177e4 5957/**
20346722 5958 * s2io_init_nic - Initialization of the adapter .
1da177e4
LT
5959 * @pdev : structure containing the PCI related information of the device.
5960 * @pre: List of PCI devices supported by the driver listed in s2io_tbl.
5961 * Description:
5962 * The function initializes an adapter identified by the pci_dec structure.
20346722 5963 * All OS related initialization including memory and device structure and
5964 * initlaization of the device private variable is done. Also the swapper
5965 * control register is initialized to enable read and write into the I/O
1da177e4
LT
5966 * registers of the device.
5967 * Return value:
5968 * returns 0 on success and negative on failure.
5969 */
5970
5971static int __devinit
5972s2io_init_nic(struct pci_dev *pdev, const struct pci_device_id *pre)
5973{
5974 nic_t *sp;
5975 struct net_device *dev;
1da177e4
LT
5976 int i, j, ret;
5977 int dma_flag = FALSE;
5978 u32 mac_up, mac_down;
5979 u64 val64 = 0, tmp64 = 0;
5980 XENA_dev_config_t __iomem *bar0 = NULL;
5981 u16 subid;
5982 mac_info_t *mac_control;
5983 struct config_param *config;
541ae68f 5984 int mode;
cc6e7c44 5985 u8 dev_intr_type = intr_type;
1da177e4 5986
20346722 5987#ifdef CONFIG_S2IO_NAPI
cc6e7c44
RA
5988 if (dev_intr_type != INTA) {
5989 DBG_PRINT(ERR_DBG, "NAPI cannot be enabled when MSI/MSI-X \
5990is enabled. Defaulting to INTA\n");
5991 dev_intr_type = INTA;
5992 }
5993 else
5994 DBG_PRINT(ERR_DBG, "NAPI support has been enabled\n");
20346722 5995#endif
1da177e4
LT
5996
5997 if ((ret = pci_enable_device(pdev))) {
5998 DBG_PRINT(ERR_DBG,
5999 "s2io_init_nic: pci_enable_device failed\n");
6000 return ret;
6001 }
6002
1e7f0bd8 6003 if (!pci_set_dma_mask(pdev, DMA_64BIT_MASK)) {
1da177e4
LT
6004 DBG_PRINT(INIT_DBG, "s2io_init_nic: Using 64bit DMA\n");
6005 dma_flag = TRUE;
1da177e4 6006 if (pci_set_consistent_dma_mask
1e7f0bd8 6007 (pdev, DMA_64BIT_MASK)) {
1da177e4
LT
6008 DBG_PRINT(ERR_DBG,
6009 "Unable to obtain 64bit DMA for \
6010 consistent allocations\n");
6011 pci_disable_device(pdev);
6012 return -ENOMEM;
6013 }
1e7f0bd8 6014 } else if (!pci_set_dma_mask(pdev, DMA_32BIT_MASK)) {
1da177e4
LT
6015 DBG_PRINT(INIT_DBG, "s2io_init_nic: Using 32bit DMA\n");
6016 } else {
6017 pci_disable_device(pdev);
6018 return -ENOMEM;
6019 }
6020
cc6e7c44
RA
6021 if ((dev_intr_type == MSI_X) &&
6022 ((pdev->device != PCI_DEVICE_ID_HERC_WIN) &&
6023 (pdev->device != PCI_DEVICE_ID_HERC_UNI))) {
6024 DBG_PRINT(ERR_DBG, "Xframe I does not support MSI_X. \
6025Defaulting to INTA\n");
6026 dev_intr_type = INTA;
6027 }
6028 if (dev_intr_type != MSI_X) {
6029 if (pci_request_regions(pdev, s2io_driver_name)) {
6030 DBG_PRINT(ERR_DBG, "Request Regions failed\n"),
6031 pci_disable_device(pdev);
6032 return -ENODEV;
6033 }
6034 }
6035 else {
6036 if (!(request_mem_region(pci_resource_start(pdev, 0),
6037 pci_resource_len(pdev, 0), s2io_driver_name))) {
6038 DBG_PRINT(ERR_DBG, "bar0 Request Regions failed\n");
6039 pci_disable_device(pdev);
6040 return -ENODEV;
6041 }
6042 if (!(request_mem_region(pci_resource_start(pdev, 2),
6043 pci_resource_len(pdev, 2), s2io_driver_name))) {
6044 DBG_PRINT(ERR_DBG, "bar1 Request Regions failed\n");
6045 release_mem_region(pci_resource_start(pdev, 0),
6046 pci_resource_len(pdev, 0));
6047 pci_disable_device(pdev);
6048 return -ENODEV;
6049 }
1da177e4
LT
6050 }
6051
6052 dev = alloc_etherdev(sizeof(nic_t));
6053 if (dev == NULL) {
6054 DBG_PRINT(ERR_DBG, "Device allocation failed\n");
6055 pci_disable_device(pdev);
6056 pci_release_regions(pdev);
6057 return -ENODEV;
6058 }
6059
6060 pci_set_master(pdev);
6061 pci_set_drvdata(pdev, dev);
6062 SET_MODULE_OWNER(dev);
6063 SET_NETDEV_DEV(dev, &pdev->dev);
6064
6065 /* Private member variable initialized to s2io NIC structure */
6066 sp = dev->priv;
6067 memset(sp, 0, sizeof(nic_t));
6068 sp->dev = dev;
6069 sp->pdev = pdev;
1da177e4 6070 sp->high_dma_flag = dma_flag;
1da177e4 6071 sp->device_enabled_once = FALSE;
da6971d8
AR
6072 if (rx_ring_mode == 1)
6073 sp->rxd_mode = RXD_MODE_1;
6074 if (rx_ring_mode == 2)
6075 sp->rxd_mode = RXD_MODE_3B;
6076 if (rx_ring_mode == 3)
6077 sp->rxd_mode = RXD_MODE_3A;
6078
cc6e7c44 6079 sp->intr_type = dev_intr_type;
1da177e4 6080
541ae68f 6081 if ((pdev->device == PCI_DEVICE_ID_HERC_WIN) ||
6082 (pdev->device == PCI_DEVICE_ID_HERC_UNI))
6083 sp->device_type = XFRAME_II_DEVICE;
6084 else
6085 sp->device_type = XFRAME_I_DEVICE;
6086
7d3d0439 6087 sp->lro = lro;
cc6e7c44 6088
1da177e4
LT
6089 /* Initialize some PCI/PCI-X fields of the NIC. */
6090 s2io_init_pci(sp);
6091
20346722 6092 /*
1da177e4 6093 * Setting the device configuration parameters.
20346722 6094 * Most of these parameters can be specified by the user during
6095 * module insertion as they are module loadable parameters. If
6096 * these parameters are not not specified during load time, they
1da177e4
LT
6097 * are initialized with default values.
6098 */
6099 mac_control = &sp->mac_control;
6100 config = &sp->config;
6101
6102 /* Tx side parameters. */
0b1f7ebe 6103 if (tx_fifo_len[0] == 0)
6104 tx_fifo_len[0] = DEFAULT_FIFO_LEN; /* Default value. */
1da177e4
LT
6105 config->tx_fifo_num = tx_fifo_num;
6106 for (i = 0; i < MAX_TX_FIFOS; i++) {
6107 config->tx_cfg[i].fifo_len = tx_fifo_len[i];
6108 config->tx_cfg[i].fifo_priority = i;
6109 }
6110
20346722 6111 /* mapping the QoS priority to the configured fifos */
6112 for (i = 0; i < MAX_TX_FIFOS; i++)
6113 config->fifo_mapping[i] = fifo_map[config->tx_fifo_num][i];
6114
1da177e4
LT
6115 config->tx_intr_type = TXD_INT_TYPE_UTILZ;
6116 for (i = 0; i < config->tx_fifo_num; i++) {
6117 config->tx_cfg[i].f_no_snoop =
6118 (NO_SNOOP_TXD | NO_SNOOP_TXD_BUFFER);
6119 if (config->tx_cfg[i].fifo_len < 65) {
6120 config->tx_intr_type = TXD_INT_TYPE_PER_LIST;
6121 break;
6122 }
6123 }
fed5eccd
AR
6124 /* + 2 because one Txd for skb->data and one Txd for UFO */
6125 config->max_txds = MAX_SKB_FRAGS + 2;
1da177e4
LT
6126
6127 /* Rx side parameters. */
0b1f7ebe 6128 if (rx_ring_sz[0] == 0)
6129 rx_ring_sz[0] = SMALL_BLK_CNT; /* Default value. */
1da177e4
LT
6130 config->rx_ring_num = rx_ring_num;
6131 for (i = 0; i < MAX_RX_RINGS; i++) {
6132 config->rx_cfg[i].num_rxd = rx_ring_sz[i] *
da6971d8 6133 (rxd_count[sp->rxd_mode] + 1);
1da177e4
LT
6134 config->rx_cfg[i].ring_priority = i;
6135 }
6136
6137 for (i = 0; i < rx_ring_num; i++) {
6138 config->rx_cfg[i].ring_org = RING_ORG_BUFF1;
6139 config->rx_cfg[i].f_no_snoop =
6140 (NO_SNOOP_RXD | NO_SNOOP_RXD_BUFFER);
6141 }
6142
6143 /* Setting Mac Control parameters */
6144 mac_control->rmac_pause_time = rmac_pause_time;
6145 mac_control->mc_pause_threshold_q0q3 = mc_pause_threshold_q0q3;
6146 mac_control->mc_pause_threshold_q4q7 = mc_pause_threshold_q4q7;
6147
6148
6149 /* Initialize Ring buffer parameters. */
6150 for (i = 0; i < config->rx_ring_num; i++)
6151 atomic_set(&sp->rx_bufs_left[i], 0);
6152
7ba013ac 6153 /* Initialize the number of ISRs currently running */
6154 atomic_set(&sp->isr_cnt, 0);
6155
1da177e4
LT
6156 /* initialize the shared memory used by the NIC and the host */
6157 if (init_shared_mem(sp)) {
6158 DBG_PRINT(ERR_DBG, "%s: Memory allocation failed\n",
0b1f7ebe 6159 __FUNCTION__);
1da177e4
LT
6160 ret = -ENOMEM;
6161 goto mem_alloc_failed;
6162 }
6163
6164 sp->bar0 = ioremap(pci_resource_start(pdev, 0),
6165 pci_resource_len(pdev, 0));
6166 if (!sp->bar0) {
6167 DBG_PRINT(ERR_DBG, "%s: S2IO: cannot remap io mem1\n",
6168 dev->name);
6169 ret = -ENOMEM;
6170 goto bar0_remap_failed;
6171 }
6172
6173 sp->bar1 = ioremap(pci_resource_start(pdev, 2),
6174 pci_resource_len(pdev, 2));
6175 if (!sp->bar1) {
6176 DBG_PRINT(ERR_DBG, "%s: S2IO: cannot remap io mem2\n",
6177 dev->name);
6178 ret = -ENOMEM;
6179 goto bar1_remap_failed;
6180 }
6181
6182 dev->irq = pdev->irq;
6183 dev->base_addr = (unsigned long) sp->bar0;
6184
6185 /* Initializing the BAR1 address as the start of the FIFO pointer. */
6186 for (j = 0; j < MAX_TX_FIFOS; j++) {
6187 mac_control->tx_FIFO_start[j] = (TxFIFO_element_t __iomem *)
6188 (sp->bar1 + (j * 0x00020000));
6189 }
6190
6191 /* Driver entry points */
6192 dev->open = &s2io_open;
6193 dev->stop = &s2io_close;
6194 dev->hard_start_xmit = &s2io_xmit;
6195 dev->get_stats = &s2io_get_stats;
6196 dev->set_multicast_list = &s2io_set_multicast;
6197 dev->do_ioctl = &s2io_ioctl;
6198 dev->change_mtu = &s2io_change_mtu;
6199 SET_ETHTOOL_OPS(dev, &netdev_ethtool_ops);
be3a6b02 6200 dev->features |= NETIF_F_HW_VLAN_TX | NETIF_F_HW_VLAN_RX;
6201 dev->vlan_rx_register = s2io_vlan_rx_register;
6202 dev->vlan_rx_kill_vid = (void *)s2io_vlan_rx_kill_vid;
20346722 6203
1da177e4
LT
6204 /*
6205 * will use eth_mac_addr() for dev->set_mac_address
6206 * mac address will be set every time dev->open() is called
6207 */
20346722 6208#if defined(CONFIG_S2IO_NAPI)
1da177e4 6209 dev->poll = s2io_poll;
20346722 6210 dev->weight = 32;
1da177e4
LT
6211#endif
6212
6213 dev->features |= NETIF_F_SG | NETIF_F_IP_CSUM;
6214 if (sp->high_dma_flag == TRUE)
6215 dev->features |= NETIF_F_HIGHDMA;
6216#ifdef NETIF_F_TSO
6217 dev->features |= NETIF_F_TSO;
6218#endif
fed5eccd
AR
6219 if (sp->device_type & XFRAME_II_DEVICE) {
6220 dev->features |= NETIF_F_UFO;
6221 dev->features |= NETIF_F_HW_CSUM;
6222 }
1da177e4
LT
6223
6224 dev->tx_timeout = &s2io_tx_watchdog;
6225 dev->watchdog_timeo = WATCH_DOG_TIMEOUT;
6226 INIT_WORK(&sp->rst_timer_task,
6227 (void (*)(void *)) s2io_restart_nic, dev);
6228 INIT_WORK(&sp->set_link_task,
6229 (void (*)(void *)) s2io_set_link, sp);
6230
e960fc5c 6231 pci_save_state(sp->pdev);
1da177e4
LT
6232
6233 /* Setting swapper control on the NIC, for proper reset operation */
6234 if (s2io_set_swapper(sp)) {
6235 DBG_PRINT(ERR_DBG, "%s:swapper settings are wrong\n",
6236 dev->name);
6237 ret = -EAGAIN;
6238 goto set_swap_failed;
6239 }
6240
541ae68f 6241 /* Verify if the Herc works on the slot its placed into */
6242 if (sp->device_type & XFRAME_II_DEVICE) {
6243 mode = s2io_verify_pci_mode(sp);
6244 if (mode < 0) {
6245 DBG_PRINT(ERR_DBG, "%s: ", __FUNCTION__);
6246 DBG_PRINT(ERR_DBG, " Unsupported PCI bus mode\n");
6247 ret = -EBADSLT;
6248 goto set_swap_failed;
6249 }
6250 }
6251
6252 /* Not needed for Herc */
6253 if (sp->device_type & XFRAME_I_DEVICE) {
6254 /*
6255 * Fix for all "FFs" MAC address problems observed on
6256 * Alpha platforms
6257 */
6258 fix_mac_address(sp);
6259 s2io_reset(sp);
6260 }
1da177e4
LT
6261
6262 /*
1da177e4
LT
6263 * MAC address initialization.
6264 * For now only one mac address will be read and used.
6265 */
6266 bar0 = sp->bar0;
6267 val64 = RMAC_ADDR_CMD_MEM_RD | RMAC_ADDR_CMD_MEM_STROBE_NEW_CMD |
6268 RMAC_ADDR_CMD_MEM_OFFSET(0 + MAC_MAC_ADDR_START_OFFSET);
6269 writeq(val64, &bar0->rmac_addr_cmd_mem);
6270 wait_for_cmd_complete(sp);
6271
6272 tmp64 = readq(&bar0->rmac_addr_data0_mem);
6273 mac_down = (u32) tmp64;
6274 mac_up = (u32) (tmp64 >> 32);
6275
6276 memset(sp->def_mac_addr[0].mac_addr, 0, sizeof(ETH_ALEN));
6277
6278 sp->def_mac_addr[0].mac_addr[3] = (u8) (mac_up);
6279 sp->def_mac_addr[0].mac_addr[2] = (u8) (mac_up >> 8);
6280 sp->def_mac_addr[0].mac_addr[1] = (u8) (mac_up >> 16);
6281 sp->def_mac_addr[0].mac_addr[0] = (u8) (mac_up >> 24);
6282 sp->def_mac_addr[0].mac_addr[5] = (u8) (mac_down >> 16);
6283 sp->def_mac_addr[0].mac_addr[4] = (u8) (mac_down >> 24);
6284
1da177e4
LT
6285 /* Set the factory defined MAC address initially */
6286 dev->addr_len = ETH_ALEN;
6287 memcpy(dev->dev_addr, sp->def_mac_addr, ETH_ALEN);
6288
6289 /*
20346722 6290 * Initialize the tasklet status and link state flags
541ae68f 6291 * and the card state parameter
1da177e4
LT
6292 */
6293 atomic_set(&(sp->card_state), 0);
6294 sp->tasklet_status = 0;
6295 sp->link_state = 0;
6296
1da177e4
LT
6297 /* Initialize spinlocks */
6298 spin_lock_init(&sp->tx_lock);
6299#ifndef CONFIG_S2IO_NAPI
6300 spin_lock_init(&sp->put_lock);
6301#endif
7ba013ac 6302 spin_lock_init(&sp->rx_lock);
1da177e4 6303
20346722 6304 /*
6305 * SXE-002: Configure link and activity LED to init state
6306 * on driver load.
1da177e4
LT
6307 */
6308 subid = sp->pdev->subsystem_device;
6309 if ((subid & 0xFF) >= 0x07) {
6310 val64 = readq(&bar0->gpio_control);
6311 val64 |= 0x0000800000000000ULL;
6312 writeq(val64, &bar0->gpio_control);
6313 val64 = 0x0411040400000000ULL;
6314 writeq(val64, (void __iomem *) bar0 + 0x2700);
6315 val64 = readq(&bar0->gpio_control);
6316 }
6317
6318 sp->rx_csum = 1; /* Rx chksum verify enabled by default */
6319
6320 if (register_netdev(dev)) {
6321 DBG_PRINT(ERR_DBG, "Device registration failed\n");
6322 ret = -ENODEV;
6323 goto register_failed;
6324 }
6325
541ae68f 6326 if (sp->device_type & XFRAME_II_DEVICE) {
6327 DBG_PRINT(ERR_DBG, "%s: Neterion Xframe II 10GbE adapter ",
6328 dev->name);
6c1792f4 6329 DBG_PRINT(ERR_DBG, "(rev %d), Version %s",
541ae68f 6330 get_xena_rev_id(sp->pdev),
6331 s2io_driver_version);
cc6e7c44
RA
6332 switch(sp->intr_type) {
6333 case INTA:
6334 DBG_PRINT(ERR_DBG, ", Intr type INTA");
6335 break;
6336 case MSI:
6337 DBG_PRINT(ERR_DBG, ", Intr type MSI");
6338 break;
6339 case MSI_X:
6340 DBG_PRINT(ERR_DBG, ", Intr type MSI-X");
6341 break;
6342 }
776bd20f 6343
6344 DBG_PRINT(ERR_DBG, "\nCopyright(c) 2002-2005 Neterion Inc.\n");
541ae68f 6345 DBG_PRINT(ERR_DBG, "MAC ADDR: %02x:%02x:%02x:%02x:%02x:%02x\n",
6346 sp->def_mac_addr[0].mac_addr[0],
6347 sp->def_mac_addr[0].mac_addr[1],
6348 sp->def_mac_addr[0].mac_addr[2],
6349 sp->def_mac_addr[0].mac_addr[3],
6350 sp->def_mac_addr[0].mac_addr[4],
6351 sp->def_mac_addr[0].mac_addr[5]);
0b1f7ebe 6352 mode = s2io_print_pci_mode(sp);
541ae68f 6353 if (mode < 0) {
6354 DBG_PRINT(ERR_DBG, " Unsupported PCI bus mode ");
6355 ret = -EBADSLT;
6356 goto set_swap_failed;
6357 }
6358 } else {
6359 DBG_PRINT(ERR_DBG, "%s: Neterion Xframe I 10GbE adapter ",
6360 dev->name);
6c1792f4 6361 DBG_PRINT(ERR_DBG, "(rev %d), Version %s",
541ae68f 6362 get_xena_rev_id(sp->pdev),
6363 s2io_driver_version);
cc6e7c44
RA
6364 switch(sp->intr_type) {
6365 case INTA:
6366 DBG_PRINT(ERR_DBG, ", Intr type INTA");
6367 break;
6368 case MSI:
6369 DBG_PRINT(ERR_DBG, ", Intr type MSI");
6370 break;
6371 case MSI_X:
6372 DBG_PRINT(ERR_DBG, ", Intr type MSI-X");
6373 break;
6374 }
776bd20f 6375 DBG_PRINT(ERR_DBG, "\nCopyright(c) 2002-2005 Neterion Inc.\n");
541ae68f 6376 DBG_PRINT(ERR_DBG, "MAC ADDR: %02x:%02x:%02x:%02x:%02x:%02x\n",
6377 sp->def_mac_addr[0].mac_addr[0],
6378 sp->def_mac_addr[0].mac_addr[1],
6379 sp->def_mac_addr[0].mac_addr[2],
6380 sp->def_mac_addr[0].mac_addr[3],
6381 sp->def_mac_addr[0].mac_addr[4],
6382 sp->def_mac_addr[0].mac_addr[5]);
6383 }
da6971d8
AR
6384 if (sp->rxd_mode == RXD_MODE_3B)
6385 DBG_PRINT(ERR_DBG, "%s: 2-Buffer mode support has been "
6386 "enabled\n",dev->name);
6387 if (sp->rxd_mode == RXD_MODE_3A)
6388 DBG_PRINT(ERR_DBG, "%s: 3-Buffer mode support has been "
6389 "enabled\n",dev->name);
541ae68f 6390
7d3d0439
RA
6391 if (sp->lro)
6392 DBG_PRINT(ERR_DBG, "%s: Large receive offload enabled\n",
6393 dev->name);
6394
7ba013ac 6395 /* Initialize device name */
6396 strcpy(sp->name, dev->name);
541ae68f 6397 if (sp->device_type & XFRAME_II_DEVICE)
6398 strcat(sp->name, ": Neterion Xframe II 10GbE adapter");
6399 else
6400 strcat(sp->name, ": Neterion Xframe I 10GbE adapter");
7ba013ac 6401
b6e3f982 6402 /* Initialize bimodal Interrupts */
6403 sp->config.bimodal = bimodal;
6404 if (!(sp->device_type & XFRAME_II_DEVICE) && bimodal) {
6405 sp->config.bimodal = 0;
6406 DBG_PRINT(ERR_DBG,"%s:Bimodal intr not supported by Xframe I\n",
6407 dev->name);
6408 }
6409
20346722 6410 /*
6411 * Make Link state as off at this point, when the Link change
6412 * interrupt comes the state will be automatically changed to
1da177e4
LT
6413 * the right state.
6414 */
6415 netif_carrier_off(dev);
1da177e4
LT
6416
6417 return 0;
6418
6419 register_failed:
6420 set_swap_failed:
6421 iounmap(sp->bar1);
6422 bar1_remap_failed:
6423 iounmap(sp->bar0);
6424 bar0_remap_failed:
6425 mem_alloc_failed:
6426 free_shared_mem(sp);
6427 pci_disable_device(pdev);
cc6e7c44
RA
6428 if (dev_intr_type != MSI_X)
6429 pci_release_regions(pdev);
6430 else {
6431 release_mem_region(pci_resource_start(pdev, 0),
6432 pci_resource_len(pdev, 0));
6433 release_mem_region(pci_resource_start(pdev, 2),
6434 pci_resource_len(pdev, 2));
6435 }
1da177e4
LT
6436 pci_set_drvdata(pdev, NULL);
6437 free_netdev(dev);
6438
6439 return ret;
6440}
6441
6442/**
20346722 6443 * s2io_rem_nic - Free the PCI device
1da177e4 6444 * @pdev: structure containing the PCI related information of the device.
20346722 6445 * Description: This function is called by the Pci subsystem to release a
1da177e4 6446 * PCI device and free up all resource held up by the device. This could
20346722 6447 * be in response to a Hot plug event or when the driver is to be removed
1da177e4
LT
6448 * from memory.
6449 */
6450
6451static void __devexit s2io_rem_nic(struct pci_dev *pdev)
6452{
6453 struct net_device *dev =
6454 (struct net_device *) pci_get_drvdata(pdev);
6455 nic_t *sp;
6456
6457 if (dev == NULL) {
6458 DBG_PRINT(ERR_DBG, "Driver Data is NULL!!\n");
6459 return;
6460 }
6461
6462 sp = dev->priv;
6463 unregister_netdev(dev);
6464
6465 free_shared_mem(sp);
6466 iounmap(sp->bar0);
6467 iounmap(sp->bar1);
6468 pci_disable_device(pdev);
cc6e7c44
RA
6469 if (sp->intr_type != MSI_X)
6470 pci_release_regions(pdev);
6471 else {
6472 release_mem_region(pci_resource_start(pdev, 0),
6473 pci_resource_len(pdev, 0));
6474 release_mem_region(pci_resource_start(pdev, 2),
6475 pci_resource_len(pdev, 2));
6476 }
1da177e4 6477 pci_set_drvdata(pdev, NULL);
1da177e4
LT
6478 free_netdev(dev);
6479}
6480
6481/**
6482 * s2io_starter - Entry point for the driver
6483 * Description: This function is the entry point for the driver. It verifies
6484 * the module loadable parameters and initializes PCI configuration space.
6485 */
6486
6487int __init s2io_starter(void)
6488{
6489 return pci_module_init(&s2io_driver);
6490}
6491
6492/**
20346722 6493 * s2io_closer - Cleanup routine for the driver
1da177e4
LT
6494 * Description: This function is the cleanup routine for the driver. It unregist * ers the driver.
6495 */
6496
26df54bf 6497static void s2io_closer(void)
1da177e4
LT
6498{
6499 pci_unregister_driver(&s2io_driver);
6500 DBG_PRINT(INIT_DBG, "cleanup done\n");
6501}
6502
6503module_init(s2io_starter);
6504module_exit(s2io_closer);
7d3d0439
RA
6505
6506static int check_L2_lro_capable(u8 *buffer, struct iphdr **ip,
6507 struct tcphdr **tcp, RxD_t *rxdp)
6508{
6509 int ip_off;
6510 u8 l2_type = (u8)((rxdp->Control_1 >> 37) & 0x7), ip_len;
6511
6512 if (!(rxdp->Control_1 & RXD_FRAME_PROTO_TCP)) {
6513 DBG_PRINT(INIT_DBG,"%s: Non-TCP frames not supported for LRO\n",
6514 __FUNCTION__);
6515 return -1;
6516 }
6517
6518 /* TODO:
6519 * By default the VLAN field in the MAC is stripped by the card, if this
6520 * feature is turned off in rx_pa_cfg register, then the ip_off field
6521 * has to be shifted by a further 2 bytes
6522 */
6523 switch (l2_type) {
6524 case 0: /* DIX type */
6525 case 4: /* DIX type with VLAN */
6526 ip_off = HEADER_ETHERNET_II_802_3_SIZE;
6527 break;
6528 /* LLC, SNAP etc are considered non-mergeable */
6529 default:
6530 return -1;
6531 }
6532
6533 *ip = (struct iphdr *)((u8 *)buffer + ip_off);
6534 ip_len = (u8)((*ip)->ihl);
6535 ip_len <<= 2;
6536 *tcp = (struct tcphdr *)((unsigned long)*ip + ip_len);
6537
6538 return 0;
6539}
6540
6541static int check_for_socket_match(lro_t *lro, struct iphdr *ip,
6542 struct tcphdr *tcp)
6543{
6544 DBG_PRINT(INFO_DBG,"%s: Been here...\n", __FUNCTION__);
6545 if ((lro->iph->saddr != ip->saddr) || (lro->iph->daddr != ip->daddr) ||
6546 (lro->tcph->source != tcp->source) || (lro->tcph->dest != tcp->dest))
6547 return -1;
6548 return 0;
6549}
6550
6551static inline int get_l4_pyld_length(struct iphdr *ip, struct tcphdr *tcp)
6552{
6553 return(ntohs(ip->tot_len) - (ip->ihl << 2) - (tcp->doff << 2));
6554}
6555
6556static void initiate_new_session(lro_t *lro, u8 *l2h,
6557 struct iphdr *ip, struct tcphdr *tcp, u32 tcp_pyld_len)
6558{
6559 DBG_PRINT(INFO_DBG,"%s: Been here...\n", __FUNCTION__);
6560 lro->l2h = l2h;
6561 lro->iph = ip;
6562 lro->tcph = tcp;
6563 lro->tcp_next_seq = tcp_pyld_len + ntohl(tcp->seq);
6564 lro->tcp_ack = ntohl(tcp->ack_seq);
6565 lro->sg_num = 1;
6566 lro->total_len = ntohs(ip->tot_len);
6567 lro->frags_len = 0;
6568 /*
6569 * check if we saw TCP timestamp. Other consistency checks have
6570 * already been done.
6571 */
6572 if (tcp->doff == 8) {
6573 u32 *ptr;
6574 ptr = (u32 *)(tcp+1);
6575 lro->saw_ts = 1;
6576 lro->cur_tsval = *(ptr+1);
6577 lro->cur_tsecr = *(ptr+2);
6578 }
6579 lro->in_use = 1;
6580}
6581
6582static void update_L3L4_header(nic_t *sp, lro_t *lro)
6583{
6584 struct iphdr *ip = lro->iph;
6585 struct tcphdr *tcp = lro->tcph;
6586 u16 nchk;
6587 StatInfo_t *statinfo = sp->mac_control.stats_info;
6588 DBG_PRINT(INFO_DBG,"%s: Been here...\n", __FUNCTION__);
6589
6590 /* Update L3 header */
6591 ip->tot_len = htons(lro->total_len);
6592 ip->check = 0;
6593 nchk = ip_fast_csum((u8 *)lro->iph, ip->ihl);
6594 ip->check = nchk;
6595
6596 /* Update L4 header */
6597 tcp->ack_seq = lro->tcp_ack;
6598 tcp->window = lro->window;
6599
6600 /* Update tsecr field if this session has timestamps enabled */
6601 if (lro->saw_ts) {
6602 u32 *ptr = (u32 *)(tcp + 1);
6603 *(ptr+2) = lro->cur_tsecr;
6604 }
6605
6606 /* Update counters required for calculation of
6607 * average no. of packets aggregated.
6608 */
6609 statinfo->sw_stat.sum_avg_pkts_aggregated += lro->sg_num;
6610 statinfo->sw_stat.num_aggregations++;
6611}
6612
6613static void aggregate_new_rx(lro_t *lro, struct iphdr *ip,
6614 struct tcphdr *tcp, u32 l4_pyld)
6615{
6616 DBG_PRINT(INFO_DBG,"%s: Been here...\n", __FUNCTION__);
6617 lro->total_len += l4_pyld;
6618 lro->frags_len += l4_pyld;
6619 lro->tcp_next_seq += l4_pyld;
6620 lro->sg_num++;
6621
6622 /* Update ack seq no. and window ad(from this pkt) in LRO object */
6623 lro->tcp_ack = tcp->ack_seq;
6624 lro->window = tcp->window;
6625
6626 if (lro->saw_ts) {
6627 u32 *ptr;
6628 /* Update tsecr and tsval from this packet */
6629 ptr = (u32 *) (tcp + 1);
6630 lro->cur_tsval = *(ptr + 1);
6631 lro->cur_tsecr = *(ptr + 2);
6632 }
6633}
6634
6635static int verify_l3_l4_lro_capable(lro_t *l_lro, struct iphdr *ip,
6636 struct tcphdr *tcp, u32 tcp_pyld_len)
6637{
7d3d0439
RA
6638 u8 *ptr;
6639
79dc1901
AM
6640 DBG_PRINT(INFO_DBG,"%s: Been here...\n", __FUNCTION__);
6641
7d3d0439
RA
6642 if (!tcp_pyld_len) {
6643 /* Runt frame or a pure ack */
6644 return -1;
6645 }
6646
6647 if (ip->ihl != 5) /* IP has options */
6648 return -1;
6649
6650 if (tcp->urg || tcp->psh || tcp->rst || tcp->syn || tcp->fin ||
6651 !tcp->ack) {
6652 /*
6653 * Currently recognize only the ack control word and
6654 * any other control field being set would result in
6655 * flushing the LRO session
6656 */
6657 return -1;
6658 }
6659
6660 /*
6661 * Allow only one TCP timestamp option. Don't aggregate if
6662 * any other options are detected.
6663 */
6664 if (tcp->doff != 5 && tcp->doff != 8)
6665 return -1;
6666
6667 if (tcp->doff == 8) {
6668 ptr = (u8 *)(tcp + 1);
6669 while (*ptr == TCPOPT_NOP)
6670 ptr++;
6671 if (*ptr != TCPOPT_TIMESTAMP || *(ptr+1) != TCPOLEN_TIMESTAMP)
6672 return -1;
6673
6674 /* Ensure timestamp value increases monotonically */
6675 if (l_lro)
6676 if (l_lro->cur_tsval > *((u32 *)(ptr+2)))
6677 return -1;
6678
6679 /* timestamp echo reply should be non-zero */
6680 if (*((u32 *)(ptr+6)) == 0)
6681 return -1;
6682 }
6683
6684 return 0;
6685}
6686
6687static int
6688s2io_club_tcp_session(u8 *buffer, u8 **tcp, u32 *tcp_len, lro_t **lro,
6689 RxD_t *rxdp, nic_t *sp)
6690{
6691 struct iphdr *ip;
6692 struct tcphdr *tcph;
6693 int ret = 0, i;
6694
6695 if (!(ret = check_L2_lro_capable(buffer, &ip, (struct tcphdr **)tcp,
6696 rxdp))) {
6697 DBG_PRINT(INFO_DBG,"IP Saddr: %x Daddr: %x\n",
6698 ip->saddr, ip->daddr);
6699 } else {
6700 return ret;
6701 }
6702
6703 tcph = (struct tcphdr *)*tcp;
6704 *tcp_len = get_l4_pyld_length(ip, tcph);
6705 for (i=0; i<MAX_LRO_SESSIONS; i++) {
6706 lro_t *l_lro = &sp->lro0_n[i];
6707 if (l_lro->in_use) {
6708 if (check_for_socket_match(l_lro, ip, tcph))
6709 continue;
6710 /* Sock pair matched */
6711 *lro = l_lro;
6712
6713 if ((*lro)->tcp_next_seq != ntohl(tcph->seq)) {
6714 DBG_PRINT(INFO_DBG, "%s:Out of order. expected "
6715 "0x%x, actual 0x%x\n", __FUNCTION__,
6716 (*lro)->tcp_next_seq,
6717 ntohl(tcph->seq));
6718
6719 sp->mac_control.stats_info->
6720 sw_stat.outof_sequence_pkts++;
6721 ret = 2;
6722 break;
6723 }
6724
6725 if (!verify_l3_l4_lro_capable(l_lro, ip, tcph,*tcp_len))
6726 ret = 1; /* Aggregate */
6727 else
6728 ret = 2; /* Flush both */
6729 break;
6730 }
6731 }
6732
6733 if (ret == 0) {
6734 /* Before searching for available LRO objects,
6735 * check if the pkt is L3/L4 aggregatable. If not
6736 * don't create new LRO session. Just send this
6737 * packet up.
6738 */
6739 if (verify_l3_l4_lro_capable(NULL, ip, tcph, *tcp_len)) {
6740 return 5;
6741 }
6742
6743 for (i=0; i<MAX_LRO_SESSIONS; i++) {
6744 lro_t *l_lro = &sp->lro0_n[i];
6745 if (!(l_lro->in_use)) {
6746 *lro = l_lro;
6747 ret = 3; /* Begin anew */
6748 break;
6749 }
6750 }
6751 }
6752
6753 if (ret == 0) { /* sessions exceeded */
6754 DBG_PRINT(INFO_DBG,"%s:All LRO sessions already in use\n",
6755 __FUNCTION__);
6756 *lro = NULL;
6757 return ret;
6758 }
6759
6760 switch (ret) {
6761 case 3:
6762 initiate_new_session(*lro, buffer, ip, tcph, *tcp_len);
6763 break;
6764 case 2:
6765 update_L3L4_header(sp, *lro);
6766 break;
6767 case 1:
6768 aggregate_new_rx(*lro, ip, tcph, *tcp_len);
6769 if ((*lro)->sg_num == sp->lro_max_aggr_per_sess) {
6770 update_L3L4_header(sp, *lro);
6771 ret = 4; /* Flush the LRO */
6772 }
6773 break;
6774 default:
6775 DBG_PRINT(ERR_DBG,"%s:Dont know, can't say!!\n",
6776 __FUNCTION__);
6777 break;
6778 }
6779
6780 return ret;
6781}
6782
6783static void clear_lro_session(lro_t *lro)
6784{
6785 static u16 lro_struct_size = sizeof(lro_t);
6786
6787 memset(lro, 0, lro_struct_size);
6788}
6789
6790static void queue_rx_frame(struct sk_buff *skb)
6791{
6792 struct net_device *dev = skb->dev;
6793
6794 skb->protocol = eth_type_trans(skb, dev);
6795#ifdef CONFIG_S2IO_NAPI
6796 netif_receive_skb(skb);
6797#else
6798 netif_rx(skb);
6799#endif
6800}
6801
6802static void lro_append_pkt(nic_t *sp, lro_t *lro, struct sk_buff *skb,
6803 u32 tcp_len)
6804{
6805 struct sk_buff *tmp, *first = lro->parent;
6806
6807 first->len += tcp_len;
6808 first->data_len = lro->frags_len;
6809 skb_pull(skb, (skb->len - tcp_len));
6810 if ((tmp = skb_shinfo(first)->frag_list)) {
6811 while (tmp->next)
6812 tmp = tmp->next;
6813 tmp->next = skb;
6814 }
6815 else
6816 skb_shinfo(first)->frag_list = skb;
6817 sp->mac_control.stats_info->sw_stat.clubbed_frms_cnt++;
6818 return;
6819}
This page took 0.792393 seconds and 5 git commands to generate.