2 * Copyright (c) 2007 Mellanox Technologies. All rights reserved.
4 * This software is available to you under a choice of one of two
5 * licenses. You may choose to be licensed under the terms of the GNU
6 * General Public License (GPL) Version 2, available from the file
7 * COPYING in the main directory of this source tree, or the
8 * OpenIB.org BSD license below:
10 * Redistribution and use in source and binary forms, with or
11 * without modification, are permitted provided that the following
14 * - Redistributions of source code must retain the above
15 * copyright notice, this list of conditions and the following
18 * - Redistributions in binary form must reproduce the above
19 * copyright notice, this list of conditions and the following
20 * disclaimer in the documentation and/or other materials
21 * provided with the distribution.
23 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
24 * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
25 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
26 * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
27 * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
28 * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
29 * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
34 #include <linux/cpumask.h>
35 #include <linux/module.h>
36 #include <linux/delay.h>
37 #include <linux/netdevice.h>
38 #include <linux/slab.h>
40 #include <linux/mlx4/driver.h>
41 #include <linux/mlx4/device.h>
42 #include <linux/mlx4/cmd.h>
46 MODULE_AUTHOR("Liran Liss, Yevgeny Petrilin");
47 MODULE_DESCRIPTION("Mellanox ConnectX HCA Ethernet driver");
48 MODULE_LICENSE("Dual BSD/GPL");
49 MODULE_VERSION(DRV_VERSION
" ("DRV_RELDATE
")");
51 static const char mlx4_en_version
[] =
52 DRV_NAME
": Mellanox ConnectX HCA Ethernet driver v"
53 DRV_VERSION
" (" DRV_RELDATE
")\n";
55 #define MLX4_EN_PARM_INT(X, def_val, desc) \
56 static unsigned int X = def_val;\
57 module_param(X , uint, 0444); \
58 MODULE_PARM_DESC(X, desc);
62 * Device scope module parameters
65 /* Enable RSS UDP traffic */
66 MLX4_EN_PARM_INT(udp_rss
, 1,
67 "Enable RSS for incomming UDP traffic or disabled (0)");
69 /* Priority pausing */
70 MLX4_EN_PARM_INT(pfctx
, 0, "Priority based Flow Control policy on TX[7:0]."
71 " Per priority bit mask");
72 MLX4_EN_PARM_INT(pfcrx
, 0, "Priority based Flow Control policy on RX[7:0]."
73 " Per priority bit mask");
75 int en_print(const char *level
, const struct mlx4_en_priv
*priv
,
76 const char *format
, ...)
82 va_start(args
, format
);
87 i
= printk("%s%s: %s: %pV",
88 level
, DRV_NAME
, priv
->dev
->name
, &vaf
);
90 i
= printk("%s%s: %s: Port %d: %pV",
91 level
, DRV_NAME
, dev_name(&priv
->mdev
->pdev
->dev
),
98 static int mlx4_en_get_profile(struct mlx4_en_dev
*mdev
)
100 struct mlx4_en_profile
*params
= &mdev
->profile
;
103 params
->udp_rss
= udp_rss
;
104 params
->num_tx_rings_p_up
= min_t(int, num_online_cpus(),
105 MLX4_EN_MAX_TX_RING_P_UP
);
106 if (params
->udp_rss
&& !(mdev
->dev
->caps
.flags
107 & MLX4_DEV_CAP_FLAG_UDP_RSS
)) {
108 mlx4_warn(mdev
, "UDP RSS is not supported on this device.\n");
111 for (i
= 1; i
<= MLX4_MAX_PORTS
; i
++) {
112 params
->prof
[i
].rx_pause
= 1;
113 params
->prof
[i
].rx_ppp
= pfcrx
;
114 params
->prof
[i
].tx_pause
= 1;
115 params
->prof
[i
].tx_ppp
= pfctx
;
116 params
->prof
[i
].tx_ring_size
= MLX4_EN_DEF_TX_RING_SIZE
;
117 params
->prof
[i
].rx_ring_size
= MLX4_EN_DEF_RX_RING_SIZE
;
118 params
->prof
[i
].tx_ring_num
= params
->num_tx_rings_p_up
*
120 params
->prof
[i
].rss_rings
= 0;
126 static void *mlx4_en_get_netdev(struct mlx4_dev
*dev
, void *ctx
, u8 port
)
128 struct mlx4_en_dev
*endev
= ctx
;
130 return endev
->pndev
[port
];
133 static void mlx4_en_event(struct mlx4_dev
*dev
, void *endev_ptr
,
134 enum mlx4_dev_event event
, unsigned long port
)
136 struct mlx4_en_dev
*mdev
= (struct mlx4_en_dev
*) endev_ptr
;
137 struct mlx4_en_priv
*priv
;
140 case MLX4_DEV_EVENT_PORT_UP
:
141 case MLX4_DEV_EVENT_PORT_DOWN
:
142 if (!mdev
->pndev
[port
])
144 priv
= netdev_priv(mdev
->pndev
[port
]);
145 /* To prevent races, we poll the link state in a separate
146 task rather than changing it here */
147 priv
->link_state
= event
;
148 queue_work(mdev
->workqueue
, &priv
->linkstate_task
);
151 case MLX4_DEV_EVENT_CATASTROPHIC_ERROR
:
152 mlx4_err(mdev
, "Internal error detected, restarting device\n");
156 if (port
< 1 || port
> dev
->caps
.num_ports
||
159 mlx4_warn(mdev
, "Unhandled event %d for port %d\n", event
,
164 static void mlx4_en_remove(struct mlx4_dev
*dev
, void *endev_ptr
)
166 struct mlx4_en_dev
*mdev
= endev_ptr
;
169 mutex_lock(&mdev
->state_lock
);
170 mdev
->device_up
= false;
171 mutex_unlock(&mdev
->state_lock
);
173 mlx4_foreach_port(i
, dev
, MLX4_PORT_TYPE_ETH
)
175 mlx4_en_destroy_netdev(mdev
->pndev
[i
]);
177 flush_workqueue(mdev
->workqueue
);
178 destroy_workqueue(mdev
->workqueue
);
179 mlx4_mr_free(dev
, &mdev
->mr
);
180 iounmap(mdev
->uar_map
);
181 mlx4_uar_free(dev
, &mdev
->priv_uar
);
182 mlx4_pd_free(dev
, mdev
->priv_pdn
);
186 static void *mlx4_en_add(struct mlx4_dev
*dev
)
188 struct mlx4_en_dev
*mdev
;
192 printk_once(KERN_INFO
"%s", mlx4_en_version
);
194 mdev
= kzalloc(sizeof *mdev
, GFP_KERNEL
);
196 dev_err(&dev
->pdev
->dev
, "Device struct alloc failed, "
202 if (mlx4_pd_alloc(dev
, &mdev
->priv_pdn
))
205 if (mlx4_uar_alloc(dev
, &mdev
->priv_uar
))
208 mdev
->uar_map
= ioremap((phys_addr_t
) mdev
->priv_uar
.pfn
<< PAGE_SHIFT
,
212 spin_lock_init(&mdev
->uar_lock
);
215 mdev
->dma_device
= &(dev
->pdev
->dev
);
216 mdev
->pdev
= dev
->pdev
;
217 mdev
->device_up
= false;
219 mdev
->LSO_support
= !!(dev
->caps
.flags
& (1 << 15));
220 if (!mdev
->LSO_support
)
221 mlx4_warn(mdev
, "LSO not supported, please upgrade to later "
222 "FW version to enable LSO\n");
224 if (mlx4_mr_alloc(mdev
->dev
, mdev
->priv_pdn
, 0, ~0ull,
225 MLX4_PERM_LOCAL_WRITE
| MLX4_PERM_LOCAL_READ
,
227 mlx4_err(mdev
, "Failed allocating memory region\n");
230 if (mlx4_mr_enable(mdev
->dev
, &mdev
->mr
)) {
231 mlx4_err(mdev
, "Failed enabling memory region\n");
235 /* Build device profile according to supplied module parameters */
236 err
= mlx4_en_get_profile(mdev
);
238 mlx4_err(mdev
, "Bad module parameters, aborting.\n");
242 /* Configure which ports to start according to module parameters */
244 mlx4_foreach_port(i
, dev
, MLX4_PORT_TYPE_ETH
)
247 mlx4_foreach_port(i
, dev
, MLX4_PORT_TYPE_ETH
) {
248 if (!dev
->caps
.comp_pool
) {
249 mdev
->profile
.prof
[i
].rx_ring_num
=
250 rounddown_pow_of_two(max_t(int, MIN_RX_RINGS
,
252 dev
->caps
.num_comp_vectors
,
255 mdev
->profile
.prof
[i
].rx_ring_num
= rounddown_pow_of_two(
256 min_t(int, dev
->caps
.comp_pool
/
257 dev
->caps
.num_ports
- 1 , MAX_MSIX_P_PORT
- 1));
261 /* Create our own workqueue for reset/multicast tasks
262 * Note: we cannot use the shared workqueue because of deadlocks caused
263 * by the rtnl lock */
264 mdev
->workqueue
= create_singlethread_workqueue("mlx4_en");
265 if (!mdev
->workqueue
) {
270 /* At this stage all non-port specific tasks are complete:
271 * mark the card state as up */
272 mutex_init(&mdev
->state_lock
);
273 mdev
->device_up
= true;
277 /* Create a netdev for each port */
278 mlx4_foreach_port(i
, dev
, MLX4_PORT_TYPE_ETH
) {
279 mlx4_info(mdev
, "Activating port:%d\n", i
);
280 if (mlx4_en_init_netdev(mdev
, i
, &mdev
->profile
.prof
[i
]))
281 mdev
->pndev
[i
] = NULL
;
286 mlx4_mr_free(dev
, &mdev
->mr
);
289 iounmap(mdev
->uar_map
);
291 mlx4_uar_free(dev
, &mdev
->priv_uar
);
293 mlx4_pd_free(dev
, mdev
->priv_pdn
);
300 static struct mlx4_interface mlx4_en_interface
= {
302 .remove
= mlx4_en_remove
,
303 .event
= mlx4_en_event
,
304 .get_dev
= mlx4_en_get_netdev
,
305 .protocol
= MLX4_PROT_ETH
,
308 static int __init
mlx4_en_init(void)
310 return mlx4_register_interface(&mlx4_en_interface
);
313 static void __exit
mlx4_en_cleanup(void)
315 mlx4_unregister_interface(&mlx4_en_interface
);
318 module_init(mlx4_en_init
);
319 module_exit(mlx4_en_cleanup
);