IB/rdmavt,hfi1,qib: Fix memory leak
[deliverable/linux.git] / drivers / infiniband / sw / rdmavt / vt.c
CommitLineData
0194621b 1/*
fe314195 2 * Copyright(c) 2016 Intel Corporation.
0194621b
DD
3 *
4 * This file is provided under a dual BSD/GPLv2 license. When using or
5 * redistributing this file, you may do so under either license.
6 *
7 * GPL LICENSE SUMMARY
8 *
9 * This program is free software; you can redistribute it and/or modify
10 * it under the terms of version 2 of the GNU General Public License as
11 * published by the Free Software Foundation.
12 *
13 * This program is distributed in the hope that it will be useful, but
14 * WITHOUT ANY WARRANTY; without even the implied warranty of
15 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
16 * General Public License for more details.
17 *
18 * BSD LICENSE
19 *
20 * Redistribution and use in source and binary forms, with or without
21 * modification, are permitted provided that the following conditions
22 * are met:
23 *
24 * - Redistributions of source code must retain the above copyright
25 * notice, this list of conditions and the following disclaimer.
26 * - Redistributions in binary form must reproduce the above copyright
27 * notice, this list of conditions and the following disclaimer in
28 * the documentation and/or other materials provided with the
29 * distribution.
30 * - Neither the name of Intel Corporation nor the names of its
31 * contributors may be used to endorse or promote products derived
32 * from this software without specific prior written permission.
33 *
34 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
35 * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
36 * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
37 * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
38 * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
39 * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
40 * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
41 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
42 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
43 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
44 * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
45 *
46 */
47
48#include <linux/module.h>
49#include <linux/kernel.h>
50#include "vt.h"
81ba39a8 51#include "trace.h"
0194621b 52
182285d0
DD
53#define RVT_UVERBS_ABI_VERSION 2
54
0194621b
DD
55MODULE_LICENSE("Dual BSD/GPL");
56MODULE_DESCRIPTION("RDMA Verbs Transport Library");
57
58static int rvt_init(void)
59{
90793f71
DD
60 /*
61 * rdmavt does not need to do anything special when it starts up. All it
62 * needs to do is sit and wait until a driver attempts registration.
63 */
0194621b
DD
64 return 0;
65}
66module_init(rvt_init);
67
68static void rvt_cleanup(void)
69{
90793f71
DD
70 /*
71 * Nothing to do at exit time either. The module won't be able to be
72 * removed until all drivers are gone which means all the dev structs
73 * are gone so there is really nothing to do.
74 */
0194621b
DD
75}
76module_exit(rvt_cleanup);
77
90793f71
DD
78/**
79 * rvt_alloc_device - allocate rdi
80 * @size: how big of a structure to allocate
81 * @nports: number of ports to allocate array slots for
82 *
83 * Use IB core device alloc to allocate space for the rdi which is assumed to be
84 * inside of the ib_device. Any extra space that drivers require should be
85 * included in size.
86 *
87 * We also allocate a port array based on the number of ports.
88 *
89 * Return: pointer to allocated rdi
90 */
ff6acd69
DD
91struct rvt_dev_info *rvt_alloc_device(size_t size, int nports)
92{
93 struct rvt_dev_info *rdi = ERR_PTR(-ENOMEM);
94
95 rdi = (struct rvt_dev_info *)ib_alloc_device(size);
96 if (!rdi)
97 return rdi;
98
99 rdi->ports = kcalloc(nports,
100 sizeof(struct rvt_ibport **),
101 GFP_KERNEL);
102 if (!rdi->ports)
103 ib_dealloc_device(&rdi->ibdev);
104
105 return rdi;
106}
107EXPORT_SYMBOL(rvt_alloc_device);
108
ea0e4ce3
JJ
109/**
110 * rvt_dealloc_device - deallocate rdi
111 * @rdi: structure to free
112 *
113 * Free a structure allocated with rvt_alloc_device()
114 */
115void rvt_dealloc_device(struct rvt_dev_info *rdi)
116{
117 kfree(rdi->ports);
118 ib_dealloc_device(&rdi->ibdev);
119}
120EXPORT_SYMBOL(rvt_dealloc_device);
121
19ef1edd
DD
122static int rvt_query_device(struct ib_device *ibdev,
123 struct ib_device_attr *props,
124 struct ib_udata *uhw)
125{
feaeb6e2
HC
126 struct rvt_dev_info *rdi = ib_to_rvt(ibdev);
127
128 if (uhw->inlen || uhw->outlen)
129 return -EINVAL;
19ef1edd 130 /*
feaeb6e2 131 * Return rvt_dev_info.dparms.props contents
19ef1edd 132 */
feaeb6e2
HC
133 *props = rdi->dparms.props;
134 return 0;
19ef1edd
DD
135}
136
137static int rvt_modify_device(struct ib_device *device,
138 int device_modify_mask,
139 struct ib_device_modify *device_modify)
140{
141 /*
90793f71
DD
142 * There is currently no need to supply this based on qib and hfi1.
143 * Future drivers may need to implement this though.
19ef1edd
DD
144 */
145
19ef1edd
DD
146 return -EOPNOTSUPP;
147}
148
765525c1
DD
149/**
150 * rvt_query_port: Passes the query port call to the driver
151 * @ibdev: Verbs IB dev
f1badc71 152 * @port_num: port number, 1 based from ib core
765525c1
DD
153 * @props: structure to hold returned properties
154 *
90793f71 155 * Return: 0 on success
765525c1 156 */
f1badc71 157static int rvt_query_port(struct ib_device *ibdev, u8 port_num,
765525c1
DD
158 struct ib_port_attr *props)
159{
61a650c1
HC
160 struct rvt_dev_info *rdi = ib_to_rvt(ibdev);
161 struct rvt_ibport *rvp;
162 int port_index = ibport_num_to_idx(ibdev, port_num);
163
164 if (port_index < 0)
f1badc71
DD
165 return -EINVAL;
166
61a650c1
HC
167 rvp = rdi->ports[port_index];
168 memset(props, 0, sizeof(*props));
169 props->sm_lid = rvp->sm_lid;
170 props->sm_sl = rvp->sm_sl;
171 props->port_cap_flags = rvp->port_cap_flags;
172 props->max_msg_sz = 0x80000000;
173 props->pkey_tbl_len = rvt_get_npkeys(rdi);
174 props->bad_pkey_cntr = rvp->pkey_violations;
175 props->qkey_viol_cntr = rvp->qkey_violations;
176 props->subnet_timeout = rvp->subnet_timeout;
177 props->init_type_reply = 0;
178
179 /* Populate the remaining ib_port_attr elements */
180 return rdi->driver_f.query_port_state(rdi, port_num, props);
765525c1
DD
181}
182
183/**
184 * rvt_modify_port
185 * @ibdev: Verbs IB dev
f1badc71 186 * @port_num: Port number, 1 based from ib core
765525c1
DD
187 * @port_modify_mask: How to change the port
188 * @props: Structure to fill in
189 *
90793f71 190 * Return: 0 on success
765525c1 191 */
f1badc71 192static int rvt_modify_port(struct ib_device *ibdev, u8 port_num,
765525c1
DD
193 int port_modify_mask, struct ib_port_modify *props)
194{
61a650c1
HC
195 struct rvt_dev_info *rdi = ib_to_rvt(ibdev);
196 struct rvt_ibport *rvp;
197 int ret = 0;
198 int port_index = ibport_num_to_idx(ibdev, port_num);
199
200 if (port_index < 0)
f1badc71
DD
201 return -EINVAL;
202
61a650c1
HC
203 rvp = rdi->ports[port_index];
204 rvp->port_cap_flags |= props->set_port_cap_mask;
205 rvp->port_cap_flags &= ~props->clr_port_cap_mask;
206
207 if (props->set_port_cap_mask || props->clr_port_cap_mask)
208 rdi->driver_f.cap_mask_chg(rdi, port_num);
209 if (port_modify_mask & IB_PORT_SHUTDOWN)
210 ret = rdi->driver_f.shut_down_port(rdi, port_num);
211 if (port_modify_mask & IB_PORT_RESET_QKEY_CNTR)
212 rvp->qkey_violations = 0;
213
214 return ret;
765525c1
DD
215}
216
30588643
DD
217/**
218 * rvt_query_pkey - Return a pkey from the table at a given index
219 * @ibdev: Verbs IB dev
f1badc71 220 * @port_num: Port number, 1 based from ib core
30588643
DD
221 * @intex: Index into pkey table
222 *
90793f71 223 * Return: 0 on failure pkey otherwise
30588643 224 */
f1badc71 225static int rvt_query_pkey(struct ib_device *ibdev, u8 port_num, u16 index,
30588643
DD
226 u16 *pkey)
227{
228 /*
229 * Driver will be responsible for keeping rvt_dev_info.pkey_table up to
230 * date. This function will just return that value. There is no need to
231 * lock, if a stale value is read and sent to the user so be it there is
232 * no way to protect against that anyway.
233 */
38ce2c6f
DD
234 struct rvt_dev_info *rdi = ib_to_rvt(ibdev);
235 int port_index;
236
f1badc71
DD
237 port_index = ibport_num_to_idx(ibdev, port_num);
238 if (port_index < 0)
38ce2c6f
DD
239 return -EINVAL;
240
f1badc71 241 if (index >= rvt_get_npkeys(rdi))
38ce2c6f
DD
242 return -EINVAL;
243
244 *pkey = rvt_get_pkey(rdi, port_index, index);
30588643
DD
245 return 0;
246}
247
2d092e11
DD
248/**
249 * rvt_query_gid - Return a gid from the table
250 * @ibdev: Verbs IB dev
f1badc71 251 * @port_num: Port number, 1 based from ib core
2d092e11
DD
252 * @index: = Index in table
253 * @gid: Gid to return
254 *
90793f71 255 * Return: 0 on success
2d092e11 256 */
f1badc71 257static int rvt_query_gid(struct ib_device *ibdev, u8 port_num,
1f024992 258 int guid_index, union ib_gid *gid)
2d092e11 259{
1f024992
DD
260 struct rvt_dev_info *rdi;
261 struct rvt_ibport *rvp;
262 int port_index;
263
2d092e11
DD
264 /*
265 * Driver is responsible for updating the guid table. Which will be used
266 * to craft the return value. This will work similar to how query_pkey()
267 * is being done.
268 */
1f024992
DD
269 port_index = ibport_num_to_idx(ibdev, port_num);
270 if (port_index < 0)
f1badc71 271 return -EINVAL;
2d092e11 272
1f024992
DD
273 rdi = ib_to_rvt(ibdev);
274 rvp = rdi->ports[port_index];
275
276 gid->global.subnet_prefix = rvp->gid_prefix;
277
278 return rdi->driver_f.get_guid_be(rdi, rvp, guid_index,
279 &gid->global.interface_id);
2d092e11
DD
280}
281
6c43cf4b
HC
282struct rvt_ucontext {
283 struct ib_ucontext ibucontext;
284};
285
286static inline struct rvt_ucontext *to_iucontext(struct ib_ucontext
287 *ibucontext)
288{
289 return container_of(ibucontext, struct rvt_ucontext, ibucontext);
290}
291
c4ed7d8b
DD
292/**
293 * rvt_alloc_ucontext - Allocate a user context
294 * @ibdev: Vers IB dev
295 * @data: User data allocated
296 */
297static struct ib_ucontext *rvt_alloc_ucontext(struct ib_device *ibdev,
298 struct ib_udata *udata)
299{
6c43cf4b
HC
300 struct rvt_ucontext *context;
301
302 context = kmalloc(sizeof(*context), GFP_KERNEL);
303 if (!context)
304 return ERR_PTR(-ENOMEM);
305 return &context->ibucontext;
c4ed7d8b
DD
306}
307
308/**
309 *rvt_dealloc_ucontext - Free a user context
310 *@context - Free this
311 */
312static int rvt_dealloc_ucontext(struct ib_ucontext *context)
313{
6c43cf4b
HC
314 kfree(to_iucontext(context));
315 return 0;
c4ed7d8b
DD
316}
317
e6a8818a
DD
318static int rvt_get_port_immutable(struct ib_device *ibdev, u8 port_num,
319 struct ib_port_immutable *immutable)
320{
61a650c1
HC
321 struct rvt_dev_info *rdi = ib_to_rvt(ibdev);
322 struct ib_port_attr attr;
323 int err, port_index;
324
325 port_index = ibport_num_to_idx(ibdev, port_num);
326 if (port_index < 0)
327 return -EINVAL;
328
329 err = rvt_query_port(ibdev, port_num, &attr);
330 if (err)
331 return err;
332
333 immutable->pkey_tbl_len = attr.pkey_tbl_len;
334 immutable->gid_tbl_len = attr.gid_tbl_len;
335 immutable->core_cap_flags = rdi->dparms.core_cap_flags;
336 immutable->max_mad_size = rdi->dparms.max_mad_size;
337
338 return 0;
e6a8818a
DD
339}
340
1348d706
DD
341enum {
342 MISC,
343 QUERY_DEVICE,
344 MODIFY_DEVICE,
345 QUERY_PORT,
346 MODIFY_PORT,
347 QUERY_PKEY,
348 QUERY_GID,
349 ALLOC_UCONTEXT,
350 DEALLOC_UCONTEXT,
351 GET_PORT_IMMUTABLE,
352 CREATE_QP,
353 MODIFY_QP,
354 DESTROY_QP,
355 QUERY_QP,
356 POST_SEND,
357 POST_RECV,
358 POST_SRQ_RECV,
359 CREATE_AH,
360 DESTROY_AH,
361 MODIFY_AH,
362 QUERY_AH,
363 CREATE_SRQ,
364 MODIFY_SRQ,
365 DESTROY_SRQ,
366 QUERY_SRQ,
367 ATTACH_MCAST,
368 DETACH_MCAST,
369 GET_DMA_MR,
370 REG_USER_MR,
371 DEREG_MR,
372 ALLOC_MR,
373 ALLOC_FMR,
374 MAP_PHYS_FMR,
375 UNMAP_FMR,
376 DEALLOC_FMR,
377 MMAP,
378 CREATE_CQ,
379 DESTROY_CQ,
380 POLL_CQ,
381 REQ_NOTFIY_CQ,
382 RESIZE_CQ,
383 ALLOC_PD,
384 DEALLOC_PD,
385 _VERB_IDX_MAX /* Must always be last! */
386};
387
388static inline int check_driver_override(struct rvt_dev_info *rdi,
389 size_t offset, void *func)
390{
391 if (!*(void **)((void *)&rdi->ibdev + offset)) {
392 *(void **)((void *)&rdi->ibdev + offset) = func;
393 return 0;
394 }
395
396 return 1;
397}
398
9debaaec 399static noinline int check_support(struct rvt_dev_info *rdi, int verb)
1348d706
DD
400{
401 switch (verb) {
402 case MISC:
403 /*
404 * These functions are not part of verbs specifically but are
405 * required for rdmavt to function.
406 */
407 if ((!rdi->driver_f.port_callback) ||
408 (!rdi->driver_f.get_card_name) ||
409 (!rdi->driver_f.get_pci_dev))
410 return -EINVAL;
411 break;
412
413 case QUERY_DEVICE:
414 check_driver_override(rdi, offsetof(struct ib_device,
415 query_device),
416 rvt_query_device);
417 break;
418
419 case MODIFY_DEVICE:
420 /*
421 * rdmavt does not support modify device currently drivers must
422 * provide.
423 */
424 if (!check_driver_override(rdi, offsetof(struct ib_device,
425 modify_device),
426 rvt_modify_device))
427 return -EOPNOTSUPP;
428 break;
429
430 case QUERY_PORT:
431 if (!check_driver_override(rdi, offsetof(struct ib_device,
432 query_port),
433 rvt_query_port))
434 if (!rdi->driver_f.query_port_state)
435 return -EINVAL;
436 break;
437
438 case MODIFY_PORT:
439 if (!check_driver_override(rdi, offsetof(struct ib_device,
440 modify_port),
441 rvt_modify_port))
442 if (!rdi->driver_f.cap_mask_chg ||
443 !rdi->driver_f.shut_down_port)
444 return -EINVAL;
445 break;
446
447 case QUERY_PKEY:
448 check_driver_override(rdi, offsetof(struct ib_device,
449 query_pkey),
450 rvt_query_pkey);
451 break;
452
453 case QUERY_GID:
454 if (!check_driver_override(rdi, offsetof(struct ib_device,
455 query_gid),
456 rvt_query_gid))
457 if (!rdi->driver_f.get_guid_be)
458 return -EINVAL;
459 break;
460
461 case ALLOC_UCONTEXT:
462 check_driver_override(rdi, offsetof(struct ib_device,
463 alloc_ucontext),
464 rvt_alloc_ucontext);
465 break;
466
467 case DEALLOC_UCONTEXT:
468 check_driver_override(rdi, offsetof(struct ib_device,
469 dealloc_ucontext),
470 rvt_dealloc_ucontext);
471 break;
472
473 case GET_PORT_IMMUTABLE:
474 check_driver_override(rdi, offsetof(struct ib_device,
475 get_port_immutable),
476 rvt_get_port_immutable);
477 break;
478
479 case CREATE_QP:
480 if (!check_driver_override(rdi, offsetof(struct ib_device,
481 create_qp),
482 rvt_create_qp))
483 if (!rdi->driver_f.qp_priv_alloc ||
484 !rdi->driver_f.qp_priv_free ||
485 !rdi->driver_f.notify_qp_reset ||
486 !rdi->driver_f.flush_qp_waiters ||
487 !rdi->driver_f.stop_send_queue ||
488 !rdi->driver_f.quiesce_qp)
489 return -EINVAL;
490 break;
491
492 case MODIFY_QP:
493 if (!check_driver_override(rdi, offsetof(struct ib_device,
494 modify_qp),
495 rvt_modify_qp))
496 if (!rdi->driver_f.notify_qp_reset ||
497 !rdi->driver_f.schedule_send ||
498 !rdi->driver_f.get_pmtu_from_attr ||
499 !rdi->driver_f.flush_qp_waiters ||
500 !rdi->driver_f.stop_send_queue ||
501 !rdi->driver_f.quiesce_qp ||
502 !rdi->driver_f.notify_error_qp ||
503 !rdi->driver_f.mtu_from_qp ||
504 !rdi->driver_f.mtu_to_path_mtu ||
505 !rdi->driver_f.shut_down_port ||
506 !rdi->driver_f.cap_mask_chg)
507 return -EINVAL;
508 break;
509
510 case DESTROY_QP:
511 if (!check_driver_override(rdi, offsetof(struct ib_device,
512 destroy_qp),
513 rvt_destroy_qp))
514 if (!rdi->driver_f.qp_priv_free ||
515 !rdi->driver_f.notify_qp_reset ||
516 !rdi->driver_f.flush_qp_waiters ||
517 !rdi->driver_f.stop_send_queue ||
518 !rdi->driver_f.quiesce_qp)
519 return -EINVAL;
520 break;
521
522 case QUERY_QP:
523 check_driver_override(rdi, offsetof(struct ib_device,
524 query_qp),
525 rvt_query_qp);
526 break;
527
528 case POST_SEND:
529 if (!check_driver_override(rdi, offsetof(struct ib_device,
530 post_send),
531 rvt_post_send))
532 if (!rdi->driver_f.schedule_send ||
533 !rdi->driver_f.do_send)
534 return -EINVAL;
535 break;
536
537 case POST_RECV:
538 check_driver_override(rdi, offsetof(struct ib_device,
539 post_recv),
540 rvt_post_recv);
541 break;
542 case POST_SRQ_RECV:
543 check_driver_override(rdi, offsetof(struct ib_device,
544 post_srq_recv),
545 rvt_post_srq_recv);
546 break;
547
548 case CREATE_AH:
549 check_driver_override(rdi, offsetof(struct ib_device,
550 create_ah),
551 rvt_create_ah);
552 break;
553
554 case DESTROY_AH:
555 check_driver_override(rdi, offsetof(struct ib_device,
556 destroy_ah),
557 rvt_destroy_ah);
558 break;
559
560 case MODIFY_AH:
561 check_driver_override(rdi, offsetof(struct ib_device,
562 modify_ah),
563 rvt_modify_ah);
564 break;
565
566 case QUERY_AH:
567 check_driver_override(rdi, offsetof(struct ib_device,
568 query_ah),
569 rvt_query_ah);
570 break;
571
572 case CREATE_SRQ:
573 check_driver_override(rdi, offsetof(struct ib_device,
574 create_srq),
575 rvt_create_srq);
576 break;
577
578 case MODIFY_SRQ:
579 check_driver_override(rdi, offsetof(struct ib_device,
580 modify_srq),
581 rvt_modify_srq);
582 break;
583
584 case DESTROY_SRQ:
585 check_driver_override(rdi, offsetof(struct ib_device,
586 destroy_srq),
587 rvt_destroy_srq);
588 break;
589
590 case QUERY_SRQ:
591 check_driver_override(rdi, offsetof(struct ib_device,
592 query_srq),
593 rvt_query_srq);
594 break;
595
596 case ATTACH_MCAST:
597 check_driver_override(rdi, offsetof(struct ib_device,
598 attach_mcast),
599 rvt_attach_mcast);
600 break;
601
602 case DETACH_MCAST:
603 check_driver_override(rdi, offsetof(struct ib_device,
604 detach_mcast),
605 rvt_detach_mcast);
606 break;
607
608 case GET_DMA_MR:
609 check_driver_override(rdi, offsetof(struct ib_device,
610 get_dma_mr),
611 rvt_get_dma_mr);
612 break;
613
614 case REG_USER_MR:
615 check_driver_override(rdi, offsetof(struct ib_device,
616 reg_user_mr),
617 rvt_reg_user_mr);
618 break;
619
620 case DEREG_MR:
621 check_driver_override(rdi, offsetof(struct ib_device,
622 dereg_mr),
623 rvt_dereg_mr);
624 break;
625
626 case ALLOC_FMR:
627 check_driver_override(rdi, offsetof(struct ib_device,
628 alloc_fmr),
629 rvt_alloc_fmr);
630 break;
631
632 case ALLOC_MR:
633 check_driver_override(rdi, offsetof(struct ib_device,
634 alloc_mr),
635 rvt_alloc_mr);
636 break;
637
638 case MAP_PHYS_FMR:
639 check_driver_override(rdi, offsetof(struct ib_device,
640 map_phys_fmr),
641 rvt_map_phys_fmr);
642 break;
643
644 case UNMAP_FMR:
645 check_driver_override(rdi, offsetof(struct ib_device,
646 unmap_fmr),
647 rvt_unmap_fmr);
648 break;
649
650 case DEALLOC_FMR:
651 check_driver_override(rdi, offsetof(struct ib_device,
652 dealloc_fmr),
653 rvt_dealloc_fmr);
654 break;
655
656 case MMAP:
657 check_driver_override(rdi, offsetof(struct ib_device,
658 mmap),
659 rvt_mmap);
660 break;
661
662 case CREATE_CQ:
663 check_driver_override(rdi, offsetof(struct ib_device,
664 create_cq),
665 rvt_create_cq);
666 break;
667
668 case DESTROY_CQ:
669 check_driver_override(rdi, offsetof(struct ib_device,
670 destroy_cq),
671 rvt_destroy_cq);
672 break;
673
674 case POLL_CQ:
675 check_driver_override(rdi, offsetof(struct ib_device,
676 poll_cq),
677 rvt_poll_cq);
678 break;
679
680 case REQ_NOTFIY_CQ:
681 check_driver_override(rdi, offsetof(struct ib_device,
682 req_notify_cq),
683 rvt_req_notify_cq);
684 break;
685
686 case RESIZE_CQ:
687 check_driver_override(rdi, offsetof(struct ib_device,
688 resize_cq),
689 rvt_resize_cq);
690 break;
691
692 case ALLOC_PD:
693 check_driver_override(rdi, offsetof(struct ib_device,
694 alloc_pd),
695 rvt_alloc_pd);
696 break;
697
698 case DEALLOC_PD:
699 check_driver_override(rdi, offsetof(struct ib_device,
700 dealloc_pd),
701 rvt_dealloc_pd);
702 break;
703
704 default:
705 return -EINVAL;
706 }
707
708 return 0;
709}
4997870a 710
90793f71
DD
711/**
712 * rvt_register_device - register a driver
713 * @rdi: main dev structure for all of rdmavt operations
714 *
715 * It is up to drivers to allocate the rdi and fill in the appropriate
716 * information.
717 *
718 * Return: 0 on success otherwise an errno.
719 */
0194621b
DD
720int rvt_register_device(struct rvt_dev_info *rdi)
721{
1348d706 722 int ret = 0, i;
7b1e2099 723
0194621b
DD
724 if (!rdi)
725 return -EINVAL;
726
1348d706
DD
727 /*
728 * Check to ensure drivers have setup the required helpers for the verbs
729 * they want rdmavt to handle
730 */
731 for (i = 0; i < _VERB_IDX_MAX; i++)
732 if (check_support(rdi, i)) {
733 pr_err("Driver support req not met at %d\n", i);
734 return -EINVAL;
735 }
736
b534875d 737
81ba39a8
DD
738 /* Once we get past here we can use rvt_pr macros and tracepoints */
739 trace_rvt_dbg(rdi, "Driver attempting registration");
822514d7 740 rvt_mmap_init(rdi);
b534875d 741
b518d3e6 742 /* Queue Pairs */
0acb0cc7
DD
743 ret = rvt_driver_qp_init(rdi);
744 if (ret) {
745 pr_err("Error in driver QP init.\n");
746 return -EINVAL;
747 }
748
4c1e4972 749 /* Address Handle */
119a8e70
KH
750 spin_lock_init(&rdi->n_ahs_lock);
751 rdi->n_ahs_allocated = 0;
4c1e4972 752
aad9158b 753 /* Shared Receive Queue */
b8f881b9 754 rvt_driver_srq_init(rdi);
aad9158b 755
9fa25171 756 /* Multicast */
4e74080b 757 rvt_driver_mcast_init(rdi);
9fa25171 758
2a055eb7 759 /* Mem Region */
7b1e2099
DD
760 ret = rvt_driver_mr_init(rdi);
761 if (ret) {
36055a06 762 pr_err("Error in driver MR init.\n");
7b1e2099
DD
763 goto bail_no_mr;
764 }
765
cf16335a 766 /* Completion queues */
6f6387ae
DD
767 ret = rvt_driver_cq_init(rdi);
768 if (ret) {
769 pr_err("Error in driver CQ init.\n");
770 goto bail_mr;
771 }
cf16335a 772
8afd32eb 773 /* DMA Operations */
c1b332bc
DD
774 rdi->ibdev.dma_ops =
775 rdi->ibdev.dma_ops ? : &rvt_default_dma_mapping_ops;
776
8afd32eb 777 /* Protection Domain */
8afd32eb
DD
778 spin_lock_init(&rdi->n_pds_lock);
779 rdi->n_pds_allocated = 0;
780
182285d0
DD
781 /*
782 * There are some things which could be set by underlying drivers but
783 * really should be up to rdmavt to set. For instance drivers can't know
784 * exactly which functions rdmavt supports, nor do they know the ABI
785 * version, so we do all of this sort of stuff here.
786 */
787 rdi->ibdev.uverbs_abi_ver = RVT_UVERBS_ABI_VERSION;
788 rdi->ibdev.uverbs_cmd_mask =
789 (1ull << IB_USER_VERBS_CMD_GET_CONTEXT) |
790 (1ull << IB_USER_VERBS_CMD_QUERY_DEVICE) |
791 (1ull << IB_USER_VERBS_CMD_QUERY_PORT) |
792 (1ull << IB_USER_VERBS_CMD_ALLOC_PD) |
793 (1ull << IB_USER_VERBS_CMD_DEALLOC_PD) |
794 (1ull << IB_USER_VERBS_CMD_CREATE_AH) |
795 (1ull << IB_USER_VERBS_CMD_MODIFY_AH) |
796 (1ull << IB_USER_VERBS_CMD_QUERY_AH) |
797 (1ull << IB_USER_VERBS_CMD_DESTROY_AH) |
798 (1ull << IB_USER_VERBS_CMD_REG_MR) |
799 (1ull << IB_USER_VERBS_CMD_DEREG_MR) |
800 (1ull << IB_USER_VERBS_CMD_CREATE_COMP_CHANNEL) |
801 (1ull << IB_USER_VERBS_CMD_CREATE_CQ) |
802 (1ull << IB_USER_VERBS_CMD_RESIZE_CQ) |
803 (1ull << IB_USER_VERBS_CMD_DESTROY_CQ) |
804 (1ull << IB_USER_VERBS_CMD_POLL_CQ) |
805 (1ull << IB_USER_VERBS_CMD_REQ_NOTIFY_CQ) |
806 (1ull << IB_USER_VERBS_CMD_CREATE_QP) |
807 (1ull << IB_USER_VERBS_CMD_QUERY_QP) |
808 (1ull << IB_USER_VERBS_CMD_MODIFY_QP) |
809 (1ull << IB_USER_VERBS_CMD_DESTROY_QP) |
810 (1ull << IB_USER_VERBS_CMD_POST_SEND) |
811 (1ull << IB_USER_VERBS_CMD_POST_RECV) |
812 (1ull << IB_USER_VERBS_CMD_ATTACH_MCAST) |
813 (1ull << IB_USER_VERBS_CMD_DETACH_MCAST) |
814 (1ull << IB_USER_VERBS_CMD_CREATE_SRQ) |
815 (1ull << IB_USER_VERBS_CMD_MODIFY_SRQ) |
816 (1ull << IB_USER_VERBS_CMD_QUERY_SRQ) |
817 (1ull << IB_USER_VERBS_CMD_DESTROY_SRQ) |
818 (1ull << IB_USER_VERBS_CMD_POST_SRQ_RECV);
819 rdi->ibdev.node_type = RDMA_NODE_IB_CA;
820 rdi->ibdev.num_comp_vectors = 1;
821
7b1e2099
DD
822 /* We are now good to announce we exist */
823 ret = ib_register_device(&rdi->ibdev, rdi->driver_f.port_callback);
824 if (ret) {
825 rvt_pr_err(rdi, "Failed to register driver with ib core.\n");
6f6387ae 826 goto bail_cq;
7b1e2099
DD
827 }
828
3711baf2
DD
829 rvt_create_mad_agents(rdi);
830
b534875d 831 rvt_pr_info(rdi, "Registration with rdmavt done.\n");
7b1e2099 832 return ret;
aec57787 833
6f6387ae
DD
834bail_cq:
835 rvt_cq_exit(rdi);
836
7b1e2099
DD
837bail_mr:
838 rvt_mr_exit(rdi);
839
840bail_no_mr:
0acb0cc7
DD
841 rvt_qp_exit(rdi);
842
7b1e2099 843 return ret;
0194621b
DD
844}
845EXPORT_SYMBOL(rvt_register_device);
846
90793f71
DD
847/**
848 * rvt_unregister_device - remove a driver
849 * @rdi: rvt dev struct
850 */
0194621b
DD
851void rvt_unregister_device(struct rvt_dev_info *rdi)
852{
81ba39a8 853 trace_rvt_dbg(rdi, "Driver is unregistering.");
0194621b
DD
854 if (!rdi)
855 return;
856
3711baf2
DD
857 rvt_free_mad_agents(rdi);
858
0194621b 859 ib_unregister_device(&rdi->ibdev);
6f6387ae 860 rvt_cq_exit(rdi);
7b1e2099 861 rvt_mr_exit(rdi);
515667f8 862 rvt_qp_exit(rdi);
0194621b
DD
863}
864EXPORT_SYMBOL(rvt_unregister_device);
f3d01bbc 865
90793f71
DD
866/**
867 * rvt_init_port - init internal data for driver port
868 * @rdi: rvt dev strut
869 * @port: rvt port
870 * @port_index: 0 based index of ports, different from IB core port num
871 *
f3d01bbc
DD
872 * Keep track of a list of ports. No need to have a detach port.
873 * They persist until the driver goes away.
90793f71
DD
874 *
875 * Return: always 0
f3d01bbc 876 */
38ce2c6f 877int rvt_init_port(struct rvt_dev_info *rdi, struct rvt_ibport *port,
f1badc71 878 int port_index, u16 *pkey_table)
f3d01bbc 879{
38ce2c6f 880
f1badc71
DD
881 rdi->ports[port_index] = port;
882 rdi->ports[port_index]->pkey_table = pkey_table;
38ce2c6f
DD
883
884 return 0;
f3d01bbc 885}
38ce2c6f 886EXPORT_SYMBOL(rvt_init_port);
This page took 0.075555 seconds and 5 git commands to generate.