IB/uverbs: Export ib_umem_get()/ib_umem_release() to modules
[deliverable/linux.git] / drivers / infiniband / hw / ehca / ehca_classes.h
CommitLineData
fab97220
HS
1/*
2 * IBM eServer eHCA Infiniband device driver for Linux on POWER
3 *
4 * Struct definition for eHCA internal structures
5 *
6 * Authors: Heiko J Schick <schickhj@de.ibm.com>
7 * Christoph Raisch <raisch@de.ibm.com>
8 *
9 * Copyright (c) 2005 IBM Corporation
10 *
11 * All rights reserved.
12 *
13 * This source code is distributed under a dual license of GPL v2.0 and OpenIB
14 * BSD.
15 *
16 * OpenIB BSD License
17 *
18 * Redistribution and use in source and binary forms, with or without
19 * modification, are permitted provided that the following conditions are met:
20 *
21 * Redistributions of source code must retain the above copyright notice, this
22 * list of conditions and the following disclaimer.
23 *
24 * Redistributions in binary form must reproduce the above copyright notice,
25 * this list of conditions and the following disclaimer in the documentation
26 * and/or other materials
27 * provided with the distribution.
28 *
29 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
30 * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
31 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
32 * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
33 * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
34 * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
35 * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR
36 * BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER
37 * IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
38 * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
39 * POSSIBILITY OF SUCH DAMAGE.
40 */
41
42#ifndef __EHCA_CLASSES_H__
43#define __EHCA_CLASSES_H__
44
fab97220
HS
45
46struct ehca_module;
47struct ehca_qp;
48struct ehca_cq;
49struct ehca_eq;
50struct ehca_mr;
51struct ehca_mw;
52struct ehca_pd;
53struct ehca_av;
54
31726798
HNN
55#include <linux/wait.h>
56
78d8d5f9
HNN
57#include <rdma/ib_verbs.h>
58#include <rdma/ib_user_verbs.h>
59
fab97220
HS
60#ifdef CONFIG_PPC64
61#include "ehca_classes_pSeries.h"
62#endif
78d8d5f9
HNN
63#include "ipz_pt_fn.h"
64#include "ehca_qes.h"
65#include "ehca_irq.h"
fab97220 66
78d8d5f9 67#define EHCA_EQE_CACHE_SIZE 20
fab97220 68
78d8d5f9
HNN
69struct ehca_eqe_cache_entry {
70 struct ehca_eqe *eqe;
71 struct ehca_cq *cq;
72};
fab97220
HS
73
74struct ehca_eq {
75 u32 length;
76 struct ipz_queue ipz_queue;
77 struct ipz_eq_handle ipz_eq_handle;
78 struct work_struct work;
79 struct h_galpas galpas;
80 int is_initialized;
81 struct ehca_pfeq pf;
82 spinlock_t spinlock;
83 struct tasklet_struct interrupt_task;
84 u32 ist;
78d8d5f9
HNN
85 spinlock_t irq_spinlock;
86 struct ehca_eqe_cache_entry eqe_cache[EHCA_EQE_CACHE_SIZE];
fab97220
HS
87};
88
89struct ehca_sport {
90 struct ib_cq *ibcq_aqp1;
91 struct ib_qp *ibqp_aqp1;
92 enum ib_rate rate;
93 enum ib_port_state port_state;
94};
95
96struct ehca_shca {
97 struct ib_device ib_device;
98 struct ibmebus_dev *ibmebus_dev;
99 u8 num_ports;
100 int hw_level;
101 struct list_head shca_list;
102 struct ipz_adapter_handle ipz_hca_handle;
103 struct ehca_sport sport[2];
104 struct ehca_eq eq;
105 struct ehca_eq neq;
106 struct ehca_mr *maxmr;
107 struct ehca_pd *pd;
108 struct h_galpas galpas;
c4ed790d 109 struct mutex modify_mutex;
fab97220
HS
110};
111
112struct ehca_pd {
113 struct ib_pd ib_pd;
114 struct ipz_pd fw_pd;
115 u32 ownpid;
116};
117
118struct ehca_qp {
119 struct ib_qp ib_qp;
120 u32 qp_type;
121 struct ipz_queue ipz_squeue;
122 struct ipz_queue ipz_rqueue;
123 struct h_galpas galpas;
124 u32 qkey;
125 u32 real_qp_num;
126 u32 token;
127 spinlock_t spinlock_s;
128 spinlock_t spinlock_r;
129 u32 sq_max_inline_data_size;
130 struct ipz_qp_handle ipz_qp_handle;
131 struct ehca_pfqp pf;
132 struct ib_qp_init_attr init_attr;
fab97220
HS
133 struct ehca_cq *send_cq;
134 struct ehca_cq *recv_cq;
135 unsigned int sqerr_purgeflag;
136 struct hlist_node list_entries;
4c34bdf5
HNN
137 /* mmap counter for resources mapped into user space */
138 u32 mm_count_squeue;
139 u32 mm_count_rqueue;
140 u32 mm_count_galpa;
fab97220
HS
141};
142
143/* must be power of 2 */
144#define QP_HASHTAB_LEN 8
145
146struct ehca_cq {
147 struct ib_cq ib_cq;
148 struct ipz_queue ipz_queue;
149 struct h_galpas galpas;
150 spinlock_t spinlock;
151 u32 cq_number;
152 u32 token;
153 u32 nr_of_entries;
154 struct ipz_cq_handle ipz_cq_handle;
155 struct ehca_pfcq pf;
156 spinlock_t cb_lock;
fab97220
HS
157 struct hlist_head qp_hashtab[QP_HASHTAB_LEN];
158 struct list_head entry;
31726798
HNN
159 u32 nr_callbacks; /* #events assigned to cpu by scaling code */
160 u32 nr_events; /* #events seen */
161 wait_queue_head_t wait_completion;
fab97220
HS
162 spinlock_t task_lock;
163 u32 ownpid;
4c34bdf5
HNN
164 /* mmap counter for resources mapped into user space */
165 u32 mm_count_queue;
166 u32 mm_count_galpa;
fab97220
HS
167};
168
169enum ehca_mr_flag {
170 EHCA_MR_FLAG_FMR = 0x80000000, /* FMR, created with ehca_alloc_fmr */
171 EHCA_MR_FLAG_MAXMR = 0x40000000, /* max-MR */
172};
173
174struct ehca_mr {
175 union {
176 struct ib_mr ib_mr; /* must always be first in ehca_mr */
177 struct ib_fmr ib_fmr; /* must always be first in ehca_mr */
178 } ib;
f7c6a7b5 179 struct ib_umem *umem;
fab97220
HS
180 spinlock_t mrlock;
181
182 enum ehca_mr_flag flags;
183 u32 num_pages; /* number of MR pages */
184 u32 num_4k; /* number of 4k "page" portions to form MR */
185 int acl; /* ACL (stored here for usage in reregister) */
186 u64 *start; /* virtual start address (stored here for */
187 /* usage in reregister) */
188 u64 size; /* size (stored here for usage in reregister) */
189 u32 fmr_page_size; /* page size for FMR */
190 u32 fmr_max_pages; /* max pages for FMR */
191 u32 fmr_max_maps; /* max outstanding maps for FMR */
192 u32 fmr_map_cnt; /* map counter for FMR */
193 /* fw specific data */
194 struct ipz_mrmw_handle ipz_mr_handle; /* MR handle for h-calls */
195 struct h_galpas galpas;
196 /* data for userspace bridge */
197 u32 nr_of_pages;
198 void *pagearray;
199};
200
201struct ehca_mw {
202 struct ib_mw ib_mw; /* gen2 mw, must always be first in ehca_mw */
203 spinlock_t mwlock;
204
205 u8 never_bound; /* indication MW was never bound */
206 struct ipz_mrmw_handle ipz_mw_handle; /* MW handle for h-calls */
207 struct h_galpas galpas;
208};
209
210enum ehca_mr_pgi_type {
211 EHCA_MR_PGI_PHYS = 1, /* type of ehca_reg_phys_mr,
212 * ehca_rereg_phys_mr,
213 * ehca_reg_internal_maxmr */
214 EHCA_MR_PGI_USER = 2, /* type of ehca_reg_user_mr */
215 EHCA_MR_PGI_FMR = 3 /* type of ehca_map_phys_fmr */
216};
217
218struct ehca_mr_pginfo {
219 enum ehca_mr_pgi_type type;
220 u64 num_pages;
221 u64 page_cnt;
222 u64 num_4k; /* number of 4k "page" portions */
223 u64 page_4k_cnt; /* counter for 4k "page" portions */
224 u64 next_4k; /* next 4k "page" portion in buffer/chunk/listelem */
225
226 /* type EHCA_MR_PGI_PHYS section */
227 int num_phys_buf;
228 struct ib_phys_buf *phys_buf_array;
229 u64 next_buf;
230
231 /* type EHCA_MR_PGI_USER section */
232 struct ib_umem *region;
233 struct ib_umem_chunk *next_chunk;
234 u64 next_nmap;
235
236 /* type EHCA_MR_PGI_FMR section */
237 u64 *page_list;
238 u64 next_listelem;
239 /* next_4k also used within EHCA_MR_PGI_FMR */
240};
241
242/* output parameters for MR/FMR hipz calls */
243struct ehca_mr_hipzout_parms {
244 struct ipz_mrmw_handle handle;
245 u32 lkey;
246 u32 rkey;
247 u64 len;
248 u64 vaddr;
249 u32 acl;
250};
251
252/* output parameters for MW hipz calls */
253struct ehca_mw_hipzout_parms {
254 struct ipz_mrmw_handle handle;
255 u32 rkey;
256};
257
258struct ehca_av {
259 struct ib_ah ib_ah;
260 struct ehca_ud_av av;
261};
262
263struct ehca_ucontext {
264 struct ib_ucontext ib_ucontext;
265};
266
fab97220
HS
267int ehca_init_pd_cache(void);
268void ehca_cleanup_pd_cache(void);
269int ehca_init_cq_cache(void);
270void ehca_cleanup_cq_cache(void);
271int ehca_init_qp_cache(void);
272void ehca_cleanup_qp_cache(void);
273int ehca_init_av_cache(void);
274void ehca_cleanup_av_cache(void);
275int ehca_init_mrmw_cache(void);
276void ehca_cleanup_mrmw_cache(void);
277
278extern spinlock_t ehca_qp_idr_lock;
279extern spinlock_t ehca_cq_idr_lock;
280extern struct idr ehca_qp_idr;
281extern struct idr ehca_cq_idr;
282
283extern int ehca_static_rate;
284extern int ehca_port_act_time;
285extern int ehca_use_hp_mr;
4fd30060 286extern int ehca_scaling_code;
fab97220
HS
287
288struct ipzu_queue_resp {
fab97220
HS
289 u32 qe_size; /* queue entry size */
290 u32 act_nr_of_sg;
291 u32 queue_length; /* queue length allocated in bytes */
292 u32 pagesize;
293 u32 toggle_state;
294 u32 dummy; /* padding for 8 byte alignment */
295};
296
297struct ehca_create_cq_resp {
298 u32 cq_number;
299 u32 token;
300 struct ipzu_queue_resp ipz_queue;
fab97220
HS
301};
302
303struct ehca_create_qp_resp {
304 u32 qp_num;
305 u32 token;
306 u32 qp_type;
307 u32 qkey;
308 /* qp_num assigned by ehca: sqp0/1 may have got different numbers */
309 u32 real_qp_num;
310 u32 dummy; /* padding for 8 byte alignment */
311 struct ipzu_queue_resp ipz_squeue;
312 struct ipzu_queue_resp ipz_rqueue;
fab97220
HS
313};
314
315struct ehca_alloc_cq_parms {
316 u32 nr_cqe;
317 u32 act_nr_of_entries;
318 u32 act_pages;
319 struct ipz_eq_handle eq_handle;
320};
321
322struct ehca_alloc_qp_parms {
323 int servicetype;
324 int sigtype;
325 int daqp_ctrl;
326 int max_send_sge;
327 int max_recv_sge;
328 int ud_av_l_key_ctl;
329
330 u16 act_nr_send_wqes;
331 u16 act_nr_recv_wqes;
332 u8 act_nr_recv_sges;
333 u8 act_nr_send_sges;
334
335 u32 nr_rq_pages;
336 u32 nr_sq_pages;
337
338 struct ipz_eq_handle ipz_eq_handle;
339 struct ipz_pd pd;
340};
341
342int ehca_cq_assign_qp(struct ehca_cq *cq, struct ehca_qp *qp);
343int ehca_cq_unassign_qp(struct ehca_cq *cq, unsigned int qp_num);
344struct ehca_qp* ehca_cq_get_qp(struct ehca_cq *cq, int qp_num);
345
346#endif
This page took 0.097116 seconds and 5 git commands to generate.