2 * IBM eServer eHCA Infiniband device driver for Linux on POWER
4 * Struct definition for eHCA internal structures
6 * Authors: Heiko J Schick <schickhj@de.ibm.com>
7 * Christoph Raisch <raisch@de.ibm.com>
9 * Copyright (c) 2005 IBM Corporation
11 * All rights reserved.
13 * This source code is distributed under a dual license of GPL v2.0 and OpenIB
18 * Redistribution and use in source and binary forms, with or without
19 * modification, are permitted provided that the following conditions are met:
21 * Redistributions of source code must retain the above copyright notice, this
22 * list of conditions and the following disclaimer.
24 * Redistributions in binary form must reproduce the above copyright notice,
25 * this list of conditions and the following disclaimer in the documentation
26 * and/or other materials
27 * provided with the distribution.
29 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
30 * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
31 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
32 * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
33 * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
34 * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
35 * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR
36 * BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER
37 * IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
38 * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
39 * POSSIBILITY OF SUCH DAMAGE.
42 #ifndef __EHCA_CLASSES_H__
43 #define __EHCA_CLASSES_H__
55 #include <linux/wait.h>
57 #include <rdma/ib_verbs.h>
58 #include <rdma/ib_user_verbs.h>
61 #include "ehca_classes_pSeries.h"
63 #include "ipz_pt_fn.h"
67 #define EHCA_EQE_CACHE_SIZE 20
69 struct ehca_eqe_cache_entry
{
76 struct ipz_queue ipz_queue
;
77 struct ipz_eq_handle ipz_eq_handle
;
78 struct work_struct work
;
79 struct h_galpas galpas
;
83 struct tasklet_struct interrupt_task
;
85 spinlock_t irq_spinlock
;
86 struct ehca_eqe_cache_entry eqe_cache
[EHCA_EQE_CACHE_SIZE
];
90 struct ib_cq
*ibcq_aqp1
;
91 struct ib_qp
*ibqp_aqp1
;
93 enum ib_port_state port_state
;
97 struct ib_device ib_device
;
98 struct ibmebus_dev
*ibmebus_dev
;
101 struct list_head shca_list
;
102 struct ipz_adapter_handle ipz_hca_handle
;
103 struct ehca_sport sport
[2];
106 struct ehca_mr
*maxmr
;
108 struct h_galpas galpas
;
109 struct mutex modify_mutex
;
121 struct ipz_queue ipz_squeue
;
122 struct ipz_queue ipz_rqueue
;
123 struct h_galpas galpas
;
127 spinlock_t spinlock_s
;
128 spinlock_t spinlock_r
;
129 u32 sq_max_inline_data_size
;
130 struct ipz_qp_handle ipz_qp_handle
;
132 struct ib_qp_init_attr init_attr
;
133 struct ehca_cq
*send_cq
;
134 struct ehca_cq
*recv_cq
;
135 unsigned int sqerr_purgeflag
;
136 struct hlist_node list_entries
;
137 /* mmap counter for resources mapped into user space */
143 /* must be power of 2 */
144 #define QP_HASHTAB_LEN 8
148 struct ipz_queue ipz_queue
;
149 struct h_galpas galpas
;
154 struct ipz_cq_handle ipz_cq_handle
;
157 struct hlist_head qp_hashtab
[QP_HASHTAB_LEN
];
158 struct list_head entry
;
159 u32 nr_callbacks
; /* #events assigned to cpu by scaling code */
160 u32 nr_events
; /* #events seen */
161 wait_queue_head_t wait_completion
;
162 spinlock_t task_lock
;
164 /* mmap counter for resources mapped into user space */
170 EHCA_MR_FLAG_FMR
= 0x80000000, /* FMR, created with ehca_alloc_fmr */
171 EHCA_MR_FLAG_MAXMR
= 0x40000000, /* max-MR */
176 struct ib_mr ib_mr
; /* must always be first in ehca_mr */
177 struct ib_fmr ib_fmr
; /* must always be first in ehca_mr */
179 struct ib_umem
*umem
;
182 enum ehca_mr_flag flags
;
183 u32 num_pages
; /* number of MR pages */
184 u32 num_4k
; /* number of 4k "page" portions to form MR */
185 int acl
; /* ACL (stored here for usage in reregister) */
186 u64
*start
; /* virtual start address (stored here for */
187 /* usage in reregister) */
188 u64 size
; /* size (stored here for usage in reregister) */
189 u32 fmr_page_size
; /* page size for FMR */
190 u32 fmr_max_pages
; /* max pages for FMR */
191 u32 fmr_max_maps
; /* max outstanding maps for FMR */
192 u32 fmr_map_cnt
; /* map counter for FMR */
193 /* fw specific data */
194 struct ipz_mrmw_handle ipz_mr_handle
; /* MR handle for h-calls */
195 struct h_galpas galpas
;
196 /* data for userspace bridge */
202 struct ib_mw ib_mw
; /* gen2 mw, must always be first in ehca_mw */
205 u8 never_bound
; /* indication MW was never bound */
206 struct ipz_mrmw_handle ipz_mw_handle
; /* MW handle for h-calls */
207 struct h_galpas galpas
;
210 enum ehca_mr_pgi_type
{
211 EHCA_MR_PGI_PHYS
= 1, /* type of ehca_reg_phys_mr,
212 * ehca_rereg_phys_mr,
213 * ehca_reg_internal_maxmr */
214 EHCA_MR_PGI_USER
= 2, /* type of ehca_reg_user_mr */
215 EHCA_MR_PGI_FMR
= 3 /* type of ehca_map_phys_fmr */
218 struct ehca_mr_pginfo
{
219 enum ehca_mr_pgi_type type
;
222 u64 num_4k
; /* number of 4k "page" portions */
223 u64 page_4k_cnt
; /* counter for 4k "page" portions */
224 u64 next_4k
; /* next 4k "page" portion in buffer/chunk/listelem */
226 /* type EHCA_MR_PGI_PHYS section */
228 struct ib_phys_buf
*phys_buf_array
;
231 /* type EHCA_MR_PGI_USER section */
232 struct ib_umem
*region
;
233 struct ib_umem_chunk
*next_chunk
;
236 /* type EHCA_MR_PGI_FMR section */
239 /* next_4k also used within EHCA_MR_PGI_FMR */
242 /* output parameters for MR/FMR hipz calls */
243 struct ehca_mr_hipzout_parms
{
244 struct ipz_mrmw_handle handle
;
252 /* output parameters for MW hipz calls */
253 struct ehca_mw_hipzout_parms
{
254 struct ipz_mrmw_handle handle
;
260 struct ehca_ud_av av
;
263 struct ehca_ucontext
{
264 struct ib_ucontext ib_ucontext
;
267 int ehca_init_pd_cache(void);
268 void ehca_cleanup_pd_cache(void);
269 int ehca_init_cq_cache(void);
270 void ehca_cleanup_cq_cache(void);
271 int ehca_init_qp_cache(void);
272 void ehca_cleanup_qp_cache(void);
273 int ehca_init_av_cache(void);
274 void ehca_cleanup_av_cache(void);
275 int ehca_init_mrmw_cache(void);
276 void ehca_cleanup_mrmw_cache(void);
278 extern spinlock_t ehca_qp_idr_lock
;
279 extern spinlock_t ehca_cq_idr_lock
;
280 extern struct idr ehca_qp_idr
;
281 extern struct idr ehca_cq_idr
;
283 extern int ehca_static_rate
;
284 extern int ehca_port_act_time
;
285 extern int ehca_use_hp_mr
;
286 extern int ehca_scaling_code
;
288 struct ipzu_queue_resp
{
289 u32 qe_size
; /* queue entry size */
291 u32 queue_length
; /* queue length allocated in bytes */
294 u32 dummy
; /* padding for 8 byte alignment */
297 struct ehca_create_cq_resp
{
300 struct ipzu_queue_resp ipz_queue
;
303 struct ehca_create_qp_resp
{
308 /* qp_num assigned by ehca: sqp0/1 may have got different numbers */
310 u32 dummy
; /* padding for 8 byte alignment */
311 struct ipzu_queue_resp ipz_squeue
;
312 struct ipzu_queue_resp ipz_rqueue
;
315 struct ehca_alloc_cq_parms
{
317 u32 act_nr_of_entries
;
319 struct ipz_eq_handle eq_handle
;
322 struct ehca_alloc_qp_parms
{
330 u16 act_nr_send_wqes
;
331 u16 act_nr_recv_wqes
;
338 struct ipz_eq_handle ipz_eq_handle
;
342 int ehca_cq_assign_qp(struct ehca_cq
*cq
, struct ehca_qp
*qp
);
343 int ehca_cq_unassign_qp(struct ehca_cq
*cq
, unsigned int qp_num
);
344 struct ehca_qp
* ehca_cq_get_qp(struct ehca_cq
*cq
, int qp_num
);