fs: use kmem_cache_zalloc instead
[pv_ops_mirror.git] / drivers / infiniband / hw / ehca / ehca_hca.c
blob4aa3ffa6a19fe425345e1bcd800dad2d6cc3d459
1 /*
2 * IBM eServer eHCA Infiniband device driver for Linux on POWER
4 * HCA query functions
6 * Authors: Heiko J Schick <schickhj@de.ibm.com>
7 * Christoph Raisch <raisch@de.ibm.com>
9 * Copyright (c) 2005 IBM Corporation
11 * All rights reserved.
13 * This source code is distributed under a dual license of GPL v2.0 and OpenIB
14 * BSD.
16 * OpenIB BSD License
18 * Redistribution and use in source and binary forms, with or without
19 * modification, are permitted provided that the following conditions are met:
21 * Redistributions of source code must retain the above copyright notice, this
22 * list of conditions and the following disclaimer.
24 * Redistributions in binary form must reproduce the above copyright notice,
25 * this list of conditions and the following disclaimer in the documentation
26 * and/or other materials
27 * provided with the distribution.
29 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
30 * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
31 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
32 * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
33 * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
34 * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
35 * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR
36 * BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER
37 * IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
38 * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
39 * POSSIBILITY OF SUCH DAMAGE.
42 #include "ehca_tools.h"
43 #include "ehca_iverbs.h"
44 #include "hcp_if.h"
46 int ehca_query_device(struct ib_device *ibdev, struct ib_device_attr *props)
48 int i, ret = 0;
49 struct ehca_shca *shca = container_of(ibdev, struct ehca_shca,
50 ib_device);
51 struct hipz_query_hca *rblock;
53 static const u32 cap_mapping[] = {
54 IB_DEVICE_RESIZE_MAX_WR, HCA_CAP_WQE_RESIZE,
55 IB_DEVICE_BAD_PKEY_CNTR, HCA_CAP_BAD_P_KEY_CTR,
56 IB_DEVICE_BAD_QKEY_CNTR, HCA_CAP_Q_KEY_VIOL_CTR,
57 IB_DEVICE_RAW_MULTI, HCA_CAP_RAW_PACKET_MCAST,
58 IB_DEVICE_AUTO_PATH_MIG, HCA_CAP_AUTO_PATH_MIG,
59 IB_DEVICE_CHANGE_PHY_PORT, HCA_CAP_SQD_RTS_PORT_CHANGE,
60 IB_DEVICE_UD_AV_PORT_ENFORCE, HCA_CAP_AH_PORT_NR_CHECK,
61 IB_DEVICE_CURR_QP_STATE_MOD, HCA_CAP_CUR_QP_STATE_MOD,
62 IB_DEVICE_SHUTDOWN_PORT, HCA_CAP_SHUTDOWN_PORT,
63 IB_DEVICE_INIT_TYPE, HCA_CAP_INIT_TYPE,
64 IB_DEVICE_PORT_ACTIVE_EVENT, HCA_CAP_PORT_ACTIVE_EVENT,
67 rblock = ehca_alloc_fw_ctrlblock(GFP_KERNEL);
68 if (!rblock) {
69 ehca_err(&shca->ib_device, "Can't allocate rblock memory.");
70 return -ENOMEM;
73 if (hipz_h_query_hca(shca->ipz_hca_handle, rblock) != H_SUCCESS) {
74 ehca_err(&shca->ib_device, "Can't query device properties");
75 ret = -EINVAL;
76 goto query_device1;
79 memset(props, 0, sizeof(struct ib_device_attr));
80 props->fw_ver = rblock->hw_ver;
81 props->max_mr_size = rblock->max_mr_size;
82 props->vendor_id = rblock->vendor_id >> 8;
83 props->vendor_part_id = rblock->vendor_part_id >> 16;
84 props->hw_ver = rblock->hw_ver;
85 props->max_qp = min_t(unsigned, rblock->max_qp, INT_MAX);
86 props->max_qp_wr = min_t(unsigned, rblock->max_wqes_wq, INT_MAX);
87 props->max_sge = min_t(unsigned, rblock->max_sge, INT_MAX);
88 props->max_sge_rd = min_t(unsigned, rblock->max_sge_rd, INT_MAX);
89 props->max_cq = min_t(unsigned, rblock->max_cq, INT_MAX);
90 props->max_cqe = min_t(unsigned, rblock->max_cqe, INT_MAX);
91 props->max_mr = min_t(unsigned, rblock->max_mr, INT_MAX);
92 props->max_mw = min_t(unsigned, rblock->max_mw, INT_MAX);
93 props->max_pd = min_t(unsigned, rblock->max_pd, INT_MAX);
94 props->max_ah = min_t(unsigned, rblock->max_ah, INT_MAX);
95 props->max_fmr = min_t(unsigned, rblock->max_mr, INT_MAX);
97 if (EHCA_BMASK_GET(HCA_CAP_SRQ, shca->hca_cap)) {
98 props->max_srq = props->max_qp;
99 props->max_srq_wr = props->max_qp_wr;
100 props->max_srq_sge = 3;
103 props->max_pkeys = 16;
104 props->local_ca_ack_delay
105 = rblock->local_ca_ack_delay;
106 props->max_raw_ipv6_qp
107 = min_t(unsigned, rblock->max_raw_ipv6_qp, INT_MAX);
108 props->max_raw_ethy_qp
109 = min_t(unsigned, rblock->max_raw_ethy_qp, INT_MAX);
110 props->max_mcast_grp
111 = min_t(unsigned, rblock->max_mcast_grp, INT_MAX);
112 props->max_mcast_qp_attach
113 = min_t(unsigned, rblock->max_mcast_qp_attach, INT_MAX);
114 props->max_total_mcast_qp_attach
115 = min_t(unsigned, rblock->max_total_mcast_qp_attach, INT_MAX);
117 /* translate device capabilities */
118 props->device_cap_flags = IB_DEVICE_SYS_IMAGE_GUID |
119 IB_DEVICE_RC_RNR_NAK_GEN | IB_DEVICE_N_NOTIFY_CQ;
120 for (i = 0; i < ARRAY_SIZE(cap_mapping); i += 2)
121 if (rblock->hca_cap_indicators & cap_mapping[i + 1])
122 props->device_cap_flags |= cap_mapping[i];
124 query_device1:
125 ehca_free_fw_ctrlblock(rblock);
127 return ret;
130 int ehca_query_port(struct ib_device *ibdev,
131 u8 port, struct ib_port_attr *props)
133 int ret = 0;
134 u64 h_ret;
135 struct ehca_shca *shca = container_of(ibdev, struct ehca_shca,
136 ib_device);
137 struct hipz_query_port *rblock;
139 rblock = ehca_alloc_fw_ctrlblock(GFP_KERNEL);
140 if (!rblock) {
141 ehca_err(&shca->ib_device, "Can't allocate rblock memory.");
142 return -ENOMEM;
145 h_ret = hipz_h_query_port(shca->ipz_hca_handle, port, rblock);
146 if (h_ret != H_SUCCESS) {
147 ehca_err(&shca->ib_device, "Can't query port properties");
148 ret = -EINVAL;
149 goto query_port1;
152 memset(props, 0, sizeof(struct ib_port_attr));
153 props->state = rblock->state;
155 switch (rblock->max_mtu) {
156 case 0x1:
157 props->active_mtu = props->max_mtu = IB_MTU_256;
158 break;
159 case 0x2:
160 props->active_mtu = props->max_mtu = IB_MTU_512;
161 break;
162 case 0x3:
163 props->active_mtu = props->max_mtu = IB_MTU_1024;
164 break;
165 case 0x4:
166 props->active_mtu = props->max_mtu = IB_MTU_2048;
167 break;
168 case 0x5:
169 props->active_mtu = props->max_mtu = IB_MTU_4096;
170 break;
171 default:
172 ehca_err(&shca->ib_device, "Unknown MTU size: %x.",
173 rblock->max_mtu);
174 break;
177 props->port_cap_flags = rblock->capability_mask;
178 props->gid_tbl_len = rblock->gid_tbl_len;
179 props->max_msg_sz = rblock->max_msg_sz;
180 props->bad_pkey_cntr = rblock->bad_pkey_cntr;
181 props->qkey_viol_cntr = rblock->qkey_viol_cntr;
182 props->pkey_tbl_len = rblock->pkey_tbl_len;
183 props->lid = rblock->lid;
184 props->sm_lid = rblock->sm_lid;
185 props->lmc = rblock->lmc;
186 props->sm_sl = rblock->sm_sl;
187 props->subnet_timeout = rblock->subnet_timeout;
188 props->init_type_reply = rblock->init_type_reply;
190 props->active_width = IB_WIDTH_12X;
191 props->active_speed = 0x1;
193 /* at the moment (logical) link state is always LINK_UP */
194 props->phys_state = 0x5;
196 query_port1:
197 ehca_free_fw_ctrlblock(rblock);
199 return ret;
202 int ehca_query_sma_attr(struct ehca_shca *shca,
203 u8 port, struct ehca_sma_attr *attr)
205 int ret = 0;
206 u64 h_ret;
207 struct hipz_query_port *rblock;
209 rblock = ehca_alloc_fw_ctrlblock(GFP_ATOMIC);
210 if (!rblock) {
211 ehca_err(&shca->ib_device, "Can't allocate rblock memory.");
212 return -ENOMEM;
215 h_ret = hipz_h_query_port(shca->ipz_hca_handle, port, rblock);
216 if (h_ret != H_SUCCESS) {
217 ehca_err(&shca->ib_device, "Can't query port properties");
218 ret = -EINVAL;
219 goto query_sma_attr1;
222 memset(attr, 0, sizeof(struct ehca_sma_attr));
224 attr->lid = rblock->lid;
225 attr->lmc = rblock->lmc;
226 attr->sm_sl = rblock->sm_sl;
227 attr->sm_lid = rblock->sm_lid;
229 attr->pkey_tbl_len = rblock->pkey_tbl_len;
230 memcpy(attr->pkeys, rblock->pkey_entries, sizeof(attr->pkeys));
232 query_sma_attr1:
233 ehca_free_fw_ctrlblock(rblock);
235 return ret;
238 int ehca_query_pkey(struct ib_device *ibdev, u8 port, u16 index, u16 *pkey)
240 int ret = 0;
241 u64 h_ret;
242 struct ehca_shca *shca;
243 struct hipz_query_port *rblock;
245 shca = container_of(ibdev, struct ehca_shca, ib_device);
246 if (index > 16) {
247 ehca_err(&shca->ib_device, "Invalid index: %x.", index);
248 return -EINVAL;
251 rblock = ehca_alloc_fw_ctrlblock(GFP_KERNEL);
252 if (!rblock) {
253 ehca_err(&shca->ib_device, "Can't allocate rblock memory.");
254 return -ENOMEM;
257 h_ret = hipz_h_query_port(shca->ipz_hca_handle, port, rblock);
258 if (h_ret != H_SUCCESS) {
259 ehca_err(&shca->ib_device, "Can't query port properties");
260 ret = -EINVAL;
261 goto query_pkey1;
264 memcpy(pkey, &rblock->pkey_entries + index, sizeof(u16));
266 query_pkey1:
267 ehca_free_fw_ctrlblock(rblock);
269 return ret;
272 int ehca_query_gid(struct ib_device *ibdev, u8 port,
273 int index, union ib_gid *gid)
275 int ret = 0;
276 u64 h_ret;
277 struct ehca_shca *shca = container_of(ibdev, struct ehca_shca,
278 ib_device);
279 struct hipz_query_port *rblock;
281 if (index > 255) {
282 ehca_err(&shca->ib_device, "Invalid index: %x.", index);
283 return -EINVAL;
286 rblock = ehca_alloc_fw_ctrlblock(GFP_KERNEL);
287 if (!rblock) {
288 ehca_err(&shca->ib_device, "Can't allocate rblock memory.");
289 return -ENOMEM;
292 h_ret = hipz_h_query_port(shca->ipz_hca_handle, port, rblock);
293 if (h_ret != H_SUCCESS) {
294 ehca_err(&shca->ib_device, "Can't query port properties");
295 ret = -EINVAL;
296 goto query_gid1;
299 memcpy(&gid->raw[0], &rblock->gid_prefix, sizeof(u64));
300 memcpy(&gid->raw[8], &rblock->guid_entries[index], sizeof(u64));
302 query_gid1:
303 ehca_free_fw_ctrlblock(rblock);
305 return ret;
308 const u32 allowed_port_caps = (
309 IB_PORT_SM | IB_PORT_LED_INFO_SUP | IB_PORT_CM_SUP |
310 IB_PORT_SNMP_TUNNEL_SUP | IB_PORT_DEVICE_MGMT_SUP |
311 IB_PORT_VENDOR_CLASS_SUP);
313 int ehca_modify_port(struct ib_device *ibdev,
314 u8 port, int port_modify_mask,
315 struct ib_port_modify *props)
317 int ret = 0;
318 struct ehca_shca *shca;
319 struct hipz_query_port *rblock;
320 u32 cap;
321 u64 hret;
323 shca = container_of(ibdev, struct ehca_shca, ib_device);
324 if ((props->set_port_cap_mask | props->clr_port_cap_mask)
325 & ~allowed_port_caps) {
326 ehca_err(&shca->ib_device, "Non-changeable bits set in masks "
327 "set=%x clr=%x allowed=%x", props->set_port_cap_mask,
328 props->clr_port_cap_mask, allowed_port_caps);
329 return -EINVAL;
332 if (mutex_lock_interruptible(&shca->modify_mutex))
333 return -ERESTARTSYS;
335 rblock = ehca_alloc_fw_ctrlblock(GFP_KERNEL);
336 if (!rblock) {
337 ehca_err(&shca->ib_device, "Can't allocate rblock memory.");
338 ret = -ENOMEM;
339 goto modify_port1;
342 hret = hipz_h_query_port(shca->ipz_hca_handle, port, rblock);
343 if (hret != H_SUCCESS) {
344 ehca_err(&shca->ib_device, "Can't query port properties");
345 ret = -EINVAL;
346 goto modify_port2;
349 cap = (rblock->capability_mask | props->set_port_cap_mask)
350 & ~props->clr_port_cap_mask;
352 hret = hipz_h_modify_port(shca->ipz_hca_handle, port,
353 cap, props->init_type, port_modify_mask);
354 if (hret != H_SUCCESS) {
355 ehca_err(&shca->ib_device, "Modify port failed h_ret=%li",
356 hret);
357 ret = -EINVAL;
360 modify_port2:
361 ehca_free_fw_ctrlblock(rblock);
363 modify_port1:
364 mutex_unlock(&shca->modify_mutex);
366 return ret;