2 * Copyright (c) 2004, 2005 Topspin Communications. All rights reserved.
3 * Copyright (c) 2005 Mellanox Technologies. All rights reserved.
5 * This software is available to you under a choice of one of two
6 * licenses. You may choose to be licensed under the terms of the GNU
7 * General Public License (GPL) Version 2, available from the file
8 * COPYING in the main directory of this source tree, or the
9 * OpenIB.org BSD license below:
11 * Redistribution and use in source and binary forms, with or
12 * without modification, are permitted provided that the following
15 * - Redistributions of source code must retain the above
16 * copyright notice, this list of conditions and the following
19 * - Redistributions in binary form must reproduce the above
20 * copyright notice, this list of conditions and the following
21 * disclaimer in the documentation and/or other materials
22 * provided with the distribution.
24 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
25 * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
26 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
27 * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
28 * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
29 * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
30 * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
34 #include <linux/slab.h>
38 int ipoib_mcast_attach(struct net_device
*dev
, u16 mlid
, union ib_gid
*mgid
, int set_qkey
)
40 struct ipoib_dev_priv
*priv
= netdev_priv(dev
);
41 struct ib_qp_attr
*qp_attr
= NULL
;
45 if (ib_find_pkey(priv
->ca
, priv
->port
, priv
->pkey
, &pkey_index
)) {
46 clear_bit(IPOIB_PKEY_ASSIGNED
, &priv
->flags
);
50 set_bit(IPOIB_PKEY_ASSIGNED
, &priv
->flags
);
54 qp_attr
= kmalloc(sizeof *qp_attr
, GFP_KERNEL
);
58 /* set correct QKey for QP */
59 qp_attr
->qkey
= priv
->qkey
;
60 ret
= ib_modify_qp(priv
->qp
, qp_attr
, IB_QP_QKEY
);
62 ipoib_warn(priv
, "failed to modify QP, ret = %d\n", ret
);
67 /* attach QP to multicast group */
68 ret
= ib_attach_mcast(priv
->qp
, mgid
, mlid
);
70 ipoib_warn(priv
, "failed to attach to multicast group, ret = %d\n", ret
);
77 int ipoib_init_qp(struct net_device
*dev
)
79 struct ipoib_dev_priv
*priv
= netdev_priv(dev
);
81 struct ib_qp_attr qp_attr
;
84 if (!test_bit(IPOIB_PKEY_ASSIGNED
, &priv
->flags
))
87 qp_attr
.qp_state
= IB_QPS_INIT
;
89 qp_attr
.port_num
= priv
->port
;
90 qp_attr
.pkey_index
= priv
->pkey_index
;
96 ret
= ib_modify_qp(priv
->qp
, &qp_attr
, attr_mask
);
98 ipoib_warn(priv
, "failed to modify QP to init, ret = %d\n", ret
);
102 qp_attr
.qp_state
= IB_QPS_RTR
;
103 /* Can't set this in a INIT->RTR transition */
104 attr_mask
&= ~IB_QP_PORT
;
105 ret
= ib_modify_qp(priv
->qp
, &qp_attr
, attr_mask
);
107 ipoib_warn(priv
, "failed to modify QP to RTR, ret = %d\n", ret
);
111 qp_attr
.qp_state
= IB_QPS_RTS
;
113 attr_mask
|= IB_QP_SQ_PSN
;
114 attr_mask
&= ~IB_QP_PKEY_INDEX
;
115 ret
= ib_modify_qp(priv
->qp
, &qp_attr
, attr_mask
);
117 ipoib_warn(priv
, "failed to modify QP to RTS, ret = %d\n", ret
);
124 qp_attr
.qp_state
= IB_QPS_RESET
;
125 if (ib_modify_qp(priv
->qp
, &qp_attr
, IB_QP_STATE
))
126 ipoib_warn(priv
, "Failed to modify QP to RESET state\n");
131 int ipoib_transport_dev_init(struct net_device
*dev
, struct ib_device
*ca
)
133 struct ipoib_dev_priv
*priv
= netdev_priv(dev
);
134 struct ib_qp_init_attr init_attr
= {
136 .max_send_wr
= ipoib_sendq_size
,
137 .max_recv_wr
= ipoib_recvq_size
,
139 .max_recv_sge
= IPOIB_UD_RX_SG
141 .sq_sig_type
= IB_SIGNAL_ALL_WR
,
144 struct ib_cq_init_attr cq_attr
= {};
149 priv
->pd
= ib_alloc_pd(priv
->ca
);
150 if (IS_ERR(priv
->pd
)) {
151 printk(KERN_WARNING
"%s: failed to allocate PD\n", ca
->name
);
156 * the various IPoIB tasks assume they will never race against
157 * themselves, so always use a single thread workqueue
159 priv
->wq
= create_singlethread_workqueue("ipoib_wq");
161 printk(KERN_WARNING
"ipoib: failed to allocate device WQ\n");
165 size
= ipoib_recvq_size
+ 1;
166 ret
= ipoib_cm_dev_init(dev
);
168 size
+= ipoib_sendq_size
;
169 if (ipoib_cm_has_srq(dev
))
170 size
+= ipoib_recvq_size
+ 1; /* 1 extra for rx_drain_qp */
172 size
+= ipoib_recvq_size
* ipoib_max_conn_qp
;
178 priv
->recv_cq
= ib_create_cq(priv
->ca
, ipoib_ib_completion
, NULL
,
180 if (IS_ERR(priv
->recv_cq
)) {
181 printk(KERN_WARNING
"%s: failed to create receive CQ\n", ca
->name
);
182 goto out_cm_dev_cleanup
;
185 cq_attr
.cqe
= ipoib_sendq_size
;
186 priv
->send_cq
= ib_create_cq(priv
->ca
, ipoib_send_comp_handler
, NULL
,
188 if (IS_ERR(priv
->send_cq
)) {
189 printk(KERN_WARNING
"%s: failed to create send CQ\n", ca
->name
);
190 goto out_free_recv_cq
;
193 if (ib_req_notify_cq(priv
->recv_cq
, IB_CQ_NEXT_COMP
))
194 goto out_free_send_cq
;
196 init_attr
.send_cq
= priv
->send_cq
;
197 init_attr
.recv_cq
= priv
->recv_cq
;
199 if (priv
->hca_caps
& IB_DEVICE_UD_TSO
)
200 init_attr
.create_flags
|= IB_QP_CREATE_IPOIB_UD_LSO
;
202 if (priv
->hca_caps
& IB_DEVICE_BLOCK_MULTICAST_LOOPBACK
)
203 init_attr
.create_flags
|= IB_QP_CREATE_BLOCK_MULTICAST_LOOPBACK
;
205 if (priv
->hca_caps
& IB_DEVICE_MANAGED_FLOW_STEERING
)
206 init_attr
.create_flags
|= IB_QP_CREATE_NETIF_QP
;
208 if (dev
->features
& NETIF_F_SG
)
209 init_attr
.cap
.max_send_sge
=
210 min_t(u32
, priv
->ca
->attrs
.max_sge
, MAX_SKB_FRAGS
+ 1);
212 priv
->qp
= ib_create_qp(priv
->pd
, &init_attr
);
213 if (IS_ERR(priv
->qp
)) {
214 printk(KERN_WARNING
"%s: failed to create QP\n", ca
->name
);
215 goto out_free_send_cq
;
218 priv
->dev
->dev_addr
[1] = (priv
->qp
->qp_num
>> 16) & 0xff;
219 priv
->dev
->dev_addr
[2] = (priv
->qp
->qp_num
>> 8) & 0xff;
220 priv
->dev
->dev_addr
[3] = (priv
->qp
->qp_num
) & 0xff;
222 for (i
= 0; i
< MAX_SKB_FRAGS
+ 1; ++i
)
223 priv
->tx_sge
[i
].lkey
= priv
->pd
->local_dma_lkey
;
225 priv
->tx_wr
.wr
.opcode
= IB_WR_SEND
;
226 priv
->tx_wr
.wr
.sg_list
= priv
->tx_sge
;
227 priv
->tx_wr
.wr
.send_flags
= IB_SEND_SIGNALED
;
229 priv
->rx_sge
[0].lkey
= priv
->pd
->local_dma_lkey
;
231 priv
->rx_sge
[0].length
= IPOIB_UD_BUF_SIZE(priv
->max_ib_mtu
);
232 priv
->rx_wr
.num_sge
= 1;
234 priv
->rx_wr
.next
= NULL
;
235 priv
->rx_wr
.sg_list
= priv
->rx_sge
;
237 priv
->max_send_sge
= init_attr
.cap
.max_send_sge
;
242 ib_destroy_cq(priv
->send_cq
);
245 ib_destroy_cq(priv
->recv_cq
);
248 ipoib_cm_dev_cleanup(dev
);
251 destroy_workqueue(priv
->wq
);
255 ib_dealloc_pd(priv
->pd
);
260 void ipoib_transport_dev_cleanup(struct net_device
*dev
)
262 struct ipoib_dev_priv
*priv
= netdev_priv(dev
);
265 if (ib_destroy_qp(priv
->qp
))
266 ipoib_warn(priv
, "ib_qp_destroy failed\n");
269 clear_bit(IPOIB_PKEY_ASSIGNED
, &priv
->flags
);
272 if (ib_destroy_cq(priv
->send_cq
))
273 ipoib_warn(priv
, "ib_cq_destroy (send) failed\n");
275 if (ib_destroy_cq(priv
->recv_cq
))
276 ipoib_warn(priv
, "ib_cq_destroy (recv) failed\n");
278 ipoib_cm_dev_cleanup(dev
);
281 flush_workqueue(priv
->wq
);
282 destroy_workqueue(priv
->wq
);
286 ib_dealloc_pd(priv
->pd
);
289 void ipoib_event(struct ib_event_handler
*handler
,
290 struct ib_event
*record
)
292 struct ipoib_dev_priv
*priv
=
293 container_of(handler
, struct ipoib_dev_priv
, event_handler
);
295 if (record
->element
.port_num
!= priv
->port
)
298 ipoib_dbg(priv
, "Event %d on device %s port %d\n", record
->event
,
299 record
->device
->name
, record
->element
.port_num
);
301 if (record
->event
== IB_EVENT_SM_CHANGE
||
302 record
->event
== IB_EVENT_CLIENT_REREGISTER
) {
303 queue_work(ipoib_workqueue
, &priv
->flush_light
);
304 } else if (record
->event
== IB_EVENT_PORT_ERR
||
305 record
->event
== IB_EVENT_PORT_ACTIVE
||
306 record
->event
== IB_EVENT_LID_CHANGE
) {
307 queue_work(ipoib_workqueue
, &priv
->flush_normal
);
308 } else if (record
->event
== IB_EVENT_PKEY_CHANGE
) {
309 queue_work(ipoib_workqueue
, &priv
->flush_heavy
);