1 /* SPDX-License-Identifier: (GPL-2.0 OR BSD-3-Clause) */
3 * Copyright(c) 2018 Intel Corporation.
6 #ifndef HFI1_TID_RDMA_H
7 #define HFI1_TID_RDMA_H
9 #include <linux/circ_buf.h>
12 /* Add a convenience helper */
13 #define CIRC_ADD(val, add, size) (((val) + (add)) & ((size) - 1))
14 #define CIRC_NEXT(val, size) CIRC_ADD(val, 1, size)
15 #define CIRC_PREV(val, size) CIRC_ADD(val, -1, size)
17 #define TID_RDMA_MIN_SEGMENT_SIZE BIT(18) /* 256 KiB (for now) */
18 #define TID_RDMA_MAX_SEGMENT_SIZE BIT(18) /* 256 KiB (for now) */
19 #define TID_RDMA_MAX_PAGES (BIT(18) >> PAGE_SHIFT)
20 #define TID_RDMA_SEGMENT_SHIFT 18
23 * Bit definitions for priv->s_flags.
24 * These bit flags overload the bit flags defined for the QP's s_flags.
25 * Due to the fact that these bit fields are used only for the QP priv
26 * s_flags, there are no collisions.
28 * HFI1_S_TID_WAIT_INTERLCK - QP is waiting for requester interlock
29 * HFI1_R_TID_WAIT_INTERLCK - QP is waiting for responder interlock
31 #define HFI1_S_TID_BUSY_SET BIT(0)
32 /* BIT(1) reserved for RVT_S_BUSY. */
33 #define HFI1_R_TID_RSC_TIMER BIT(2)
34 /* BIT(3) reserved for RVT_S_RESP_PENDING. */
35 /* BIT(4) reserved for RVT_S_ACK_PENDING. */
36 #define HFI1_S_TID_WAIT_INTERLCK BIT(5)
37 #define HFI1_R_TID_WAIT_INTERLCK BIT(6)
38 /* BIT(7) - BIT(15) reserved for RVT_S_WAIT_*. */
39 /* BIT(16) reserved for RVT_S_SEND_ONE */
40 #define HFI1_S_TID_RETRY_TIMER BIT(17)
41 /* BIT(18) reserved for RVT_S_ECN. */
42 #define HFI1_R_TID_SW_PSN BIT(19)
43 /* BIT(26) reserved for HFI1_S_WAIT_HALT */
44 /* BIT(27) reserved for HFI1_S_WAIT_TID_RESP */
45 /* BIT(28) reserved for HFI1_S_WAIT_TID_SPACE */
48 * Unlike regular IB RDMA VERBS, which do not require an entry
49 * in the s_ack_queue, TID RDMA WRITE requests do because they
51 * Therefore, the s_ack_queue needs to be extended by a certain
52 * amount. The key point is that the queue needs to be extended
53 * without letting the "user" know so they user doesn't end up
54 * using these extra entries.
56 #define HFI1_TID_RDMA_WRITE_CNT 8
58 struct tid_rdma_params
{
59 struct rcu_head rcu_head
;
70 struct tid_rdma_qp_params
{
71 struct work_struct trigger_work
;
72 struct tid_rdma_params local
;
73 struct tid_rdma_params __rcu
*remote
;
76 /* Track state for each hardware flow */
77 struct tid_flow_state
{
84 enum tid_rdma_req_state
{
85 TID_REQUEST_INACTIVE
= 0,
87 TID_REQUEST_INIT_RESEND
,
90 TID_REQUEST_RESEND_ACTIVE
,
97 struct tid_rdma_request
{
99 struct hfi1_ctxtdata
*rcd
;
101 struct rvt_swqe
*swqe
;
102 struct rvt_ack_entry
*ack
;
105 struct tid_rdma_flow
*flows
; /* array of tid flows */
106 struct rvt_sge_state ss
; /* SGE state for TID RDMA requests */
107 u16 n_flows
; /* size of the flow buffer window */
108 u16 setup_head
; /* flow index we are setting up */
109 u16 clear_tail
; /* flow index we are clearing */
110 u16 flow_idx
; /* flow index most recently set up */
115 u32 r_ack_psn
; /* next expected ack PSN */
116 u32 r_flow_psn
; /* IB PSN of next segment start */
117 u32 r_last_acked
; /* IB PSN of last ACK'ed packet */
118 u32 s_next_psn
; /* IB PSN of next segment start for read */
120 u32 total_segs
; /* segments required to complete a request */
121 u32 cur_seg
; /* index of current segment */
122 u32 comp_seg
; /* index of last completed segment */
123 u32 ack_seg
; /* index of last ack'ed segment */
124 u32 alloc_seg
; /* index of next segment to be allocated */
125 u32 isge
; /* index of "current" sge */
126 u32 ack_pending
; /* num acks pending for this request */
128 enum tid_rdma_req_state state
;
132 * When header suppression is used, PSNs associated with a "flow" are
133 * relevant (and not the PSNs maintained by verbs). Track per-flow
134 * PSNs here for a TID RDMA segment.
139 u32 resp_ib_psn
; /* The IB PSN of the response for this flow */
140 u32 generation
; /* generation of flow */
141 u32 spsn
; /* starting PSN in TID space */
142 u32 lpsn
; /* last PSN in TID space */
143 u32 r_next_psn
; /* next PSN to be received (in TID space) */
145 /* For tid rdma read */
146 u32 ib_spsn
; /* starting PSN in Verbs space */
147 u32 ib_lpsn
; /* last PSn in Verbs space */
150 struct tid_rdma_pageset
{
151 dma_addr_t addr
: 48; /* Only needed for the first page */
158 * kern_tid_node - used for managing TID's in TID groups
160 * @grp_idx: rcd relative index to tid_group
161 * @map: grp->map captured prior to programming this TID group in HW
162 * @cnt: Only @cnt of available group entries are actually programmed
164 struct kern_tid_node
{
165 struct tid_group
*grp
;
170 /* Overall info for a TID RDMA segment */
171 struct tid_rdma_flow
{
173 * While a TID RDMA segment is being transferred, it uses a QP number
174 * from the "KDETH section of QP numbers" (which is different from the
175 * QP number that originated the request). Bits 11-15 of these QP
176 * numbers identify the "TID flow" for the segment.
178 struct flow_state flow_state
;
179 struct tid_rdma_request
*req
;
192 struct kern_tid_node tnode
[TID_RDMA_MAX_PAGES
];
193 struct tid_rdma_pageset pagesets
[TID_RDMA_MAX_PAGES
];
194 u32 tid_entry
[TID_RDMA_MAX_PAGES
];
197 enum tid_rnr_nak_state
{
198 TID_RNR_NAK_INIT
= 0,
203 bool tid_rdma_conn_req(struct rvt_qp
*qp
, u64
*data
);
204 bool tid_rdma_conn_reply(struct rvt_qp
*qp
, u64 data
);
205 bool tid_rdma_conn_resp(struct rvt_qp
*qp
, u64
*data
);
206 void tid_rdma_conn_error(struct rvt_qp
*qp
);
207 void tid_rdma_opfn_init(struct rvt_qp
*qp
, struct tid_rdma_params
*p
);
209 int hfi1_kern_exp_rcv_init(struct hfi1_ctxtdata
*rcd
, int reinit
);
210 int hfi1_kern_exp_rcv_setup(struct tid_rdma_request
*req
,
211 struct rvt_sge_state
*ss
, bool *last
);
212 int hfi1_kern_exp_rcv_clear(struct tid_rdma_request
*req
);
213 void hfi1_kern_exp_rcv_clear_all(struct tid_rdma_request
*req
);
214 void __trdma_clean_swqe(struct rvt_qp
*qp
, struct rvt_swqe
*wqe
);
217 * trdma_clean_swqe - clean flows for swqe if large send queue
221 static inline void trdma_clean_swqe(struct rvt_qp
*qp
, struct rvt_swqe
*wqe
)
225 __trdma_clean_swqe(qp
, wqe
);
228 void hfi1_kern_read_tid_flow_free(struct rvt_qp
*qp
);
230 int hfi1_qp_priv_init(struct rvt_dev_info
*rdi
, struct rvt_qp
*qp
,
231 struct ib_qp_init_attr
*init_attr
);
232 void hfi1_qp_priv_tid_free(struct rvt_dev_info
*rdi
, struct rvt_qp
*qp
);
234 void hfi1_tid_rdma_flush_wait(struct rvt_qp
*qp
);
236 int hfi1_kern_setup_hw_flow(struct hfi1_ctxtdata
*rcd
, struct rvt_qp
*qp
);
237 void hfi1_kern_clear_hw_flow(struct hfi1_ctxtdata
*rcd
, struct rvt_qp
*qp
);
238 void hfi1_kern_init_ctxt_generations(struct hfi1_ctxtdata
*rcd
);
241 u64
hfi1_access_sw_tid_wait(const struct cntr_entry
*entry
,
242 void *context
, int vl
, int mode
, u64 data
);
244 u32
hfi1_build_tid_rdma_read_packet(struct rvt_swqe
*wqe
,
245 struct ib_other_headers
*ohdr
,
246 u32
*bth1
, u32
*bth2
, u32
*len
);
247 u32
hfi1_build_tid_rdma_read_req(struct rvt_qp
*qp
, struct rvt_swqe
*wqe
,
248 struct ib_other_headers
*ohdr
, u32
*bth1
,
249 u32
*bth2
, u32
*len
);
250 void hfi1_rc_rcv_tid_rdma_read_req(struct hfi1_packet
*packet
);
251 u32
hfi1_build_tid_rdma_read_resp(struct rvt_qp
*qp
, struct rvt_ack_entry
*e
,
252 struct ib_other_headers
*ohdr
, u32
*bth0
,
253 u32
*bth1
, u32
*bth2
, u32
*len
, bool *last
);
254 void hfi1_rc_rcv_tid_rdma_read_resp(struct hfi1_packet
*packet
);
255 bool hfi1_handle_kdeth_eflags(struct hfi1_ctxtdata
*rcd
,
256 struct hfi1_pportdata
*ppd
,
257 struct hfi1_packet
*packet
);
258 void hfi1_tid_rdma_restart_req(struct rvt_qp
*qp
, struct rvt_swqe
*wqe
,
260 void hfi1_qp_kern_exp_rcv_clear_all(struct rvt_qp
*qp
);
261 bool hfi1_tid_rdma_wqe_interlock(struct rvt_qp
*qp
, struct rvt_swqe
*wqe
);
263 void setup_tid_rdma_wqe(struct rvt_qp
*qp
, struct rvt_swqe
*wqe
);
264 static inline void hfi1_setup_tid_rdma_wqe(struct rvt_qp
*qp
,
265 struct rvt_swqe
*wqe
)
268 (wqe
->wr
.opcode
== IB_WR_RDMA_READ
||
269 wqe
->wr
.opcode
== IB_WR_RDMA_WRITE
) &&
270 wqe
->length
>= TID_RDMA_MIN_SEGMENT_SIZE
)
271 setup_tid_rdma_wqe(qp
, wqe
);
274 u32
hfi1_build_tid_rdma_write_req(struct rvt_qp
*qp
, struct rvt_swqe
*wqe
,
275 struct ib_other_headers
*ohdr
,
276 u32
*bth1
, u32
*bth2
, u32
*len
);
278 void hfi1_rc_rcv_tid_rdma_write_req(struct hfi1_packet
*packet
);
280 u32
hfi1_build_tid_rdma_write_resp(struct rvt_qp
*qp
, struct rvt_ack_entry
*e
,
281 struct ib_other_headers
*ohdr
, u32
*bth1
,
283 struct rvt_sge_state
**ss
);
285 void hfi1_del_tid_reap_timer(struct rvt_qp
*qp
);
287 void hfi1_rc_rcv_tid_rdma_write_resp(struct hfi1_packet
*packet
);
289 bool hfi1_build_tid_rdma_packet(struct rvt_swqe
*wqe
,
290 struct ib_other_headers
*ohdr
,
291 u32
*bth1
, u32
*bth2
, u32
*len
);
293 void hfi1_rc_rcv_tid_rdma_write_data(struct hfi1_packet
*packet
);
295 u32
hfi1_build_tid_rdma_write_ack(struct rvt_qp
*qp
, struct rvt_ack_entry
*e
,
296 struct ib_other_headers
*ohdr
, u16 iflow
,
297 u32
*bth1
, u32
*bth2
);
299 void hfi1_rc_rcv_tid_rdma_ack(struct hfi1_packet
*packet
);
301 void hfi1_add_tid_retry_timer(struct rvt_qp
*qp
);
302 void hfi1_del_tid_retry_timer(struct rvt_qp
*qp
);
304 u32
hfi1_build_tid_rdma_resync(struct rvt_qp
*qp
, struct rvt_swqe
*wqe
,
305 struct ib_other_headers
*ohdr
, u32
*bth1
,
306 u32
*bth2
, u16 fidx
);
308 void hfi1_rc_rcv_tid_rdma_resync(struct hfi1_packet
*packet
);
310 struct hfi1_pkt_state
;
311 int hfi1_make_tid_rdma_pkt(struct rvt_qp
*qp
, struct hfi1_pkt_state
*ps
);
313 void _hfi1_do_tid_send(struct work_struct
*work
);
315 bool hfi1_schedule_tid_send(struct rvt_qp
*qp
);
317 bool hfi1_tid_rdma_ack_interlock(struct rvt_qp
*qp
, struct rvt_ack_entry
*e
);
319 #endif /* HFI1_TID_RDMA_H */