2 * cxgb3i_offload.c: Chelsio S3xx iscsi offloaded tcp connection management
4 * Copyright (C) 2003-2008 Chelsio Communications. All rights reserved.
6 * This program is distributed in the hope that it will be useful, but WITHOUT
7 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
8 * FITNESS FOR A PARTICULAR PURPOSE. See the LICENSE file included in this
9 * release for licensing terms and conditions.
11 * Written by: Dimitris Michailidis (dm@chelsio.com)
12 * Karen Xie (kxie@chelsio.com)
15 #include <linux/if_vlan.h>
16 #include <linux/version.h>
18 #include "cxgb3_defs.h"
19 #include "cxgb3_ctl_defs.h"
20 #include "firmware_exports.h"
21 #include "cxgb3i_offload.h"
22 #include "cxgb3i_pdu.h"
23 #include "cxgb3i_ddp.h"
25 #ifdef __DEBUG_C3CN_CONN__
26 #define c3cn_conn_debug cxgb3i_log_debug
28 #define c3cn_conn_debug(fmt...)
31 #ifdef __DEBUG_C3CN_TX__
32 #define c3cn_tx_debug cxgb3i_log_debug
34 #define c3cn_tx_debug(fmt...)
37 #ifdef __DEBUG_C3CN_RX__
38 #define c3cn_rx_debug cxgb3i_log_debug
40 #define c3cn_rx_debug(fmt...)
44 * module parameters releated to offloaded iscsi connection
46 static int cxgb3_rcv_win
= 256 * 1024;
47 module_param(cxgb3_rcv_win
, int, 0644);
48 MODULE_PARM_DESC(cxgb3_rcv_win
, "TCP receive window in bytes (default=256KB)");
50 static int cxgb3_snd_win
= 128 * 1024;
51 module_param(cxgb3_snd_win
, int, 0644);
52 MODULE_PARM_DESC(cxgb3_snd_win
, "TCP send window in bytes (default=128KB)");
54 static int cxgb3_rx_credit_thres
= 10 * 1024;
55 module_param(cxgb3_rx_credit_thres
, int, 0644);
56 MODULE_PARM_DESC(rx_credit_thres
,
57 "RX credits return threshold in bytes (default=10KB)");
59 static unsigned int cxgb3_max_connect
= 8 * 1024;
60 module_param(cxgb3_max_connect
, uint
, 0644);
61 MODULE_PARM_DESC(cxgb3_max_connect
, "Max. # of connections (default=8092)");
63 static unsigned int cxgb3_sport_base
= 20000;
64 module_param(cxgb3_sport_base
, uint
, 0644);
65 MODULE_PARM_DESC(cxgb3_sport_base
, "starting port number (default=20000)");
68 * cxgb3i tcp connection data(per adapter) list
70 static LIST_HEAD(cdata_list
);
71 static DEFINE_RWLOCK(cdata_rwlock
);
73 static int c3cn_push_tx_frames(struct s3_conn
*c3cn
, int req_completion
);
74 static void c3cn_release_offload_resources(struct s3_conn
*c3cn
);
77 * iscsi source port management
79 * Find a free source port in the port allocation map. We use a very simple
80 * rotor scheme to look for the next free port.
82 * If a source port has been specified make sure that it doesn't collide with
83 * our normal source port allocation map. If it's outside the range of our
84 * allocation/deallocation scheme just let them use it.
86 * If the source port is outside our allocation range, the caller is
87 * responsible for keeping track of their port usage.
89 static int c3cn_get_port(struct s3_conn
*c3cn
, struct cxgb3i_sdev_data
*cdata
)
97 if (c3cn
->saddr
.sin_port
) {
98 cxgb3i_log_error("connect, sin_port NON-ZERO %u.\n",
99 c3cn
->saddr
.sin_port
);
103 spin_lock_bh(&cdata
->lock
);
104 start
= idx
= cdata
->sport_next
;
106 if (++idx
>= cxgb3_max_connect
)
108 if (!cdata
->sport_conn
[idx
]) {
109 c3cn
->saddr
.sin_port
= htons(cxgb3_sport_base
+ idx
);
110 cdata
->sport_next
= idx
;
111 cdata
->sport_conn
[idx
] = c3cn
;
112 spin_unlock_bh(&cdata
->lock
);
114 c3cn_conn_debug("%s reserve port %u.\n",
116 cxgb3_sport_base
+ idx
);
119 } while (idx
!= start
);
120 spin_unlock_bh(&cdata
->lock
);
123 return -EADDRNOTAVAIL
;
126 static void c3cn_put_port(struct s3_conn
*c3cn
)
131 if (c3cn
->saddr
.sin_port
) {
132 struct cxgb3i_sdev_data
*cdata
= CXGB3_SDEV_DATA(c3cn
->cdev
);
133 int idx
= ntohs(c3cn
->saddr
.sin_port
) - cxgb3_sport_base
;
135 c3cn
->saddr
.sin_port
= 0;
136 if (idx
< 0 || idx
>= cxgb3_max_connect
)
138 spin_lock_bh(&cdata
->lock
);
139 cdata
->sport_conn
[idx
] = NULL
;
140 spin_unlock_bh(&cdata
->lock
);
141 c3cn_conn_debug("%s, release port %u.\n",
142 cdata
->cdev
->name
, cxgb3_sport_base
+ idx
);
146 static inline void c3cn_set_flag(struct s3_conn
*c3cn
, enum c3cn_flags flag
)
148 __set_bit(flag
, &c3cn
->flags
);
149 c3cn_conn_debug("c3cn 0x%p, set %d, s %u, f 0x%lx.\n",
150 c3cn
, flag
, c3cn
->state
, c3cn
->flags
);
153 static inline void c3cn_clear_flag(struct s3_conn
*c3cn
, enum c3cn_flags flag
)
155 __clear_bit(flag
, &c3cn
->flags
);
156 c3cn_conn_debug("c3cn 0x%p, clear %d, s %u, f 0x%lx.\n",
157 c3cn
, flag
, c3cn
->state
, c3cn
->flags
);
160 static inline int c3cn_flag(struct s3_conn
*c3cn
, enum c3cn_flags flag
)
164 return test_bit(flag
, &c3cn
->flags
);
167 static void c3cn_set_state(struct s3_conn
*c3cn
, int state
)
169 c3cn_conn_debug("c3cn 0x%p state -> %u.\n", c3cn
, state
);
173 static inline void c3cn_hold(struct s3_conn
*c3cn
)
175 atomic_inc(&c3cn
->refcnt
);
178 static inline void c3cn_put(struct s3_conn
*c3cn
)
180 if (atomic_dec_and_test(&c3cn
->refcnt
)) {
181 c3cn_conn_debug("free c3cn 0x%p, s %u, f 0x%lx.\n",
182 c3cn
, c3cn
->state
, c3cn
->flags
);
187 static void c3cn_closed(struct s3_conn
*c3cn
)
189 c3cn_conn_debug("c3cn 0x%p, state %u, flag 0x%lx.\n",
190 c3cn
, c3cn
->state
, c3cn
->flags
);
193 c3cn_release_offload_resources(c3cn
);
194 c3cn_set_state(c3cn
, C3CN_STATE_CLOSED
);
195 cxgb3i_conn_closing(c3cn
);
199 * CPL (Chelsio Protocol Language) defines a message passing interface between
200 * the host driver and T3 asic.
201 * The section below implments CPLs that related to iscsi tcp connection
202 * open/close/abort and data send/receive.
206 * CPL connection active open request: host ->
208 static unsigned int find_best_mtu(const struct t3c_data
*d
, unsigned short mtu
)
212 while (i
< d
->nmtus
- 1 && d
->mtus
[i
+ 1] <= mtu
)
217 static unsigned int select_mss(struct s3_conn
*c3cn
, unsigned int pmtu
)
220 struct dst_entry
*dst
= c3cn
->dst_cache
;
221 struct t3cdev
*cdev
= c3cn
->cdev
;
222 const struct t3c_data
*td
= T3C_DATA(cdev
);
223 u16 advmss
= dst_metric(dst
, RTAX_ADVMSS
);
225 if (advmss
> pmtu
- 40)
227 if (advmss
< td
->mtus
[0] - 40)
228 advmss
= td
->mtus
[0] - 40;
229 idx
= find_best_mtu(td
, advmss
+ 40);
233 static inline int compute_wscale(int win
)
236 while (wscale
< 14 && (65535<<wscale
) < win
)
241 static inline unsigned int calc_opt0h(struct s3_conn
*c3cn
)
243 int wscale
= compute_wscale(cxgb3_rcv_win
);
244 return V_KEEP_ALIVE(1) |
246 V_WND_SCALE(wscale
) |
247 V_MSS_IDX(c3cn
->mss_idx
);
250 static inline unsigned int calc_opt0l(struct s3_conn
*c3cn
)
252 return V_ULP_MODE(ULP_MODE_ISCSI
) |
253 V_RCV_BUFSIZ(cxgb3_rcv_win
>>10);
256 static void make_act_open_req(struct s3_conn
*c3cn
, struct sk_buff
*skb
,
257 unsigned int atid
, const struct l2t_entry
*e
)
259 struct cpl_act_open_req
*req
;
261 c3cn_conn_debug("c3cn 0x%p, atid 0x%x.\n", c3cn
, atid
);
263 skb
->priority
= CPL_PRIORITY_SETUP
;
264 req
= (struct cpl_act_open_req
*)__skb_put(skb
, sizeof(*req
));
265 req
->wr
.wr_hi
= htonl(V_WR_OP(FW_WROPCODE_FORWARD
));
266 OPCODE_TID(req
) = htonl(MK_OPCODE_TID(CPL_ACT_OPEN_REQ
, atid
));
267 req
->local_port
= c3cn
->saddr
.sin_port
;
268 req
->peer_port
= c3cn
->daddr
.sin_port
;
269 req
->local_ip
= c3cn
->saddr
.sin_addr
.s_addr
;
270 req
->peer_ip
= c3cn
->daddr
.sin_addr
.s_addr
;
271 req
->opt0h
= htonl(calc_opt0h(c3cn
) | V_L2T_IDX(e
->idx
) |
272 V_TX_CHANNEL(e
->smt_idx
));
273 req
->opt0l
= htonl(calc_opt0l(c3cn
));
277 static void fail_act_open(struct s3_conn
*c3cn
, int errno
)
279 c3cn_conn_debug("c3cn 0x%p, state %u, flag 0x%lx.\n",
280 c3cn
, c3cn
->state
, c3cn
->flags
);
285 static void act_open_req_arp_failure(struct t3cdev
*dev
, struct sk_buff
*skb
)
287 struct s3_conn
*c3cn
= (struct s3_conn
*)skb
->sk
;
289 c3cn_conn_debug("c3cn 0x%p, state %u.\n", c3cn
, c3cn
->state
);
292 spin_lock_bh(&c3cn
->lock
);
293 if (c3cn
->state
== C3CN_STATE_CONNECTING
)
294 fail_act_open(c3cn
, -EHOSTUNREACH
);
295 spin_unlock_bh(&c3cn
->lock
);
301 * CPL connection close request: host ->
303 * Close a connection by sending a CPL_CLOSE_CON_REQ message and queue it to
304 * the write queue (i.e., after any unsent txt data).
306 static void skb_entail(struct s3_conn
*c3cn
, struct sk_buff
*skb
,
309 skb_tcp_seq(skb
) = c3cn
->write_seq
;
310 skb_flags(skb
) = flags
;
311 __skb_queue_tail(&c3cn
->write_queue
, skb
);
314 static void send_close_req(struct s3_conn
*c3cn
)
316 struct sk_buff
*skb
= c3cn
->cpl_close
;
317 struct cpl_close_con_req
*req
= (struct cpl_close_con_req
*)skb
->head
;
318 unsigned int tid
= c3cn
->tid
;
320 c3cn_conn_debug("c3cn 0x%p, state 0x%x, flag 0x%lx.\n",
321 c3cn
, c3cn
->state
, c3cn
->flags
);
323 c3cn
->cpl_close
= NULL
;
325 req
->wr
.wr_hi
= htonl(V_WR_OP(FW_WROPCODE_OFLD_CLOSE_CON
));
326 req
->wr
.wr_lo
= htonl(V_WR_TID(tid
));
327 OPCODE_TID(req
) = htonl(MK_OPCODE_TID(CPL_CLOSE_CON_REQ
, tid
));
328 req
->rsvd
= htonl(c3cn
->write_seq
);
330 skb_entail(c3cn
, skb
, C3CB_FLAG_NO_APPEND
);
331 if (c3cn
->state
!= C3CN_STATE_CONNECTING
)
332 c3cn_push_tx_frames(c3cn
, 1);
336 * CPL connection abort request: host ->
338 * Send an ABORT_REQ message. Makes sure we do not send multiple ABORT_REQs
339 * for the same connection and also that we do not try to send a message
340 * after the connection has closed.
342 static void abort_arp_failure(struct t3cdev
*cdev
, struct sk_buff
*skb
)
344 struct cpl_abort_req
*req
= cplhdr(skb
);
346 c3cn_conn_debug("tdev 0x%p.\n", cdev
);
348 req
->cmd
= CPL_ABORT_NO_RST
;
349 cxgb3_ofld_send(cdev
, skb
);
352 static inline void c3cn_purge_write_queue(struct s3_conn
*c3cn
)
356 while ((skb
= __skb_dequeue(&c3cn
->write_queue
)))
360 static void send_abort_req(struct s3_conn
*c3cn
)
362 struct sk_buff
*skb
= c3cn
->cpl_abort_req
;
363 struct cpl_abort_req
*req
;
364 unsigned int tid
= c3cn
->tid
;
366 if (unlikely(c3cn
->state
== C3CN_STATE_ABORTING
) || !skb
||
370 c3cn_set_state(c3cn
, C3CN_STATE_ABORTING
);
372 c3cn_conn_debug("c3cn 0x%p, flag ABORT_RPL + ABORT_SHUT.\n", c3cn
);
374 c3cn_set_flag(c3cn
, C3CN_ABORT_RPL_PENDING
);
376 /* Purge the send queue so we don't send anything after an abort. */
377 c3cn_purge_write_queue(c3cn
);
379 c3cn
->cpl_abort_req
= NULL
;
380 req
= (struct cpl_abort_req
*)skb
->head
;
382 skb
->priority
= CPL_PRIORITY_DATA
;
383 set_arp_failure_handler(skb
, abort_arp_failure
);
385 req
->wr
.wr_hi
= htonl(V_WR_OP(FW_WROPCODE_OFLD_HOST_ABORT_CON_REQ
));
386 req
->wr
.wr_lo
= htonl(V_WR_TID(tid
));
387 OPCODE_TID(req
) = htonl(MK_OPCODE_TID(CPL_ABORT_REQ
, tid
));
388 req
->rsvd0
= htonl(c3cn
->snd_nxt
);
389 req
->rsvd1
= !c3cn_flag(c3cn
, C3CN_TX_DATA_SENT
);
390 req
->cmd
= CPL_ABORT_SEND_RST
;
392 l2t_send(c3cn
->cdev
, skb
, c3cn
->l2t
);
396 * CPL connection abort reply: host ->
398 * Send an ABORT_RPL message in response of the ABORT_REQ received.
400 static void send_abort_rpl(struct s3_conn
*c3cn
, int rst_status
)
402 struct sk_buff
*skb
= c3cn
->cpl_abort_rpl
;
403 struct cpl_abort_rpl
*rpl
= (struct cpl_abort_rpl
*)skb
->head
;
405 c3cn
->cpl_abort_rpl
= NULL
;
407 skb
->priority
= CPL_PRIORITY_DATA
;
408 rpl
->wr
.wr_hi
= htonl(V_WR_OP(FW_WROPCODE_OFLD_HOST_ABORT_CON_RPL
));
409 rpl
->wr
.wr_lo
= htonl(V_WR_TID(c3cn
->tid
));
410 OPCODE_TID(rpl
) = htonl(MK_OPCODE_TID(CPL_ABORT_RPL
, c3cn
->tid
));
411 rpl
->cmd
= rst_status
;
413 cxgb3_ofld_send(c3cn
->cdev
, skb
);
417 * CPL connection rx data ack: host ->
418 * Send RX credits through an RX_DATA_ACK CPL message. Returns the number of
421 static u32
send_rx_credits(struct s3_conn
*c3cn
, u32 credits
, u32 dack
)
424 struct cpl_rx_data_ack
*req
;
426 skb
= alloc_skb(sizeof(*req
), GFP_ATOMIC
);
430 req
= (struct cpl_rx_data_ack
*)__skb_put(skb
, sizeof(*req
));
431 req
->wr
.wr_hi
= htonl(V_WR_OP(FW_WROPCODE_FORWARD
));
432 OPCODE_TID(req
) = htonl(MK_OPCODE_TID(CPL_RX_DATA_ACK
, c3cn
->tid
));
433 req
->credit_dack
= htonl(dack
| V_RX_CREDITS(credits
));
434 skb
->priority
= CPL_PRIORITY_ACK
;
435 cxgb3_ofld_send(c3cn
->cdev
, skb
);
440 * CPL connection tx data: host ->
442 * Send iscsi PDU via TX_DATA CPL message. Returns the number of
444 * Each TX_DATA consumes work request credit (wrs), so we need to keep track of
445 * how many we've used so far and how many are pending (i.e., yet ack'ed by T3).
449 * For ULP connections HW may inserts digest bytes into the pdu. Those digest
450 * bytes are not sent by the host but are part of the TCP payload and therefore
451 * consume TCP sequence space.
453 static const unsigned int cxgb3_ulp_extra_len
[] = { 0, 4, 4, 8 };
454 static inline unsigned int ulp_extra_len(const struct sk_buff
*skb
)
456 return cxgb3_ulp_extra_len
[skb_ulp_mode(skb
) & 3];
459 static unsigned int wrlen __read_mostly
;
462 * The number of WRs needed for an skb depends on the number of fragments
463 * in the skb and whether it has any payload in its main body. This maps the
464 * length of the gather list represented by an skb into the # of necessary WRs.
465 * The extra two fragments are for iscsi bhs and payload padding.
467 #define SKB_WR_LIST_SIZE (MAX_SKB_FRAGS + 2)
468 static unsigned int skb_wrs
[SKB_WR_LIST_SIZE
] __read_mostly
;
470 static void s3_init_wr_tab(unsigned int wr_len
)
474 if (skb_wrs
[1]) /* already initialized */
477 for (i
= 1; i
< SKB_WR_LIST_SIZE
; i
++) {
478 int sgl_len
= (3 * i
) / 2 + (i
& 1);
481 skb_wrs
[i
] = (sgl_len
<= wr_len
482 ? 1 : 1 + (sgl_len
- 2) / (wr_len
- 1));
488 static inline void reset_wr_list(struct s3_conn
*c3cn
)
490 c3cn
->wr_pending_head
= c3cn
->wr_pending_tail
= NULL
;
494 * Add a WR to a connections's list of pending WRs. This is a singly-linked
495 * list of sk_buffs operating as a FIFO. The head is kept in wr_pending_head
496 * and the tail in wr_pending_tail.
498 static inline void enqueue_wr(struct s3_conn
*c3cn
,
501 skb_tx_wr_next(skb
) = NULL
;
504 * We want to take an extra reference since both us and the driver
505 * need to free the packet before it's really freed. We know there's
506 * just one user currently so we use atomic_set rather than skb_get
507 * to avoid the atomic op.
509 atomic_set(&skb
->users
, 2);
511 if (!c3cn
->wr_pending_head
)
512 c3cn
->wr_pending_head
= skb
;
514 skb_tx_wr_next(c3cn
->wr_pending_tail
) = skb
;
515 c3cn
->wr_pending_tail
= skb
;
518 static int count_pending_wrs(struct s3_conn
*c3cn
)
521 const struct sk_buff
*skb
= c3cn
->wr_pending_head
;
525 skb
= skb_tx_wr_next(skb
);
530 static inline struct sk_buff
*peek_wr(const struct s3_conn
*c3cn
)
532 return c3cn
->wr_pending_head
;
535 static inline void free_wr_skb(struct sk_buff
*skb
)
540 static inline struct sk_buff
*dequeue_wr(struct s3_conn
*c3cn
)
542 struct sk_buff
*skb
= c3cn
->wr_pending_head
;
545 /* Don't bother clearing the tail */
546 c3cn
->wr_pending_head
= skb_tx_wr_next(skb
);
547 skb_tx_wr_next(skb
) = NULL
;
552 static void purge_wr_queue(struct s3_conn
*c3cn
)
555 while ((skb
= dequeue_wr(c3cn
)) != NULL
)
559 static inline void make_tx_data_wr(struct s3_conn
*c3cn
, struct sk_buff
*skb
,
560 int len
, int req_completion
)
562 struct tx_data_wr
*req
;
564 skb_reset_transport_header(skb
);
565 req
= (struct tx_data_wr
*)__skb_push(skb
, sizeof(*req
));
566 req
->wr_hi
= htonl(V_WR_OP(FW_WROPCODE_OFLD_TX_DATA
) |
567 (req_completion
? F_WR_COMPL
: 0));
568 req
->wr_lo
= htonl(V_WR_TID(c3cn
->tid
));
569 req
->sndseq
= htonl(c3cn
->snd_nxt
);
570 /* len includes the length of any HW ULP additions */
571 req
->len
= htonl(len
);
572 req
->param
= htonl(V_TX_PORT(c3cn
->l2t
->smt_idx
));
573 /* V_TX_ULP_SUBMODE sets both the mode and submode */
574 req
->flags
= htonl(V_TX_ULP_SUBMODE(skb_ulp_mode(skb
)) |
575 V_TX_SHOVE((skb_peek(&c3cn
->write_queue
) ? 0 : 1)));
577 if (!c3cn_flag(c3cn
, C3CN_TX_DATA_SENT
)) {
578 req
->flags
|= htonl(V_TX_ACK_PAGES(2) | F_TX_INIT
|
579 V_TX_CPU_IDX(c3cn
->qset
));
580 /* Sendbuffer is in units of 32KB. */
581 req
->param
|= htonl(V_TX_SNDBUF(cxgb3_snd_win
>> 15));
582 c3cn_set_flag(c3cn
, C3CN_TX_DATA_SENT
);
587 * c3cn_push_tx_frames -- start transmit
588 * @c3cn: the offloaded connection
589 * @req_completion: request wr_ack or not
591 * Prepends TX_DATA_WR or CPL_CLOSE_CON_REQ headers to buffers waiting in a
592 * connection's send queue and sends them on to T3. Must be called with the
593 * connection's lock held. Returns the amount of send buffer space that was
594 * freed as a result of sending queued data to T3.
596 static void arp_failure_discard(struct t3cdev
*cdev
, struct sk_buff
*skb
)
601 static int c3cn_push_tx_frames(struct s3_conn
*c3cn
, int req_completion
)
606 struct cxgb3i_sdev_data
*cdata
;
608 if (unlikely(c3cn
->state
== C3CN_STATE_CONNECTING
||
609 c3cn
->state
== C3CN_STATE_CLOSE_WAIT_1
||
610 c3cn
->state
>= C3CN_STATE_ABORTING
)) {
611 c3cn_tx_debug("c3cn 0x%p, in closing state %u.\n",
617 cdata
= CXGB3_SDEV_DATA(cdev
);
619 while (c3cn
->wr_avail
620 && (skb
= skb_peek(&c3cn
->write_queue
)) != NULL
) {
621 int len
= skb
->len
; /* length before skb_push */
622 int frags
= skb_shinfo(skb
)->nr_frags
+ (len
!= skb
->data_len
);
623 int wrs_needed
= skb_wrs
[frags
];
625 if (wrs_needed
> 1 && len
+ sizeof(struct tx_data_wr
) <= wrlen
)
628 WARN_ON(frags
>= SKB_WR_LIST_SIZE
|| wrs_needed
< 1);
630 if (c3cn
->wr_avail
< wrs_needed
) {
631 c3cn_tx_debug("c3cn 0x%p, skb len %u/%u, frag %u, "
633 c3cn
, skb
->len
, skb
->data_len
, frags
,
634 wrs_needed
, c3cn
->wr_avail
);
638 __skb_unlink(skb
, &c3cn
->write_queue
);
639 skb
->priority
= CPL_PRIORITY_DATA
;
640 skb
->csum
= wrs_needed
; /* remember this until the WR_ACK */
641 c3cn
->wr_avail
-= wrs_needed
;
642 c3cn
->wr_unacked
+= wrs_needed
;
643 enqueue_wr(c3cn
, skb
);
645 c3cn_tx_debug("c3cn 0x%p, enqueue, skb len %u/%u, frag %u, "
646 "wr %d, left %u, unack %u.\n",
647 c3cn
, skb
->len
, skb
->data_len
, frags
,
648 wrs_needed
, c3cn
->wr_avail
, c3cn
->wr_unacked
);
651 if (likely(skb_flags(skb
) & C3CB_FLAG_NEED_HDR
)) {
652 if ((req_completion
&&
653 c3cn
->wr_unacked
== wrs_needed
) ||
654 (skb_flags(skb
) & C3CB_FLAG_COMPL
) ||
655 c3cn
->wr_unacked
>= c3cn
->wr_max
/ 2) {
657 c3cn
->wr_unacked
= 0;
659 len
+= ulp_extra_len(skb
);
660 make_tx_data_wr(c3cn
, skb
, len
, req_completion
);
661 c3cn
->snd_nxt
+= len
;
662 skb_flags(skb
) &= ~C3CB_FLAG_NEED_HDR
;
665 total_size
+= skb
->truesize
;
666 set_arp_failure_handler(skb
, arp_failure_discard
);
667 l2t_send(cdev
, skb
, c3cn
->l2t
);
673 * process_cpl_msg: -> host
674 * Top-level CPL message processing used by most CPL messages that
675 * pertain to connections.
677 static inline void process_cpl_msg(void (*fn
)(struct s3_conn
*,
679 struct s3_conn
*c3cn
,
682 spin_lock_bh(&c3cn
->lock
);
684 spin_unlock_bh(&c3cn
->lock
);
688 * process_cpl_msg_ref: -> host
689 * Similar to process_cpl_msg() but takes an extra connection reference around
690 * the call to the handler. Should be used if the handler may drop a
691 * connection reference.
693 static inline void process_cpl_msg_ref(void (*fn
) (struct s3_conn
*,
695 struct s3_conn
*c3cn
,
699 process_cpl_msg(fn
, c3cn
, skb
);
704 * Process a CPL_ACT_ESTABLISH message: -> host
705 * Updates connection state from an active establish CPL message. Runs with
706 * the connection lock held.
709 static inline void s3_free_atid(struct t3cdev
*cdev
, unsigned int tid
)
711 struct s3_conn
*c3cn
= cxgb3_free_atid(cdev
, tid
);
716 static void c3cn_established(struct s3_conn
*c3cn
, u32 snd_isn
,
719 c3cn_conn_debug("c3cn 0x%p, state %u.\n", c3cn
, c3cn
->state
);
721 c3cn
->write_seq
= c3cn
->snd_nxt
= c3cn
->snd_una
= snd_isn
;
724 * Causes the first RX_DATA_ACK to supply any Rx credits we couldn't
727 if (cxgb3_rcv_win
> (M_RCV_BUFSIZ
<< 10))
728 c3cn
->rcv_wup
-= cxgb3_rcv_win
- (M_RCV_BUFSIZ
<< 10);
730 dst_confirm(c3cn
->dst_cache
);
734 c3cn_set_state(c3cn
, C3CN_STATE_ESTABLISHED
);
737 static void process_act_establish(struct s3_conn
*c3cn
, struct sk_buff
*skb
)
739 struct cpl_act_establish
*req
= cplhdr(skb
);
740 u32 rcv_isn
= ntohl(req
->rcv_isn
); /* real RCV_ISN + 1 */
742 c3cn_conn_debug("c3cn 0x%p, state %u, flag 0x%lx.\n",
743 c3cn
, c3cn
->state
, c3cn
->flags
);
745 if (unlikely(c3cn
->state
!= C3CN_STATE_CONNECTING
))
746 cxgb3i_log_error("TID %u expected SYN_SENT, got EST., s %u\n",
747 c3cn
->tid
, c3cn
->state
);
749 c3cn
->copied_seq
= c3cn
->rcv_wup
= c3cn
->rcv_nxt
= rcv_isn
;
750 c3cn_established(c3cn
, ntohl(req
->snd_isn
), ntohs(req
->tcp_opt
));
754 if (unlikely(c3cn_flag(c3cn
, C3CN_ACTIVE_CLOSE_NEEDED
)))
755 /* upper layer has requested closing */
756 send_abort_req(c3cn
);
758 if (skb_queue_len(&c3cn
->write_queue
))
759 c3cn_push_tx_frames(c3cn
, 1);
760 cxgb3i_conn_tx_open(c3cn
);
764 static int do_act_establish(struct t3cdev
*cdev
, struct sk_buff
*skb
,
767 struct cpl_act_establish
*req
= cplhdr(skb
);
768 unsigned int tid
= GET_TID(req
);
769 unsigned int atid
= G_PASS_OPEN_TID(ntohl(req
->tos_tid
));
770 struct s3_conn
*c3cn
= ctx
;
771 struct cxgb3i_sdev_data
*cdata
= CXGB3_SDEV_DATA(cdev
);
773 c3cn_conn_debug("rcv, tid 0x%x, c3cn 0x%p, s %u, f 0x%lx.\n",
774 tid
, c3cn
, c3cn
->state
, c3cn
->flags
);
778 cxgb3_insert_tid(cdata
->cdev
, cdata
->client
, c3cn
, tid
);
779 s3_free_atid(cdev
, atid
);
781 c3cn
->qset
= G_QNUM(ntohl(skb
->csum
));
783 process_cpl_msg(process_act_establish
, c3cn
, skb
);
788 * Process a CPL_ACT_OPEN_RPL message: -> host
789 * Handle active open failures.
791 static int act_open_rpl_status_to_errno(int status
)
794 case CPL_ERR_CONN_RESET
:
795 return -ECONNREFUSED
;
796 case CPL_ERR_ARP_MISS
:
797 return -EHOSTUNREACH
;
798 case CPL_ERR_CONN_TIMEDOUT
:
800 case CPL_ERR_TCAM_FULL
:
802 case CPL_ERR_CONN_EXIST
:
803 cxgb3i_log_error("ACTIVE_OPEN_RPL: 4-tuple in use\n");
810 static void act_open_retry_timer(unsigned long data
)
813 struct s3_conn
*c3cn
= (struct s3_conn
*)data
;
815 c3cn_conn_debug("c3cn 0x%p, state %u.\n", c3cn
, c3cn
->state
);
817 spin_lock_bh(&c3cn
->lock
);
818 skb
= alloc_skb(sizeof(struct cpl_act_open_req
), GFP_ATOMIC
);
820 fail_act_open(c3cn
, -ENOMEM
);
822 skb
->sk
= (struct sock
*)c3cn
;
823 set_arp_failure_handler(skb
, act_open_req_arp_failure
);
824 make_act_open_req(c3cn
, skb
, c3cn
->tid
, c3cn
->l2t
);
825 l2t_send(c3cn
->cdev
, skb
, c3cn
->l2t
);
827 spin_unlock_bh(&c3cn
->lock
);
831 static void process_act_open_rpl(struct s3_conn
*c3cn
, struct sk_buff
*skb
)
833 struct cpl_act_open_rpl
*rpl
= cplhdr(skb
);
835 c3cn_conn_debug("c3cn 0x%p, state %u, flag 0x%lx.\n",
836 c3cn
, c3cn
->state
, c3cn
->flags
);
838 if (rpl
->status
== CPL_ERR_CONN_EXIST
&&
839 c3cn
->retry_timer
.function
!= act_open_retry_timer
) {
840 c3cn
->retry_timer
.function
= act_open_retry_timer
;
841 if (!mod_timer(&c3cn
->retry_timer
, jiffies
+ HZ
/ 2))
844 fail_act_open(c3cn
, act_open_rpl_status_to_errno(rpl
->status
));
848 static int do_act_open_rpl(struct t3cdev
*cdev
, struct sk_buff
*skb
, void *ctx
)
850 struct s3_conn
*c3cn
= ctx
;
851 struct cpl_act_open_rpl
*rpl
= cplhdr(skb
);
853 c3cn_conn_debug("rcv, status 0x%x, c3cn 0x%p, s %u, f 0x%lx.\n",
854 rpl
->status
, c3cn
, c3cn
->state
, c3cn
->flags
);
856 if (rpl
->status
!= CPL_ERR_TCAM_FULL
&&
857 rpl
->status
!= CPL_ERR_CONN_EXIST
&&
858 rpl
->status
!= CPL_ERR_ARP_MISS
)
859 cxgb3_queue_tid_release(cdev
, GET_TID(rpl
));
861 process_cpl_msg_ref(process_act_open_rpl
, c3cn
, skb
);
866 * Process PEER_CLOSE CPL messages: -> host
869 static void process_peer_close(struct s3_conn
*c3cn
, struct sk_buff
*skb
)
871 c3cn_conn_debug("c3cn 0x%p, state %u, flag 0x%lx.\n",
872 c3cn
, c3cn
->state
, c3cn
->flags
);
874 if (c3cn_flag(c3cn
, C3CN_ABORT_RPL_PENDING
))
877 switch (c3cn
->state
) {
878 case C3CN_STATE_ESTABLISHED
:
879 c3cn_set_state(c3cn
, C3CN_STATE_PASSIVE_CLOSE
);
881 case C3CN_STATE_ACTIVE_CLOSE
:
882 c3cn_set_state(c3cn
, C3CN_STATE_CLOSE_WAIT_2
);
884 case C3CN_STATE_CLOSE_WAIT_1
:
887 case C3CN_STATE_ABORTING
:
890 cxgb3i_log_error("%s: peer close, TID %u in bad state %u\n",
891 c3cn
->cdev
->name
, c3cn
->tid
, c3cn
->state
);
894 cxgb3i_conn_closing(c3cn
);
899 static int do_peer_close(struct t3cdev
*cdev
, struct sk_buff
*skb
, void *ctx
)
901 struct s3_conn
*c3cn
= ctx
;
903 c3cn_conn_debug("rcv, c3cn 0x%p, s %u, f 0x%lx.\n",
904 c3cn
, c3cn
->state
, c3cn
->flags
);
905 process_cpl_msg_ref(process_peer_close
, c3cn
, skb
);
910 * Process CLOSE_CONN_RPL CPL message: -> host
911 * Process a peer ACK to our FIN.
913 static void process_close_con_rpl(struct s3_conn
*c3cn
, struct sk_buff
*skb
)
915 struct cpl_close_con_rpl
*rpl
= cplhdr(skb
);
917 c3cn_conn_debug("c3cn 0x%p, state %u, flag 0x%lx.\n",
918 c3cn
, c3cn
->state
, c3cn
->flags
);
920 c3cn
->snd_una
= ntohl(rpl
->snd_nxt
) - 1; /* exclude FIN */
922 if (c3cn_flag(c3cn
, C3CN_ABORT_RPL_PENDING
))
925 switch (c3cn
->state
) {
926 case C3CN_STATE_ACTIVE_CLOSE
:
927 c3cn_set_state(c3cn
, C3CN_STATE_CLOSE_WAIT_1
);
929 case C3CN_STATE_CLOSE_WAIT_1
:
930 case C3CN_STATE_CLOSE_WAIT_2
:
933 case C3CN_STATE_ABORTING
:
936 cxgb3i_log_error("%s: close_rpl, TID %u in bad state %u\n",
937 c3cn
->cdev
->name
, c3cn
->tid
, c3cn
->state
);
944 static int do_close_con_rpl(struct t3cdev
*cdev
, struct sk_buff
*skb
,
947 struct s3_conn
*c3cn
= ctx
;
949 c3cn_conn_debug("rcv, c3cn 0x%p, s %u, f 0x%lx.\n",
950 c3cn
, c3cn
->state
, c3cn
->flags
);
952 process_cpl_msg_ref(process_close_con_rpl
, c3cn
, skb
);
957 * Process ABORT_REQ_RSS CPL message: -> host
958 * Process abort requests. If we are waiting for an ABORT_RPL we ignore this
959 * request except that we need to reply to it.
962 static int abort_status_to_errno(struct s3_conn
*c3cn
, int abort_reason
,
965 switch (abort_reason
) {
966 case CPL_ERR_BAD_SYN
: /* fall through */
967 case CPL_ERR_CONN_RESET
:
968 return c3cn
->state
> C3CN_STATE_ESTABLISHED
?
969 -EPIPE
: -ECONNRESET
;
970 case CPL_ERR_XMIT_TIMEDOUT
:
971 case CPL_ERR_PERSIST_TIMEDOUT
:
972 case CPL_ERR_FINWAIT2_TIMEDOUT
:
973 case CPL_ERR_KEEPALIVE_TIMEDOUT
:
980 static void process_abort_req(struct s3_conn
*c3cn
, struct sk_buff
*skb
)
982 int rst_status
= CPL_ABORT_NO_RST
;
983 const struct cpl_abort_req_rss
*req
= cplhdr(skb
);
985 c3cn_conn_debug("c3cn 0x%p, state %u, flag 0x%lx.\n",
986 c3cn
, c3cn
->state
, c3cn
->flags
);
988 if (!c3cn_flag(c3cn
, C3CN_ABORT_REQ_RCVD
)) {
989 c3cn_set_flag(c3cn
, C3CN_ABORT_REQ_RCVD
);
990 c3cn_set_state(c3cn
, C3CN_STATE_ABORTING
);
995 c3cn_clear_flag(c3cn
, C3CN_ABORT_REQ_RCVD
);
996 send_abort_rpl(c3cn
, rst_status
);
998 if (!c3cn_flag(c3cn
, C3CN_ABORT_RPL_PENDING
)) {
1000 abort_status_to_errno(c3cn
, req
->status
, &rst_status
);
1005 static int do_abort_req(struct t3cdev
*cdev
, struct sk_buff
*skb
, void *ctx
)
1007 const struct cpl_abort_req_rss
*req
= cplhdr(skb
);
1008 struct s3_conn
*c3cn
= ctx
;
1010 c3cn_conn_debug("rcv, c3cn 0x%p, s 0x%x, f 0x%lx.\n",
1011 c3cn
, c3cn
->state
, c3cn
->flags
);
1013 if (req
->status
== CPL_ERR_RTX_NEG_ADVICE
||
1014 req
->status
== CPL_ERR_PERSIST_NEG_ADVICE
) {
1019 process_cpl_msg_ref(process_abort_req
, c3cn
, skb
);
1024 * Process ABORT_RPL_RSS CPL message: -> host
1025 * Process abort replies. We only process these messages if we anticipate
1026 * them as the coordination between SW and HW in this area is somewhat lacking
1027 * and sometimes we get ABORT_RPLs after we are done with the connection that
1028 * originated the ABORT_REQ.
1030 static void process_abort_rpl(struct s3_conn
*c3cn
, struct sk_buff
*skb
)
1032 c3cn_conn_debug("c3cn 0x%p, state %u, flag 0x%lx.\n",
1033 c3cn
, c3cn
->state
, c3cn
->flags
);
1035 if (c3cn_flag(c3cn
, C3CN_ABORT_RPL_PENDING
)) {
1036 if (!c3cn_flag(c3cn
, C3CN_ABORT_RPL_RCVD
))
1037 c3cn_set_flag(c3cn
, C3CN_ABORT_RPL_RCVD
);
1039 c3cn_clear_flag(c3cn
, C3CN_ABORT_RPL_RCVD
);
1040 c3cn_clear_flag(c3cn
, C3CN_ABORT_RPL_PENDING
);
1041 if (c3cn_flag(c3cn
, C3CN_ABORT_REQ_RCVD
))
1042 cxgb3i_log_error("%s tid %u, ABORT_RPL_RSS\n",
1043 c3cn
->cdev
->name
, c3cn
->tid
);
1050 static int do_abort_rpl(struct t3cdev
*cdev
, struct sk_buff
*skb
, void *ctx
)
1052 struct cpl_abort_rpl_rss
*rpl
= cplhdr(skb
);
1053 struct s3_conn
*c3cn
= ctx
;
1055 c3cn_conn_debug("rcv, status 0x%x, c3cn 0x%p, s %u, 0x%lx.\n",
1056 rpl
->status
, c3cn
, c3cn
? c3cn
->state
: 0,
1057 c3cn
? c3cn
->flags
: 0UL);
1060 * Ignore replies to post-close aborts indicating that the abort was
1061 * requested too late. These connections are terminated when we get
1062 * PEER_CLOSE or CLOSE_CON_RPL and by the time the abort_rpl_rss
1063 * arrives the TID is either no longer used or it has been recycled.
1065 if (rpl
->status
== CPL_ERR_ABORT_FAILED
)
1069 * Sometimes we've already closed the connection, e.g., a post-close
1070 * abort races with ABORT_REQ_RSS, the latter frees the connection
1071 * expecting the ABORT_REQ will fail with CPL_ERR_ABORT_FAILED,
1072 * but FW turns the ABORT_REQ into a regular one and so we get
1073 * ABORT_RPL_RSS with status 0 and no connection.
1078 process_cpl_msg_ref(process_abort_rpl
, c3cn
, skb
);
1087 * Process RX_ISCSI_HDR CPL message: -> host
1088 * Handle received PDUs, the payload could be DDP'ed. If not, the payload
1089 * follow after the bhs.
1091 static void process_rx_iscsi_hdr(struct s3_conn
*c3cn
, struct sk_buff
*skb
)
1093 struct cpl_iscsi_hdr
*hdr_cpl
= cplhdr(skb
);
1094 struct cpl_iscsi_hdr_norss data_cpl
;
1095 struct cpl_rx_data_ddp_norss ddp_cpl
;
1096 unsigned int hdr_len
, data_len
, status
;
1100 if (unlikely(c3cn
->state
>= C3CN_STATE_PASSIVE_CLOSE
)) {
1101 if (c3cn
->state
!= C3CN_STATE_ABORTING
)
1102 send_abort_req(c3cn
);
1107 skb_tcp_seq(skb
) = ntohl(hdr_cpl
->seq
);
1110 skb_reset_transport_header(skb
);
1111 __skb_pull(skb
, sizeof(struct cpl_iscsi_hdr
));
1113 len
= hdr_len
= ntohs(hdr_cpl
->len
);
1114 /* msg coalesce is off or not enough data received */
1115 if (skb
->len
<= hdr_len
) {
1116 cxgb3i_log_error("%s: TID %u, ISCSI_HDR, skb len %u < %u.\n",
1117 c3cn
->cdev
->name
, c3cn
->tid
,
1122 err
= skb_copy_bits(skb
, skb
->len
- sizeof(ddp_cpl
), &ddp_cpl
,
1127 skb_ulp_mode(skb
) = ULP2_FLAG_DATA_READY
;
1128 skb_rx_pdulen(skb
) = ntohs(ddp_cpl
.len
);
1129 skb_rx_ddigest(skb
) = ntohl(ddp_cpl
.ulp_crc
);
1130 status
= ntohl(ddp_cpl
.ddp_status
);
1132 c3cn_rx_debug("rx skb 0x%p, len %u, pdulen %u, ddp status 0x%x.\n",
1133 skb
, skb
->len
, skb_rx_pdulen(skb
), status
);
1135 if (status
& (1 << RX_DDP_STATUS_HCRC_SHIFT
))
1136 skb_ulp_mode(skb
) |= ULP2_FLAG_HCRC_ERROR
;
1137 if (status
& (1 << RX_DDP_STATUS_DCRC_SHIFT
))
1138 skb_ulp_mode(skb
) |= ULP2_FLAG_DCRC_ERROR
;
1139 if (status
& (1 << RX_DDP_STATUS_PAD_SHIFT
))
1140 skb_ulp_mode(skb
) |= ULP2_FLAG_PAD_ERROR
;
1142 if (skb
->len
> (hdr_len
+ sizeof(ddp_cpl
))) {
1143 err
= skb_copy_bits(skb
, hdr_len
, &data_cpl
, sizeof(data_cpl
));
1146 data_len
= ntohs(data_cpl
.len
);
1147 len
+= sizeof(data_cpl
) + data_len
;
1148 } else if (status
& (1 << RX_DDP_STATUS_DDP_SHIFT
))
1149 skb_ulp_mode(skb
) |= ULP2_FLAG_DATA_DDPED
;
1151 c3cn
->rcv_nxt
= ntohl(ddp_cpl
.seq
) + skb_rx_pdulen(skb
);
1152 __pskb_trim(skb
, len
);
1153 __skb_queue_tail(&c3cn
->receive_queue
, skb
);
1154 cxgb3i_conn_pdu_ready(c3cn
);
1159 send_abort_req(c3cn
);
1163 static int do_iscsi_hdr(struct t3cdev
*t3dev
, struct sk_buff
*skb
, void *ctx
)
1165 struct s3_conn
*c3cn
= ctx
;
1167 process_cpl_msg(process_rx_iscsi_hdr
, c3cn
, skb
);
1172 * Process TX_DATA_ACK CPL messages: -> host
1173 * Process an acknowledgment of WR completion. Advance snd_una and send the
1174 * next batch of work requests from the write queue.
1176 static void check_wr_invariants(struct s3_conn
*c3cn
)
1178 int pending
= count_pending_wrs(c3cn
);
1180 if (unlikely(c3cn
->wr_avail
+ pending
!= c3cn
->wr_max
))
1181 cxgb3i_log_error("TID %u: credit imbalance: avail %u, "
1182 "pending %u, total should be %u\n",
1183 c3cn
->tid
, c3cn
->wr_avail
, pending
,
1187 static void process_wr_ack(struct s3_conn
*c3cn
, struct sk_buff
*skb
)
1189 struct cpl_wr_ack
*hdr
= cplhdr(skb
);
1190 unsigned int credits
= ntohs(hdr
->credits
);
1191 u32 snd_una
= ntohl(hdr
->snd_una
);
1193 c3cn_tx_debug("%u WR credits, avail %u, unack %u, TID %u, state %u.\n",
1194 credits
, c3cn
->wr_avail
, c3cn
->wr_unacked
,
1195 c3cn
->tid
, c3cn
->state
);
1197 c3cn
->wr_avail
+= credits
;
1198 if (c3cn
->wr_unacked
> c3cn
->wr_max
- c3cn
->wr_avail
)
1199 c3cn
->wr_unacked
= c3cn
->wr_max
- c3cn
->wr_avail
;
1202 struct sk_buff
*p
= peek_wr(c3cn
);
1205 cxgb3i_log_error("%u WR_ACK credits for TID %u with "
1206 "nothing pending, state %u\n",
1207 credits
, c3cn
->tid
, c3cn
->state
);
1210 if (unlikely(credits
< p
->csum
)) {
1211 struct tx_data_wr
*w
= cplhdr(p
);
1212 cxgb3i_log_error("TID %u got %u WR credits need %u, "
1213 "len %u, main body %u, frags %u, "
1214 "seq # %u, ACK una %u, ACK nxt %u, "
1215 "WR_AVAIL %u, WRs pending %u\n",
1216 c3cn
->tid
, credits
, p
->csum
, p
->len
,
1217 p
->len
- p
->data_len
,
1218 skb_shinfo(p
)->nr_frags
,
1219 ntohl(w
->sndseq
), snd_una
,
1220 ntohl(hdr
->snd_nxt
), c3cn
->wr_avail
,
1221 count_pending_wrs(c3cn
) - credits
);
1231 check_wr_invariants(c3cn
);
1233 if (unlikely(before(snd_una
, c3cn
->snd_una
))) {
1234 cxgb3i_log_error("TID %u, unexpected sequence # %u in WR_ACK "
1236 c3cn
->tid
, snd_una
, c3cn
->snd_una
);
1240 if (c3cn
->snd_una
!= snd_una
) {
1241 c3cn
->snd_una
= snd_una
;
1242 dst_confirm(c3cn
->dst_cache
);
1245 if (skb_queue_len(&c3cn
->write_queue
)) {
1246 if (c3cn_push_tx_frames(c3cn
, 0))
1247 cxgb3i_conn_tx_open(c3cn
);
1249 cxgb3i_conn_tx_open(c3cn
);
1254 static int do_wr_ack(struct t3cdev
*cdev
, struct sk_buff
*skb
, void *ctx
)
1256 struct s3_conn
*c3cn
= ctx
;
1258 process_cpl_msg(process_wr_ack
, c3cn
, skb
);
1263 * for each connection, pre-allocate skbs needed for close/abort requests. So
1264 * that we can service the request right away.
1266 static void c3cn_free_cpl_skbs(struct s3_conn
*c3cn
)
1268 if (c3cn
->cpl_close
)
1269 kfree_skb(c3cn
->cpl_close
);
1270 if (c3cn
->cpl_abort_req
)
1271 kfree_skb(c3cn
->cpl_abort_req
);
1272 if (c3cn
->cpl_abort_rpl
)
1273 kfree_skb(c3cn
->cpl_abort_rpl
);
1276 static int c3cn_alloc_cpl_skbs(struct s3_conn
*c3cn
)
1278 c3cn
->cpl_close
= alloc_skb(sizeof(struct cpl_close_con_req
),
1280 if (!c3cn
->cpl_close
)
1282 skb_put(c3cn
->cpl_close
, sizeof(struct cpl_close_con_req
));
1284 c3cn
->cpl_abort_req
= alloc_skb(sizeof(struct cpl_abort_req
),
1286 if (!c3cn
->cpl_abort_req
)
1288 skb_put(c3cn
->cpl_abort_req
, sizeof(struct cpl_abort_req
));
1290 c3cn
->cpl_abort_rpl
= alloc_skb(sizeof(struct cpl_abort_rpl
),
1292 if (!c3cn
->cpl_abort_rpl
)
1294 skb_put(c3cn
->cpl_abort_rpl
, sizeof(struct cpl_abort_rpl
));
1299 c3cn_free_cpl_skbs(c3cn
);
1304 * c3cn_release_offload_resources - release offload resource
1305 * @c3cn: the offloaded iscsi tcp connection.
1306 * Release resources held by an offload connection (TID, L2T entry, etc.)
1308 static void c3cn_release_offload_resources(struct s3_conn
*c3cn
)
1310 struct t3cdev
*cdev
= c3cn
->cdev
;
1311 unsigned int tid
= c3cn
->tid
;
1314 c3cn_free_cpl_skbs(c3cn
);
1316 if (c3cn
->wr_avail
!= c3cn
->wr_max
) {
1317 purge_wr_queue(c3cn
);
1318 reset_wr_list(c3cn
);
1323 l2t_release(L2DATA(cdev
), c3cn
->l2t
);
1326 if (c3cn
->state
== C3CN_STATE_CONNECTING
)
1328 s3_free_atid(cdev
, tid
);
1331 cxgb3_remove_tid(cdev
, (void *)c3cn
, tid
);
1336 c3cn
->dst_cache
= NULL
;
1341 * cxgb3i_c3cn_create - allocate and initialize an s3_conn structure
1342 * returns the s3_conn structure allocated.
1344 struct s3_conn
*cxgb3i_c3cn_create(void)
1346 struct s3_conn
*c3cn
;
1348 c3cn
= kzalloc(sizeof(*c3cn
), GFP_KERNEL
);
1352 /* pre-allocate close/abort cpl, so we don't need to wait for memory
1353 when close/abort is requested. */
1354 if (c3cn_alloc_cpl_skbs(c3cn
) < 0)
1357 c3cn_conn_debug("alloc c3cn 0x%p.\n", c3cn
);
1360 spin_lock_init(&c3cn
->lock
);
1361 atomic_set(&c3cn
->refcnt
, 1);
1362 skb_queue_head_init(&c3cn
->receive_queue
);
1363 skb_queue_head_init(&c3cn
->write_queue
);
1364 setup_timer(&c3cn
->retry_timer
, NULL
, (unsigned long)c3cn
);
1365 rwlock_init(&c3cn
->callback_lock
);
1374 static void c3cn_active_close(struct s3_conn
*c3cn
)
1379 c3cn_conn_debug("c3cn 0x%p, state %u, flag 0x%lx.\n",
1380 c3cn
, c3cn
->state
, c3cn
->flags
);
1382 dst_confirm(c3cn
->dst_cache
);
1385 spin_lock_bh(&c3cn
->lock
);
1387 data_lost
= skb_queue_len(&c3cn
->receive_queue
);
1388 __skb_queue_purge(&c3cn
->receive_queue
);
1390 switch (c3cn
->state
) {
1391 case C3CN_STATE_CLOSED
:
1392 case C3CN_STATE_ACTIVE_CLOSE
:
1393 case C3CN_STATE_CLOSE_WAIT_1
:
1394 case C3CN_STATE_CLOSE_WAIT_2
:
1395 case C3CN_STATE_ABORTING
:
1396 /* nothing need to be done */
1398 case C3CN_STATE_CONNECTING
:
1399 /* defer until cpl_act_open_rpl or cpl_act_establish */
1400 c3cn_set_flag(c3cn
, C3CN_ACTIVE_CLOSE_NEEDED
);
1402 case C3CN_STATE_ESTABLISHED
:
1404 c3cn_set_state(c3cn
, C3CN_STATE_ACTIVE_CLOSE
);
1406 case C3CN_STATE_PASSIVE_CLOSE
:
1408 c3cn_set_state(c3cn
, C3CN_STATE_CLOSE_WAIT_2
);
1414 /* Unread data was tossed, zap the connection. */
1415 send_abort_req(c3cn
);
1417 send_close_req(c3cn
);
1420 spin_unlock_bh(&c3cn
->lock
);
1425 * cxgb3i_c3cn_release - close and release an iscsi tcp connection and any
1427 * @c3cn: the iscsi tcp connection
1429 void cxgb3i_c3cn_release(struct s3_conn
*c3cn
)
1431 c3cn_conn_debug("c3cn 0x%p, s %u, f 0x%lx.\n",
1432 c3cn
, c3cn
->state
, c3cn
->flags
);
1433 if (unlikely(c3cn
->state
== C3CN_STATE_CONNECTING
))
1434 c3cn_set_flag(c3cn
, C3CN_ACTIVE_CLOSE_NEEDED
);
1435 else if (likely(c3cn
->state
!= C3CN_STATE_CLOSED
))
1436 c3cn_active_close(c3cn
);
1440 static int is_cxgb3_dev(struct net_device
*dev
)
1442 struct cxgb3i_sdev_data
*cdata
;
1443 struct net_device
*ndev
= dev
;
1445 if (dev
->priv_flags
& IFF_802_1Q_VLAN
)
1446 ndev
= vlan_dev_real_dev(dev
);
1448 write_lock(&cdata_rwlock
);
1449 list_for_each_entry(cdata
, &cdata_list
, list
) {
1450 struct adap_ports
*ports
= &cdata
->ports
;
1453 for (i
= 0; i
< ports
->nports
; i
++)
1454 if (ndev
== ports
->lldevs
[i
]) {
1455 write_unlock(&cdata_rwlock
);
1459 write_unlock(&cdata_rwlock
);
1464 * cxgb3_egress_dev - return the cxgb3 egress device
1465 * @root_dev: the root device anchoring the search
1466 * @c3cn: the connection used to determine egress port in bonding mode
1467 * @context: in bonding mode, indicates a connection set up or failover
1469 * Return egress device or NULL if the egress device isn't one of our ports.
1471 static struct net_device
*cxgb3_egress_dev(struct net_device
*root_dev
,
1472 struct s3_conn
*c3cn
,
1476 if (root_dev
->priv_flags
& IFF_802_1Q_VLAN
)
1477 root_dev
= vlan_dev_real_dev(root_dev
);
1478 else if (is_cxgb3_dev(root_dev
))
1486 static struct rtable
*find_route(struct net_device
*dev
,
1487 __be32 saddr
, __be32 daddr
,
1488 __be16 sport
, __be16 dport
)
1492 .oif
= dev
? dev
->ifindex
: 0,
1498 .proto
= IPPROTO_TCP
,
1502 .dport
= dport
} } };
1504 if (ip_route_output_flow(&init_net
, &rt
, &fl
, NULL
, 0))
1510 * Assign offload parameters to some connection fields.
1512 static void init_offload_conn(struct s3_conn
*c3cn
,
1513 struct t3cdev
*cdev
,
1514 struct dst_entry
*dst
)
1516 BUG_ON(c3cn
->cdev
!= cdev
);
1517 c3cn
->wr_max
= c3cn
->wr_avail
= T3C_DATA(cdev
)->max_wrs
- 1;
1518 c3cn
->wr_unacked
= 0;
1519 c3cn
->mss_idx
= select_mss(c3cn
, dst_mtu(dst
));
1521 reset_wr_list(c3cn
);
1524 static int initiate_act_open(struct s3_conn
*c3cn
, struct net_device
*dev
)
1526 struct cxgb3i_sdev_data
*cdata
= NDEV2CDATA(dev
);
1527 struct t3cdev
*cdev
= cdata
->cdev
;
1528 struct dst_entry
*dst
= c3cn
->dst_cache
;
1529 struct sk_buff
*skb
;
1531 c3cn_conn_debug("c3cn 0x%p, state %u, flag 0x%lx.\n",
1532 c3cn
, c3cn
->state
, c3cn
->flags
);
1534 * Initialize connection data. Note that the flags and ULP mode are
1535 * initialized higher up ...
1539 c3cn
->tid
= cxgb3_alloc_atid(cdev
, cdata
->client
, c3cn
);
1544 c3cn
->l2t
= t3_l2t_get(cdev
, dst
->neighbour
, dev
);
1548 skb
= alloc_skb(sizeof(struct cpl_act_open_req
), GFP_KERNEL
);
1552 skb
->sk
= (struct sock
*)c3cn
;
1553 set_arp_failure_handler(skb
, act_open_req_arp_failure
);
1557 init_offload_conn(c3cn
, cdev
, dst
);
1560 make_act_open_req(c3cn
, skb
, c3cn
->tid
, c3cn
->l2t
);
1561 l2t_send(cdev
, skb
, c3cn
->l2t
);
1565 l2t_release(L2DATA(cdev
), c3cn
->l2t
);
1567 s3_free_atid(cdev
, c3cn
->tid
);
1574 * cxgb3i_find_dev - find the interface associated with the given address
1575 * @ipaddr: ip address
1577 static struct net_device
*
1578 cxgb3i_find_dev(struct net_device
*dev
, __be32 ipaddr
)
1584 memset(&fl
, 0, sizeof(fl
));
1585 fl
.nl_u
.ip4_u
.daddr
= ipaddr
;
1587 err
= ip_route_output_key(dev
? dev_net(dev
) : &init_net
, &rt
, &fl
);
1589 return (&rt
->u
.dst
)->dev
;
1595 * cxgb3i_c3cn_connect - initiates an iscsi tcp connection to a given address
1596 * @c3cn: the iscsi tcp connection
1597 * @usin: destination address
1599 * return 0 if active open request is sent, < 0 otherwise.
1601 int cxgb3i_c3cn_connect(struct net_device
*dev
, struct s3_conn
*c3cn
,
1602 struct sockaddr_in
*usin
)
1605 struct cxgb3i_sdev_data
*cdata
;
1606 struct t3cdev
*cdev
;
1608 struct net_device
*dstdev
;
1611 c3cn_conn_debug("c3cn 0x%p, dev 0x%p.\n", c3cn
, dev
);
1613 if (usin
->sin_family
!= AF_INET
)
1614 return -EAFNOSUPPORT
;
1616 c3cn
->daddr
.sin_port
= usin
->sin_port
;
1617 c3cn
->daddr
.sin_addr
.s_addr
= usin
->sin_addr
.s_addr
;
1619 dstdev
= cxgb3i_find_dev(dev
, usin
->sin_addr
.s_addr
);
1620 if (!dstdev
|| !is_cxgb3_dev(dstdev
))
1621 return -ENETUNREACH
;
1623 if (dstdev
->priv_flags
& IFF_802_1Q_VLAN
)
1626 rt
= find_route(dev
, c3cn
->saddr
.sin_addr
.s_addr
,
1627 c3cn
->daddr
.sin_addr
.s_addr
,
1628 c3cn
->saddr
.sin_port
,
1629 c3cn
->daddr
.sin_port
);
1631 c3cn_conn_debug("NO route to 0x%x, port %u, dev %s.\n",
1632 c3cn
->daddr
.sin_addr
.s_addr
,
1633 ntohs(c3cn
->daddr
.sin_port
),
1634 dev
? dev
->name
: "any");
1635 return -ENETUNREACH
;
1638 if (rt
->rt_flags
& (RTCF_MULTICAST
| RTCF_BROADCAST
)) {
1639 c3cn_conn_debug("multi-cast route to 0x%x, port %u, dev %s.\n",
1640 c3cn
->daddr
.sin_addr
.s_addr
,
1641 ntohs(c3cn
->daddr
.sin_port
),
1642 dev
? dev
->name
: "any");
1644 return -ENETUNREACH
;
1647 if (!c3cn
->saddr
.sin_addr
.s_addr
)
1648 c3cn
->saddr
.sin_addr
.s_addr
= rt
->rt_src
;
1650 /* now commit destination to connection */
1651 c3cn
->dst_cache
= &rt
->u
.dst
;
1653 /* try to establish an offloaded connection */
1654 dev
= cxgb3_egress_dev(c3cn
->dst_cache
->dev
, c3cn
, 0);
1656 c3cn_conn_debug("c3cn 0x%p, egress dev NULL.\n", c3cn
);
1657 return -ENETUNREACH
;
1659 cdata
= NDEV2CDATA(dev
);
1662 /* get a source port if one hasn't been provided */
1663 err
= c3cn_get_port(c3cn
, cdata
);
1667 c3cn_conn_debug("c3cn 0x%p get port %u.\n",
1668 c3cn
, ntohs(c3cn
->saddr
.sin_port
));
1670 sipv4
= cxgb3i_get_private_ipv4addr(dev
);
1672 c3cn_conn_debug("c3cn 0x%p, iscsi ip not configured.\n", c3cn
);
1673 sipv4
= c3cn
->saddr
.sin_addr
.s_addr
;
1674 cxgb3i_set_private_ipv4addr(dev
, sipv4
);
1676 c3cn
->saddr
.sin_addr
.s_addr
= sipv4
;
1678 c3cn_conn_debug("c3cn 0x%p, %pI4,%u-%pI4,%u SYN_SENT.\n",
1680 &c3cn
->saddr
.sin_addr
.s_addr
,
1681 ntohs(c3cn
->saddr
.sin_port
),
1682 &c3cn
->daddr
.sin_addr
.s_addr
,
1683 ntohs(c3cn
->daddr
.sin_port
));
1685 c3cn_set_state(c3cn
, C3CN_STATE_CONNECTING
);
1686 if (!initiate_act_open(c3cn
, dev
))
1690 * If we get here, we don't have an offload connection so simply
1696 * This trashes the connection and releases the local port,
1699 c3cn_conn_debug("c3cn 0x%p -> CLOSED.\n", c3cn
);
1700 c3cn_set_state(c3cn
, C3CN_STATE_CLOSED
);
1702 c3cn_put_port(c3cn
);
1707 * cxgb3i_c3cn_rx_credits - ack received tcp data.
1708 * @c3cn: iscsi tcp connection
1709 * @copied: # of bytes processed
1711 * Called after some received data has been read. It returns RX credits
1712 * to the HW for the amount of data processed.
1714 void cxgb3i_c3cn_rx_credits(struct s3_conn
*c3cn
, int copied
)
1716 struct t3cdev
*cdev
;
1718 u32 credits
, dack
= 0;
1720 if (c3cn
->state
!= C3CN_STATE_ESTABLISHED
)
1723 credits
= c3cn
->copied_seq
- c3cn
->rcv_wup
;
1724 if (unlikely(!credits
))
1729 if (unlikely(cxgb3_rx_credit_thres
== 0))
1732 dack
= F_RX_DACK_CHANGE
| V_RX_DACK_MODE(1);
1735 * For coalescing to work effectively ensure the receive window has
1736 * at least 16KB left.
1738 must_send
= credits
+ 16384 >= cxgb3_rcv_win
;
1740 if (must_send
|| credits
>= cxgb3_rx_credit_thres
)
1741 c3cn
->rcv_wup
+= send_rx_credits(c3cn
, credits
, dack
);
1745 * cxgb3i_c3cn_send_pdus - send the skbs containing iscsi pdus
1746 * @c3cn: iscsi tcp connection
1747 * @skb: skb contains the iscsi pdu
1749 * Add a list of skbs to a connection send queue. The skbs must comply with
1750 * the max size limit of the device and have a headroom of at least
1751 * TX_HEADER_LEN bytes.
1752 * Return # of bytes queued.
1754 int cxgb3i_c3cn_send_pdus(struct s3_conn
*c3cn
, struct sk_buff
*skb
)
1756 struct sk_buff
*next
;
1757 int err
, copied
= 0;
1759 spin_lock_bh(&c3cn
->lock
);
1761 if (c3cn
->state
!= C3CN_STATE_ESTABLISHED
) {
1762 c3cn_tx_debug("c3cn 0x%p, not in est. state %u.\n",
1769 c3cn_tx_debug("c3cn 0x%p, err %d.\n", c3cn
, c3cn
->err
);
1774 if (c3cn
->write_seq
- c3cn
->snd_una
>= cxgb3_snd_win
) {
1775 c3cn_tx_debug("c3cn 0x%p, snd %u - %u > %u.\n",
1776 c3cn
, c3cn
->write_seq
, c3cn
->snd_una
,
1783 int frags
= skb_shinfo(skb
)->nr_frags
+
1784 (skb
->len
!= skb
->data_len
);
1786 if (unlikely(skb_headroom(skb
) < TX_HEADER_LEN
)) {
1787 c3cn_tx_debug("c3cn 0x%p, skb head.\n", c3cn
);
1792 if (frags
>= SKB_WR_LIST_SIZE
) {
1793 cxgb3i_log_error("c3cn 0x%p, tx frags %d, len %u,%u.\n",
1794 c3cn
, skb_shinfo(skb
)->nr_frags
,
1795 skb
->len
, skb
->data_len
);
1802 skb_entail(c3cn
, skb
, C3CB_FLAG_NO_APPEND
| C3CB_FLAG_NEED_HDR
);
1804 c3cn
->write_seq
+= skb
->len
+ ulp_extra_len(skb
);
1808 if (likely(skb_queue_len(&c3cn
->write_queue
)))
1809 c3cn_push_tx_frames(c3cn
, 1);
1810 spin_unlock_bh(&c3cn
->lock
);
1814 if (copied
== 0 && err
== -EPIPE
)
1815 copied
= c3cn
->err
? c3cn
->err
: -EPIPE
;
1821 static void sdev_data_cleanup(struct cxgb3i_sdev_data
*cdata
)
1823 struct adap_ports
*ports
= &cdata
->ports
;
1824 struct s3_conn
*c3cn
;
1827 for (i
= 0; i
< cxgb3_max_connect
; i
++) {
1828 if (cdata
->sport_conn
[i
]) {
1829 c3cn
= cdata
->sport_conn
[i
];
1830 cdata
->sport_conn
[i
] = NULL
;
1832 spin_lock_bh(&c3cn
->lock
);
1834 c3cn_set_flag(c3cn
, C3CN_OFFLOAD_DOWN
);
1836 spin_unlock_bh(&c3cn
->lock
);
1840 for (i
= 0; i
< ports
->nports
; i
++)
1841 NDEV2CDATA(ports
->lldevs
[i
]) = NULL
;
1843 cxgb3i_free_big_mem(cdata
);
1846 void cxgb3i_sdev_cleanup(void)
1848 struct cxgb3i_sdev_data
*cdata
;
1850 write_lock(&cdata_rwlock
);
1851 list_for_each_entry(cdata
, &cdata_list
, list
) {
1852 list_del(&cdata
->list
);
1853 sdev_data_cleanup(cdata
);
1855 write_unlock(&cdata_rwlock
);
1858 int cxgb3i_sdev_init(cxgb3_cpl_handler_func
*cpl_handlers
)
1860 cpl_handlers
[CPL_ACT_ESTABLISH
] = do_act_establish
;
1861 cpl_handlers
[CPL_ACT_OPEN_RPL
] = do_act_open_rpl
;
1862 cpl_handlers
[CPL_PEER_CLOSE
] = do_peer_close
;
1863 cpl_handlers
[CPL_ABORT_REQ_RSS
] = do_abort_req
;
1864 cpl_handlers
[CPL_ABORT_RPL_RSS
] = do_abort_rpl
;
1865 cpl_handlers
[CPL_CLOSE_CON_RPL
] = do_close_con_rpl
;
1866 cpl_handlers
[CPL_TX_DMA_ACK
] = do_wr_ack
;
1867 cpl_handlers
[CPL_ISCSI_HDR
] = do_iscsi_hdr
;
1869 if (cxgb3_max_connect
> CXGB3I_MAX_CONN
)
1870 cxgb3_max_connect
= CXGB3I_MAX_CONN
;
1875 * cxgb3i_sdev_add - allocate and initialize resources for each adapter found
1876 * @cdev: t3cdev adapter
1877 * @client: cxgb3 driver client
1879 void cxgb3i_sdev_add(struct t3cdev
*cdev
, struct cxgb3_client
*client
)
1881 struct cxgb3i_sdev_data
*cdata
;
1882 struct ofld_page_info rx_page_info
;
1883 unsigned int wr_len
;
1884 int mapsize
= cxgb3_max_connect
* sizeof(struct s3_conn
*);
1887 cdata
= cxgb3i_alloc_big_mem(sizeof(*cdata
) + mapsize
, GFP_KERNEL
);
1889 cxgb3i_log_warn("t3dev 0x%p, offload up, OOM %d.\n",
1894 if (cdev
->ctl(cdev
, GET_WR_LEN
, &wr_len
) < 0 ||
1895 cdev
->ctl(cdev
, GET_PORTS
, &cdata
->ports
) < 0 ||
1896 cdev
->ctl(cdev
, GET_RX_PAGE_INFO
, &rx_page_info
) < 0) {
1897 cxgb3i_log_warn("t3dev 0x%p, offload up, ioctl failed.\n",
1902 s3_init_wr_tab(wr_len
);
1904 spin_lock_init(&cdata
->lock
);
1905 INIT_LIST_HEAD(&cdata
->list
);
1907 cdata
->client
= client
;
1909 for (i
= 0; i
< cdata
->ports
.nports
; i
++)
1910 NDEV2CDATA(cdata
->ports
.lldevs
[i
]) = cdata
;
1912 write_lock(&cdata_rwlock
);
1913 list_add_tail(&cdata
->list
, &cdata_list
);
1914 write_unlock(&cdata_rwlock
);
1916 cxgb3i_log_info("t3dev 0x%p, offload up, added.\n", cdev
);
1920 cxgb3i_free_big_mem(cdata
);
1924 * cxgb3i_sdev_remove - free the allocated resources for the adapter
1925 * @cdev: t3cdev adapter
1927 void cxgb3i_sdev_remove(struct t3cdev
*cdev
)
1929 struct cxgb3i_sdev_data
*cdata
= CXGB3_SDEV_DATA(cdev
);
1931 cxgb3i_log_info("t3dev 0x%p, offload down, remove.\n", cdev
);
1933 write_lock(&cdata_rwlock
);
1934 list_del(&cdata
->list
);
1935 write_unlock(&cdata_rwlock
);
1937 sdev_data_cleanup(cdata
);