2 * Copyright (c) 2012 Intel Corporation. All rights reserved.
3 * Copyright (c) 2007 - 2012 QLogic Corporation. All rights reserved.
5 * This software is available to you under a choice of one of two
6 * licenses. You may choose to be licensed under the terms of the GNU
7 * General Public License (GPL) Version 2, available from the file
8 * COPYING in the main directory of this source tree, or the
9 * OpenIB.org BSD license below:
11 * Redistribution and use in source and binary forms, with or
12 * without modification, are permitted provided that the following
15 * - Redistributions of source code must retain the above
16 * copyright notice, this list of conditions and the following
19 * - Redistributions in binary form must reproduce the above
20 * copyright notice, this list of conditions and the following
21 * disclaimer in the documentation and/or other materials
22 * provided with the distribution.
24 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
25 * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
26 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
27 * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
28 * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
29 * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
30 * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
34 #include <linux/spinlock.h>
35 #include <linux/netdevice.h>
36 #include <linux/moduleparam.h>
39 #include "qib_common.h"
41 /* default pio off, sdma on */
42 static ushort sdma_descq_cnt
= 256;
43 module_param_named(sdma_descq_cnt
, sdma_descq_cnt
, ushort
, S_IRUGO
);
44 MODULE_PARM_DESC(sdma_descq_cnt
, "Number of SDMA descq entries");
47 * Bits defined in the send DMA descriptor.
49 #define SDMA_DESC_LAST (1ULL << 11)
50 #define SDMA_DESC_FIRST (1ULL << 12)
51 #define SDMA_DESC_DMA_HEAD (1ULL << 13)
52 #define SDMA_DESC_USE_LARGE_BUF (1ULL << 14)
53 #define SDMA_DESC_INTR (1ULL << 15)
54 #define SDMA_DESC_COUNT_LSB 16
55 #define SDMA_DESC_GEN_LSB 30
57 /* declare all statics here rather than keep sorting */
58 static int alloc_sdma(struct qib_pportdata
*);
59 static void sdma_complete(struct kref
*);
60 static void sdma_finalput(struct qib_sdma_state
*);
61 static void sdma_get(struct qib_sdma_state
*);
62 static void sdma_put(struct qib_sdma_state
*);
63 static void sdma_set_state(struct qib_pportdata
*, enum qib_sdma_states
);
64 static void sdma_start_sw_clean_up(struct qib_pportdata
*);
65 static void sdma_sw_clean_up_task(unsigned long);
66 static void unmap_desc(struct qib_pportdata
*, unsigned);
68 static void sdma_get(struct qib_sdma_state
*ss
)
73 static void sdma_complete(struct kref
*kref
)
75 struct qib_sdma_state
*ss
=
76 container_of(kref
, struct qib_sdma_state
, kref
);
81 static void sdma_put(struct qib_sdma_state
*ss
)
83 kref_put(&ss
->kref
, sdma_complete
);
86 static void sdma_finalput(struct qib_sdma_state
*ss
)
89 wait_for_completion(&ss
->comp
);
93 * Complete all the sdma requests on the active list, in the correct
94 * order, and with appropriate processing. Called when cleaning up
95 * after sdma shutdown, and when new sdma requests are submitted for
96 * a link that is down. This matches what is done for requests
97 * that complete normally, it's just the full list.
99 * Must be called with sdma_lock held
101 static void clear_sdma_activelist(struct qib_pportdata
*ppd
)
103 struct qib_sdma_txreq
*txp
, *txp_next
;
105 list_for_each_entry_safe(txp
, txp_next
, &ppd
->sdma_activelist
, list
) {
106 list_del_init(&txp
->list
);
107 if (txp
->flags
& QIB_SDMA_TXREQ_F_FREEDESC
) {
110 idx
= txp
->start_idx
;
111 while (idx
!= txp
->next_descq_idx
) {
112 unmap_desc(ppd
, idx
);
113 if (++idx
== ppd
->sdma_descq_cnt
)
118 (*txp
->callback
)(txp
, QIB_SDMA_TXREQ_S_ABORTED
);
122 static void sdma_sw_clean_up_task(unsigned long opaque
)
124 struct qib_pportdata
*ppd
= (struct qib_pportdata
*) opaque
;
127 spin_lock_irqsave(&ppd
->sdma_lock
, flags
);
130 * At this point, the following should always be true:
131 * - We are halted, so no more descriptors are getting retired.
132 * - We are not running, so no one is submitting new work.
133 * - Only we can send the e40_sw_cleaned, so we can't start
134 * running again until we say so. So, the active list and
135 * descq are ours to play with.
138 /* Process all retired requests. */
139 qib_sdma_make_progress(ppd
);
141 clear_sdma_activelist(ppd
);
144 * Resync count of added and removed. It is VERY important that
145 * sdma_descq_removed NEVER decrement - user_sdma depends on it.
147 ppd
->sdma_descq_removed
= ppd
->sdma_descq_added
;
150 * Reset our notion of head and tail.
151 * Note that the HW registers will be reset when switching states
152 * due to calling __qib_sdma_process_event() below.
154 ppd
->sdma_descq_tail
= 0;
155 ppd
->sdma_descq_head
= 0;
156 ppd
->sdma_head_dma
[0] = 0;
157 ppd
->sdma_generation
= 0;
159 __qib_sdma_process_event(ppd
, qib_sdma_event_e40_sw_cleaned
);
161 spin_unlock_irqrestore(&ppd
->sdma_lock
, flags
);
165 * This is called when changing to state qib_sdma_state_s10_hw_start_up_wait
166 * as a result of send buffer errors or send DMA descriptor errors.
167 * We want to disarm the buffers in these cases.
169 static void sdma_hw_start_up(struct qib_pportdata
*ppd
)
171 struct qib_sdma_state
*ss
= &ppd
->sdma_state
;
174 for (bufno
= ss
->first_sendbuf
; bufno
< ss
->last_sendbuf
; ++bufno
)
175 ppd
->dd
->f_sendctrl(ppd
, QIB_SENDCTRL_DISARM_BUF(bufno
));
177 ppd
->dd
->f_sdma_hw_start_up(ppd
);
180 static void sdma_sw_tear_down(struct qib_pportdata
*ppd
)
182 struct qib_sdma_state
*ss
= &ppd
->sdma_state
;
184 /* Releasing this reference means the state machine has stopped. */
188 static void sdma_start_sw_clean_up(struct qib_pportdata
*ppd
)
190 tasklet_hi_schedule(&ppd
->sdma_sw_clean_up_task
);
193 static void sdma_set_state(struct qib_pportdata
*ppd
,
194 enum qib_sdma_states next_state
)
196 struct qib_sdma_state
*ss
= &ppd
->sdma_state
;
197 struct sdma_set_state_action
*action
= ss
->set_state_action
;
200 /* debugging bookkeeping */
201 ss
->previous_state
= ss
->current_state
;
202 ss
->previous_op
= ss
->current_op
;
204 ss
->current_state
= next_state
;
206 if (action
[next_state
].op_enable
)
207 op
|= QIB_SDMA_SENDCTRL_OP_ENABLE
;
209 if (action
[next_state
].op_intenable
)
210 op
|= QIB_SDMA_SENDCTRL_OP_INTENABLE
;
212 if (action
[next_state
].op_halt
)
213 op
|= QIB_SDMA_SENDCTRL_OP_HALT
;
215 if (action
[next_state
].op_drain
)
216 op
|= QIB_SDMA_SENDCTRL_OP_DRAIN
;
218 if (action
[next_state
].go_s99_running_tofalse
)
219 ss
->go_s99_running
= 0;
221 if (action
[next_state
].go_s99_running_totrue
)
222 ss
->go_s99_running
= 1;
226 ppd
->dd
->f_sdma_sendctrl(ppd
, ss
->current_op
);
229 static void unmap_desc(struct qib_pportdata
*ppd
, unsigned head
)
231 __le64
*descqp
= &ppd
->sdma_descq
[head
].qw
[0];
236 desc
[0] = le64_to_cpu(descqp
[0]);
237 desc
[1] = le64_to_cpu(descqp
[1]);
239 addr
= (desc
[1] << 32) | (desc
[0] >> 32);
240 len
= (desc
[0] >> 14) & (0x7ffULL
<< 2);
241 dma_unmap_single(&ppd
->dd
->pcidev
->dev
, addr
, len
, DMA_TO_DEVICE
);
244 static int alloc_sdma(struct qib_pportdata
*ppd
)
246 ppd
->sdma_descq_cnt
= sdma_descq_cnt
;
247 if (!ppd
->sdma_descq_cnt
)
248 ppd
->sdma_descq_cnt
= 256;
250 /* Allocate memory for SendDMA descriptor FIFO */
251 ppd
->sdma_descq
= dma_alloc_coherent(&ppd
->dd
->pcidev
->dev
,
252 ppd
->sdma_descq_cnt
* sizeof(u64
[2]), &ppd
->sdma_descq_phys
,
255 if (!ppd
->sdma_descq
) {
257 "failed to allocate SendDMA descriptor FIFO memory\n");
261 /* Allocate memory for DMA of head register to memory */
262 ppd
->sdma_head_dma
= dma_alloc_coherent(&ppd
->dd
->pcidev
->dev
,
263 PAGE_SIZE
, &ppd
->sdma_head_phys
, GFP_KERNEL
);
264 if (!ppd
->sdma_head_dma
) {
266 "failed to allocate SendDMA head memory\n");
269 ppd
->sdma_head_dma
[0] = 0;
273 dma_free_coherent(&ppd
->dd
->pcidev
->dev
,
274 ppd
->sdma_descq_cnt
* sizeof(u64
[2]), (void *)ppd
->sdma_descq
,
275 ppd
->sdma_descq_phys
);
276 ppd
->sdma_descq
= NULL
;
277 ppd
->sdma_descq_phys
= 0;
279 ppd
->sdma_descq_cnt
= 0;
283 static void free_sdma(struct qib_pportdata
*ppd
)
285 struct qib_devdata
*dd
= ppd
->dd
;
287 if (ppd
->sdma_head_dma
) {
288 dma_free_coherent(&dd
->pcidev
->dev
, PAGE_SIZE
,
289 (void *)ppd
->sdma_head_dma
,
290 ppd
->sdma_head_phys
);
291 ppd
->sdma_head_dma
= NULL
;
292 ppd
->sdma_head_phys
= 0;
295 if (ppd
->sdma_descq
) {
296 dma_free_coherent(&dd
->pcidev
->dev
,
297 ppd
->sdma_descq_cnt
* sizeof(u64
[2]),
298 ppd
->sdma_descq
, ppd
->sdma_descq_phys
);
299 ppd
->sdma_descq
= NULL
;
300 ppd
->sdma_descq_phys
= 0;
304 static inline void make_sdma_desc(struct qib_pportdata
*ppd
,
305 u64
*sdmadesc
, u64 addr
, u64 dwlen
,
310 /* SDmaPhyAddr[47:32] */
311 sdmadesc
[1] = addr
>> 32;
312 /* SDmaPhyAddr[31:0] */
313 sdmadesc
[0] = (addr
& 0xfffffffcULL
) << 32;
314 /* SDmaGeneration[1:0] */
315 sdmadesc
[0] |= (ppd
->sdma_generation
& 3ULL) <<
317 /* SDmaDwordCount[10:0] */
318 sdmadesc
[0] |= (dwlen
& 0x7ffULL
) << SDMA_DESC_COUNT_LSB
;
319 /* SDmaBufOffset[12:2] */
320 sdmadesc
[0] |= dwoffset
& 0x7ffULL
;
323 /* sdma_lock must be held */
324 int qib_sdma_make_progress(struct qib_pportdata
*ppd
)
326 struct list_head
*lp
= NULL
;
327 struct qib_sdma_txreq
*txp
= NULL
;
328 struct qib_devdata
*dd
= ppd
->dd
;
333 hwhead
= dd
->f_sdma_gethead(ppd
);
335 /* The reason for some of the complexity of this code is that
336 * not all descriptors have corresponding txps. So, we have to
337 * be able to skip over descs until we wander into the range of
338 * the next txp on the list.
341 if (!list_empty(&ppd
->sdma_activelist
)) {
342 lp
= ppd
->sdma_activelist
.next
;
343 txp
= list_entry(lp
, struct qib_sdma_txreq
, list
);
344 idx
= txp
->start_idx
;
347 while (ppd
->sdma_descq_head
!= hwhead
) {
348 /* if desc is part of this txp, unmap if needed */
349 if (txp
&& (txp
->flags
& QIB_SDMA_TXREQ_F_FREEDESC
) &&
350 (idx
== ppd
->sdma_descq_head
)) {
351 unmap_desc(ppd
, ppd
->sdma_descq_head
);
352 if (++idx
== ppd
->sdma_descq_cnt
)
356 /* increment dequed desc count */
357 ppd
->sdma_descq_removed
++;
359 /* advance head, wrap if needed */
360 if (++ppd
->sdma_descq_head
== ppd
->sdma_descq_cnt
)
361 ppd
->sdma_descq_head
= 0;
363 /* if now past this txp's descs, do the callback */
364 if (txp
&& txp
->next_descq_idx
== ppd
->sdma_descq_head
) {
365 /* remove from active list */
366 list_del_init(&txp
->list
);
368 (*txp
->callback
)(txp
, QIB_SDMA_TXREQ_S_OK
);
369 /* see if there is another txp */
370 if (list_empty(&ppd
->sdma_activelist
))
373 lp
= ppd
->sdma_activelist
.next
;
374 txp
= list_entry(lp
, struct qib_sdma_txreq
,
376 idx
= txp
->start_idx
;
382 qib_verbs_sdma_desc_avail(ppd
, qib_sdma_descq_freecnt(ppd
));
387 * This is called from interrupt context.
389 void qib_sdma_intr(struct qib_pportdata
*ppd
)
393 spin_lock_irqsave(&ppd
->sdma_lock
, flags
);
395 __qib_sdma_intr(ppd
);
397 spin_unlock_irqrestore(&ppd
->sdma_lock
, flags
);
400 void __qib_sdma_intr(struct qib_pportdata
*ppd
)
402 if (__qib_sdma_running(ppd
)) {
403 qib_sdma_make_progress(ppd
);
404 if (!list_empty(&ppd
->sdma_userpending
))
405 qib_user_sdma_send_desc(ppd
, &ppd
->sdma_userpending
);
409 int qib_setup_sdma(struct qib_pportdata
*ppd
)
411 struct qib_devdata
*dd
= ppd
->dd
;
415 ret
= alloc_sdma(ppd
);
419 /* set consistent sdma state */
420 ppd
->dd
->f_sdma_init_early(ppd
);
421 spin_lock_irqsave(&ppd
->sdma_lock
, flags
);
422 sdma_set_state(ppd
, qib_sdma_state_s00_hw_down
);
423 spin_unlock_irqrestore(&ppd
->sdma_lock
, flags
);
425 /* set up reference counting */
426 kref_init(&ppd
->sdma_state
.kref
);
427 init_completion(&ppd
->sdma_state
.comp
);
429 ppd
->sdma_generation
= 0;
430 ppd
->sdma_descq_head
= 0;
431 ppd
->sdma_descq_removed
= 0;
432 ppd
->sdma_descq_added
= 0;
434 ppd
->sdma_intrequest
= 0;
435 INIT_LIST_HEAD(&ppd
->sdma_userpending
);
437 INIT_LIST_HEAD(&ppd
->sdma_activelist
);
439 tasklet_init(&ppd
->sdma_sw_clean_up_task
, sdma_sw_clean_up_task
,
442 ret
= dd
->f_init_sdma_regs(ppd
);
446 qib_sdma_process_event(ppd
, qib_sdma_event_e10_go_hw_start
);
451 qib_teardown_sdma(ppd
);
456 void qib_teardown_sdma(struct qib_pportdata
*ppd
)
458 qib_sdma_process_event(ppd
, qib_sdma_event_e00_go_hw_down
);
461 * This waits for the state machine to exit so it is not
462 * necessary to kill the sdma_sw_clean_up_task to make sure
465 sdma_finalput(&ppd
->sdma_state
);
470 int qib_sdma_running(struct qib_pportdata
*ppd
)
475 spin_lock_irqsave(&ppd
->sdma_lock
, flags
);
476 ret
= __qib_sdma_running(ppd
);
477 spin_unlock_irqrestore(&ppd
->sdma_lock
, flags
);
483 * Complete a request when sdma not running; likely only request
484 * but to simplify the code, always queue it, then process the full
485 * activelist. We process the entire list to ensure that this particular
486 * request does get it's callback, but in the correct order.
487 * Must be called with sdma_lock held
489 static void complete_sdma_err_req(struct qib_pportdata
*ppd
,
490 struct qib_verbs_txreq
*tx
)
492 struct qib_qp_priv
*priv
= tx
->qp
->priv
;
494 atomic_inc(&priv
->s_dma_busy
);
495 /* no sdma descriptors, so no unmap_desc */
496 tx
->txreq
.start_idx
= 0;
497 tx
->txreq
.next_descq_idx
= 0;
498 list_add_tail(&tx
->txreq
.list
, &ppd
->sdma_activelist
);
499 clear_sdma_activelist(ppd
);
503 * This function queues one IB packet onto the send DMA queue per call.
504 * The caller is responsible for checking:
505 * 1) The number of send DMA descriptor entries is less than the size of
506 * the descriptor queue.
507 * 2) The IB SGE addresses and lengths are 32-bit aligned
508 * (except possibly the last SGE's length)
509 * 3) The SGE addresses are suitable for passing to dma_map_single().
511 int qib_sdma_verbs_send(struct qib_pportdata
*ppd
,
512 struct rvt_sge_state
*ss
, u32 dwords
,
513 struct qib_verbs_txreq
*tx
)
524 struct qib_qp_priv
*priv
;
526 spin_lock_irqsave(&ppd
->sdma_lock
, flags
);
529 if (unlikely(!__qib_sdma_running(ppd
))) {
530 complete_sdma_err_req(ppd
, tx
);
534 if (tx
->txreq
.sg_count
> qib_sdma_descq_freecnt(ppd
)) {
535 if (qib_sdma_make_progress(ppd
))
537 if (ppd
->dd
->flags
& QIB_HAS_SDMA_TIMEOUT
)
538 ppd
->dd
->f_sdma_set_desc_cnt(ppd
,
539 ppd
->sdma_descq_cnt
/ 2);
543 dwoffset
= tx
->hdr_dwords
;
544 make_sdma_desc(ppd
, sdmadesc
, (u64
) tx
->txreq
.addr
, dwoffset
, 0);
546 sdmadesc
[0] |= SDMA_DESC_FIRST
;
547 if (tx
->txreq
.flags
& QIB_SDMA_TXREQ_F_USELARGEBUF
)
548 sdmadesc
[0] |= SDMA_DESC_USE_LARGE_BUF
;
550 /* write to the descq */
551 tail
= ppd
->sdma_descq_tail
;
552 descqp
= &ppd
->sdma_descq
[tail
].qw
[0];
553 *descqp
++ = cpu_to_le64(sdmadesc
[0]);
554 *descqp
++ = cpu_to_le64(sdmadesc
[1]);
556 /* increment the tail */
557 if (++tail
== ppd
->sdma_descq_cnt
) {
559 descqp
= &ppd
->sdma_descq
[0].qw
[0];
560 ++ppd
->sdma_generation
;
563 tx
->txreq
.start_idx
= tail
;
568 u32 len
= rvt_get_sge_length(sge
, dwords
<< 2);
571 addr
= dma_map_single(&ppd
->dd
->pcidev
->dev
, sge
->vaddr
,
572 dw
<< 2, DMA_TO_DEVICE
);
573 if (dma_mapping_error(&ppd
->dd
->pcidev
->dev
, addr
)) {
578 make_sdma_desc(ppd
, sdmadesc
, (u64
) addr
, dw
, dwoffset
);
579 /* SDmaUseLargeBuf has to be set in every descriptor */
580 if (tx
->txreq
.flags
& QIB_SDMA_TXREQ_F_USELARGEBUF
)
581 sdmadesc
[0] |= SDMA_DESC_USE_LARGE_BUF
;
582 /* write to the descq */
583 *descqp
++ = cpu_to_le64(sdmadesc
[0]);
584 *descqp
++ = cpu_to_le64(sdmadesc
[1]);
586 /* increment the tail */
587 if (++tail
== ppd
->sdma_descq_cnt
) {
589 descqp
= &ppd
->sdma_descq
[0].qw
[0];
590 ++ppd
->sdma_generation
;
592 rvt_update_sge(ss
, len
, false);
598 descqp
= &ppd
->sdma_descq
[ppd
->sdma_descq_cnt
].qw
[0];
600 descqp
[0] |= cpu_to_le64(SDMA_DESC_LAST
);
601 if (tx
->txreq
.flags
& QIB_SDMA_TXREQ_F_HEADTOHOST
)
602 descqp
[0] |= cpu_to_le64(SDMA_DESC_DMA_HEAD
);
603 if (tx
->txreq
.flags
& QIB_SDMA_TXREQ_F_INTREQ
)
604 descqp
[0] |= cpu_to_le64(SDMA_DESC_INTR
);
606 atomic_inc(&priv
->s_dma_busy
);
607 tx
->txreq
.next_descq_idx
= tail
;
608 ppd
->dd
->f_sdma_update_tail(ppd
, tail
);
609 ppd
->sdma_descq_added
+= tx
->txreq
.sg_count
;
610 list_add_tail(&tx
->txreq
.list
, &ppd
->sdma_activelist
);
616 tail
= ppd
->sdma_descq_cnt
- 1;
619 if (tail
== ppd
->sdma_descq_tail
)
621 unmap_desc(ppd
, tail
);
626 spin_lock(&qp
->r_lock
);
627 spin_lock(&qp
->s_lock
);
628 if (qp
->ibqp
.qp_type
== IB_QPT_RC
) {
629 /* XXX what about error sending RDMA read responses? */
630 if (ib_rvt_state_ops
[qp
->state
] & RVT_PROCESS_RECV_OK
)
631 rvt_error_qp(qp
, IB_WC_GENERAL_ERR
);
632 } else if (qp
->s_wqe
)
633 rvt_send_complete(qp
, qp
->s_wqe
, IB_WC_GENERAL_ERR
);
634 spin_unlock(&qp
->s_lock
);
635 spin_unlock(&qp
->r_lock
);
636 /* return zero to process the next send work request */
642 spin_lock(&qp
->s_lock
);
643 if (ib_rvt_state_ops
[qp
->state
] & RVT_PROCESS_RECV_OK
) {
644 struct qib_ibdev
*dev
;
647 * If we couldn't queue the DMA request, save the info
648 * and try again later rather than destroying the
649 * buffer and undoing the side effects of the copy.
654 dev
= &ppd
->dd
->verbs_dev
;
655 spin_lock(&dev
->rdi
.pending_lock
);
656 if (list_empty(&priv
->iowait
)) {
657 struct qib_ibport
*ibp
;
659 ibp
= &ppd
->ibport_data
;
660 ibp
->rvp
.n_dmawait
++;
661 qp
->s_flags
|= RVT_S_WAIT_DMA_DESC
;
662 list_add_tail(&priv
->iowait
, &dev
->dmawait
);
664 spin_unlock(&dev
->rdi
.pending_lock
);
665 qp
->s_flags
&= ~RVT_S_BUSY
;
666 spin_unlock(&qp
->s_lock
);
669 spin_unlock(&qp
->s_lock
);
673 spin_unlock_irqrestore(&ppd
->sdma_lock
, flags
);
678 * sdma_lock should be acquired before calling this routine
680 void dump_sdma_state(struct qib_pportdata
*ppd
)
682 struct qib_sdma_desc
*descq
;
683 struct qib_sdma_txreq
*txp
, *txpnext
;
687 u16 gen
, dwlen
, dwoffset
;
690 head
= ppd
->sdma_descq_head
;
691 tail
= ppd
->sdma_descq_tail
;
692 cnt
= qib_sdma_descq_freecnt(ppd
);
693 descq
= ppd
->sdma_descq
;
695 qib_dev_porterr(ppd
->dd
, ppd
->port
,
696 "SDMA ppd->sdma_descq_head: %u\n", head
);
697 qib_dev_porterr(ppd
->dd
, ppd
->port
,
698 "SDMA ppd->sdma_descq_tail: %u\n", tail
);
699 qib_dev_porterr(ppd
->dd
, ppd
->port
,
700 "SDMA sdma_descq_freecnt: %u\n", cnt
);
702 /* print info for each entry in the descriptor queue */
703 while (head
!= tail
) {
704 char flags
[6] = { 'x', 'x', 'x', 'x', 'x', 0 };
706 descqp
= &descq
[head
].qw
[0];
707 desc
[0] = le64_to_cpu(descqp
[0]);
708 desc
[1] = le64_to_cpu(descqp
[1]);
709 flags
[0] = (desc
[0] & 1<<15) ? 'I' : '-';
710 flags
[1] = (desc
[0] & 1<<14) ? 'L' : 'S';
711 flags
[2] = (desc
[0] & 1<<13) ? 'H' : '-';
712 flags
[3] = (desc
[0] & 1<<12) ? 'F' : '-';
713 flags
[4] = (desc
[0] & 1<<11) ? 'L' : '-';
714 addr
= (desc
[1] << 32) | ((desc
[0] >> 32) & 0xfffffffcULL
);
715 gen
= (desc
[0] >> 30) & 3ULL;
716 dwlen
= (desc
[0] >> 14) & (0x7ffULL
<< 2);
717 dwoffset
= (desc
[0] & 0x7ffULL
) << 2;
718 qib_dev_porterr(ppd
->dd
, ppd
->port
,
719 "SDMA sdmadesc[%u]: flags:%s addr:0x%016llx gen:%u len:%u bytes offset:%u bytes\n",
720 head
, flags
, addr
, gen
, dwlen
, dwoffset
);
721 if (++head
== ppd
->sdma_descq_cnt
)
725 /* print dma descriptor indices from the TX requests */
726 list_for_each_entry_safe(txp
, txpnext
, &ppd
->sdma_activelist
,
728 qib_dev_porterr(ppd
->dd
, ppd
->port
,
729 "SDMA txp->start_idx: %u txp->next_descq_idx: %u\n",
730 txp
->start_idx
, txp
->next_descq_idx
);
733 void qib_sdma_process_event(struct qib_pportdata
*ppd
,
734 enum qib_sdma_events event
)
738 spin_lock_irqsave(&ppd
->sdma_lock
, flags
);
740 __qib_sdma_process_event(ppd
, event
);
742 if (ppd
->sdma_state
.current_state
== qib_sdma_state_s99_running
)
743 qib_verbs_sdma_desc_avail(ppd
, qib_sdma_descq_freecnt(ppd
));
745 spin_unlock_irqrestore(&ppd
->sdma_lock
, flags
);
748 void __qib_sdma_process_event(struct qib_pportdata
*ppd
,
749 enum qib_sdma_events event
)
751 struct qib_sdma_state
*ss
= &ppd
->sdma_state
;
753 switch (ss
->current_state
) {
754 case qib_sdma_state_s00_hw_down
:
756 case qib_sdma_event_e00_go_hw_down
:
758 case qib_sdma_event_e30_go_running
:
760 * If down, but running requested (usually result
761 * of link up, then we need to start up.
762 * This can happen when hw down is requested while
763 * bringing the link up with traffic active on
765 ss
->go_s99_running
= 1;
766 /* fall through -- and start dma engine */
767 case qib_sdma_event_e10_go_hw_start
:
768 /* This reference means the state machine is started */
769 sdma_get(&ppd
->sdma_state
);
771 qib_sdma_state_s10_hw_start_up_wait
);
773 case qib_sdma_event_e20_hw_started
:
775 case qib_sdma_event_e40_sw_cleaned
:
776 sdma_sw_tear_down(ppd
);
778 case qib_sdma_event_e50_hw_cleaned
:
780 case qib_sdma_event_e60_hw_halted
:
782 case qib_sdma_event_e70_go_idle
:
784 case qib_sdma_event_e7220_err_halted
:
786 case qib_sdma_event_e7322_err_halted
:
788 case qib_sdma_event_e90_timer_tick
:
793 case qib_sdma_state_s10_hw_start_up_wait
:
795 case qib_sdma_event_e00_go_hw_down
:
796 sdma_set_state(ppd
, qib_sdma_state_s00_hw_down
);
797 sdma_sw_tear_down(ppd
);
799 case qib_sdma_event_e10_go_hw_start
:
801 case qib_sdma_event_e20_hw_started
:
802 sdma_set_state(ppd
, ss
->go_s99_running
?
803 qib_sdma_state_s99_running
:
804 qib_sdma_state_s20_idle
);
806 case qib_sdma_event_e30_go_running
:
807 ss
->go_s99_running
= 1;
809 case qib_sdma_event_e40_sw_cleaned
:
811 case qib_sdma_event_e50_hw_cleaned
:
813 case qib_sdma_event_e60_hw_halted
:
815 case qib_sdma_event_e70_go_idle
:
816 ss
->go_s99_running
= 0;
818 case qib_sdma_event_e7220_err_halted
:
820 case qib_sdma_event_e7322_err_halted
:
822 case qib_sdma_event_e90_timer_tick
:
827 case qib_sdma_state_s20_idle
:
829 case qib_sdma_event_e00_go_hw_down
:
830 sdma_set_state(ppd
, qib_sdma_state_s00_hw_down
);
831 sdma_sw_tear_down(ppd
);
833 case qib_sdma_event_e10_go_hw_start
:
835 case qib_sdma_event_e20_hw_started
:
837 case qib_sdma_event_e30_go_running
:
838 sdma_set_state(ppd
, qib_sdma_state_s99_running
);
839 ss
->go_s99_running
= 1;
841 case qib_sdma_event_e40_sw_cleaned
:
843 case qib_sdma_event_e50_hw_cleaned
:
845 case qib_sdma_event_e60_hw_halted
:
847 case qib_sdma_event_e70_go_idle
:
849 case qib_sdma_event_e7220_err_halted
:
851 case qib_sdma_event_e7322_err_halted
:
853 case qib_sdma_event_e90_timer_tick
:
858 case qib_sdma_state_s30_sw_clean_up_wait
:
860 case qib_sdma_event_e00_go_hw_down
:
861 sdma_set_state(ppd
, qib_sdma_state_s00_hw_down
);
863 case qib_sdma_event_e10_go_hw_start
:
865 case qib_sdma_event_e20_hw_started
:
867 case qib_sdma_event_e30_go_running
:
868 ss
->go_s99_running
= 1;
870 case qib_sdma_event_e40_sw_cleaned
:
872 qib_sdma_state_s10_hw_start_up_wait
);
873 sdma_hw_start_up(ppd
);
875 case qib_sdma_event_e50_hw_cleaned
:
877 case qib_sdma_event_e60_hw_halted
:
879 case qib_sdma_event_e70_go_idle
:
880 ss
->go_s99_running
= 0;
882 case qib_sdma_event_e7220_err_halted
:
884 case qib_sdma_event_e7322_err_halted
:
886 case qib_sdma_event_e90_timer_tick
:
891 case qib_sdma_state_s40_hw_clean_up_wait
:
893 case qib_sdma_event_e00_go_hw_down
:
894 sdma_set_state(ppd
, qib_sdma_state_s00_hw_down
);
895 sdma_start_sw_clean_up(ppd
);
897 case qib_sdma_event_e10_go_hw_start
:
899 case qib_sdma_event_e20_hw_started
:
901 case qib_sdma_event_e30_go_running
:
902 ss
->go_s99_running
= 1;
904 case qib_sdma_event_e40_sw_cleaned
:
906 case qib_sdma_event_e50_hw_cleaned
:
908 qib_sdma_state_s30_sw_clean_up_wait
);
909 sdma_start_sw_clean_up(ppd
);
911 case qib_sdma_event_e60_hw_halted
:
913 case qib_sdma_event_e70_go_idle
:
914 ss
->go_s99_running
= 0;
916 case qib_sdma_event_e7220_err_halted
:
918 case qib_sdma_event_e7322_err_halted
:
920 case qib_sdma_event_e90_timer_tick
:
925 case qib_sdma_state_s50_hw_halt_wait
:
927 case qib_sdma_event_e00_go_hw_down
:
928 sdma_set_state(ppd
, qib_sdma_state_s00_hw_down
);
929 sdma_start_sw_clean_up(ppd
);
931 case qib_sdma_event_e10_go_hw_start
:
933 case qib_sdma_event_e20_hw_started
:
935 case qib_sdma_event_e30_go_running
:
936 ss
->go_s99_running
= 1;
938 case qib_sdma_event_e40_sw_cleaned
:
940 case qib_sdma_event_e50_hw_cleaned
:
942 case qib_sdma_event_e60_hw_halted
:
944 qib_sdma_state_s40_hw_clean_up_wait
);
945 ppd
->dd
->f_sdma_hw_clean_up(ppd
);
947 case qib_sdma_event_e70_go_idle
:
948 ss
->go_s99_running
= 0;
950 case qib_sdma_event_e7220_err_halted
:
952 case qib_sdma_event_e7322_err_halted
:
954 case qib_sdma_event_e90_timer_tick
:
959 case qib_sdma_state_s99_running
:
961 case qib_sdma_event_e00_go_hw_down
:
962 sdma_set_state(ppd
, qib_sdma_state_s00_hw_down
);
963 sdma_start_sw_clean_up(ppd
);
965 case qib_sdma_event_e10_go_hw_start
:
967 case qib_sdma_event_e20_hw_started
:
969 case qib_sdma_event_e30_go_running
:
971 case qib_sdma_event_e40_sw_cleaned
:
973 case qib_sdma_event_e50_hw_cleaned
:
975 case qib_sdma_event_e60_hw_halted
:
977 qib_sdma_state_s30_sw_clean_up_wait
);
978 sdma_start_sw_clean_up(ppd
);
980 case qib_sdma_event_e70_go_idle
:
981 sdma_set_state(ppd
, qib_sdma_state_s50_hw_halt_wait
);
982 ss
->go_s99_running
= 0;
984 case qib_sdma_event_e7220_err_halted
:
986 qib_sdma_state_s30_sw_clean_up_wait
);
987 sdma_start_sw_clean_up(ppd
);
989 case qib_sdma_event_e7322_err_halted
:
990 sdma_set_state(ppd
, qib_sdma_state_s50_hw_halt_wait
);
992 case qib_sdma_event_e90_timer_tick
:
998 ss
->last_event
= event
;