2 * Copyright (c) 2007, 2008 QLogic Corporation. All rights reserved.
4 * This software is available to you under a choice of one of two
5 * licenses. You may choose to be licensed under the terms of the GNU
6 * General Public License (GPL) Version 2, available from the file
7 * COPYING in the main directory of this source tree, or the
8 * OpenIB.org BSD license below:
10 * Redistribution and use in source and binary forms, with or
11 * without modification, are permitted provided that the following
14 * - Redistributions of source code must retain the above
15 * copyright notice, this list of conditions and the following
18 * - Redistributions in binary form must reproduce the above
19 * copyright notice, this list of conditions and the following
20 * disclaimer in the documentation and/or other materials
21 * provided with the distribution.
23 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
24 * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
25 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
26 * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
27 * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
28 * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
29 * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
33 #include <linux/spinlock.h>
34 #include <linux/gfp.h>
36 #include "ipath_kernel.h"
37 #include "ipath_verbs.h"
38 #include "ipath_common.h"
40 #define SDMA_DESCQ_SZ PAGE_SIZE /* 256 entries per 4KB page */
42 static void vl15_watchdog_enq(struct ipath_devdata
*dd
)
44 /* ipath_sdma_lock must already be held */
45 if (atomic_inc_return(&dd
->ipath_sdma_vl15_count
) == 1) {
46 unsigned long interval
= (HZ
+ 19) / 20;
47 dd
->ipath_sdma_vl15_timer
.expires
= jiffies
+ interval
;
48 add_timer(&dd
->ipath_sdma_vl15_timer
);
52 static void vl15_watchdog_deq(struct ipath_devdata
*dd
)
54 /* ipath_sdma_lock must already be held */
55 if (atomic_dec_return(&dd
->ipath_sdma_vl15_count
) != 0) {
56 unsigned long interval
= (HZ
+ 19) / 20;
57 mod_timer(&dd
->ipath_sdma_vl15_timer
, jiffies
+ interval
);
59 del_timer(&dd
->ipath_sdma_vl15_timer
);
63 static void vl15_watchdog_timeout(unsigned long opaque
)
65 struct ipath_devdata
*dd
= (struct ipath_devdata
*)opaque
;
67 if (atomic_read(&dd
->ipath_sdma_vl15_count
) != 0) {
68 ipath_dbg("vl15 watchdog timeout - clearing\n");
69 ipath_cancel_sends(dd
, 1);
72 ipath_dbg("vl15 watchdog timeout - "
73 "condition already cleared\n");
77 static void unmap_desc(struct ipath_devdata
*dd
, unsigned head
)
79 __le64
*descqp
= &dd
->ipath_sdma_descq
[head
].qw
[0];
84 desc
[0] = le64_to_cpu(descqp
[0]);
85 desc
[1] = le64_to_cpu(descqp
[1]);
87 addr
= (desc
[1] << 32) | (desc
[0] >> 32);
88 len
= (desc
[0] >> 14) & (0x7ffULL
<< 2);
89 dma_unmap_single(&dd
->pcidev
->dev
, addr
, len
, DMA_TO_DEVICE
);
93 * ipath_sdma_lock should be locked before calling this.
95 int ipath_sdma_make_progress(struct ipath_devdata
*dd
)
97 struct list_head
*lp
= NULL
;
98 struct ipath_sdma_txreq
*txp
= NULL
;
103 if (!list_empty(&dd
->ipath_sdma_activelist
)) {
104 lp
= dd
->ipath_sdma_activelist
.next
;
105 txp
= list_entry(lp
, struct ipath_sdma_txreq
, list
);
106 start_idx
= txp
->start_idx
;
110 * Read the SDMA head register in order to know that the
111 * interrupt clear has been written to the chip.
112 * Otherwise, we may not get an interrupt for the last
113 * descriptor in the queue.
115 dmahead
= (u16
)ipath_read_kreg32(dd
, dd
->ipath_kregs
->kr_senddmahead
);
116 /* sanity check return value for error handling (chip reset, etc.) */
117 if (dmahead
>= dd
->ipath_sdma_descq_cnt
)
120 while (dd
->ipath_sdma_descq_head
!= dmahead
) {
121 if (txp
&& txp
->flags
& IPATH_SDMA_TXREQ_F_FREEDESC
&&
122 dd
->ipath_sdma_descq_head
== start_idx
) {
123 unmap_desc(dd
, dd
->ipath_sdma_descq_head
);
125 if (start_idx
== dd
->ipath_sdma_descq_cnt
)
129 /* increment free count and head */
130 dd
->ipath_sdma_descq_removed
++;
131 if (++dd
->ipath_sdma_descq_head
== dd
->ipath_sdma_descq_cnt
)
132 dd
->ipath_sdma_descq_head
= 0;
134 if (txp
&& txp
->next_descq_idx
== dd
->ipath_sdma_descq_head
) {
135 /* move to notify list */
136 if (txp
->flags
& IPATH_SDMA_TXREQ_F_VL15
)
137 vl15_watchdog_deq(dd
);
138 list_move_tail(lp
, &dd
->ipath_sdma_notifylist
);
139 if (!list_empty(&dd
->ipath_sdma_activelist
)) {
140 lp
= dd
->ipath_sdma_activelist
.next
;
141 txp
= list_entry(lp
, struct ipath_sdma_txreq
,
143 start_idx
= txp
->start_idx
;
153 tasklet_hi_schedule(&dd
->ipath_sdma_notify_task
);
159 static void ipath_sdma_notify(struct ipath_devdata
*dd
, struct list_head
*list
)
161 struct ipath_sdma_txreq
*txp
, *txp_next
;
163 list_for_each_entry_safe(txp
, txp_next
, list
, list
) {
164 list_del_init(&txp
->list
);
167 (*txp
->callback
)(txp
->callback_cookie
,
168 txp
->callback_status
);
172 static void sdma_notify_taskbody(struct ipath_devdata
*dd
)
175 struct list_head list
;
177 INIT_LIST_HEAD(&list
);
179 spin_lock_irqsave(&dd
->ipath_sdma_lock
, flags
);
181 list_splice_init(&dd
->ipath_sdma_notifylist
, &list
);
183 spin_unlock_irqrestore(&dd
->ipath_sdma_lock
, flags
);
185 ipath_sdma_notify(dd
, &list
);
188 * The IB verbs layer needs to see the callback before getting
189 * the call to ipath_ib_piobufavail() because the callback
190 * handles releasing resources the next send will need.
191 * Otherwise, we could do these calls in
192 * ipath_sdma_make_progress().
194 ipath_ib_piobufavail(dd
->verbs_dev
);
197 static void sdma_notify_task(unsigned long opaque
)
199 struct ipath_devdata
*dd
= (struct ipath_devdata
*)opaque
;
201 if (!test_bit(IPATH_SDMA_SHUTDOWN
, &dd
->ipath_sdma_status
))
202 sdma_notify_taskbody(dd
);
205 static void dump_sdma_state(struct ipath_devdata
*dd
)
209 reg
= ipath_read_kreg64(dd
, dd
->ipath_kregs
->kr_senddmastatus
);
210 ipath_cdbg(VERBOSE
, "kr_senddmastatus: 0x%016lx\n", reg
);
212 reg
= ipath_read_kreg64(dd
, dd
->ipath_kregs
->kr_sendctrl
);
213 ipath_cdbg(VERBOSE
, "kr_sendctrl: 0x%016lx\n", reg
);
215 reg
= ipath_read_kreg64(dd
, dd
->ipath_kregs
->kr_senddmabufmask0
);
216 ipath_cdbg(VERBOSE
, "kr_senddmabufmask0: 0x%016lx\n", reg
);
218 reg
= ipath_read_kreg64(dd
, dd
->ipath_kregs
->kr_senddmabufmask1
);
219 ipath_cdbg(VERBOSE
, "kr_senddmabufmask1: 0x%016lx\n", reg
);
221 reg
= ipath_read_kreg64(dd
, dd
->ipath_kregs
->kr_senddmabufmask2
);
222 ipath_cdbg(VERBOSE
, "kr_senddmabufmask2: 0x%016lx\n", reg
);
224 reg
= ipath_read_kreg64(dd
, dd
->ipath_kregs
->kr_senddmatail
);
225 ipath_cdbg(VERBOSE
, "kr_senddmatail: 0x%016lx\n", reg
);
227 reg
= ipath_read_kreg64(dd
, dd
->ipath_kregs
->kr_senddmahead
);
228 ipath_cdbg(VERBOSE
, "kr_senddmahead: 0x%016lx\n", reg
);
231 static void sdma_abort_task(unsigned long opaque
)
233 struct ipath_devdata
*dd
= (struct ipath_devdata
*) opaque
;
237 if (test_bit(IPATH_SDMA_SHUTDOWN
, &dd
->ipath_sdma_status
))
240 spin_lock_irqsave(&dd
->ipath_sdma_lock
, flags
);
242 status
= dd
->ipath_sdma_status
& IPATH_SDMA_ABORT_MASK
;
245 if (status
== IPATH_SDMA_ABORT_NONE
)
248 /* ipath_sdma_abort() is done, waiting for interrupt */
249 if (status
== IPATH_SDMA_ABORT_DISARMED
) {
250 if (jiffies
< dd
->ipath_sdma_abort_intr_timeout
)
251 goto resched_noprint
;
252 /* give up, intr got lost somewhere */
253 ipath_dbg("give up waiting for SDMADISABLED intr\n");
254 __set_bit(IPATH_SDMA_DISABLED
, &dd
->ipath_sdma_status
);
255 status
= IPATH_SDMA_ABORT_ABORTED
;
258 /* everything is stopped, time to clean up and restart */
259 if (status
== IPATH_SDMA_ABORT_ABORTED
) {
260 struct ipath_sdma_txreq
*txp
, *txpnext
;
264 hwstatus
= ipath_read_kreg64(dd
,
265 dd
->ipath_kregs
->kr_senddmastatus
);
267 if ((hwstatus
& (IPATH_SDMA_STATUS_SCORE_BOARD_DRAIN_IN_PROG
|
268 IPATH_SDMA_STATUS_ABORT_IN_PROG
|
269 IPATH_SDMA_STATUS_INTERNAL_SDMA_ENABLE
)) ||
270 !(hwstatus
& IPATH_SDMA_STATUS_SCB_EMPTY
)) {
271 if (dd
->ipath_sdma_reset_wait
> 0) {
272 /* not done shutting down sdma */
273 --dd
->ipath_sdma_reset_wait
;
276 ipath_cdbg(VERBOSE
, "gave up waiting for quiescent "
277 "status after SDMA reset, continuing\n");
281 /* dequeue all "sent" requests */
282 list_for_each_entry_safe(txp
, txpnext
,
283 &dd
->ipath_sdma_activelist
, list
) {
284 txp
->callback_status
= IPATH_SDMA_TXREQ_S_ABORTED
;
285 if (txp
->flags
& IPATH_SDMA_TXREQ_F_VL15
)
286 vl15_watchdog_deq(dd
);
287 list_move_tail(&txp
->list
, &dd
->ipath_sdma_notifylist
);
291 tasklet_hi_schedule(&dd
->ipath_sdma_notify_task
);
293 /* reset our notion of head and tail */
294 dd
->ipath_sdma_descq_tail
= 0;
295 dd
->ipath_sdma_descq_head
= 0;
296 dd
->ipath_sdma_head_dma
[0] = 0;
297 dd
->ipath_sdma_generation
= 0;
298 dd
->ipath_sdma_descq_removed
= dd
->ipath_sdma_descq_added
;
300 /* Reset SendDmaLenGen */
301 ipath_write_kreg(dd
, dd
->ipath_kregs
->kr_senddmalengen
,
302 (u64
) dd
->ipath_sdma_descq_cnt
| (1ULL << 18));
304 /* done with sdma state for a bit */
305 spin_unlock_irqrestore(&dd
->ipath_sdma_lock
, flags
);
308 * Don't restart sdma here (with the exception
309 * below). Wait until link is up to ACTIVE. VL15 MADs
310 * used to bring the link up use PIO, and multiple link
311 * transitions otherwise cause the sdma engine to be
312 * stopped and started multiple times.
313 * The disable is done here, including the shadow,
314 * so the state is kept consistent.
315 * See ipath_restart_sdma() for the actual starting
318 spin_lock_irqsave(&dd
->ipath_sendctrl_lock
, flags
);
319 dd
->ipath_sendctrl
&= ~INFINIPATH_S_SDMAENABLE
;
320 ipath_write_kreg(dd
, dd
->ipath_kregs
->kr_sendctrl
,
322 ipath_read_kreg64(dd
, dd
->ipath_kregs
->kr_scratch
);
323 spin_unlock_irqrestore(&dd
->ipath_sendctrl_lock
, flags
);
325 /* make sure I see next message */
326 dd
->ipath_sdma_abort_jiffies
= 0;
329 * Not everything that takes SDMA offline is a link
330 * status change. If the link was up, restart SDMA.
332 if (dd
->ipath_flags
& IPATH_LINKACTIVE
)
333 ipath_restart_sdma(dd
);
340 * for now, keep spinning
341 * JAG - this is bad to just have default be a loop without
344 if (jiffies
> dd
->ipath_sdma_abort_jiffies
) {
345 ipath_dbg("looping with status 0x%08lx\n",
346 dd
->ipath_sdma_status
);
347 dd
->ipath_sdma_abort_jiffies
= jiffies
+ 5 * HZ
;
350 spin_unlock_irqrestore(&dd
->ipath_sdma_lock
, flags
);
351 if (!test_bit(IPATH_SDMA_SHUTDOWN
, &dd
->ipath_sdma_status
))
352 tasklet_hi_schedule(&dd
->ipath_sdma_abort_task
);
356 spin_unlock_irqrestore(&dd
->ipath_sdma_lock
, flags
);
362 * This is called from interrupt context.
364 void ipath_sdma_intr(struct ipath_devdata
*dd
)
368 spin_lock_irqsave(&dd
->ipath_sdma_lock
, flags
);
370 (void) ipath_sdma_make_progress(dd
);
372 spin_unlock_irqrestore(&dd
->ipath_sdma_lock
, flags
);
375 static int alloc_sdma(struct ipath_devdata
*dd
)
379 /* Allocate memory for SendDMA descriptor FIFO */
380 dd
->ipath_sdma_descq
= dma_alloc_coherent(&dd
->pcidev
->dev
,
381 SDMA_DESCQ_SZ
, &dd
->ipath_sdma_descq_phys
, GFP_KERNEL
);
383 if (!dd
->ipath_sdma_descq
) {
384 ipath_dev_err(dd
, "failed to allocate SendDMA descriptor "
390 dd
->ipath_sdma_descq_cnt
=
391 SDMA_DESCQ_SZ
/ sizeof(struct ipath_sdma_desc
);
393 /* Allocate memory for DMA of head register to memory */
394 dd
->ipath_sdma_head_dma
= dma_alloc_coherent(&dd
->pcidev
->dev
,
395 PAGE_SIZE
, &dd
->ipath_sdma_head_phys
, GFP_KERNEL
);
396 if (!dd
->ipath_sdma_head_dma
) {
397 ipath_dev_err(dd
, "failed to allocate SendDMA head memory\n");
401 dd
->ipath_sdma_head_dma
[0] = 0;
403 init_timer(&dd
->ipath_sdma_vl15_timer
);
404 dd
->ipath_sdma_vl15_timer
.function
= vl15_watchdog_timeout
;
405 dd
->ipath_sdma_vl15_timer
.data
= (unsigned long)dd
;
406 atomic_set(&dd
->ipath_sdma_vl15_count
, 0);
411 dma_free_coherent(&dd
->pcidev
->dev
, SDMA_DESCQ_SZ
,
412 (void *)dd
->ipath_sdma_descq
, dd
->ipath_sdma_descq_phys
);
413 dd
->ipath_sdma_descq
= NULL
;
414 dd
->ipath_sdma_descq_phys
= 0;
419 int setup_sdma(struct ipath_devdata
*dd
)
424 u64 senddmabufmask
[3] = { 0 };
427 ret
= alloc_sdma(dd
);
431 if (!dd
->ipath_sdma_descq
) {
432 ipath_dev_err(dd
, "SendDMA memory not allocated\n");
437 * Set initial status as if we had been up, then gone down.
438 * This lets initial start on transition to ACTIVE be the
439 * same as restart after link flap.
441 dd
->ipath_sdma_status
= IPATH_SDMA_ABORT_ABORTED
;
442 dd
->ipath_sdma_abort_jiffies
= 0;
443 dd
->ipath_sdma_generation
= 0;
444 dd
->ipath_sdma_descq_tail
= 0;
445 dd
->ipath_sdma_descq_head
= 0;
446 dd
->ipath_sdma_descq_removed
= 0;
447 dd
->ipath_sdma_descq_added
= 0;
449 /* Set SendDmaBase */
450 ipath_write_kreg(dd
, dd
->ipath_kregs
->kr_senddmabase
,
451 dd
->ipath_sdma_descq_phys
);
452 /* Set SendDmaLenGen */
453 tmp64
= dd
->ipath_sdma_descq_cnt
;
454 tmp64
|= 1<<18; /* enable generation checking */
455 ipath_write_kreg(dd
, dd
->ipath_kregs
->kr_senddmalengen
, tmp64
);
456 /* Set SendDmaTail */
457 ipath_write_kreg(dd
, dd
->ipath_kregs
->kr_senddmatail
,
458 dd
->ipath_sdma_descq_tail
);
459 /* Set SendDmaHeadAddr */
460 ipath_write_kreg(dd
, dd
->ipath_kregs
->kr_senddmaheadaddr
,
461 dd
->ipath_sdma_head_phys
);
464 * Reserve all the former "kernel" piobufs, using high number range
465 * so we get as many 4K buffers as possible
467 n
= dd
->ipath_piobcnt2k
+ dd
->ipath_piobcnt4k
;
468 i
= dd
->ipath_lastport_piobuf
+ dd
->ipath_pioreserved
;
469 ipath_chg_pioavailkernel(dd
, i
, n
- i
, 0);
471 unsigned word
= i
/ 64;
472 unsigned bit
= i
& 63;
474 senddmabufmask
[word
] |= 1ULL << bit
;
476 ipath_write_kreg(dd
, dd
->ipath_kregs
->kr_senddmabufmask0
,
478 ipath_write_kreg(dd
, dd
->ipath_kregs
->kr_senddmabufmask1
,
480 ipath_write_kreg(dd
, dd
->ipath_kregs
->kr_senddmabufmask2
,
483 INIT_LIST_HEAD(&dd
->ipath_sdma_activelist
);
484 INIT_LIST_HEAD(&dd
->ipath_sdma_notifylist
);
486 tasklet_init(&dd
->ipath_sdma_notify_task
, sdma_notify_task
,
488 tasklet_init(&dd
->ipath_sdma_abort_task
, sdma_abort_task
,
492 * No use to turn on SDMA here, as link is probably not ACTIVE
493 * Just mark it RUNNING and enable the interrupt, and let the
494 * ipath_restart_sdma() on link transition to ACTIVE actually
497 spin_lock_irqsave(&dd
->ipath_sendctrl_lock
, flags
);
498 dd
->ipath_sendctrl
|= INFINIPATH_S_SDMAINTENABLE
;
499 ipath_write_kreg(dd
, dd
->ipath_kregs
->kr_sendctrl
, dd
->ipath_sendctrl
);
500 ipath_read_kreg64(dd
, dd
->ipath_kregs
->kr_scratch
);
501 __set_bit(IPATH_SDMA_RUNNING
, &dd
->ipath_sdma_status
);
502 spin_unlock_irqrestore(&dd
->ipath_sendctrl_lock
, flags
);
508 void teardown_sdma(struct ipath_devdata
*dd
)
510 struct ipath_sdma_txreq
*txp
, *txpnext
;
512 dma_addr_t sdma_head_phys
= 0;
513 dma_addr_t sdma_descq_phys
= 0;
514 void *sdma_descq
= NULL
;
515 void *sdma_head_dma
= NULL
;
517 spin_lock_irqsave(&dd
->ipath_sdma_lock
, flags
);
518 __clear_bit(IPATH_SDMA_RUNNING
, &dd
->ipath_sdma_status
);
519 __set_bit(IPATH_SDMA_ABORTING
, &dd
->ipath_sdma_status
);
520 __set_bit(IPATH_SDMA_SHUTDOWN
, &dd
->ipath_sdma_status
);
521 spin_unlock_irqrestore(&dd
->ipath_sdma_lock
, flags
);
523 tasklet_kill(&dd
->ipath_sdma_abort_task
);
524 tasklet_kill(&dd
->ipath_sdma_notify_task
);
527 spin_lock_irqsave(&dd
->ipath_sendctrl_lock
, flags
);
528 dd
->ipath_sendctrl
&= ~INFINIPATH_S_SDMAENABLE
;
529 ipath_write_kreg(dd
, dd
->ipath_kregs
->kr_sendctrl
,
531 ipath_read_kreg64(dd
, dd
->ipath_kregs
->kr_scratch
);
532 spin_unlock_irqrestore(&dd
->ipath_sendctrl_lock
, flags
);
534 spin_lock_irqsave(&dd
->ipath_sdma_lock
, flags
);
535 /* dequeue all "sent" requests */
536 list_for_each_entry_safe(txp
, txpnext
, &dd
->ipath_sdma_activelist
,
538 txp
->callback_status
= IPATH_SDMA_TXREQ_S_SHUTDOWN
;
539 if (txp
->flags
& IPATH_SDMA_TXREQ_F_VL15
)
540 vl15_watchdog_deq(dd
);
541 list_move_tail(&txp
->list
, &dd
->ipath_sdma_notifylist
);
543 spin_unlock_irqrestore(&dd
->ipath_sdma_lock
, flags
);
545 sdma_notify_taskbody(dd
);
547 del_timer_sync(&dd
->ipath_sdma_vl15_timer
);
549 spin_lock_irqsave(&dd
->ipath_sdma_lock
, flags
);
551 dd
->ipath_sdma_abort_jiffies
= 0;
553 ipath_write_kreg(dd
, dd
->ipath_kregs
->kr_senddmabase
, 0);
554 ipath_write_kreg(dd
, dd
->ipath_kregs
->kr_senddmalengen
, 0);
555 ipath_write_kreg(dd
, dd
->ipath_kregs
->kr_senddmatail
, 0);
556 ipath_write_kreg(dd
, dd
->ipath_kregs
->kr_senddmaheadaddr
, 0);
557 ipath_write_kreg(dd
, dd
->ipath_kregs
->kr_senddmabufmask0
, 0);
558 ipath_write_kreg(dd
, dd
->ipath_kregs
->kr_senddmabufmask1
, 0);
559 ipath_write_kreg(dd
, dd
->ipath_kregs
->kr_senddmabufmask2
, 0);
561 if (dd
->ipath_sdma_head_dma
) {
562 sdma_head_dma
= (void *) dd
->ipath_sdma_head_dma
;
563 sdma_head_phys
= dd
->ipath_sdma_head_phys
;
564 dd
->ipath_sdma_head_dma
= NULL
;
565 dd
->ipath_sdma_head_phys
= 0;
568 if (dd
->ipath_sdma_descq
) {
569 sdma_descq
= dd
->ipath_sdma_descq
;
570 sdma_descq_phys
= dd
->ipath_sdma_descq_phys
;
571 dd
->ipath_sdma_descq
= NULL
;
572 dd
->ipath_sdma_descq_phys
= 0;
575 spin_unlock_irqrestore(&dd
->ipath_sdma_lock
, flags
);
578 dma_free_coherent(&dd
->pcidev
->dev
, PAGE_SIZE
,
579 sdma_head_dma
, sdma_head_phys
);
582 dma_free_coherent(&dd
->pcidev
->dev
, SDMA_DESCQ_SZ
,
583 sdma_descq
, sdma_descq_phys
);
587 * [Re]start SDMA, if we use it, and it's not already OK.
588 * This is called on transition to link ACTIVE, either the first or
591 void ipath_restart_sdma(struct ipath_devdata
*dd
)
596 if (!(dd
->ipath_flags
& IPATH_HAS_SEND_DMA
))
600 * First, make sure we should, which is to say,
601 * check that we are "RUNNING" (not in teardown)
604 spin_lock_irqsave(&dd
->ipath_sdma_lock
, flags
);
605 if (!test_bit(IPATH_SDMA_RUNNING
, &dd
->ipath_sdma_status
)
606 || test_bit(IPATH_SDMA_SHUTDOWN
, &dd
->ipath_sdma_status
))
609 __clear_bit(IPATH_SDMA_DISABLED
, &dd
->ipath_sdma_status
);
610 __clear_bit(IPATH_SDMA_DISARMED
, &dd
->ipath_sdma_status
);
611 __clear_bit(IPATH_SDMA_ABORTING
, &dd
->ipath_sdma_status
);
613 spin_unlock_irqrestore(&dd
->ipath_sdma_lock
, flags
);
615 ipath_dbg("invalid attempt to restart SDMA, status 0x%08lx\n",
616 dd
->ipath_sdma_status
);
619 spin_lock_irqsave(&dd
->ipath_sendctrl_lock
, flags
);
621 * First clear, just to be safe. Enable is only done
622 * in chip on 0->1 transition
624 dd
->ipath_sendctrl
&= ~INFINIPATH_S_SDMAENABLE
;
625 ipath_write_kreg(dd
, dd
->ipath_kregs
->kr_sendctrl
, dd
->ipath_sendctrl
);
626 ipath_read_kreg64(dd
, dd
->ipath_kregs
->kr_scratch
);
627 dd
->ipath_sendctrl
|= INFINIPATH_S_SDMAENABLE
;
628 ipath_write_kreg(dd
, dd
->ipath_kregs
->kr_sendctrl
, dd
->ipath_sendctrl
);
629 ipath_read_kreg64(dd
, dd
->ipath_kregs
->kr_scratch
);
630 spin_unlock_irqrestore(&dd
->ipath_sendctrl_lock
, flags
);
632 /* notify upper layers */
633 ipath_ib_piobufavail(dd
->verbs_dev
);
639 static inline void make_sdma_desc(struct ipath_devdata
*dd
,
640 u64
*sdmadesc
, u64 addr
, u64 dwlen
, u64 dwoffset
)
643 /* SDmaPhyAddr[47:32] */
644 sdmadesc
[1] = addr
>> 32;
645 /* SDmaPhyAddr[31:0] */
646 sdmadesc
[0] = (addr
& 0xfffffffcULL
) << 32;
647 /* SDmaGeneration[1:0] */
648 sdmadesc
[0] |= (dd
->ipath_sdma_generation
& 3ULL) << 30;
649 /* SDmaDwordCount[10:0] */
650 sdmadesc
[0] |= (dwlen
& 0x7ffULL
) << 16;
651 /* SDmaBufOffset[12:2] */
652 sdmadesc
[0] |= dwoffset
& 0x7ffULL
;
656 * This function queues one IB packet onto the send DMA queue per call.
657 * The caller is responsible for checking:
658 * 1) The number of send DMA descriptor entries is less than the size of
659 * the descriptor queue.
660 * 2) The IB SGE addresses and lengths are 32-bit aligned
661 * (except possibly the last SGE's length)
662 * 3) The SGE addresses are suitable for passing to dma_map_single().
664 int ipath_sdma_verbs_send(struct ipath_devdata
*dd
,
665 struct ipath_sge_state
*ss
, u32 dwords
,
666 struct ipath_verbs_txreq
*tx
)
670 struct ipath_sge
*sge
;
678 if ((tx
->map_len
+ (dwords
<<2)) > dd
->ipath_ibmaxlen
) {
679 ipath_dbg("packet size %X > ibmax %X, fail\n",
680 tx
->map_len
+ (dwords
<<2), dd
->ipath_ibmaxlen
);
685 spin_lock_irqsave(&dd
->ipath_sdma_lock
, flags
);
688 if (unlikely(test_bit(IPATH_SDMA_ABORTING
, &dd
->ipath_sdma_status
))) {
693 if (tx
->txreq
.sg_count
> ipath_sdma_descq_freecnt(dd
)) {
694 if (ipath_sdma_make_progress(dd
))
700 addr
= dma_map_single(&dd
->pcidev
->dev
, tx
->txreq
.map_addr
,
701 tx
->map_len
, DMA_TO_DEVICE
);
702 if (dma_mapping_error(&dd
->pcidev
->dev
, addr
))
705 dwoffset
= tx
->map_len
>> 2;
706 make_sdma_desc(dd
, sdmadesc
, (u64
) addr
, dwoffset
, 0);
709 sdmadesc
[0] |= 1ULL << 12;
710 if (tx
->txreq
.flags
& IPATH_SDMA_TXREQ_F_USELARGEBUF
)
711 sdmadesc
[0] |= 1ULL << 14; /* SDmaUseLargeBuf */
713 /* write to the descq */
714 tail
= dd
->ipath_sdma_descq_tail
;
715 descqp
= &dd
->ipath_sdma_descq
[tail
].qw
[0];
716 *descqp
++ = cpu_to_le64(sdmadesc
[0]);
717 *descqp
++ = cpu_to_le64(sdmadesc
[1]);
719 if (tx
->txreq
.flags
& IPATH_SDMA_TXREQ_F_FREEDESC
)
720 tx
->txreq
.start_idx
= tail
;
722 /* increment the tail */
723 if (++tail
== dd
->ipath_sdma_descq_cnt
) {
725 descqp
= &dd
->ipath_sdma_descq
[0].qw
[0];
726 ++dd
->ipath_sdma_generation
;
735 if (len
> sge
->length
)
737 if (len
> sge
->sge_length
)
738 len
= sge
->sge_length
;
741 addr
= dma_map_single(&dd
->pcidev
->dev
, sge
->vaddr
, dw
<< 2,
743 if (dma_mapping_error(&dd
->pcidev
->dev
, addr
))
745 make_sdma_desc(dd
, sdmadesc
, (u64
) addr
, dw
, dwoffset
);
746 /* SDmaUseLargeBuf has to be set in every descriptor */
747 if (tx
->txreq
.flags
& IPATH_SDMA_TXREQ_F_USELARGEBUF
)
748 sdmadesc
[0] |= 1ULL << 14;
749 /* write to the descq */
750 *descqp
++ = cpu_to_le64(sdmadesc
[0]);
751 *descqp
++ = cpu_to_le64(sdmadesc
[1]);
753 /* increment the tail */
754 if (++tail
== dd
->ipath_sdma_descq_cnt
) {
756 descqp
= &dd
->ipath_sdma_descq
[0].qw
[0];
757 ++dd
->ipath_sdma_generation
;
761 sge
->sge_length
-= len
;
762 if (sge
->sge_length
== 0) {
764 *sge
= *ss
->sg_list
++;
765 } else if (sge
->length
== 0 && sge
->mr
!= NULL
) {
766 if (++sge
->n
>= IPATH_SEGSZ
) {
767 if (++sge
->m
>= sge
->mr
->mapsz
)
772 sge
->mr
->map
[sge
->m
]->segs
[sge
->n
].vaddr
;
774 sge
->mr
->map
[sge
->m
]->segs
[sge
->n
].length
;
782 descqp
= &dd
->ipath_sdma_descq
[dd
->ipath_sdma_descq_cnt
].qw
[0];
785 descqp
[0] |= cpu_to_le64(1ULL << 11);
786 if (tx
->txreq
.flags
& IPATH_SDMA_TXREQ_F_INTREQ
) {
788 descqp
[0] |= cpu_to_le64(1ULL << 15);
791 /* Commit writes to memory and advance the tail on the chip */
793 ipath_write_kreg(dd
, dd
->ipath_kregs
->kr_senddmatail
, tail
);
795 tx
->txreq
.next_descq_idx
= tail
;
796 tx
->txreq
.callback_status
= IPATH_SDMA_TXREQ_S_OK
;
797 dd
->ipath_sdma_descq_tail
= tail
;
798 dd
->ipath_sdma_descq_added
+= tx
->txreq
.sg_count
;
799 list_add_tail(&tx
->txreq
.list
, &dd
->ipath_sdma_activelist
);
800 if (tx
->txreq
.flags
& IPATH_SDMA_TXREQ_F_VL15
)
801 vl15_watchdog_enq(dd
);
805 while (tail
!= dd
->ipath_sdma_descq_tail
) {
807 tail
= dd
->ipath_sdma_descq_cnt
- 1;
810 unmap_desc(dd
, tail
);
815 spin_unlock_irqrestore(&dd
->ipath_sdma_lock
, flags
);