1 // SPDX-License-Identifier: GPL-2.0-only
3 * MOXA ART SoCs DMA Engine support.
5 * Copyright (C) 2013 Jonas Jensen
7 * Jonas Jensen <jonas.jensen@gmail.com>
10 #include <linux/dmaengine.h>
11 #include <linux/dma-mapping.h>
12 #include <linux/err.h>
13 #include <linux/init.h>
14 #include <linux/interrupt.h>
15 #include <linux/list.h>
16 #include <linux/module.h>
17 #include <linux/platform_device.h>
18 #include <linux/slab.h>
19 #include <linux/spinlock.h>
20 #include <linux/of_address.h>
21 #include <linux/of_irq.h>
22 #include <linux/of_dma.h>
23 #include <linux/bitops.h>
25 #include <asm/cacheflush.h>
27 #include "dmaengine.h"
30 #define APB_DMA_MAX_CHANNEL 4
32 #define REG_OFF_ADDRESS_SOURCE 0
33 #define REG_OFF_ADDRESS_DEST 4
34 #define REG_OFF_CYCLES 8
35 #define REG_OFF_CTRL 12
36 #define REG_OFF_CHAN_SIZE 16
38 #define APB_DMA_ENABLE BIT(0)
39 #define APB_DMA_FIN_INT_STS BIT(1)
40 #define APB_DMA_FIN_INT_EN BIT(2)
41 #define APB_DMA_BURST_MODE BIT(3)
42 #define APB_DMA_ERR_INT_STS BIT(4)
43 #define APB_DMA_ERR_INT_EN BIT(5)
49 #define APB_DMA_SOURCE_SELECT 0x40
50 #define APB_DMA_DEST_SELECT 0x80
52 #define APB_DMA_SOURCE 0x100
53 #define APB_DMA_DEST 0x1000
55 #define APB_DMA_SOURCE_MASK 0x700
56 #define APB_DMA_DEST_MASK 0x7000
60 * 001: +1 (Burst=0), +4 (Burst=1)
61 * 010: +2 (Burst=0), +8 (Burst=1)
62 * 011: +4 (Burst=0), +16 (Burst=1)
63 * 101: -1 (Burst=0), -4 (Burst=1)
64 * 110: -2 (Burst=0), -8 (Burst=1)
65 * 111: -4 (Burst=0), -16 (Burst=1)
67 #define APB_DMA_SOURCE_INC_0 0
68 #define APB_DMA_SOURCE_INC_1_4 0x100
69 #define APB_DMA_SOURCE_INC_2_8 0x200
70 #define APB_DMA_SOURCE_INC_4_16 0x300
71 #define APB_DMA_SOURCE_DEC_1_4 0x500
72 #define APB_DMA_SOURCE_DEC_2_8 0x600
73 #define APB_DMA_SOURCE_DEC_4_16 0x700
74 #define APB_DMA_DEST_INC_0 0
75 #define APB_DMA_DEST_INC_1_4 0x1000
76 #define APB_DMA_DEST_INC_2_8 0x2000
77 #define APB_DMA_DEST_INC_4_16 0x3000
78 #define APB_DMA_DEST_DEC_1_4 0x5000
79 #define APB_DMA_DEST_DEC_2_8 0x6000
80 #define APB_DMA_DEST_DEC_4_16 0x7000
83 * Request signal select source/destination address for DMA hardware handshake.
85 * The request line number is a property of the DMA controller itself,
86 * e.g. MMC must always request channels where dma_slave_config->slave_id is 5.
88 * 0: No request / Grant signal
89 * 1-15: Request / Grant signal
91 #define APB_DMA_SOURCE_REQ_NO 0x1000000
92 #define APB_DMA_SOURCE_REQ_NO_MASK 0xf000000
93 #define APB_DMA_DEST_REQ_NO 0x10000
94 #define APB_DMA_DEST_REQ_NO_MASK 0xf0000
96 #define APB_DMA_DATA_WIDTH 0x100000
97 #define APB_DMA_DATA_WIDTH_MASK 0x300000
99 * Data width of transfer:
105 #define APB_DMA_DATA_WIDTH_4 0
106 #define APB_DMA_DATA_WIDTH_2 0x100000
107 #define APB_DMA_DATA_WIDTH_1 0x200000
109 #define APB_DMA_CYCLES_MASK 0x00ffffff
111 #define MOXART_DMA_DATA_TYPE_S8 0x00
112 #define MOXART_DMA_DATA_TYPE_S16 0x01
113 #define MOXART_DMA_DATA_TYPE_S32 0x02
121 enum dma_transfer_direction dma_dir
;
124 unsigned int dma_cycles
;
125 struct virt_dma_desc vd
;
127 struct moxart_sg sg
[] __counted_by(sglen
);
131 struct virt_dma_chan vc
;
134 struct moxart_desc
*desc
;
136 struct dma_slave_config cfg
;
141 unsigned int line_reqno
;
145 struct moxart_dmadev
{
146 struct dma_device dma_slave
;
147 struct moxart_chan slave_chans
[APB_DMA_MAX_CHANNEL
];
151 static const unsigned int es_bytes
[] = {
152 [MOXART_DMA_DATA_TYPE_S8
] = 1,
153 [MOXART_DMA_DATA_TYPE_S16
] = 2,
154 [MOXART_DMA_DATA_TYPE_S32
] = 4,
157 static struct device
*chan2dev(struct dma_chan
*chan
)
159 return &chan
->dev
->device
;
162 static inline struct moxart_chan
*to_moxart_dma_chan(struct dma_chan
*c
)
164 return container_of(c
, struct moxart_chan
, vc
.chan
);
167 static inline struct moxart_desc
*to_moxart_dma_desc(
168 struct dma_async_tx_descriptor
*t
)
170 return container_of(t
, struct moxart_desc
, vd
.tx
);
173 static void moxart_dma_desc_free(struct virt_dma_desc
*vd
)
175 kfree(container_of(vd
, struct moxart_desc
, vd
));
178 static int moxart_terminate_all(struct dma_chan
*chan
)
180 struct moxart_chan
*ch
= to_moxart_dma_chan(chan
);
185 dev_dbg(chan2dev(chan
), "%s: ch=%p\n", __func__
, ch
);
187 spin_lock_irqsave(&ch
->vc
.lock
, flags
);
190 moxart_dma_desc_free(&ch
->desc
->vd
);
194 ctrl
= readl(ch
->base
+ REG_OFF_CTRL
);
195 ctrl
&= ~(APB_DMA_ENABLE
| APB_DMA_FIN_INT_EN
| APB_DMA_ERR_INT_EN
);
196 writel(ctrl
, ch
->base
+ REG_OFF_CTRL
);
198 vchan_get_all_descriptors(&ch
->vc
, &head
);
199 spin_unlock_irqrestore(&ch
->vc
.lock
, flags
);
200 vchan_dma_desc_free_list(&ch
->vc
, &head
);
205 static int moxart_slave_config(struct dma_chan
*chan
,
206 struct dma_slave_config
*cfg
)
208 struct moxart_chan
*ch
= to_moxart_dma_chan(chan
);
213 ctrl
= readl(ch
->base
+ REG_OFF_CTRL
);
214 ctrl
|= APB_DMA_BURST_MODE
;
215 ctrl
&= ~(APB_DMA_DEST_MASK
| APB_DMA_SOURCE_MASK
);
216 ctrl
&= ~(APB_DMA_DEST_REQ_NO_MASK
| APB_DMA_SOURCE_REQ_NO_MASK
);
218 switch (ch
->cfg
.src_addr_width
) {
219 case DMA_SLAVE_BUSWIDTH_1_BYTE
:
220 ctrl
|= APB_DMA_DATA_WIDTH_1
;
221 if (ch
->cfg
.direction
!= DMA_MEM_TO_DEV
)
222 ctrl
|= APB_DMA_DEST_INC_1_4
;
224 ctrl
|= APB_DMA_SOURCE_INC_1_4
;
226 case DMA_SLAVE_BUSWIDTH_2_BYTES
:
227 ctrl
|= APB_DMA_DATA_WIDTH_2
;
228 if (ch
->cfg
.direction
!= DMA_MEM_TO_DEV
)
229 ctrl
|= APB_DMA_DEST_INC_2_8
;
231 ctrl
|= APB_DMA_SOURCE_INC_2_8
;
233 case DMA_SLAVE_BUSWIDTH_4_BYTES
:
234 ctrl
&= ~APB_DMA_DATA_WIDTH
;
235 if (ch
->cfg
.direction
!= DMA_MEM_TO_DEV
)
236 ctrl
|= APB_DMA_DEST_INC_4_16
;
238 ctrl
|= APB_DMA_SOURCE_INC_4_16
;
244 if (ch
->cfg
.direction
== DMA_MEM_TO_DEV
) {
245 ctrl
&= ~APB_DMA_DEST_SELECT
;
246 ctrl
|= APB_DMA_SOURCE_SELECT
;
247 ctrl
|= (ch
->line_reqno
<< 16 &
248 APB_DMA_DEST_REQ_NO_MASK
);
250 ctrl
|= APB_DMA_DEST_SELECT
;
251 ctrl
&= ~APB_DMA_SOURCE_SELECT
;
252 ctrl
|= (ch
->line_reqno
<< 24 &
253 APB_DMA_SOURCE_REQ_NO_MASK
);
256 writel(ctrl
, ch
->base
+ REG_OFF_CTRL
);
261 static struct dma_async_tx_descriptor
*moxart_prep_slave_sg(
262 struct dma_chan
*chan
, struct scatterlist
*sgl
,
263 unsigned int sg_len
, enum dma_transfer_direction dir
,
264 unsigned long tx_flags
, void *context
)
266 struct moxart_chan
*ch
= to_moxart_dma_chan(chan
);
267 struct moxart_desc
*d
;
268 enum dma_slave_buswidth dev_width
;
270 struct scatterlist
*sgent
;
274 if (!is_slave_direction(dir
)) {
275 dev_err(chan2dev(chan
), "%s: invalid DMA direction\n",
280 if (dir
== DMA_DEV_TO_MEM
) {
281 dev_addr
= ch
->cfg
.src_addr
;
282 dev_width
= ch
->cfg
.src_addr_width
;
284 dev_addr
= ch
->cfg
.dst_addr
;
285 dev_width
= ch
->cfg
.dst_addr_width
;
289 case DMA_SLAVE_BUSWIDTH_1_BYTE
:
290 es
= MOXART_DMA_DATA_TYPE_S8
;
292 case DMA_SLAVE_BUSWIDTH_2_BYTES
:
293 es
= MOXART_DMA_DATA_TYPE_S16
;
295 case DMA_SLAVE_BUSWIDTH_4_BYTES
:
296 es
= MOXART_DMA_DATA_TYPE_S32
;
299 dev_err(chan2dev(chan
), "%s: unsupported data width (%u)\n",
300 __func__
, dev_width
);
304 d
= kzalloc(struct_size(d
, sg
, sg_len
), GFP_ATOMIC
);
310 d
->dev_addr
= dev_addr
;
313 for_each_sg(sgl
, sgent
, sg_len
, i
) {
314 d
->sg
[i
].addr
= sg_dma_address(sgent
);
315 d
->sg
[i
].len
= sg_dma_len(sgent
);
320 return vchan_tx_prep(&ch
->vc
, &d
->vd
, tx_flags
);
323 static struct dma_chan
*moxart_of_xlate(struct of_phandle_args
*dma_spec
,
324 struct of_dma
*ofdma
)
326 struct moxart_dmadev
*mdc
= ofdma
->of_dma_data
;
327 struct dma_chan
*chan
;
328 struct moxart_chan
*ch
;
330 chan
= dma_get_any_slave_channel(&mdc
->dma_slave
);
334 ch
= to_moxart_dma_chan(chan
);
335 ch
->line_reqno
= dma_spec
->args
[0];
340 static int moxart_alloc_chan_resources(struct dma_chan
*chan
)
342 struct moxart_chan
*ch
= to_moxart_dma_chan(chan
);
344 dev_dbg(chan2dev(chan
), "%s: allocating channel #%u\n",
345 __func__
, ch
->ch_num
);
351 static void moxart_free_chan_resources(struct dma_chan
*chan
)
353 struct moxart_chan
*ch
= to_moxart_dma_chan(chan
);
355 vchan_free_chan_resources(&ch
->vc
);
357 dev_dbg(chan2dev(chan
), "%s: freeing channel #%u\n",
358 __func__
, ch
->ch_num
);
362 static void moxart_dma_set_params(struct moxart_chan
*ch
, dma_addr_t src_addr
,
365 writel(src_addr
, ch
->base
+ REG_OFF_ADDRESS_SOURCE
);
366 writel(dst_addr
, ch
->base
+ REG_OFF_ADDRESS_DEST
);
369 static void moxart_set_transfer_params(struct moxart_chan
*ch
, unsigned int len
)
371 struct moxart_desc
*d
= ch
->desc
;
372 unsigned int sglen_div
= es_bytes
[d
->es
];
374 d
->dma_cycles
= len
>> sglen_div
;
377 * There are 4 cycles on 64 bytes copied, i.e. one cycle copies 16
378 * bytes ( when width is APB_DMAB_DATA_WIDTH_4 ).
380 writel(d
->dma_cycles
, ch
->base
+ REG_OFF_CYCLES
);
382 dev_dbg(chan2dev(&ch
->vc
.chan
), "%s: set %u DMA cycles (len=%u)\n",
383 __func__
, d
->dma_cycles
, len
);
386 static void moxart_start_dma(struct moxart_chan
*ch
)
390 ctrl
= readl(ch
->base
+ REG_OFF_CTRL
);
391 ctrl
|= (APB_DMA_ENABLE
| APB_DMA_FIN_INT_EN
| APB_DMA_ERR_INT_EN
);
392 writel(ctrl
, ch
->base
+ REG_OFF_CTRL
);
395 static void moxart_dma_start_sg(struct moxart_chan
*ch
, unsigned int idx
)
397 struct moxart_desc
*d
= ch
->desc
;
398 struct moxart_sg
*sg
= ch
->desc
->sg
+ idx
;
400 if (ch
->desc
->dma_dir
== DMA_MEM_TO_DEV
)
401 moxart_dma_set_params(ch
, sg
->addr
, d
->dev_addr
);
402 else if (ch
->desc
->dma_dir
== DMA_DEV_TO_MEM
)
403 moxart_dma_set_params(ch
, d
->dev_addr
, sg
->addr
);
405 moxart_set_transfer_params(ch
, sg
->len
);
407 moxart_start_dma(ch
);
410 static void moxart_dma_start_desc(struct dma_chan
*chan
)
412 struct moxart_chan
*ch
= to_moxart_dma_chan(chan
);
413 struct virt_dma_desc
*vd
;
415 vd
= vchan_next_desc(&ch
->vc
);
424 ch
->desc
= to_moxart_dma_desc(&vd
->tx
);
427 moxart_dma_start_sg(ch
, 0);
430 static void moxart_issue_pending(struct dma_chan
*chan
)
432 struct moxart_chan
*ch
= to_moxart_dma_chan(chan
);
435 spin_lock_irqsave(&ch
->vc
.lock
, flags
);
436 if (vchan_issue_pending(&ch
->vc
) && !ch
->desc
)
437 moxart_dma_start_desc(chan
);
438 spin_unlock_irqrestore(&ch
->vc
.lock
, flags
);
441 static size_t moxart_dma_desc_size(struct moxart_desc
*d
,
442 unsigned int completed_sgs
)
447 for (size
= i
= completed_sgs
; i
< d
->sglen
; i
++)
448 size
+= d
->sg
[i
].len
;
453 static size_t moxart_dma_desc_size_in_flight(struct moxart_chan
*ch
)
456 unsigned int completed_cycles
, cycles
;
458 size
= moxart_dma_desc_size(ch
->desc
, ch
->sgidx
);
459 cycles
= readl(ch
->base
+ REG_OFF_CYCLES
);
460 completed_cycles
= (ch
->desc
->dma_cycles
- cycles
);
461 size
-= completed_cycles
<< es_bytes
[ch
->desc
->es
];
463 dev_dbg(chan2dev(&ch
->vc
.chan
), "%s: size=%zu\n", __func__
, size
);
468 static enum dma_status
moxart_tx_status(struct dma_chan
*chan
,
470 struct dma_tx_state
*txstate
)
472 struct moxart_chan
*ch
= to_moxart_dma_chan(chan
);
473 struct virt_dma_desc
*vd
;
474 struct moxart_desc
*d
;
479 * dma_cookie_status() assigns initial residue value.
481 ret
= dma_cookie_status(chan
, cookie
, txstate
);
483 spin_lock_irqsave(&ch
->vc
.lock
, flags
);
484 vd
= vchan_find_desc(&ch
->vc
, cookie
);
486 d
= to_moxart_dma_desc(&vd
->tx
);
487 txstate
->residue
= moxart_dma_desc_size(d
, 0);
488 } else if (ch
->desc
&& ch
->desc
->vd
.tx
.cookie
== cookie
) {
489 txstate
->residue
= moxart_dma_desc_size_in_flight(ch
);
491 spin_unlock_irqrestore(&ch
->vc
.lock
, flags
);
499 static void moxart_dma_init(struct dma_device
*dma
, struct device
*dev
)
501 dma
->device_prep_slave_sg
= moxart_prep_slave_sg
;
502 dma
->device_alloc_chan_resources
= moxart_alloc_chan_resources
;
503 dma
->device_free_chan_resources
= moxart_free_chan_resources
;
504 dma
->device_issue_pending
= moxart_issue_pending
;
505 dma
->device_tx_status
= moxart_tx_status
;
506 dma
->device_config
= moxart_slave_config
;
507 dma
->device_terminate_all
= moxart_terminate_all
;
510 INIT_LIST_HEAD(&dma
->channels
);
513 static irqreturn_t
moxart_dma_interrupt(int irq
, void *devid
)
515 struct moxart_dmadev
*mc
= devid
;
516 struct moxart_chan
*ch
= &mc
->slave_chans
[0];
520 dev_dbg(chan2dev(&ch
->vc
.chan
), "%s\n", __func__
);
522 for (i
= 0; i
< APB_DMA_MAX_CHANNEL
; i
++, ch
++) {
526 ctrl
= readl(ch
->base
+ REG_OFF_CTRL
);
528 dev_dbg(chan2dev(&ch
->vc
.chan
), "%s: ch=%p ch->base=%p ctrl=%x\n",
529 __func__
, ch
, ch
->base
, ctrl
);
531 if (ctrl
& APB_DMA_FIN_INT_STS
) {
532 ctrl
&= ~APB_DMA_FIN_INT_STS
;
534 spin_lock(&ch
->vc
.lock
);
535 if (++ch
->sgidx
< ch
->desc
->sglen
) {
536 moxart_dma_start_sg(ch
, ch
->sgidx
);
538 vchan_cookie_complete(&ch
->desc
->vd
);
539 moxart_dma_start_desc(&ch
->vc
.chan
);
541 spin_unlock(&ch
->vc
.lock
);
545 if (ctrl
& APB_DMA_ERR_INT_STS
) {
546 ctrl
&= ~APB_DMA_ERR_INT_STS
;
550 writel(ctrl
, ch
->base
+ REG_OFF_CTRL
);
556 static int moxart_probe(struct platform_device
*pdev
)
558 struct device
*dev
= &pdev
->dev
;
559 struct device_node
*node
= dev
->of_node
;
560 void __iomem
*dma_base_addr
;
563 struct moxart_chan
*ch
;
564 struct moxart_dmadev
*mdc
;
566 mdc
= devm_kzalloc(dev
, sizeof(*mdc
), GFP_KERNEL
);
570 irq
= irq_of_parse_and_map(node
, 0);
572 dev_err(dev
, "no IRQ resource\n");
576 dma_base_addr
= devm_platform_ioremap_resource(pdev
, 0);
577 if (IS_ERR(dma_base_addr
))
578 return PTR_ERR(dma_base_addr
);
580 dma_cap_zero(mdc
->dma_slave
.cap_mask
);
581 dma_cap_set(DMA_SLAVE
, mdc
->dma_slave
.cap_mask
);
582 dma_cap_set(DMA_PRIVATE
, mdc
->dma_slave
.cap_mask
);
584 moxart_dma_init(&mdc
->dma_slave
, dev
);
586 ch
= &mdc
->slave_chans
[0];
587 for (i
= 0; i
< APB_DMA_MAX_CHANNEL
; i
++, ch
++) {
589 ch
->base
= dma_base_addr
+ i
* REG_OFF_CHAN_SIZE
;
592 ch
->vc
.desc_free
= moxart_dma_desc_free
;
593 vchan_init(&ch
->vc
, &mdc
->dma_slave
);
595 dev_dbg(dev
, "%s: chs[%d]: ch->ch_num=%u ch->base=%p\n",
596 __func__
, i
, ch
->ch_num
, ch
->base
);
599 platform_set_drvdata(pdev
, mdc
);
601 ret
= devm_request_irq(dev
, irq
, moxart_dma_interrupt
, 0,
602 "moxart-dma-engine", mdc
);
604 dev_err(dev
, "devm_request_irq failed\n");
609 ret
= dma_async_device_register(&mdc
->dma_slave
);
611 dev_err(dev
, "dma_async_device_register failed\n");
615 ret
= of_dma_controller_register(node
, moxart_of_xlate
, mdc
);
617 dev_err(dev
, "of_dma_controller_register failed\n");
618 dma_async_device_unregister(&mdc
->dma_slave
);
622 dev_dbg(dev
, "%s: IRQ=%u\n", __func__
, irq
);
627 static void moxart_remove(struct platform_device
*pdev
)
629 struct moxart_dmadev
*m
= platform_get_drvdata(pdev
);
631 devm_free_irq(&pdev
->dev
, m
->irq
, m
);
633 dma_async_device_unregister(&m
->dma_slave
);
635 if (pdev
->dev
.of_node
)
636 of_dma_controller_free(pdev
->dev
.of_node
);
639 static const struct of_device_id moxart_dma_match
[] = {
640 { .compatible
= "moxa,moxart-dma" },
643 MODULE_DEVICE_TABLE(of
, moxart_dma_match
);
645 static struct platform_driver moxart_driver
= {
646 .probe
= moxart_probe
,
647 .remove
= moxart_remove
,
649 .name
= "moxart-dma-engine",
650 .of_match_table
= moxart_dma_match
,
654 static int moxart_init(void)
656 return platform_driver_register(&moxart_driver
);
658 subsys_initcall(moxart_init
);
660 static void __exit
moxart_exit(void)
662 platform_driver_unregister(&moxart_driver
);
664 module_exit(moxart_exit
);
666 MODULE_AUTHOR("Jonas Jensen <jonas.jensen@gmail.com>");
667 MODULE_DESCRIPTION("MOXART DMA engine driver");
668 MODULE_LICENSE("GPL v2");