1 // SPDX-License-Identifier: GPL-2.0
3 * MUSB OTG driver - support for Mentor's DMA controller
5 * Copyright 2005 Mentor Graphics Corporation
6 * Copyright (C) 2005-2007 by Texas Instruments
8 #include <linux/device.h>
9 #include <linux/interrupt.h>
10 #include <linux/platform_device.h>
11 #include <linux/slab.h>
12 #include "musb_core.h"
15 #define MUSB_HSDMA_CHANNEL_OFFSET(_bchannel, _offset) \
16 (MUSB_HSDMA_BASE + (_bchannel << 4) + _offset)
18 #define musb_read_hsdma_addr(mbase, bchannel) \
20 MUSB_HSDMA_CHANNEL_OFFSET(bchannel, MUSB_HSDMA_ADDRESS))
22 #define musb_write_hsdma_addr(mbase, bchannel, addr) \
24 MUSB_HSDMA_CHANNEL_OFFSET(bchannel, MUSB_HSDMA_ADDRESS), \
27 #define musb_read_hsdma_count(mbase, bchannel) \
29 MUSB_HSDMA_CHANNEL_OFFSET(bchannel, MUSB_HSDMA_COUNT))
31 #define musb_write_hsdma_count(mbase, bchannel, len) \
33 MUSB_HSDMA_CHANNEL_OFFSET(bchannel, MUSB_HSDMA_COUNT), \
35 /* control register (16-bit): */
36 #define MUSB_HSDMA_ENABLE_SHIFT 0
37 #define MUSB_HSDMA_TRANSMIT_SHIFT 1
38 #define MUSB_HSDMA_MODE1_SHIFT 2
39 #define MUSB_HSDMA_IRQENABLE_SHIFT 3
40 #define MUSB_HSDMA_ENDPOINT_SHIFT 4
41 #define MUSB_HSDMA_BUSERROR_SHIFT 8
42 #define MUSB_HSDMA_BURSTMODE_SHIFT 9
43 #define MUSB_HSDMA_BURSTMODE (3 << MUSB_HSDMA_BURSTMODE_SHIFT)
44 #define MUSB_HSDMA_BURSTMODE_UNSPEC 0
45 #define MUSB_HSDMA_BURSTMODE_INCR4 1
46 #define MUSB_HSDMA_BURSTMODE_INCR8 2
47 #define MUSB_HSDMA_BURSTMODE_INCR16 3
49 #define MUSB_HSDMA_CHANNELS 8
51 struct musb_dma_controller
;
53 struct musb_dma_channel
{
54 struct dma_channel channel
;
55 struct musb_dma_controller
*controller
;
64 struct musb_dma_controller
{
65 struct dma_controller controller
;
66 struct musb_dma_channel channel
[MUSB_HSDMA_CHANNELS
];
74 static void dma_channel_release(struct dma_channel
*channel
);
76 static void dma_controller_stop(struct musb_dma_controller
*controller
)
78 struct musb
*musb
= controller
->private_data
;
79 struct dma_channel
*channel
;
82 if (controller
->used_channels
!= 0) {
83 dev_err(musb
->controller
,
84 "Stopping DMA controller while channel active\n");
86 for (bit
= 0; bit
< MUSB_HSDMA_CHANNELS
; bit
++) {
87 if (controller
->used_channels
& (1 << bit
)) {
88 channel
= &controller
->channel
[bit
].channel
;
89 dma_channel_release(channel
);
91 if (!controller
->used_channels
)
98 static struct dma_channel
*dma_channel_allocate(struct dma_controller
*c
,
99 struct musb_hw_ep
*hw_ep
, u8 transmit
)
101 struct musb_dma_controller
*controller
= container_of(c
,
102 struct musb_dma_controller
, controller
);
103 struct musb_dma_channel
*musb_channel
= NULL
;
104 struct dma_channel
*channel
= NULL
;
107 for (bit
= 0; bit
< MUSB_HSDMA_CHANNELS
; bit
++) {
108 if (!(controller
->used_channels
& (1 << bit
))) {
109 controller
->used_channels
|= (1 << bit
);
110 musb_channel
= &(controller
->channel
[bit
]);
111 musb_channel
->controller
= controller
;
112 musb_channel
->idx
= bit
;
113 musb_channel
->epnum
= hw_ep
->epnum
;
114 musb_channel
->transmit
= transmit
;
115 channel
= &(musb_channel
->channel
);
116 channel
->private_data
= musb_channel
;
117 channel
->status
= MUSB_DMA_STATUS_FREE
;
118 channel
->max_len
= 0x100000;
119 /* Tx => mode 1; Rx => mode 0 */
120 channel
->desired_mode
= transmit
;
121 channel
->actual_len
= 0;
129 static void dma_channel_release(struct dma_channel
*channel
)
131 struct musb_dma_channel
*musb_channel
= channel
->private_data
;
133 channel
->actual_len
= 0;
134 musb_channel
->start_addr
= 0;
135 musb_channel
->len
= 0;
137 musb_channel
->controller
->used_channels
&=
138 ~(1 << musb_channel
->idx
);
140 channel
->status
= MUSB_DMA_STATUS_UNKNOWN
;
143 static void configure_channel(struct dma_channel
*channel
,
144 u16 packet_sz
, u8 mode
,
145 dma_addr_t dma_addr
, u32 len
)
147 struct musb_dma_channel
*musb_channel
= channel
->private_data
;
148 struct musb_dma_controller
*controller
= musb_channel
->controller
;
149 struct musb
*musb
= controller
->private_data
;
150 void __iomem
*mbase
= controller
->base
;
151 u8 bchannel
= musb_channel
->idx
;
154 musb_dbg(musb
, "%p, pkt_sz %d, addr %pad, len %d, mode %d",
155 channel
, packet_sz
, &dma_addr
, len
, mode
);
158 csr
|= 1 << MUSB_HSDMA_MODE1_SHIFT
;
159 BUG_ON(len
< packet_sz
);
161 csr
|= MUSB_HSDMA_BURSTMODE_INCR16
162 << MUSB_HSDMA_BURSTMODE_SHIFT
;
164 csr
|= (musb_channel
->epnum
<< MUSB_HSDMA_ENDPOINT_SHIFT
)
165 | (1 << MUSB_HSDMA_ENABLE_SHIFT
)
166 | (1 << MUSB_HSDMA_IRQENABLE_SHIFT
)
167 | (musb_channel
->transmit
168 ? (1 << MUSB_HSDMA_TRANSMIT_SHIFT
)
172 musb_write_hsdma_addr(mbase
, bchannel
, dma_addr
);
173 musb_write_hsdma_count(mbase
, bchannel
, len
);
175 /* control (this should start things) */
177 MUSB_HSDMA_CHANNEL_OFFSET(bchannel
, MUSB_HSDMA_CONTROL
),
181 static int dma_channel_program(struct dma_channel
*channel
,
182 u16 packet_sz
, u8 mode
,
183 dma_addr_t dma_addr
, u32 len
)
185 struct musb_dma_channel
*musb_channel
= channel
->private_data
;
186 struct musb_dma_controller
*controller
= musb_channel
->controller
;
187 struct musb
*musb
= controller
->private_data
;
189 musb_dbg(musb
, "ep%d-%s pkt_sz %d, dma_addr %pad length %d, mode %d",
191 musb_channel
->transmit
? "Tx" : "Rx",
192 packet_sz
, &dma_addr
, len
, mode
);
194 BUG_ON(channel
->status
== MUSB_DMA_STATUS_UNKNOWN
||
195 channel
->status
== MUSB_DMA_STATUS_BUSY
);
198 * The DMA engine in RTL1.8 and above cannot handle
199 * DMA addresses that are not aligned to a 4 byte boundary.
200 * It ends up masking the last two bits of the address
201 * programmed in DMA_ADDR.
203 * Fail such DMA transfers, so that the backup PIO mode
204 * can carry out the transfer
206 if ((musb
->hwvers
>= MUSB_HWVERS_1800
) && (dma_addr
% 4))
209 channel
->actual_len
= 0;
210 musb_channel
->start_addr
= dma_addr
;
211 musb_channel
->len
= len
;
212 musb_channel
->max_packet_sz
= packet_sz
;
213 channel
->status
= MUSB_DMA_STATUS_BUSY
;
215 configure_channel(channel
, packet_sz
, mode
, dma_addr
, len
);
220 static int dma_channel_abort(struct dma_channel
*channel
)
222 struct musb_dma_channel
*musb_channel
= channel
->private_data
;
223 void __iomem
*mbase
= musb_channel
->controller
->base
;
224 struct musb
*musb
= musb_channel
->controller
->private_data
;
226 u8 bchannel
= musb_channel
->idx
;
230 if (channel
->status
== MUSB_DMA_STATUS_BUSY
) {
231 if (musb_channel
->transmit
) {
232 offset
= musb
->io
.ep_offset(musb_channel
->epnum
,
236 * The programming guide says that we must clear
237 * the DMAENAB bit before the DMAMODE bit...
239 csr
= musb_readw(mbase
, offset
);
240 csr
&= ~(MUSB_TXCSR_AUTOSET
| MUSB_TXCSR_DMAENAB
);
241 musb_writew(mbase
, offset
, csr
);
242 csr
&= ~MUSB_TXCSR_DMAMODE
;
243 musb_writew(mbase
, offset
, csr
);
245 offset
= musb
->io
.ep_offset(musb_channel
->epnum
,
248 csr
= musb_readw(mbase
, offset
);
249 csr
&= ~(MUSB_RXCSR_AUTOCLEAR
|
252 musb_writew(mbase
, offset
, csr
);
256 MUSB_HSDMA_CHANNEL_OFFSET(bchannel
, MUSB_HSDMA_CONTROL
),
258 musb_write_hsdma_addr(mbase
, bchannel
, 0);
259 musb_write_hsdma_count(mbase
, bchannel
, 0);
260 channel
->status
= MUSB_DMA_STATUS_FREE
;
266 irqreturn_t
dma_controller_irq(int irq
, void *private_data
)
268 struct musb_dma_controller
*controller
= private_data
;
269 struct musb
*musb
= controller
->private_data
;
270 struct musb_dma_channel
*musb_channel
;
271 struct dma_channel
*channel
;
273 void __iomem
*mbase
= controller
->base
;
275 irqreturn_t retval
= IRQ_NONE
;
285 spin_lock_irqsave(&musb
->lock
, flags
);
287 int_hsdma
= musb_clearb(mbase
, MUSB_HSDMA_INTR
);
290 musb_dbg(musb
, "spurious DMA irq");
292 for (bchannel
= 0; bchannel
< MUSB_HSDMA_CHANNELS
; bchannel
++) {
293 musb_channel
= (struct musb_dma_channel
*)
294 &(controller
->channel
[bchannel
]);
295 channel
= &musb_channel
->channel
;
296 if (channel
->status
== MUSB_DMA_STATUS_BUSY
) {
297 count
= musb_read_hsdma_count(mbase
, bchannel
);
300 int_hsdma
|= (1 << bchannel
);
304 musb_dbg(musb
, "int_hsdma = 0x%x", int_hsdma
);
310 for (bchannel
= 0; bchannel
< MUSB_HSDMA_CHANNELS
; bchannel
++) {
311 if (int_hsdma
& (1 << bchannel
)) {
312 musb_channel
= (struct musb_dma_channel
*)
313 &(controller
->channel
[bchannel
]);
314 channel
= &musb_channel
->channel
;
316 csr
= musb_readw(mbase
,
317 MUSB_HSDMA_CHANNEL_OFFSET(bchannel
,
318 MUSB_HSDMA_CONTROL
));
320 if (csr
& (1 << MUSB_HSDMA_BUSERROR_SHIFT
)) {
321 musb_channel
->channel
.status
=
322 MUSB_DMA_STATUS_BUS_ABORT
;
326 addr
= musb_read_hsdma_addr(mbase
,
328 channel
->actual_len
= addr
329 - musb_channel
->start_addr
;
331 musb_dbg(musb
, "ch %p, 0x%x -> 0x%x (%zu / %d) %s",
332 channel
, musb_channel
->start_addr
,
333 addr
, channel
->actual_len
,
336 < musb_channel
->len
) ?
337 "=> reconfig 0" : "=> complete");
339 devctl
= musb_readb(mbase
, MUSB_DEVCTL
);
341 channel
->status
= MUSB_DMA_STATUS_FREE
;
344 if (musb_channel
->transmit
&&
345 (!channel
->desired_mode
||
346 (channel
->actual_len
%
347 musb_channel
->max_packet_sz
))) {
348 u8 epnum
= musb_channel
->epnum
;
349 int offset
= musb
->io
.ep_offset(epnum
,
354 * The programming guide says that we
355 * must clear DMAENAB before DMAMODE.
357 musb_ep_select(mbase
, epnum
);
358 txcsr
= musb_readw(mbase
, offset
);
359 if (channel
->desired_mode
== 1) {
360 txcsr
&= ~(MUSB_TXCSR_DMAENAB
361 | MUSB_TXCSR_AUTOSET
);
362 musb_writew(mbase
, offset
, txcsr
);
363 /* Send out the packet */
364 txcsr
&= ~MUSB_TXCSR_DMAMODE
;
365 txcsr
|= MUSB_TXCSR_DMAENAB
;
367 txcsr
|= MUSB_TXCSR_TXPKTRDY
;
368 musb_writew(mbase
, offset
, txcsr
);
370 musb_dma_completion(musb
, musb_channel
->epnum
,
371 musb_channel
->transmit
);
376 retval
= IRQ_HANDLED
;
378 spin_unlock_irqrestore(&musb
->lock
, flags
);
381 EXPORT_SYMBOL_GPL(dma_controller_irq
);
383 void musbhs_dma_controller_destroy(struct dma_controller
*c
)
385 struct musb_dma_controller
*controller
= container_of(c
,
386 struct musb_dma_controller
, controller
);
388 dma_controller_stop(controller
);
391 free_irq(controller
->irq
, c
);
395 EXPORT_SYMBOL_GPL(musbhs_dma_controller_destroy
);
397 static struct musb_dma_controller
*
398 dma_controller_alloc(struct musb
*musb
, void __iomem
*base
)
400 struct musb_dma_controller
*controller
;
402 controller
= kzalloc(sizeof(*controller
), GFP_KERNEL
);
406 controller
->channel_count
= MUSB_HSDMA_CHANNELS
;
407 controller
->private_data
= musb
;
408 controller
->base
= base
;
410 controller
->controller
.channel_alloc
= dma_channel_allocate
;
411 controller
->controller
.channel_release
= dma_channel_release
;
412 controller
->controller
.channel_program
= dma_channel_program
;
413 controller
->controller
.channel_abort
= dma_channel_abort
;
417 struct dma_controller
*
418 musbhs_dma_controller_create(struct musb
*musb
, void __iomem
*base
)
420 struct musb_dma_controller
*controller
;
421 struct device
*dev
= musb
->controller
;
422 struct platform_device
*pdev
= to_platform_device(dev
);
423 int irq
= platform_get_irq_byname(pdev
, "dma");
426 dev_err(dev
, "No DMA interrupt line!\n");
430 controller
= dma_controller_alloc(musb
, base
);
434 if (request_irq(irq
, dma_controller_irq
, 0,
435 dev_name(musb
->controller
), controller
)) {
436 dev_err(dev
, "request_irq %d failed!\n", irq
);
437 musb_dma_controller_destroy(&controller
->controller
);
442 controller
->irq
= irq
;
444 return &controller
->controller
;
446 EXPORT_SYMBOL_GPL(musbhs_dma_controller_create
);
448 struct dma_controller
*
449 musbhs_dma_controller_create_noirq(struct musb
*musb
, void __iomem
*base
)
451 struct musb_dma_controller
*controller
;
453 controller
= dma_controller_alloc(musb
, base
);
457 return &controller
->controller
;
459 EXPORT_SYMBOL_GPL(musbhs_dma_controller_create_noirq
);