[PATCH] W1: possible cleanups
[linux-2.6/verdex.git] / include / asm-ppc / ppc4xx_dma.h
blob935d1e05366b4cb7cc89e4c40d657dfec34423e7
1 /*
2 * include/asm-ppc/ppc4xx_dma.h
4 * IBM PPC4xx DMA engine library
6 * Copyright 2000-2004 MontaVista Software Inc.
8 * Cleaned up a bit more, Matt Porter <mporter@kernel.crashing.org>
10 * Original code by Armin Kuster <akuster@mvista.com>
11 * and Pete Popov <ppopov@mvista.com>
13 * This program is free software; you can redistribute it and/or modify it
14 * under the terms of the GNU General Public License as published by the
15 * Free Software Foundation; either version 2 of the License, or (at your
16 * option) any later version.
18 * You should have received a copy of the GNU General Public License along
19 * with this program; if not, write to the Free Software Foundation, Inc.,
20 * 675 Mass Ave, Cambridge, MA 02139, USA.
23 #ifdef __KERNEL__
24 #ifndef __ASMPPC_PPC4xx_DMA_H
25 #define __ASMPPC_PPC4xx_DMA_H
27 #include <linux/types.h>
28 #include <asm/mmu.h>
29 #include <asm/ibm4xx.h>
31 #undef DEBUG_4xxDMA
33 #define MAX_PPC4xx_DMA_CHANNELS 4
36 * Function return status codes
37 * These values are used to indicate whether or not the function
38 * call was successful, or a bad/invalid parameter was passed.
40 #define DMA_STATUS_GOOD 0
41 #define DMA_STATUS_BAD_CHANNEL 1
42 #define DMA_STATUS_BAD_HANDLE 2
43 #define DMA_STATUS_BAD_MODE 3
44 #define DMA_STATUS_NULL_POINTER 4
45 #define DMA_STATUS_OUT_OF_MEMORY 5
46 #define DMA_STATUS_SGL_LIST_EMPTY 6
47 #define DMA_STATUS_GENERAL_ERROR 7
48 #define DMA_STATUS_CHANNEL_NOTFREE 8
50 #define DMA_CHANNEL_BUSY 0x80000000
53 * These indicate status as returned from the DMA Status Register.
55 #define DMA_STATUS_NO_ERROR 0
56 #define DMA_STATUS_CS 1 /* Count Status */
57 #define DMA_STATUS_TS 2 /* Transfer Status */
58 #define DMA_STATUS_DMA_ERROR 3 /* DMA Error Occurred */
59 #define DMA_STATUS_DMA_BUSY 4 /* The channel is busy */
63 * DMA Channel Control Registers
66 #ifdef CONFIG_44x
67 #define PPC4xx_DMA_64BIT
68 #define DMA_CR_OFFSET 1
69 #else
70 #define DMA_CR_OFFSET 0
71 #endif
73 #define DMA_CE_ENABLE (1<<31) /* DMA Channel Enable */
74 #define SET_DMA_CE_ENABLE(x) (((x)&0x1)<<31)
75 #define GET_DMA_CE_ENABLE(x) (((x)&DMA_CE_ENABLE)>>31)
77 #define DMA_CIE_ENABLE (1<<30) /* DMA Channel Interrupt Enable */
78 #define SET_DMA_CIE_ENABLE(x) (((x)&0x1)<<30)
79 #define GET_DMA_CIE_ENABLE(x) (((x)&DMA_CIE_ENABLE)>>30)
81 #define DMA_TD (1<<29)
82 #define SET_DMA_TD(x) (((x)&0x1)<<29)
83 #define GET_DMA_TD(x) (((x)&DMA_TD)>>29)
85 #define DMA_PL (1<<28) /* Peripheral Location */
86 #define SET_DMA_PL(x) (((x)&0x1)<<28)
87 #define GET_DMA_PL(x) (((x)&DMA_PL)>>28)
89 #define EXTERNAL_PERIPHERAL 0
90 #define INTERNAL_PERIPHERAL 1
92 #define SET_DMA_PW(x) (((x)&0x3)<<(26-DMA_CR_OFFSET)) /* Peripheral Width */
93 #define DMA_PW_MASK SET_DMA_PW(3)
94 #define PW_8 0
95 #define PW_16 1
96 #define PW_32 2
97 #define PW_64 3
98 /* FIXME: Add PW_128 support for 440GP DMA block */
99 #define GET_DMA_PW(x) (((x)&DMA_PW_MASK)>>(26-DMA_CR_OFFSET))
101 #define DMA_DAI (1<<(25-DMA_CR_OFFSET)) /* Destination Address Increment */
102 #define SET_DMA_DAI(x) (((x)&0x1)<<(25-DMA_CR_OFFSET))
104 #define DMA_SAI (1<<(24-DMA_CR_OFFSET)) /* Source Address Increment */
105 #define SET_DMA_SAI(x) (((x)&0x1)<<(24-DMA_CR_OFFSET))
107 #define DMA_BEN (1<<(23-DMA_CR_OFFSET)) /* Buffer Enable */
108 #define SET_DMA_BEN(x) (((x)&0x1)<<(23-DMA_CR_OFFSET))
110 #define SET_DMA_TM(x) (((x)&0x3)<<(21-DMA_CR_OFFSET)) /* Transfer Mode */
111 #define DMA_TM_MASK SET_DMA_TM(3)
112 #define TM_PERIPHERAL 0 /* Peripheral */
113 #define TM_RESERVED 1 /* Reserved */
114 #define TM_S_MM 2 /* Memory to Memory */
115 #define TM_D_MM 3 /* Device Paced Memory to Memory */
116 #define GET_DMA_TM(x) (((x)&DMA_TM_MASK)>>(21-DMA_CR_OFFSET))
118 #define SET_DMA_PSC(x) (((x)&0x3)<<(19-DMA_CR_OFFSET)) /* Peripheral Setup Cycles */
119 #define DMA_PSC_MASK SET_DMA_PSC(3)
120 #define GET_DMA_PSC(x) (((x)&DMA_PSC_MASK)>>(19-DMA_CR_OFFSET))
122 #define SET_DMA_PWC(x) (((x)&0x3F)<<(13-DMA_CR_OFFSET)) /* Peripheral Wait Cycles */
123 #define DMA_PWC_MASK SET_DMA_PWC(0x3F)
124 #define GET_DMA_PWC(x) (((x)&DMA_PWC_MASK)>>(13-DMA_CR_OFFSET))
126 #define SET_DMA_PHC(x) (((x)&0x7)<<(10-DMA_CR_OFFSET)) /* Peripheral Hold Cycles */
127 #define DMA_PHC_MASK SET_DMA_PHC(0x7)
128 #define GET_DMA_PHC(x) (((x)&DMA_PHC_MASK)>>(10-DMA_CR_OFFSET))
130 #define DMA_ETD_OUTPUT (1<<(9-DMA_CR_OFFSET)) /* EOT pin is a TC output */
131 #define SET_DMA_ETD(x) (((x)&0x1)<<(9-DMA_CR_OFFSET))
133 #define DMA_TCE_ENABLE (1<<(8-DMA_CR_OFFSET))
134 #define SET_DMA_TCE(x) (((x)&0x1)<<(8-DMA_CR_OFFSET))
136 #define DMA_DEC (1<<(2)) /* Address Decrement */
137 #define SET_DMA_DEC(x) (((x)&0x1)<<2)
138 #define GET_DMA_DEC(x) (((x)&DMA_DEC)>>2)
142 * Transfer Modes
143 * These modes are defined in a way that makes it possible to
144 * simply "or" in the value in the control register.
147 #define DMA_MODE_MM (SET_DMA_TM(TM_S_MM)) /* memory to memory */
149 /* Device-paced memory to memory, */
150 /* device is at source address */
151 #define DMA_MODE_MM_DEVATSRC (DMA_TD | SET_DMA_TM(TM_D_MM))
153 /* Device-paced memory to memory, */
154 /* device is at destination address */
155 #define DMA_MODE_MM_DEVATDST (SET_DMA_TM(TM_D_MM))
157 /* 405gp/440gp */
158 #define SET_DMA_PREFETCH(x) (((x)&0x3)<<(4-DMA_CR_OFFSET)) /* Memory Read Prefetch */
159 #define DMA_PREFETCH_MASK SET_DMA_PREFETCH(3)
160 #define PREFETCH_1 0 /* Prefetch 1 Double Word */
161 #define PREFETCH_2 1
162 #define PREFETCH_4 2
163 #define GET_DMA_PREFETCH(x) (((x)&DMA_PREFETCH_MASK)>>(4-DMA_CR_OFFSET))
165 #define DMA_PCE (1<<(3-DMA_CR_OFFSET)) /* Parity Check Enable */
166 #define SET_DMA_PCE(x) (((x)&0x1)<<(3-DMA_CR_OFFSET))
167 #define GET_DMA_PCE(x) (((x)&DMA_PCE)>>(3-DMA_CR_OFFSET))
169 /* stb3x */
171 #define DMA_ECE_ENABLE (1<<5)
172 #define SET_DMA_ECE(x) (((x)&0x1)<<5)
173 #define GET_DMA_ECE(x) (((x)&DMA_ECE_ENABLE)>>5)
175 #define DMA_TCD_DISABLE (1<<4)
176 #define SET_DMA_TCD(x) (((x)&0x1)<<4)
177 #define GET_DMA_TCD(x) (((x)&DMA_TCD_DISABLE)>>4)
179 typedef uint32_t sgl_handle_t;
181 #ifdef CONFIG_PPC4xx_EDMA
183 #define SGL_LIST_SIZE 4096
184 #define DMA_PPC4xx_SIZE SGL_LIST_SIZE
186 #define SET_DMA_PRIORITY(x) (((x)&0x3)<<(6-DMA_CR_OFFSET)) /* DMA Channel Priority */
187 #define DMA_PRIORITY_MASK SET_DMA_PRIORITY(3)
188 #define PRIORITY_LOW 0
189 #define PRIORITY_MID_LOW 1
190 #define PRIORITY_MID_HIGH 2
191 #define PRIORITY_HIGH 3
192 #define GET_DMA_PRIORITY(x) (((x)&DMA_PRIORITY_MASK)>>(6-DMA_CR_OFFSET))
195 * DMA Polarity Configuration Register
197 #define DMAReq_ActiveLow(chan) (1<<(31-(chan*3)))
198 #define DMAAck_ActiveLow(chan) (1<<(30-(chan*3)))
199 #define EOT_ActiveLow(chan) (1<<(29-(chan*3))) /* End of Transfer */
202 * DMA Sleep Mode Register
204 #define SLEEP_MODE_ENABLE (1<<21)
207 * DMA Status Register
209 #define DMA_CS0 (1<<31) /* Terminal Count has been reached */
210 #define DMA_CS1 (1<<30)
211 #define DMA_CS2 (1<<29)
212 #define DMA_CS3 (1<<28)
214 #define DMA_TS0 (1<<27) /* End of Transfer has been requested */
215 #define DMA_TS1 (1<<26)
216 #define DMA_TS2 (1<<25)
217 #define DMA_TS3 (1<<24)
219 #define DMA_CH0_ERR (1<<23) /* DMA Chanel 0 Error */
220 #define DMA_CH1_ERR (1<<22)
221 #define DMA_CH2_ERR (1<<21)
222 #define DMA_CH3_ERR (1<<20)
224 #define DMA_IN_DMA_REQ0 (1<<19) /* Internal DMA Request is pending */
225 #define DMA_IN_DMA_REQ1 (1<<18)
226 #define DMA_IN_DMA_REQ2 (1<<17)
227 #define DMA_IN_DMA_REQ3 (1<<16)
229 #define DMA_EXT_DMA_REQ0 (1<<15) /* External DMA Request is pending */
230 #define DMA_EXT_DMA_REQ1 (1<<14)
231 #define DMA_EXT_DMA_REQ2 (1<<13)
232 #define DMA_EXT_DMA_REQ3 (1<<12)
234 #define DMA_CH0_BUSY (1<<11) /* DMA Channel 0 Busy */
235 #define DMA_CH1_BUSY (1<<10)
236 #define DMA_CH2_BUSY (1<<9)
237 #define DMA_CH3_BUSY (1<<8)
239 #define DMA_SG0 (1<<7) /* DMA Channel 0 Scatter/Gather in progress */
240 #define DMA_SG1 (1<<6)
241 #define DMA_SG2 (1<<5)
242 #define DMA_SG3 (1<<4)
244 /* DMA Channel Count Register */
245 #define DMA_CTC_BTEN (1<<23) /* Burst Enable/Disable bit */
246 #define DMA_CTC_BSIZ_MSK (3<<21) /* Mask of the Burst size bits */
247 #define DMA_CTC_BSIZ_2 (0)
248 #define DMA_CTC_BSIZ_4 (1<<21)
249 #define DMA_CTC_BSIZ_8 (2<<21)
250 #define DMA_CTC_BSIZ_16 (3<<21)
253 * DMA SG Command Register
255 #define SSG_ENABLE(chan) (1<<(31-chan)) /* Start Scatter Gather */
256 #define SSG_MASK_ENABLE(chan) (1<<(15-chan)) /* Enable writing to SSG0 bit */
259 * DMA Scatter/Gather Descriptor Bit fields
261 #define SG_LINK (1<<31) /* Link */
262 #define SG_TCI_ENABLE (1<<29) /* Enable Terminal Count Interrupt */
263 #define SG_ETI_ENABLE (1<<28) /* Enable End of Transfer Interrupt */
264 #define SG_ERI_ENABLE (1<<27) /* Enable Error Interrupt */
265 #define SG_COUNT_MASK 0xFFFF /* Count Field */
267 #define SET_DMA_CONTROL \
268 (SET_DMA_CIE_ENABLE(p_init->int_enable) | /* interrupt enable */ \
269 SET_DMA_BEN(p_init->buffer_enable) | /* buffer enable */\
270 SET_DMA_ETD(p_init->etd_output) | /* end of transfer pin */ \
271 SET_DMA_TCE(p_init->tce_enable) | /* terminal count enable */ \
272 SET_DMA_PL(p_init->pl) | /* peripheral location */ \
273 SET_DMA_DAI(p_init->dai) | /* dest addr increment */ \
274 SET_DMA_SAI(p_init->sai) | /* src addr increment */ \
275 SET_DMA_PRIORITY(p_init->cp) | /* channel priority */ \
276 SET_DMA_PW(p_init->pwidth) | /* peripheral/bus width */ \
277 SET_DMA_PSC(p_init->psc) | /* peripheral setup cycles */ \
278 SET_DMA_PWC(p_init->pwc) | /* peripheral wait cycles */ \
279 SET_DMA_PHC(p_init->phc) | /* peripheral hold cycles */ \
280 SET_DMA_PREFETCH(p_init->pf) /* read prefetch */)
282 #define GET_DMA_POLARITY(chan) (DMAReq_ActiveLow(chan) | DMAAck_ActiveLow(chan) | EOT_ActiveLow(chan))
284 #elif defined(CONFIG_STB03xxx) /* stb03xxx */
286 #define DMA_PPC4xx_SIZE 4096
289 * DMA Status Register
292 #define SET_DMA_PRIORITY(x) (((x)&0x00800001)) /* DMA Channel Priority */
293 #define DMA_PRIORITY_MASK 0x00800001
294 #define PRIORITY_LOW 0x00000000
295 #define PRIORITY_MID_LOW 0x00000001
296 #define PRIORITY_MID_HIGH 0x00800000
297 #define PRIORITY_HIGH 0x00800001
298 #define GET_DMA_PRIORITY(x) (((((x)&DMA_PRIORITY_MASK) &0x00800000) >> 22 ) | (((x)&DMA_PRIORITY_MASK) &0x00000001))
300 #define DMA_CS0 (1<<31) /* Terminal Count has been reached */
301 #define DMA_CS1 (1<<30)
302 #define DMA_CS2 (1<<29)
303 #define DMA_CS3 (1<<28)
305 #define DMA_TS0 (1<<27) /* End of Transfer has been requested */
306 #define DMA_TS1 (1<<26)
307 #define DMA_TS2 (1<<25)
308 #define DMA_TS3 (1<<24)
310 #define DMA_CH0_ERR (1<<23) /* DMA Chanel 0 Error */
311 #define DMA_CH1_ERR (1<<22)
312 #define DMA_CH2_ERR (1<<21)
313 #define DMA_CH3_ERR (1<<20)
315 #define DMA_CT0 (1<<19) /* Chained transfere */
317 #define DMA_IN_DMA_REQ0 (1<<18) /* Internal DMA Request is pending */
318 #define DMA_IN_DMA_REQ1 (1<<17)
319 #define DMA_IN_DMA_REQ2 (1<<16)
320 #define DMA_IN_DMA_REQ3 (1<<15)
322 #define DMA_EXT_DMA_REQ0 (1<<14) /* External DMA Request is pending */
323 #define DMA_EXT_DMA_REQ1 (1<<13)
324 #define DMA_EXT_DMA_REQ2 (1<<12)
325 #define DMA_EXT_DMA_REQ3 (1<<11)
327 #define DMA_CH0_BUSY (1<<10) /* DMA Channel 0 Busy */
328 #define DMA_CH1_BUSY (1<<9)
329 #define DMA_CH2_BUSY (1<<8)
330 #define DMA_CH3_BUSY (1<<7)
332 #define DMA_CT1 (1<<6) /* Chained transfere */
333 #define DMA_CT2 (1<<5)
334 #define DMA_CT3 (1<<4)
336 #define DMA_CH_ENABLE (1<<7)
337 #define SET_DMA_CH(x) (((x)&0x1)<<7)
338 #define GET_DMA_CH(x) (((x)&DMA_CH_ENABLE)>>7)
340 /* STBx25xxx dma unique */
341 /* enable device port on a dma channel
342 * example ext 0 on dma 1
345 #define SSP0_RECV 15
346 #define SSP0_XMIT 14
347 #define EXT_DMA_0 12
348 #define SC1_XMIT 11
349 #define SC1_RECV 10
350 #define EXT_DMA_2 9
351 #define EXT_DMA_3 8
352 #define SERIAL2_XMIT 7
353 #define SERIAL2_RECV 6
354 #define SC0_XMIT 5
355 #define SC0_RECV 4
356 #define SERIAL1_XMIT 3
357 #define SERIAL1_RECV 2
358 #define SERIAL0_XMIT 1
359 #define SERIAL0_RECV 0
361 #define DMA_CHAN_0 1
362 #define DMA_CHAN_1 2
363 #define DMA_CHAN_2 3
364 #define DMA_CHAN_3 4
366 /* end STBx25xx */
369 * Bit 30 must be one for Redwoods, otherwise transfers may receive errors.
371 #define DMA_CR_MB0 0x2
373 #define SET_DMA_CONTROL \
374 (SET_DMA_CIE_ENABLE(p_init->int_enable) | /* interrupt enable */ \
375 SET_DMA_ETD(p_init->etd_output) | /* end of transfer pin */ \
376 SET_DMA_TCE(p_init->tce_enable) | /* terminal count enable */ \
377 SET_DMA_PL(p_init->pl) | /* peripheral location */ \
378 SET_DMA_DAI(p_init->dai) | /* dest addr increment */ \
379 SET_DMA_SAI(p_init->sai) | /* src addr increment */ \
380 SET_DMA_PRIORITY(p_init->cp) | /* channel priority */ \
381 SET_DMA_PW(p_init->pwidth) | /* peripheral/bus width */ \
382 SET_DMA_PSC(p_init->psc) | /* peripheral setup cycles */ \
383 SET_DMA_PWC(p_init->pwc) | /* peripheral wait cycles */ \
384 SET_DMA_PHC(p_init->phc) | /* peripheral hold cycles */ \
385 SET_DMA_TCD(p_init->tcd_disable) | /* TC chain mode disable */ \
386 SET_DMA_ECE(p_init->ece_enable) | /* ECE chanin mode enable */ \
387 SET_DMA_CH(p_init->ch_enable) | /* Chain enable */ \
388 DMA_CR_MB0 /* must be one */)
390 #define GET_DMA_POLARITY(chan) chan
392 #endif
394 typedef struct {
395 unsigned short in_use; /* set when channel is being used, clr when
396 * available.
399 * Valid polarity settings:
400 * DMAReq_ActiveLow(n)
401 * DMAAck_ActiveLow(n)
402 * EOT_ActiveLow(n)
404 * n is 0 to max dma chans
406 unsigned int polarity;
408 char buffer_enable; /* Boolean: buffer enable */
409 char tce_enable; /* Boolean: terminal count enable */
410 char etd_output; /* Boolean: eot pin is a tc output */
411 char pce; /* Boolean: parity check enable */
414 * Peripheral location:
415 * INTERNAL_PERIPHERAL (UART0 on the 405GP)
416 * EXTERNAL_PERIPHERAL
418 char pl; /* internal/external peripheral */
421 * Valid pwidth settings:
422 * PW_8
423 * PW_16
424 * PW_32
425 * PW_64
427 unsigned int pwidth;
429 char dai; /* Boolean: dst address increment */
430 char sai; /* Boolean: src address increment */
433 * Valid psc settings: 0-3
435 unsigned int psc; /* Peripheral Setup Cycles */
438 * Valid pwc settings:
439 * 0-63
441 unsigned int pwc; /* Peripheral Wait Cycles */
444 * Valid phc settings:
445 * 0-7
447 unsigned int phc; /* Peripheral Hold Cycles */
450 * Valid cp (channel priority) settings:
451 * PRIORITY_LOW
452 * PRIORITY_MID_LOW
453 * PRIORITY_MID_HIGH
454 * PRIORITY_HIGH
456 unsigned int cp; /* channel priority */
459 * Valid pf (memory read prefetch) settings:
461 * PREFETCH_1
462 * PREFETCH_2
463 * PREFETCH_4
465 unsigned int pf; /* memory read prefetch */
468 * Boolean: channel interrupt enable
469 * NOTE: for sgl transfers, only the last descriptor will be setup to
470 * interrupt.
472 char int_enable;
474 char shift; /* easy access to byte_count shift, based on */
475 /* the width of the channel */
477 uint32_t control; /* channel control word */
479 /* These variabled are used ONLY in single dma transfers */
480 unsigned int mode; /* transfer mode */
481 phys_addr_t addr;
482 char ce; /* channel enable */
483 #ifdef CONFIG_STB03xxx
484 char ch_enable;
485 char tcd_disable;
486 char ece_enable;
487 char td; /* transfer direction */
488 #endif
490 char int_on_final_sg;/* for scatter/gather - only interrupt on last sg */
491 } ppc_dma_ch_t;
494 * PPC44x DMA implementations have a slightly different
495 * descriptor layout. Probably moved about due to the
496 * change to 64-bit addresses and link pointer. I don't
497 * know why they didn't just leave control_count after
498 * the dst_addr.
500 #ifdef PPC4xx_DMA_64BIT
501 typedef struct {
502 uint32_t control;
503 uint32_t control_count;
504 phys_addr_t src_addr;
505 phys_addr_t dst_addr;
506 phys_addr_t next;
507 } ppc_sgl_t;
508 #else
509 typedef struct {
510 uint32_t control;
511 phys_addr_t src_addr;
512 phys_addr_t dst_addr;
513 uint32_t control_count;
514 uint32_t next;
515 } ppc_sgl_t;
516 #endif
518 typedef struct {
519 unsigned int dmanr;
520 uint32_t control; /* channel ctrl word; loaded from each descrptr */
521 uint32_t sgl_control; /* LK, TCI, ETI, and ERI bits in sgl descriptor */
522 dma_addr_t dma_addr; /* dma (physical) address of this list */
523 ppc_sgl_t *phead;
524 dma_addr_t phead_dma;
525 ppc_sgl_t *ptail;
526 dma_addr_t ptail_dma;
527 } sgl_list_info_t;
529 typedef struct {
530 phys_addr_t *src_addr;
531 phys_addr_t *dst_addr;
532 phys_addr_t dma_src_addr;
533 phys_addr_t dma_dst_addr;
534 } pci_alloc_desc_t;
536 extern ppc_dma_ch_t dma_channels[];
539 * The DMA API are in ppc4xx_dma.c and ppc4xx_sgdma.c
541 extern int ppc4xx_init_dma_channel(unsigned int, ppc_dma_ch_t *);
542 extern int ppc4xx_get_channel_config(unsigned int, ppc_dma_ch_t *);
543 extern int ppc4xx_set_channel_priority(unsigned int, unsigned int);
544 extern unsigned int ppc4xx_get_peripheral_width(unsigned int);
545 extern void ppc4xx_set_sg_addr(int, phys_addr_t);
546 extern int ppc4xx_add_dma_sgl(sgl_handle_t, phys_addr_t, phys_addr_t, unsigned int);
547 extern void ppc4xx_enable_dma_sgl(sgl_handle_t);
548 extern void ppc4xx_disable_dma_sgl(sgl_handle_t);
549 extern int ppc4xx_get_dma_sgl_residue(sgl_handle_t, phys_addr_t *, phys_addr_t *);
550 extern int ppc4xx_delete_dma_sgl_element(sgl_handle_t, phys_addr_t *, phys_addr_t *);
551 extern int ppc4xx_alloc_dma_handle(sgl_handle_t *, unsigned int, unsigned int);
552 extern void ppc4xx_free_dma_handle(sgl_handle_t);
553 extern int ppc4xx_get_dma_status(void);
554 extern int ppc4xx_enable_burst(unsigned int);
555 extern int ppc4xx_disable_burst(unsigned int);
556 extern int ppc4xx_set_burst_size(unsigned int, unsigned int);
557 extern void ppc4xx_set_src_addr(int dmanr, phys_addr_t src_addr);
558 extern void ppc4xx_set_dst_addr(int dmanr, phys_addr_t dst_addr);
559 extern void ppc4xx_enable_dma(unsigned int dmanr);
560 extern void ppc4xx_disable_dma(unsigned int dmanr);
561 extern void ppc4xx_set_dma_count(unsigned int dmanr, unsigned int count);
562 extern int ppc4xx_get_dma_residue(unsigned int dmanr);
563 extern void ppc4xx_set_dma_addr2(unsigned int dmanr, phys_addr_t src_dma_addr,
564 phys_addr_t dst_dma_addr);
565 extern int ppc4xx_enable_dma_interrupt(unsigned int dmanr);
566 extern int ppc4xx_disable_dma_interrupt(unsigned int dmanr);
567 extern int ppc4xx_clr_dma_status(unsigned int dmanr);
568 extern int ppc4xx_map_dma_port(unsigned int dmanr, unsigned int ocp_dma,short dma_chan);
569 extern int ppc4xx_disable_dma_port(unsigned int dmanr, unsigned int ocp_dma,short dma_chan);
570 extern int ppc4xx_set_dma_mode(unsigned int dmanr, unsigned int mode);
572 /* These are in kernel/dma.c: */
574 /* reserve a DMA channel */
575 extern int request_dma(unsigned int dmanr, const char *device_id);
576 /* release it again */
577 extern void free_dma(unsigned int dmanr);
578 #endif
579 #endif /* __KERNEL__ */