2 * Dynamic DMA mapping support.
4 * On i386 there is no hardware dynamic DMA address translation,
5 * so consistent alloc/free are merely page allocation/freeing.
6 * The rest of the dynamic DMA mapping interface is implemented
10 #include <linux/types.h>
12 #include <linux/string.h>
13 #include <linux/pci.h>
16 struct dma_coherent_mem
{
21 unsigned long *bitmap
;
24 void *dma_alloc_coherent(struct device
*dev
, size_t size
,
25 dma_addr_t
*dma_handle
, unsigned int __nocast gfp
)
28 struct dma_coherent_mem
*mem
= dev
? dev
->dma_mem
: NULL
;
29 int order
= get_order(size
);
30 /* ignore region specifiers */
31 gfp
&= ~(__GFP_DMA
| __GFP_HIGHMEM
);
34 int page
= bitmap_find_free_region(mem
->bitmap
, mem
->size
,
37 *dma_handle
= mem
->device_base
+ (page
<< PAGE_SHIFT
);
38 ret
= mem
->virt_base
+ (page
<< PAGE_SHIFT
);
42 if (mem
->flags
& DMA_MEMORY_EXCLUSIVE
)
46 if (dev
== NULL
|| (dev
->coherent_dma_mask
< 0xffffffff))
49 ret
= (void *)__get_free_pages(gfp
, order
);
53 *dma_handle
= virt_to_phys(ret
);
58 void dma_free_coherent(struct device
*dev
, size_t size
,
59 void *vaddr
, dma_addr_t dma_handle
)
61 struct dma_coherent_mem
*mem
= dev
? dev
->dma_mem
: NULL
;
62 int order
= get_order(size
);
64 if (mem
&& vaddr
>= mem
->virt_base
&& vaddr
< (mem
->virt_base
+ (mem
->size
<< PAGE_SHIFT
))) {
65 int page
= (vaddr
- mem
->virt_base
) >> PAGE_SHIFT
;
67 bitmap_release_region(mem
->bitmap
, page
, order
);
69 free_pages((unsigned long)vaddr
, order
);
72 int dma_declare_coherent_memory(struct device
*dev
, dma_addr_t bus_addr
,
73 dma_addr_t device_addr
, size_t size
, int flags
)
75 void __iomem
*mem_base
;
76 int pages
= size
>> PAGE_SHIFT
;
77 int bitmap_size
= (pages
+ 31)/32;
79 if ((flags
& (DMA_MEMORY_MAP
| DMA_MEMORY_IO
)) == 0)
86 /* FIXME: this routine just ignores DMA_MEMORY_INCLUDES_CHILDREN */
88 mem_base
= ioremap(bus_addr
, size
);
92 dev
->dma_mem
= kmalloc(sizeof(struct dma_coherent_mem
), GFP_KERNEL
);
95 memset(dev
->dma_mem
, 0, sizeof(struct dma_coherent_mem
));
96 dev
->dma_mem
->bitmap
= kmalloc(bitmap_size
, GFP_KERNEL
);
97 if (!dev
->dma_mem
->bitmap
)
99 memset(dev
->dma_mem
->bitmap
, 0, bitmap_size
);
101 dev
->dma_mem
->virt_base
= mem_base
;
102 dev
->dma_mem
->device_base
= device_addr
;
103 dev
->dma_mem
->size
= pages
;
104 dev
->dma_mem
->flags
= flags
;
106 if (flags
& DMA_MEMORY_MAP
)
107 return DMA_MEMORY_MAP
;
109 return DMA_MEMORY_IO
;
112 kfree(dev
->dma_mem
->bitmap
);
116 EXPORT_SYMBOL(dma_declare_coherent_memory
);
118 void dma_release_declared_memory(struct device
*dev
)
120 struct dma_coherent_mem
*mem
= dev
->dma_mem
;
125 iounmap(mem
->virt_base
);
129 EXPORT_SYMBOL(dma_release_declared_memory
);
131 void *dma_mark_declared_memory_occupied(struct device
*dev
,
132 dma_addr_t device_addr
, size_t size
)
134 struct dma_coherent_mem
*mem
= dev
->dma_mem
;
135 int pages
= (size
+ (device_addr
& ~PAGE_MASK
) + PAGE_SIZE
- 1) >> PAGE_SHIFT
;
139 return ERR_PTR(-EINVAL
);
141 pos
= (device_addr
- mem
->device_base
) >> PAGE_SHIFT
;
142 err
= bitmap_allocate_region(mem
->bitmap
, pos
, get_order(pages
));
145 return mem
->virt_base
+ (pos
<< PAGE_SHIFT
);
147 EXPORT_SYMBOL(dma_mark_declared_memory_occupied
);