2 * arch/sh/mm/consistent.c
4 * Copyright (C) 2004 - 2007 Paul Mundt
6 * Declared coherent memory functions based on arch/x86/kernel/pci-dma_32.c
8 * This file is subject to the terms and conditions of the GNU General Public
9 * License. See the file "COPYING" in the main directory of this archive
13 #include <linux/platform_device.h>
14 #include <linux/dma-mapping.h>
15 #include <asm/cacheflush.h>
16 #include <asm/addrspace.h>
19 void *dma_alloc_coherent(struct device
*dev
, size_t size
,
20 dma_addr_t
*dma_handle
, gfp_t gfp
)
22 void *ret
, *ret_nocache
;
23 int order
= get_order(size
);
25 if (dma_alloc_from_coherent(dev
, size
, dma_handle
, &ret
))
28 ret
= (void *)__get_free_pages(gfp
, order
);
34 * Pages from the page allocator may have data present in
35 * cache. So flush the cache before using uncached memory.
37 dma_cache_sync(dev
, ret
, size
, DMA_BIDIRECTIONAL
);
39 ret_nocache
= (void __force
*)ioremap_nocache(virt_to_phys(ret
), size
);
41 free_pages((unsigned long)ret
, order
);
45 *dma_handle
= virt_to_phys(ret
);
48 EXPORT_SYMBOL(dma_alloc_coherent
);
50 void dma_free_coherent(struct device
*dev
, size_t size
,
51 void *vaddr
, dma_addr_t dma_handle
)
53 int order
= get_order(size
);
55 if (!dma_release_from_coherent(dev
, order
, vaddr
)) {
56 WARN_ON(irqs_disabled()); /* for portability */
57 free_pages((unsigned long)phys_to_virt(dma_handle
), order
);
61 EXPORT_SYMBOL(dma_free_coherent
);
63 void dma_cache_sync(struct device
*dev
, void *vaddr
, size_t size
,
64 enum dma_data_direction direction
)
69 void *p1addr
= (void*) P1SEGADDR((unsigned long)vaddr
);
73 case DMA_FROM_DEVICE
: /* invalidate only */
74 __flush_invalidate_region(p1addr
, size
);
76 case DMA_TO_DEVICE
: /* writeback only */
77 __flush_wback_region(p1addr
, size
);
79 case DMA_BIDIRECTIONAL
: /* writeback and invalidate */
80 __flush_purge_region(p1addr
, size
);
86 EXPORT_SYMBOL(dma_cache_sync
);
88 static int __init
memchunk_setup(char *str
)
90 return 1; /* accept anything that begins with "memchunk." */
92 __setup("memchunk.", memchunk_setup
);
94 static void __init
memchunk_cmdline_override(char *name
, unsigned long *sizep
)
96 char *p
= boot_command_line
;
99 while ((p
= strstr(p
, "memchunk."))) {
100 p
+= 9; /* strlen("memchunk.") */
101 if (!strncmp(name
, p
, k
) && p
[k
] == '=') {
103 *sizep
= memparse(p
, NULL
);
104 pr_info("%s: forcing memory chunk size to 0x%08lx\n",
111 int __init
platform_resource_setup_memory(struct platform_device
*pdev
,
112 char *name
, unsigned long memsize
)
115 dma_addr_t dma_handle
;
118 r
= pdev
->resource
+ pdev
->num_resources
- 1;
120 pr_warning("%s: unable to find empty space for resource\n",
125 memchunk_cmdline_override(name
, &memsize
);
129 buf
= dma_alloc_coherent(NULL
, memsize
, &dma_handle
, GFP_KERNEL
);
131 pr_warning("%s: unable to allocate memory\n", name
);
135 memset(buf
, 0, memsize
);
137 r
->flags
= IORESOURCE_MEM
;
138 r
->start
= dma_handle
;
139 r
->end
= r
->start
+ memsize
- 1;