Merge tag 'for_linus' of git://git.kernel.org/pub/scm/linux/kernel/git/mst/vhost
[cris-mirror.git] / arch / mips / include / asm / mach-loongson64 / dma-coherence.h
blobb1b575f5c6c10288f4827f736bcd5b857697a711
1 /*
2 * This file is subject to the terms and conditions of the GNU General Public
3 * License. See the file "COPYING" in the main directory of this archive
4 * for more details.
6 * Copyright (C) 2006, 07 Ralf Baechle <ralf@linux-mips.org>
7 * Copyright (C) 2007 Lemote, Inc. & Institute of Computing Technology
8 * Author: Fuxin Zhang, zhangfx@lemote.com
11 #ifndef __ASM_MACH_LOONGSON64_DMA_COHERENCE_H
12 #define __ASM_MACH_LOONGSON64_DMA_COHERENCE_H
14 #ifdef CONFIG_SWIOTLB
15 #include <linux/swiotlb.h>
16 #endif
18 struct device;
20 static inline bool dma_capable(struct device *dev, dma_addr_t addr, size_t size)
22 if (!dev->dma_mask)
23 return false;
25 return addr + size - 1 <= *dev->dma_mask;
28 extern dma_addr_t phys_to_dma(struct device *dev, phys_addr_t paddr);
29 extern phys_addr_t dma_to_phys(struct device *dev, dma_addr_t daddr);
30 static inline dma_addr_t plat_map_dma_mem(struct device *dev, void *addr,
31 size_t size)
33 #ifdef CONFIG_CPU_LOONGSON3
34 return phys_to_dma(dev, virt_to_phys(addr));
35 #else
36 return virt_to_phys(addr) | 0x80000000;
37 #endif
40 static inline dma_addr_t plat_map_dma_mem_page(struct device *dev,
41 struct page *page)
43 #ifdef CONFIG_CPU_LOONGSON3
44 return phys_to_dma(dev, page_to_phys(page));
45 #else
46 return page_to_phys(page) | 0x80000000;
47 #endif
50 static inline unsigned long plat_dma_addr_to_phys(struct device *dev,
51 dma_addr_t dma_addr)
53 #if defined(CONFIG_CPU_LOONGSON3) && defined(CONFIG_64BIT)
54 return dma_to_phys(dev, dma_addr);
55 #elif defined(CONFIG_CPU_LOONGSON2F) && defined(CONFIG_64BIT)
56 return (dma_addr > 0x8fffffff) ? dma_addr : (dma_addr & 0x0fffffff);
57 #else
58 return dma_addr & 0x7fffffff;
59 #endif
62 static inline void plat_unmap_dma_mem(struct device *dev, dma_addr_t dma_addr,
63 size_t size, enum dma_data_direction direction)
67 static inline int plat_dma_supported(struct device *dev, u64 mask)
70 * we fall back to GFP_DMA when the mask isn't all 1s,
71 * so we can't guarantee allocations that must be
72 * within a tighter range than GFP_DMA..
74 if (mask < DMA_BIT_MASK(24))
75 return 0;
77 return 1;
80 static inline int plat_device_is_coherent(struct device *dev)
82 #ifdef CONFIG_DMA_NONCOHERENT
83 return 0;
84 #else
85 return 1;
86 #endif /* CONFIG_DMA_NONCOHERENT */
89 static inline void plat_post_dma_flush(struct device *dev)
93 #endif /* __ASM_MACH_LOONGSON64_DMA_COHERENCE_H */