1 /* Fallback functions when the main IOMMU code is not compiled in. This
2 code is roughly equivalent to i386. */
3 #include <linux/dma-mapping.h>
4 #include <linux/scatterlist.h>
5 #include <linux/string.h>
6 #include <linux/init.h>
10 #include <asm/processor.h>
11 #include <asm/iommu.h>
15 check_addr(char *name
, struct device
*hwdev
, dma_addr_t bus
, size_t size
)
17 if (hwdev
&& !dma_capable(hwdev
, bus
, size
)) {
18 if (*hwdev
->dma_mask
>= DMA_BIT_MASK(32))
20 "nommu_%s: overflow %Lx+%zu of device mask %Lx\n",
21 name
, (long long)bus
, size
,
22 (long long)*hwdev
->dma_mask
);
28 static dma_addr_t
nommu_map_page(struct device
*dev
, struct page
*page
,
29 unsigned long offset
, size_t size
,
30 enum dma_data_direction dir
,
31 struct dma_attrs
*attrs
)
33 dma_addr_t bus
= page_to_phys(page
) + offset
;
35 if (!check_addr("map_single", dev
, bus
, size
))
36 return bad_dma_address
;
37 flush_write_buffers();
41 /* Map a set of buffers described by scatterlist in streaming
42 * mode for DMA. This is the scatter-gather version of the
43 * above pci_map_single interface. Here the scatter gather list
44 * elements are each tagged with the appropriate dma address
45 * and length. They are obtained via sg_dma_{address,length}(SG).
47 * NOTE: An implementation may be able to use a smaller number of
48 * DMA address/length pairs than there are SG table elements.
49 * (for example via virtual mapping capabilities)
50 * The routine returns the number of addr/length pairs actually
51 * used, at most nents.
53 * Device ownership issues as mentioned above for pci_map_single are
56 static int nommu_map_sg(struct device
*hwdev
, struct scatterlist
*sg
,
57 int nents
, enum dma_data_direction dir
,
58 struct dma_attrs
*attrs
)
60 struct scatterlist
*s
;
63 WARN_ON(nents
== 0 || sg
[0].length
== 0);
65 for_each_sg(sg
, s
, nents
, i
) {
67 s
->dma_address
= sg_phys(s
);
68 if (!check_addr("map_sg", hwdev
, s
->dma_address
, s
->length
))
70 s
->dma_length
= s
->length
;
72 flush_write_buffers();
76 static void nommu_free_coherent(struct device
*dev
, size_t size
, void *vaddr
,
79 free_pages((unsigned long)vaddr
, get_order(size
));
82 static void nommu_sync_single_for_device(struct device
*dev
,
83 dma_addr_t addr
, size_t size
,
84 enum dma_data_direction dir
)
86 flush_write_buffers();
90 static void nommu_sync_sg_for_device(struct device
*dev
,
91 struct scatterlist
*sg
, int nelems
,
92 enum dma_data_direction dir
)
94 flush_write_buffers();
97 struct dma_map_ops nommu_dma_ops
= {
98 .alloc_coherent
= dma_generic_alloc_coherent
,
99 .free_coherent
= nommu_free_coherent
,
100 .map_sg
= nommu_map_sg
,
101 .map_page
= nommu_map_page
,
102 .sync_single_for_device
= nommu_sync_single_for_device
,
103 .sync_sg_for_device
= nommu_sync_sg_for_device
,
107 void __init
no_iommu_init(void)
112 force_iommu
= 0; /* no HW IOMMU */
113 dma_ops
= &nommu_dma_ops
;