1 // SPDX-License-Identifier: GPL-2.0
3 * Re-map IO memory to kernel address space so that we can access it.
4 * This is needed for high PCI addresses that aren't mapped in the
5 * 640k-1MB IO memory area on PC's
7 * (C) Copyright 1995 1996 Linus Torvalds
9 #include <linux/vmalloc.h>
11 #include <linux/sched.h>
13 #include <linux/export.h>
14 #include <asm/cacheflush.h>
15 #include <asm/pgtable.h>
17 #ifdef CONFIG_HAVE_ARCH_HUGE_VMAP
18 static int __read_mostly ioremap_p4d_capable
;
19 static int __read_mostly ioremap_pud_capable
;
20 static int __read_mostly ioremap_pmd_capable
;
21 static int __read_mostly ioremap_huge_disabled
;
23 static int __init
set_nohugeiomap(char *str
)
25 ioremap_huge_disabled
= 1;
28 early_param("nohugeiomap", set_nohugeiomap
);
30 void __init
ioremap_huge_init(void)
32 if (!ioremap_huge_disabled
) {
33 if (arch_ioremap_p4d_supported())
34 ioremap_p4d_capable
= 1;
35 if (arch_ioremap_pud_supported())
36 ioremap_pud_capable
= 1;
37 if (arch_ioremap_pmd_supported())
38 ioremap_pmd_capable
= 1;
42 static inline int ioremap_p4d_enabled(void)
44 return ioremap_p4d_capable
;
47 static inline int ioremap_pud_enabled(void)
49 return ioremap_pud_capable
;
52 static inline int ioremap_pmd_enabled(void)
54 return ioremap_pmd_capable
;
57 #else /* !CONFIG_HAVE_ARCH_HUGE_VMAP */
58 static inline int ioremap_p4d_enabled(void) { return 0; }
59 static inline int ioremap_pud_enabled(void) { return 0; }
60 static inline int ioremap_pmd_enabled(void) { return 0; }
61 #endif /* CONFIG_HAVE_ARCH_HUGE_VMAP */
63 static int ioremap_pte_range(pmd_t
*pmd
, unsigned long addr
,
64 unsigned long end
, phys_addr_t phys_addr
, pgprot_t prot
)
69 pfn
= phys_addr
>> PAGE_SHIFT
;
70 pte
= pte_alloc_kernel(pmd
, addr
);
74 BUG_ON(!pte_none(*pte
));
75 set_pte_at(&init_mm
, addr
, pte
, pfn_pte(pfn
, prot
));
77 } while (pte
++, addr
+= PAGE_SIZE
, addr
!= end
);
81 static int ioremap_try_huge_pmd(pmd_t
*pmd
, unsigned long addr
,
82 unsigned long end
, phys_addr_t phys_addr
,
85 if (!ioremap_pmd_enabled())
88 if ((end
- addr
) != PMD_SIZE
)
91 if (!IS_ALIGNED(addr
, PMD_SIZE
))
94 if (!IS_ALIGNED(phys_addr
, PMD_SIZE
))
97 if (pmd_present(*pmd
) && !pmd_free_pte_page(pmd
, addr
))
100 return pmd_set_huge(pmd
, phys_addr
, prot
);
103 static inline int ioremap_pmd_range(pud_t
*pud
, unsigned long addr
,
104 unsigned long end
, phys_addr_t phys_addr
, pgprot_t prot
)
109 pmd
= pmd_alloc(&init_mm
, pud
, addr
);
113 next
= pmd_addr_end(addr
, end
);
115 if (ioremap_try_huge_pmd(pmd
, addr
, next
, phys_addr
, prot
))
118 if (ioremap_pte_range(pmd
, addr
, next
, phys_addr
, prot
))
120 } while (pmd
++, phys_addr
+= (next
- addr
), addr
= next
, addr
!= end
);
124 static int ioremap_try_huge_pud(pud_t
*pud
, unsigned long addr
,
125 unsigned long end
, phys_addr_t phys_addr
,
128 if (!ioremap_pud_enabled())
131 if ((end
- addr
) != PUD_SIZE
)
134 if (!IS_ALIGNED(addr
, PUD_SIZE
))
137 if (!IS_ALIGNED(phys_addr
, PUD_SIZE
))
140 if (pud_present(*pud
) && !pud_free_pmd_page(pud
, addr
))
143 return pud_set_huge(pud
, phys_addr
, prot
);
146 static inline int ioremap_pud_range(p4d_t
*p4d
, unsigned long addr
,
147 unsigned long end
, phys_addr_t phys_addr
, pgprot_t prot
)
152 pud
= pud_alloc(&init_mm
, p4d
, addr
);
156 next
= pud_addr_end(addr
, end
);
158 if (ioremap_try_huge_pud(pud
, addr
, next
, phys_addr
, prot
))
161 if (ioremap_pmd_range(pud
, addr
, next
, phys_addr
, prot
))
163 } while (pud
++, phys_addr
+= (next
- addr
), addr
= next
, addr
!= end
);
167 static int ioremap_try_huge_p4d(p4d_t
*p4d
, unsigned long addr
,
168 unsigned long end
, phys_addr_t phys_addr
,
171 if (!ioremap_p4d_enabled())
174 if ((end
- addr
) != P4D_SIZE
)
177 if (!IS_ALIGNED(addr
, P4D_SIZE
))
180 if (!IS_ALIGNED(phys_addr
, P4D_SIZE
))
183 if (p4d_present(*p4d
) && !p4d_free_pud_page(p4d
, addr
))
186 return p4d_set_huge(p4d
, phys_addr
, prot
);
189 static inline int ioremap_p4d_range(pgd_t
*pgd
, unsigned long addr
,
190 unsigned long end
, phys_addr_t phys_addr
, pgprot_t prot
)
195 p4d
= p4d_alloc(&init_mm
, pgd
, addr
);
199 next
= p4d_addr_end(addr
, end
);
201 if (ioremap_try_huge_p4d(p4d
, addr
, next
, phys_addr
, prot
))
204 if (ioremap_pud_range(p4d
, addr
, next
, phys_addr
, prot
))
206 } while (p4d
++, phys_addr
+= (next
- addr
), addr
= next
, addr
!= end
);
210 int ioremap_page_range(unsigned long addr
,
211 unsigned long end
, phys_addr_t phys_addr
, pgprot_t prot
)
222 pgd
= pgd_offset_k(addr
);
224 next
= pgd_addr_end(addr
, end
);
225 err
= ioremap_p4d_range(pgd
, addr
, next
, phys_addr
, prot
);
228 } while (pgd
++, phys_addr
+= (next
- addr
), addr
= next
, addr
!= end
);
230 flush_cache_vmap(start
, end
);
235 #ifdef CONFIG_GENERIC_IOREMAP
236 void __iomem
*ioremap_prot(phys_addr_t addr
, size_t size
, unsigned long prot
)
238 unsigned long offset
, vaddr
;
239 phys_addr_t last_addr
;
240 struct vm_struct
*area
;
242 /* Disallow wrap-around or zero size */
243 last_addr
= addr
+ size
- 1;
244 if (!size
|| last_addr
< addr
)
247 /* Page-align mappings */
248 offset
= addr
& (~PAGE_MASK
);
250 size
= PAGE_ALIGN(size
+ offset
);
252 area
= get_vm_area_caller(size
, VM_IOREMAP
,
253 __builtin_return_address(0));
256 vaddr
= (unsigned long)area
->addr
;
258 if (ioremap_page_range(vaddr
, vaddr
+ size
, addr
, __pgprot(prot
))) {
263 return (void __iomem
*)(vaddr
+ offset
);
265 EXPORT_SYMBOL(ioremap_prot
);
267 void iounmap(volatile void __iomem
*addr
)
269 vunmap((void *)((unsigned long)addr
& PAGE_MASK
));
271 EXPORT_SYMBOL(iounmap
);
272 #endif /* CONFIG_GENERIC_IOREMAP */