4 * Copyright (C) 2001 Deep Blue Solutions Ltd.
5 * Copyright (C) 2012 ARM Ltd.
7 * This program is free software; you can redistribute it and/or modify
8 * it under the terms of the GNU General Public License version 2 as
9 * published by the Free Software Foundation.
11 * This program is distributed in the hope that it will be useful,
12 * but WITHOUT ANY WARRANTY; without even the implied warranty of
13 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
14 * GNU General Public License for more details.
16 * You should have received a copy of the GNU General Public License
17 * along with this program. If not, see <http://www.gnu.org/licenses/>.
20 #include <linux/errno.h>
21 #include <linux/linkage.h>
22 #include <linux/init.h>
23 #include <asm/assembler.h>
24 #include <asm/cpufeature.h>
25 #include <asm/alternative.h>
26 #include <asm/asm-uaccess.h>
29 * flush_icache_range(start,end)
31 * Ensure that the I and D caches are coherent within specified region.
32 * This is typically used when code has been written to a memory region,
33 * and will be executed.
35 * - start - virtual start address of region
36 * - end - virtual end address of region
38 ENTRY(__flush_icache_range)
42 * __flush_cache_user_range(start,end)
44 * Ensure that the I and D caches are coherent within specified region.
45 * This is typically used when code has been written to a memory region,
46 * and will be executed.
48 * - start - virtual start address of region
49 * - end - virtual end address of region
51 ENTRY(__flush_cache_user_range)
52 uaccess_ttbr0_enable x2, x3, x4
53 alternative_if ARM64_HAS_CACHE_IDC
56 alternative_else_nop_endif
57 dcache_line_size x2, x3
61 user_alt 9f, "dc cvau, x4", "dc civac, x4", ARM64_WORKAROUND_CLEAN_CACHE
68 alternative_if ARM64_HAS_CACHE_DIC
71 alternative_else_nop_endif
72 invalidate_icache_by_line x0, x1, x2, x3, 9f
75 uaccess_ttbr0_disable x1, x2
80 ENDPROC(__flush_icache_range)
81 ENDPROC(__flush_cache_user_range)
84 * invalidate_icache_range(start,end)
86 * Ensure that the I cache is invalid within specified region.
88 * - start - virtual start address of region
89 * - end - virtual end address of region
91 ENTRY(invalidate_icache_range)
92 alternative_if ARM64_HAS_CACHE_DIC
96 alternative_else_nop_endif
98 uaccess_ttbr0_enable x2, x3, x4
100 invalidate_icache_by_line x0, x1, x2, x3, 2f
103 uaccess_ttbr0_disable x1, x2
108 ENDPROC(invalidate_icache_range)
111 * __flush_dcache_area(kaddr, size)
113 * Ensure that any D-cache lines for the interval [kaddr, kaddr+size)
114 * are cleaned and invalidated to the PoC.
116 * - kaddr - kernel address
117 * - size - size in question
119 ENTRY(__flush_dcache_area)
120 dcache_by_line_op civac, sy, x0, x1, x2, x3
122 ENDPIPROC(__flush_dcache_area)
125 * __clean_dcache_area_pou(kaddr, size)
127 * Ensure that any D-cache lines for the interval [kaddr, kaddr+size)
128 * are cleaned to the PoU.
130 * - kaddr - kernel address
131 * - size - size in question
133 ENTRY(__clean_dcache_area_pou)
134 alternative_if ARM64_HAS_CACHE_IDC
137 alternative_else_nop_endif
138 dcache_by_line_op cvau, ish, x0, x1, x2, x3
140 ENDPROC(__clean_dcache_area_pou)
143 * __inval_dcache_area(kaddr, size)
145 * Ensure that any D-cache lines for the interval [kaddr, kaddr+size)
146 * are invalidated. Any partial lines at the ends of the interval are
147 * also cleaned to PoC to prevent data loss.
149 * - kaddr - kernel address
150 * - size - size in question
152 ENTRY(__inval_dcache_area)
156 * __dma_inv_area(start, size)
157 * - start - virtual start address of region
158 * - size - size in question
162 dcache_line_size x2, x3
164 tst x1, x3 // end cache line aligned?
167 dc civac, x1 // clean & invalidate D / U line
168 1: tst x0, x3 // start cache line aligned?
171 dc civac, x0 // clean & invalidate D / U line
173 2: dc ivac, x0 // invalidate D / U line
179 ENDPIPROC(__inval_dcache_area)
180 ENDPROC(__dma_inv_area)
183 * __clean_dcache_area_poc(kaddr, size)
185 * Ensure that any D-cache lines for the interval [kaddr, kaddr+size)
186 * are cleaned to the PoC.
188 * - kaddr - kernel address
189 * - size - size in question
191 ENTRY(__clean_dcache_area_poc)
195 * __dma_clean_area(start, size)
196 * - start - virtual start address of region
197 * - size - size in question
200 dcache_by_line_op cvac, sy, x0, x1, x2, x3
202 ENDPIPROC(__clean_dcache_area_poc)
203 ENDPROC(__dma_clean_area)
206 * __clean_dcache_area_pop(kaddr, size)
208 * Ensure that any D-cache lines for the interval [kaddr, kaddr+size)
209 * are cleaned to the PoP.
211 * - kaddr - kernel address
212 * - size - size in question
214 ENTRY(__clean_dcache_area_pop)
215 alternative_if_not ARM64_HAS_DCPOP
216 b __clean_dcache_area_poc
217 alternative_else_nop_endif
218 dcache_by_line_op cvap, sy, x0, x1, x2, x3
220 ENDPIPROC(__clean_dcache_area_pop)
223 * __dma_flush_area(start, size)
225 * clean & invalidate D / U line
227 * - start - virtual start address of region
228 * - size - size in question
230 ENTRY(__dma_flush_area)
231 dcache_by_line_op civac, sy, x0, x1, x2, x3
233 ENDPIPROC(__dma_flush_area)
236 * __dma_map_area(start, size, dir)
237 * - start - kernel virtual start address
238 * - size - size of region
239 * - dir - DMA direction
241 ENTRY(__dma_map_area)
242 cmp w2, #DMA_FROM_DEVICE
245 ENDPIPROC(__dma_map_area)
248 * __dma_unmap_area(start, size, dir)
249 * - start - kernel virtual start address
250 * - size - size of region
251 * - dir - DMA direction
253 ENTRY(__dma_unmap_area)
254 cmp w2, #DMA_TO_DEVICE
257 ENDPIPROC(__dma_unmap_area)