1 /* SPDX-License-Identifier: GPL-2.0-only */
3 * Copyright (C) 2012 ARM Ltd.
5 #ifndef __ASM_PGTABLE_HWDEF_H
6 #define __ASM_PGTABLE_HWDEF_H
8 #include <asm/memory.h>
11 * Number of page-table levels required to address 'va_bits' wide
12 * address, without section mapping. We resolve the top (va_bits - PAGE_SHIFT)
13 * bits with (PAGE_SHIFT - 3) bits at each page table level. Hence:
15 * levels = DIV_ROUND_UP((va_bits - PAGE_SHIFT), (PAGE_SHIFT - 3))
17 * where DIV_ROUND_UP(n, d) => (((n) + (d) - 1) / (d))
19 * We cannot include linux/kernel.h which defines DIV_ROUND_UP here
20 * due to build issues. So we open code DIV_ROUND_UP here:
22 * ((((va_bits) - PAGE_SHIFT) + (PAGE_SHIFT - 3) - 1) / (PAGE_SHIFT - 3))
24 * which gets simplified as :
26 #define ARM64_HW_PGTABLE_LEVELS(va_bits) (((va_bits) - 4) / (PAGE_SHIFT - 3))
29 * Size mapped by an entry at level n ( 0 <= n <= 3)
30 * We map (PAGE_SHIFT - 3) at all translation levels and PAGE_SHIFT bits
31 * in the final page. The maximum number of translation levels supported by
32 * the architecture is 4. Hence, starting at at level n, we have further
33 * ((4 - n) - 1) levels of translation excluding the offset within the page.
34 * So, the total number of bits mapped by an entry at level n is :
36 * ((4 - n) - 1) * (PAGE_SHIFT - 3) + PAGE_SHIFT
38 * Rearranging it a bit we get :
39 * (4 - n) * (PAGE_SHIFT - 3) + 3
41 #define ARM64_HW_PGTABLE_LEVEL_SHIFT(n) ((PAGE_SHIFT - 3) * (4 - (n)) + 3)
43 #define PTRS_PER_PTE (1 << (PAGE_SHIFT - 3))
46 * PMD_SHIFT determines the size a level 2 page table entry can map.
48 #if CONFIG_PGTABLE_LEVELS > 2
49 #define PMD_SHIFT ARM64_HW_PGTABLE_LEVEL_SHIFT(2)
50 #define PMD_SIZE (_AC(1, UL) << PMD_SHIFT)
51 #define PMD_MASK (~(PMD_SIZE-1))
52 #define PTRS_PER_PMD PTRS_PER_PTE
56 * PUD_SHIFT determines the size a level 1 page table entry can map.
58 #if CONFIG_PGTABLE_LEVELS > 3
59 #define PUD_SHIFT ARM64_HW_PGTABLE_LEVEL_SHIFT(1)
60 #define PUD_SIZE (_AC(1, UL) << PUD_SHIFT)
61 #define PUD_MASK (~(PUD_SIZE-1))
62 #define PTRS_PER_PUD PTRS_PER_PTE
66 * PGDIR_SHIFT determines the size a top-level page table entry can map
67 * (depending on the configuration, this level can be 0, 1 or 2).
69 #define PGDIR_SHIFT ARM64_HW_PGTABLE_LEVEL_SHIFT(4 - CONFIG_PGTABLE_LEVELS)
70 #define PGDIR_SIZE (_AC(1, UL) << PGDIR_SHIFT)
71 #define PGDIR_MASK (~(PGDIR_SIZE-1))
72 #define PTRS_PER_PGD (1 << (MAX_USER_VA_BITS - PGDIR_SHIFT))
75 * Section address mask and size definitions.
77 #define SECTION_SHIFT PMD_SHIFT
78 #define SECTION_SIZE (_AC(1, UL) << SECTION_SHIFT)
79 #define SECTION_MASK (~(SECTION_SIZE-1))
82 * Contiguous page definitions.
84 #ifdef CONFIG_ARM64_64K_PAGES
85 #define CONT_PTE_SHIFT 5
86 #define CONT_PMD_SHIFT 5
87 #elif defined(CONFIG_ARM64_16K_PAGES)
88 #define CONT_PTE_SHIFT 7
89 #define CONT_PMD_SHIFT 5
91 #define CONT_PTE_SHIFT 4
92 #define CONT_PMD_SHIFT 4
95 #define CONT_PTES (1 << CONT_PTE_SHIFT)
96 #define CONT_PTE_SIZE (CONT_PTES * PAGE_SIZE)
97 #define CONT_PTE_MASK (~(CONT_PTE_SIZE - 1))
98 #define CONT_PMDS (1 << CONT_PMD_SHIFT)
99 #define CONT_PMD_SIZE (CONT_PMDS * PMD_SIZE)
100 #define CONT_PMD_MASK (~(CONT_PMD_SIZE - 1))
101 /* the the numerical offset of the PTE within a range of CONT_PTES */
102 #define CONT_RANGE_OFFSET(addr) (((addr)>>PAGE_SHIFT)&(CONT_PTES-1))
105 * Hardware page table definitions.
107 * Level 1 descriptor (PUD).
109 #define PUD_TYPE_TABLE (_AT(pudval_t, 3) << 0)
110 #define PUD_TABLE_BIT (_AT(pudval_t, 1) << 1)
111 #define PUD_TYPE_MASK (_AT(pudval_t, 3) << 0)
112 #define PUD_TYPE_SECT (_AT(pudval_t, 1) << 0)
115 * Level 2 descriptor (PMD).
117 #define PMD_TYPE_MASK (_AT(pmdval_t, 3) << 0)
118 #define PMD_TYPE_TABLE (_AT(pmdval_t, 3) << 0)
119 #define PMD_TYPE_SECT (_AT(pmdval_t, 1) << 0)
120 #define PMD_TABLE_BIT (_AT(pmdval_t, 1) << 1)
125 #define PMD_SECT_VALID (_AT(pmdval_t, 1) << 0)
126 #define PMD_SECT_USER (_AT(pmdval_t, 1) << 6) /* AP[1] */
127 #define PMD_SECT_RDONLY (_AT(pmdval_t, 1) << 7) /* AP[2] */
128 #define PMD_SECT_S (_AT(pmdval_t, 3) << 8)
129 #define PMD_SECT_AF (_AT(pmdval_t, 1) << 10)
130 #define PMD_SECT_NG (_AT(pmdval_t, 1) << 11)
131 #define PMD_SECT_CONT (_AT(pmdval_t, 1) << 52)
132 #define PMD_SECT_PXN (_AT(pmdval_t, 1) << 53)
133 #define PMD_SECT_UXN (_AT(pmdval_t, 1) << 54)
136 * AttrIndx[2:0] encoding (mapping attributes defined in the MAIR* registers).
138 #define PMD_ATTRINDX(t) (_AT(pmdval_t, (t)) << 2)
139 #define PMD_ATTRINDX_MASK (_AT(pmdval_t, 7) << 2)
142 * Level 3 descriptor (PTE).
144 #define PTE_VALID (_AT(pteval_t, 1) << 0)
145 #define PTE_TYPE_MASK (_AT(pteval_t, 3) << 0)
146 #define PTE_TYPE_PAGE (_AT(pteval_t, 3) << 0)
147 #define PTE_TABLE_BIT (_AT(pteval_t, 1) << 1)
148 #define PTE_USER (_AT(pteval_t, 1) << 6) /* AP[1] */
149 #define PTE_RDONLY (_AT(pteval_t, 1) << 7) /* AP[2] */
150 #define PTE_SHARED (_AT(pteval_t, 3) << 8) /* SH[1:0], inner shareable */
151 #define PTE_AF (_AT(pteval_t, 1) << 10) /* Access Flag */
152 #define PTE_NG (_AT(pteval_t, 1) << 11) /* nG */
153 #define PTE_DBM (_AT(pteval_t, 1) << 51) /* Dirty Bit Management */
154 #define PTE_CONT (_AT(pteval_t, 1) << 52) /* Contiguous range */
155 #define PTE_PXN (_AT(pteval_t, 1) << 53) /* Privileged XN */
156 #define PTE_UXN (_AT(pteval_t, 1) << 54) /* User XN */
157 #define PTE_HYP_XN (_AT(pteval_t, 1) << 54) /* HYP XN */
159 #define PTE_ADDR_LOW (((_AT(pteval_t, 1) << (48 - PAGE_SHIFT)) - 1) << PAGE_SHIFT)
160 #ifdef CONFIG_ARM64_PA_BITS_52
161 #define PTE_ADDR_HIGH (_AT(pteval_t, 0xf) << 12)
162 #define PTE_ADDR_MASK (PTE_ADDR_LOW | PTE_ADDR_HIGH)
164 #define PTE_ADDR_MASK PTE_ADDR_LOW
168 * AttrIndx[2:0] encoding (mapping attributes defined in the MAIR* registers).
170 #define PTE_ATTRINDX(t) (_AT(pteval_t, (t)) << 2)
171 #define PTE_ATTRINDX_MASK (_AT(pteval_t, 7) << 2)
174 * 2nd stage PTE definitions
176 #define PTE_S2_RDONLY (_AT(pteval_t, 1) << 6) /* HAP[2:1] */
177 #define PTE_S2_RDWR (_AT(pteval_t, 3) << 6) /* HAP[2:1] */
178 #define PTE_S2_XN (_AT(pteval_t, 2) << 53) /* XN[1:0] */
180 #define PMD_S2_RDONLY (_AT(pmdval_t, 1) << 6) /* HAP[2:1] */
181 #define PMD_S2_RDWR (_AT(pmdval_t, 3) << 6) /* HAP[2:1] */
182 #define PMD_S2_XN (_AT(pmdval_t, 2) << 53) /* XN[1:0] */
184 #define PUD_S2_RDONLY (_AT(pudval_t, 1) << 6) /* HAP[2:1] */
185 #define PUD_S2_RDWR (_AT(pudval_t, 3) << 6) /* HAP[2:1] */
186 #define PUD_S2_XN (_AT(pudval_t, 2) << 53) /* XN[1:0] */
189 * Memory Attribute override for Stage-2 (MemAttr[3:0])
191 #define PTE_S2_MEMATTR(t) (_AT(pteval_t, (t)) << 2)
192 #define PTE_S2_MEMATTR_MASK (_AT(pteval_t, 0xf) << 2)
195 * EL2/HYP PTE/PMD definitions
197 #define PMD_HYP PMD_SECT_USER
198 #define PTE_HYP PTE_USER
201 * Highest possible physical address supported.
203 #define PHYS_MASK_SHIFT (CONFIG_ARM64_PA_BITS)
204 #define PHYS_MASK ((UL(1) << PHYS_MASK_SHIFT) - 1)
206 #define TTBR_CNP_BIT (UL(1) << 0)
211 #define TCR_T0SZ_OFFSET 0
212 #define TCR_T1SZ_OFFSET 16
213 #define TCR_T0SZ(x) ((UL(64) - (x)) << TCR_T0SZ_OFFSET)
214 #define TCR_T1SZ(x) ((UL(64) - (x)) << TCR_T1SZ_OFFSET)
215 #define TCR_TxSZ(x) (TCR_T0SZ(x) | TCR_T1SZ(x))
216 #define TCR_TxSZ_WIDTH 6
217 #define TCR_T0SZ_MASK (((UL(1) << TCR_TxSZ_WIDTH) - 1) << TCR_T0SZ_OFFSET)
219 #define TCR_EPD0_SHIFT 7
220 #define TCR_EPD0_MASK (UL(1) << TCR_EPD0_SHIFT)
221 #define TCR_IRGN0_SHIFT 8
222 #define TCR_IRGN0_MASK (UL(3) << TCR_IRGN0_SHIFT)
223 #define TCR_IRGN0_NC (UL(0) << TCR_IRGN0_SHIFT)
224 #define TCR_IRGN0_WBWA (UL(1) << TCR_IRGN0_SHIFT)
225 #define TCR_IRGN0_WT (UL(2) << TCR_IRGN0_SHIFT)
226 #define TCR_IRGN0_WBnWA (UL(3) << TCR_IRGN0_SHIFT)
228 #define TCR_EPD1_SHIFT 23
229 #define TCR_EPD1_MASK (UL(1) << TCR_EPD1_SHIFT)
230 #define TCR_IRGN1_SHIFT 24
231 #define TCR_IRGN1_MASK (UL(3) << TCR_IRGN1_SHIFT)
232 #define TCR_IRGN1_NC (UL(0) << TCR_IRGN1_SHIFT)
233 #define TCR_IRGN1_WBWA (UL(1) << TCR_IRGN1_SHIFT)
234 #define TCR_IRGN1_WT (UL(2) << TCR_IRGN1_SHIFT)
235 #define TCR_IRGN1_WBnWA (UL(3) << TCR_IRGN1_SHIFT)
237 #define TCR_IRGN_NC (TCR_IRGN0_NC | TCR_IRGN1_NC)
238 #define TCR_IRGN_WBWA (TCR_IRGN0_WBWA | TCR_IRGN1_WBWA)
239 #define TCR_IRGN_WT (TCR_IRGN0_WT | TCR_IRGN1_WT)
240 #define TCR_IRGN_WBnWA (TCR_IRGN0_WBnWA | TCR_IRGN1_WBnWA)
241 #define TCR_IRGN_MASK (TCR_IRGN0_MASK | TCR_IRGN1_MASK)
244 #define TCR_ORGN0_SHIFT 10
245 #define TCR_ORGN0_MASK (UL(3) << TCR_ORGN0_SHIFT)
246 #define TCR_ORGN0_NC (UL(0) << TCR_ORGN0_SHIFT)
247 #define TCR_ORGN0_WBWA (UL(1) << TCR_ORGN0_SHIFT)
248 #define TCR_ORGN0_WT (UL(2) << TCR_ORGN0_SHIFT)
249 #define TCR_ORGN0_WBnWA (UL(3) << TCR_ORGN0_SHIFT)
251 #define TCR_ORGN1_SHIFT 26
252 #define TCR_ORGN1_MASK (UL(3) << TCR_ORGN1_SHIFT)
253 #define TCR_ORGN1_NC (UL(0) << TCR_ORGN1_SHIFT)
254 #define TCR_ORGN1_WBWA (UL(1) << TCR_ORGN1_SHIFT)
255 #define TCR_ORGN1_WT (UL(2) << TCR_ORGN1_SHIFT)
256 #define TCR_ORGN1_WBnWA (UL(3) << TCR_ORGN1_SHIFT)
258 #define TCR_ORGN_NC (TCR_ORGN0_NC | TCR_ORGN1_NC)
259 #define TCR_ORGN_WBWA (TCR_ORGN0_WBWA | TCR_ORGN1_WBWA)
260 #define TCR_ORGN_WT (TCR_ORGN0_WT | TCR_ORGN1_WT)
261 #define TCR_ORGN_WBnWA (TCR_ORGN0_WBnWA | TCR_ORGN1_WBnWA)
262 #define TCR_ORGN_MASK (TCR_ORGN0_MASK | TCR_ORGN1_MASK)
264 #define TCR_SH0_SHIFT 12
265 #define TCR_SH0_MASK (UL(3) << TCR_SH0_SHIFT)
266 #define TCR_SH0_INNER (UL(3) << TCR_SH0_SHIFT)
268 #define TCR_SH1_SHIFT 28
269 #define TCR_SH1_MASK (UL(3) << TCR_SH1_SHIFT)
270 #define TCR_SH1_INNER (UL(3) << TCR_SH1_SHIFT)
271 #define TCR_SHARED (TCR_SH0_INNER | TCR_SH1_INNER)
273 #define TCR_TG0_SHIFT 14
274 #define TCR_TG0_MASK (UL(3) << TCR_TG0_SHIFT)
275 #define TCR_TG0_4K (UL(0) << TCR_TG0_SHIFT)
276 #define TCR_TG0_64K (UL(1) << TCR_TG0_SHIFT)
277 #define TCR_TG0_16K (UL(2) << TCR_TG0_SHIFT)
279 #define TCR_TG1_SHIFT 30
280 #define TCR_TG1_MASK (UL(3) << TCR_TG1_SHIFT)
281 #define TCR_TG1_16K (UL(1) << TCR_TG1_SHIFT)
282 #define TCR_TG1_4K (UL(2) << TCR_TG1_SHIFT)
283 #define TCR_TG1_64K (UL(3) << TCR_TG1_SHIFT)
285 #define TCR_IPS_SHIFT 32
286 #define TCR_IPS_MASK (UL(7) << TCR_IPS_SHIFT)
287 #define TCR_A1 (UL(1) << 22)
288 #define TCR_ASID16 (UL(1) << 36)
289 #define TCR_TBI0 (UL(1) << 37)
290 #define TCR_TBI1 (UL(1) << 38)
291 #define TCR_HA (UL(1) << 39)
292 #define TCR_HD (UL(1) << 40)
293 #define TCR_NFD0 (UL(1) << 53)
294 #define TCR_NFD1 (UL(1) << 54)
299 #ifdef CONFIG_ARM64_PA_BITS_52
301 * This should be GENMASK_ULL(47, 2).
302 * TTBR_ELx[1] is RES0 in this configuration.
304 #define TTBR_BADDR_MASK_52 (((UL(1) << 46) - 1) << 2)
307 #ifdef CONFIG_ARM64_USER_VA_BITS_52
308 /* Must be at least 64-byte aligned to prevent corruption of the TTBR */
309 #define TTBR1_BADDR_4852_OFFSET (((UL(1) << (52 - PGDIR_SHIFT)) - \
310 (UL(1) << (48 - PGDIR_SHIFT))) * 8)