drm/panthor: Don't add write fences to the shared BOs
[drm/drm-misc.git] / arch / riscv / mm / ptdump.c
blob9d5f657a251b328dffb0ea807b4576ab15cc4e92
1 // SPDX-License-Identifier: GPL-2.0-only
2 /*
3 * Copyright (C) 2019 SiFive
4 */
6 #include <linux/efi.h>
7 #include <linux/init.h>
8 #include <linux/debugfs.h>
9 #include <linux/memory_hotplug.h>
10 #include <linux/seq_file.h>
11 #include <linux/ptdump.h>
13 #include <linux/pgtable.h>
14 #include <asm/kasan.h>
16 #define pt_dump_seq_printf(m, fmt, args...) \
17 ({ \
18 if (m) \
19 seq_printf(m, fmt, ##args); \
22 #define pt_dump_seq_puts(m, fmt) \
23 ({ \
24 if (m) \
25 seq_printf(m, fmt); \
29 * The page dumper groups page table entries of the same type into a single
30 * description. It uses pg_state to track the range information while
31 * iterating over the pte entries. When the continuity is broken it then
32 * dumps out a description of the range.
34 struct pg_state {
35 struct ptdump_state ptdump;
36 struct seq_file *seq;
37 const struct addr_marker *marker;
38 unsigned long start_address;
39 unsigned long start_pa;
40 unsigned long last_pa;
41 int level;
42 u64 current_prot;
43 bool check_wx;
44 unsigned long wx_pages;
47 /* Address marker */
48 struct addr_marker {
49 unsigned long start_address;
50 const char *name;
53 /* Private information for debugfs */
54 struct ptd_mm_info {
55 struct mm_struct *mm;
56 const struct addr_marker *markers;
57 unsigned long base_addr;
58 unsigned long end;
61 enum address_markers_idx {
62 FIXMAP_START_NR,
63 FIXMAP_END_NR,
64 PCI_IO_START_NR,
65 PCI_IO_END_NR,
66 #ifdef CONFIG_SPARSEMEM_VMEMMAP
67 VMEMMAP_START_NR,
68 VMEMMAP_END_NR,
69 #endif
70 VMALLOC_START_NR,
71 VMALLOC_END_NR,
72 PAGE_OFFSET_NR,
73 #ifdef CONFIG_KASAN
74 KASAN_SHADOW_START_NR,
75 KASAN_SHADOW_END_NR,
76 #endif
77 #ifdef CONFIG_64BIT
78 MODULES_MAPPING_NR,
79 KERNEL_MAPPING_NR,
80 #endif
81 END_OF_SPACE_NR
84 static struct addr_marker address_markers[] = {
85 {0, "Fixmap start"},
86 {0, "Fixmap end"},
87 {0, "PCI I/O start"},
88 {0, "PCI I/O end"},
89 #ifdef CONFIG_SPARSEMEM_VMEMMAP
90 {0, "vmemmap start"},
91 {0, "vmemmap end"},
92 #endif
93 {0, "vmalloc() area"},
94 {0, "vmalloc() end"},
95 {0, "Linear mapping"},
96 #ifdef CONFIG_KASAN
97 {0, "Kasan shadow start"},
98 {0, "Kasan shadow end"},
99 #endif
100 #ifdef CONFIG_64BIT
101 {0, "Modules/BPF mapping"},
102 {0, "Kernel mapping"},
103 #endif
104 {-1, NULL},
107 static struct ptd_mm_info kernel_ptd_info = {
108 .mm = &init_mm,
109 .markers = address_markers,
110 .base_addr = 0,
111 .end = ULONG_MAX,
114 #ifdef CONFIG_EFI
115 static struct addr_marker efi_addr_markers[] = {
116 { 0, "UEFI runtime start" },
117 { SZ_1G, "UEFI runtime end" },
118 { -1, NULL }
121 static struct ptd_mm_info efi_ptd_info = {
122 .mm = &efi_mm,
123 .markers = efi_addr_markers,
124 .base_addr = 0,
125 .end = SZ_2G,
127 #endif
129 /* Page Table Entry */
130 struct prot_bits {
131 u64 mask;
132 const char *set;
133 const char *clear;
136 static const struct prot_bits pte_bits[] = {
138 #ifdef CONFIG_64BIT
139 .mask = _PAGE_NAPOT,
140 .set = "N",
141 .clear = ".",
142 }, {
143 .mask = _PAGE_MTMASK_SVPBMT,
144 .set = "MT(%s)",
145 .clear = " .. ",
146 }, {
147 #endif
148 .mask = _PAGE_SOFT,
149 .set = "RSW(%d)",
150 .clear = " .. ",
151 }, {
152 .mask = _PAGE_DIRTY,
153 .set = "D",
154 .clear = ".",
155 }, {
156 .mask = _PAGE_ACCESSED,
157 .set = "A",
158 .clear = ".",
159 }, {
160 .mask = _PAGE_GLOBAL,
161 .set = "G",
162 .clear = ".",
163 }, {
164 .mask = _PAGE_USER,
165 .set = "U",
166 .clear = ".",
167 }, {
168 .mask = _PAGE_EXEC,
169 .set = "X",
170 .clear = ".",
171 }, {
172 .mask = _PAGE_WRITE,
173 .set = "W",
174 .clear = ".",
175 }, {
176 .mask = _PAGE_READ,
177 .set = "R",
178 .clear = ".",
179 }, {
180 .mask = _PAGE_PRESENT,
181 .set = "V",
182 .clear = ".",
186 /* Page Level */
187 struct pg_level {
188 const char *name;
189 u64 mask;
192 static struct pg_level pg_level[] = {
193 { /* pgd */
194 .name = "PGD",
195 }, { /* p4d */
196 .name = (CONFIG_PGTABLE_LEVELS > 4) ? "P4D" : "PGD",
197 }, { /* pud */
198 .name = (CONFIG_PGTABLE_LEVELS > 3) ? "PUD" : "PGD",
199 }, { /* pmd */
200 .name = (CONFIG_PGTABLE_LEVELS > 2) ? "PMD" : "PGD",
201 }, { /* pte */
202 .name = "PTE",
206 static void dump_prot(struct pg_state *st)
208 unsigned int i;
210 for (i = 0; i < ARRAY_SIZE(pte_bits); i++) {
211 char s[7];
212 unsigned long val;
214 val = st->current_prot & pte_bits[i].mask;
215 if (val) {
216 if (pte_bits[i].mask == _PAGE_SOFT)
217 sprintf(s, pte_bits[i].set, val >> 8);
218 #ifdef CONFIG_64BIT
219 else if (pte_bits[i].mask == _PAGE_MTMASK_SVPBMT) {
220 if (val == _PAGE_NOCACHE_SVPBMT)
221 sprintf(s, pte_bits[i].set, "NC");
222 else if (val == _PAGE_IO_SVPBMT)
223 sprintf(s, pte_bits[i].set, "IO");
224 else
225 sprintf(s, pte_bits[i].set, "??");
227 #endif
228 else
229 sprintf(s, "%s", pte_bits[i].set);
230 } else {
231 sprintf(s, "%s", pte_bits[i].clear);
234 pt_dump_seq_printf(st->seq, " %s", s);
238 #ifdef CONFIG_64BIT
239 #define ADDR_FORMAT "0x%016lx"
240 #else
241 #define ADDR_FORMAT "0x%08lx"
242 #endif
243 static void dump_addr(struct pg_state *st, unsigned long addr)
245 static const char units[] = "KMGTPE";
246 const char *unit = units;
247 unsigned long delta;
249 pt_dump_seq_printf(st->seq, ADDR_FORMAT "-" ADDR_FORMAT " ",
250 st->start_address, addr);
252 pt_dump_seq_printf(st->seq, " " ADDR_FORMAT " ", st->start_pa);
253 delta = (addr - st->start_address) >> 10;
255 while (!(delta & 1023) && unit[1]) {
256 delta >>= 10;
257 unit++;
260 pt_dump_seq_printf(st->seq, "%9lu%c %s", delta, *unit,
261 pg_level[st->level].name);
264 static void note_prot_wx(struct pg_state *st, unsigned long addr)
266 if (!st->check_wx)
267 return;
269 if ((st->current_prot & (_PAGE_WRITE | _PAGE_EXEC)) !=
270 (_PAGE_WRITE | _PAGE_EXEC))
271 return;
273 WARN_ONCE(1, "riscv/mm: Found insecure W+X mapping at address %p/%pS\n",
274 (void *)st->start_address, (void *)st->start_address);
276 st->wx_pages += (addr - st->start_address) / PAGE_SIZE;
279 static void note_page(struct ptdump_state *pt_st, unsigned long addr,
280 int level, u64 val)
282 struct pg_state *st = container_of(pt_st, struct pg_state, ptdump);
283 u64 pa = PFN_PHYS(pte_pfn(__pte(val)));
284 u64 prot = 0;
286 if (level >= 0)
287 prot = val & pg_level[level].mask;
289 if (st->level == -1) {
290 st->level = level;
291 st->current_prot = prot;
292 st->start_address = addr;
293 st->start_pa = pa;
294 st->last_pa = pa;
295 pt_dump_seq_printf(st->seq, "---[ %s ]---\n", st->marker->name);
296 } else if (prot != st->current_prot ||
297 level != st->level || addr >= st->marker[1].start_address) {
298 if (st->current_prot) {
299 note_prot_wx(st, addr);
300 dump_addr(st, addr);
301 dump_prot(st);
302 pt_dump_seq_puts(st->seq, "\n");
305 while (addr >= st->marker[1].start_address) {
306 st->marker++;
307 pt_dump_seq_printf(st->seq, "---[ %s ]---\n",
308 st->marker->name);
311 st->start_address = addr;
312 st->start_pa = pa;
313 st->last_pa = pa;
314 st->current_prot = prot;
315 st->level = level;
316 } else {
317 st->last_pa = pa;
321 static void ptdump_walk(struct seq_file *s, struct ptd_mm_info *pinfo)
323 struct pg_state st = {
324 .seq = s,
325 .marker = pinfo->markers,
326 .level = -1,
327 .ptdump = {
328 .note_page = note_page,
329 .range = (struct ptdump_range[]) {
330 {pinfo->base_addr, pinfo->end},
331 {0, 0}
336 ptdump_walk_pgd(&st.ptdump, pinfo->mm, NULL);
339 bool ptdump_check_wx(void)
341 struct pg_state st = {
342 .seq = NULL,
343 .marker = (struct addr_marker[]) {
344 {0, NULL},
345 {-1, NULL},
347 .level = -1,
348 .check_wx = true,
349 .ptdump = {
350 .note_page = note_page,
351 .range = (struct ptdump_range[]) {
352 {KERN_VIRT_START, ULONG_MAX},
353 {0, 0}
358 ptdump_walk_pgd(&st.ptdump, &init_mm, NULL);
360 if (st.wx_pages) {
361 pr_warn("Checked W+X mappings: failed, %lu W+X pages found\n",
362 st.wx_pages);
364 return false;
365 } else {
366 pr_info("Checked W+X mappings: passed, no W+X pages found\n");
368 return true;
372 static int ptdump_show(struct seq_file *m, void *v)
374 get_online_mems();
375 ptdump_walk(m, m->private);
376 put_online_mems();
378 return 0;
381 DEFINE_SHOW_ATTRIBUTE(ptdump);
383 static int __init ptdump_init(void)
385 unsigned int i, j;
387 address_markers[FIXMAP_START_NR].start_address = FIXADDR_START;
388 address_markers[FIXMAP_END_NR].start_address = FIXADDR_TOP;
389 address_markers[PCI_IO_START_NR].start_address = PCI_IO_START;
390 address_markers[PCI_IO_END_NR].start_address = PCI_IO_END;
391 #ifdef CONFIG_SPARSEMEM_VMEMMAP
392 address_markers[VMEMMAP_START_NR].start_address = VMEMMAP_START;
393 address_markers[VMEMMAP_END_NR].start_address = VMEMMAP_END;
394 #endif
395 address_markers[VMALLOC_START_NR].start_address = VMALLOC_START;
396 address_markers[VMALLOC_END_NR].start_address = VMALLOC_END;
397 address_markers[PAGE_OFFSET_NR].start_address = PAGE_OFFSET;
398 #ifdef CONFIG_KASAN
399 address_markers[KASAN_SHADOW_START_NR].start_address = KASAN_SHADOW_START;
400 address_markers[KASAN_SHADOW_END_NR].start_address = KASAN_SHADOW_END;
401 #endif
402 #ifdef CONFIG_64BIT
403 address_markers[MODULES_MAPPING_NR].start_address = MODULES_VADDR;
404 address_markers[KERNEL_MAPPING_NR].start_address = kernel_map.virt_addr;
405 #endif
407 kernel_ptd_info.base_addr = KERN_VIRT_START;
409 pg_level[1].name = pgtable_l5_enabled ? "P4D" : "PGD";
410 pg_level[2].name = pgtable_l4_enabled ? "PUD" : "PGD";
412 for (i = 0; i < ARRAY_SIZE(pg_level); i++)
413 for (j = 0; j < ARRAY_SIZE(pte_bits); j++)
414 pg_level[i].mask |= pte_bits[j].mask;
416 debugfs_create_file("kernel_page_tables", 0400, NULL, &kernel_ptd_info,
417 &ptdump_fops);
418 #ifdef CONFIG_EFI
419 if (efi_enabled(EFI_RUNTIME_SERVICES))
420 debugfs_create_file("efi_page_tables", 0400, NULL, &efi_ptd_info,
421 &ptdump_fops);
422 #endif
424 return 0;
427 device_initcall(ptdump_init);