1 // SPDX-License-Identifier: GPL-2.0
3 * KCSAN short boot-time selftests.
5 * Copyright (C) 2019, Google LLC.
8 #define pr_fmt(fmt) "kcsan: " fmt
10 #include <linux/atomic.h>
11 #include <linux/bitops.h>
12 #include <linux/init.h>
13 #include <linux/kcsan-checks.h>
14 #include <linux/kernel.h>
15 #include <linux/printk.h>
16 #include <linux/random.h>
17 #include <linux/sched.h>
18 #include <linux/spinlock.h>
19 #include <linux/types.h>
23 #define ITERS_PER_TEST 2000
26 * Test watchpoint encode and decode: check that encoding some access's info,
27 * and then subsequent decode preserves the access's info.
29 static bool __init
test_encode_decode(void)
33 for (i
= 0; i
< ITERS_PER_TEST
; ++i
) {
34 size_t size
= get_random_u32_inclusive(1, MAX_ENCODABLE_SIZE
);
35 bool is_write
= !!get_random_u32_below(2);
36 unsigned long verif_masked_addr
;
37 long encoded_watchpoint
;
42 get_random_bytes(&addr
, sizeof(addr
));
46 if (WARN_ON(!check_encodable(addr
, size
)))
49 encoded_watchpoint
= encode_watchpoint(addr
, size
, is_write
);
51 /* Check special watchpoints */
52 if (WARN_ON(decode_watchpoint(INVALID_WATCHPOINT
, &verif_masked_addr
, &verif_size
, &verif_is_write
)))
54 if (WARN_ON(decode_watchpoint(CONSUMED_WATCHPOINT
, &verif_masked_addr
, &verif_size
, &verif_is_write
)))
57 /* Check decoding watchpoint returns same data */
58 if (WARN_ON(!decode_watchpoint(encoded_watchpoint
, &verif_masked_addr
, &verif_size
, &verif_is_write
)))
60 if (WARN_ON(verif_masked_addr
!= (addr
& WATCHPOINT_ADDR_MASK
)))
62 if (WARN_ON(verif_size
!= size
))
64 if (WARN_ON(is_write
!= verif_is_write
))
69 pr_err("%s fail: %s %zu bytes @ %lx -> encoded: %lx -> %s %zu bytes @ %lx\n",
70 __func__
, is_write
? "write" : "read", size
, addr
, encoded_watchpoint
,
71 verif_is_write
? "write" : "read", verif_size
, verif_masked_addr
);
78 /* Test access matching function. */
79 static bool __init
test_matching_access(void)
81 if (WARN_ON(!matching_access(10, 1, 10, 1)))
83 if (WARN_ON(!matching_access(10, 2, 11, 1)))
85 if (WARN_ON(!matching_access(10, 1, 9, 2)))
87 if (WARN_ON(matching_access(10, 1, 11, 1)))
89 if (WARN_ON(matching_access(9, 1, 10, 1)))
93 * An access of size 0 could match another access, as demonstrated here.
94 * Rather than add more comparisons to 'matching_access()', which would
95 * end up in the fast-path for *all* checks, check_access() simply
96 * returns for all accesses of size 0.
98 if (WARN_ON(!matching_access(8, 8, 12, 0)))
105 * Correct memory barrier instrumentation is critical to avoiding false
106 * positives: simple test to check at boot certain barriers are always properly
107 * instrumented. See kcsan_test for a more complete test.
109 static DEFINE_SPINLOCK(test_spinlock
);
110 static bool __init
test_barrier(void)
112 #ifdef CONFIG_KCSAN_WEAK_MEMORY
113 struct kcsan_scoped_access
*reorder_access
= ¤t
->kcsan_ctx
.reorder_access
;
115 struct kcsan_scoped_access
*reorder_access
= NULL
;
118 arch_spinlock_t arch_spinlock
= __ARCH_SPIN_LOCK_UNLOCKED
;
122 if (!reorder_access
|| !IS_ENABLED(CONFIG_SMP
))
125 #define __KCSAN_CHECK_BARRIER(access_type, barrier, name) \
127 reorder_access->type = (access_type) | KCSAN_ACCESS_SCOPED; \
128 reorder_access->size = 1; \
130 if (reorder_access->size != 0) { \
131 pr_err("improperly instrumented type=(" #access_type "): " name "\n"); \
135 #define KCSAN_CHECK_READ_BARRIER(b) __KCSAN_CHECK_BARRIER(0, b, #b)
136 #define KCSAN_CHECK_WRITE_BARRIER(b) __KCSAN_CHECK_BARRIER(KCSAN_ACCESS_WRITE, b, #b)
137 #define KCSAN_CHECK_RW_BARRIER(b) __KCSAN_CHECK_BARRIER(KCSAN_ACCESS_WRITE | KCSAN_ACCESS_COMPOUND, b, #b)
139 kcsan_nestable_atomic_begin(); /* No watchpoints in called functions. */
141 KCSAN_CHECK_READ_BARRIER(mb());
142 KCSAN_CHECK_READ_BARRIER(rmb());
143 KCSAN_CHECK_READ_BARRIER(smp_mb());
144 KCSAN_CHECK_READ_BARRIER(smp_rmb());
145 KCSAN_CHECK_READ_BARRIER(dma_rmb());
146 KCSAN_CHECK_READ_BARRIER(smp_mb__before_atomic());
147 KCSAN_CHECK_READ_BARRIER(smp_mb__after_atomic());
148 KCSAN_CHECK_READ_BARRIER(smp_mb__after_spinlock());
149 KCSAN_CHECK_READ_BARRIER(smp_store_mb(test_var
, 0));
150 KCSAN_CHECK_READ_BARRIER(smp_store_release(&test_var
, 0));
151 KCSAN_CHECK_READ_BARRIER(xchg(&test_var
, 0));
152 KCSAN_CHECK_READ_BARRIER(xchg_release(&test_var
, 0));
153 KCSAN_CHECK_READ_BARRIER(cmpxchg(&test_var
, 0, 0));
154 KCSAN_CHECK_READ_BARRIER(cmpxchg_release(&test_var
, 0, 0));
155 KCSAN_CHECK_READ_BARRIER(atomic_set_release(&dummy
, 0));
156 KCSAN_CHECK_READ_BARRIER(atomic_add_return(1, &dummy
));
157 KCSAN_CHECK_READ_BARRIER(atomic_add_return_release(1, &dummy
));
158 KCSAN_CHECK_READ_BARRIER(atomic_fetch_add(1, &dummy
));
159 KCSAN_CHECK_READ_BARRIER(atomic_fetch_add_release(1, &dummy
));
160 KCSAN_CHECK_READ_BARRIER(test_and_set_bit(0, &test_var
));
161 KCSAN_CHECK_READ_BARRIER(test_and_clear_bit(0, &test_var
));
162 KCSAN_CHECK_READ_BARRIER(test_and_change_bit(0, &test_var
));
163 KCSAN_CHECK_READ_BARRIER(clear_bit_unlock(0, &test_var
));
164 KCSAN_CHECK_READ_BARRIER(__clear_bit_unlock(0, &test_var
));
165 arch_spin_lock(&arch_spinlock
);
166 KCSAN_CHECK_READ_BARRIER(arch_spin_unlock(&arch_spinlock
));
167 spin_lock(&test_spinlock
);
168 KCSAN_CHECK_READ_BARRIER(spin_unlock(&test_spinlock
));
170 KCSAN_CHECK_WRITE_BARRIER(mb());
171 KCSAN_CHECK_WRITE_BARRIER(wmb());
172 KCSAN_CHECK_WRITE_BARRIER(smp_mb());
173 KCSAN_CHECK_WRITE_BARRIER(smp_wmb());
174 KCSAN_CHECK_WRITE_BARRIER(dma_wmb());
175 KCSAN_CHECK_WRITE_BARRIER(smp_mb__before_atomic());
176 KCSAN_CHECK_WRITE_BARRIER(smp_mb__after_atomic());
177 KCSAN_CHECK_WRITE_BARRIER(smp_mb__after_spinlock());
178 KCSAN_CHECK_WRITE_BARRIER(smp_store_mb(test_var
, 0));
179 KCSAN_CHECK_WRITE_BARRIER(smp_store_release(&test_var
, 0));
180 KCSAN_CHECK_WRITE_BARRIER(xchg(&test_var
, 0));
181 KCSAN_CHECK_WRITE_BARRIER(xchg_release(&test_var
, 0));
182 KCSAN_CHECK_WRITE_BARRIER(cmpxchg(&test_var
, 0, 0));
183 KCSAN_CHECK_WRITE_BARRIER(cmpxchg_release(&test_var
, 0, 0));
184 KCSAN_CHECK_WRITE_BARRIER(atomic_set_release(&dummy
, 0));
185 KCSAN_CHECK_WRITE_BARRIER(atomic_add_return(1, &dummy
));
186 KCSAN_CHECK_WRITE_BARRIER(atomic_add_return_release(1, &dummy
));
187 KCSAN_CHECK_WRITE_BARRIER(atomic_fetch_add(1, &dummy
));
188 KCSAN_CHECK_WRITE_BARRIER(atomic_fetch_add_release(1, &dummy
));
189 KCSAN_CHECK_WRITE_BARRIER(test_and_set_bit(0, &test_var
));
190 KCSAN_CHECK_WRITE_BARRIER(test_and_clear_bit(0, &test_var
));
191 KCSAN_CHECK_WRITE_BARRIER(test_and_change_bit(0, &test_var
));
192 KCSAN_CHECK_WRITE_BARRIER(clear_bit_unlock(0, &test_var
));
193 KCSAN_CHECK_WRITE_BARRIER(__clear_bit_unlock(0, &test_var
));
194 arch_spin_lock(&arch_spinlock
);
195 KCSAN_CHECK_WRITE_BARRIER(arch_spin_unlock(&arch_spinlock
));
196 spin_lock(&test_spinlock
);
197 KCSAN_CHECK_WRITE_BARRIER(spin_unlock(&test_spinlock
));
199 KCSAN_CHECK_RW_BARRIER(mb());
200 KCSAN_CHECK_RW_BARRIER(wmb());
201 KCSAN_CHECK_RW_BARRIER(rmb());
202 KCSAN_CHECK_RW_BARRIER(smp_mb());
203 KCSAN_CHECK_RW_BARRIER(smp_wmb());
204 KCSAN_CHECK_RW_BARRIER(smp_rmb());
205 KCSAN_CHECK_RW_BARRIER(dma_wmb());
206 KCSAN_CHECK_RW_BARRIER(dma_rmb());
207 KCSAN_CHECK_RW_BARRIER(smp_mb__before_atomic());
208 KCSAN_CHECK_RW_BARRIER(smp_mb__after_atomic());
209 KCSAN_CHECK_RW_BARRIER(smp_mb__after_spinlock());
210 KCSAN_CHECK_RW_BARRIER(smp_store_mb(test_var
, 0));
211 KCSAN_CHECK_RW_BARRIER(smp_store_release(&test_var
, 0));
212 KCSAN_CHECK_RW_BARRIER(xchg(&test_var
, 0));
213 KCSAN_CHECK_RW_BARRIER(xchg_release(&test_var
, 0));
214 KCSAN_CHECK_RW_BARRIER(cmpxchg(&test_var
, 0, 0));
215 KCSAN_CHECK_RW_BARRIER(cmpxchg_release(&test_var
, 0, 0));
216 KCSAN_CHECK_RW_BARRIER(atomic_set_release(&dummy
, 0));
217 KCSAN_CHECK_RW_BARRIER(atomic_add_return(1, &dummy
));
218 KCSAN_CHECK_RW_BARRIER(atomic_add_return_release(1, &dummy
));
219 KCSAN_CHECK_RW_BARRIER(atomic_fetch_add(1, &dummy
));
220 KCSAN_CHECK_RW_BARRIER(atomic_fetch_add_release(1, &dummy
));
221 KCSAN_CHECK_RW_BARRIER(test_and_set_bit(0, &test_var
));
222 KCSAN_CHECK_RW_BARRIER(test_and_clear_bit(0, &test_var
));
223 KCSAN_CHECK_RW_BARRIER(test_and_change_bit(0, &test_var
));
224 KCSAN_CHECK_RW_BARRIER(clear_bit_unlock(0, &test_var
));
225 KCSAN_CHECK_RW_BARRIER(__clear_bit_unlock(0, &test_var
));
226 arch_spin_lock(&arch_spinlock
);
227 KCSAN_CHECK_RW_BARRIER(arch_spin_unlock(&arch_spinlock
));
228 spin_lock(&test_spinlock
);
229 KCSAN_CHECK_RW_BARRIER(spin_unlock(&test_spinlock
));
230 KCSAN_CHECK_RW_BARRIER(xor_unlock_is_negative_byte(1, &test_var
));
231 KCSAN_CHECK_READ_BARRIER(xor_unlock_is_negative_byte(1, &test_var
));
232 KCSAN_CHECK_WRITE_BARRIER(xor_unlock_is_negative_byte(1, &test_var
));
233 kcsan_nestable_atomic_end();
238 static int __init
kcsan_selftest(void)
243 #define RUN_TEST(do_test) \
249 pr_err("selftest: " #do_test " failed"); \
252 RUN_TEST(test_encode_decode
);
253 RUN_TEST(test_matching_access
);
254 RUN_TEST(test_barrier
);
256 pr_info("selftest: %d/%d tests passed\n", passed
, total
);
258 panic("selftests failed");
261 postcore_initcall(kcsan_selftest
);