1 // SPDX-License-Identifier: GPL-2.0-or-later
3 * test_kprobes.c - simple sanity test for k*probes
5 * Copyright IBM Corp. 2008
8 #include <linux/kernel.h>
9 #include <linux/kprobes.h>
10 #include <linux/random.h>
11 #include <kunit/test.h>
15 static u32 rand1
, preh_val
, posth_val
;
16 static u32 (*target
)(u32 value
);
17 static u32 (*recursed_target
)(u32 value
);
18 static u32 (*target2
)(u32 value
);
19 static struct kunit
*current_test
;
21 static unsigned long (*internal_target
)(void);
22 static unsigned long (*stacktrace_target
)(void);
23 static unsigned long (*stacktrace_driver
)(void);
24 static unsigned long target_return_address
[2];
26 static noinline u32
kprobe_target(u32 value
)
28 return (value
/ div_factor
);
31 static noinline u32
kprobe_recursed_target(u32 value
)
33 return (value
/ div_factor
);
36 static int kp_pre_handler(struct kprobe
*p
, struct pt_regs
*regs
)
38 KUNIT_EXPECT_FALSE(current_test
, preemptible());
40 preh_val
= recursed_target(rand1
);
44 static void kp_post_handler(struct kprobe
*p
, struct pt_regs
*regs
,
47 u32 expval
= recursed_target(rand1
);
49 KUNIT_EXPECT_FALSE(current_test
, preemptible());
50 KUNIT_EXPECT_EQ(current_test
, preh_val
, expval
);
52 posth_val
= preh_val
+ div_factor
;
55 static struct kprobe kp
= {
56 .symbol_name
= "kprobe_target",
57 .pre_handler
= kp_pre_handler
,
58 .post_handler
= kp_post_handler
61 static void test_kprobe(struct kunit
*test
)
64 KUNIT_EXPECT_EQ(test
, 0, register_kprobe(&kp
));
66 unregister_kprobe(&kp
);
67 KUNIT_EXPECT_NE(test
, 0, preh_val
);
68 KUNIT_EXPECT_NE(test
, 0, posth_val
);
71 static noinline u32
kprobe_target2(u32 value
)
73 return (value
/ div_factor
) + 1;
76 static noinline
unsigned long kprobe_stacktrace_internal_target(void)
78 if (!target_return_address
[0])
79 target_return_address
[0] = (unsigned long)__builtin_return_address(0);
80 return target_return_address
[0];
83 static noinline
unsigned long kprobe_stacktrace_target(void)
85 if (!target_return_address
[1])
86 target_return_address
[1] = (unsigned long)__builtin_return_address(0);
91 return target_return_address
[1];
94 static noinline
unsigned long kprobe_stacktrace_driver(void)
96 if (stacktrace_target
)
99 /* This is for preventing inlining the function */
100 return (unsigned long)__builtin_return_address(0);
103 static int kp_pre_handler2(struct kprobe
*p
, struct pt_regs
*regs
)
105 preh_val
= (rand1
/ div_factor
) + 1;
109 static void kp_post_handler2(struct kprobe
*p
, struct pt_regs
*regs
,
112 KUNIT_EXPECT_EQ(current_test
, preh_val
, (rand1
/ div_factor
) + 1);
113 posth_val
= preh_val
+ div_factor
;
116 static struct kprobe kp2
= {
117 .symbol_name
= "kprobe_target2",
118 .pre_handler
= kp_pre_handler2
,
119 .post_handler
= kp_post_handler2
122 static void test_kprobes(struct kunit
*test
)
124 struct kprobe
*kps
[2] = {&kp
, &kp2
};
128 /* addr and flags should be cleard for reusing kprobe. */
132 KUNIT_EXPECT_EQ(test
, 0, register_kprobes(kps
, 2));
137 KUNIT_EXPECT_NE(test
, 0, preh_val
);
138 KUNIT_EXPECT_NE(test
, 0, posth_val
);
144 KUNIT_EXPECT_NE(test
, 0, preh_val
);
145 KUNIT_EXPECT_NE(test
, 0, posth_val
);
146 unregister_kprobes(kps
, 2);
149 static struct kprobe kp_missed
= {
150 .symbol_name
= "kprobe_recursed_target",
151 .pre_handler
= kp_pre_handler
,
152 .post_handler
= kp_post_handler
,
155 static void test_kprobe_missed(struct kunit
*test
)
161 KUNIT_EXPECT_EQ(test
, 0, register_kprobe(&kp_missed
));
163 recursed_target(rand1
);
165 KUNIT_EXPECT_EQ(test
, 2, kp_missed
.nmissed
);
166 KUNIT_EXPECT_NE(test
, 0, preh_val
);
167 KUNIT_EXPECT_NE(test
, 0, posth_val
);
169 unregister_kprobe(&kp_missed
);
172 #ifdef CONFIG_KRETPROBES
175 static int entry_handler(struct kretprobe_instance
*ri
, struct pt_regs
*regs
)
177 KUNIT_EXPECT_FALSE(current_test
, preemptible());
178 krph_val
= (rand1
/ div_factor
);
182 static int return_handler(struct kretprobe_instance
*ri
, struct pt_regs
*regs
)
184 unsigned long ret
= regs_return_value(regs
);
186 KUNIT_EXPECT_FALSE(current_test
, preemptible());
187 KUNIT_EXPECT_EQ(current_test
, ret
, rand1
/ div_factor
);
188 KUNIT_EXPECT_NE(current_test
, krph_val
, 0);
193 static struct kretprobe rp
= {
194 .handler
= return_handler
,
195 .entry_handler
= entry_handler
,
196 .kp
.symbol_name
= "kprobe_target"
199 static void test_kretprobe(struct kunit
*test
)
202 KUNIT_EXPECT_EQ(test
, 0, register_kretprobe(&rp
));
204 unregister_kretprobe(&rp
);
205 KUNIT_EXPECT_EQ(test
, krph_val
, rand1
);
208 static int return_handler2(struct kretprobe_instance
*ri
, struct pt_regs
*regs
)
210 unsigned long ret
= regs_return_value(regs
);
212 KUNIT_EXPECT_EQ(current_test
, ret
, (rand1
/ div_factor
) + 1);
213 KUNIT_EXPECT_NE(current_test
, krph_val
, 0);
218 static struct kretprobe rp2
= {
219 .handler
= return_handler2
,
220 .entry_handler
= entry_handler
,
221 .kp
.symbol_name
= "kprobe_target2"
224 static void test_kretprobes(struct kunit
*test
)
226 struct kretprobe
*rps
[2] = {&rp
, &rp2
};
229 /* addr and flags should be cleard for reusing kprobe. */
232 KUNIT_EXPECT_EQ(test
, 0, register_kretprobes(rps
, 2));
236 KUNIT_EXPECT_EQ(test
, krph_val
, rand1
);
240 KUNIT_EXPECT_EQ(test
, krph_val
, rand1
);
241 unregister_kretprobes(rps
, 2);
244 #ifdef CONFIG_ARCH_CORRECT_STACKTRACE_ON_KRETPROBE
245 #define STACK_BUF_SIZE 16
246 static unsigned long stack_buf
[STACK_BUF_SIZE
];
248 static int stacktrace_return_handler(struct kretprobe_instance
*ri
, struct pt_regs
*regs
)
250 unsigned long retval
= regs_return_value(regs
);
253 KUNIT_EXPECT_FALSE(current_test
, preemptible());
254 KUNIT_EXPECT_EQ(current_test
, retval
, target_return_address
[1]);
257 * Test stacktrace inside the kretprobe handler, this will involves
258 * kretprobe trampoline, but must include correct return address
259 * of the target function.
261 ret
= stack_trace_save(stack_buf
, STACK_BUF_SIZE
, 0);
262 KUNIT_EXPECT_NE(current_test
, ret
, 0);
264 for (i
= 0; i
< ret
; i
++) {
265 if (stack_buf
[i
] == target_return_address
[1])
268 KUNIT_EXPECT_NE(current_test
, i
, ret
);
270 #if !IS_MODULE(CONFIG_KPROBES_SANITY_TEST)
272 * Test stacktrace from pt_regs at the return address. Thus the stack
273 * trace must start from the target return address.
275 ret
= stack_trace_save_regs(regs
, stack_buf
, STACK_BUF_SIZE
, 0);
276 KUNIT_EXPECT_NE(current_test
, ret
, 0);
277 KUNIT_EXPECT_EQ(current_test
, stack_buf
[0], target_return_address
[1]);
283 static struct kretprobe rp3
= {
284 .handler
= stacktrace_return_handler
,
285 .kp
.symbol_name
= "kprobe_stacktrace_target"
288 static void test_stacktrace_on_kretprobe(struct kunit
*test
)
290 unsigned long myretaddr
= (unsigned long)__builtin_return_address(0);
297 * Run the stacktrace_driver() to record correct return address in
298 * stacktrace_target() and ensure stacktrace_driver() call is not
299 * inlined by checking the return address of stacktrace_driver()
300 * and the return address of this function is different.
302 KUNIT_ASSERT_NE(test
, myretaddr
, stacktrace_driver());
304 KUNIT_ASSERT_EQ(test
, 0, register_kretprobe(&rp3
));
305 KUNIT_ASSERT_NE(test
, myretaddr
, stacktrace_driver());
306 unregister_kretprobe(&rp3
);
309 static int stacktrace_internal_return_handler(struct kretprobe_instance
*ri
, struct pt_regs
*regs
)
311 unsigned long retval
= regs_return_value(regs
);
314 KUNIT_EXPECT_FALSE(current_test
, preemptible());
315 KUNIT_EXPECT_EQ(current_test
, retval
, target_return_address
[0]);
318 * Test stacktrace inside the kretprobe handler for nested case.
319 * The unwinder will find the kretprobe_trampoline address on the
320 * return address, and kretprobe must solve that.
322 ret
= stack_trace_save(stack_buf
, STACK_BUF_SIZE
, 0);
323 KUNIT_EXPECT_NE(current_test
, ret
, 0);
325 for (i
= 0; i
< ret
- 1; i
++) {
326 if (stack_buf
[i
] == target_return_address
[0]) {
327 KUNIT_EXPECT_EQ(current_test
, stack_buf
[i
+ 1], target_return_address
[1]);
331 KUNIT_EXPECT_NE(current_test
, i
, ret
);
333 #if !IS_MODULE(CONFIG_KPROBES_SANITY_TEST)
334 /* Ditto for the regs version. */
335 ret
= stack_trace_save_regs(regs
, stack_buf
, STACK_BUF_SIZE
, 0);
336 KUNIT_EXPECT_NE(current_test
, ret
, 0);
337 KUNIT_EXPECT_EQ(current_test
, stack_buf
[0], target_return_address
[0]);
338 KUNIT_EXPECT_EQ(current_test
, stack_buf
[1], target_return_address
[1]);
344 static struct kretprobe rp4
= {
345 .handler
= stacktrace_internal_return_handler
,
346 .kp
.symbol_name
= "kprobe_stacktrace_internal_target"
349 static void test_stacktrace_on_nested_kretprobe(struct kunit
*test
)
351 unsigned long myretaddr
= (unsigned long)__builtin_return_address(0);
352 struct kretprobe
*rps
[2] = {&rp3
, &rp4
};
358 //KUNIT_ASSERT_NE(test, myretaddr, stacktrace_driver());
360 KUNIT_ASSERT_EQ(test
, 0, register_kretprobes(rps
, 2));
361 KUNIT_ASSERT_NE(test
, myretaddr
, stacktrace_driver());
362 unregister_kretprobes(rps
, 2);
364 #endif /* CONFIG_ARCH_CORRECT_STACKTRACE_ON_KRETPROBE */
366 #endif /* CONFIG_KRETPROBES */
368 static int kprobes_test_init(struct kunit
*test
)
370 target
= kprobe_target
;
371 target2
= kprobe_target2
;
372 recursed_target
= kprobe_recursed_target
;
373 stacktrace_target
= kprobe_stacktrace_target
;
374 internal_target
= kprobe_stacktrace_internal_target
;
375 stacktrace_driver
= kprobe_stacktrace_driver
;
376 rand1
= get_random_u32_above(div_factor
);
380 static struct kunit_case kprobes_testcases
[] = {
381 KUNIT_CASE(test_kprobe
),
382 KUNIT_CASE(test_kprobes
),
383 KUNIT_CASE(test_kprobe_missed
),
384 #ifdef CONFIG_KRETPROBES
385 KUNIT_CASE(test_kretprobe
),
386 KUNIT_CASE(test_kretprobes
),
387 #ifdef CONFIG_ARCH_CORRECT_STACKTRACE_ON_KRETPROBE
388 KUNIT_CASE(test_stacktrace_on_kretprobe
),
389 KUNIT_CASE(test_stacktrace_on_nested_kretprobe
),
395 static struct kunit_suite kprobes_test_suite
= {
396 .name
= "kprobes_test",
397 .init
= kprobes_test_init
,
398 .test_cases
= kprobes_testcases
,
401 kunit_test_suites(&kprobes_test_suite
);
403 MODULE_DESCRIPTION("simple sanity test for k*probes");
404 MODULE_LICENSE("GPL");