2 * SuperH process tracing
4 * Copyright (C) 1999, 2000 Kaz Kojima & Niibe Yutaka
5 * Copyright (C) 2002 - 2009 Paul Mundt
7 * Audit support by Yuichi Nakamura <ynakam@hitachisoft.jp>
9 * This file is subject to the terms and conditions of the GNU General Public
10 * License. See the file "COPYING" in the main directory of this archive
13 #include <linux/kernel.h>
14 #include <linux/sched.h>
16 #include <linux/smp.h>
17 #include <linux/errno.h>
18 #include <linux/ptrace.h>
19 #include <linux/user.h>
20 #include <linux/security.h>
21 #include <linux/signal.h>
23 #include <linux/audit.h>
24 #include <linux/seccomp.h>
25 #include <linux/tracehook.h>
26 #include <linux/elf.h>
27 #include <linux/regset.h>
28 #include <linux/hw_breakpoint.h>
29 #include <asm/uaccess.h>
30 #include <asm/pgtable.h>
31 #include <asm/system.h>
32 #include <asm/processor.h>
33 #include <asm/mmu_context.h>
34 #include <asm/syscalls.h>
37 #define CREATE_TRACE_POINTS
38 #include <trace/events/syscalls.h>
41 * This routine will get a word off of the process kernel stack.
43 static inline int get_stack_long(struct task_struct
*task
, int offset
)
47 stack
= (unsigned char *)task_pt_regs(task
);
49 return (*((int *)stack
));
53 * This routine will put a word on the process kernel stack.
55 static inline int put_stack_long(struct task_struct
*task
, int offset
,
60 stack
= (unsigned char *)task_pt_regs(task
);
62 *(unsigned long *) stack
= data
;
66 void ptrace_triggered(struct perf_event
*bp
,
67 struct perf_sample_data
*data
, struct pt_regs
*regs
)
69 struct perf_event_attr attr
;
72 * Disable the breakpoint request here since ptrace has defined a
73 * one-shot behaviour for breakpoint exceptions.
77 modify_user_hw_breakpoint(bp
, &attr
);
80 static int set_single_step(struct task_struct
*tsk
, unsigned long addr
)
82 struct thread_struct
*thread
= &tsk
->thread
;
83 struct perf_event
*bp
;
84 struct perf_event_attr attr
;
86 bp
= thread
->ptrace_bps
[0];
88 ptrace_breakpoint_init(&attr
);
91 attr
.bp_len
= HW_BREAKPOINT_LEN_2
;
92 attr
.bp_type
= HW_BREAKPOINT_R
;
94 bp
= register_user_hw_breakpoint(&attr
, ptrace_triggered
,
99 thread
->ptrace_bps
[0] = bp
;
105 /* reenable breakpoint */
106 attr
.disabled
= false;
107 err
= modify_user_hw_breakpoint(bp
, &attr
);
115 void user_enable_single_step(struct task_struct
*child
)
117 unsigned long pc
= get_stack_long(child
, offsetof(struct pt_regs
, pc
));
119 set_tsk_thread_flag(child
, TIF_SINGLESTEP
);
121 if (ptrace_get_breakpoints(child
) < 0)
124 set_single_step(child
, pc
);
125 ptrace_put_breakpoints(child
);
128 void user_disable_single_step(struct task_struct
*child
)
130 clear_tsk_thread_flag(child
, TIF_SINGLESTEP
);
134 * Called by kernel/ptrace.c when detaching..
136 * Make sure single step bits etc are not set.
138 void ptrace_disable(struct task_struct
*child
)
140 user_disable_single_step(child
);
143 static int genregs_get(struct task_struct
*target
,
144 const struct user_regset
*regset
,
145 unsigned int pos
, unsigned int count
,
146 void *kbuf
, void __user
*ubuf
)
148 const struct pt_regs
*regs
= task_pt_regs(target
);
151 ret
= user_regset_copyout(&pos
, &count
, &kbuf
, &ubuf
,
153 0, 16 * sizeof(unsigned long));
155 /* PC, PR, SR, GBR, MACH, MACL, TRA */
156 ret
= user_regset_copyout(&pos
, &count
, &kbuf
, &ubuf
,
158 offsetof(struct pt_regs
, pc
),
159 sizeof(struct pt_regs
));
161 ret
= user_regset_copyout_zero(&pos
, &count
, &kbuf
, &ubuf
,
162 sizeof(struct pt_regs
), -1);
167 static int genregs_set(struct task_struct
*target
,
168 const struct user_regset
*regset
,
169 unsigned int pos
, unsigned int count
,
170 const void *kbuf
, const void __user
*ubuf
)
172 struct pt_regs
*regs
= task_pt_regs(target
);
175 ret
= user_regset_copyin(&pos
, &count
, &kbuf
, &ubuf
,
177 0, 16 * sizeof(unsigned long));
178 if (!ret
&& count
> 0)
179 ret
= user_regset_copyin(&pos
, &count
, &kbuf
, &ubuf
,
181 offsetof(struct pt_regs
, pc
),
182 sizeof(struct pt_regs
));
184 ret
= user_regset_copyin_ignore(&pos
, &count
, &kbuf
, &ubuf
,
185 sizeof(struct pt_regs
), -1);
191 int fpregs_get(struct task_struct
*target
,
192 const struct user_regset
*regset
,
193 unsigned int pos
, unsigned int count
,
194 void *kbuf
, void __user
*ubuf
)
198 ret
= init_fpu(target
);
202 if ((boot_cpu_data
.flags
& CPU_HAS_FPU
))
203 return user_regset_copyout(&pos
, &count
, &kbuf
, &ubuf
,
204 &target
->thread
.xstate
->hardfpu
, 0, -1);
206 return user_regset_copyout(&pos
, &count
, &kbuf
, &ubuf
,
207 &target
->thread
.xstate
->softfpu
, 0, -1);
210 static int fpregs_set(struct task_struct
*target
,
211 const struct user_regset
*regset
,
212 unsigned int pos
, unsigned int count
,
213 const void *kbuf
, const void __user
*ubuf
)
217 ret
= init_fpu(target
);
221 set_stopped_child_used_math(target
);
223 if ((boot_cpu_data
.flags
& CPU_HAS_FPU
))
224 return user_regset_copyin(&pos
, &count
, &kbuf
, &ubuf
,
225 &target
->thread
.xstate
->hardfpu
, 0, -1);
227 return user_regset_copyin(&pos
, &count
, &kbuf
, &ubuf
,
228 &target
->thread
.xstate
->softfpu
, 0, -1);
231 static int fpregs_active(struct task_struct
*target
,
232 const struct user_regset
*regset
)
234 return tsk_used_math(target
) ? regset
->n
: 0;
239 static int dspregs_get(struct task_struct
*target
,
240 const struct user_regset
*regset
,
241 unsigned int pos
, unsigned int count
,
242 void *kbuf
, void __user
*ubuf
)
244 const struct pt_dspregs
*regs
=
245 (struct pt_dspregs
*)&target
->thread
.dsp_status
.dsp_regs
;
248 ret
= user_regset_copyout(&pos
, &count
, &kbuf
, &ubuf
, regs
,
249 0, sizeof(struct pt_dspregs
));
251 ret
= user_regset_copyout_zero(&pos
, &count
, &kbuf
, &ubuf
,
252 sizeof(struct pt_dspregs
), -1);
257 static int dspregs_set(struct task_struct
*target
,
258 const struct user_regset
*regset
,
259 unsigned int pos
, unsigned int count
,
260 const void *kbuf
, const void __user
*ubuf
)
262 struct pt_dspregs
*regs
=
263 (struct pt_dspregs
*)&target
->thread
.dsp_status
.dsp_regs
;
266 ret
= user_regset_copyin(&pos
, &count
, &kbuf
, &ubuf
, regs
,
267 0, sizeof(struct pt_dspregs
));
269 ret
= user_regset_copyin_ignore(&pos
, &count
, &kbuf
, &ubuf
,
270 sizeof(struct pt_dspregs
), -1);
275 static int dspregs_active(struct task_struct
*target
,
276 const struct user_regset
*regset
)
278 struct pt_regs
*regs
= task_pt_regs(target
);
280 return regs
->sr
& SR_DSP
? regset
->n
: 0;
284 const struct pt_regs_offset regoffset_table
[] = {
295 REGS_OFFSET_NAME(10),
296 REGS_OFFSET_NAME(11),
297 REGS_OFFSET_NAME(12),
298 REGS_OFFSET_NAME(13),
299 REGS_OFFSET_NAME(14),
300 REGS_OFFSET_NAME(15),
304 REG_OFFSET_NAME(gbr
),
305 REG_OFFSET_NAME(mach
),
306 REG_OFFSET_NAME(macl
),
307 REG_OFFSET_NAME(tra
),
312 * These are our native regset flavours.
324 static const struct user_regset sh_regsets
[] = {
328 * PC, PR, SR, GBR, MACH, MACL, TRA
331 .core_note_type
= NT_PRSTATUS
,
333 .size
= sizeof(long),
334 .align
= sizeof(long),
341 .core_note_type
= NT_PRFPREG
,
342 .n
= sizeof(struct user_fpu_struct
) / sizeof(long),
343 .size
= sizeof(long),
344 .align
= sizeof(long),
347 .active
= fpregs_active
,
353 .n
= sizeof(struct pt_dspregs
) / sizeof(long),
354 .size
= sizeof(long),
355 .align
= sizeof(long),
358 .active
= dspregs_active
,
363 static const struct user_regset_view user_sh_native_view
= {
366 .regsets
= sh_regsets
,
367 .n
= ARRAY_SIZE(sh_regsets
),
370 const struct user_regset_view
*task_user_regset_view(struct task_struct
*task
)
372 return &user_sh_native_view
;
375 long arch_ptrace(struct task_struct
*child
, long request
,
376 unsigned long addr
, unsigned long data
)
378 unsigned long __user
*datap
= (unsigned long __user
*)data
;
382 /* read the word at location addr in the USER area. */
383 case PTRACE_PEEKUSR
: {
387 if ((addr
& 3) || addr
< 0 ||
388 addr
> sizeof(struct user
) - 3)
391 if (addr
< sizeof(struct pt_regs
))
392 tmp
= get_stack_long(child
, addr
);
393 else if (addr
>= offsetof(struct user
, fpu
) &&
394 addr
< offsetof(struct user
, u_fpvalid
)) {
395 if (!tsk_used_math(child
)) {
396 if (addr
== offsetof(struct user
, fpu
.fpscr
))
402 ret
= init_fpu(child
);
405 index
= addr
- offsetof(struct user
, fpu
);
406 tmp
= ((unsigned long *)child
->thread
.xstate
)
409 } else if (addr
== offsetof(struct user
, u_fpvalid
))
410 tmp
= !!tsk_used_math(child
);
411 else if (addr
== PT_TEXT_ADDR
)
412 tmp
= child
->mm
->start_code
;
413 else if (addr
== PT_DATA_ADDR
)
414 tmp
= child
->mm
->start_data
;
415 else if (addr
== PT_TEXT_END_ADDR
)
416 tmp
= child
->mm
->end_code
;
417 else if (addr
== PT_TEXT_LEN
)
418 tmp
= child
->mm
->end_code
- child
->mm
->start_code
;
421 ret
= put_user(tmp
, datap
);
425 case PTRACE_POKEUSR
: /* write the word at location addr in the USER area */
427 if ((addr
& 3) || addr
< 0 ||
428 addr
> sizeof(struct user
) - 3)
431 if (addr
< sizeof(struct pt_regs
))
432 ret
= put_stack_long(child
, addr
, data
);
433 else if (addr
>= offsetof(struct user
, fpu
) &&
434 addr
< offsetof(struct user
, u_fpvalid
)) {
436 ret
= init_fpu(child
);
439 index
= addr
- offsetof(struct user
, fpu
);
440 set_stopped_child_used_math(child
);
441 ((unsigned long *)child
->thread
.xstate
)
444 } else if (addr
== offsetof(struct user
, u_fpvalid
)) {
445 conditional_stopped_child_used_math(data
, child
);
451 return copy_regset_to_user(child
, &user_sh_native_view
,
453 0, sizeof(struct pt_regs
),
456 return copy_regset_from_user(child
, &user_sh_native_view
,
458 0, sizeof(struct pt_regs
),
461 case PTRACE_GETFPREGS
:
462 return copy_regset_to_user(child
, &user_sh_native_view
,
464 0, sizeof(struct user_fpu_struct
),
466 case PTRACE_SETFPREGS
:
467 return copy_regset_from_user(child
, &user_sh_native_view
,
469 0, sizeof(struct user_fpu_struct
),
473 case PTRACE_GETDSPREGS
:
474 return copy_regset_to_user(child
, &user_sh_native_view
,
476 0, sizeof(struct pt_dspregs
),
478 case PTRACE_SETDSPREGS
:
479 return copy_regset_from_user(child
, &user_sh_native_view
,
481 0, sizeof(struct pt_dspregs
),
485 ret
= ptrace_request(child
, request
, addr
, data
);
492 static inline int audit_arch(void)
496 #ifdef CONFIG_CPU_LITTLE_ENDIAN
497 arch
|= __AUDIT_ARCH_LE
;
503 asmlinkage
long do_syscall_trace_enter(struct pt_regs
*regs
)
507 secure_computing(regs
->regs
[0]);
509 if (test_thread_flag(TIF_SYSCALL_TRACE
) &&
510 tracehook_report_syscall_entry(regs
))
512 * Tracing decided this syscall should not happen.
513 * We'll return a bogus call number to get an ENOSYS
514 * error, but leave the original number in regs->regs[0].
518 if (unlikely(test_thread_flag(TIF_SYSCALL_TRACEPOINT
)))
519 trace_sys_enter(regs
, regs
->regs
[0]);
521 if (unlikely(current
->audit_context
))
522 audit_syscall_entry(audit_arch(), regs
->regs
[3],
523 regs
->regs
[4], regs
->regs
[5],
524 regs
->regs
[6], regs
->regs
[7]);
526 return ret
?: regs
->regs
[0];
529 asmlinkage
void do_syscall_trace_leave(struct pt_regs
*regs
)
533 if (unlikely(current
->audit_context
))
534 audit_syscall_exit(AUDITSC_RESULT(regs
->regs
[0]),
537 if (unlikely(test_thread_flag(TIF_SYSCALL_TRACEPOINT
)))
538 trace_sys_exit(regs
, regs
->regs
[0]);
540 step
= test_thread_flag(TIF_SINGLESTEP
);
541 if (step
|| test_thread_flag(TIF_SYSCALL_TRACE
))
542 tracehook_report_syscall_exit(regs
, step
);