2 * Copyright (C) 2002 Jeff Dike (jdike@karaya.com)
3 * Licensed under the GPL
6 #include "linux/sched.h"
7 #include "linux/signal.h"
8 #include "linux/kernel.h"
9 #include "linux/interrupt.h"
10 #include "linux/ptrace.h"
11 #include "asm/system.h"
12 #include "asm/pgalloc.h"
13 #include "asm/ptrace.h"
14 #include "asm/tlbflush.h"
16 #include "kern_util.h"
17 #include "user_util.h"
20 #include "sigcontext.h"
24 #include "mode_kern.h"
28 void switch_to_tt(void *prev
, void *next
)
30 struct task_struct
*from
, *to
, *prev_sched
;
32 int err
, vtalrm
, alrm
, prof
, cpu
;
38 cpu
= task_thread_info(from
)->cpu
;
40 forward_interrupts(to
->thread
.mode
.tt
.extern_pid
);
42 forward_ipi(cpu_data
[cpu
].ipi_pipe
[0], to
->thread
.mode
.tt
.extern_pid
);
44 local_irq_save(flags
);
46 vtalrm
= change_sig(SIGVTALRM
, 0);
47 alrm
= change_sig(SIGALRM
, 0);
48 prof
= change_sig(SIGPROF
, 0);
50 forward_pending_sigio(to
->thread
.mode
.tt
.extern_pid
);
54 /* Notice that here we "up" the semaphore on which "to" is waiting, and
55 * below (the read) we wait on this semaphore (which is implemented by
56 * switch_pipe) and go sleeping. Thus, after that, we have resumed in
57 * "to", and can't use any more the value of "from" (which is outdated),
58 * nor the value in "to" (since it was the task which stole us the CPU,
59 * which we don't care about). */
61 err
= os_write_file(to
->thread
.mode
.tt
.switch_pipe
[1], &c
, sizeof(c
));
63 panic("write of switch_pipe failed, err = %d", -err
);
65 if(from
->thread
.mode
.tt
.switch_pipe
[0] == -1)
66 os_kill_process(os_getpid(), 0);
68 err
= os_read_file(from
->thread
.mode
.tt
.switch_pipe
[0], &c
, sizeof(c
));
70 panic("read of switch_pipe failed, errno = %d", -err
);
72 /* If the process that we have just scheduled away from has exited,
73 * then it needs to be killed here. The reason is that, even though
74 * it will kill itself when it next runs, that may be too late. Its
75 * stack will be freed, possibly before then, and if that happens,
76 * we have a use-after-free situation. So, it gets killed here
77 * in case it has not already killed itself.
79 prev_sched
= current
->thread
.prev_sched
;
80 if(prev_sched
->thread
.mode
.tt
.switch_pipe
[0] == -1)
81 os_kill_process(prev_sched
->thread
.mode
.tt
.extern_pid
, 1);
83 change_sig(SIGVTALRM
, vtalrm
);
84 change_sig(SIGALRM
, alrm
);
85 change_sig(SIGPROF
, prof
);
87 arch_switch_to_tt(prev_sched
, current
);
90 local_irq_restore(flags
);
93 void release_thread_tt(struct task_struct
*task
)
95 int pid
= task
->thread
.mode
.tt
.extern_pid
;
98 * We first have to kill the other process, before
99 * closing its switch_pipe. Else it might wake up
100 * and receive "EOF" before we could kill it.
102 if(os_getpid() != pid
)
103 os_kill_process(pid
, 0);
105 os_close_file(task
->thread
.mode
.tt
.switch_pipe
[0]);
106 os_close_file(task
->thread
.mode
.tt
.switch_pipe
[1]);
107 /* use switch_pipe as flag: thread is released */
108 task
->thread
.mode
.tt
.switch_pipe
[0] = -1;
111 void suspend_new_thread(int fd
)
116 os_stop_process(os_getpid());
117 err
= os_read_file(fd
, &c
, sizeof(c
));
119 panic("read failed in suspend_new_thread, err = %d", -err
);
122 void schedule_tail(task_t
*prev
);
124 static void new_thread_handler(int sig
)
126 unsigned long disable
;
130 fn
= current
->thread
.request
.u
.thread
.proc
;
131 arg
= current
->thread
.request
.u
.thread
.arg
;
133 UPT_SC(¤t
->thread
.regs
.regs
) = (void *) (&sig
+ 1);
134 disable
= (1 << (SIGVTALRM
- 1)) | (1 << (SIGALRM
- 1)) |
135 (1 << (SIGIO
- 1)) | (1 << (SIGPROF
- 1));
136 SC_SIGMASK(UPT_SC(¤t
->thread
.regs
.regs
)) &= ~disable
;
138 suspend_new_thread(current
->thread
.mode
.tt
.switch_pipe
[0]);
141 if(current
->thread
.prev_sched
!= NULL
)
142 schedule_tail(current
->thread
.prev_sched
);
143 current
->thread
.prev_sched
= NULL
;
145 init_new_thread_signals(1);
147 free_page(current
->thread
.temp_stack
);
148 set_cmdline("(kernel thread)");
150 change_sig(SIGUSR1
, 1);
151 change_sig(SIGPROF
, 1);
153 if(!run_kernel_thread(fn
, arg
, ¤t
->thread
.exec_buf
))
156 /* XXX No set_user_mode here because a newly execed process will
157 * immediately segfault on its non-existent IP, coming straight back
158 * to the signal handler, which will call set_user_mode on its way
159 * out. This should probably change since it's confusing.
163 static int new_thread_proc(void *stack
)
165 /* local_irq_disable is needed to block out signals until this thread is
166 * properly scheduled. Otherwise, the tracing thread will get mighty
167 * upset about any signals that arrive before that.
168 * This has the complication that it sets the saved signal mask in
169 * the sigcontext to block signals. This gets restored when this
170 * thread (or a descendant, since they get a copy of this sigcontext)
171 * returns to userspace.
172 * So, this is compensated for elsewhere.
173 * XXX There is still a small window until local_irq_disable() actually
174 * finishes where signals are possible - shouldn't be a problem in
175 * practice since SIGIO hasn't been forwarded here yet, and the
176 * local_irq_disable should finish before a SIGVTALRM has time to be
181 init_new_thread_stack(stack
, new_thread_handler
);
182 os_usr1_process(os_getpid());
183 change_sig(SIGUSR1
, 1);
187 /* Signal masking - signals are blocked at the start of fork_tramp. They
188 * are re-enabled when finish_fork_handler is entered by fork_tramp hitting
189 * itself with a SIGUSR1. set_user_mode has to be run with SIGUSR1 off,
190 * so it is blocked before it's called. They are re-enabled on sigreturn
191 * despite the fact that they were blocked when the SIGUSR1 was issued because
192 * copy_thread copies the parent's sigcontext, including the signal mask
193 * onto the signal frame.
196 void finish_fork_handler(int sig
)
198 UPT_SC(¤t
->thread
.regs
.regs
) = (void *) (&sig
+ 1);
199 suspend_new_thread(current
->thread
.mode
.tt
.switch_pipe
[0]);
202 if(current
->thread
.prev_sched
!= NULL
)
203 schedule_tail(current
->thread
.prev_sched
);
204 current
->thread
.prev_sched
= NULL
;
207 change_sig(SIGVTALRM
, 1);
209 if(current
->mm
!= current
->parent
->mm
)
210 protect_memory(uml_reserved
, high_physmem
- uml_reserved
, 1,
212 task_protections((unsigned long) current_thread
);
214 free_page(current
->thread
.temp_stack
);
216 change_sig(SIGUSR1
, 0);
217 set_user_mode(current
);
220 int fork_tramp(void *stack
)
224 init_new_thread_stack(stack
, finish_fork_handler
);
226 os_usr1_process(os_getpid());
227 change_sig(SIGUSR1
, 1);
231 int copy_thread_tt(int nr
, unsigned long clone_flags
, unsigned long sp
,
232 unsigned long stack_top
, struct task_struct
* p
,
233 struct pt_regs
*regs
)
235 int (*tramp
)(void *);
239 if(current
->thread
.forking
)
242 tramp
= new_thread_proc
;
243 p
->thread
.request
.u
.thread
= current
->thread
.request
.u
.thread
;
246 err
= os_pipe(p
->thread
.mode
.tt
.switch_pipe
, 1, 1);
248 printk("copy_thread : pipe failed, err = %d\n", -err
);
252 stack
= alloc_stack(0, 0);
254 printk(KERN_ERR
"copy_thread : failed to allocate "
255 "temporary stack\n");
259 clone_flags
&= CLONE_VM
;
260 p
->thread
.temp_stack
= stack
;
261 new_pid
= start_fork_tramp(task_stack_page(p
), stack
, clone_flags
, tramp
);
263 printk(KERN_ERR
"copy_thread : clone failed - errno = %d\n",
268 if(current
->thread
.forking
){
269 sc_to_sc(UPT_SC(&p
->thread
.regs
.regs
), UPT_SC(®s
->regs
));
270 SC_SET_SYSCALL_RETURN(UPT_SC(&p
->thread
.regs
.regs
), 0);
272 SC_SP(UPT_SC(&p
->thread
.regs
.regs
)) = sp
;
274 p
->thread
.mode
.tt
.extern_pid
= new_pid
;
276 current
->thread
.request
.op
= OP_FORK
;
277 current
->thread
.request
.u
.fork
.pid
= new_pid
;
278 os_usr1_process(os_getpid());
280 /* Enable the signal and then disable it to ensure that it is handled
281 * here, and nowhere else.
283 change_sig(SIGUSR1
, 1);
285 change_sig(SIGUSR1
, 0);
292 current
->thread
.request
.op
= OP_REBOOT
;
293 os_usr1_process(os_getpid());
294 change_sig(SIGUSR1
, 1);
299 current
->thread
.request
.op
= OP_HALT
;
300 os_usr1_process(os_getpid());
301 change_sig(SIGUSR1
, 1);
304 void kill_off_processes_tt(void)
306 struct task_struct
*p
;
311 if(p
->thread
.mode
.tt
.extern_pid
!= me
)
312 os_kill_process(p
->thread
.mode
.tt
.extern_pid
, 0);
314 if(init_task
.thread
.mode
.tt
.extern_pid
!= me
)
315 os_kill_process(init_task
.thread
.mode
.tt
.extern_pid
, 0);
318 void initial_thread_cb_tt(void (*proc
)(void *), void *arg
)
320 if(os_getpid() == tracing_pid
){
324 current
->thread
.request
.op
= OP_CB
;
325 current
->thread
.request
.u
.cb
.proc
= proc
;
326 current
->thread
.request
.u
.cb
.arg
= arg
;
327 os_usr1_process(os_getpid());
328 change_sig(SIGUSR1
, 1);
330 change_sig(SIGUSR1
, 0);
334 int do_proc_op(void *t
, int proc_id
)
336 struct task_struct
*task
;
337 struct thread_struct
*thread
;
341 thread
= &task
->thread
;
342 op
= thread
->request
.op
;
348 pid
= thread
->request
.u
.exec
.pid
;
349 do_exec(thread
->mode
.tt
.extern_pid
, pid
);
350 thread
->mode
.tt
.extern_pid
= pid
;
351 cpu_tasks
[task_thread_info(task
)->cpu
].pid
= pid
;
354 attach_process(thread
->request
.u
.fork
.pid
);
357 (*thread
->request
.u
.cb
.proc
)(thread
->request
.u
.cb
.arg
);
363 tracer_panic("Bad op in do_proc_op");
366 thread
->request
.op
= OP_NONE
;
370 void init_idle_tt(void)
375 extern void start_kernel(void);
377 static int start_kernel_proc(void *unused
)
384 cpu_tasks
[0].pid
= pid
;
385 cpu_tasks
[0].task
= current
;
387 cpu_online_map
= cpumask_of_cpu(0);
389 if(debug
) os_stop_process(pid
);
394 void set_tracing(void *task
, int tracing
)
396 ((struct task_struct
*) task
)->thread
.mode
.tt
.tracing
= tracing
;
399 int is_tracing(void *t
)
401 return (((struct task_struct
*) t
)->thread
.mode
.tt
.tracing
);
404 int set_user_mode(void *t
)
406 struct task_struct
*task
;
408 task
= t
? t
: current
;
409 if(task
->thread
.mode
.tt
.tracing
)
411 task
->thread
.request
.op
= OP_TRACE_ON
;
412 os_usr1_process(os_getpid());
416 void set_init_pid(int pid
)
420 init_task
.thread
.mode
.tt
.extern_pid
= pid
;
421 err
= os_pipe(init_task
.thread
.mode
.tt
.switch_pipe
, 1, 1);
423 panic("Can't create switch pipe for init_task, errno = %d",
427 int start_uml_tt(void)
432 pages
= (1 << CONFIG_KERNEL_STACK_ORDER
);
433 sp
= task_stack_page(&init_task
) +
434 pages
* PAGE_SIZE
- sizeof(unsigned long);
435 return(tracer(start_kernel_proc
, sp
));
438 int external_pid_tt(struct task_struct
*task
)
440 return(task
->thread
.mode
.tt
.extern_pid
);
443 int thread_pid_tt(struct task_struct
*task
)
445 return(task
->thread
.mode
.tt
.extern_pid
);
448 int is_valid_pid(int pid
)
450 struct task_struct
*task
;
452 read_lock(&tasklist_lock
);
453 for_each_process(task
){
454 if(task
->thread
.mode
.tt
.extern_pid
== pid
){
455 read_unlock(&tasklist_lock
);
459 read_unlock(&tasklist_lock
);