sched: make early bootup sched_clock() use safer
[wrt350n-kernel.git] / fs / binfmt_aout.c
bloba1bb2244cac772dd31dd8caff79b99fcead5ac3d
1 /*
2 * linux/fs/binfmt_aout.c
4 * Copyright (C) 1991, 1992, 1996 Linus Torvalds
5 */
7 #include <linux/module.h>
9 #include <linux/time.h>
10 #include <linux/kernel.h>
11 #include <linux/mm.h>
12 #include <linux/mman.h>
13 #include <linux/a.out.h>
14 #include <linux/errno.h>
15 #include <linux/signal.h>
16 #include <linux/string.h>
17 #include <linux/fs.h>
18 #include <linux/file.h>
19 #include <linux/stat.h>
20 #include <linux/fcntl.h>
21 #include <linux/ptrace.h>
22 #include <linux/user.h>
23 #include <linux/slab.h>
24 #include <linux/binfmts.h>
25 #include <linux/personality.h>
26 #include <linux/init.h>
28 #include <asm/system.h>
29 #include <asm/uaccess.h>
30 #include <asm/cacheflush.h>
31 #include <asm/a.out-core.h>
33 static int load_aout_binary(struct linux_binprm *, struct pt_regs * regs);
34 static int load_aout_library(struct file*);
35 static int aout_core_dump(long signr, struct pt_regs *regs, struct file *file, unsigned long limit);
37 static struct linux_binfmt aout_format = {
38 .module = THIS_MODULE,
39 .load_binary = load_aout_binary,
40 .load_shlib = load_aout_library,
41 .core_dump = aout_core_dump,
42 .min_coredump = PAGE_SIZE
45 #define BAD_ADDR(x) ((unsigned long)(x) >= TASK_SIZE)
47 static int set_brk(unsigned long start, unsigned long end)
49 start = PAGE_ALIGN(start);
50 end = PAGE_ALIGN(end);
51 if (end > start) {
52 unsigned long addr;
53 down_write(&current->mm->mmap_sem);
54 addr = do_brk(start, end - start);
55 up_write(&current->mm->mmap_sem);
56 if (BAD_ADDR(addr))
57 return addr;
59 return 0;
63 * These are the only things you should do on a core-file: use only these
64 * macros to write out all the necessary info.
67 static int dump_write(struct file *file, const void *addr, int nr)
69 return file->f_op->write(file, addr, nr, &file->f_pos) == nr;
72 #define DUMP_WRITE(addr, nr) \
73 if (!dump_write(file, (void *)(addr), (nr))) \
74 goto end_coredump;
76 #define DUMP_SEEK(offset) \
77 if (file->f_op->llseek) { \
78 if (file->f_op->llseek(file,(offset),0) != (offset)) \
79 goto end_coredump; \
80 } else file->f_pos = (offset)
83 * Routine writes a core dump image in the current directory.
84 * Currently only a stub-function.
86 * Note that setuid/setgid files won't make a core-dump if the uid/gid
87 * changed due to the set[u|g]id. It's enforced by the "current->mm->dumpable"
88 * field, which also makes sure the core-dumps won't be recursive if the
89 * dumping of the process results in another error..
92 static int aout_core_dump(long signr, struct pt_regs *regs, struct file *file, unsigned long limit)
94 mm_segment_t fs;
95 int has_dumped = 0;
96 unsigned long dump_start, dump_size;
97 struct user dump;
98 #if defined(__alpha__)
99 # define START_DATA(u) (u.start_data)
100 #elif defined(__arm__)
101 # define START_DATA(u) ((u.u_tsize << PAGE_SHIFT) + u.start_code)
102 #elif defined(__sparc__)
103 # define START_DATA(u) (u.u_tsize)
104 #elif defined(__i386__) || defined(__mc68000__) || defined(__arch_um__)
105 # define START_DATA(u) (u.u_tsize << PAGE_SHIFT)
106 #endif
107 #ifdef __sparc__
108 # define START_STACK(u) ((regs->u_regs[UREG_FP]) & ~(PAGE_SIZE - 1))
109 #else
110 # define START_STACK(u) (u.start_stack)
111 #endif
113 fs = get_fs();
114 set_fs(KERNEL_DS);
115 has_dumped = 1;
116 current->flags |= PF_DUMPCORE;
117 strncpy(dump.u_comm, current->comm, sizeof(dump.u_comm));
118 #ifndef __sparc__
119 dump.u_ar0 = offsetof(struct user, regs);
120 #endif
121 dump.signal = signr;
122 aout_dump_thread(regs, &dump);
124 /* If the size of the dump file exceeds the rlimit, then see what would happen
125 if we wrote the stack, but not the data area. */
126 #ifdef __sparc__
127 if ((dump.u_dsize + dump.u_ssize) > limit)
128 dump.u_dsize = 0;
129 #else
130 if ((dump.u_dsize + dump.u_ssize+1) * PAGE_SIZE > limit)
131 dump.u_dsize = 0;
132 #endif
134 /* Make sure we have enough room to write the stack and data areas. */
135 #ifdef __sparc__
136 if (dump.u_ssize > limit)
137 dump.u_ssize = 0;
138 #else
139 if ((dump.u_ssize + 1) * PAGE_SIZE > limit)
140 dump.u_ssize = 0;
141 #endif
143 /* make sure we actually have a data and stack area to dump */
144 set_fs(USER_DS);
145 #ifdef __sparc__
146 if (!access_ok(VERIFY_READ, (void __user *)START_DATA(dump), dump.u_dsize))
147 dump.u_dsize = 0;
148 if (!access_ok(VERIFY_READ, (void __user *)START_STACK(dump), dump.u_ssize))
149 dump.u_ssize = 0;
150 #else
151 if (!access_ok(VERIFY_READ, (void __user *)START_DATA(dump), dump.u_dsize << PAGE_SHIFT))
152 dump.u_dsize = 0;
153 if (!access_ok(VERIFY_READ, (void __user *)START_STACK(dump), dump.u_ssize << PAGE_SHIFT))
154 dump.u_ssize = 0;
155 #endif
157 set_fs(KERNEL_DS);
158 /* struct user */
159 DUMP_WRITE(&dump,sizeof(dump));
160 /* Now dump all of the user data. Include malloced stuff as well */
161 #ifndef __sparc__
162 DUMP_SEEK(PAGE_SIZE);
163 #endif
164 /* now we start writing out the user space info */
165 set_fs(USER_DS);
166 /* Dump the data area */
167 if (dump.u_dsize != 0) {
168 dump_start = START_DATA(dump);
169 #ifdef __sparc__
170 dump_size = dump.u_dsize;
171 #else
172 dump_size = dump.u_dsize << PAGE_SHIFT;
173 #endif
174 DUMP_WRITE(dump_start,dump_size);
176 /* Now prepare to dump the stack area */
177 if (dump.u_ssize != 0) {
178 dump_start = START_STACK(dump);
179 #ifdef __sparc__
180 dump_size = dump.u_ssize;
181 #else
182 dump_size = dump.u_ssize << PAGE_SHIFT;
183 #endif
184 DUMP_WRITE(dump_start,dump_size);
186 /* Finally dump the task struct. Not be used by gdb, but could be useful */
187 set_fs(KERNEL_DS);
188 DUMP_WRITE(current,sizeof(*current));
189 end_coredump:
190 set_fs(fs);
191 return has_dumped;
195 * create_aout_tables() parses the env- and arg-strings in new user
196 * memory and creates the pointer tables from them, and puts their
197 * addresses on the "stack", returning the new stack pointer value.
199 static unsigned long __user *create_aout_tables(char __user *p, struct linux_binprm * bprm)
201 char __user * __user *argv;
202 char __user * __user *envp;
203 unsigned long __user *sp;
204 int argc = bprm->argc;
205 int envc = bprm->envc;
207 sp = (void __user *)((-(unsigned long)sizeof(char *)) & (unsigned long) p);
208 #ifdef __sparc__
209 /* This imposes the proper stack alignment for a new process. */
210 sp = (void __user *) (((unsigned long) sp) & ~7);
211 if ((envc+argc+3)&1) --sp;
212 #endif
213 #ifdef __alpha__
214 /* whee.. test-programs are so much fun. */
215 put_user(0, --sp);
216 put_user(0, --sp);
217 if (bprm->loader) {
218 put_user(0, --sp);
219 put_user(0x3eb, --sp);
220 put_user(bprm->loader, --sp);
221 put_user(0x3ea, --sp);
223 put_user(bprm->exec, --sp);
224 put_user(0x3e9, --sp);
225 #endif
226 sp -= envc+1;
227 envp = (char __user * __user *) sp;
228 sp -= argc+1;
229 argv = (char __user * __user *) sp;
230 #if defined(__i386__) || defined(__mc68000__) || defined(__arm__) || defined(__arch_um__)
231 put_user((unsigned long) envp,--sp);
232 put_user((unsigned long) argv,--sp);
233 #endif
234 put_user(argc,--sp);
235 current->mm->arg_start = (unsigned long) p;
236 while (argc-->0) {
237 char c;
238 put_user(p,argv++);
239 do {
240 get_user(c,p++);
241 } while (c);
243 put_user(NULL,argv);
244 current->mm->arg_end = current->mm->env_start = (unsigned long) p;
245 while (envc-->0) {
246 char c;
247 put_user(p,envp++);
248 do {
249 get_user(c,p++);
250 } while (c);
252 put_user(NULL,envp);
253 current->mm->env_end = (unsigned long) p;
254 return sp;
258 * These are the functions used to load a.out style executables and shared
259 * libraries. There is no binary dependent code anywhere else.
262 static int load_aout_binary(struct linux_binprm * bprm, struct pt_regs * regs)
264 struct exec ex;
265 unsigned long error;
266 unsigned long fd_offset;
267 unsigned long rlim;
268 int retval;
270 ex = *((struct exec *) bprm->buf); /* exec-header */
271 if ((N_MAGIC(ex) != ZMAGIC && N_MAGIC(ex) != OMAGIC &&
272 N_MAGIC(ex) != QMAGIC && N_MAGIC(ex) != NMAGIC) ||
273 N_TRSIZE(ex) || N_DRSIZE(ex) ||
274 i_size_read(bprm->file->f_path.dentry->d_inode) < ex.a_text+ex.a_data+N_SYMSIZE(ex)+N_TXTOFF(ex)) {
275 return -ENOEXEC;
279 * Requires a mmap handler. This prevents people from using a.out
280 * as part of an exploit attack against /proc-related vulnerabilities.
282 if (!bprm->file->f_op || !bprm->file->f_op->mmap)
283 return -ENOEXEC;
285 fd_offset = N_TXTOFF(ex);
287 /* Check initial limits. This avoids letting people circumvent
288 * size limits imposed on them by creating programs with large
289 * arrays in the data or bss.
291 rlim = current->signal->rlim[RLIMIT_DATA].rlim_cur;
292 if (rlim >= RLIM_INFINITY)
293 rlim = ~0;
294 if (ex.a_data + ex.a_bss > rlim)
295 return -ENOMEM;
297 /* Flush all traces of the currently running executable */
298 retval = flush_old_exec(bprm);
299 if (retval)
300 return retval;
302 /* OK, This is the point of no return */
303 #if defined(__alpha__)
304 SET_AOUT_PERSONALITY(bprm, ex);
305 #elif defined(__sparc__)
306 set_personality(PER_SUNOS);
307 #if !defined(__sparc_v9__)
308 memcpy(&current->thread.core_exec, &ex, sizeof(struct exec));
309 #endif
310 #else
311 set_personality(PER_LINUX);
312 #endif
314 current->mm->end_code = ex.a_text +
315 (current->mm->start_code = N_TXTADDR(ex));
316 current->mm->end_data = ex.a_data +
317 (current->mm->start_data = N_DATADDR(ex));
318 current->mm->brk = ex.a_bss +
319 (current->mm->start_brk = N_BSSADDR(ex));
320 current->mm->free_area_cache = current->mm->mmap_base;
321 current->mm->cached_hole_size = 0;
323 compute_creds(bprm);
324 current->flags &= ~PF_FORKNOEXEC;
325 #ifdef __sparc__
326 if (N_MAGIC(ex) == NMAGIC) {
327 loff_t pos = fd_offset;
328 /* Fuck me plenty... */
329 /* <AOL></AOL> */
330 down_write(&current->mm->mmap_sem);
331 error = do_brk(N_TXTADDR(ex), ex.a_text);
332 up_write(&current->mm->mmap_sem);
333 bprm->file->f_op->read(bprm->file, (char *) N_TXTADDR(ex),
334 ex.a_text, &pos);
335 down_write(&current->mm->mmap_sem);
336 error = do_brk(N_DATADDR(ex), ex.a_data);
337 up_write(&current->mm->mmap_sem);
338 bprm->file->f_op->read(bprm->file, (char *) N_DATADDR(ex),
339 ex.a_data, &pos);
340 goto beyond_if;
342 #endif
344 if (N_MAGIC(ex) == OMAGIC) {
345 unsigned long text_addr, map_size;
346 loff_t pos;
348 text_addr = N_TXTADDR(ex);
350 #if defined(__alpha__) || defined(__sparc__)
351 pos = fd_offset;
352 map_size = ex.a_text+ex.a_data + PAGE_SIZE - 1;
353 #else
354 pos = 32;
355 map_size = ex.a_text+ex.a_data;
356 #endif
357 down_write(&current->mm->mmap_sem);
358 error = do_brk(text_addr & PAGE_MASK, map_size);
359 up_write(&current->mm->mmap_sem);
360 if (error != (text_addr & PAGE_MASK)) {
361 send_sig(SIGKILL, current, 0);
362 return error;
365 error = bprm->file->f_op->read(bprm->file,
366 (char __user *)text_addr,
367 ex.a_text+ex.a_data, &pos);
368 if ((signed long)error < 0) {
369 send_sig(SIGKILL, current, 0);
370 return error;
373 flush_icache_range(text_addr, text_addr+ex.a_text+ex.a_data);
374 } else {
375 static unsigned long error_time, error_time2;
376 if ((ex.a_text & 0xfff || ex.a_data & 0xfff) &&
377 (N_MAGIC(ex) != NMAGIC) && (jiffies-error_time2) > 5*HZ)
379 printk(KERN_NOTICE "executable not page aligned\n");
380 error_time2 = jiffies;
383 if ((fd_offset & ~PAGE_MASK) != 0 &&
384 (jiffies-error_time) > 5*HZ)
386 printk(KERN_WARNING
387 "fd_offset is not page aligned. Please convert program: %s\n",
388 bprm->file->f_path.dentry->d_name.name);
389 error_time = jiffies;
392 if (!bprm->file->f_op->mmap||((fd_offset & ~PAGE_MASK) != 0)) {
393 loff_t pos = fd_offset;
394 down_write(&current->mm->mmap_sem);
395 do_brk(N_TXTADDR(ex), ex.a_text+ex.a_data);
396 up_write(&current->mm->mmap_sem);
397 bprm->file->f_op->read(bprm->file,
398 (char __user *)N_TXTADDR(ex),
399 ex.a_text+ex.a_data, &pos);
400 flush_icache_range((unsigned long) N_TXTADDR(ex),
401 (unsigned long) N_TXTADDR(ex) +
402 ex.a_text+ex.a_data);
403 goto beyond_if;
406 down_write(&current->mm->mmap_sem);
407 error = do_mmap(bprm->file, N_TXTADDR(ex), ex.a_text,
408 PROT_READ | PROT_EXEC,
409 MAP_FIXED | MAP_PRIVATE | MAP_DENYWRITE | MAP_EXECUTABLE,
410 fd_offset);
411 up_write(&current->mm->mmap_sem);
413 if (error != N_TXTADDR(ex)) {
414 send_sig(SIGKILL, current, 0);
415 return error;
418 down_write(&current->mm->mmap_sem);
419 error = do_mmap(bprm->file, N_DATADDR(ex), ex.a_data,
420 PROT_READ | PROT_WRITE | PROT_EXEC,
421 MAP_FIXED | MAP_PRIVATE | MAP_DENYWRITE | MAP_EXECUTABLE,
422 fd_offset + ex.a_text);
423 up_write(&current->mm->mmap_sem);
424 if (error != N_DATADDR(ex)) {
425 send_sig(SIGKILL, current, 0);
426 return error;
429 beyond_if:
430 set_binfmt(&aout_format);
432 retval = set_brk(current->mm->start_brk, current->mm->brk);
433 if (retval < 0) {
434 send_sig(SIGKILL, current, 0);
435 return retval;
438 retval = setup_arg_pages(bprm, STACK_TOP, EXSTACK_DEFAULT);
439 if (retval < 0) {
440 /* Someone check-me: is this error path enough? */
441 send_sig(SIGKILL, current, 0);
442 return retval;
445 current->mm->start_stack =
446 (unsigned long) create_aout_tables((char __user *) bprm->p, bprm);
447 #ifdef __alpha__
448 regs->gp = ex.a_gpvalue;
449 #endif
450 start_thread(regs, ex.a_entry, current->mm->start_stack);
451 if (unlikely(current->ptrace & PT_PTRACED)) {
452 if (current->ptrace & PT_TRACE_EXEC)
453 ptrace_notify ((PTRACE_EVENT_EXEC << 8) | SIGTRAP);
454 else
455 send_sig(SIGTRAP, current, 0);
457 return 0;
460 static int load_aout_library(struct file *file)
462 struct inode * inode;
463 unsigned long bss, start_addr, len;
464 unsigned long error;
465 int retval;
466 struct exec ex;
468 inode = file->f_path.dentry->d_inode;
470 retval = -ENOEXEC;
471 error = kernel_read(file, 0, (char *) &ex, sizeof(ex));
472 if (error != sizeof(ex))
473 goto out;
475 /* We come in here for the regular a.out style of shared libraries */
476 if ((N_MAGIC(ex) != ZMAGIC && N_MAGIC(ex) != QMAGIC) || N_TRSIZE(ex) ||
477 N_DRSIZE(ex) || ((ex.a_entry & 0xfff) && N_MAGIC(ex) == ZMAGIC) ||
478 i_size_read(inode) < ex.a_text+ex.a_data+N_SYMSIZE(ex)+N_TXTOFF(ex)) {
479 goto out;
483 * Requires a mmap handler. This prevents people from using a.out
484 * as part of an exploit attack against /proc-related vulnerabilities.
486 if (!file->f_op || !file->f_op->mmap)
487 goto out;
489 if (N_FLAGS(ex))
490 goto out;
492 /* For QMAGIC, the starting address is 0x20 into the page. We mask
493 this off to get the starting address for the page */
495 start_addr = ex.a_entry & 0xfffff000;
497 if ((N_TXTOFF(ex) & ~PAGE_MASK) != 0) {
498 static unsigned long error_time;
499 loff_t pos = N_TXTOFF(ex);
501 if ((jiffies-error_time) > 5*HZ)
503 printk(KERN_WARNING
504 "N_TXTOFF is not page aligned. Please convert library: %s\n",
505 file->f_path.dentry->d_name.name);
506 error_time = jiffies;
508 down_write(&current->mm->mmap_sem);
509 do_brk(start_addr, ex.a_text + ex.a_data + ex.a_bss);
510 up_write(&current->mm->mmap_sem);
512 file->f_op->read(file, (char __user *)start_addr,
513 ex.a_text + ex.a_data, &pos);
514 flush_icache_range((unsigned long) start_addr,
515 (unsigned long) start_addr + ex.a_text + ex.a_data);
517 retval = 0;
518 goto out;
520 /* Now use mmap to map the library into memory. */
521 down_write(&current->mm->mmap_sem);
522 error = do_mmap(file, start_addr, ex.a_text + ex.a_data,
523 PROT_READ | PROT_WRITE | PROT_EXEC,
524 MAP_FIXED | MAP_PRIVATE | MAP_DENYWRITE,
525 N_TXTOFF(ex));
526 up_write(&current->mm->mmap_sem);
527 retval = error;
528 if (error != start_addr)
529 goto out;
531 len = PAGE_ALIGN(ex.a_text + ex.a_data);
532 bss = ex.a_text + ex.a_data + ex.a_bss;
533 if (bss > len) {
534 down_write(&current->mm->mmap_sem);
535 error = do_brk(start_addr + len, bss - len);
536 up_write(&current->mm->mmap_sem);
537 retval = error;
538 if (error != start_addr + len)
539 goto out;
541 retval = 0;
542 out:
543 return retval;
546 static int __init init_aout_binfmt(void)
548 return register_binfmt(&aout_format);
551 static void __exit exit_aout_binfmt(void)
553 unregister_binfmt(&aout_format);
556 core_initcall(init_aout_binfmt);
557 module_exit(exit_aout_binfmt);
558 MODULE_LICENSE("GPL");