4 * Linux architectural port borrowing liberally from similar works of
5 * others. All original copyrights apply as per the original source
8 * Modifications for the OpenRISC architecture:
9 * Copyright (C) 2003 Matjaz Breskvar <phoenix@bsemi.com>
10 * Copyright (C) 2005 Gyorgy Jeney <nog@bsemi.com>
11 * Copyright (C) 2010-2011 Jonas Bonn <jonas@southpole.se>
13 * This program is free software; you can redistribute it and/or
14 * modify it under the terms of the GNU General Public License
15 * as published by the Free Software Foundation; either version
16 * 2 of the License, or (at your option) any later version.
19 #include <linux/linkage.h>
21 #include <asm/processor.h>
22 #include <asm/unistd.h>
23 #include <asm/thread_info.h>
24 #include <asm/errno.h>
25 #include <asm/spr_defs.h>
28 #include <asm/pgtable.h>
29 #include <asm/asm-offsets.h>
31 #define DISABLE_INTERRUPTS(t1,t2) \
32 l.mfspr t2,r0,SPR_SR ;\
33 l.movhi t1,hi(~(SPR_SR_IEE|SPR_SR_TEE)) ;\
34 l.ori t1,t1,lo(~(SPR_SR_IEE|SPR_SR_TEE)) ;\
38 #define ENABLE_INTERRUPTS(t1) \
39 l.mfspr t1,r0,SPR_SR ;\
40 l.ori t1,t1,lo(SPR_SR_IEE|SPR_SR_TEE) ;\
43 /* =========================================================[ macros ]=== */
46 * We need to disable interrupts at beginning of RESTORE_ALL
47 * since interrupt might come in after we've loaded EPC return address
48 * and overwrite EPC with address somewhere in RESTORE_ALL
49 * which is of course wrong!
53 DISABLE_INTERRUPTS(r3,r4) ;\
55 l.mtspr r0,r3,SPR_EPCR_BASE ;\
57 l.mtspr r0,r3,SPR_ESR_BASE ;\
58 l.lwz r2,PT_GPR2(r1) ;\
59 l.lwz r3,PT_GPR3(r1) ;\
60 l.lwz r4,PT_GPR4(r1) ;\
61 l.lwz r5,PT_GPR5(r1) ;\
62 l.lwz r6,PT_GPR6(r1) ;\
63 l.lwz r7,PT_GPR7(r1) ;\
64 l.lwz r8,PT_GPR8(r1) ;\
65 l.lwz r9,PT_GPR9(r1) ;\
66 l.lwz r10,PT_GPR10(r1) ;\
67 l.lwz r11,PT_GPR11(r1) ;\
68 l.lwz r12,PT_GPR12(r1) ;\
69 l.lwz r13,PT_GPR13(r1) ;\
70 l.lwz r14,PT_GPR14(r1) ;\
71 l.lwz r15,PT_GPR15(r1) ;\
72 l.lwz r16,PT_GPR16(r1) ;\
73 l.lwz r17,PT_GPR17(r1) ;\
74 l.lwz r18,PT_GPR18(r1) ;\
75 l.lwz r19,PT_GPR19(r1) ;\
76 l.lwz r20,PT_GPR20(r1) ;\
77 l.lwz r21,PT_GPR21(r1) ;\
78 l.lwz r22,PT_GPR22(r1) ;\
79 l.lwz r23,PT_GPR23(r1) ;\
80 l.lwz r24,PT_GPR24(r1) ;\
81 l.lwz r25,PT_GPR25(r1) ;\
82 l.lwz r26,PT_GPR26(r1) ;\
83 l.lwz r27,PT_GPR27(r1) ;\
84 l.lwz r28,PT_GPR28(r1) ;\
85 l.lwz r29,PT_GPR29(r1) ;\
86 l.lwz r30,PT_GPR30(r1) ;\
87 l.lwz r31,PT_GPR31(r1) ;\
92 #define EXCEPTION_ENTRY(handler) \
95 /* r1, EPCR, ESR a already saved */ ;\
96 l.sw PT_GPR2(r1),r2 ;\
97 l.sw PT_GPR3(r1),r3 ;\
98 l.sw PT_ORIG_GPR11(r1),r11 ;\
99 /* r4 already save */ ;\
100 l.sw PT_GPR5(r1),r5 ;\
101 l.sw PT_GPR6(r1),r6 ;\
102 l.sw PT_GPR7(r1),r7 ;\
103 l.sw PT_GPR8(r1),r8 ;\
104 l.sw PT_GPR9(r1),r9 ;\
105 /* r10 already saved */ ;\
106 l.sw PT_GPR11(r1),r11 ;\
107 /* r12 already saved */ ;\
108 l.sw PT_GPR13(r1),r13 ;\
109 l.sw PT_GPR14(r1),r14 ;\
110 l.sw PT_GPR15(r1),r15 ;\
111 l.sw PT_GPR16(r1),r16 ;\
112 l.sw PT_GPR17(r1),r17 ;\
113 l.sw PT_GPR18(r1),r18 ;\
114 l.sw PT_GPR19(r1),r19 ;\
115 l.sw PT_GPR20(r1),r20 ;\
116 l.sw PT_GPR21(r1),r21 ;\
117 l.sw PT_GPR22(r1),r22 ;\
118 l.sw PT_GPR23(r1),r23 ;\
119 l.sw PT_GPR24(r1),r24 ;\
120 l.sw PT_GPR25(r1),r25 ;\
121 l.sw PT_GPR26(r1),r26 ;\
122 l.sw PT_GPR27(r1),r27 ;\
123 l.sw PT_GPR28(r1),r28 ;\
124 l.sw PT_GPR29(r1),r29 ;\
125 /* r30 already save */ ;\
126 /* l.sw PT_GPR30(r1),r30*/ ;\
127 l.sw PT_GPR31(r1),r31 ;\
128 l.sw PT_SYSCALLNO(r1),r0
130 #define UNHANDLED_EXCEPTION(handler,vector) \
133 /* r1, EPCR, ESR already saved */ ;\
134 l.sw PT_GPR2(r1),r2 ;\
135 l.sw PT_GPR3(r1),r3 ;\
136 l.sw PT_ORIG_GPR11(r1),r11 ;\
137 l.sw PT_GPR5(r1),r5 ;\
138 l.sw PT_GPR6(r1),r6 ;\
139 l.sw PT_GPR7(r1),r7 ;\
140 l.sw PT_GPR8(r1),r8 ;\
141 l.sw PT_GPR9(r1),r9 ;\
142 /* r10 already saved */ ;\
143 l.sw PT_GPR11(r1),r11 ;\
144 /* r12 already saved */ ;\
145 l.sw PT_GPR13(r1),r13 ;\
146 l.sw PT_GPR14(r1),r14 ;\
147 l.sw PT_GPR15(r1),r15 ;\
148 l.sw PT_GPR16(r1),r16 ;\
149 l.sw PT_GPR17(r1),r17 ;\
150 l.sw PT_GPR18(r1),r18 ;\
151 l.sw PT_GPR19(r1),r19 ;\
152 l.sw PT_GPR20(r1),r20 ;\
153 l.sw PT_GPR21(r1),r21 ;\
154 l.sw PT_GPR22(r1),r22 ;\
155 l.sw PT_GPR23(r1),r23 ;\
156 l.sw PT_GPR24(r1),r24 ;\
157 l.sw PT_GPR25(r1),r25 ;\
158 l.sw PT_GPR26(r1),r26 ;\
159 l.sw PT_GPR27(r1),r27 ;\
160 l.sw PT_GPR28(r1),r28 ;\
161 l.sw PT_GPR29(r1),r29 ;\
162 /* r31 already saved */ ;\
163 l.sw PT_GPR30(r1),r30 ;\
164 /* l.sw PT_GPR31(r1),r31 */ ;\
165 l.sw PT_SYSCALLNO(r1),r0 ;\
167 /* r4 is exception EA */ ;\
168 l.addi r5,r0,vector ;\
169 l.jal unhandled_exception ;\
171 l.j _ret_from_exception ;\
175 * NOTE: one should never assume that SPR_EPC, SPR_ESR, SPR_EEAR
176 * contain the same values as when exception we're handling
177 * occured. in fact they never do. if you need them use
178 * values saved on stack (for SPR_EPC, SPR_ESR) or content
179 * of r4 (for SPR_EEAR). for details look at EXCEPTION_HANDLE()
180 * in 'arch/or32/kernel/head.S'
183 /* =====================================================[ exceptions] === */
185 /* ---[ 0x100: RESET exception ]----------------------------------------- */
187 EXCEPTION_ENTRY(_tng_kernel_start)
191 /* ---[ 0x200: BUS exception ]------------------------------------------- */
193 EXCEPTION_ENTRY(_bus_fault_handler)
194 /* r4: EA of fault (set by EXCEPTION_HANDLE) */
196 l.addi r3,r1,0 /* pt_regs */
198 l.j _ret_from_exception
201 /* ---[ 0x300: Data Page Fault exception ]------------------------------- */
203 EXCEPTION_ENTRY(_data_page_fault_handler)
204 /* set up parameters for do_page_fault */
205 l.addi r3,r1,0 // pt_regs
206 /* r4 set be EXCEPTION_HANDLE */ // effective address of fault
207 l.ori r5,r0,0x300 // exception vector
212 * all this can be written much simpler. look at
213 * DTLB miss handler in the CONFIG_GUARD_PROTECTED_CORE part
215 #ifdef CONFIG_OPENRISC_NO_SPR_SR_DSX
216 l.lwz r6,PT_PC(r3) // address of an offending insn
217 l.lwz r6,0(r6) // instruction that caused pf
219 l.srli r6,r6,26 // check opcode for jump insn
222 l.sfeqi r6,1 // l.jal
224 l.sfeqi r6,3 // l.bnf
228 l.sfeqi r6,0x11 // l.jr
230 l.sfeqi r6,0x12 // l.jalr
239 l.lwz r6,PT_PC(r3) // address of an offending insn
241 l.lwz r6,0(r6) // instruction that caused pf
242 l.srli r6,r6,26 // get opcode
247 l.mfspr r6,r0,SPR_SR // SR
248 // l.lwz r6,PT_SR(r3) // ESR
249 l.andi r6,r6,SPR_SR_DSX // check for delay slot exception
250 l.sfeqi r6,0x1 // exception happened in delay slot
252 l.lwz r6,PT_PC(r3) // address of an offending insn
254 l.addi r6,r6,4 // offending insn is in delay slot
256 l.lwz r6,0(r6) // instruction that caused pf
257 l.srli r6,r6,26 // check opcode for write access
260 l.sfgeui r6,0x34 // check opcode for write access
264 l.ori r6,r0,0x1 // write access
267 1: l.ori r6,r0,0x0 // !write access
270 /* call fault.c handler in or32/mm/fault.c */
273 l.j _ret_from_exception
276 /* ---[ 0x400: Insn Page Fault exception ]------------------------------- */
278 EXCEPTION_ENTRY(_insn_page_fault_handler)
279 /* set up parameters for do_page_fault */
280 l.addi r3,r1,0 // pt_regs
281 /* r4 set be EXCEPTION_HANDLE */ // effective address of fault
282 l.ori r5,r0,0x400 // exception vector
283 l.ori r6,r0,0x0 // !write access
285 /* call fault.c handler in or32/mm/fault.c */
288 l.j _ret_from_exception
292 /* ---[ 0x500: Timer exception ]----------------------------------------- */
294 EXCEPTION_ENTRY(_timer_handler)
295 l.jal timer_interrupt
296 l.addi r3,r1,0 /* pt_regs */
301 /* ---[ 0x600: Aligment exception ]-------------------------------------- */
303 EXCEPTION_ENTRY(_alignment_handler)
304 /* r4: EA of fault (set by EXCEPTION_HANDLE) */
305 l.jal do_unaligned_access
306 l.addi r3,r1,0 /* pt_regs */
308 l.j _ret_from_exception
312 EXCEPTION_ENTRY(_aligment_handler)
313 // l.mfspr r2,r0,SPR_EEAR_BASE /* Load the efective addres */
315 // l.mfspr r5,r0,SPR_EPCR_BASE /* Load the insn address */
318 l.lwz r3,0(r5) /* Load insn */
319 l.srli r4,r3,26 /* Shift left to get the insn opcode */
321 l.sfeqi r4,0x00 /* Check if the load/store insn is in delay slot */
335 l.addi r5,r5,4 /* Increment PC to get return insn address */
338 l.slli r4,r3,6 /* Get the signed extended jump length */
341 l.lwz r3,4(r5) /* Load the real load/store insn */
343 l.add r5,r5,r4 /* Calculate jump target address */
346 l.srli r4,r3,26 /* Shift left to get the insn opcode */
349 l.slli r4,r3,9 /* Shift to get the reg nb */
352 l.lwz r3,4(r5) /* Load the real load/store insn */
354 l.add r4,r4,r1 /* Load the jump register value from the stack */
357 l.srli r4,r3,26 /* Shift left to get the insn opcode */
361 // l.mtspr r0,r5,SPR_EPCR_BASE
378 1: l.j 1b /* I don't know what to do */
462 /* ---[ 0x700: Illegal insn exception ]---------------------------------- */
464 EXCEPTION_ENTRY(_illegal_instruction_handler)
465 /* r4: EA of fault (set by EXCEPTION_HANDLE) */
466 l.jal do_illegal_instruction
467 l.addi r3,r1,0 /* pt_regs */
469 l.j _ret_from_exception
472 /* ---[ 0x800: External interrupt exception ]---------------------------- */
474 EXCEPTION_ENTRY(_external_irq_handler)
475 #ifdef CONFIG_OPENRISC_ESR_EXCEPTION_BUG_CHECK
476 l.lwz r4,PT_SR(r1) // were interrupts enabled ?
477 l.andi r4,r4,SPR_SR_IEE
479 l.bnf 1f // ext irq enabled, all ok.
490 .section .rodata, "a"
492 .string "\n\rESR interrupt bug: in _external_irq_handler (ESR %x)\n\r"
496 l.ori r4,r4,SPR_SR_IEE // fix the bug
501 l.movhi r8,hi(do_IRQ)
502 l.ori r8,r8,lo(do_IRQ)
508 /* ---[ 0x900: DTLB miss exception ]------------------------------------- */
511 /* ---[ 0xa00: ITLB miss exception ]------------------------------------- */
514 /* ---[ 0xb00: Range exception ]----------------------------------------- */
516 UNHANDLED_EXCEPTION(_vector_0xb00,0xb00)
518 /* ---[ 0xc00: Syscall exception ]--------------------------------------- */
521 * Syscalls are a special type of exception in that they are
522 * _explicitly_ invoked by userspace and can therefore be
523 * held to conform to the same ABI as normal functions with
524 * respect to whether registers are preserved across the call
528 /* Upon syscall entry we just save the callee-saved registers
529 * and not the call-clobbered ones.
532 _string_syscall_return:
533 .string "syscall return %ld \n\r\0"
536 ENTRY(_sys_call_handler)
537 /* syscalls run with interrupts enabled */
538 ENABLE_INTERRUPTS(r29) // enable interrupts, r29 is temp
540 /* r1, EPCR, ESR a already saved */
542 /* r3-r8 must be saved because syscall restart relies
543 * on us being able to restart the syscall args... technically
544 * they should be clobbered, otherwise
547 /* r4 already saved */
548 /* r4 holds the EEAR address of the fault, load the original r4 */
555 /* r10 already saved */
556 l.sw PT_GPR11(r1),r11
557 l.sw PT_ORIG_GPR11(r1),r11
558 /* r12,r13 already saved */
560 /* r14-r28 (even) aren't touched by the syscall fast path below
561 * so we don't need to save them. However, the functions that return
562 * to userspace via a call to switch() DO need to save these because
563 * switch() effectively clobbers them... saving these registers for
564 * such functions is handled in their syscall wrappers (see fork, vfork,
567 /* r30 is the only register we clobber in the fast path */
568 /* r30 already saved */
569 /* l.sw PT_GPR30(r1),r30 */
570 /* This is used by do_signal to determine whether to check for
571 * syscall restart or not */
572 l.sw PT_SYSCALLNO(r1),r11
574 _syscall_check_trace_enter:
575 /* If TIF_SYSCALL_TRACE is set, then we want to do syscall tracing */
576 l.lwz r30,TI_FLAGS(r10)
577 l.andi r30,r30,_TIF_SYSCALL_TRACE
579 l.bf _syscall_trace_enter
583 /* Ensure that the syscall number is reasonable */
584 l.sfgeui r11,__NR_syscalls
589 l.movhi r29,hi(sys_call_table)
590 l.ori r29,r29,lo(sys_call_table)
599 /* All syscalls return here... just pay attention to ret_from_fork
600 * which does it in a round-about way.
602 l.sw PT_GPR11(r1),r11 // save return value
606 l.movhi r3,hi(_string_syscall_return)
607 l.ori r3,r3,lo(_string_syscall_return)
612 l.movhi r27,hi(printk)
613 l.ori r27,r27,lo(printk)
619 _syscall_check_trace_leave:
620 /* r30 is a callee-saved register so this should still hold the
621 * _TIF_SYSCALL_TRACE flag from _syscall_check_trace_enter above...
622 * _syscall_trace_leave expects syscall result to be in pt_regs->r11.
625 l.bf _syscall_trace_leave
628 /* This is where the exception-return code begins... interrupts need to be
629 * disabled the rest of the way here because we can't afford to miss any
630 * interrupts that set NEED_RESCHED or SIGNALPENDING... really true? */
633 /* Here we need to disable interrupts */
634 DISABLE_INTERRUPTS(r27,r29)
635 l.lwz r30,TI_FLAGS(r10)
636 l.andi r30,r30,_TIF_WORK_MASK
639 l.bnf _syscall_resume_userspace
642 /* Work pending follows a different return path, so we need to
643 * make sure that all the call-saved registers get into pt_regs
644 * before branching...
646 l.sw PT_GPR14(r1),r14
647 l.sw PT_GPR16(r1),r16
648 l.sw PT_GPR18(r1),r18
649 l.sw PT_GPR20(r1),r20
650 l.sw PT_GPR22(r1),r22
651 l.sw PT_GPR24(r1),r24
652 l.sw PT_GPR26(r1),r26
653 l.sw PT_GPR28(r1),r28
655 /* _work_pending needs to be called with interrupts disabled */
659 _syscall_resume_userspace:
660 // ENABLE_INTERRUPTS(r29)
663 /* This is the hot path for returning to userspace from a syscall. If there's
664 * work to be done and the branch to _work_pending was taken above, then the
665 * return to userspace will be done via the normal exception return path...
666 * that path restores _all_ registers and will overwrite the "clobbered"
667 * registers with whatever garbage is in pt_regs -- that's OK because those
668 * registers are clobbered anyway and because the extra work is insignificant
669 * in the context of the extra work that _work_pending is doing.
671 /* Once again, syscalls are special and only guarantee to preserve the
672 * same registers as a normal function call */
674 /* The assumption here is that the registers r14-r28 (even) are untouched and
675 * don't need to be restored... be sure that that's really the case!
678 /* This is still too much... we should only be restoring what we actually
679 * clobbered... we should even be using 'scratch' (odd) regs above so that
680 * we don't need to restore anything, hardly...
686 /* r3-r8 are technically clobbered, but syscall restart needs these
697 l.lwz r10,PT_GPR10(r1)
698 l.lwz r11,PT_GPR11(r1)
700 /* r30 is the only register we clobber in the fast path */
701 l.lwz r30,PT_GPR30(r1)
703 /* Here we use r13-r19 (odd) as scratch regs */
707 /* Interrupts need to be disabled for setting EPCR and ESR
708 * so that another interrupt doesn't come in here and clobber
709 * them before we can use them for our l.rfe */
710 DISABLE_INTERRUPTS(r17,r19)
711 l.mtspr r0,r13,SPR_EPCR_BASE
712 l.mtspr r0,r15,SPR_ESR_BASE
716 * Keep the below tracing and error handling out of the hot path...
719 _syscall_trace_enter:
720 /* Here we pass pt_regs to do_syscall_trace_enter. Make sure
721 * that function is really getting all the info it needs as
722 * pt_regs isn't a complete set of userspace regs, just the
723 * ones relevant to the syscall...
725 * Note use of delay slot for setting argument.
727 l.jal do_syscall_trace_enter
730 /* Restore arguments (not preserved across do_syscall_trace_enter)
731 * so that we can do the syscall for real and return to the syscall
734 l.lwz r11,PT_SYSCALLNO(r1)
744 _syscall_trace_leave:
745 l.jal do_syscall_trace_leave
748 l.j _syscall_check_work
752 /* Here we effectively pretend to have executed an imaginary
753 * syscall that returns -ENOSYS and then return to the regular
755 * Note that "return value" is set in the delay slot...
758 l.addi r11,r0,-ENOSYS
760 /******* END SYSCALL HANDLING *******/
762 /* ---[ 0xd00: Trap exception ]------------------------------------------ */
764 UNHANDLED_EXCEPTION(_vector_0xd00,0xd00)
766 /* ---[ 0xe00: Trap exception ]------------------------------------------ */
768 EXCEPTION_ENTRY(_trap_handler)
769 /* r4: EA of fault (set by EXCEPTION_HANDLE) */
771 l.addi r3,r1,0 /* pt_regs */
773 l.j _ret_from_exception
776 /* ---[ 0xf00: Reserved exception ]-------------------------------------- */
778 UNHANDLED_EXCEPTION(_vector_0xf00,0xf00)
780 /* ---[ 0x1000: Reserved exception ]------------------------------------- */
782 UNHANDLED_EXCEPTION(_vector_0x1000,0x1000)
784 /* ---[ 0x1100: Reserved exception ]------------------------------------- */
786 UNHANDLED_EXCEPTION(_vector_0x1100,0x1100)
788 /* ---[ 0x1200: Reserved exception ]------------------------------------- */
790 UNHANDLED_EXCEPTION(_vector_0x1200,0x1200)
792 /* ---[ 0x1300: Reserved exception ]------------------------------------- */
794 UNHANDLED_EXCEPTION(_vector_0x1300,0x1300)
796 /* ---[ 0x1400: Reserved exception ]------------------------------------- */
798 UNHANDLED_EXCEPTION(_vector_0x1400,0x1400)
800 /* ---[ 0x1500: Reserved exception ]------------------------------------- */
802 UNHANDLED_EXCEPTION(_vector_0x1500,0x1500)
804 /* ---[ 0x1600: Reserved exception ]------------------------------------- */
806 UNHANDLED_EXCEPTION(_vector_0x1600,0x1600)
808 /* ---[ 0x1700: Reserved exception ]------------------------------------- */
810 UNHANDLED_EXCEPTION(_vector_0x1700,0x1700)
812 /* ---[ 0x1800: Reserved exception ]------------------------------------- */
814 UNHANDLED_EXCEPTION(_vector_0x1800,0x1800)
816 /* ---[ 0x1900: Reserved exception ]------------------------------------- */
818 UNHANDLED_EXCEPTION(_vector_0x1900,0x1900)
820 /* ---[ 0x1a00: Reserved exception ]------------------------------------- */
822 UNHANDLED_EXCEPTION(_vector_0x1a00,0x1a00)
824 /* ---[ 0x1b00: Reserved exception ]------------------------------------- */
826 UNHANDLED_EXCEPTION(_vector_0x1b00,0x1b00)
828 /* ---[ 0x1c00: Reserved exception ]------------------------------------- */
830 UNHANDLED_EXCEPTION(_vector_0x1c00,0x1c00)
832 /* ---[ 0x1d00: Reserved exception ]------------------------------------- */
834 UNHANDLED_EXCEPTION(_vector_0x1d00,0x1d00)
836 /* ---[ 0x1e00: Reserved exception ]------------------------------------- */
838 UNHANDLED_EXCEPTION(_vector_0x1e00,0x1e00)
840 /* ---[ 0x1f00: Reserved exception ]------------------------------------- */
842 UNHANDLED_EXCEPTION(_vector_0x1f00,0x1f00)
844 /* ========================================================[ return ] === */
848 * if (current_thread_info->flags & _TIF_NEED_RESCHED)
851 l.lwz r5,TI_FLAGS(r10)
852 l.andi r3,r5,_TIF_NEED_RESCHED
854 l.bnf _work_notifysig
858 l.j _resume_userspace
861 /* Handle pending signals and notify-resume requests.
862 * do_notify_resume must be passed the latest pushed pt_regs, not
863 * necessarily the "userspace" ones. Also, pt_regs->syscallno
864 * must be set so that the syscall restart functionality works.
867 l.jal do_notify_resume
868 l.ori r3,r1,0 /* pt_regs */
871 DISABLE_INTERRUPTS(r3,r4)
872 l.lwz r3,TI_FLAGS(r10)
873 l.andi r3,r3,_TIF_WORK_MASK
880 /* This returns to userspace code */
883 ENTRY(_ret_from_intr)
884 ENTRY(_ret_from_exception)
886 l.andi r3,r4,SPR_SR_SM
890 l.j _resume_userspace
897 /* _syscall_returns expect r11 to contain return value */
898 l.lwz r11,PT_GPR11(r1)
900 /* The syscall fast path return expects call-saved registers
901 * r12-r28 to be untouched, so we restore them here as they
902 * will have been effectively clobbered when arriving here
903 * via the call to switch()
905 l.lwz r12,PT_GPR12(r1)
906 l.lwz r14,PT_GPR14(r1)
907 l.lwz r16,PT_GPR16(r1)
908 l.lwz r18,PT_GPR18(r1)
909 l.lwz r20,PT_GPR20(r1)
910 l.lwz r22,PT_GPR22(r1)
911 l.lwz r24,PT_GPR24(r1)
912 l.lwz r26,PT_GPR26(r1)
913 l.lwz r28,PT_GPR28(r1)
918 /* Since syscalls don't save call-clobbered registers, the args to
919 * kernel_thread_helper will need to be passed through callee-saved
920 * registers and copied to the parameter registers when the thread
923 * See arch/openrisc/kernel/process.c:
924 * The args are passed as follows:
925 * arg1 (r3) : passed in r20
926 * arg2 (r4) : passed in r22
929 ENTRY(_kernel_thread_helper)
932 l.movhi r31,hi(kernel_thread_helper)
933 l.ori r31,r31,lo(kernel_thread_helper)
938 /* ========================================================[ switch ] === */
941 * This routine switches between two different tasks. The process
942 * state of one is saved on its kernel stack. Then the state
943 * of the other is restored from its kernel stack. The memory
944 * management hardware is updated to the second process's state.
945 * Finally, we can return to the second process, via the 'return'.
947 * Note: there are two ways to get to the "going out" portion
948 * of this code; either by coming in via the entry (_switch)
949 * or via "fork" which must set up an environment equivalent
950 * to the "_switch" path. If you change this (or in particular, the
951 * SAVE_REGS macro), you'll have to change the fork code also.
955 /* _switch MUST never lay on page boundry, cause it runs from
956 * effective addresses and beeing interrupted by iTLB miss would kill it.
957 * dTLB miss seams to never accour in the bad place since data accesses
958 * are from task structures which are always page aligned.
960 * The problem happens in RESTORE_ALL_NO_R11 where we first set the EPCR
961 * register, then load the previous register values and only at the end call
962 * the l.rfe instruction. If get TLB miss in beetwen the EPCR register gets
963 * garbled and we end up calling l.rfe with the wrong EPCR. (same probably
966 * To avoid this problems it is sufficient to align _switch to
967 * some nice round number smaller than it's size...
970 /* ABI rules apply here... we either enter _switch via schedule() or via
971 * an imaginary call to which we shall return at return_from_fork. Either
972 * way, we are a function call and only need to preserve the callee-saved
973 * registers when we return. As such, we don't need to save the registers
974 * on the stack that we won't be returning as they were...
979 /* We don't store SR as _switch only gets called in a context where
980 * the SR will be the same going in and coming out... */
982 /* Set up new pt_regs struct for saving task state */
983 l.addi r1,r1,-(INT_FRAME_SIZE)
985 /* No need to store r1/PT_SP as it goes into KSP below */
988 /* This is wrong, r12 shouldn't be here... but GCC is broken for the time being
989 * and expects r12 to be callee-saved... */
990 l.sw PT_GPR12(r1),r12
991 l.sw PT_GPR14(r1),r14
992 l.sw PT_GPR16(r1),r16
993 l.sw PT_GPR18(r1),r18
994 l.sw PT_GPR20(r1),r20
995 l.sw PT_GPR22(r1),r22
996 l.sw PT_GPR24(r1),r24
997 l.sw PT_GPR26(r1),r26
998 l.sw PT_GPR28(r1),r28
999 l.sw PT_GPR30(r1),r30
1001 l.addi r11,r10,0 /* Save old 'current' to 'last' return value*/
1003 /* We use thread_info->ksp for storing the address of the above
1004 * structure so that we can get back to it later... we don't want
1005 * to lose the value of thread_info->ksp, though, so store it as
1006 * pt_regs->sp so that we can easily restore it when we are made
1010 /* Save the old value of thread_info->ksp as pt_regs->sp */
1011 l.lwz r29,TI_KSP(r10)
1014 /* Swap kernel stack pointers */
1015 l.sw TI_KSP(r10),r1 /* Save old stack pointer */
1016 l.or r10,r4,r0 /* Set up new current_thread_info */
1017 l.lwz r1,TI_KSP(r10) /* Load new stack pointer */
1019 /* Restore the old value of thread_info->ksp */
1021 l.sw TI_KSP(r10),r29
1023 /* ...and restore the registers, except r11 because the return value
1024 * has already been set above.
1026 l.lwz r2,PT_GPR2(r1)
1027 l.lwz r9,PT_GPR9(r1)
1028 /* No need to restore r10 */
1029 /* ...and do not restore r11 */
1031 /* This is wrong, r12 shouldn't be here... but GCC is broken for the time being
1032 * and expects r12 to be callee-saved... */
1033 l.lwz r12,PT_GPR12(r1)
1034 l.lwz r14,PT_GPR14(r1)
1035 l.lwz r16,PT_GPR16(r1)
1036 l.lwz r18,PT_GPR18(r1)
1037 l.lwz r20,PT_GPR20(r1)
1038 l.lwz r22,PT_GPR22(r1)
1039 l.lwz r24,PT_GPR24(r1)
1040 l.lwz r26,PT_GPR26(r1)
1041 l.lwz r28,PT_GPR28(r1)
1042 l.lwz r30,PT_GPR30(r1)
1044 /* Unwind stack to pre-switch state */
1045 l.addi r1,r1,(INT_FRAME_SIZE)
1047 /* Return via the link-register back to where we 'came from', where that can be
1048 * either schedule() or return_from_fork()... */
1052 /* ==================================================================== */
1054 /* These all use the delay slot for setting the argument register, so the
1055 * jump is always happening after the l.addi instruction.
1057 * These are all just wrappers that don't touch the link-register r9, so the
1058 * return from the "real" syscall function will return back to the syscall
1059 * code that did the l.jal that brought us here.
1062 /* fork requires that we save all the callee-saved registers because they
1063 * are all effectively clobbered by the call to _switch. Here we store
1064 * all the registers that aren't touched by the syscall fast path and thus
1065 * weren't saved there.
1068 _fork_save_extra_regs_and_call:
1069 l.sw PT_GPR14(r1),r14
1070 l.sw PT_GPR16(r1),r16
1071 l.sw PT_GPR18(r1),r18
1072 l.sw PT_GPR20(r1),r20
1073 l.sw PT_GPR22(r1),r22
1074 l.sw PT_GPR24(r1),r24
1075 l.sw PT_GPR26(r1),r26
1077 l.sw PT_GPR28(r1),r28
1080 l.movhi r29,hi(_sys_clone)
1081 l.ori r29,r29,lo(_sys_clone)
1082 l.j _fork_save_extra_regs_and_call
1086 l.movhi r29,hi(_sys_fork)
1087 l.ori r29,r29,lo(_sys_fork)
1088 l.j _fork_save_extra_regs_and_call
1095 ENTRY(sys_sigaltstack)
1096 l.j _sys_sigaltstack
1099 ENTRY(sys_rt_sigreturn)
1100 l.j _sys_rt_sigreturn
1103 /* This is a catch-all syscall for atomic instructions for the OpenRISC 1000.
1104 * The functions takes a variable number of parameters depending on which
1105 * particular flavour of atomic you want... parameter 1 is a flag identifying
1106 * the atomic in question. Currently, this function implements the
1107 * following variants:
1113 * Atomically exchange the values in pointers 1 and 2.
1117 ENTRY(sys_or1k_atomic)
1118 /* FIXME: This ignores r3 and always does an XCHG */
1119 DISABLE_INTERRUPTS(r17,r19)
1124 ENABLE_INTERRUPTS(r17)
1128 /* ============================================================[ EOF ]=== */