treewide: remove redundant IS_ERR() before error code check
[linux/fpc-iii.git] / arch / powerpc / kvm / book3s_32_sr.S
blobe3ab9df6cf199c1151464ac08cc62460d08e210f
1 /* SPDX-License-Identifier: GPL-2.0-only */
2 /*
3  *
4  * Copyright SUSE Linux Products GmbH 2009
5  *
6  * Authors: Alexander Graf <agraf@suse.de>
7  */
9 /******************************************************************************
10  *                                                                            *
11  *                               Entry code                                   *
12  *                                                                            *
13  *****************************************************************************/
15 .macro LOAD_GUEST_SEGMENTS
17         /* Required state:
18          *
19          * MSR = ~IR|DR
20          * R1 = host R1
21          * R2 = host R2
22          * R3 = shadow vcpu
23          * all other volatile GPRS = free except R4, R6
24          * SVCPU[CR]  = guest CR
25          * SVCPU[XER] = guest XER
26          * SVCPU[CTR] = guest CTR
27          * SVCPU[LR]  = guest LR
28          */
30 #define XCHG_SR(n)      lwz     r9, (SVCPU_SR+(n*4))(r3);  \
31                         mtsr    n, r9
33         XCHG_SR(0)
34         XCHG_SR(1)
35         XCHG_SR(2)
36         XCHG_SR(3)
37         XCHG_SR(4)
38         XCHG_SR(5)
39         XCHG_SR(6)
40         XCHG_SR(7)
41         XCHG_SR(8)
42         XCHG_SR(9)
43         XCHG_SR(10)
44         XCHG_SR(11)
45         XCHG_SR(12)
46         XCHG_SR(13)
47         XCHG_SR(14)
48         XCHG_SR(15)
50         /* Clear BATs. */
52 #define KVM_KILL_BAT(n, reg)            \
53         mtspr   SPRN_IBAT##n##U,reg;    \
54         mtspr   SPRN_IBAT##n##L,reg;    \
55         mtspr   SPRN_DBAT##n##U,reg;    \
56         mtspr   SPRN_DBAT##n##L,reg;    \
58         li      r9, 0
59         KVM_KILL_BAT(0, r9)
60         KVM_KILL_BAT(1, r9)
61         KVM_KILL_BAT(2, r9)
62         KVM_KILL_BAT(3, r9)
64 .endm
66 /******************************************************************************
67  *                                                                            *
68  *                               Exit code                                    *
69  *                                                                            *
70  *****************************************************************************/
72 .macro LOAD_HOST_SEGMENTS
74         /* Register usage at this point:
75          *
76          * R1         = host R1
77          * R2         = host R2
78          * R12        = exit handler id
79          * R13        = shadow vcpu - SHADOW_VCPU_OFF
80          * SVCPU.*    = guest *
81          * SVCPU[CR]  = guest CR
82          * SVCPU[XER] = guest XER
83          * SVCPU[CTR] = guest CTR
84          * SVCPU[LR]  = guest LR
85          *
86          */
88         /* Restore BATs */
90         /* We only overwrite the upper part, so we only restoree
91            the upper part. */
92 #define KVM_LOAD_BAT(n, reg, RA, RB)    \
93         lwz     RA,(n*16)+0(reg);       \
94         lwz     RB,(n*16)+4(reg);       \
95         mtspr   SPRN_IBAT##n##U,RA;     \
96         mtspr   SPRN_IBAT##n##L,RB;     \
97         lwz     RA,(n*16)+8(reg);       \
98         lwz     RB,(n*16)+12(reg);      \
99         mtspr   SPRN_DBAT##n##U,RA;     \
100         mtspr   SPRN_DBAT##n##L,RB;     \
102         lis     r9, BATS@ha
103         addi    r9, r9, BATS@l
104         tophys(r9, r9)
105         KVM_LOAD_BAT(0, r9, r10, r11)
106         KVM_LOAD_BAT(1, r9, r10, r11)
107         KVM_LOAD_BAT(2, r9, r10, r11)
108         KVM_LOAD_BAT(3, r9, r10, r11)
110         /* Restore Segment Registers */
112         /* 0xc - 0xf */
114         li      r0, 4
115         mtctr   r0
116         LOAD_REG_IMMEDIATE(r3, 0x20000000 | (0x111 * 0xc))
117         lis     r4, 0xc000
118 3:      mtsrin  r3, r4
119         addi    r3, r3, 0x111     /* increment VSID */
120         addis   r4, r4, 0x1000    /* address of next segment */
121         bdnz    3b
123         /* 0x0 - 0xb */
125         /* 'current->mm' needs to be in r4 */
126         tophys(r4, r2)
127         lwz     r4, MM(r4)
128         tophys(r4, r4)
129         /* This only clobbers r0, r3, r4 and r5 */
130         bl      switch_mmu_context
132 .endm