2 * Copyright (C) 2012 - Virtual Open Systems and Columbia University
3 * Author: Christoffer Dall <c.dall@virtualopensystems.com>
5 * This program is free software; you can redistribute it and/or modify
6 * it under the terms of the GNU General Public License, version 2, as
7 * published by the Free Software Foundation.
9 * This program is distributed in the hope that it will be useful,
10 * but WITHOUT ANY WARRANTY; without even the implied warranty of
11 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
12 * GNU General Public License for more details.
14 * You should have received a copy of the GNU General Public License
15 * along with this program; if not, write to the Free Software
16 * Foundation, 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA.
20 #include <linux/kvm_host.h>
21 #include <asm/kvm_arm.h>
22 #include <asm/kvm_emulate.h>
23 #include <asm/opcodes.h>
24 #include <trace/events/kvm.h>
28 #define VCPU_NR_MODES 6
29 #define VCPU_REG_OFFSET_USR 0
30 #define VCPU_REG_OFFSET_FIQ 1
31 #define VCPU_REG_OFFSET_IRQ 2
32 #define VCPU_REG_OFFSET_SVC 3
33 #define VCPU_REG_OFFSET_ABT 4
34 #define VCPU_REG_OFFSET_UND 5
35 #define REG_OFFSET(_reg) \
36 (offsetof(struct kvm_regs, _reg) / sizeof(u32))
38 #define USR_REG_OFFSET(_num) REG_OFFSET(usr_regs.uregs[_num])
40 static const unsigned long vcpu_reg_offsets
[VCPU_NR_MODES
][15] = {
41 /* USR/SYS Registers */
42 [VCPU_REG_OFFSET_USR
] = {
43 USR_REG_OFFSET(0), USR_REG_OFFSET(1), USR_REG_OFFSET(2),
44 USR_REG_OFFSET(3), USR_REG_OFFSET(4), USR_REG_OFFSET(5),
45 USR_REG_OFFSET(6), USR_REG_OFFSET(7), USR_REG_OFFSET(8),
46 USR_REG_OFFSET(9), USR_REG_OFFSET(10), USR_REG_OFFSET(11),
47 USR_REG_OFFSET(12), USR_REG_OFFSET(13), USR_REG_OFFSET(14),
51 [VCPU_REG_OFFSET_FIQ
] = {
52 USR_REG_OFFSET(0), USR_REG_OFFSET(1), USR_REG_OFFSET(2),
53 USR_REG_OFFSET(3), USR_REG_OFFSET(4), USR_REG_OFFSET(5),
54 USR_REG_OFFSET(6), USR_REG_OFFSET(7),
55 REG_OFFSET(fiq_regs
[0]), /* r8 */
56 REG_OFFSET(fiq_regs
[1]), /* r9 */
57 REG_OFFSET(fiq_regs
[2]), /* r10 */
58 REG_OFFSET(fiq_regs
[3]), /* r11 */
59 REG_OFFSET(fiq_regs
[4]), /* r12 */
60 REG_OFFSET(fiq_regs
[5]), /* r13 */
61 REG_OFFSET(fiq_regs
[6]), /* r14 */
65 [VCPU_REG_OFFSET_IRQ
] = {
66 USR_REG_OFFSET(0), USR_REG_OFFSET(1), USR_REG_OFFSET(2),
67 USR_REG_OFFSET(3), USR_REG_OFFSET(4), USR_REG_OFFSET(5),
68 USR_REG_OFFSET(6), USR_REG_OFFSET(7), USR_REG_OFFSET(8),
69 USR_REG_OFFSET(9), USR_REG_OFFSET(10), USR_REG_OFFSET(11),
71 REG_OFFSET(irq_regs
[0]), /* r13 */
72 REG_OFFSET(irq_regs
[1]), /* r14 */
76 [VCPU_REG_OFFSET_SVC
] = {
77 USR_REG_OFFSET(0), USR_REG_OFFSET(1), USR_REG_OFFSET(2),
78 USR_REG_OFFSET(3), USR_REG_OFFSET(4), USR_REG_OFFSET(5),
79 USR_REG_OFFSET(6), USR_REG_OFFSET(7), USR_REG_OFFSET(8),
80 USR_REG_OFFSET(9), USR_REG_OFFSET(10), USR_REG_OFFSET(11),
82 REG_OFFSET(svc_regs
[0]), /* r13 */
83 REG_OFFSET(svc_regs
[1]), /* r14 */
87 [VCPU_REG_OFFSET_ABT
] = {
88 USR_REG_OFFSET(0), USR_REG_OFFSET(1), USR_REG_OFFSET(2),
89 USR_REG_OFFSET(3), USR_REG_OFFSET(4), USR_REG_OFFSET(5),
90 USR_REG_OFFSET(6), USR_REG_OFFSET(7), USR_REG_OFFSET(8),
91 USR_REG_OFFSET(9), USR_REG_OFFSET(10), USR_REG_OFFSET(11),
93 REG_OFFSET(abt_regs
[0]), /* r13 */
94 REG_OFFSET(abt_regs
[1]), /* r14 */
98 [VCPU_REG_OFFSET_UND
] = {
99 USR_REG_OFFSET(0), USR_REG_OFFSET(1), USR_REG_OFFSET(2),
100 USR_REG_OFFSET(3), USR_REG_OFFSET(4), USR_REG_OFFSET(5),
101 USR_REG_OFFSET(6), USR_REG_OFFSET(7), USR_REG_OFFSET(8),
102 USR_REG_OFFSET(9), USR_REG_OFFSET(10), USR_REG_OFFSET(11),
104 REG_OFFSET(und_regs
[0]), /* r13 */
105 REG_OFFSET(und_regs
[1]), /* r14 */
110 * Return a pointer to the register number valid in the current mode of
113 unsigned long *vcpu_reg(struct kvm_vcpu
*vcpu
, u8 reg_num
)
115 unsigned long *reg_array
= (unsigned long *)&vcpu
->arch
.ctxt
.gp_regs
;
116 unsigned long mode
= *vcpu_cpsr(vcpu
) & MODE_MASK
;
119 case USR_MODE
...SVC_MODE
:
120 mode
&= ~MODE32_BIT
; /* 0 ... 3 */
124 mode
= VCPU_REG_OFFSET_ABT
;
128 mode
= VCPU_REG_OFFSET_UND
;
132 mode
= VCPU_REG_OFFSET_USR
;
139 return reg_array
+ vcpu_reg_offsets
[mode
][reg_num
];
143 * Return the SPSR for the current mode of the virtual CPU.
145 unsigned long *vcpu_spsr(struct kvm_vcpu
*vcpu
)
147 unsigned long mode
= *vcpu_cpsr(vcpu
) & MODE_MASK
;
150 return &vcpu
->arch
.ctxt
.gp_regs
.KVM_ARM_SVC_spsr
;
152 return &vcpu
->arch
.ctxt
.gp_regs
.KVM_ARM_ABT_spsr
;
154 return &vcpu
->arch
.ctxt
.gp_regs
.KVM_ARM_UND_spsr
;
156 return &vcpu
->arch
.ctxt
.gp_regs
.KVM_ARM_IRQ_spsr
;
158 return &vcpu
->arch
.ctxt
.gp_regs
.KVM_ARM_FIQ_spsr
;
164 /******************************************************************************
165 * Inject exceptions into the guest
169 * kvm_inject_vabt - inject an async abort / SError into the guest
170 * @vcpu: The VCPU to receive the exception
172 * It is assumed that this code is called from the VCPU thread and that the
173 * VCPU therefore is not currently executing guest code.
175 void kvm_inject_vabt(struct kvm_vcpu
*vcpu
)
177 vcpu_set_hcr(vcpu
, vcpu_get_hcr(vcpu
) | HCR_VA
);