2 * Copyright (C) 2012 - Virtual Open Systems and Columbia University
3 * Author: Christoffer Dall <c.dall@virtualopensystems.com>
5 * This program is free software; you can redistribute it and/or modify
6 * it under the terms of the GNU General Public License, version 2, as
7 * published by the Free Software Foundation.
9 * This program is distributed in the hope that it will be useful,
10 * but WITHOUT ANY WARRANTY; without even the implied warranty of
11 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
12 * GNU General Public License for more details.
14 * You should have received a copy of the GNU General Public License
15 * along with this program; if not, write to the Free Software
16 * Foundation, 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA.
19 #include <linux/kvm_host.h>
20 #include <asm/kvm_mmio.h>
21 #include <asm/kvm_emulate.h>
22 #include <trace/events/kvm.h>
26 static void mmio_write_buf(char *buf
, unsigned int len
, unsigned long data
)
55 memcpy(buf
, datap
, len
);
58 static unsigned long mmio_read_buf(char *buf
, unsigned int len
)
60 unsigned long data
= 0;
72 memcpy(&tmp
.hword
, buf
, len
);
76 memcpy(&tmp
.word
, buf
, len
);
80 memcpy(&tmp
.dword
, buf
, len
);
89 * kvm_handle_mmio_return -- Handle MMIO loads after user space emulation
90 * @vcpu: The VCPU pointer
91 * @run: The VCPU run struct containing the mmio data
93 * This should only be called after returning from userspace for MMIO load
96 int kvm_handle_mmio_return(struct kvm_vcpu
*vcpu
, struct kvm_run
*run
)
102 if (!run
->mmio
.is_write
) {
104 if (len
> sizeof(unsigned long))
107 data
= mmio_read_buf(run
->mmio
.data
, len
);
109 if (vcpu
->arch
.mmio_decode
.sign_extend
&&
110 len
< sizeof(unsigned long)) {
111 mask
= 1U << ((len
* 8) - 1);
112 data
= (data
^ mask
) - mask
;
115 trace_kvm_mmio(KVM_TRACE_MMIO_READ
, len
, run
->mmio
.phys_addr
,
117 data
= vcpu_data_host_to_guest(vcpu
, data
, len
);
118 *vcpu_reg(vcpu
, vcpu
->arch
.mmio_decode
.rt
) = data
;
124 static int decode_hsr(struct kvm_vcpu
*vcpu
, bool *is_write
, int *len
)
130 if (kvm_vcpu_dabt_isextabt(vcpu
)) {
131 /* cache operation on I/O addr, tell guest unsupported */
132 kvm_inject_dabt(vcpu
, kvm_vcpu_get_hfar(vcpu
));
136 if (kvm_vcpu_dabt_iss1tw(vcpu
)) {
137 /* page table accesses IO mem: tell guest to fix its TTBR */
138 kvm_inject_dabt(vcpu
, kvm_vcpu_get_hfar(vcpu
));
142 access_size
= kvm_vcpu_dabt_get_as(vcpu
);
143 if (unlikely(access_size
< 0))
146 *is_write
= kvm_vcpu_dabt_iswrite(vcpu
);
147 sign_extend
= kvm_vcpu_dabt_issext(vcpu
);
148 rt
= kvm_vcpu_dabt_get_rd(vcpu
);
151 vcpu
->arch
.mmio_decode
.sign_extend
= sign_extend
;
152 vcpu
->arch
.mmio_decode
.rt
= rt
;
155 * The MMIO instruction is emulated and should not be re-executed
158 kvm_skip_instr(vcpu
, kvm_vcpu_trap_il_is32bit(vcpu
));
162 int io_mem_abort(struct kvm_vcpu
*vcpu
, struct kvm_run
*run
,
163 phys_addr_t fault_ipa
)
173 * Prepare MMIO operation. First decode the syndrome data we get
174 * from the CPU. Then try if some in-kernel emulation feels
175 * responsible, otherwise let user space do its magic.
177 if (kvm_vcpu_dabt_isvalid(vcpu
)) {
178 ret
= decode_hsr(vcpu
, &is_write
, &len
);
182 kvm_err("load/store instruction decoding not implemented\n");
186 rt
= vcpu
->arch
.mmio_decode
.rt
;
189 data
= vcpu_data_guest_to_host(vcpu
, *vcpu_reg(vcpu
, rt
), len
);
191 trace_kvm_mmio(KVM_TRACE_MMIO_WRITE
, len
, fault_ipa
, data
);
192 mmio_write_buf(data_buf
, len
, data
);
194 ret
= kvm_io_bus_write(vcpu
, KVM_MMIO_BUS
, fault_ipa
, len
,
197 trace_kvm_mmio(KVM_TRACE_MMIO_READ_UNSATISFIED
, len
,
200 ret
= kvm_io_bus_read(vcpu
, KVM_MMIO_BUS
, fault_ipa
, len
,
204 /* Now prepare kvm_run for the potential return to userland. */
205 run
->mmio
.is_write
= is_write
;
206 run
->mmio
.phys_addr
= fault_ipa
;
208 memcpy(run
->mmio
.data
, data_buf
, len
);
211 /* We handled the access successfully in the kernel. */
212 kvm_handle_mmio_return(vcpu
, run
);
216 run
->exit_reason
= KVM_EXIT_MMIO
;