2 * This file is subject to the terms and conditions of the GNU General Public
3 * License. See the file "COPYING" in the main directory of this archive
6 * KVM/MIPS: Binary Patching for privileged instructions, reduces traps.
8 * Copyright (C) 2012 MIPS Technologies, Inc. All rights reserved.
9 * Authors: Sanjay Lal <sanjayl@kymasys.com>
12 #include <linux/errno.h>
13 #include <linux/err.h>
14 #include <linux/highmem.h>
15 #include <linux/kvm_host.h>
16 #include <linux/uaccess.h>
17 #include <linux/vmalloc.h>
19 #include <linux/bootmem.h>
20 #include <asm/cacheflush.h>
25 * kvm_mips_trans_replace() - Replace trapping instruction in guest memory.
27 * @opc: PC of instruction to replace.
28 * @replace: Instruction to write
30 static int kvm_mips_trans_replace(struct kvm_vcpu
*vcpu
, u32
*opc
,
31 union mips_instruction replace
)
33 unsigned long vaddr
= (unsigned long)opc
;
37 /* The GVA page table is still active so use the Linux TLB handlers */
38 kvm_trap_emul_gva_lockless_begin(vcpu
);
39 err
= put_user(replace
.word
, opc
);
40 kvm_trap_emul_gva_lockless_end(vcpu
);
44 * We write protect clean pages in GVA page table so normal
45 * Linux TLB mod handler doesn't silently dirty the page.
46 * Its also possible we raced with a GVA invalidation.
47 * Try to force the page to become dirty.
49 err
= kvm_trap_emul_gva_fault(vcpu
, vaddr
, true);
51 kvm_info("%s: Address unwriteable: %p\n",
57 * Try again. This will likely trigger a TLB refill, which will
58 * fetch the new dirty entry from the GVA page table, which
59 * should then succeed.
63 __local_flush_icache_user_range(vaddr
, vaddr
+ 4);
68 int kvm_mips_trans_cache_index(union mips_instruction inst
, u32
*opc
,
69 struct kvm_vcpu
*vcpu
)
71 union mips_instruction nop_inst
= { 0 };
73 /* Replace the CACHE instruction, with a NOP */
74 return kvm_mips_trans_replace(vcpu
, opc
, nop_inst
);
78 * Address based CACHE instructions are transformed into synci(s). A little
79 * heavy for just D-cache invalidates, but avoids an expensive trap
81 int kvm_mips_trans_cache_va(union mips_instruction inst
, u32
*opc
,
82 struct kvm_vcpu
*vcpu
)
84 union mips_instruction synci_inst
= { 0 };
86 synci_inst
.i_format
.opcode
= bcond_op
;
87 synci_inst
.i_format
.rs
= inst
.i_format
.rs
;
88 synci_inst
.i_format
.rt
= synci_op
;
90 synci_inst
.i_format
.simmediate
= inst
.spec3_format
.simmediate
;
92 synci_inst
.i_format
.simmediate
= inst
.i_format
.simmediate
;
94 return kvm_mips_trans_replace(vcpu
, opc
, synci_inst
);
97 int kvm_mips_trans_mfc0(union mips_instruction inst
, u32
*opc
,
98 struct kvm_vcpu
*vcpu
)
100 union mips_instruction mfc0_inst
= { 0 };
103 rd
= inst
.c0r_format
.rd
;
104 sel
= inst
.c0r_format
.sel
;
106 if (rd
== MIPS_CP0_ERRCTL
&& sel
== 0) {
107 mfc0_inst
.r_format
.opcode
= spec_op
;
108 mfc0_inst
.r_format
.rd
= inst
.c0r_format
.rt
;
109 mfc0_inst
.r_format
.func
= add_op
;
111 mfc0_inst
.i_format
.opcode
= lw_op
;
112 mfc0_inst
.i_format
.rt
= inst
.c0r_format
.rt
;
113 mfc0_inst
.i_format
.simmediate
= KVM_GUEST_COMMPAGE_ADDR
|
114 offsetof(struct kvm_mips_commpage
, cop0
.reg
[rd
][sel
]);
115 #ifdef CONFIG_CPU_BIG_ENDIAN
116 if (sizeof(vcpu
->arch
.cop0
->reg
[0][0]) == 8)
117 mfc0_inst
.i_format
.simmediate
|= 4;
121 return kvm_mips_trans_replace(vcpu
, opc
, mfc0_inst
);
124 int kvm_mips_trans_mtc0(union mips_instruction inst
, u32
*opc
,
125 struct kvm_vcpu
*vcpu
)
127 union mips_instruction mtc0_inst
= { 0 };
130 rd
= inst
.c0r_format
.rd
;
131 sel
= inst
.c0r_format
.sel
;
133 mtc0_inst
.i_format
.opcode
= sw_op
;
134 mtc0_inst
.i_format
.rt
= inst
.c0r_format
.rt
;
135 mtc0_inst
.i_format
.simmediate
= KVM_GUEST_COMMPAGE_ADDR
|
136 offsetof(struct kvm_mips_commpage
, cop0
.reg
[rd
][sel
]);
137 #ifdef CONFIG_CPU_BIG_ENDIAN
138 if (sizeof(vcpu
->arch
.cop0
->reg
[0][0]) == 8)
139 mtc0_inst
.i_format
.simmediate
|= 4;
142 return kvm_mips_trans_replace(vcpu
, opc
, mtc0_inst
);