1 // SPDX-License-Identifier: GPL-2.0-only
4 * Copyright SUSE Linux Products GmbH 2009
6 * Authors: Alexander Graf <agraf@suse.de>
9 #include <linux/types.h>
10 #include <linux/string.h>
11 #include <linux/kvm.h>
12 #include <linux/kvm_host.h>
13 #include <linux/highmem.h>
15 #include <asm/kvm_ppc.h>
16 #include <asm/kvm_book3s.h>
18 /* #define DEBUG_MMU */
19 /* #define DEBUG_MMU_PTE */
20 /* #define DEBUG_MMU_PTE_IP 0xfff14c40 */
23 #define dprintk(X...) printk(KERN_INFO X)
25 #define dprintk(X...) do { } while(0)
29 #define dprintk_pte(X...) printk(KERN_INFO X)
31 #define dprintk_pte(X...) do { } while(0)
34 #define PTEG_FLAG_ACCESSED 0x00000100
35 #define PTEG_FLAG_DIRTY 0x00000080
40 static inline bool check_debug_ip(struct kvm_vcpu
*vcpu
)
42 #ifdef DEBUG_MMU_PTE_IP
43 return vcpu
->arch
.regs
.nip
== DEBUG_MMU_PTE_IP
;
49 static inline u32
sr_vsid(u32 sr_raw
)
51 return sr_raw
& 0x0fffffff;
54 static inline bool sr_valid(u32 sr_raw
)
56 return (sr_raw
& 0x80000000) ? false : true;
59 static inline bool sr_ks(u32 sr_raw
)
61 return (sr_raw
& 0x40000000) ? true: false;
64 static inline bool sr_kp(u32 sr_raw
)
66 return (sr_raw
& 0x20000000) ? true: false;
69 static int kvmppc_mmu_book3s_32_xlate_bat(struct kvm_vcpu
*vcpu
, gva_t eaddr
,
70 struct kvmppc_pte
*pte
, bool data
,
72 static int kvmppc_mmu_book3s_32_esid_to_vsid(struct kvm_vcpu
*vcpu
, ulong esid
,
75 static u32
find_sr(struct kvm_vcpu
*vcpu
, gva_t eaddr
)
77 return kvmppc_get_sr(vcpu
, (eaddr
>> 28) & 0xf);
80 static u64
kvmppc_mmu_book3s_32_ea_to_vp(struct kvm_vcpu
*vcpu
, gva_t eaddr
,
84 struct kvmppc_pte pte
;
86 if (!kvmppc_mmu_book3s_32_xlate_bat(vcpu
, eaddr
, &pte
, data
, false))
89 kvmppc_mmu_book3s_32_esid_to_vsid(vcpu
, eaddr
>> SID_SHIFT
, &vsid
);
90 return (((u64
)eaddr
>> 12) & 0xffff) | (vsid
<< 16);
93 static hva_t
kvmppc_mmu_book3s_32_get_pteg(struct kvm_vcpu
*vcpu
,
97 struct kvmppc_vcpu_book3s
*vcpu_book3s
= to_book3s(vcpu
);
98 u32 page
, hash
, pteg
, htabmask
;
101 page
= (eaddr
& 0x0FFFFFFF) >> 12;
102 htabmask
= ((vcpu_book3s
->sdr1
& 0x1FF) << 16) | 0xFFC0;
104 hash
= ((sr_vsid(sre
) ^ page
) << 6);
109 pteg
= (vcpu_book3s
->sdr1
& 0xffff0000) | hash
;
111 dprintk("MMU: pc=0x%lx eaddr=0x%lx sdr1=0x%llx pteg=0x%x vsid=0x%x\n",
112 kvmppc_get_pc(vcpu
), eaddr
, vcpu_book3s
->sdr1
, pteg
,
115 r
= gfn_to_hva(vcpu
->kvm
, pteg
>> PAGE_SHIFT
);
116 if (kvm_is_error_hva(r
))
118 return r
| (pteg
& ~PAGE_MASK
);
121 static u32
kvmppc_mmu_book3s_32_get_ptem(u32 sre
, gva_t eaddr
, bool primary
)
123 return ((eaddr
& 0x0fffffff) >> 22) | (sr_vsid(sre
) << 7) |
124 (primary
? 0 : 0x40) | 0x80000000;
127 static int kvmppc_mmu_book3s_32_xlate_bat(struct kvm_vcpu
*vcpu
, gva_t eaddr
,
128 struct kvmppc_pte
*pte
, bool data
,
131 struct kvmppc_vcpu_book3s
*vcpu_book3s
= to_book3s(vcpu
);
132 struct kvmppc_bat
*bat
;
135 for (i
= 0; i
< 8; i
++) {
137 bat
= &vcpu_book3s
->dbat
[i
];
139 bat
= &vcpu_book3s
->ibat
[i
];
141 if (kvmppc_get_msr(vcpu
) & MSR_PR
) {
149 if (check_debug_ip(vcpu
))
151 dprintk_pte("%cBAT %02d: 0x%lx - 0x%x (0x%x)\n",
152 data
? 'd' : 'i', i
, eaddr
, bat
->bepi
,
155 if ((eaddr
& bat
->bepi_mask
) == bat
->bepi
) {
157 kvmppc_mmu_book3s_32_esid_to_vsid(vcpu
,
158 eaddr
>> SID_SHIFT
, &vsid
);
160 pte
->vpage
= (((u64
)eaddr
>> 12) & 0xffff) | vsid
;
162 pte
->raddr
= bat
->brpn
| (eaddr
& ~bat
->bepi_mask
);
163 pte
->may_read
= bat
->pp
;
164 pte
->may_write
= bat
->pp
> 1;
165 pte
->may_execute
= true;
166 if (!pte
->may_read
) {
167 printk(KERN_INFO
"BAT is not readable!\n");
170 if (iswrite
&& !pte
->may_write
) {
171 dprintk_pte("BAT is read-only!\n");
182 static int kvmppc_mmu_book3s_32_xlate_pte(struct kvm_vcpu
*vcpu
, gva_t eaddr
,
183 struct kvmppc_pte
*pte
, bool data
,
184 bool iswrite
, bool primary
)
194 sre
= find_sr(vcpu
, eaddr
);
196 dprintk_pte("SR 0x%lx: vsid=0x%x, raw=0x%x\n", eaddr
>> 28,
199 pte
->vpage
= kvmppc_mmu_book3s_32_ea_to_vp(vcpu
, eaddr
, data
);
201 ptegp
= kvmppc_mmu_book3s_32_get_pteg(vcpu
, sre
, eaddr
, primary
);
202 if (kvm_is_error_hva(ptegp
)) {
203 printk(KERN_INFO
"KVM: Invalid PTEG!\n");
207 ptem
= kvmppc_mmu_book3s_32_get_ptem(sre
, eaddr
, primary
);
209 if(copy_from_user(pteg
, (void __user
*)ptegp
, sizeof(pteg
))) {
210 printk_ratelimited(KERN_ERR
211 "KVM: Can't copy data from 0x%lx!\n", ptegp
);
215 for (i
=0; i
<16; i
+=2) {
216 pte0
= be32_to_cpu(pteg
[i
]);
217 pte1
= be32_to_cpu(pteg
[i
+ 1]);
221 pte
->raddr
= (pte1
& ~(0xFFFULL
)) | (eaddr
& 0xFFF);
224 if ((sr_kp(sre
) && (kvmppc_get_msr(vcpu
) & MSR_PR
)) ||
225 (sr_ks(sre
) && !(kvmppc_get_msr(vcpu
) & MSR_PR
)))
228 pte
->may_write
= false;
229 pte
->may_read
= false;
230 pte
->may_execute
= true;
236 pte
->may_write
= true;
241 pte
->may_read
= true;
245 dprintk_pte("MMU: Found PTE -> %x %x - %x\n",
252 /* Update PTE C and A bits, so the guest's swapper knows we used the
256 char __user
*addr
= (char __user
*) (ptegp
+ (i
+1) * sizeof(u32
));
259 * Use single-byte writes to update the HPTE, to
260 * conform to what real hardware does.
262 if (pte
->may_read
&& !(pte_r
& PTEG_FLAG_ACCESSED
)) {
263 pte_r
|= PTEG_FLAG_ACCESSED
;
264 put_user(pte_r
>> 8, addr
+ 2);
266 if (iswrite
&& pte
->may_write
&& !(pte_r
& PTEG_FLAG_DIRTY
)) {
267 pte_r
|= PTEG_FLAG_DIRTY
;
268 put_user(pte_r
, addr
+ 3);
270 if (!pte
->may_read
|| (iswrite
&& !pte
->may_write
))
277 if (check_debug_ip(vcpu
)) {
278 dprintk_pte("KVM MMU: No PTE found (sdr1=0x%llx ptegp=0x%lx)\n",
279 to_book3s(vcpu
)->sdr1
, ptegp
);
280 for (i
=0; i
<16; i
+=2) {
281 dprintk_pte(" %02d: 0x%x - 0x%x (0x%x)\n",
282 i
, be32_to_cpu(pteg
[i
]),
283 be32_to_cpu(pteg
[i
+1]), ptem
);
290 static int kvmppc_mmu_book3s_32_xlate(struct kvm_vcpu
*vcpu
, gva_t eaddr
,
291 struct kvmppc_pte
*pte
, bool data
,
295 ulong mp_ea
= vcpu
->arch
.magic_page_ea
;
298 pte
->page_size
= MMU_PAGE_4K
;
300 /* Magic page override */
301 if (unlikely(mp_ea
) &&
302 unlikely((eaddr
& ~0xfffULL
) == (mp_ea
& ~0xfffULL
)) &&
303 !(kvmppc_get_msr(vcpu
) & MSR_PR
)) {
304 pte
->vpage
= kvmppc_mmu_book3s_32_ea_to_vp(vcpu
, eaddr
, data
);
305 pte
->raddr
= vcpu
->arch
.magic_page_pa
| (pte
->raddr
& 0xfff);
306 pte
->raddr
&= KVM_PAM
;
307 pte
->may_execute
= true;
308 pte
->may_read
= true;
309 pte
->may_write
= true;
314 r
= kvmppc_mmu_book3s_32_xlate_bat(vcpu
, eaddr
, pte
, data
, iswrite
);
316 r
= kvmppc_mmu_book3s_32_xlate_pte(vcpu
, eaddr
, pte
,
317 data
, iswrite
, true);
319 r
= kvmppc_mmu_book3s_32_xlate_pte(vcpu
, eaddr
, pte
,
320 data
, iswrite
, false);
326 static u32
kvmppc_mmu_book3s_32_mfsrin(struct kvm_vcpu
*vcpu
, u32 srnum
)
328 return kvmppc_get_sr(vcpu
, srnum
);
331 static void kvmppc_mmu_book3s_32_mtsrin(struct kvm_vcpu
*vcpu
, u32 srnum
,
334 kvmppc_set_sr(vcpu
, srnum
, value
);
335 kvmppc_mmu_map_segment(vcpu
, srnum
<< SID_SHIFT
);
338 static void kvmppc_mmu_book3s_32_tlbie(struct kvm_vcpu
*vcpu
, ulong ea
, bool large
)
343 /* flush this VA on all cpus */
344 kvm_for_each_vcpu(i
, v
, vcpu
->kvm
)
345 kvmppc_mmu_pte_flush(v
, ea
, 0x0FFFF000);
348 static int kvmppc_mmu_book3s_32_esid_to_vsid(struct kvm_vcpu
*vcpu
, ulong esid
,
351 ulong ea
= esid
<< SID_SHIFT
;
354 u64 msr
= kvmppc_get_msr(vcpu
);
356 if (msr
& (MSR_DR
|MSR_IR
)) {
357 sr
= find_sr(vcpu
, ea
);
362 /* In case we only have one of MSR_IR or MSR_DR set, let's put
363 that in the real-mode context (and hope RM doesn't access
365 switch (msr
& (MSR_DR
|MSR_IR
)) {
367 *vsid
= VSID_REAL
| esid
;
370 *vsid
= VSID_REAL_IR
| gvsid
;
373 *vsid
= VSID_REAL_DR
| gvsid
;
379 *vsid
= VSID_BAT
| gvsid
;
391 static bool kvmppc_mmu_book3s_32_is_dcbz32(struct kvm_vcpu
*vcpu
)
397 void kvmppc_mmu_book3s_32_init(struct kvm_vcpu
*vcpu
)
399 struct kvmppc_mmu
*mmu
= &vcpu
->arch
.mmu
;
401 mmu
->mtsrin
= kvmppc_mmu_book3s_32_mtsrin
;
402 mmu
->mfsrin
= kvmppc_mmu_book3s_32_mfsrin
;
403 mmu
->xlate
= kvmppc_mmu_book3s_32_xlate
;
404 mmu
->tlbie
= kvmppc_mmu_book3s_32_tlbie
;
405 mmu
->esid_to_vsid
= kvmppc_mmu_book3s_32_esid_to_vsid
;
406 mmu
->ea_to_vp
= kvmppc_mmu_book3s_32_ea_to_vp
;
407 mmu
->is_dcbz32
= kvmppc_mmu_book3s_32_is_dcbz32
;