2 * This program is free software; you can redistribute it and/or modify
3 * it under the terms of the GNU General Public License, version 2, as
4 * published by the Free Software Foundation.
6 * This program is distributed in the hope that it will be useful,
7 * but WITHOUT ANY WARRANTY; without even the implied warranty of
8 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
9 * GNU General Public License for more details.
11 * You should have received a copy of the GNU General Public License
12 * along with this program; if not, write to the Free Software
13 * Foundation, 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA.
15 * Copyright SUSE Linux Products GmbH 2009
17 * Authors: Alexander Graf <agraf@suse.de>
20 #include <linux/types.h>
21 #include <linux/string.h>
22 #include <linux/kvm.h>
23 #include <linux/kvm_host.h>
24 #include <linux/highmem.h>
26 #include <asm/tlbflush.h>
27 #include <asm/kvm_ppc.h>
28 #include <asm/kvm_book3s.h>
30 /* #define DEBUG_MMU */
31 /* #define DEBUG_MMU_PTE */
32 /* #define DEBUG_MMU_PTE_IP 0xfff14c40 */
35 #define dprintk(X...) printk(KERN_INFO X)
37 #define dprintk(X...) do { } while(0)
41 #define dprintk_pte(X...) printk(KERN_INFO X)
43 #define dprintk_pte(X...) do { } while(0)
46 #define PTEG_FLAG_ACCESSED 0x00000100
47 #define PTEG_FLAG_DIRTY 0x00000080
52 static inline bool check_debug_ip(struct kvm_vcpu
*vcpu
)
54 #ifdef DEBUG_MMU_PTE_IP
55 return vcpu
->arch
.pc
== DEBUG_MMU_PTE_IP
;
61 static inline u32
sr_vsid(u32 sr_raw
)
63 return sr_raw
& 0x0fffffff;
66 static inline bool sr_valid(u32 sr_raw
)
68 return (sr_raw
& 0x80000000) ? false : true;
71 static inline bool sr_ks(u32 sr_raw
)
73 return (sr_raw
& 0x40000000) ? true: false;
76 static inline bool sr_kp(u32 sr_raw
)
78 return (sr_raw
& 0x20000000) ? true: false;
81 static int kvmppc_mmu_book3s_32_xlate_bat(struct kvm_vcpu
*vcpu
, gva_t eaddr
,
82 struct kvmppc_pte
*pte
, bool data
,
84 static int kvmppc_mmu_book3s_32_esid_to_vsid(struct kvm_vcpu
*vcpu
, ulong esid
,
87 static u32
find_sr(struct kvm_vcpu
*vcpu
, gva_t eaddr
)
89 return kvmppc_get_sr(vcpu
, (eaddr
>> 28) & 0xf);
92 static u64
kvmppc_mmu_book3s_32_ea_to_vp(struct kvm_vcpu
*vcpu
, gva_t eaddr
,
96 struct kvmppc_pte pte
;
98 if (!kvmppc_mmu_book3s_32_xlate_bat(vcpu
, eaddr
, &pte
, data
, false))
101 kvmppc_mmu_book3s_32_esid_to_vsid(vcpu
, eaddr
>> SID_SHIFT
, &vsid
);
102 return (((u64
)eaddr
>> 12) & 0xffff) | (vsid
<< 16);
105 static void kvmppc_mmu_book3s_32_reset_msr(struct kvm_vcpu
*vcpu
)
107 kvmppc_set_msr(vcpu
, 0);
110 static hva_t
kvmppc_mmu_book3s_32_get_pteg(struct kvm_vcpu
*vcpu
,
111 u32 sre
, gva_t eaddr
,
114 struct kvmppc_vcpu_book3s
*vcpu_book3s
= to_book3s(vcpu
);
115 u32 page
, hash
, pteg
, htabmask
;
118 page
= (eaddr
& 0x0FFFFFFF) >> 12;
119 htabmask
= ((vcpu_book3s
->sdr1
& 0x1FF) << 16) | 0xFFC0;
121 hash
= ((sr_vsid(sre
) ^ page
) << 6);
126 pteg
= (vcpu_book3s
->sdr1
& 0xffff0000) | hash
;
128 dprintk("MMU: pc=0x%lx eaddr=0x%lx sdr1=0x%llx pteg=0x%x vsid=0x%x\n",
129 kvmppc_get_pc(vcpu
), eaddr
, vcpu_book3s
->sdr1
, pteg
,
132 r
= gfn_to_hva(vcpu
->kvm
, pteg
>> PAGE_SHIFT
);
133 if (kvm_is_error_hva(r
))
135 return r
| (pteg
& ~PAGE_MASK
);
138 static u32
kvmppc_mmu_book3s_32_get_ptem(u32 sre
, gva_t eaddr
, bool primary
)
140 return ((eaddr
& 0x0fffffff) >> 22) | (sr_vsid(sre
) << 7) |
141 (primary
? 0 : 0x40) | 0x80000000;
144 static int kvmppc_mmu_book3s_32_xlate_bat(struct kvm_vcpu
*vcpu
, gva_t eaddr
,
145 struct kvmppc_pte
*pte
, bool data
,
148 struct kvmppc_vcpu_book3s
*vcpu_book3s
= to_book3s(vcpu
);
149 struct kvmppc_bat
*bat
;
152 for (i
= 0; i
< 8; i
++) {
154 bat
= &vcpu_book3s
->dbat
[i
];
156 bat
= &vcpu_book3s
->ibat
[i
];
158 if (kvmppc_get_msr(vcpu
) & MSR_PR
) {
166 if (check_debug_ip(vcpu
))
168 dprintk_pte("%cBAT %02d: 0x%lx - 0x%x (0x%x)\n",
169 data
? 'd' : 'i', i
, eaddr
, bat
->bepi
,
172 if ((eaddr
& bat
->bepi_mask
) == bat
->bepi
) {
174 kvmppc_mmu_book3s_32_esid_to_vsid(vcpu
,
175 eaddr
>> SID_SHIFT
, &vsid
);
177 pte
->vpage
= (((u64
)eaddr
>> 12) & 0xffff) | vsid
;
179 pte
->raddr
= bat
->brpn
| (eaddr
& ~bat
->bepi_mask
);
180 pte
->may_read
= bat
->pp
;
181 pte
->may_write
= bat
->pp
> 1;
182 pte
->may_execute
= true;
183 if (!pte
->may_read
) {
184 printk(KERN_INFO
"BAT is not readable!\n");
187 if (iswrite
&& !pte
->may_write
) {
188 dprintk_pte("BAT is read-only!\n");
199 static int kvmppc_mmu_book3s_32_xlate_pte(struct kvm_vcpu
*vcpu
, gva_t eaddr
,
200 struct kvmppc_pte
*pte
, bool data
,
201 bool iswrite
, bool primary
)
211 sre
= find_sr(vcpu
, eaddr
);
213 dprintk_pte("SR 0x%lx: vsid=0x%x, raw=0x%x\n", eaddr
>> 28,
216 pte
->vpage
= kvmppc_mmu_book3s_32_ea_to_vp(vcpu
, eaddr
, data
);
218 ptegp
= kvmppc_mmu_book3s_32_get_pteg(vcpu
, sre
, eaddr
, primary
);
219 if (kvm_is_error_hva(ptegp
)) {
220 printk(KERN_INFO
"KVM: Invalid PTEG!\n");
224 ptem
= kvmppc_mmu_book3s_32_get_ptem(sre
, eaddr
, primary
);
226 if(copy_from_user(pteg
, (void __user
*)ptegp
, sizeof(pteg
))) {
227 printk(KERN_ERR
"KVM: Can't copy data from 0x%lx!\n", ptegp
);
231 for (i
=0; i
<16; i
+=2) {
232 pte0
= be32_to_cpu(pteg
[i
]);
233 pte1
= be32_to_cpu(pteg
[i
+ 1]);
237 pte
->raddr
= (pte1
& ~(0xFFFULL
)) | (eaddr
& 0xFFF);
240 if ((sr_kp(sre
) && (kvmppc_get_msr(vcpu
) & MSR_PR
)) ||
241 (sr_ks(sre
) && !(kvmppc_get_msr(vcpu
) & MSR_PR
)))
244 pte
->may_write
= false;
245 pte
->may_read
= false;
246 pte
->may_execute
= true;
252 pte
->may_write
= true;
256 pte
->may_read
= true;
260 dprintk_pte("MMU: Found PTE -> %x %x - %x\n",
267 /* Update PTE C and A bits, so the guest's swapper knows we used the
271 char __user
*addr
= (char __user
*) (ptegp
+ (i
+1) * sizeof(u32
));
274 * Use single-byte writes to update the HPTE, to
275 * conform to what real hardware does.
277 if (pte
->may_read
&& !(pte_r
& PTEG_FLAG_ACCESSED
)) {
278 pte_r
|= PTEG_FLAG_ACCESSED
;
279 put_user(pte_r
>> 8, addr
+ 2);
281 if (iswrite
&& pte
->may_write
&& !(pte_r
& PTEG_FLAG_DIRTY
)) {
282 pte_r
|= PTEG_FLAG_DIRTY
;
283 put_user(pte_r
, addr
+ 3);
285 if (!pte
->may_read
|| (iswrite
&& !pte
->may_write
))
292 if (check_debug_ip(vcpu
)) {
293 dprintk_pte("KVM MMU: No PTE found (sdr1=0x%llx ptegp=0x%lx)\n",
294 to_book3s(vcpu
)->sdr1
, ptegp
);
295 for (i
=0; i
<16; i
+=2) {
296 dprintk_pte(" %02d: 0x%x - 0x%x (0x%x)\n",
297 i
, be32_to_cpu(pteg
[i
]),
298 be32_to_cpu(pteg
[i
+1]), ptem
);
305 static int kvmppc_mmu_book3s_32_xlate(struct kvm_vcpu
*vcpu
, gva_t eaddr
,
306 struct kvmppc_pte
*pte
, bool data
,
310 ulong mp_ea
= vcpu
->arch
.magic_page_ea
;
313 pte
->page_size
= MMU_PAGE_4K
;
315 /* Magic page override */
316 if (unlikely(mp_ea
) &&
317 unlikely((eaddr
& ~0xfffULL
) == (mp_ea
& ~0xfffULL
)) &&
318 !(kvmppc_get_msr(vcpu
) & MSR_PR
)) {
319 pte
->vpage
= kvmppc_mmu_book3s_32_ea_to_vp(vcpu
, eaddr
, data
);
320 pte
->raddr
= vcpu
->arch
.magic_page_pa
| (pte
->raddr
& 0xfff);
321 pte
->raddr
&= KVM_PAM
;
322 pte
->may_execute
= true;
323 pte
->may_read
= true;
324 pte
->may_write
= true;
329 r
= kvmppc_mmu_book3s_32_xlate_bat(vcpu
, eaddr
, pte
, data
, iswrite
);
331 r
= kvmppc_mmu_book3s_32_xlate_pte(vcpu
, eaddr
, pte
,
332 data
, iswrite
, true);
334 r
= kvmppc_mmu_book3s_32_xlate_pte(vcpu
, eaddr
, pte
,
335 data
, iswrite
, false);
341 static u32
kvmppc_mmu_book3s_32_mfsrin(struct kvm_vcpu
*vcpu
, u32 srnum
)
343 return kvmppc_get_sr(vcpu
, srnum
);
346 static void kvmppc_mmu_book3s_32_mtsrin(struct kvm_vcpu
*vcpu
, u32 srnum
,
349 kvmppc_set_sr(vcpu
, srnum
, value
);
350 kvmppc_mmu_map_segment(vcpu
, srnum
<< SID_SHIFT
);
353 static void kvmppc_mmu_book3s_32_tlbie(struct kvm_vcpu
*vcpu
, ulong ea
, bool large
)
358 /* flush this VA on all cpus */
359 kvm_for_each_vcpu(i
, v
, vcpu
->kvm
)
360 kvmppc_mmu_pte_flush(v
, ea
, 0x0FFFF000);
363 static int kvmppc_mmu_book3s_32_esid_to_vsid(struct kvm_vcpu
*vcpu
, ulong esid
,
366 ulong ea
= esid
<< SID_SHIFT
;
369 u64 msr
= kvmppc_get_msr(vcpu
);
371 if (msr
& (MSR_DR
|MSR_IR
)) {
372 sr
= find_sr(vcpu
, ea
);
377 /* In case we only have one of MSR_IR or MSR_DR set, let's put
378 that in the real-mode context (and hope RM doesn't access
380 switch (msr
& (MSR_DR
|MSR_IR
)) {
382 *vsid
= VSID_REAL
| esid
;
385 *vsid
= VSID_REAL_IR
| gvsid
;
388 *vsid
= VSID_REAL_DR
| gvsid
;
394 *vsid
= VSID_BAT
| gvsid
;
406 static bool kvmppc_mmu_book3s_32_is_dcbz32(struct kvm_vcpu
*vcpu
)
412 void kvmppc_mmu_book3s_32_init(struct kvm_vcpu
*vcpu
)
414 struct kvmppc_mmu
*mmu
= &vcpu
->arch
.mmu
;
416 mmu
->mtsrin
= kvmppc_mmu_book3s_32_mtsrin
;
417 mmu
->mfsrin
= kvmppc_mmu_book3s_32_mfsrin
;
418 mmu
->xlate
= kvmppc_mmu_book3s_32_xlate
;
419 mmu
->reset_msr
= kvmppc_mmu_book3s_32_reset_msr
;
420 mmu
->tlbie
= kvmppc_mmu_book3s_32_tlbie
;
421 mmu
->esid_to_vsid
= kvmppc_mmu_book3s_32_esid_to_vsid
;
422 mmu
->ea_to_vp
= kvmppc_mmu_book3s_32_ea_to_vp
;
423 mmu
->is_dcbz32
= kvmppc_mmu_book3s_32_is_dcbz32
;