2 * Copyright (C) 2010,2012 Freescale Semiconductor, Inc. All rights reserved.
4 * Author: Varun Sethi, <varun.sethi@freescale.com>
7 * This file is derived from arch/powerpc/kvm/e500.c,
8 * by Yu Liu <yu.liu@freescale.com>.
10 * This program is free software; you can redistribute it and/or modify
11 * it under the terms of the GNU General Public License, version 2, as
12 * published by the Free Software Foundation.
15 #include <linux/kvm_host.h>
16 #include <linux/slab.h>
17 #include <linux/err.h>
18 #include <linux/export.h>
19 #include <linux/miscdevice.h>
20 #include <linux/module.h>
23 #include <asm/cputable.h>
24 #include <asm/tlbflush.h>
25 #include <asm/kvm_ppc.h>
26 #include <asm/dbell.h>
31 void kvmppc_set_pending_interrupt(struct kvm_vcpu
*vcpu
, enum int_class type
)
33 enum ppc_dbell dbell_type
;
37 case INT_CLASS_NONCRIT
:
38 dbell_type
= PPC_G_DBELL
;
41 dbell_type
= PPC_G_DBELL_CRIT
;
44 dbell_type
= PPC_G_DBELL_MC
;
47 WARN_ONCE(1, "%s: unknown int type %d\n", __func__
, type
);
52 tag
= PPC_DBELL_LPID(get_lpid(vcpu
)) | vcpu
->vcpu_id
;
54 ppc_msgsnd(dbell_type
, 0, tag
);
58 /* gtlbe must not be mapped by more than one host tlb entry */
59 void kvmppc_e500_tlbil_one(struct kvmppc_vcpu_e500
*vcpu_e500
,
60 struct kvm_book3e_206_tlb_entry
*gtlbe
)
67 ts
= get_tlb_ts(gtlbe
);
68 tid
= get_tlb_tid(gtlbe
);
70 /* We search the host TLB to invalidate its shadow TLB entry */
71 val
= (tid
<< 16) | ts
;
72 eaddr
= get_tlb_eaddr(gtlbe
);
74 local_irq_save(flags
);
76 mtspr(SPRN_MAS6
, val
);
77 mtspr(SPRN_MAS5
, MAS5_SGS
| get_lpid(&vcpu_e500
->vcpu
));
79 asm volatile("tlbsx 0, %[eaddr]\n" : : [eaddr
] "r" (eaddr
));
80 val
= mfspr(SPRN_MAS1
);
81 if (val
& MAS1_VALID
) {
82 mtspr(SPRN_MAS1
, val
& ~MAS1_VALID
);
83 asm volatile("tlbwe");
86 /* NOTE: tlbsx also updates mas8, so clear it for host tlbwe */
90 local_irq_restore(flags
);
93 void kvmppc_e500_tlbil_all(struct kvmppc_vcpu_e500
*vcpu_e500
)
97 local_irq_save(flags
);
98 mtspr(SPRN_MAS5
, MAS5_SGS
| get_lpid(&vcpu_e500
->vcpu
));
99 asm volatile("tlbilxlpid");
101 local_irq_restore(flags
);
104 void kvmppc_set_pid(struct kvm_vcpu
*vcpu
, u32 pid
)
106 vcpu
->arch
.pid
= pid
;
109 void kvmppc_mmu_msr_notify(struct kvm_vcpu
*vcpu
, u32 old_msr
)
113 /* We use two lpids per VM */
114 static DEFINE_PER_CPU(struct kvm_vcpu
*[KVMPPC_NR_LPIDS
], last_vcpu_of_lpid
);
116 static void kvmppc_core_vcpu_load_e500mc(struct kvm_vcpu
*vcpu
, int cpu
)
118 struct kvmppc_vcpu_e500
*vcpu_e500
= to_e500(vcpu
);
120 kvmppc_booke_vcpu_load(vcpu
, cpu
);
122 mtspr(SPRN_LPID
, get_lpid(vcpu
));
123 mtspr(SPRN_EPCR
, vcpu
->arch
.shadow_epcr
);
124 mtspr(SPRN_GPIR
, vcpu
->vcpu_id
);
125 mtspr(SPRN_MSRP
, vcpu
->arch
.shadow_msrp
);
126 vcpu
->arch
.eplc
= EPC_EGS
| (get_lpid(vcpu
) << EPC_ELPID_SHIFT
);
127 vcpu
->arch
.epsc
= vcpu
->arch
.eplc
;
128 mtspr(SPRN_EPLC
, vcpu
->arch
.eplc
);
129 mtspr(SPRN_EPSC
, vcpu
->arch
.epsc
);
131 mtspr(SPRN_GIVPR
, vcpu
->arch
.ivpr
);
132 mtspr(SPRN_GIVOR2
, vcpu
->arch
.ivor
[BOOKE_IRQPRIO_DATA_STORAGE
]);
133 mtspr(SPRN_GIVOR8
, vcpu
->arch
.ivor
[BOOKE_IRQPRIO_SYSCALL
]);
134 mtspr(SPRN_GSPRG0
, (unsigned long)vcpu
->arch
.shared
->sprg0
);
135 mtspr(SPRN_GSPRG1
, (unsigned long)vcpu
->arch
.shared
->sprg1
);
136 mtspr(SPRN_GSPRG2
, (unsigned long)vcpu
->arch
.shared
->sprg2
);
137 mtspr(SPRN_GSPRG3
, (unsigned long)vcpu
->arch
.shared
->sprg3
);
139 mtspr(SPRN_GSRR0
, vcpu
->arch
.shared
->srr0
);
140 mtspr(SPRN_GSRR1
, vcpu
->arch
.shared
->srr1
);
142 mtspr(SPRN_GEPR
, vcpu
->arch
.epr
);
143 mtspr(SPRN_GDEAR
, vcpu
->arch
.shared
->dar
);
144 mtspr(SPRN_GESR
, vcpu
->arch
.shared
->esr
);
146 if (vcpu
->arch
.oldpir
!= mfspr(SPRN_PIR
) ||
147 __this_cpu_read(last_vcpu_of_lpid
[get_lpid(vcpu
)]) != vcpu
) {
148 kvmppc_e500_tlbil_all(vcpu_e500
);
149 __this_cpu_write(last_vcpu_of_lpid
[get_lpid(vcpu
)], vcpu
);
153 static void kvmppc_core_vcpu_put_e500mc(struct kvm_vcpu
*vcpu
)
155 vcpu
->arch
.eplc
= mfspr(SPRN_EPLC
);
156 vcpu
->arch
.epsc
= mfspr(SPRN_EPSC
);
158 vcpu
->arch
.shared
->sprg0
= mfspr(SPRN_GSPRG0
);
159 vcpu
->arch
.shared
->sprg1
= mfspr(SPRN_GSPRG1
);
160 vcpu
->arch
.shared
->sprg2
= mfspr(SPRN_GSPRG2
);
161 vcpu
->arch
.shared
->sprg3
= mfspr(SPRN_GSPRG3
);
163 vcpu
->arch
.shared
->srr0
= mfspr(SPRN_GSRR0
);
164 vcpu
->arch
.shared
->srr1
= mfspr(SPRN_GSRR1
);
166 vcpu
->arch
.epr
= mfspr(SPRN_GEPR
);
167 vcpu
->arch
.shared
->dar
= mfspr(SPRN_GDEAR
);
168 vcpu
->arch
.shared
->esr
= mfspr(SPRN_GESR
);
170 vcpu
->arch
.oldpir
= mfspr(SPRN_PIR
);
172 kvmppc_booke_vcpu_put(vcpu
);
175 int kvmppc_core_check_processor_compat(void)
179 if (strcmp(cur_cpu_spec
->cpu_name
, "e500mc") == 0)
181 else if (strcmp(cur_cpu_spec
->cpu_name
, "e5500") == 0)
183 #ifdef CONFIG_ALTIVEC
185 * Since guests have the priviledge to enable AltiVec, we need AltiVec
186 * support in the host to save/restore their context.
187 * Don't use CPU_FTR_ALTIVEC to identify cores with AltiVec unit
188 * because it's cleared in the absence of CONFIG_ALTIVEC!
190 else if (strcmp(cur_cpu_spec
->cpu_name
, "e6500") == 0)
199 int kvmppc_core_vcpu_setup(struct kvm_vcpu
*vcpu
)
201 struct kvmppc_vcpu_e500
*vcpu_e500
= to_e500(vcpu
);
203 vcpu
->arch
.shadow_epcr
= SPRN_EPCR_DSIGS
| SPRN_EPCR_DGTMI
| \
206 vcpu
->arch
.shadow_epcr
|= SPRN_EPCR_ICM
;
208 vcpu
->arch
.shadow_msrp
= MSRP_UCLEP
| MSRP_PMMP
;
210 vcpu
->arch
.pvr
= mfspr(SPRN_PVR
);
211 vcpu_e500
->svr
= mfspr(SPRN_SVR
);
213 vcpu
->arch
.cpu_type
= KVM_CPU_E500MC
;
218 static int kvmppc_core_get_sregs_e500mc(struct kvm_vcpu
*vcpu
,
219 struct kvm_sregs
*sregs
)
221 struct kvmppc_vcpu_e500
*vcpu_e500
= to_e500(vcpu
);
223 sregs
->u
.e
.features
|= KVM_SREGS_E_ARCH206_MMU
| KVM_SREGS_E_PM
|
225 sregs
->u
.e
.impl_id
= KVM_SREGS_E_IMPL_FSL
;
227 sregs
->u
.e
.impl
.fsl
.features
= 0;
228 sregs
->u
.e
.impl
.fsl
.svr
= vcpu_e500
->svr
;
229 sregs
->u
.e
.impl
.fsl
.hid0
= vcpu_e500
->hid0
;
230 sregs
->u
.e
.impl
.fsl
.mcar
= vcpu_e500
->mcar
;
232 kvmppc_get_sregs_e500_tlb(vcpu
, sregs
);
234 sregs
->u
.e
.ivor_high
[3] =
235 vcpu
->arch
.ivor
[BOOKE_IRQPRIO_PERFORMANCE_MONITOR
];
236 sregs
->u
.e
.ivor_high
[4] = vcpu
->arch
.ivor
[BOOKE_IRQPRIO_DBELL
];
237 sregs
->u
.e
.ivor_high
[5] = vcpu
->arch
.ivor
[BOOKE_IRQPRIO_DBELL_CRIT
];
239 return kvmppc_get_sregs_ivor(vcpu
, sregs
);
242 static int kvmppc_core_set_sregs_e500mc(struct kvm_vcpu
*vcpu
,
243 struct kvm_sregs
*sregs
)
245 struct kvmppc_vcpu_e500
*vcpu_e500
= to_e500(vcpu
);
248 if (sregs
->u
.e
.impl_id
== KVM_SREGS_E_IMPL_FSL
) {
249 vcpu_e500
->svr
= sregs
->u
.e
.impl
.fsl
.svr
;
250 vcpu_e500
->hid0
= sregs
->u
.e
.impl
.fsl
.hid0
;
251 vcpu_e500
->mcar
= sregs
->u
.e
.impl
.fsl
.mcar
;
254 ret
= kvmppc_set_sregs_e500_tlb(vcpu
, sregs
);
258 if (!(sregs
->u
.e
.features
& KVM_SREGS_E_IVOR
))
261 if (sregs
->u
.e
.features
& KVM_SREGS_E_PM
) {
262 vcpu
->arch
.ivor
[BOOKE_IRQPRIO_PERFORMANCE_MONITOR
] =
263 sregs
->u
.e
.ivor_high
[3];
266 if (sregs
->u
.e
.features
& KVM_SREGS_E_PC
) {
267 vcpu
->arch
.ivor
[BOOKE_IRQPRIO_DBELL
] =
268 sregs
->u
.e
.ivor_high
[4];
269 vcpu
->arch
.ivor
[BOOKE_IRQPRIO_DBELL_CRIT
] =
270 sregs
->u
.e
.ivor_high
[5];
273 return kvmppc_set_sregs_ivor(vcpu
, sregs
);
276 static int kvmppc_get_one_reg_e500mc(struct kvm_vcpu
*vcpu
, u64 id
,
277 union kvmppc_one_reg
*val
)
282 case KVM_REG_PPC_SPRG9
:
283 *val
= get_reg_val(id
, vcpu
->arch
.sprg9
);
286 r
= kvmppc_get_one_reg_e500_tlb(vcpu
, id
, val
);
292 static int kvmppc_set_one_reg_e500mc(struct kvm_vcpu
*vcpu
, u64 id
,
293 union kvmppc_one_reg
*val
)
298 case KVM_REG_PPC_SPRG9
:
299 vcpu
->arch
.sprg9
= set_reg_val(id
, *val
);
302 r
= kvmppc_set_one_reg_e500_tlb(vcpu
, id
, val
);
308 static struct kvm_vcpu
*kvmppc_core_vcpu_create_e500mc(struct kvm
*kvm
,
311 struct kvmppc_vcpu_e500
*vcpu_e500
;
312 struct kvm_vcpu
*vcpu
;
315 vcpu_e500
= kmem_cache_zalloc(kvm_vcpu_cache
, GFP_KERNEL
);
320 vcpu
= &vcpu_e500
->vcpu
;
322 /* Invalid PIR value -- this LPID dosn't have valid state on any cpu */
323 vcpu
->arch
.oldpir
= 0xffffffff;
325 err
= kvm_vcpu_init(vcpu
, kvm
, id
);
329 err
= kvmppc_e500_tlb_init(vcpu_e500
);
333 vcpu
->arch
.shared
= (void *)__get_free_page(GFP_KERNEL
| __GFP_ZERO
);
334 if (!vcpu
->arch
.shared
)
340 kvmppc_e500_tlb_uninit(vcpu_e500
);
342 kvm_vcpu_uninit(vcpu
);
345 kmem_cache_free(kvm_vcpu_cache
, vcpu_e500
);
350 static void kvmppc_core_vcpu_free_e500mc(struct kvm_vcpu
*vcpu
)
352 struct kvmppc_vcpu_e500
*vcpu_e500
= to_e500(vcpu
);
354 free_page((unsigned long)vcpu
->arch
.shared
);
355 kvmppc_e500_tlb_uninit(vcpu_e500
);
356 kvm_vcpu_uninit(vcpu
);
357 kmem_cache_free(kvm_vcpu_cache
, vcpu_e500
);
360 static int kvmppc_core_init_vm_e500mc(struct kvm
*kvm
)
364 lpid
= kvmppc_alloc_lpid();
369 * Use two lpids per VM on cores with two threads like e6500. Use
370 * even numbers to speedup vcpu lpid computation with consecutive lpids
371 * per VM. vm1 will use lpids 2 and 3, vm2 lpids 4 and 5, and so on.
373 if (threads_per_core
== 2)
376 kvm
->arch
.lpid
= lpid
;
380 static void kvmppc_core_destroy_vm_e500mc(struct kvm
*kvm
)
382 int lpid
= kvm
->arch
.lpid
;
384 if (threads_per_core
== 2)
387 kvmppc_free_lpid(lpid
);
390 static struct kvmppc_ops kvm_ops_e500mc
= {
391 .get_sregs
= kvmppc_core_get_sregs_e500mc
,
392 .set_sregs
= kvmppc_core_set_sregs_e500mc
,
393 .get_one_reg
= kvmppc_get_one_reg_e500mc
,
394 .set_one_reg
= kvmppc_set_one_reg_e500mc
,
395 .vcpu_load
= kvmppc_core_vcpu_load_e500mc
,
396 .vcpu_put
= kvmppc_core_vcpu_put_e500mc
,
397 .vcpu_create
= kvmppc_core_vcpu_create_e500mc
,
398 .vcpu_free
= kvmppc_core_vcpu_free_e500mc
,
399 .mmu_destroy
= kvmppc_mmu_destroy_e500
,
400 .init_vm
= kvmppc_core_init_vm_e500mc
,
401 .destroy_vm
= kvmppc_core_destroy_vm_e500mc
,
402 .emulate_op
= kvmppc_core_emulate_op_e500
,
403 .emulate_mtspr
= kvmppc_core_emulate_mtspr_e500
,
404 .emulate_mfspr
= kvmppc_core_emulate_mfspr_e500
,
407 static int __init
kvmppc_e500mc_init(void)
411 r
= kvmppc_booke_init();
416 * Use two lpids per VM on dual threaded processors like e6500
417 * to workarround the lack of tlb write conditional instruction.
418 * Expose half the number of available hardware lpids to the lpid
421 kvmppc_init_lpid(KVMPPC_NR_LPIDS
/threads_per_core
);
422 kvmppc_claim_lpid(0); /* host */
424 r
= kvm_init(NULL
, sizeof(struct kvmppc_vcpu_e500
), 0, THIS_MODULE
);
427 kvm_ops_e500mc
.owner
= THIS_MODULE
;
428 kvmppc_pr_ops
= &kvm_ops_e500mc
;
434 static void __exit
kvmppc_e500mc_exit(void)
436 kvmppc_pr_ops
= NULL
;
440 module_init(kvmppc_e500mc_init
);
441 module_exit(kvmppc_e500mc_exit
);
442 MODULE_ALIAS_MISCDEV(KVM_MINOR
);
443 MODULE_ALIAS("devname:kvm");