1 // SPDX-License-Identifier: GPL-2.0
3 * in-kernel handling for sie intercepts
5 * Copyright IBM Corp. 2008, 2020
7 * Author(s): Carsten Otte <cotte@de.ibm.com>
8 * Christian Borntraeger <borntraeger@de.ibm.com>
11 #include <linux/kvm_host.h>
12 #include <linux/errno.h>
13 #include <linux/pagemap.h>
15 #include <asm/asm-offsets.h>
17 #include <asm/sysinfo.h>
23 #include "trace-s390.h"
25 u8
kvm_s390_get_ilen(struct kvm_vcpu
*vcpu
)
27 struct kvm_s390_sie_block
*sie_block
= vcpu
->arch
.sie_block
;
30 switch (vcpu
->arch
.sie_block
->icptcode
) {
36 /* instruction only stored for these icptcodes */
37 ilen
= insn_length(vcpu
->arch
.sie_block
->ipa
>> 8);
38 /* Use the length of the EXECUTE instruction if necessary */
39 if (sie_block
->icptstatus
& 1) {
40 ilen
= (sie_block
->icptstatus
>> 4) & 0x6;
46 /* bit 1+2 of pgmilc are the ilc, so we directly get ilen */
47 ilen
= vcpu
->arch
.sie_block
->pgmilc
& 0x6;
53 static int handle_stop(struct kvm_vcpu
*vcpu
)
55 struct kvm_s390_local_interrupt
*li
= &vcpu
->arch
.local_int
;
57 uint8_t flags
, stop_pending
;
59 vcpu
->stat
.exit_stop_request
++;
61 /* delay the stop if any non-stop irq is pending */
62 if (kvm_s390_vcpu_has_irq(vcpu
, 1))
65 /* avoid races with the injection/SIGP STOP code */
67 flags
= li
->irq
.stop
.flags
;
68 stop_pending
= kvm_s390_is_stop_irq_pending(vcpu
);
69 spin_unlock(&li
->lock
);
71 trace_kvm_s390_stop_request(stop_pending
, flags
);
75 if (flags
& KVM_S390_STOP_FLAG_STORE_STATUS
) {
76 rc
= kvm_s390_vcpu_store_status(vcpu
,
77 KVM_S390_STORE_STATUS_NOADDR
);
83 * no need to check the return value of vcpu_stop as it can only have
84 * an error for protvirt, but protvirt means user cpu state
86 if (!kvm_s390_user_cpu_state_ctrl(vcpu
->kvm
))
87 kvm_s390_vcpu_stop(vcpu
);
91 static int handle_validity(struct kvm_vcpu
*vcpu
)
93 int viwhy
= vcpu
->arch
.sie_block
->ipb
>> 16;
95 vcpu
->stat
.exit_validity
++;
96 trace_kvm_s390_intercept_validity(vcpu
, viwhy
);
97 KVM_EVENT(3, "validity intercept 0x%x for pid %u (kvm 0x%pK)", viwhy
,
98 current
->pid
, vcpu
->kvm
);
100 /* do not warn on invalid runtime instrumentation mode */
101 WARN_ONCE(viwhy
!= 0x44, "kvm: unhandled validity intercept 0x%x\n",
106 static int handle_instruction(struct kvm_vcpu
*vcpu
)
108 vcpu
->stat
.exit_instruction
++;
109 trace_kvm_s390_intercept_instruction(vcpu
,
110 vcpu
->arch
.sie_block
->ipa
,
111 vcpu
->arch
.sie_block
->ipb
);
113 switch (vcpu
->arch
.sie_block
->ipa
>> 8) {
115 return kvm_s390_handle_01(vcpu
);
117 return kvm_s390_handle_lpsw(vcpu
);
119 return kvm_s390_handle_diag(vcpu
);
121 return kvm_s390_handle_aa(vcpu
);
123 return kvm_s390_handle_sigp(vcpu
);
125 return kvm_s390_handle_b2(vcpu
);
127 return kvm_s390_handle_stctl(vcpu
);
129 return kvm_s390_handle_lctl(vcpu
);
131 return kvm_s390_handle_b9(vcpu
);
133 return kvm_s390_handle_e3(vcpu
);
135 return kvm_s390_handle_e5(vcpu
);
137 return kvm_s390_handle_eb(vcpu
);
143 static int inject_prog_on_prog_intercept(struct kvm_vcpu
*vcpu
)
145 struct kvm_s390_pgm_info pgm_info
= {
146 .code
= vcpu
->arch
.sie_block
->iprcc
,
147 /* the PSW has already been rewound */
148 .flags
= KVM_S390_PGM_FLAGS_NO_REWIND
,
151 switch (vcpu
->arch
.sie_block
->iprcc
& ~PGM_PER
) {
152 case PGM_AFX_TRANSLATION
:
153 case PGM_ASX_TRANSLATION
:
154 case PGM_EX_TRANSLATION
:
155 case PGM_LFX_TRANSLATION
:
156 case PGM_LSTE_SEQUENCE
:
157 case PGM_LSX_TRANSLATION
:
158 case PGM_LX_TRANSLATION
:
159 case PGM_PRIMARY_AUTHORITY
:
160 case PGM_SECONDARY_AUTHORITY
:
161 case PGM_SPACE_SWITCH
:
162 pgm_info
.trans_exc_code
= vcpu
->arch
.sie_block
->tecmc
;
164 case PGM_ALEN_TRANSLATION
:
165 case PGM_ALE_SEQUENCE
:
166 case PGM_ASTE_INSTANCE
:
167 case PGM_ASTE_SEQUENCE
:
168 case PGM_ASTE_VALIDITY
:
169 case PGM_EXTENDED_AUTHORITY
:
170 pgm_info
.exc_access_id
= vcpu
->arch
.sie_block
->eai
;
173 case PGM_PAGE_TRANSLATION
:
174 case PGM_REGION_FIRST_TRANS
:
175 case PGM_REGION_SECOND_TRANS
:
176 case PGM_REGION_THIRD_TRANS
:
177 case PGM_SEGMENT_TRANSLATION
:
178 pgm_info
.trans_exc_code
= vcpu
->arch
.sie_block
->tecmc
;
179 pgm_info
.exc_access_id
= vcpu
->arch
.sie_block
->eai
;
180 pgm_info
.op_access_id
= vcpu
->arch
.sie_block
->oai
;
183 pgm_info
.mon_class_nr
= vcpu
->arch
.sie_block
->mcn
;
184 pgm_info
.mon_code
= vcpu
->arch
.sie_block
->tecmc
;
186 case PGM_VECTOR_PROCESSING
:
188 pgm_info
.data_exc_code
= vcpu
->arch
.sie_block
->dxc
;
191 pgm_info
.trans_exc_code
= vcpu
->arch
.sie_block
->tecmc
;
192 pgm_info
.exc_access_id
= vcpu
->arch
.sie_block
->eai
;
198 if (vcpu
->arch
.sie_block
->iprcc
& PGM_PER
) {
199 pgm_info
.per_code
= vcpu
->arch
.sie_block
->perc
;
200 pgm_info
.per_atmid
= vcpu
->arch
.sie_block
->peratmid
;
201 pgm_info
.per_address
= vcpu
->arch
.sie_block
->peraddr
;
202 pgm_info
.per_access_id
= vcpu
->arch
.sie_block
->peraid
;
204 return kvm_s390_inject_prog_irq(vcpu
, &pgm_info
);
208 * restore ITDB to program-interruption TDB in guest lowcore
209 * and set TX abort indication if required
211 static int handle_itdb(struct kvm_vcpu
*vcpu
)
213 struct kvm_s390_itdb
*itdb
;
216 if (!IS_TE_ENABLED(vcpu
) || !IS_ITDB_VALID(vcpu
))
218 if (current
->thread
.per_flags
& PER_FLAG_NO_TE
)
220 itdb
= (struct kvm_s390_itdb
*)vcpu
->arch
.sie_block
->itdba
;
221 rc
= write_guest_lc(vcpu
, __LC_PGM_TDB
, itdb
, sizeof(*itdb
));
224 memset(itdb
, 0, sizeof(*itdb
));
229 #define per_event(vcpu) (vcpu->arch.sie_block->iprcc & PGM_PER)
231 static int handle_prog(struct kvm_vcpu
*vcpu
)
236 vcpu
->stat
.exit_program_interruption
++;
239 * Intercept 8 indicates a loop of specification exceptions
240 * for protected guests.
242 if (kvm_s390_pv_cpu_is_protected(vcpu
))
245 if (guestdbg_enabled(vcpu
) && per_event(vcpu
)) {
246 rc
= kvm_s390_handle_per_event(vcpu
);
249 /* the interrupt might have been filtered out completely */
250 if (vcpu
->arch
.sie_block
->iprcc
== 0)
254 trace_kvm_s390_intercept_prog(vcpu
, vcpu
->arch
.sie_block
->iprcc
);
255 if (vcpu
->arch
.sie_block
->iprcc
== PGM_SPECIFICATION
) {
256 rc
= read_guest_lc(vcpu
, __LC_PGM_NEW_PSW
, &psw
, sizeof(psw_t
));
259 /* Avoid endless loops of specification exceptions */
260 if (!is_valid_psw(&psw
))
263 rc
= handle_itdb(vcpu
);
267 return inject_prog_on_prog_intercept(vcpu
);
271 * handle_external_interrupt - used for external interruption interceptions
273 * This interception only occurs if the CPUSTAT_EXT_INT bit was set, or if
274 * the new PSW does not have external interrupts disabled. In the first case,
275 * we've got to deliver the interrupt manually, and in the second case, we
276 * drop to userspace to handle the situation there.
278 static int handle_external_interrupt(struct kvm_vcpu
*vcpu
)
280 u16 eic
= vcpu
->arch
.sie_block
->eic
;
281 struct kvm_s390_irq irq
;
285 vcpu
->stat
.exit_external_interrupt
++;
287 rc
= read_guest_lc(vcpu
, __LC_EXT_NEW_PSW
, &newpsw
, sizeof(psw_t
));
290 /* We can not handle clock comparator or timer interrupt with bad PSW */
291 if ((eic
== EXT_IRQ_CLK_COMP
|| eic
== EXT_IRQ_CPU_TIMER
) &&
292 (newpsw
.mask
& PSW_MASK_EXT
))
296 case EXT_IRQ_CLK_COMP
:
297 irq
.type
= KVM_S390_INT_CLOCK_COMP
;
299 case EXT_IRQ_CPU_TIMER
:
300 irq
.type
= KVM_S390_INT_CPU_TIMER
;
302 case EXT_IRQ_EXTERNAL_CALL
:
303 irq
.type
= KVM_S390_INT_EXTERNAL_CALL
;
304 irq
.u
.extcall
.code
= vcpu
->arch
.sie_block
->extcpuaddr
;
305 rc
= kvm_s390_inject_vcpu(vcpu
, &irq
);
306 /* ignore if another external call is already pending */
314 return kvm_s390_inject_vcpu(vcpu
, &irq
);
318 * Handle MOVE PAGE partial execution interception.
320 * This interception can only happen for guests with DAT disabled and
321 * addresses that are currently not mapped in the host. Thus we try to
322 * set up the mappings for the corresponding user pages here (or throw
323 * addressing exceptions in case of illegal guest addresses).
325 static int handle_mvpg_pei(struct kvm_vcpu
*vcpu
)
327 unsigned long srcaddr
, dstaddr
;
330 kvm_s390_get_regs_rre(vcpu
, ®1
, ®2
);
332 /* Make sure that the source is paged-in */
333 rc
= guest_translate_address(vcpu
, vcpu
->run
->s
.regs
.gprs
[reg2
],
334 reg2
, &srcaddr
, GACC_FETCH
);
336 return kvm_s390_inject_prog_cond(vcpu
, rc
);
337 rc
= kvm_arch_fault_in_page(vcpu
, srcaddr
, 0);
341 /* Make sure that the destination is paged-in */
342 rc
= guest_translate_address(vcpu
, vcpu
->run
->s
.regs
.gprs
[reg1
],
343 reg1
, &dstaddr
, GACC_STORE
);
345 return kvm_s390_inject_prog_cond(vcpu
, rc
);
346 rc
= kvm_arch_fault_in_page(vcpu
, dstaddr
, 1);
350 kvm_s390_retry_instr(vcpu
);
355 static int handle_partial_execution(struct kvm_vcpu
*vcpu
)
357 vcpu
->stat
.exit_pei
++;
359 if (vcpu
->arch
.sie_block
->ipa
== 0xb254) /* MVPG */
360 return handle_mvpg_pei(vcpu
);
361 if (vcpu
->arch
.sie_block
->ipa
>> 8 == 0xae) /* SIGP */
362 return kvm_s390_handle_sigp_pei(vcpu
);
368 * Handle the sthyi instruction that provides the guest with system
369 * information, like current CPU resources available at each level of
372 int handle_sthyi(struct kvm_vcpu
*vcpu
)
374 int reg1
, reg2
, r
= 0;
375 u64 code
, addr
, cc
= 0, rc
= 0;
376 struct sthyi_sctns
*sctns
= NULL
;
378 if (!test_kvm_facility(vcpu
->kvm
, 74))
379 return kvm_s390_inject_program_int(vcpu
, PGM_OPERATION
);
381 kvm_s390_get_regs_rre(vcpu
, ®1
, ®2
);
382 code
= vcpu
->run
->s
.regs
.gprs
[reg1
];
383 addr
= vcpu
->run
->s
.regs
.gprs
[reg2
];
385 vcpu
->stat
.instruction_sthyi
++;
386 VCPU_EVENT(vcpu
, 3, "STHYI: fc: %llu addr: 0x%016llx", code
, addr
);
387 trace_kvm_s390_handle_sthyi(vcpu
, code
, addr
);
389 if (reg1
== reg2
|| reg1
& 1 || reg2
& 1)
390 return kvm_s390_inject_program_int(vcpu
, PGM_SPECIFICATION
);
398 if (!kvm_s390_pv_cpu_is_protected(vcpu
) && (addr
& ~PAGE_MASK
))
399 return kvm_s390_inject_program_int(vcpu
, PGM_SPECIFICATION
);
401 sctns
= (void *)get_zeroed_page(GFP_KERNEL
);
405 cc
= sthyi_fill(sctns
, &rc
);
409 if (kvm_s390_pv_cpu_is_protected(vcpu
)) {
410 memcpy((void *)(sida_origin(vcpu
->arch
.sie_block
)),
413 r
= write_guest(vcpu
, addr
, reg2
, sctns
, PAGE_SIZE
);
415 free_page((unsigned long)sctns
);
416 return kvm_s390_inject_prog_cond(vcpu
, r
);
421 free_page((unsigned long)sctns
);
422 vcpu
->run
->s
.regs
.gprs
[reg2
+ 1] = rc
;
423 kvm_s390_set_psw_cc(vcpu
, cc
);
427 static int handle_operexc(struct kvm_vcpu
*vcpu
)
429 psw_t oldpsw
, newpsw
;
432 vcpu
->stat
.exit_operation_exception
++;
433 trace_kvm_s390_handle_operexc(vcpu
, vcpu
->arch
.sie_block
->ipa
,
434 vcpu
->arch
.sie_block
->ipb
);
436 if (vcpu
->arch
.sie_block
->ipa
== 0xb256)
437 return handle_sthyi(vcpu
);
439 if (vcpu
->arch
.sie_block
->ipa
== 0 && vcpu
->kvm
->arch
.user_instr0
)
441 rc
= read_guest_lc(vcpu
, __LC_PGM_NEW_PSW
, &newpsw
, sizeof(psw_t
));
445 * Avoid endless loops of operation exceptions, if the pgm new
446 * PSW will cause a new operation exception.
447 * The heuristic checks if the pgm new psw is within 6 bytes before
448 * the faulting psw address (with same DAT, AS settings) and the
449 * new psw is not a wait psw and the fault was not triggered by
452 oldpsw
= vcpu
->arch
.sie_block
->gpsw
;
453 if (oldpsw
.addr
- newpsw
.addr
<= 6 &&
454 !(newpsw
.mask
& PSW_MASK_WAIT
) &&
455 !(oldpsw
.mask
& PSW_MASK_PSTATE
) &&
456 (newpsw
.mask
& PSW_MASK_ASC
) == (oldpsw
.mask
& PSW_MASK_ASC
) &&
457 (newpsw
.mask
& PSW_MASK_DAT
) == (oldpsw
.mask
& PSW_MASK_DAT
))
460 return kvm_s390_inject_program_int(vcpu
, PGM_OPERATION
);
463 static int handle_pv_spx(struct kvm_vcpu
*vcpu
)
465 u32 pref
= *(u32
*)vcpu
->arch
.sie_block
->sidad
;
467 kvm_s390_set_prefix(vcpu
, pref
);
468 trace_kvm_s390_handle_prefix(vcpu
, 1, pref
);
472 static int handle_pv_sclp(struct kvm_vcpu
*vcpu
)
474 struct kvm_s390_float_interrupt
*fi
= &vcpu
->kvm
->arch
.float_int
;
476 spin_lock(&fi
->lock
);
479 * a: an sccb answering interrupt was already pending or in flight.
480 * As the sccb value is not known we can simply set some value to
481 * trigger delivery of a saved SCCB. UV will then use its saved
482 * copy of the SCCB value.
483 * b: an error SCCB interrupt needs to be injected so we also inject
484 * a fake SCCB address. Firmware will use the proper one.
485 * This makes sure, that both errors and real sccb returns will only
486 * be delivered after a notification intercept (instruction has
487 * finished) but not after others.
489 fi
->srv_signal
.ext_params
|= 0x43000;
490 set_bit(IRQ_PEND_EXT_SERVICE
, &fi
->pending_irqs
);
491 clear_bit(IRQ_PEND_EXT_SERVICE
, &fi
->masked_irqs
);
492 spin_unlock(&fi
->lock
);
496 static int handle_pv_uvc(struct kvm_vcpu
*vcpu
)
498 struct uv_cb_share
*guest_uvcb
= (void *)vcpu
->arch
.sie_block
->sidad
;
499 struct uv_cb_cts uvcb
= {
500 .header
.cmd
= UVC_CMD_UNPIN_PAGE_SHARED
,
501 .header
.len
= sizeof(uvcb
),
502 .guest_handle
= kvm_s390_pv_get_handle(vcpu
->kvm
),
503 .gaddr
= guest_uvcb
->paddr
,
507 if (guest_uvcb
->header
.cmd
!= UVC_CMD_REMOVE_SHARED_ACCESS
) {
508 WARN_ONCE(1, "Unexpected notification intercept for UVC 0x%x\n",
509 guest_uvcb
->header
.cmd
);
512 rc
= gmap_make_secure(vcpu
->arch
.gmap
, uvcb
.gaddr
, &uvcb
);
514 * If the unpin did not succeed, the guest will exit again for the UVC
515 * and we will retry the unpin.
522 static int handle_pv_notification(struct kvm_vcpu
*vcpu
)
524 if (vcpu
->arch
.sie_block
->ipa
== 0xb210)
525 return handle_pv_spx(vcpu
);
526 if (vcpu
->arch
.sie_block
->ipa
== 0xb220)
527 return handle_pv_sclp(vcpu
);
528 if (vcpu
->arch
.sie_block
->ipa
== 0xb9a4)
529 return handle_pv_uvc(vcpu
);
531 return handle_instruction(vcpu
);
534 int kvm_handle_sie_intercept(struct kvm_vcpu
*vcpu
)
538 if (kvm_is_ucontrol(vcpu
->kvm
))
541 switch (vcpu
->arch
.sie_block
->icptcode
) {
543 vcpu
->stat
.exit_external_request
++;
546 vcpu
->stat
.exit_io_request
++;
549 rc
= handle_instruction(vcpu
);
552 return handle_prog(vcpu
);
554 return handle_external_interrupt(vcpu
);
556 return kvm_s390_handle_wait(vcpu
);
558 return handle_validity(vcpu
);
560 return handle_stop(vcpu
);
562 rc
= handle_operexc(vcpu
);
565 rc
= handle_partial_execution(vcpu
);
568 rc
= kvm_s390_skey_check_enable(vcpu
);
571 case ICPT_INT_ENABLE
:
573 * PSW bit 13 or a CR (0, 6, 14) changed and we might
574 * now be able to deliver interrupts. The pre-run code
575 * will take care of this.
580 rc
= handle_instruction(vcpu
);
583 rc
= handle_pv_notification(vcpu
);
587 gmap_convert_to_secure(vcpu
->arch
.gmap
,
588 kvm_s390_get_prefix(vcpu
));
589 gmap_convert_to_secure(vcpu
->arch
.gmap
,
590 kvm_s390_get_prefix(vcpu
) + PAGE_SIZE
);
596 /* process PER, also if the instrution is processed in user space */
597 if (vcpu
->arch
.sie_block
->icptstatus
& 0x02 &&
598 (!rc
|| rc
== -EOPNOTSUPP
))
599 per_rc
= kvm_s390_handle_per_ifetch_icpt(vcpu
);
600 return per_rc
? per_rc
: rc
;