2 * RISC-V Emulation Helpers for QEMU.
4 * Copyright (c) 2016-2017 Sagar Karandikar, sagark@eecs.berkeley.edu
5 * Copyright (c) 2017-2018 SiFive, Inc.
6 * Copyright (c) 2022 VRULL GmbH
8 * This program is free software; you can redistribute it and/or modify it
9 * under the terms and conditions of the GNU General Public License,
10 * version 2 or later, as published by the Free Software Foundation.
12 * This program is distributed in the hope it will be useful, but WITHOUT
13 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
14 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
17 * You should have received a copy of the GNU General Public License along with
18 * this program. If not, see <http://www.gnu.org/licenses/>.
21 #include "qemu/osdep.h"
23 #include "internals.h"
24 #include "exec/exec-all.h"
25 #include "exec/cpu_ldst.h"
26 #include "exec/helper-proto.h"
28 /* Exceptions processing helpers */
29 G_NORETURN
void riscv_raise_exception(CPURISCVState
*env
,
30 uint32_t exception
, uintptr_t pc
)
32 CPUState
*cs
= env_cpu(env
);
33 cs
->exception_index
= exception
;
34 cpu_loop_exit_restore(cs
, pc
);
37 void helper_raise_exception(CPURISCVState
*env
, uint32_t exception
)
39 riscv_raise_exception(env
, exception
, 0);
42 target_ulong
helper_csrr(CPURISCVState
*env
, int csr
)
45 * The seed CSR must be accessed with a read-write instruction. A
46 * read-only instruction such as CSRRS/CSRRC with rs1=x0 or CSRRSI/
47 * CSRRCI with uimm=0 will raise an illegal instruction exception.
49 if (csr
== CSR_SEED
) {
50 riscv_raise_exception(env
, RISCV_EXCP_ILLEGAL_INST
, GETPC());
54 RISCVException ret
= riscv_csrr(env
, csr
, &val
);
56 if (ret
!= RISCV_EXCP_NONE
) {
57 riscv_raise_exception(env
, ret
, GETPC());
62 void helper_csrw(CPURISCVState
*env
, int csr
, target_ulong src
)
64 target_ulong mask
= env
->xl
== MXL_RV32
? UINT32_MAX
: (target_ulong
)-1;
65 RISCVException ret
= riscv_csrrw(env
, csr
, NULL
, src
, mask
);
67 if (ret
!= RISCV_EXCP_NONE
) {
68 riscv_raise_exception(env
, ret
, GETPC());
72 target_ulong
helper_csrrw(CPURISCVState
*env
, int csr
,
73 target_ulong src
, target_ulong write_mask
)
76 RISCVException ret
= riscv_csrrw(env
, csr
, &val
, src
, write_mask
);
78 if (ret
!= RISCV_EXCP_NONE
) {
79 riscv_raise_exception(env
, ret
, GETPC());
84 target_ulong
helper_csrr_i128(CPURISCVState
*env
, int csr
)
86 Int128 rv
= int128_zero();
87 RISCVException ret
= riscv_csrr_i128(env
, csr
, &rv
);
89 if (ret
!= RISCV_EXCP_NONE
) {
90 riscv_raise_exception(env
, ret
, GETPC());
93 env
->retxh
= int128_gethi(rv
);
94 return int128_getlo(rv
);
97 void helper_csrw_i128(CPURISCVState
*env
, int csr
,
98 target_ulong srcl
, target_ulong srch
)
100 RISCVException ret
= riscv_csrrw_i128(env
, csr
, NULL
,
101 int128_make128(srcl
, srch
),
104 if (ret
!= RISCV_EXCP_NONE
) {
105 riscv_raise_exception(env
, ret
, GETPC());
109 target_ulong
helper_csrrw_i128(CPURISCVState
*env
, int csr
,
110 target_ulong srcl
, target_ulong srch
,
111 target_ulong maskl
, target_ulong maskh
)
113 Int128 rv
= int128_zero();
114 RISCVException ret
= riscv_csrrw_i128(env
, csr
, &rv
,
115 int128_make128(srcl
, srch
),
116 int128_make128(maskl
, maskh
));
118 if (ret
!= RISCV_EXCP_NONE
) {
119 riscv_raise_exception(env
, ret
, GETPC());
122 env
->retxh
= int128_gethi(rv
);
123 return int128_getlo(rv
);
130 * Raise virtual exceptions and illegal instruction exceptions for
131 * Zicbo[mz] instructions based on the settings of [mhs]envcfg as
132 * specified in section 2.5.1 of the CMO specification.
134 static void check_zicbo_envcfg(CPURISCVState
*env
, target_ulong envbits
,
137 #ifndef CONFIG_USER_ONLY
138 if ((env
->priv
< PRV_M
) && !get_field(env
->menvcfg
, envbits
)) {
139 riscv_raise_exception(env
, RISCV_EXCP_ILLEGAL_INST
, ra
);
142 if (env
->virt_enabled
&&
143 (((env
->priv
<= PRV_S
) && !get_field(env
->henvcfg
, envbits
)) ||
144 ((env
->priv
< PRV_S
) && !get_field(env
->senvcfg
, envbits
)))) {
145 riscv_raise_exception(env
, RISCV_EXCP_VIRT_INSTRUCTION_FAULT
, ra
);
148 if ((env
->priv
< PRV_S
) && !get_field(env
->senvcfg
, envbits
)) {
149 riscv_raise_exception(env
, RISCV_EXCP_ILLEGAL_INST
, ra
);
154 void helper_cbo_zero(CPURISCVState
*env
, target_ulong address
)
156 RISCVCPU
*cpu
= env_archcpu(env
);
157 uint16_t cbozlen
= cpu
->cfg
.cboz_blocksize
;
158 int mmu_idx
= riscv_env_mmu_index(env
, false);
159 uintptr_t ra
= GETPC();
162 check_zicbo_envcfg(env
, MENVCFG_CBZE
, ra
);
164 /* Mask off low-bits to align-down to the cache-block. */
165 address
&= ~(cbozlen
- 1);
168 * cbo.zero requires MMU_DATA_STORE access. Do a probe_write()
169 * to raise any exceptions, including PMP.
171 mem
= probe_write(env
, address
, cbozlen
, mmu_idx
, ra
);
174 memset(mem
, 0, cbozlen
);
177 * This means that we're dealing with an I/O page. Section 4.2
178 * of cmobase v1.0.1 says:
180 * "Cache-block zero instructions store zeros independently
181 * of whether data from the underlying memory locations are
184 * Write zeros in address + cbozlen regardless of not being
187 for (int i
= 0; i
< cbozlen
; i
++) {
188 cpu_stb_mmuidx_ra(env
, address
+ i
, 0, mmu_idx
, ra
);
194 * check_zicbom_access
196 * Check access permissions (LOAD, STORE or FETCH as specified in
197 * section 2.5.2 of the CMO specification) for Zicbom, raising
198 * either store page-fault (non-virtualized) or store guest-page
199 * fault (virtualized).
201 static void check_zicbom_access(CPURISCVState
*env
,
202 target_ulong address
,
205 RISCVCPU
*cpu
= env_archcpu(env
);
206 int mmu_idx
= riscv_env_mmu_index(env
, false);
207 uint16_t cbomlen
= cpu
->cfg
.cbom_blocksize
;
211 /* Mask off low-bits to align-down to the cache-block. */
212 address
&= ~(cbomlen
- 1);
215 * Section 2.5.2 of cmobase v1.0.1:
217 * "A cache-block management instruction is permitted to
218 * access the specified cache block whenever a load instruction
219 * or store instruction is permitted to access the corresponding
220 * physical addresses. If neither a load instruction nor store
221 * instruction is permitted to access the physical addresses,
222 * but an instruction fetch is permitted to access the physical
223 * addresses, whether a cache-block management instruction is
224 * permitted to access the cache block is UNSPECIFIED."
226 ret
= probe_access_flags(env
, address
, cbomlen
, MMU_DATA_LOAD
,
227 mmu_idx
, true, &phost
, ra
);
228 if (ret
!= TLB_INVALID_MASK
) {
229 /* Success: readable */
234 * Since not readable, must be writable. On failure, store
235 * fault/store guest amo fault will be raised by
236 * riscv_cpu_tlb_fill(). PMP exceptions will be caught
239 probe_write(env
, address
, cbomlen
, mmu_idx
, ra
);
242 void helper_cbo_clean_flush(CPURISCVState
*env
, target_ulong address
)
244 uintptr_t ra
= GETPC();
245 check_zicbo_envcfg(env
, MENVCFG_CBCFE
, ra
);
246 check_zicbom_access(env
, address
, ra
);
248 /* We don't emulate the cache-hierarchy, so we're done. */
251 void helper_cbo_inval(CPURISCVState
*env
, target_ulong address
)
253 uintptr_t ra
= GETPC();
254 check_zicbo_envcfg(env
, MENVCFG_CBIE
, ra
);
255 check_zicbom_access(env
, address
, ra
);
257 /* We don't emulate the cache-hierarchy, so we're done. */
260 #ifndef CONFIG_USER_ONLY
262 target_ulong
helper_sret(CPURISCVState
*env
)
265 target_ulong prev_priv
, prev_virt
= env
->virt_enabled
;
267 if (!(env
->priv
>= PRV_S
)) {
268 riscv_raise_exception(env
, RISCV_EXCP_ILLEGAL_INST
, GETPC());
271 target_ulong retpc
= env
->sepc
;
272 if (!riscv_has_ext(env
, RVC
) && (retpc
& 0x3)) {
273 riscv_raise_exception(env
, RISCV_EXCP_INST_ADDR_MIS
, GETPC());
276 if (get_field(env
->mstatus
, MSTATUS_TSR
) && !(env
->priv
>= PRV_M
)) {
277 riscv_raise_exception(env
, RISCV_EXCP_ILLEGAL_INST
, GETPC());
280 if (env
->virt_enabled
&& get_field(env
->hstatus
, HSTATUS_VTSR
)) {
281 riscv_raise_exception(env
, RISCV_EXCP_VIRT_INSTRUCTION_FAULT
, GETPC());
284 mstatus
= env
->mstatus
;
285 prev_priv
= get_field(mstatus
, MSTATUS_SPP
);
286 mstatus
= set_field(mstatus
, MSTATUS_SIE
,
287 get_field(mstatus
, MSTATUS_SPIE
));
288 mstatus
= set_field(mstatus
, MSTATUS_SPIE
, 1);
289 mstatus
= set_field(mstatus
, MSTATUS_SPP
, PRV_U
);
290 if (env
->priv_ver
>= PRIV_VERSION_1_12_0
) {
291 mstatus
= set_field(mstatus
, MSTATUS_MPRV
, 0);
293 env
->mstatus
= mstatus
;
295 if (riscv_has_ext(env
, RVH
) && !env
->virt_enabled
) {
296 /* We support Hypervisor extensions and virtulisation is disabled */
297 target_ulong hstatus
= env
->hstatus
;
299 prev_virt
= get_field(hstatus
, HSTATUS_SPV
);
301 hstatus
= set_field(hstatus
, HSTATUS_SPV
, 0);
303 env
->hstatus
= hstatus
;
306 riscv_cpu_swap_hypervisor_regs(env
);
310 riscv_cpu_set_mode(env
, prev_priv
, prev_virt
);
315 target_ulong
helper_mret(CPURISCVState
*env
)
317 if (!(env
->priv
>= PRV_M
)) {
318 riscv_raise_exception(env
, RISCV_EXCP_ILLEGAL_INST
, GETPC());
321 target_ulong retpc
= env
->mepc
;
322 if (!riscv_has_ext(env
, RVC
) && (retpc
& 0x3)) {
323 riscv_raise_exception(env
, RISCV_EXCP_INST_ADDR_MIS
, GETPC());
326 uint64_t mstatus
= env
->mstatus
;
327 target_ulong prev_priv
= get_field(mstatus
, MSTATUS_MPP
);
329 if (riscv_cpu_cfg(env
)->pmp
&&
330 !pmp_get_num_rules(env
) && (prev_priv
!= PRV_M
)) {
331 riscv_raise_exception(env
, RISCV_EXCP_INST_ACCESS_FAULT
, GETPC());
334 target_ulong prev_virt
= get_field(env
->mstatus
, MSTATUS_MPV
) &&
335 (prev_priv
!= PRV_M
);
336 mstatus
= set_field(mstatus
, MSTATUS_MIE
,
337 get_field(mstatus
, MSTATUS_MPIE
));
338 mstatus
= set_field(mstatus
, MSTATUS_MPIE
, 1);
339 mstatus
= set_field(mstatus
, MSTATUS_MPP
,
340 riscv_has_ext(env
, RVU
) ? PRV_U
: PRV_M
);
341 mstatus
= set_field(mstatus
, MSTATUS_MPV
, 0);
342 if ((env
->priv_ver
>= PRIV_VERSION_1_12_0
) && (prev_priv
!= PRV_M
)) {
343 mstatus
= set_field(mstatus
, MSTATUS_MPRV
, 0);
345 env
->mstatus
= mstatus
;
347 if (riscv_has_ext(env
, RVH
) && prev_virt
) {
348 riscv_cpu_swap_hypervisor_regs(env
);
351 riscv_cpu_set_mode(env
, prev_priv
, prev_virt
);
356 void helper_wfi(CPURISCVState
*env
)
358 CPUState
*cs
= env_cpu(env
);
359 bool rvs
= riscv_has_ext(env
, RVS
);
360 bool prv_u
= env
->priv
== PRV_U
;
361 bool prv_s
= env
->priv
== PRV_S
;
363 if (((prv_s
|| (!rvs
&& prv_u
)) && get_field(env
->mstatus
, MSTATUS_TW
)) ||
364 (rvs
&& prv_u
&& !env
->virt_enabled
)) {
365 riscv_raise_exception(env
, RISCV_EXCP_ILLEGAL_INST
, GETPC());
366 } else if (env
->virt_enabled
&&
367 (prv_u
|| (prv_s
&& get_field(env
->hstatus
, HSTATUS_VTW
)))) {
368 riscv_raise_exception(env
, RISCV_EXCP_VIRT_INSTRUCTION_FAULT
, GETPC());
371 cs
->exception_index
= EXCP_HLT
;
376 void helper_wrs_nto(CPURISCVState
*env
)
378 if (env
->virt_enabled
&& (env
->priv
== PRV_S
|| env
->priv
== PRV_U
) &&
379 get_field(env
->hstatus
, HSTATUS_VTW
) &&
380 !get_field(env
->mstatus
, MSTATUS_TW
)) {
381 riscv_raise_exception(env
, RISCV_EXCP_VIRT_INSTRUCTION_FAULT
, GETPC());
382 } else if (env
->priv
!= PRV_M
&& get_field(env
->mstatus
, MSTATUS_TW
)) {
383 riscv_raise_exception(env
, RISCV_EXCP_ILLEGAL_INST
, GETPC());
387 void helper_tlb_flush(CPURISCVState
*env
)
389 CPUState
*cs
= env_cpu(env
);
390 if (!env
->virt_enabled
&&
391 (env
->priv
== PRV_U
||
392 (env
->priv
== PRV_S
&& get_field(env
->mstatus
, MSTATUS_TVM
)))) {
393 riscv_raise_exception(env
, RISCV_EXCP_ILLEGAL_INST
, GETPC());
394 } else if (env
->virt_enabled
&&
395 (env
->priv
== PRV_U
|| get_field(env
->hstatus
, HSTATUS_VTVM
))) {
396 riscv_raise_exception(env
, RISCV_EXCP_VIRT_INSTRUCTION_FAULT
, GETPC());
402 void helper_tlb_flush_all(CPURISCVState
*env
)
404 CPUState
*cs
= env_cpu(env
);
405 tlb_flush_all_cpus_synced(cs
);
408 void helper_hyp_tlb_flush(CPURISCVState
*env
)
410 CPUState
*cs
= env_cpu(env
);
412 if (env
->virt_enabled
) {
413 riscv_raise_exception(env
, RISCV_EXCP_VIRT_INSTRUCTION_FAULT
, GETPC());
416 if (env
->priv
== PRV_M
||
417 (env
->priv
== PRV_S
&& !env
->virt_enabled
)) {
422 riscv_raise_exception(env
, RISCV_EXCP_ILLEGAL_INST
, GETPC());
425 void helper_hyp_gvma_tlb_flush(CPURISCVState
*env
)
427 if (env
->priv
== PRV_S
&& !env
->virt_enabled
&&
428 get_field(env
->mstatus
, MSTATUS_TVM
)) {
429 riscv_raise_exception(env
, RISCV_EXCP_ILLEGAL_INST
, GETPC());
432 helper_hyp_tlb_flush(env
);
435 static int check_access_hlsv(CPURISCVState
*env
, bool x
, uintptr_t ra
)
437 if (env
->priv
== PRV_M
) {
439 } else if (env
->virt_enabled
) {
440 riscv_raise_exception(env
, RISCV_EXCP_VIRT_INSTRUCTION_FAULT
, ra
);
441 } else if (env
->priv
== PRV_U
&& !get_field(env
->hstatus
, HSTATUS_HU
)) {
442 riscv_raise_exception(env
, RISCV_EXCP_ILLEGAL_INST
, ra
);
445 int mode
= get_field(env
->hstatus
, HSTATUS_SPVP
);
446 if (!x
&& mode
== PRV_S
&& get_field(env
->vsstatus
, MSTATUS_SUM
)) {
449 return mode
| MMU_2STAGE_BIT
;
452 target_ulong
helper_hyp_hlv_bu(CPURISCVState
*env
, target_ulong addr
)
454 uintptr_t ra
= GETPC();
455 int mmu_idx
= check_access_hlsv(env
, false, ra
);
456 MemOpIdx oi
= make_memop_idx(MO_UB
, mmu_idx
);
458 return cpu_ldb_mmu(env
, addr
, oi
, ra
);
461 target_ulong
helper_hyp_hlv_hu(CPURISCVState
*env
, target_ulong addr
)
463 uintptr_t ra
= GETPC();
464 int mmu_idx
= check_access_hlsv(env
, false, ra
);
465 MemOpIdx oi
= make_memop_idx(MO_TEUW
, mmu_idx
);
467 return cpu_ldw_mmu(env
, addr
, oi
, ra
);
470 target_ulong
helper_hyp_hlv_wu(CPURISCVState
*env
, target_ulong addr
)
472 uintptr_t ra
= GETPC();
473 int mmu_idx
= check_access_hlsv(env
, false, ra
);
474 MemOpIdx oi
= make_memop_idx(MO_TEUL
, mmu_idx
);
476 return cpu_ldl_mmu(env
, addr
, oi
, ra
);
479 target_ulong
helper_hyp_hlv_d(CPURISCVState
*env
, target_ulong addr
)
481 uintptr_t ra
= GETPC();
482 int mmu_idx
= check_access_hlsv(env
, false, ra
);
483 MemOpIdx oi
= make_memop_idx(MO_TEUQ
, mmu_idx
);
485 return cpu_ldq_mmu(env
, addr
, oi
, ra
);
488 void helper_hyp_hsv_b(CPURISCVState
*env
, target_ulong addr
, target_ulong val
)
490 uintptr_t ra
= GETPC();
491 int mmu_idx
= check_access_hlsv(env
, false, ra
);
492 MemOpIdx oi
= make_memop_idx(MO_UB
, mmu_idx
);
494 cpu_stb_mmu(env
, addr
, val
, oi
, ra
);
497 void helper_hyp_hsv_h(CPURISCVState
*env
, target_ulong addr
, target_ulong val
)
499 uintptr_t ra
= GETPC();
500 int mmu_idx
= check_access_hlsv(env
, false, ra
);
501 MemOpIdx oi
= make_memop_idx(MO_TEUW
, mmu_idx
);
503 cpu_stw_mmu(env
, addr
, val
, oi
, ra
);
506 void helper_hyp_hsv_w(CPURISCVState
*env
, target_ulong addr
, target_ulong val
)
508 uintptr_t ra
= GETPC();
509 int mmu_idx
= check_access_hlsv(env
, false, ra
);
510 MemOpIdx oi
= make_memop_idx(MO_TEUL
, mmu_idx
);
512 cpu_stl_mmu(env
, addr
, val
, oi
, ra
);
515 void helper_hyp_hsv_d(CPURISCVState
*env
, target_ulong addr
, target_ulong val
)
517 uintptr_t ra
= GETPC();
518 int mmu_idx
= check_access_hlsv(env
, false, ra
);
519 MemOpIdx oi
= make_memop_idx(MO_TEUQ
, mmu_idx
);
521 cpu_stq_mmu(env
, addr
, val
, oi
, ra
);
525 * TODO: These implementations are not quite correct. They perform the
526 * access using execute permission just fine, but the final PMP check
527 * is supposed to have read permission as well. Without replicating
528 * a fair fraction of cputlb.c, fixing this requires adding new mmu_idx
529 * which would imply that exact check in tlb_fill.
531 target_ulong
helper_hyp_hlvx_hu(CPURISCVState
*env
, target_ulong addr
)
533 uintptr_t ra
= GETPC();
534 int mmu_idx
= check_access_hlsv(env
, true, ra
);
535 MemOpIdx oi
= make_memop_idx(MO_TEUW
, mmu_idx
);
537 return cpu_ldw_code_mmu(env
, addr
, oi
, GETPC());
540 target_ulong
helper_hyp_hlvx_wu(CPURISCVState
*env
, target_ulong addr
)
542 uintptr_t ra
= GETPC();
543 int mmu_idx
= check_access_hlsv(env
, true, ra
);
544 MemOpIdx oi
= make_memop_idx(MO_TEUL
, mmu_idx
);
546 return cpu_ldl_code_mmu(env
, addr
, oi
, ra
);
549 #endif /* !CONFIG_USER_ONLY */