2 * Copyright (c) 2011, Max Filippov, Open Source and Linux Lab.
5 * Redistribution and use in source and binary forms, with or without
6 * modification, are permitted provided that the following conditions are met:
7 * * Redistributions of source code must retain the above copyright
8 * notice, this list of conditions and the following disclaimer.
9 * * Redistributions in binary form must reproduce the above copyright
10 * notice, this list of conditions and the following disclaimer in the
11 * documentation and/or other materials provided with the distribution.
12 * * Neither the name of the Open Source and Linux Lab nor the
13 * names of its contributors may be used to endorse or promote products
14 * derived from this software without specific prior written permission.
16 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
17 * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
18 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
19 * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY
20 * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
21 * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
22 * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
23 * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
24 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
25 * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
28 #include "qemu/osdep.h"
29 #include "qemu/units.h"
31 #include "exec/exec-all.h"
32 #include "exec/gdbstub.h"
33 #include "qemu/host-utils.h"
34 #if !defined(CONFIG_USER_ONLY)
35 #include "hw/loader.h"
38 static struct XtensaConfigList
*xtensa_cores
;
40 static void xtensa_core_class_init(ObjectClass
*oc
, void *data
)
42 CPUClass
*cc
= CPU_CLASS(oc
);
43 XtensaCPUClass
*xcc
= XTENSA_CPU_CLASS(oc
);
44 const XtensaConfig
*config
= data
;
48 /* Use num_core_regs to see only non-privileged registers in an unmodified
49 * gdb. Use num_regs to see all registers. gdb modification is required
50 * for that: reset bit 0 in the 'flags' field of the registers definitions
51 * in the gdb/xtensa-config.c inside gdb source tree or inside gdb overlay.
53 cc
->gdb_num_core_regs
= config
->gdb_regmap
.num_regs
;
56 static void init_libisa(XtensaConfig
*config
)
62 config
->isa
= xtensa_isa_init(config
->isa_internal
, NULL
, NULL
);
63 assert(xtensa_isa_maxlength(config
->isa
) <= MAX_INSN_LENGTH
);
64 opcodes
= xtensa_isa_num_opcodes(config
->isa
);
65 formats
= xtensa_isa_num_formats(config
->isa
);
66 config
->opcode_ops
= g_new(XtensaOpcodeOps
*, opcodes
);
68 for (i
= 0; i
< formats
; ++i
) {
69 assert(xtensa_format_num_slots(config
->isa
, i
) <= MAX_INSN_SLOTS
);
72 for (i
= 0; i
< opcodes
; ++i
) {
73 const char *opc_name
= xtensa_opcode_name(config
->isa
, i
);
74 XtensaOpcodeOps
*ops
= NULL
;
76 assert(xtensa_opcode_num_operands(config
->isa
, i
) <= MAX_OPCODE_ARGS
);
77 if (!config
->opcode_translators
) {
78 ops
= xtensa_find_opcode_ops(&xtensa_core_opcodes
, opc_name
);
80 for (j
= 0; !ops
&& config
->opcode_translators
[j
]; ++j
) {
81 ops
= xtensa_find_opcode_ops(config
->opcode_translators
[j
],
88 "opcode translator not found for %s's opcode '%s'\n",
89 config
->name
, opc_name
);
92 config
->opcode_ops
[i
] = ops
;
96 void xtensa_finalize_config(XtensaConfig
*config
)
98 if (config
->isa_internal
) {
102 if (config
->gdb_regmap
.num_regs
== 0 ||
103 config
->gdb_regmap
.num_core_regs
== 0) {
105 unsigned n_core_regs
= 0;
107 xtensa_count_regs(config
, &n_regs
, &n_core_regs
);
108 if (config
->gdb_regmap
.num_regs
== 0) {
109 config
->gdb_regmap
.num_regs
= n_regs
;
111 if (config
->gdb_regmap
.num_core_regs
== 0) {
112 config
->gdb_regmap
.num_core_regs
= n_core_regs
;
117 void xtensa_register_core(XtensaConfigList
*node
)
120 .parent
= TYPE_XTENSA_CPU
,
121 .class_init
= xtensa_core_class_init
,
122 .class_data
= (void *)node
->config
,
125 node
->next
= xtensa_cores
;
127 type
.name
= g_strdup_printf(XTENSA_CPU_TYPE_NAME("%s"), node
->config
->name
);
128 type_register(&type
);
129 g_free((gpointer
)type
.name
);
132 static uint32_t check_hw_breakpoints(CPUXtensaState
*env
)
136 for (i
= 0; i
< env
->config
->ndbreak
; ++i
) {
137 if (env
->cpu_watchpoint
[i
] &&
138 env
->cpu_watchpoint
[i
]->flags
& BP_WATCHPOINT_HIT
) {
139 return DEBUGCAUSE_DB
| (i
<< DEBUGCAUSE_DBNUM_SHIFT
);
145 void xtensa_breakpoint_handler(CPUState
*cs
)
147 XtensaCPU
*cpu
= XTENSA_CPU(cs
);
148 CPUXtensaState
*env
= &cpu
->env
;
150 if (cs
->watchpoint_hit
) {
151 if (cs
->watchpoint_hit
->flags
& BP_CPU
) {
154 cs
->watchpoint_hit
= NULL
;
155 cause
= check_hw_breakpoints(env
);
157 debug_exception_env(env
, cause
);
159 cpu_loop_exit_noexc(cs
);
164 void xtensa_cpu_list(FILE *f
, fprintf_function cpu_fprintf
)
166 XtensaConfigList
*core
= xtensa_cores
;
167 cpu_fprintf(f
, "Available CPUs:\n");
168 for (; core
; core
= core
->next
) {
169 cpu_fprintf(f
, " %s\n", core
->config
->name
);
173 hwaddr
xtensa_cpu_get_phys_page_debug(CPUState
*cs
, vaddr addr
)
175 #ifndef CONFIG_USER_ONLY
176 XtensaCPU
*cpu
= XTENSA_CPU(cs
);
181 if (xtensa_get_physical_addr(&cpu
->env
, false, addr
, 0, 0,
182 &paddr
, &page_size
, &access
) == 0) {
185 if (xtensa_get_physical_addr(&cpu
->env
, false, addr
, 2, 0,
186 &paddr
, &page_size
, &access
) == 0) {
195 #ifndef CONFIG_USER_ONLY
197 static uint32_t relocated_vector(CPUXtensaState
*env
, uint32_t vector
)
199 if (xtensa_option_enabled(env
->config
,
200 XTENSA_OPTION_RELOCATABLE_VECTOR
)) {
201 return vector
- env
->config
->vecbase
+ env
->sregs
[VECBASE
];
208 * Handle penging IRQ.
209 * For the high priority interrupt jump to the corresponding interrupt vector.
210 * For the level-1 interrupt convert it to either user, kernel or double
211 * exception with the 'level-1 interrupt' exception cause.
213 static void handle_interrupt(CPUXtensaState
*env
)
215 int level
= env
->pending_irq_level
;
217 if (level
> xtensa_get_cintlevel(env
) &&
218 level
<= env
->config
->nlevel
&&
219 (env
->config
->level_mask
[level
] &
221 env
->sregs
[INTENABLE
])) {
222 CPUState
*cs
= CPU(xtensa_env_get_cpu(env
));
225 env
->sregs
[EPC1
+ level
- 1] = env
->pc
;
226 env
->sregs
[EPS2
+ level
- 2] = env
->sregs
[PS
];
228 (env
->sregs
[PS
] & ~PS_INTLEVEL
) | level
| PS_EXCM
;
229 env
->pc
= relocated_vector(env
,
230 env
->config
->interrupt_vector
[level
]);
232 env
->sregs
[EXCCAUSE
] = LEVEL1_INTERRUPT_CAUSE
;
234 if (env
->sregs
[PS
] & PS_EXCM
) {
235 if (env
->config
->ndepc
) {
236 env
->sregs
[DEPC
] = env
->pc
;
238 env
->sregs
[EPC1
] = env
->pc
;
240 cs
->exception_index
= EXC_DOUBLE
;
242 env
->sregs
[EPC1
] = env
->pc
;
243 cs
->exception_index
=
244 (env
->sregs
[PS
] & PS_UM
) ? EXC_USER
: EXC_KERNEL
;
246 env
->sregs
[PS
] |= PS_EXCM
;
248 env
->exception_taken
= 1;
252 /* Called from cpu_handle_interrupt with BQL held */
253 void xtensa_cpu_do_interrupt(CPUState
*cs
)
255 XtensaCPU
*cpu
= XTENSA_CPU(cs
);
256 CPUXtensaState
*env
= &cpu
->env
;
258 if (cs
->exception_index
== EXC_IRQ
) {
259 qemu_log_mask(CPU_LOG_INT
,
260 "%s(EXC_IRQ) level = %d, cintlevel = %d, "
261 "pc = %08x, a0 = %08x, ps = %08x, "
262 "intset = %08x, intenable = %08x, "
264 __func__
, env
->pending_irq_level
, xtensa_get_cintlevel(env
),
265 env
->pc
, env
->regs
[0], env
->sregs
[PS
],
266 env
->sregs
[INTSET
], env
->sregs
[INTENABLE
],
268 handle_interrupt(env
);
271 switch (cs
->exception_index
) {
272 case EXC_WINDOW_OVERFLOW4
:
273 case EXC_WINDOW_UNDERFLOW4
:
274 case EXC_WINDOW_OVERFLOW8
:
275 case EXC_WINDOW_UNDERFLOW8
:
276 case EXC_WINDOW_OVERFLOW12
:
277 case EXC_WINDOW_UNDERFLOW12
:
282 qemu_log_mask(CPU_LOG_INT
, "%s(%d) "
283 "pc = %08x, a0 = %08x, ps = %08x, ccount = %08x\n",
284 __func__
, cs
->exception_index
,
285 env
->pc
, env
->regs
[0], env
->sregs
[PS
], env
->sregs
[CCOUNT
]);
286 if (env
->config
->exception_vector
[cs
->exception_index
]) {
287 env
->pc
= relocated_vector(env
,
288 env
->config
->exception_vector
[cs
->exception_index
]);
289 env
->exception_taken
= 1;
291 qemu_log_mask(CPU_LOG_INT
, "%s(pc = %08x) bad exception_index: %d\n",
292 __func__
, env
->pc
, cs
->exception_index
);
300 qemu_log("%s(pc = %08x) unknown exception_index: %d\n",
301 __func__
, env
->pc
, cs
->exception_index
);
304 check_interrupts(env
);
307 void xtensa_cpu_do_interrupt(CPUState
*cs
)
312 bool xtensa_cpu_exec_interrupt(CPUState
*cs
, int interrupt_request
)
314 if (interrupt_request
& CPU_INTERRUPT_HARD
) {
315 cs
->exception_index
= EXC_IRQ
;
316 xtensa_cpu_do_interrupt(cs
);
322 #ifdef CONFIG_USER_ONLY
324 int xtensa_cpu_handle_mmu_fault(CPUState
*cs
, vaddr address
, int size
, int rw
,
327 XtensaCPU
*cpu
= XTENSA_CPU(cs
);
328 CPUXtensaState
*env
= &cpu
->env
;
330 qemu_log_mask(CPU_LOG_INT
,
331 "%s: rw = %d, address = 0x%08" VADDR_PRIx
", size = %d\n",
332 __func__
, rw
, address
, size
);
333 env
->sregs
[EXCVADDR
] = address
;
334 env
->sregs
[EXCCAUSE
] = rw
? STORE_PROHIBITED_CAUSE
: LOAD_PROHIBITED_CAUSE
;
335 cs
->exception_index
= EXC_USER
;
341 static void reset_tlb_mmu_all_ways(CPUXtensaState
*env
,
342 const xtensa_tlb
*tlb
, xtensa_tlb_entry entry
[][MAX_TLB_WAY_SIZE
])
346 for (wi
= 0; wi
< tlb
->nways
; ++wi
) {
347 for (ei
= 0; ei
< tlb
->way_size
[wi
]; ++ei
) {
348 entry
[wi
][ei
].asid
= 0;
349 entry
[wi
][ei
].variable
= true;
354 static void reset_tlb_mmu_ways56(CPUXtensaState
*env
,
355 const xtensa_tlb
*tlb
, xtensa_tlb_entry entry
[][MAX_TLB_WAY_SIZE
])
357 if (!tlb
->varway56
) {
358 static const xtensa_tlb_entry way5
[] = {
373 static const xtensa_tlb_entry way6
[] = {
388 memcpy(entry
[5], way5
, sizeof(way5
));
389 memcpy(entry
[6], way6
, sizeof(way6
));
392 for (ei
= 0; ei
< 8; ++ei
) {
393 entry
[6][ei
].vaddr
= ei
<< 29;
394 entry
[6][ei
].paddr
= ei
<< 29;
395 entry
[6][ei
].asid
= 1;
396 entry
[6][ei
].attr
= 3;
401 static void reset_tlb_region_way0(CPUXtensaState
*env
,
402 xtensa_tlb_entry entry
[][MAX_TLB_WAY_SIZE
])
406 for (ei
= 0; ei
< 8; ++ei
) {
407 entry
[0][ei
].vaddr
= ei
<< 29;
408 entry
[0][ei
].paddr
= ei
<< 29;
409 entry
[0][ei
].asid
= 1;
410 entry
[0][ei
].attr
= 2;
411 entry
[0][ei
].variable
= true;
415 void reset_mmu(CPUXtensaState
*env
)
417 if (xtensa_option_enabled(env
->config
, XTENSA_OPTION_MMU
)) {
418 env
->sregs
[RASID
] = 0x04030201;
419 env
->sregs
[ITLBCFG
] = 0;
420 env
->sregs
[DTLBCFG
] = 0;
421 env
->autorefill_idx
= 0;
422 reset_tlb_mmu_all_ways(env
, &env
->config
->itlb
, env
->itlb
);
423 reset_tlb_mmu_all_ways(env
, &env
->config
->dtlb
, env
->dtlb
);
424 reset_tlb_mmu_ways56(env
, &env
->config
->itlb
, env
->itlb
);
425 reset_tlb_mmu_ways56(env
, &env
->config
->dtlb
, env
->dtlb
);
427 reset_tlb_region_way0(env
, env
->itlb
);
428 reset_tlb_region_way0(env
, env
->dtlb
);
432 static unsigned get_ring(const CPUXtensaState
*env
, uint8_t asid
)
435 for (i
= 0; i
< 4; ++i
) {
436 if (((env
->sregs
[RASID
] >> i
* 8) & 0xff) == asid
) {
444 * Lookup xtensa TLB for the given virtual address.
447 * \param pwi: [out] way index
448 * \param pei: [out] entry index
449 * \param pring: [out] access ring
450 * \return 0 if ok, exception cause code otherwise
452 int xtensa_tlb_lookup(const CPUXtensaState
*env
, uint32_t addr
, bool dtlb
,
453 uint32_t *pwi
, uint32_t *pei
, uint8_t *pring
)
455 const xtensa_tlb
*tlb
= dtlb
?
456 &env
->config
->dtlb
: &env
->config
->itlb
;
457 const xtensa_tlb_entry (*entry
)[MAX_TLB_WAY_SIZE
] = dtlb
?
458 env
->dtlb
: env
->itlb
;
463 for (wi
= 0; wi
< tlb
->nways
; ++wi
) {
466 split_tlb_entry_spec_way(env
, addr
, dtlb
, &vpn
, wi
, &ei
);
467 if (entry
[wi
][ei
].vaddr
== vpn
&& entry
[wi
][ei
].asid
) {
468 unsigned ring
= get_ring(env
, entry
[wi
][ei
].asid
);
472 LOAD_STORE_TLB_MULTI_HIT_CAUSE
:
473 INST_TLB_MULTI_HIT_CAUSE
;
482 (dtlb
? LOAD_STORE_TLB_MISS_CAUSE
: INST_TLB_MISS_CAUSE
);
486 * Convert MMU ATTR to PAGE_{READ,WRITE,EXEC} mask.
489 static unsigned mmu_attr_to_access(uint32_t attr
)
499 access
|= PAGE_WRITE
;
502 switch (attr
& 0xc) {
504 access
|= PAGE_CACHE_BYPASS
;
508 access
|= PAGE_CACHE_WB
;
512 access
|= PAGE_CACHE_WT
;
515 } else if (attr
== 13) {
516 access
|= PAGE_READ
| PAGE_WRITE
| PAGE_CACHE_ISOLATE
;
522 * Convert region protection ATTR to PAGE_{READ,WRITE,EXEC} mask.
525 static unsigned region_attr_to_access(uint32_t attr
)
527 static const unsigned access
[16] = {
528 [0] = PAGE_READ
| PAGE_WRITE
| PAGE_CACHE_WT
,
529 [1] = PAGE_READ
| PAGE_WRITE
| PAGE_EXEC
| PAGE_CACHE_WT
,
530 [2] = PAGE_READ
| PAGE_WRITE
| PAGE_EXEC
| PAGE_CACHE_BYPASS
,
531 [3] = PAGE_EXEC
| PAGE_CACHE_WB
,
532 [4] = PAGE_READ
| PAGE_WRITE
| PAGE_EXEC
| PAGE_CACHE_WB
,
533 [5] = PAGE_READ
| PAGE_WRITE
| PAGE_EXEC
| PAGE_CACHE_WB
,
534 [14] = PAGE_READ
| PAGE_WRITE
| PAGE_CACHE_ISOLATE
,
537 return access
[attr
& 0xf];
541 * Convert cacheattr to PAGE_{READ,WRITE,EXEC} mask.
542 * See ISA, A.2.14 The Cache Attribute Register
544 static unsigned cacheattr_attr_to_access(uint32_t attr
)
546 static const unsigned access
[16] = {
547 [0] = PAGE_READ
| PAGE_WRITE
| PAGE_CACHE_WT
,
548 [1] = PAGE_READ
| PAGE_WRITE
| PAGE_EXEC
| PAGE_CACHE_WT
,
549 [2] = PAGE_READ
| PAGE_WRITE
| PAGE_EXEC
| PAGE_CACHE_BYPASS
,
550 [3] = PAGE_EXEC
| PAGE_CACHE_WB
,
551 [4] = PAGE_READ
| PAGE_WRITE
| PAGE_EXEC
| PAGE_CACHE_WB
,
552 [14] = PAGE_READ
| PAGE_WRITE
| PAGE_CACHE_ISOLATE
,
555 return access
[attr
& 0xf];
558 static bool is_access_granted(unsigned access
, int is_write
)
562 return access
& PAGE_READ
;
565 return access
& PAGE_WRITE
;
568 return access
& PAGE_EXEC
;
575 static bool get_pte(CPUXtensaState
*env
, uint32_t vaddr
, uint32_t *pte
);
577 static int get_physical_addr_mmu(CPUXtensaState
*env
, bool update_tlb
,
578 uint32_t vaddr
, int is_write
, int mmu_idx
,
579 uint32_t *paddr
, uint32_t *page_size
, unsigned *access
,
582 bool dtlb
= is_write
!= 2;
588 const xtensa_tlb_entry
*entry
= NULL
;
589 xtensa_tlb_entry tmp_entry
;
590 int ret
= xtensa_tlb_lookup(env
, vaddr
, dtlb
, &wi
, &ei
, &ring
);
592 if ((ret
== INST_TLB_MISS_CAUSE
|| ret
== LOAD_STORE_TLB_MISS_CAUSE
) &&
593 may_lookup_pt
&& get_pte(env
, vaddr
, &pte
)) {
594 ring
= (pte
>> 4) & 0x3;
596 split_tlb_entry_spec_way(env
, vaddr
, dtlb
, &vpn
, wi
, &ei
);
599 wi
= ++env
->autorefill_idx
& 0x3;
600 xtensa_tlb_set_entry(env
, dtlb
, wi
, ei
, vpn
, pte
);
601 env
->sregs
[EXCVADDR
] = vaddr
;
602 qemu_log_mask(CPU_LOG_MMU
, "%s: autorefill(%08x): %08x -> %08x\n",
603 __func__
, vaddr
, vpn
, pte
);
605 xtensa_tlb_set_entry_mmu(env
, &tmp_entry
, dtlb
, wi
, ei
, vpn
, pte
);
615 entry
= xtensa_tlb_get_entry(env
, dtlb
, wi
, ei
);
618 if (ring
< mmu_idx
) {
620 LOAD_STORE_PRIVILEGE_CAUSE
:
621 INST_FETCH_PRIVILEGE_CAUSE
;
624 *access
= mmu_attr_to_access(entry
->attr
) &
625 ~(dtlb
? PAGE_EXEC
: PAGE_READ
| PAGE_WRITE
);
626 if (!is_access_granted(*access
, is_write
)) {
629 STORE_PROHIBITED_CAUSE
:
630 LOAD_PROHIBITED_CAUSE
) :
631 INST_FETCH_PROHIBITED_CAUSE
;
634 *paddr
= entry
->paddr
| (vaddr
& ~xtensa_tlb_get_addr_mask(env
, dtlb
, wi
));
635 *page_size
= ~xtensa_tlb_get_addr_mask(env
, dtlb
, wi
) + 1;
640 static bool get_pte(CPUXtensaState
*env
, uint32_t vaddr
, uint32_t *pte
)
642 CPUState
*cs
= CPU(xtensa_env_get_cpu(env
));
647 (env
->sregs
[PTEVADDR
] | (vaddr
>> 10)) & 0xfffffffc;
648 int ret
= get_physical_addr_mmu(env
, false, pt_vaddr
, 0, 0,
649 &paddr
, &page_size
, &access
, false);
652 qemu_log_mask(CPU_LOG_MMU
,
653 "%s: autorefill(%08x): PTE va = %08x, pa = %08x\n",
654 __func__
, vaddr
, pt_vaddr
, paddr
);
656 qemu_log_mask(CPU_LOG_MMU
,
657 "%s: autorefill(%08x): PTE va = %08x, failed (%d)\n",
658 __func__
, vaddr
, pt_vaddr
, ret
);
664 *pte
= address_space_ldl(cs
->as
, paddr
, MEMTXATTRS_UNSPECIFIED
,
666 if (result
!= MEMTX_OK
) {
667 qemu_log_mask(CPU_LOG_MMU
,
668 "%s: couldn't load PTE: transaction failed (%u)\n",
669 __func__
, (unsigned)result
);
676 static int get_physical_addr_region(CPUXtensaState
*env
,
677 uint32_t vaddr
, int is_write
, int mmu_idx
,
678 uint32_t *paddr
, uint32_t *page_size
, unsigned *access
)
680 bool dtlb
= is_write
!= 2;
682 uint32_t ei
= (vaddr
>> 29) & 0x7;
683 const xtensa_tlb_entry
*entry
=
684 xtensa_tlb_get_entry(env
, dtlb
, wi
, ei
);
686 *access
= region_attr_to_access(entry
->attr
);
687 if (!is_access_granted(*access
, is_write
)) {
690 STORE_PROHIBITED_CAUSE
:
691 LOAD_PROHIBITED_CAUSE
) :
692 INST_FETCH_PROHIBITED_CAUSE
;
695 *paddr
= entry
->paddr
| (vaddr
& ~REGION_PAGE_MASK
);
696 *page_size
= ~REGION_PAGE_MASK
+ 1;
702 * Convert virtual address to physical addr.
703 * MMU may issue pagewalk and change xtensa autorefill TLB way entry.
705 * \return 0 if ok, exception cause code otherwise
707 int xtensa_get_physical_addr(CPUXtensaState
*env
, bool update_tlb
,
708 uint32_t vaddr
, int is_write
, int mmu_idx
,
709 uint32_t *paddr
, uint32_t *page_size
, unsigned *access
)
711 if (xtensa_option_enabled(env
->config
, XTENSA_OPTION_MMU
)) {
712 return get_physical_addr_mmu(env
, update_tlb
,
713 vaddr
, is_write
, mmu_idx
, paddr
, page_size
, access
, true);
714 } else if (xtensa_option_bits_enabled(env
->config
,
715 XTENSA_OPTION_BIT(XTENSA_OPTION_REGION_PROTECTION
) |
716 XTENSA_OPTION_BIT(XTENSA_OPTION_REGION_TRANSLATION
))) {
717 return get_physical_addr_region(env
, vaddr
, is_write
, mmu_idx
,
718 paddr
, page_size
, access
);
721 *page_size
= TARGET_PAGE_SIZE
;
722 *access
= cacheattr_attr_to_access(
723 env
->sregs
[CACHEATTR
] >> ((vaddr
& 0xe0000000) >> 27));
728 static void dump_tlb(FILE *f
, fprintf_function cpu_fprintf
,
729 CPUXtensaState
*env
, bool dtlb
)
732 const xtensa_tlb
*conf
=
733 dtlb
? &env
->config
->dtlb
: &env
->config
->itlb
;
734 unsigned (*attr_to_access
)(uint32_t) =
735 xtensa_option_enabled(env
->config
, XTENSA_OPTION_MMU
) ?
736 mmu_attr_to_access
: region_attr_to_access
;
738 for (wi
= 0; wi
< conf
->nways
; ++wi
) {
739 uint32_t sz
= ~xtensa_tlb_get_addr_mask(env
, dtlb
, wi
) + 1;
741 bool print_header
= true;
743 if (sz
>= 0x100000) {
751 for (ei
= 0; ei
< conf
->way_size
[wi
]; ++ei
) {
752 const xtensa_tlb_entry
*entry
=
753 xtensa_tlb_get_entry(env
, dtlb
, wi
, ei
);
756 static const char * const cache_text
[8] = {
757 [PAGE_CACHE_BYPASS
>> PAGE_CACHE_SHIFT
] = "Bypass",
758 [PAGE_CACHE_WT
>> PAGE_CACHE_SHIFT
] = "WT",
759 [PAGE_CACHE_WB
>> PAGE_CACHE_SHIFT
] = "WB",
760 [PAGE_CACHE_ISOLATE
>> PAGE_CACHE_SHIFT
] = "Isolate",
762 unsigned access
= attr_to_access(entry
->attr
);
763 unsigned cache_idx
= (access
& PAGE_CACHE_MASK
) >>
767 print_header
= false;
768 cpu_fprintf(f
, "Way %u (%d %s)\n", wi
, sz
, sz_text
);
770 "\tVaddr Paddr ASID Attr RWX Cache\n"
771 "\t---------- ---------- ---- ---- --- -------\n");
774 "\t0x%08x 0x%08x 0x%02x 0x%02x %c%c%c %-7s\n",
779 (access
& PAGE_READ
) ? 'R' : '-',
780 (access
& PAGE_WRITE
) ? 'W' : '-',
781 (access
& PAGE_EXEC
) ? 'X' : '-',
782 cache_text
[cache_idx
] ? cache_text
[cache_idx
] :
789 void dump_mmu(FILE *f
, fprintf_function cpu_fprintf
, CPUXtensaState
*env
)
791 if (xtensa_option_bits_enabled(env
->config
,
792 XTENSA_OPTION_BIT(XTENSA_OPTION_REGION_PROTECTION
) |
793 XTENSA_OPTION_BIT(XTENSA_OPTION_REGION_TRANSLATION
) |
794 XTENSA_OPTION_BIT(XTENSA_OPTION_MMU
))) {
796 cpu_fprintf(f
, "ITLB:\n");
797 dump_tlb(f
, cpu_fprintf
, env
, false);
798 cpu_fprintf(f
, "\nDTLB:\n");
799 dump_tlb(f
, cpu_fprintf
, env
, true);
801 cpu_fprintf(f
, "No TLB for this CPU core\n");
805 void xtensa_runstall(CPUXtensaState
*env
, bool runstall
)
807 CPUState
*cpu
= CPU(xtensa_env_get_cpu(env
));
809 env
->runstall
= runstall
;
810 cpu
->halted
= runstall
;
812 cpu_interrupt(cpu
, CPU_INTERRUPT_HALT
);
814 cpu_reset_interrupt(cpu
, CPU_INTERRUPT_HALT
);