pseries: Add a routine to find a stable "default" vty and use it
[qemu/agraf.git] / cpu-defs.h
blob57a709b6795ddc555df045bc16849742eb2fb23d
1 /*
2 * common defines for all CPUs
4 * Copyright (c) 2003 Fabrice Bellard
6 * This library is free software; you can redistribute it and/or
7 * modify it under the terms of the GNU Lesser General Public
8 * License as published by the Free Software Foundation; either
9 * version 2 of the License, or (at your option) any later version.
11 * This library is distributed in the hope that it will be useful,
12 * but WITHOUT ANY WARRANTY; without even the implied warranty of
13 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
14 * Lesser General Public License for more details.
16 * You should have received a copy of the GNU Lesser General Public
17 * License along with this library; if not, see <http://www.gnu.org/licenses/>.
19 #ifndef CPU_DEFS_H
20 #define CPU_DEFS_H
22 #ifndef NEED_CPU_H
23 #error cpu.h included from common code
24 #endif
26 #include "config.h"
27 #include <setjmp.h>
28 #include <inttypes.h>
29 #include <signal.h>
30 #include "osdep.h"
31 #include "qemu-queue.h"
32 #include "targphys.h"
34 #ifndef TARGET_LONG_BITS
35 #error TARGET_LONG_BITS must be defined before including this header
36 #endif
38 #define TARGET_LONG_SIZE (TARGET_LONG_BITS / 8)
40 typedef int16_t target_short __attribute__ ((aligned(TARGET_SHORT_ALIGNMENT)));
41 typedef uint16_t target_ushort __attribute__((aligned(TARGET_SHORT_ALIGNMENT)));
42 typedef int32_t target_int __attribute__((aligned(TARGET_INT_ALIGNMENT)));
43 typedef uint32_t target_uint __attribute__((aligned(TARGET_INT_ALIGNMENT)));
44 typedef int64_t target_llong __attribute__((aligned(TARGET_LLONG_ALIGNMENT)));
45 typedef uint64_t target_ullong __attribute__((aligned(TARGET_LLONG_ALIGNMENT)));
46 /* target_ulong is the type of a virtual address */
47 #if TARGET_LONG_SIZE == 4
48 typedef int32_t target_long __attribute__((aligned(TARGET_LONG_ALIGNMENT)));
49 typedef uint32_t target_ulong __attribute__((aligned(TARGET_LONG_ALIGNMENT)));
50 #define TARGET_FMT_lx "%08x"
51 #define TARGET_FMT_ld "%d"
52 #define TARGET_FMT_lu "%u"
53 #elif TARGET_LONG_SIZE == 8
54 typedef int64_t target_long __attribute__((aligned(TARGET_LONG_ALIGNMENT)));
55 typedef uint64_t target_ulong __attribute__((aligned(TARGET_LONG_ALIGNMENT)));
56 #define TARGET_FMT_lx "%016" PRIx64
57 #define TARGET_FMT_ld "%" PRId64
58 #define TARGET_FMT_lu "%" PRIu64
59 #else
60 #error TARGET_LONG_SIZE undefined
61 #endif
63 #define HOST_LONG_SIZE (HOST_LONG_BITS / 8)
65 #define EXCP_INTERRUPT 0x10000 /* async interruption */
66 #define EXCP_HLT 0x10001 /* hlt instruction reached */
67 #define EXCP_DEBUG 0x10002 /* cpu stopped after a breakpoint or singlestep */
68 #define EXCP_HALTED 0x10003 /* cpu is halted (waiting for external event) */
70 #define TB_JMP_CACHE_BITS 12
71 #define TB_JMP_CACHE_SIZE (1 << TB_JMP_CACHE_BITS)
73 /* Only the bottom TB_JMP_PAGE_BITS of the jump cache hash bits vary for
74 addresses on the same page. The top bits are the same. This allows
75 TLB invalidation to quickly clear a subset of the hash table. */
76 #define TB_JMP_PAGE_BITS (TB_JMP_CACHE_BITS / 2)
77 #define TB_JMP_PAGE_SIZE (1 << TB_JMP_PAGE_BITS)
78 #define TB_JMP_ADDR_MASK (TB_JMP_PAGE_SIZE - 1)
79 #define TB_JMP_PAGE_MASK (TB_JMP_CACHE_SIZE - TB_JMP_PAGE_SIZE)
81 #if !defined(CONFIG_USER_ONLY)
82 #define CPU_TLB_BITS 8
83 #define CPU_TLB_SIZE (1 << CPU_TLB_BITS)
85 #if HOST_LONG_BITS == 32 && TARGET_LONG_BITS == 32
86 #define CPU_TLB_ENTRY_BITS 4
87 #else
88 #define CPU_TLB_ENTRY_BITS 5
89 #endif
91 typedef struct CPUTLBEntry {
92 /* bit TARGET_LONG_BITS to TARGET_PAGE_BITS : virtual address
93 bit TARGET_PAGE_BITS-1..4 : Nonzero for accesses that should not
94 go directly to ram.
95 bit 3 : indicates that the entry is invalid
96 bit 2..0 : zero
98 target_ulong addr_read;
99 target_ulong addr_write;
100 target_ulong addr_code;
101 /* Addend to virtual address to get host address. IO accesses
102 use the corresponding iotlb value. */
103 unsigned long addend;
104 /* padding to get a power of two size */
105 uint8_t dummy[(1 << CPU_TLB_ENTRY_BITS) -
106 (sizeof(target_ulong) * 3 +
107 ((-sizeof(target_ulong) * 3) & (sizeof(unsigned long) - 1)) +
108 sizeof(unsigned long))];
109 } CPUTLBEntry;
111 extern int CPUTLBEntry_wrong_size[sizeof(CPUTLBEntry) == (1 << CPU_TLB_ENTRY_BITS) ? 1 : -1];
113 #define CPU_COMMON_TLB \
114 /* The meaning of the MMU modes is defined in the target code. */ \
115 CPUTLBEntry tlb_table[NB_MMU_MODES][CPU_TLB_SIZE]; \
116 target_phys_addr_t iotlb[NB_MMU_MODES][CPU_TLB_SIZE]; \
117 target_ulong tlb_flush_addr; \
118 target_ulong tlb_flush_mask;
120 #else
122 #define CPU_COMMON_TLB
124 #endif
127 #ifdef HOST_WORDS_BIGENDIAN
128 typedef struct icount_decr_u16 {
129 uint16_t high;
130 uint16_t low;
131 } icount_decr_u16;
132 #else
133 typedef struct icount_decr_u16 {
134 uint16_t low;
135 uint16_t high;
136 } icount_decr_u16;
137 #endif
139 struct kvm_run;
140 struct KVMState;
141 struct qemu_work_item;
143 typedef struct CPUBreakpoint {
144 target_ulong pc;
145 int flags; /* BP_* */
146 QTAILQ_ENTRY(CPUBreakpoint) entry;
147 } CPUBreakpoint;
149 typedef struct CPUWatchpoint {
150 target_ulong vaddr;
151 target_ulong len_mask;
152 int flags; /* BP_* */
153 QTAILQ_ENTRY(CPUWatchpoint) entry;
154 } CPUWatchpoint;
156 #ifdef _WIN32
157 #define CPU_COMMON_THREAD \
158 void *hThread;
160 #else
161 #define CPU_COMMON_THREAD
162 #endif
164 #define CPU_TEMP_BUF_NLONGS 128
165 #define CPU_COMMON \
166 struct TranslationBlock *current_tb; /* currently executing TB */ \
167 /* soft mmu support */ \
168 /* in order to avoid passing too many arguments to the MMIO \
169 helpers, we store some rarely used information in the CPU \
170 context) */ \
171 unsigned long mem_io_pc; /* host pc at which the memory was \
172 accessed */ \
173 target_ulong mem_io_vaddr; /* target virtual addr at which the \
174 memory was accessed */ \
175 uint32_t halted; /* Nonzero if the CPU is in suspend state */ \
176 uint32_t interrupt_request; \
177 volatile sig_atomic_t exit_request; \
178 CPU_COMMON_TLB \
179 struct TranslationBlock *tb_jmp_cache[TB_JMP_CACHE_SIZE]; \
180 /* buffer for temporaries in the code generator */ \
181 long temp_buf[CPU_TEMP_BUF_NLONGS]; \
183 int64_t icount_extra; /* Instructions until next timer event. */ \
184 /* Number of cycles left, with interrupt flag in high bit. \
185 This allows a single read-compare-cbranch-write sequence to test \
186 for both decrementer underflow and exceptions. */ \
187 union { \
188 uint32_t u32; \
189 icount_decr_u16 u16; \
190 } icount_decr; \
191 uint32_t can_do_io; /* nonzero if memory mapped IO is safe. */ \
193 /* from this point: preserved by CPU reset */ \
194 /* ice debug support */ \
195 QTAILQ_HEAD(breakpoints_head, CPUBreakpoint) breakpoints; \
196 int singlestep_enabled; \
198 QTAILQ_HEAD(watchpoints_head, CPUWatchpoint) watchpoints; \
199 CPUWatchpoint *watchpoint_hit; \
201 struct GDBRegisterState *gdb_regs; \
203 /* Core interrupt code */ \
204 jmp_buf jmp_env; \
205 int exception_index; \
207 CPUState *next_cpu; /* next CPU sharing TB cache */ \
208 int cpu_index; /* CPU index (informative) */ \
209 uint32_t host_tid; /* host thread ID */ \
210 int numa_node; /* NUMA node this cpu is belonging to */ \
211 int nr_cores; /* number of cores within this CPU package */ \
212 int nr_threads;/* number of threads within this CPU */ \
213 int running; /* Nonzero if cpu is currently running(usermode). */ \
214 int thread_id; \
215 /* user data */ \
216 void *opaque; \
218 uint32_t created; \
219 uint32_t stop; /* Stop request */ \
220 uint32_t stopped; /* Artificially stopped */ \
221 struct QemuThread *thread; \
222 CPU_COMMON_THREAD \
223 struct QemuCond *halt_cond; \
224 int thread_kicked; \
225 struct qemu_work_item *queued_work_first, *queued_work_last; \
226 const char *cpu_model_str; \
227 struct KVMState *kvm_state; \
228 struct kvm_run *kvm_run; \
229 int kvm_fd; \
230 int kvm_vcpu_dirty;
232 #endif