spice/qxl: zap spice 0.4 migration compatibility bits
[qemu/mdroth.git] / cpu-common.h
blob6d4a898ad14881336148c69836574bb24f9945a9
1 #ifndef CPU_COMMON_H
2 #define CPU_COMMON_H 1
4 /* CPU interfaces that are target indpendent. */
6 #if defined(__arm__) || defined(__sparc__) || defined(__mips__) || defined(__hppa__) || defined(__ia64__)
7 #define WORDS_ALIGNED
8 #endif
10 #ifdef TARGET_PHYS_ADDR_BITS
11 #include "targphys.h"
12 #endif
14 #ifndef NEED_CPU_H
15 #include "poison.h"
16 #endif
18 #include "bswap.h"
19 #include "qemu-queue.h"
21 #if !defined(CONFIG_USER_ONLY)
23 enum device_endian {
24 DEVICE_NATIVE_ENDIAN,
25 DEVICE_BIG_ENDIAN,
26 DEVICE_LITTLE_ENDIAN,
29 /* address in the RAM (different from a physical address) */
30 typedef unsigned long ram_addr_t;
32 /* memory API */
34 typedef void CPUWriteMemoryFunc(void *opaque, target_phys_addr_t addr, uint32_t value);
35 typedef uint32_t CPUReadMemoryFunc(void *opaque, target_phys_addr_t addr);
37 void cpu_register_physical_memory_offset(target_phys_addr_t start_addr,
38 ram_addr_t size,
39 ram_addr_t phys_offset,
40 ram_addr_t region_offset);
41 static inline void cpu_register_physical_memory(target_phys_addr_t start_addr,
42 ram_addr_t size,
43 ram_addr_t phys_offset)
45 cpu_register_physical_memory_offset(start_addr, size, phys_offset, 0);
48 ram_addr_t cpu_get_physical_page_desc(target_phys_addr_t addr);
49 ram_addr_t qemu_ram_alloc_from_ptr(DeviceState *dev, const char *name,
50 ram_addr_t size, void *host);
51 ram_addr_t qemu_ram_alloc(DeviceState *dev, const char *name, ram_addr_t size);
52 void qemu_ram_free(ram_addr_t addr);
53 /* This should only be used for ram local to a device. */
54 void *qemu_get_ram_ptr(ram_addr_t addr);
55 /* Same but slower, to use for migration, where the order of
56 * RAMBlocks must not change. */
57 void *qemu_safe_ram_ptr(ram_addr_t addr);
58 /* This should not be used by devices. */
59 int qemu_ram_addr_from_host(void *ptr, ram_addr_t *ram_addr);
60 ram_addr_t qemu_ram_addr_from_host_nofail(void *ptr);
62 int cpu_register_io_memory(CPUReadMemoryFunc * const *mem_read,
63 CPUWriteMemoryFunc * const *mem_write,
64 void *opaque, enum device_endian endian);
65 void cpu_unregister_io_memory(int table_address);
67 void cpu_physical_memory_rw(target_phys_addr_t addr, uint8_t *buf,
68 int len, int is_write);
69 static inline void cpu_physical_memory_read(target_phys_addr_t addr,
70 uint8_t *buf, int len)
72 cpu_physical_memory_rw(addr, buf, len, 0);
74 static inline void cpu_physical_memory_write(target_phys_addr_t addr,
75 const uint8_t *buf, int len)
77 cpu_physical_memory_rw(addr, (uint8_t *)buf, len, 1);
79 void *cpu_physical_memory_map(target_phys_addr_t addr,
80 target_phys_addr_t *plen,
81 int is_write);
82 void cpu_physical_memory_unmap(void *buffer, target_phys_addr_t len,
83 int is_write, target_phys_addr_t access_len);
84 void *cpu_register_map_client(void *opaque, void (*callback)(void *opaque));
85 void cpu_unregister_map_client(void *cookie);
87 struct CPUPhysMemoryClient;
88 typedef struct CPUPhysMemoryClient CPUPhysMemoryClient;
89 struct CPUPhysMemoryClient {
90 void (*set_memory)(struct CPUPhysMemoryClient *client,
91 target_phys_addr_t start_addr,
92 ram_addr_t size,
93 ram_addr_t phys_offset);
94 int (*sync_dirty_bitmap)(struct CPUPhysMemoryClient *client,
95 target_phys_addr_t start_addr,
96 target_phys_addr_t end_addr);
97 int (*migration_log)(struct CPUPhysMemoryClient *client,
98 int enable);
99 QLIST_ENTRY(CPUPhysMemoryClient) list;
102 void cpu_register_phys_memory_client(CPUPhysMemoryClient *);
103 void cpu_unregister_phys_memory_client(CPUPhysMemoryClient *);
105 /* Coalesced MMIO regions are areas where write operations can be reordered.
106 * This usually implies that write operations are side-effect free. This allows
107 * batching which can make a major impact on performance when using
108 * virtualization.
110 void qemu_register_coalesced_mmio(target_phys_addr_t addr, ram_addr_t size);
112 void qemu_unregister_coalesced_mmio(target_phys_addr_t addr, ram_addr_t size);
114 void qemu_flush_coalesced_mmio_buffer(void);
116 uint32_t ldub_phys(target_phys_addr_t addr);
117 uint32_t lduw_phys(target_phys_addr_t addr);
118 uint32_t ldl_phys(target_phys_addr_t addr);
119 uint64_t ldq_phys(target_phys_addr_t addr);
120 void stl_phys_notdirty(target_phys_addr_t addr, uint32_t val);
121 void stq_phys_notdirty(target_phys_addr_t addr, uint64_t val);
122 void stb_phys(target_phys_addr_t addr, uint32_t val);
123 void stw_phys(target_phys_addr_t addr, uint32_t val);
124 void stl_phys(target_phys_addr_t addr, uint32_t val);
125 void stq_phys(target_phys_addr_t addr, uint64_t val);
127 void cpu_physical_memory_write_rom(target_phys_addr_t addr,
128 const uint8_t *buf, int len);
130 #define IO_MEM_SHIFT 3
132 #define IO_MEM_RAM (0 << IO_MEM_SHIFT) /* hardcoded offset */
133 #define IO_MEM_ROM (1 << IO_MEM_SHIFT) /* hardcoded offset */
134 #define IO_MEM_UNASSIGNED (2 << IO_MEM_SHIFT)
135 #define IO_MEM_NOTDIRTY (3 << IO_MEM_SHIFT)
137 /* Acts like a ROM when read and like a device when written. */
138 #define IO_MEM_ROMD (1)
139 #define IO_MEM_SUBPAGE (2)
141 #endif
143 #endif /* !CPU_COMMON_H */