1 /* SPDX-License-Identifier: GPL-2.0-only */
3 * Copyright (C) 2015, 2016 ARM Ltd.
5 #ifndef __KVM_ARM_VGIC_MMIO_H__
6 #define __KVM_ARM_VGIC_MMIO_H__
8 struct vgic_register_region
{
9 unsigned int reg_offset
;
11 unsigned int bits_per_irq
;
12 unsigned int access_flags
;
14 unsigned long (*read
)(struct kvm_vcpu
*vcpu
, gpa_t addr
,
16 unsigned long (*its_read
)(struct kvm
*kvm
, struct vgic_its
*its
,
17 gpa_t addr
, unsigned int len
);
20 void (*write
)(struct kvm_vcpu
*vcpu
, gpa_t addr
,
21 unsigned int len
, unsigned long val
);
22 void (*its_write
)(struct kvm
*kvm
, struct vgic_its
*its
,
23 gpa_t addr
, unsigned int len
,
26 unsigned long (*uaccess_read
)(struct kvm_vcpu
*vcpu
, gpa_t addr
,
29 int (*uaccess_write
)(struct kvm_vcpu
*vcpu
, gpa_t addr
,
30 unsigned int len
, unsigned long val
);
31 int (*uaccess_its_write
)(struct kvm
*kvm
, struct vgic_its
*its
,
32 gpa_t addr
, unsigned int len
,
37 extern struct kvm_io_device_ops kvm_io_gic_ops
;
39 #define VGIC_ACCESS_8bit 1
40 #define VGIC_ACCESS_32bit 2
41 #define VGIC_ACCESS_64bit 4
44 * Generate a mask that covers the number of bytes required to address
45 * up to 1024 interrupts, each represented by <bits> bits. This assumes
46 * that <bits> is a power of two.
48 #define VGIC_ADDR_IRQ_MASK(bits) (((bits) * 1024 / 8) - 1)
51 * (addr & mask) gives us the _byte_ offset for the INT ID.
52 * We multiply this by 8 the get the _bit_ offset, then divide this by
53 * the number of bits to learn the actual INT ID.
54 * But instead of a division (which requires a "long long div" implementation),
55 * we shift by the binary logarithm of <bits>.
56 * This assumes that <bits> is a power of two.
58 #define VGIC_ADDR_TO_INTID(addr, bits) (((addr) & VGIC_ADDR_IRQ_MASK(bits)) * \
62 * Some VGIC registers store per-IRQ information, with a different number
63 * of bits per IRQ. For those registers this macro is used.
64 * The _WITH_LENGTH version instantiates registers with a fixed length
65 * and is mutually exclusive with the _PER_IRQ version.
67 #define REGISTER_DESC_WITH_BITS_PER_IRQ(off, rd, wr, ur, uw, bpi, acc) \
70 .bits_per_irq = bpi, \
71 .len = bpi * 1024 / 8, \
72 .access_flags = acc, \
76 .uaccess_write = uw, \
79 #define REGISTER_DESC_WITH_LENGTH(off, rd, wr, length, acc) \
84 .access_flags = acc, \
89 #define REGISTER_DESC_WITH_LENGTH_UACCESS(off, rd, wr, urd, uwr, length, acc) \
94 .access_flags = acc, \
97 .uaccess_read = urd, \
98 .uaccess_write = uwr, \
101 int kvm_vgic_register_mmio_region(struct kvm
*kvm
, struct kvm_vcpu
*vcpu
,
102 struct vgic_register_region
*reg_desc
,
103 struct vgic_io_device
*region
,
104 int nr_irqs
, bool offset_private
);
106 unsigned long vgic_data_mmio_bus_to_host(const void *val
, unsigned int len
);
108 void vgic_data_host_to_mmio_bus(void *buf
, unsigned int len
,
111 unsigned long extract_bytes(u64 data
, unsigned int offset
,
114 u64
update_64bit_reg(u64 reg
, unsigned int offset
, unsigned int len
,
117 unsigned long vgic_mmio_read_raz(struct kvm_vcpu
*vcpu
,
118 gpa_t addr
, unsigned int len
);
120 unsigned long vgic_mmio_read_rao(struct kvm_vcpu
*vcpu
,
121 gpa_t addr
, unsigned int len
);
123 void vgic_mmio_write_wi(struct kvm_vcpu
*vcpu
, gpa_t addr
,
124 unsigned int len
, unsigned long val
);
126 int vgic_mmio_uaccess_write_wi(struct kvm_vcpu
*vcpu
, gpa_t addr
,
127 unsigned int len
, unsigned long val
);
129 unsigned long vgic_mmio_read_group(struct kvm_vcpu
*vcpu
, gpa_t addr
,
132 void vgic_mmio_write_group(struct kvm_vcpu
*vcpu
, gpa_t addr
,
133 unsigned int len
, unsigned long val
);
135 unsigned long vgic_mmio_read_enable(struct kvm_vcpu
*vcpu
,
136 gpa_t addr
, unsigned int len
);
138 void vgic_mmio_write_senable(struct kvm_vcpu
*vcpu
,
139 gpa_t addr
, unsigned int len
,
142 void vgic_mmio_write_cenable(struct kvm_vcpu
*vcpu
,
143 gpa_t addr
, unsigned int len
,
146 unsigned long vgic_mmio_read_pending(struct kvm_vcpu
*vcpu
,
147 gpa_t addr
, unsigned int len
);
149 void vgic_mmio_write_spending(struct kvm_vcpu
*vcpu
,
150 gpa_t addr
, unsigned int len
,
153 void vgic_mmio_write_cpending(struct kvm_vcpu
*vcpu
,
154 gpa_t addr
, unsigned int len
,
157 unsigned long vgic_mmio_read_active(struct kvm_vcpu
*vcpu
,
158 gpa_t addr
, unsigned int len
);
160 void vgic_mmio_write_cactive(struct kvm_vcpu
*vcpu
,
161 gpa_t addr
, unsigned int len
,
164 void vgic_mmio_write_sactive(struct kvm_vcpu
*vcpu
,
165 gpa_t addr
, unsigned int len
,
168 int vgic_mmio_uaccess_write_cactive(struct kvm_vcpu
*vcpu
,
169 gpa_t addr
, unsigned int len
,
172 int vgic_mmio_uaccess_write_sactive(struct kvm_vcpu
*vcpu
,
173 gpa_t addr
, unsigned int len
,
176 unsigned long vgic_mmio_read_priority(struct kvm_vcpu
*vcpu
,
177 gpa_t addr
, unsigned int len
);
179 void vgic_mmio_write_priority(struct kvm_vcpu
*vcpu
,
180 gpa_t addr
, unsigned int len
,
183 unsigned long vgic_mmio_read_config(struct kvm_vcpu
*vcpu
,
184 gpa_t addr
, unsigned int len
);
186 void vgic_mmio_write_config(struct kvm_vcpu
*vcpu
,
187 gpa_t addr
, unsigned int len
,
190 int vgic_uaccess(struct kvm_vcpu
*vcpu
, struct vgic_io_device
*dev
,
191 bool is_write
, int offset
, u32
*val
);
193 u64
vgic_read_irq_line_level_info(struct kvm_vcpu
*vcpu
, u32 intid
);
195 void vgic_write_irq_line_level_info(struct kvm_vcpu
*vcpu
, u32 intid
,
198 unsigned int vgic_v2_init_dist_iodev(struct vgic_io_device
*dev
);
200 unsigned int vgic_v3_init_dist_iodev(struct vgic_io_device
*dev
);
202 u64
vgic_sanitise_outer_cacheability(u64 reg
);
203 u64
vgic_sanitise_inner_cacheability(u64 reg
);
204 u64
vgic_sanitise_shareability(u64 reg
);
205 u64
vgic_sanitise_field(u64 reg
, u64 field_mask
, int field_shift
,
206 u64 (*sanitise_fn
)(u64
));
208 /* Find the proper register handler entry given a certain address offset */
209 const struct vgic_register_region
*
210 vgic_find_mmio_region(const struct vgic_register_region
*regions
,
211 int nr_regions
, unsigned int offset
);