qapi: add QMP dispatch functions
[qemu/mdroth.git] / hw / pl080.c
blobdd8139ba9620bce527ef0c7d9200cedf3c20e77f
1 /*
2 * Arm PrimeCell PL080/PL081 DMA controller
4 * Copyright (c) 2006 CodeSourcery.
5 * Written by Paul Brook
7 * This code is licenced under the GPL.
8 */
10 #include "sysbus.h"
12 #define PL080_MAX_CHANNELS 8
13 #define PL080_CONF_E 0x1
14 #define PL080_CONF_M1 0x2
15 #define PL080_CONF_M2 0x4
17 #define PL080_CCONF_H 0x40000
18 #define PL080_CCONF_A 0x20000
19 #define PL080_CCONF_L 0x10000
20 #define PL080_CCONF_ITC 0x08000
21 #define PL080_CCONF_IE 0x04000
22 #define PL080_CCONF_E 0x00001
24 #define PL080_CCTRL_I 0x80000000
25 #define PL080_CCTRL_DI 0x08000000
26 #define PL080_CCTRL_SI 0x04000000
27 #define PL080_CCTRL_D 0x02000000
28 #define PL080_CCTRL_S 0x01000000
30 typedef struct {
31 uint32_t src;
32 uint32_t dest;
33 uint32_t lli;
34 uint32_t ctrl;
35 uint32_t conf;
36 } pl080_channel;
38 typedef struct {
39 SysBusDevice busdev;
40 uint8_t tc_int;
41 uint8_t tc_mask;
42 uint8_t err_int;
43 uint8_t err_mask;
44 uint32_t conf;
45 uint32_t sync;
46 uint32_t req_single;
47 uint32_t req_burst;
48 pl080_channel chan[PL080_MAX_CHANNELS];
49 int nchannels;
50 /* Flag to avoid recursive DMA invocations. */
51 int running;
52 qemu_irq irq;
53 } pl080_state;
55 static const VMStateDescription vmstate_pl080_channel = {
56 .name = "pl080_channel",
57 .version_id = 1,
58 .minimum_version_id = 1,
59 .fields = (VMStateField[]) {
60 VMSTATE_UINT32(src, pl080_channel),
61 VMSTATE_UINT32(dest, pl080_channel),
62 VMSTATE_UINT32(lli, pl080_channel),
63 VMSTATE_UINT32(ctrl, pl080_channel),
64 VMSTATE_UINT32(conf, pl080_channel),
65 VMSTATE_END_OF_LIST()
69 static const VMStateDescription vmstate_pl080 = {
70 .name = "pl080",
71 .version_id = 1,
72 .minimum_version_id = 1,
73 .fields = (VMStateField[]) {
74 VMSTATE_UINT8(tc_int, pl080_state),
75 VMSTATE_UINT8(tc_mask, pl080_state),
76 VMSTATE_UINT8(err_int, pl080_state),
77 VMSTATE_UINT8(err_mask, pl080_state),
78 VMSTATE_UINT32(conf, pl080_state),
79 VMSTATE_UINT32(sync, pl080_state),
80 VMSTATE_UINT32(req_single, pl080_state),
81 VMSTATE_UINT32(req_burst, pl080_state),
82 VMSTATE_UINT8(tc_int, pl080_state),
83 VMSTATE_UINT8(tc_int, pl080_state),
84 VMSTATE_UINT8(tc_int, pl080_state),
85 VMSTATE_STRUCT_ARRAY(chan, pl080_state, PL080_MAX_CHANNELS,
86 1, vmstate_pl080_channel, pl080_channel),
87 VMSTATE_INT32(running, pl080_state),
88 VMSTATE_END_OF_LIST()
92 static const unsigned char pl080_id[] =
93 { 0x80, 0x10, 0x04, 0x0a, 0x0d, 0xf0, 0x05, 0xb1 };
95 static const unsigned char pl081_id[] =
96 { 0x81, 0x10, 0x04, 0x0a, 0x0d, 0xf0, 0x05, 0xb1 };
98 static void pl080_update(pl080_state *s)
100 if ((s->tc_int & s->tc_mask)
101 || (s->err_int & s->err_mask))
102 qemu_irq_raise(s->irq);
103 else
104 qemu_irq_lower(s->irq);
107 static void pl080_run(pl080_state *s)
109 int c;
110 int flow;
111 pl080_channel *ch;
112 int swidth;
113 int dwidth;
114 int xsize;
115 int n;
116 int src_id;
117 int dest_id;
118 int size;
119 uint8_t buff[4];
120 uint32_t req;
122 s->tc_mask = 0;
123 for (c = 0; c < s->nchannels; c++) {
124 if (s->chan[c].conf & PL080_CCONF_ITC)
125 s->tc_mask |= 1 << c;
126 if (s->chan[c].conf & PL080_CCONF_IE)
127 s->err_mask |= 1 << c;
130 if ((s->conf & PL080_CONF_E) == 0)
131 return;
133 hw_error("DMA active\n");
134 /* If we are already in the middle of a DMA operation then indicate that
135 there may be new DMA requests and return immediately. */
136 if (s->running) {
137 s->running++;
138 return;
140 s->running = 1;
141 while (s->running) {
142 for (c = 0; c < s->nchannels; c++) {
143 ch = &s->chan[c];
144 again:
145 /* Test if thiws channel has any pending DMA requests. */
146 if ((ch->conf & (PL080_CCONF_H | PL080_CCONF_E))
147 != PL080_CCONF_E)
148 continue;
149 flow = (ch->conf >> 11) & 7;
150 if (flow >= 4) {
151 hw_error(
152 "pl080_run: Peripheral flow control not implemented\n");
154 src_id = (ch->conf >> 1) & 0x1f;
155 dest_id = (ch->conf >> 6) & 0x1f;
156 size = ch->ctrl & 0xfff;
157 req = s->req_single | s->req_burst;
158 switch (flow) {
159 case 0:
160 break;
161 case 1:
162 if ((req & (1u << dest_id)) == 0)
163 size = 0;
164 break;
165 case 2:
166 if ((req & (1u << src_id)) == 0)
167 size = 0;
168 break;
169 case 3:
170 if ((req & (1u << src_id)) == 0
171 || (req & (1u << dest_id)) == 0)
172 size = 0;
173 break;
175 if (!size)
176 continue;
178 /* Transfer one element. */
179 /* ??? Should transfer multiple elements for a burst request. */
180 /* ??? Unclear what the proper behavior is when source and
181 destination widths are different. */
182 swidth = 1 << ((ch->ctrl >> 18) & 7);
183 dwidth = 1 << ((ch->ctrl >> 21) & 7);
184 for (n = 0; n < dwidth; n+= swidth) {
185 cpu_physical_memory_read(ch->src, buff + n, swidth);
186 if (ch->ctrl & PL080_CCTRL_SI)
187 ch->src += swidth;
189 xsize = (dwidth < swidth) ? swidth : dwidth;
190 /* ??? This may pad the value incorrectly for dwidth < 32. */
191 for (n = 0; n < xsize; n += dwidth) {
192 cpu_physical_memory_write(ch->dest + n, buff + n, dwidth);
193 if (ch->ctrl & PL080_CCTRL_DI)
194 ch->dest += swidth;
197 size--;
198 ch->ctrl = (ch->ctrl & 0xfffff000) | size;
199 if (size == 0) {
200 /* Transfer complete. */
201 if (ch->lli) {
202 ch->src = ldl_le_phys(ch->lli);
203 ch->dest = ldl_le_phys(ch->lli + 4);
204 ch->ctrl = ldl_le_phys(ch->lli + 12);
205 ch->lli = ldl_le_phys(ch->lli + 8);
206 } else {
207 ch->conf &= ~PL080_CCONF_E;
209 if (ch->ctrl & PL080_CCTRL_I) {
210 s->tc_int |= 1 << c;
213 goto again;
215 if (--s->running)
216 s->running = 1;
220 static uint32_t pl080_read(void *opaque, target_phys_addr_t offset)
222 pl080_state *s = (pl080_state *)opaque;
223 uint32_t i;
224 uint32_t mask;
226 if (offset >= 0xfe0 && offset < 0x1000) {
227 if (s->nchannels == 8) {
228 return pl080_id[(offset - 0xfe0) >> 2];
229 } else {
230 return pl081_id[(offset - 0xfe0) >> 2];
233 if (offset >= 0x100 && offset < 0x200) {
234 i = (offset & 0xe0) >> 5;
235 if (i >= s->nchannels)
236 goto bad_offset;
237 switch (offset >> 2) {
238 case 0: /* SrcAddr */
239 return s->chan[i].src;
240 case 1: /* DestAddr */
241 return s->chan[i].dest;
242 case 2: /* LLI */
243 return s->chan[i].lli;
244 case 3: /* Control */
245 return s->chan[i].ctrl;
246 case 4: /* Configuration */
247 return s->chan[i].conf;
248 default:
249 goto bad_offset;
252 switch (offset >> 2) {
253 case 0: /* IntStatus */
254 return (s->tc_int & s->tc_mask) | (s->err_int & s->err_mask);
255 case 1: /* IntTCStatus */
256 return (s->tc_int & s->tc_mask);
257 case 3: /* IntErrorStatus */
258 return (s->err_int & s->err_mask);
259 case 5: /* RawIntTCStatus */
260 return s->tc_int;
261 case 6: /* RawIntErrorStatus */
262 return s->err_int;
263 case 7: /* EnbldChns */
264 mask = 0;
265 for (i = 0; i < s->nchannels; i++) {
266 if (s->chan[i].conf & PL080_CCONF_E)
267 mask |= 1 << i;
269 return mask;
270 case 8: /* SoftBReq */
271 case 9: /* SoftSReq */
272 case 10: /* SoftLBReq */
273 case 11: /* SoftLSReq */
274 /* ??? Implement these. */
275 return 0;
276 case 12: /* Configuration */
277 return s->conf;
278 case 13: /* Sync */
279 return s->sync;
280 default:
281 bad_offset:
282 hw_error("pl080_read: Bad offset %x\n", (int)offset);
283 return 0;
287 static void pl080_write(void *opaque, target_phys_addr_t offset,
288 uint32_t value)
290 pl080_state *s = (pl080_state *)opaque;
291 int i;
293 if (offset >= 0x100 && offset < 0x200) {
294 i = (offset & 0xe0) >> 5;
295 if (i >= s->nchannels)
296 goto bad_offset;
297 switch (offset >> 2) {
298 case 0: /* SrcAddr */
299 s->chan[i].src = value;
300 break;
301 case 1: /* DestAddr */
302 s->chan[i].dest = value;
303 break;
304 case 2: /* LLI */
305 s->chan[i].lli = value;
306 break;
307 case 3: /* Control */
308 s->chan[i].ctrl = value;
309 break;
310 case 4: /* Configuration */
311 s->chan[i].conf = value;
312 pl080_run(s);
313 break;
316 switch (offset >> 2) {
317 case 2: /* IntTCClear */
318 s->tc_int &= ~value;
319 break;
320 case 4: /* IntErrorClear */
321 s->err_int &= ~value;
322 break;
323 case 8: /* SoftBReq */
324 case 9: /* SoftSReq */
325 case 10: /* SoftLBReq */
326 case 11: /* SoftLSReq */
327 /* ??? Implement these. */
328 hw_error("pl080_write: Soft DMA not implemented\n");
329 break;
330 case 12: /* Configuration */
331 s->conf = value;
332 if (s->conf & (PL080_CONF_M1 | PL080_CONF_M1)) {
333 hw_error("pl080_write: Big-endian DMA not implemented\n");
335 pl080_run(s);
336 break;
337 case 13: /* Sync */
338 s->sync = value;
339 break;
340 default:
341 bad_offset:
342 hw_error("pl080_write: Bad offset %x\n", (int)offset);
344 pl080_update(s);
347 static CPUReadMemoryFunc * const pl080_readfn[] = {
348 pl080_read,
349 pl080_read,
350 pl080_read
353 static CPUWriteMemoryFunc * const pl080_writefn[] = {
354 pl080_write,
355 pl080_write,
356 pl080_write
359 static int pl08x_init(SysBusDevice *dev, int nchannels)
361 int iomemtype;
362 pl080_state *s = FROM_SYSBUS(pl080_state, dev);
364 iomemtype = cpu_register_io_memory(pl080_readfn,
365 pl080_writefn, s,
366 DEVICE_NATIVE_ENDIAN);
367 sysbus_init_mmio(dev, 0x1000, iomemtype);
368 sysbus_init_irq(dev, &s->irq);
369 s->nchannels = nchannels;
370 return 0;
373 static int pl080_init(SysBusDevice *dev)
375 return pl08x_init(dev, 8);
378 static int pl081_init(SysBusDevice *dev)
380 return pl08x_init(dev, 2);
383 static SysBusDeviceInfo pl080_info = {
384 .init = pl080_init,
385 .qdev.name = "pl080",
386 .qdev.size = sizeof(pl080_state),
387 .qdev.vmsd = &vmstate_pl080,
388 .qdev.no_user = 1,
391 static SysBusDeviceInfo pl081_info = {
392 .init = pl081_init,
393 .qdev.name = "pl081",
394 .qdev.size = sizeof(pl080_state),
395 .qdev.vmsd = &vmstate_pl080,
396 .qdev.no_user = 1,
399 /* The PL080 and PL081 are the same except for the number of channels
400 they implement (8 and 2 respectively). */
401 static void pl080_register_devices(void)
403 sysbus_register_withprop(&pl080_info);
404 sysbus_register_withprop(&pl081_info);
407 device_init(pl080_register_devices)