4 * handles SCLP event types
5 * - Signal Quiesce - system power down
6 * - ASCII Console Data - VT220 read and write
8 * Copyright IBM, Corp. 2012
11 * Heinz Graalfs <graalfs@de.ibm.com>
13 * This work is licensed under the terms of the GNU GPL, version 2 or (at your
14 * option) any later version. See the COPYING file in the top-level directory.
18 #include "qemu/osdep.h"
19 #include "qapi/error.h"
20 #include "qemu/module.h"
22 #include "hw/s390x/sclp.h"
23 #include "migration/vmstate.h"
24 #include "hw/s390x/event-facility.h"
26 typedef struct SCLPEventsBus
{
30 /* we need to save 32 bit chunks for compatibility */
32 #define RECV_MASK_LOWER 1
33 #define RECV_MASK_UPPER 0
34 #else /* little endian host */
35 #define RECV_MASK_LOWER 0
36 #define RECV_MASK_UPPER 1
39 struct SCLPEventFacility
{
40 SysBusDevice parent_obj
;
42 SCLPEvent quiesce
, cpu_hotplug
;
43 /* guest's receive mask */
45 uint32_t receive_mask_pieces
[2];
46 sccb_mask_t receive_mask
;
49 * when false, we keep the same broken, backwards compatible behaviour as
50 * before, allowing only masks of size exactly 4; when true, we implement
51 * the architecture correctly, allowing all valid mask sizes. Needed for
52 * migration toward older versions.
54 bool allow_all_mask_sizes
;
55 /* length of the receive mask */
59 /* return true if any child has event pending set */
60 static bool event_pending(SCLPEventFacility
*ef
)
64 SCLPEventClass
*event_class
;
66 QTAILQ_FOREACH(kid
, &ef
->sbus
.qbus
.children
, sibling
) {
67 event
= SCLP_EVENT(kid
->child
);
68 event_class
= SCLP_EVENT_GET_CLASS(event
);
69 if (event
->event_pending
&&
70 event_class
->get_send_mask() & ef
->receive_mask
) {
77 static sccb_mask_t
get_host_send_mask(SCLPEventFacility
*ef
)
81 SCLPEventClass
*child
;
85 QTAILQ_FOREACH(kid
, &ef
->sbus
.qbus
.children
, sibling
) {
86 DeviceState
*qdev
= kid
->child
;
87 child
= SCLP_EVENT_GET_CLASS((SCLPEvent
*) qdev
);
88 mask
|= child
->get_send_mask();
93 static sccb_mask_t
get_host_receive_mask(SCLPEventFacility
*ef
)
97 SCLPEventClass
*child
;
101 QTAILQ_FOREACH(kid
, &ef
->sbus
.qbus
.children
, sibling
) {
102 DeviceState
*qdev
= kid
->child
;
103 child
= SCLP_EVENT_GET_CLASS((SCLPEvent
*) qdev
);
104 mask
|= child
->get_receive_mask();
109 static uint16_t write_event_length_check(SCCB
*sccb
)
113 EventBufferHeader
*event
;
114 WriteEventData
*wed
= (WriteEventData
*) sccb
;
116 event
= (EventBufferHeader
*) &wed
->ebh
;
117 for (slen
= sccb_data_len(sccb
); slen
> 0; slen
-= elen
) {
118 elen
= be16_to_cpu(event
->length
);
119 if (elen
< sizeof(*event
) || elen
> slen
) {
120 return SCLP_RC_EVENT_BUFFER_SYNTAX_ERROR
;
122 event
= (void *) event
+ elen
;
125 return SCLP_RC_INCONSISTENT_LENGTHS
;
127 return SCLP_RC_NORMAL_COMPLETION
;
130 static uint16_t handle_write_event_buf(SCLPEventFacility
*ef
,
131 EventBufferHeader
*event_buf
, SCCB
*sccb
)
138 rc
= SCLP_RC_INVALID_FUNCTION
;
140 QTAILQ_FOREACH(kid
, &ef
->sbus
.qbus
.children
, sibling
) {
141 DeviceState
*qdev
= kid
->child
;
142 event
= (SCLPEvent
*) qdev
;
143 ec
= SCLP_EVENT_GET_CLASS(event
);
145 if (ec
->write_event_data
&&
146 ec
->can_handle_event(event_buf
->type
)) {
147 rc
= ec
->write_event_data(event
, event_buf
);
154 static uint16_t handle_sccb_write_events(SCLPEventFacility
*ef
, SCCB
*sccb
)
159 EventBufferHeader
*event_buf
;
160 WriteEventData
*wed
= (WriteEventData
*) sccb
;
162 event_buf
= &wed
->ebh
;
163 rc
= SCLP_RC_NORMAL_COMPLETION
;
165 /* loop over all contained event buffers */
166 for (slen
= sccb_data_len(sccb
); slen
> 0; slen
-= elen
) {
167 elen
= be16_to_cpu(event_buf
->length
);
169 /* in case of a previous error mark all trailing buffers
171 if (rc
!= SCLP_RC_NORMAL_COMPLETION
) {
172 event_buf
->flags
&= ~(SCLP_EVENT_BUFFER_ACCEPTED
);
174 rc
= handle_write_event_buf(ef
, event_buf
, sccb
);
176 event_buf
= (void *) event_buf
+ elen
;
181 static void write_event_data(SCLPEventFacility
*ef
, SCCB
*sccb
)
183 if (sccb
->h
.function_code
!= SCLP_FC_NORMAL_WRITE
) {
184 sccb
->h
.response_code
= cpu_to_be16(SCLP_RC_INVALID_FUNCTION
);
187 if (be16_to_cpu(sccb
->h
.length
) < 8) {
188 sccb
->h
.response_code
= cpu_to_be16(SCLP_RC_INSUFFICIENT_SCCB_LENGTH
);
191 /* first do a sanity check of the write events */
192 sccb
->h
.response_code
= cpu_to_be16(write_event_length_check(sccb
));
194 /* if no early error, then execute */
195 if (sccb
->h
.response_code
== be16_to_cpu(SCLP_RC_NORMAL_COMPLETION
)) {
196 sccb
->h
.response_code
=
197 cpu_to_be16(handle_sccb_write_events(ef
, sccb
));
201 static uint16_t handle_sccb_read_events(SCLPEventFacility
*ef
, SCCB
*sccb
,
210 EventBufferHeader
*event_buf
;
211 ReadEventData
*red
= (ReadEventData
*) sccb
;
213 event_buf
= &red
->ebh
;
214 event_buf
->length
= 0;
215 slen
= sccb_data_len(sccb
);
217 rc
= SCLP_RC_NO_EVENT_BUFFERS_STORED
;
219 QTAILQ_FOREACH(kid
, &ef
->sbus
.qbus
.children
, sibling
) {
220 DeviceState
*qdev
= kid
->child
;
221 event
= (SCLPEvent
*) qdev
;
222 ec
= SCLP_EVENT_GET_CLASS(event
);
224 if (mask
& ec
->get_send_mask()) {
225 if (ec
->read_event_data(event
, event_buf
, &slen
)) {
226 elen
= be16_to_cpu(event_buf
->length
);
227 event_buf
= (EventBufferHeader
*) ((char *)event_buf
+ elen
);
228 rc
= SCLP_RC_NORMAL_COMPLETION
;
233 if (sccb
->h
.control_mask
[2] & SCLP_VARIABLE_LENGTH_RESPONSE
) {
234 /* architecture suggests to reset variable-length-response bit */
235 sccb
->h
.control_mask
[2] &= ~SCLP_VARIABLE_LENGTH_RESPONSE
;
236 /* with a new length value */
237 sccb
->h
.length
= cpu_to_be16(SCCB_SIZE
- slen
);
242 /* copy up to src_len bytes and fill the rest of dst with zeroes */
243 static void copy_mask(uint8_t *dst
, uint8_t *src
, uint16_t dst_len
,
248 for (i
= 0; i
< dst_len
; i
++) {
249 dst
[i
] = i
< src_len
? src
[i
] : 0;
253 static void read_event_data(SCLPEventFacility
*ef
, SCCB
*sccb
)
255 sccb_mask_t sclp_active_selection_mask
;
256 sccb_mask_t sclp_cp_receive_mask
;
258 ReadEventData
*red
= (ReadEventData
*) sccb
;
260 if (be16_to_cpu(sccb
->h
.length
) != SCCB_SIZE
) {
261 sccb
->h
.response_code
= cpu_to_be16(SCLP_RC_INSUFFICIENT_SCCB_LENGTH
);
265 switch (sccb
->h
.function_code
) {
266 case SCLP_UNCONDITIONAL_READ
:
267 sccb
->h
.response_code
= cpu_to_be16(
268 handle_sccb_read_events(ef
, sccb
, ef
->receive_mask
));
270 case SCLP_SELECTIVE_READ
:
271 /* get active selection mask */
272 sclp_cp_receive_mask
= ef
->receive_mask
;
274 copy_mask((uint8_t *)&sclp_active_selection_mask
, (uint8_t *)&red
->mask
,
275 sizeof(sclp_active_selection_mask
), ef
->mask_length
);
276 sclp_active_selection_mask
= be64_to_cpu(sclp_active_selection_mask
);
277 if (!sclp_cp_receive_mask
||
278 (sclp_active_selection_mask
& ~sclp_cp_receive_mask
)) {
279 sccb
->h
.response_code
=
280 cpu_to_be16(SCLP_RC_INVALID_SELECTION_MASK
);
282 sccb
->h
.response_code
= cpu_to_be16(
283 handle_sccb_read_events(ef
, sccb
, sclp_active_selection_mask
));
287 sccb
->h
.response_code
= cpu_to_be16(SCLP_RC_INVALID_FUNCTION
);
291 static void write_event_mask(SCLPEventFacility
*ef
, SCCB
*sccb
)
293 WriteEventMask
*we_mask
= (WriteEventMask
*) sccb
;
294 uint16_t mask_length
= be16_to_cpu(we_mask
->mask_length
);
295 sccb_mask_t tmp_mask
;
297 if (!mask_length
|| (mask_length
> SCLP_EVENT_MASK_LEN_MAX
) ||
298 ((mask_length
!= 4) && !ef
->allow_all_mask_sizes
)) {
299 sccb
->h
.response_code
= cpu_to_be16(SCLP_RC_INVALID_MASK_LENGTH
);
304 * Note: We currently only support masks up to 8 byte length;
305 * the remainder is filled up with zeroes. Older Linux
306 * kernels use a 4 byte mask length, newer ones can use both
307 * 8 or 4 depending on what is available on the host.
310 /* keep track of the guest's capability masks */
311 copy_mask((uint8_t *)&tmp_mask
, WEM_CP_RECEIVE_MASK(we_mask
, mask_length
),
312 sizeof(tmp_mask
), mask_length
);
313 ef
->receive_mask
= be64_to_cpu(tmp_mask
);
315 /* return the SCLP's capability masks to the guest */
316 tmp_mask
= cpu_to_be64(get_host_receive_mask(ef
));
317 copy_mask(WEM_RECEIVE_MASK(we_mask
, mask_length
), (uint8_t *)&tmp_mask
,
318 mask_length
, sizeof(tmp_mask
));
319 tmp_mask
= cpu_to_be64(get_host_send_mask(ef
));
320 copy_mask(WEM_SEND_MASK(we_mask
, mask_length
), (uint8_t *)&tmp_mask
,
321 mask_length
, sizeof(tmp_mask
));
323 sccb
->h
.response_code
= cpu_to_be16(SCLP_RC_NORMAL_COMPLETION
);
324 ef
->mask_length
= mask_length
;
327 /* qemu object creation and initialization functions */
329 #define TYPE_SCLP_EVENTS_BUS "s390-sclp-events-bus"
331 static const TypeInfo sclp_events_bus_info
= {
332 .name
= TYPE_SCLP_EVENTS_BUS
,
336 static void command_handler(SCLPEventFacility
*ef
, SCCB
*sccb
, uint64_t code
)
338 switch (code
& SCLP_CMD_CODE_MASK
) {
339 case SCLP_CMD_READ_EVENT_DATA
:
340 read_event_data(ef
, sccb
);
342 case SCLP_CMD_WRITE_EVENT_DATA
:
343 write_event_data(ef
, sccb
);
345 case SCLP_CMD_WRITE_EVENT_MASK
:
346 write_event_mask(ef
, sccb
);
351 static bool vmstate_event_facility_mask64_needed(void *opaque
)
353 SCLPEventFacility
*ef
= opaque
;
355 return (ef
->receive_mask
& 0xFFFFFFFF) != 0;
358 static bool vmstate_event_facility_mask_length_needed(void *opaque
)
360 SCLPEventFacility
*ef
= opaque
;
362 return ef
->allow_all_mask_sizes
;
365 static const VMStateDescription vmstate_event_facility_mask64
= {
366 .name
= "vmstate-event-facility/mask64",
368 .minimum_version_id
= 0,
369 .needed
= vmstate_event_facility_mask64_needed
,
370 .fields
= (const VMStateField
[]) {
371 VMSTATE_UINT32(receive_mask_pieces
[RECV_MASK_LOWER
], SCLPEventFacility
),
372 VMSTATE_END_OF_LIST()
376 static const VMStateDescription vmstate_event_facility_mask_length
= {
377 .name
= "vmstate-event-facility/mask_length",
379 .minimum_version_id
= 0,
380 .needed
= vmstate_event_facility_mask_length_needed
,
381 .fields
= (const VMStateField
[]) {
382 VMSTATE_UINT16(mask_length
, SCLPEventFacility
),
383 VMSTATE_END_OF_LIST()
387 static const VMStateDescription vmstate_event_facility
= {
388 .name
= "vmstate-event-facility",
390 .minimum_version_id
= 0,
391 .fields
= (const VMStateField
[]) {
392 VMSTATE_UINT32(receive_mask_pieces
[RECV_MASK_UPPER
], SCLPEventFacility
),
393 VMSTATE_END_OF_LIST()
395 .subsections
= (const VMStateDescription
* const []) {
396 &vmstate_event_facility_mask64
,
397 &vmstate_event_facility_mask_length
,
402 static void sclp_event_set_allow_all_mask_sizes(Object
*obj
, bool value
,
405 SCLPEventFacility
*ef
= (SCLPEventFacility
*)obj
;
407 ef
->allow_all_mask_sizes
= value
;
410 static bool sclp_event_get_allow_all_mask_sizes(Object
*obj
, Error
**errp
)
412 SCLPEventFacility
*ef
= (SCLPEventFacility
*)obj
;
414 return ef
->allow_all_mask_sizes
;
417 static void init_event_facility(Object
*obj
)
419 SCLPEventFacility
*event_facility
= EVENT_FACILITY(obj
);
420 DeviceState
*sdev
= DEVICE(obj
);
422 event_facility
->mask_length
= 4;
423 event_facility
->allow_all_mask_sizes
= true;
424 object_property_add_bool(obj
, "allow_all_mask_sizes",
425 sclp_event_get_allow_all_mask_sizes
,
426 sclp_event_set_allow_all_mask_sizes
);
428 /* Spawn a new bus for SCLP events */
429 qbus_init(&event_facility
->sbus
, sizeof(event_facility
->sbus
),
430 TYPE_SCLP_EVENTS_BUS
, sdev
, NULL
);
432 object_initialize_child(obj
, TYPE_SCLP_QUIESCE
,
433 &event_facility
->quiesce
,
436 object_initialize_child(obj
, TYPE_SCLP_CPU_HOTPLUG
,
437 &event_facility
->cpu_hotplug
,
438 TYPE_SCLP_CPU_HOTPLUG
);
441 static void realize_event_facility(DeviceState
*dev
, Error
**errp
)
443 SCLPEventFacility
*event_facility
= EVENT_FACILITY(dev
);
445 if (!qdev_realize(DEVICE(&event_facility
->quiesce
),
446 BUS(&event_facility
->sbus
), errp
)) {
449 if (!qdev_realize(DEVICE(&event_facility
->cpu_hotplug
),
450 BUS(&event_facility
->sbus
), errp
)) {
451 qdev_unrealize(DEVICE(&event_facility
->quiesce
));
456 static void reset_event_facility(DeviceState
*dev
)
458 SCLPEventFacility
*sdev
= EVENT_FACILITY(dev
);
460 sdev
->receive_mask
= 0;
463 static void init_event_facility_class(ObjectClass
*klass
, void *data
)
465 SysBusDeviceClass
*sbdc
= SYS_BUS_DEVICE_CLASS(klass
);
466 DeviceClass
*dc
= DEVICE_CLASS(sbdc
);
467 SCLPEventFacilityClass
*k
= EVENT_FACILITY_CLASS(dc
);
469 dc
->realize
= realize_event_facility
;
470 device_class_set_legacy_reset(dc
, reset_event_facility
);
471 dc
->vmsd
= &vmstate_event_facility
;
472 set_bit(DEVICE_CATEGORY_MISC
, dc
->categories
);
473 k
->command_handler
= command_handler
;
474 k
->event_pending
= event_pending
;
477 static const TypeInfo sclp_event_facility_info
= {
478 .name
= TYPE_SCLP_EVENT_FACILITY
,
479 .parent
= TYPE_SYS_BUS_DEVICE
,
480 .instance_init
= init_event_facility
,
481 .instance_size
= sizeof(SCLPEventFacility
),
482 .class_init
= init_event_facility_class
,
483 .class_size
= sizeof(SCLPEventFacilityClass
),
486 static void event_realize(DeviceState
*qdev
, Error
**errp
)
488 SCLPEvent
*event
= SCLP_EVENT(qdev
);
489 SCLPEventClass
*child
= SCLP_EVENT_GET_CLASS(event
);
492 int rc
= child
->init(event
);
494 error_setg(errp
, "SCLP event initialization failed.");
500 static void event_class_init(ObjectClass
*klass
, void *data
)
502 DeviceClass
*dc
= DEVICE_CLASS(klass
);
504 dc
->bus_type
= TYPE_SCLP_EVENTS_BUS
;
505 dc
->realize
= event_realize
;
508 static const TypeInfo sclp_event_type_info
= {
509 .name
= TYPE_SCLP_EVENT
,
510 .parent
= TYPE_DEVICE
,
511 .instance_size
= sizeof(SCLPEvent
),
512 .class_init
= event_class_init
,
513 .class_size
= sizeof(SCLPEventClass
),
517 static void register_types(void)
519 type_register_static(&sclp_events_bus_info
);
520 type_register_static(&sclp_event_facility_info
);
521 type_register_static(&sclp_event_type_info
);
524 type_init(register_types
)
526 BusState
*sclp_get_event_facility_bus(SCLPEventFacility
*ef
)
528 return BUS(&ef
->sbus
);