1 /******************************************************************************
4 * Low level code to talks to Xen Store: ringbuffer and event channel.
6 * Copyright (C) 2005 Rusty Russell, IBM Corporation
8 * This program is free software; you can redistribute it and/or
9 * modify it under the terms of the GNU General Public License version 2
10 * as published by the Free Software Foundation; or, when distributed
11 * separately from the Linux kernel or incorporated into other
12 * software packages, subject to the following license:
14 * Permission is hereby granted, free of charge, to any person obtaining a copy
15 * of this source file (the "Software"), to deal in the Software without
16 * restriction, including without limitation the rights to use, copy, modify,
17 * merge, publish, distribute, sublicense, and/or sell copies of the Software,
18 * and to permit persons to whom the Software is furnished to do so, subject to
19 * the following conditions:
21 * The above copyright notice and this permission notice shall be included in
22 * all copies or substantial portions of the Software.
24 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
25 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
26 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
27 * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
28 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
29 * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
33 #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
35 #include <linux/wait.h>
36 #include <linux/interrupt.h>
37 #include <linux/kthread.h>
38 #include <linux/sched.h>
39 #include <linux/err.h>
40 #include <xen/xenbus.h>
41 #include <asm/xen/hypervisor.h>
42 #include <xen/events.h>
46 /* A list of replies. Currently only one will ever be outstanding. */
47 LIST_HEAD(xs_reply_list
);
49 /* A list of write requests. */
50 LIST_HEAD(xb_write_list
);
51 DECLARE_WAIT_QUEUE_HEAD(xb_waitq
);
52 DEFINE_MUTEX(xb_write_mutex
);
54 /* Protect xenbus reader thread against save/restore. */
55 DEFINE_MUTEX(xs_response_mutex
);
57 static int xenbus_irq
;
58 static struct task_struct
*xenbus_task
;
60 static DECLARE_WORK(probe_work
, xenbus_probe
);
63 static irqreturn_t
wake_waiting(int irq
, void *unused
)
65 if (unlikely(xenstored_ready
== 0)) {
67 schedule_work(&probe_work
);
74 static int check_indexes(XENSTORE_RING_IDX cons
, XENSTORE_RING_IDX prod
)
76 return ((prod
- cons
) <= XENSTORE_RING_SIZE
);
79 static void *get_output_chunk(XENSTORE_RING_IDX cons
,
80 XENSTORE_RING_IDX prod
,
81 char *buf
, uint32_t *len
)
83 *len
= XENSTORE_RING_SIZE
- MASK_XENSTORE_IDX(prod
);
84 if ((XENSTORE_RING_SIZE
- (prod
- cons
)) < *len
)
85 *len
= XENSTORE_RING_SIZE
- (prod
- cons
);
86 return buf
+ MASK_XENSTORE_IDX(prod
);
89 static const void *get_input_chunk(XENSTORE_RING_IDX cons
,
90 XENSTORE_RING_IDX prod
,
91 const char *buf
, uint32_t *len
)
93 *len
= XENSTORE_RING_SIZE
- MASK_XENSTORE_IDX(cons
);
94 if ((prod
- cons
) < *len
)
96 return buf
+ MASK_XENSTORE_IDX(cons
);
99 static int xb_data_to_write(void)
101 struct xenstore_domain_interface
*intf
= xen_store_interface
;
103 return (intf
->req_prod
- intf
->req_cons
) != XENSTORE_RING_SIZE
&&
104 !list_empty(&xb_write_list
);
108 * xb_write - low level write
109 * @data: buffer to send
110 * @len: length of buffer
112 * Returns number of bytes written or -err.
114 static int xb_write(const void *data
, unsigned int len
)
116 struct xenstore_domain_interface
*intf
= xen_store_interface
;
117 XENSTORE_RING_IDX cons
, prod
;
118 unsigned int bytes
= 0;
124 /* Read indexes, then verify. */
125 cons
= intf
->req_cons
;
126 prod
= intf
->req_prod
;
127 if (!check_indexes(cons
, prod
)) {
128 intf
->req_cons
= intf
->req_prod
= 0;
131 if (!xb_data_to_write())
134 /* Must write data /after/ reading the consumer index. */
137 dst
= get_output_chunk(cons
, prod
, intf
->req
, &avail
);
143 memcpy(dst
, data
, avail
);
148 /* Other side must not see new producer until data is there. */
150 intf
->req_prod
+= avail
;
152 /* Implies mb(): other side will see the updated producer. */
153 if (prod
<= intf
->req_cons
)
154 notify_remote_via_evtchn(xen_store_evtchn
);
160 static int xb_data_to_read(void)
162 struct xenstore_domain_interface
*intf
= xen_store_interface
;
163 return (intf
->rsp_cons
!= intf
->rsp_prod
);
166 static int xb_read(void *data
, unsigned int len
)
168 struct xenstore_domain_interface
*intf
= xen_store_interface
;
169 XENSTORE_RING_IDX cons
, prod
;
170 unsigned int bytes
= 0;
176 /* Read indexes, then verify. */
177 cons
= intf
->rsp_cons
;
178 prod
= intf
->rsp_prod
;
182 if (!check_indexes(cons
, prod
)) {
183 intf
->rsp_cons
= intf
->rsp_prod
= 0;
187 src
= get_input_chunk(cons
, prod
, intf
->rsp
, &avail
);
193 /* Must read data /after/ reading the producer index. */
196 memcpy(data
, src
, avail
);
201 /* Other side must not see free space until we've copied out */
203 intf
->rsp_cons
+= avail
;
205 /* Implies mb(): other side will see the updated consumer. */
206 if (intf
->rsp_prod
- cons
>= XENSTORE_RING_SIZE
)
207 notify_remote_via_evtchn(xen_store_evtchn
);
213 static int process_msg(void)
216 struct xsd_sockmsg msg
;
220 struct xs_watch_event
*watch
;
226 struct xb_req_data
*req
;
236 * We must disallow save/restore while reading a message.
237 * A partial read across s/r leaves us out of sync with
239 * xs_response_mutex is locked as long as we are processing one
240 * message. state.in_msg will be true as long as we are holding
243 mutex_lock(&xs_response_mutex
);
245 if (!xb_data_to_read()) {
246 /* We raced with save/restore: pending data 'gone'. */
247 mutex_unlock(&xs_response_mutex
);
248 state
.in_msg
= false;
254 if (state
.read
!= sizeof(state
.msg
)) {
255 err
= xb_read((void *)&state
.msg
+ state
.read
,
256 sizeof(state
.msg
) - state
.read
);
260 if (state
.read
!= sizeof(state
.msg
))
262 if (state
.msg
.len
> XENSTORE_PAYLOAD_MAX
) {
268 len
= state
.msg
.len
+ 1;
269 if (state
.msg
.type
== XS_WATCH_EVENT
)
270 len
+= sizeof(*state
.watch
);
272 state
.alloc
= kmalloc(len
, GFP_NOIO
| __GFP_HIGH
);
276 if (state
.msg
.type
== XS_WATCH_EVENT
)
277 state
.body
= state
.watch
->body
;
279 state
.body
= state
.alloc
;
280 state
.in_hdr
= false;
284 err
= xb_read(state
.body
+ state
.read
, state
.msg
.len
- state
.read
);
289 if (state
.read
!= state
.msg
.len
)
292 state
.body
[state
.msg
.len
] = '\0';
294 if (state
.msg
.type
== XS_WATCH_EVENT
) {
295 state
.watch
->len
= state
.msg
.len
;
296 err
= xs_watch_msg(state
.watch
);
299 mutex_lock(&xb_write_mutex
);
300 list_for_each_entry(req
, &xs_reply_list
, list
) {
301 if (req
->msg
.req_id
== state
.msg
.req_id
) {
302 list_del(&req
->list
);
307 mutex_unlock(&xb_write_mutex
);
311 if (req
->state
== xb_req_state_wait_reply
) {
312 req
->msg
.req_id
= req
->caller_req_id
;
313 req
->msg
.type
= state
.msg
.type
;
314 req
->msg
.len
= state
.msg
.len
;
315 req
->body
= state
.body
;
316 req
->state
= xb_req_state_got_reply
;
322 mutex_unlock(&xs_response_mutex
);
324 state
.in_msg
= false;
329 mutex_unlock(&xs_response_mutex
);
330 state
.in_msg
= false;
336 static int process_writes(void)
339 struct xb_req_data
*req
;
341 unsigned int written
;
347 if (!xb_data_to_write())
350 mutex_lock(&xb_write_mutex
);
353 state
.req
= list_first_entry(&xb_write_list
,
354 struct xb_req_data
, list
);
359 if (state
.req
->state
== xb_req_state_aborted
)
362 while (state
.idx
< state
.req
->num_vecs
) {
364 base
= &state
.req
->msg
;
365 len
= sizeof(state
.req
->msg
);
367 base
= state
.req
->vec
[state
.idx
].iov_base
;
368 len
= state
.req
->vec
[state
.idx
].iov_len
;
370 err
= xb_write(base
+ state
.written
, len
- state
.written
);
373 state
.written
+= err
;
374 if (state
.written
!= len
)
381 list_del(&state
.req
->list
);
382 state
.req
->state
= xb_req_state_wait_reply
;
383 list_add_tail(&state
.req
->list
, &xs_reply_list
);
387 mutex_unlock(&xb_write_mutex
);
392 state
.req
->msg
.type
= XS_ERROR
;
393 state
.req
->err
= err
;
394 list_del(&state
.req
->list
);
395 if (state
.req
->state
== xb_req_state_aborted
)
398 state
.req
->state
= xb_req_state_got_reply
;
399 wake_up(&state
.req
->wq
);
402 mutex_unlock(&xb_write_mutex
);
409 static int xb_thread_work(void)
411 return xb_data_to_read() || xb_data_to_write();
414 static int xenbus_thread(void *unused
)
418 while (!kthread_should_stop()) {
419 if (wait_event_interruptible(xb_waitq
, xb_thread_work()))
426 pr_warn_ratelimited("error %d while reading message\n",
429 err
= process_writes();
431 pr_warn_ratelimited("error %d while writing message\n",
440 * xb_init_comms - Set up interrupt handler off store event channel.
442 int xb_init_comms(void)
444 struct xenstore_domain_interface
*intf
= xen_store_interface
;
446 if (intf
->req_prod
!= intf
->req_cons
)
447 pr_err("request ring is not quiescent (%08x:%08x)!\n",
448 intf
->req_cons
, intf
->req_prod
);
450 if (intf
->rsp_prod
!= intf
->rsp_cons
) {
451 pr_warn("response ring is not quiescent (%08x:%08x): fixing up\n",
452 intf
->rsp_cons
, intf
->rsp_prod
);
455 intf
->rsp_cons
= intf
->rsp_prod
;
459 /* Already have an irq; assume we're resuming */
460 rebind_evtchn_irq(xen_store_evtchn
, xenbus_irq
);
464 err
= bind_evtchn_to_irqhandler(xen_store_evtchn
, wake_waiting
,
465 0, "xenbus", &xb_waitq
);
467 pr_err("request irq failed %i\n", err
);
474 xenbus_task
= kthread_run(xenbus_thread
, NULL
,
476 if (IS_ERR(xenbus_task
))
477 return PTR_ERR(xenbus_task
);
484 void xb_deinit_comms(void)
486 unbind_from_irqhandler(xenbus_irq
, &xb_waitq
);