2 * Xen event channels (FIFO-based ABI)
4 * Copyright (C) 2013 Citrix Systems R&D ltd.
6 * This source code is free software; you can redistribute it and/or
7 * modify it under the terms of the GNU General Public License as
8 * published by the Free Software Foundation; either version 2 of the
9 * License, or (at your option) any later version.
11 * Or, when distributed separately from the Linux kernel or
12 * incorporated into other software packages, subject to the following
15 * Permission is hereby granted, free of charge, to any person obtaining a copy
16 * of this source file (the "Software"), to deal in the Software without
17 * restriction, including without limitation the rights to use, copy, modify,
18 * merge, publish, distribute, sublicense, and/or sell copies of the Software,
19 * and to permit persons to whom the Software is furnished to do so, subject to
20 * the following conditions:
22 * The above copyright notice and this permission notice shall be included in
23 * all copies or substantial portions of the Software.
25 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
26 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
27 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
28 * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
29 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
30 * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
34 #define pr_fmt(fmt) "xen:" KBUILD_MODNAME ": " fmt
36 #include <linux/linkage.h>
37 #include <linux/interrupt.h>
38 #include <linux/irq.h>
39 #include <linux/smp.h>
40 #include <linux/percpu.h>
41 #include <linux/cpu.h>
43 #include <asm/barrier.h>
44 #include <asm/sync_bitops.h>
45 #include <asm/xen/hypercall.h>
46 #include <asm/xen/hypervisor.h>
49 #include <xen/xen-ops.h>
50 #include <xen/events.h>
51 #include <xen/interface/xen.h>
52 #include <xen/interface/event_channel.h>
55 #include "events_internal.h"
57 #define EVENT_WORDS_PER_PAGE (XEN_PAGE_SIZE / sizeof(event_word_t))
58 #define MAX_EVENT_ARRAY_PAGES (EVTCHN_FIFO_NR_CHANNELS / EVENT_WORDS_PER_PAGE)
60 struct evtchn_fifo_queue
{
61 uint32_t head
[EVTCHN_FIFO_MAX_QUEUES
];
64 static DEFINE_PER_CPU(struct evtchn_fifo_control_block
*, cpu_control_block
);
65 static DEFINE_PER_CPU(struct evtchn_fifo_queue
, cpu_queue
);
66 static event_word_t
*event_array
[MAX_EVENT_ARRAY_PAGES
] __read_mostly
;
67 static unsigned event_array_pages __read_mostly
;
70 * sync_set_bit() and friends must be unsigned long aligned.
72 #if BITS_PER_LONG > 32
74 #define BM(w) (unsigned long *)((unsigned long)w & ~0x7UL)
75 #define EVTCHN_FIFO_BIT(b, w) \
76 (((unsigned long)w & 0x4UL) ? (EVTCHN_FIFO_ ##b + 32) : EVTCHN_FIFO_ ##b)
80 #define BM(w) ((unsigned long *)(w))
81 #define EVTCHN_FIFO_BIT(b, w) EVTCHN_FIFO_ ##b
85 static inline event_word_t
*event_word_from_port(unsigned port
)
87 unsigned i
= port
/ EVENT_WORDS_PER_PAGE
;
89 return event_array
[i
] + port
% EVENT_WORDS_PER_PAGE
;
92 static unsigned evtchn_fifo_max_channels(void)
94 return EVTCHN_FIFO_NR_CHANNELS
;
97 static unsigned evtchn_fifo_nr_channels(void)
99 return event_array_pages
* EVENT_WORDS_PER_PAGE
;
102 static int init_control_block(int cpu
,
103 struct evtchn_fifo_control_block
*control_block
)
105 struct evtchn_fifo_queue
*q
= &per_cpu(cpu_queue
, cpu
);
106 struct evtchn_init_control init_control
;
109 /* Reset the control block and the local HEADs. */
110 clear_page(control_block
);
111 for (i
= 0; i
< EVTCHN_FIFO_MAX_QUEUES
; i
++)
114 init_control
.control_gfn
= virt_to_gfn(control_block
);
115 init_control
.offset
= 0;
116 init_control
.vcpu
= xen_vcpu_nr(cpu
);
118 return HYPERVISOR_event_channel_op(EVTCHNOP_init_control
, &init_control
);
121 static void free_unused_array_pages(void)
125 for (i
= event_array_pages
; i
< MAX_EVENT_ARRAY_PAGES
; i
++) {
128 free_page((unsigned long)event_array
[i
]);
129 event_array
[i
] = NULL
;
133 static void init_array_page(event_word_t
*array_page
)
137 for (i
= 0; i
< EVENT_WORDS_PER_PAGE
; i
++)
138 array_page
[i
] = 1 << EVTCHN_FIFO_MASKED
;
141 static int evtchn_fifo_setup(struct irq_info
*info
)
143 unsigned port
= info
->evtchn
;
144 unsigned new_array_pages
;
147 new_array_pages
= port
/ EVENT_WORDS_PER_PAGE
+ 1;
149 if (new_array_pages
> MAX_EVENT_ARRAY_PAGES
)
152 while (event_array_pages
< new_array_pages
) {
154 struct evtchn_expand_array expand_array
;
156 /* Might already have a page if we've resumed. */
157 array_page
= event_array
[event_array_pages
];
159 array_page
= (void *)__get_free_page(GFP_KERNEL
);
160 if (array_page
== NULL
) {
164 event_array
[event_array_pages
] = array_page
;
167 /* Mask all events in this page before adding it. */
168 init_array_page(array_page
);
170 expand_array
.array_gfn
= virt_to_gfn(array_page
);
172 ret
= HYPERVISOR_event_channel_op(EVTCHNOP_expand_array
, &expand_array
);
181 if (event_array_pages
== 0)
182 panic("xen: unable to expand event array with initial page (%d)\n", ret
);
184 pr_err("unable to expand event array (%d)\n", ret
);
185 free_unused_array_pages();
189 static void evtchn_fifo_bind_to_cpu(struct irq_info
*info
, unsigned cpu
)
194 static void evtchn_fifo_clear_pending(unsigned port
)
196 event_word_t
*word
= event_word_from_port(port
);
197 sync_clear_bit(EVTCHN_FIFO_BIT(PENDING
, word
), BM(word
));
200 static void evtchn_fifo_set_pending(unsigned port
)
202 event_word_t
*word
= event_word_from_port(port
);
203 sync_set_bit(EVTCHN_FIFO_BIT(PENDING
, word
), BM(word
));
206 static bool evtchn_fifo_is_pending(unsigned port
)
208 event_word_t
*word
= event_word_from_port(port
);
209 return sync_test_bit(EVTCHN_FIFO_BIT(PENDING
, word
), BM(word
));
212 static bool evtchn_fifo_test_and_set_mask(unsigned port
)
214 event_word_t
*word
= event_word_from_port(port
);
215 return sync_test_and_set_bit(EVTCHN_FIFO_BIT(MASKED
, word
), BM(word
));
218 static void evtchn_fifo_mask(unsigned port
)
220 event_word_t
*word
= event_word_from_port(port
);
221 sync_set_bit(EVTCHN_FIFO_BIT(MASKED
, word
), BM(word
));
224 static bool evtchn_fifo_is_masked(unsigned port
)
226 event_word_t
*word
= event_word_from_port(port
);
227 return sync_test_bit(EVTCHN_FIFO_BIT(MASKED
, word
), BM(word
));
230 * Clear MASKED, spinning if BUSY is set.
232 static void clear_masked(volatile event_word_t
*word
)
234 event_word_t
new, old
, w
;
239 old
= w
& ~(1 << EVTCHN_FIFO_BUSY
);
240 new = old
& ~(1 << EVTCHN_FIFO_MASKED
);
241 w
= sync_cmpxchg(word
, old
, new);
245 static void evtchn_fifo_unmask(unsigned port
)
247 event_word_t
*word
= event_word_from_port(port
);
249 BUG_ON(!irqs_disabled());
252 if (evtchn_fifo_is_pending(port
)) {
253 struct evtchn_unmask unmask
= { .port
= port
};
254 (void)HYPERVISOR_event_channel_op(EVTCHNOP_unmask
, &unmask
);
258 static uint32_t clear_linked(volatile event_word_t
*word
)
260 event_word_t
new, old
, w
;
266 new = (w
& ~((1 << EVTCHN_FIFO_LINKED
)
267 | EVTCHN_FIFO_LINK_MASK
));
268 } while ((w
= sync_cmpxchg(word
, old
, new)) != old
);
270 return w
& EVTCHN_FIFO_LINK_MASK
;
273 static void handle_irq_for_port(unsigned port
)
277 irq
= get_evtchn_to_irq(port
);
279 generic_handle_irq(irq
);
282 static void consume_one_event(unsigned cpu
,
283 struct evtchn_fifo_control_block
*control_block
,
284 unsigned priority
, unsigned long *ready
,
287 struct evtchn_fifo_queue
*q
= &per_cpu(cpu_queue
, cpu
);
292 head
= q
->head
[priority
];
295 * Reached the tail last time? Read the new HEAD from the
299 virt_rmb(); /* Ensure word is up-to-date before reading head. */
300 head
= control_block
->head
[priority
];
304 word
= event_word_from_port(port
);
305 head
= clear_linked(word
);
308 * If the link is non-zero, there are more events in the
309 * queue, otherwise the queue is empty.
311 * If the queue is empty, clear this priority from our local
312 * copy of the ready word.
315 clear_bit(priority
, ready
);
317 if (evtchn_fifo_is_pending(port
) && !evtchn_fifo_is_masked(port
)) {
319 pr_warn("Dropping pending event for port %u\n", port
);
321 handle_irq_for_port(port
);
324 q
->head
[priority
] = head
;
327 static void __evtchn_fifo_handle_events(unsigned cpu
, bool drop
)
329 struct evtchn_fifo_control_block
*control_block
;
333 control_block
= per_cpu(cpu_control_block
, cpu
);
335 ready
= xchg(&control_block
->ready
, 0);
338 q
= find_first_bit(&ready
, EVTCHN_FIFO_MAX_QUEUES
);
339 consume_one_event(cpu
, control_block
, q
, &ready
, drop
);
340 ready
|= xchg(&control_block
->ready
, 0);
344 static void evtchn_fifo_handle_events(unsigned cpu
)
346 __evtchn_fifo_handle_events(cpu
, false);
349 static void evtchn_fifo_resume(void)
353 for_each_possible_cpu(cpu
) {
354 void *control_block
= per_cpu(cpu_control_block
, cpu
);
361 * If this CPU is offline, take the opportunity to
362 * free the control block while it is not being
365 if (!cpu_online(cpu
)) {
366 free_page((unsigned long)control_block
);
367 per_cpu(cpu_control_block
, cpu
) = NULL
;
371 ret
= init_control_block(cpu
, control_block
);
377 * The event array starts out as empty again and is extended
378 * as normal when events are bound. The existing pages will
381 event_array_pages
= 0;
384 static const struct evtchn_ops evtchn_ops_fifo
= {
385 .max_channels
= evtchn_fifo_max_channels
,
386 .nr_channels
= evtchn_fifo_nr_channels
,
387 .setup
= evtchn_fifo_setup
,
388 .bind_to_cpu
= evtchn_fifo_bind_to_cpu
,
389 .clear_pending
= evtchn_fifo_clear_pending
,
390 .set_pending
= evtchn_fifo_set_pending
,
391 .is_pending
= evtchn_fifo_is_pending
,
392 .test_and_set_mask
= evtchn_fifo_test_and_set_mask
,
393 .mask
= evtchn_fifo_mask
,
394 .unmask
= evtchn_fifo_unmask
,
395 .handle_events
= evtchn_fifo_handle_events
,
396 .resume
= evtchn_fifo_resume
,
399 static int evtchn_fifo_alloc_control_block(unsigned cpu
)
401 void *control_block
= NULL
;
404 control_block
= (void *)__get_free_page(GFP_KERNEL
);
405 if (control_block
== NULL
)
408 ret
= init_control_block(cpu
, control_block
);
412 per_cpu(cpu_control_block
, cpu
) = control_block
;
417 free_page((unsigned long)control_block
);
421 static int evtchn_fifo_cpu_notification(struct notifier_block
*self
,
422 unsigned long action
,
425 int cpu
= (long)hcpu
;
430 if (!per_cpu(cpu_control_block
, cpu
))
431 ret
= evtchn_fifo_alloc_control_block(cpu
);
434 __evtchn_fifo_handle_events(cpu
, true);
439 return ret
< 0 ? NOTIFY_BAD
: NOTIFY_OK
;
442 static struct notifier_block evtchn_fifo_cpu_notifier
= {
443 .notifier_call
= evtchn_fifo_cpu_notification
,
446 int __init
xen_evtchn_fifo_init(void)
451 ret
= evtchn_fifo_alloc_control_block(cpu
);
455 pr_info("Using FIFO-based ABI\n");
457 evtchn_ops
= &evtchn_ops_fifo
;
459 register_cpu_notifier(&evtchn_fifo_cpu_notifier
);