1 /*P:200 This contains all the /dev/lguest code, whereby the userspace launcher
2 * controls and communicates with the Guest. For example, the first write will
3 * tell us the Guest's memory layout and entry point. A read will run the
4 * Guest until something happens, such as a signal or the Guest doing a NOTIFY
7 #include <linux/uaccess.h>
8 #include <linux/miscdevice.h>
10 #include <linux/sched.h>
11 #include <linux/eventfd.h>
12 #include <linux/file.h>
16 * Before we move on, let's jump ahead and look at what the kernel does when
17 * it needs to look up the eventfds. That will complete our picture of how we
20 * The notification value is in cpu->pending_notify: we return true if it went
23 bool send_notify_to_eventfd(struct lg_cpu
*cpu
)
26 struct lg_eventfd_map
*map
;
29 * This "rcu_read_lock()" helps track when someone is still looking at
30 * the (RCU-using) eventfds array. It's not actually a lock at all;
31 * indeed it's a noop in many configurations. (You didn't expect me to
32 * explain all the RCU secrets here, did you?)
36 * rcu_dereference is the counter-side of rcu_assign_pointer(); it
37 * makes sure we don't access the memory pointed to by
38 * cpu->lg->eventfds before cpu->lg->eventfds is set. Sounds crazy,
39 * but Alpha allows this! Paul McKenney points out that a really
40 * aggressive compiler could have the same effect:
41 * http://lists.ozlabs.org/pipermail/lguest/2009-July/001560.html
43 * So play safe, use rcu_dereference to get the rcu-protected pointer:
45 map
= rcu_dereference(cpu
->lg
->eventfds
);
47 * Simple array search: even if they add an eventfd while we do this,
48 * we'll continue to use the old array and just won't see the new one.
50 for (i
= 0; i
< map
->num
; i
++) {
51 if (map
->map
[i
].addr
== cpu
->pending_notify
) {
52 eventfd_signal(map
->map
[i
].event
, 1);
53 cpu
->pending_notify
= 0;
57 /* We're done with the rcu-protected variable cpu->lg->eventfds. */
60 /* If we cleared the notification, it's because we found a match. */
61 return cpu
->pending_notify
== 0;
65 * One of the more tricksy tricks in the Linux Kernel is a technique called
66 * Read Copy Update. Since one point of lguest is to teach lguest journeyers
67 * about kernel coding, I use it here. (In case you're curious, other purposes
68 * include learning about virtualization and instilling a deep appreciation for
69 * simplicity and puppies).
71 * We keep a simple array which maps LHCALL_NOTIFY values to eventfds, but we
72 * add new eventfds without ever blocking readers from accessing the array.
73 * The current Launcher only does this during boot, so that never happens. But
74 * Read Copy Update is cool, and adding a lock risks damaging even more puppies
75 * than this code does.
77 * We allocate a brand new one-larger array, copy the old one and add our new
78 * element. Then we make the lg eventfd pointer point to the new array.
79 * That's the easy part: now we need to free the old one, but we need to make
80 * sure no slow CPU somewhere is still looking at it. That's what
81 * synchronize_rcu does for us: waits until every CPU has indicated that it has
82 * moved on to know it's no longer using the old one.
84 * If that's unclear, see http://en.wikipedia.org/wiki/Read-copy-update.
86 static int add_eventfd(struct lguest
*lg
, unsigned long addr
, int fd
)
88 struct lg_eventfd_map
*new, *old
= lg
->eventfds
;
91 * We don't allow notifications on value 0 anyway (pending_notify of
92 * 0 means "nothing pending").
98 * Replace the old array with the new one, carefully: others can
99 * be accessing it at the same time.
101 new = kmalloc(sizeof(*new) + sizeof(new->map
[0]) * (old
->num
+ 1),
106 /* First make identical copy. */
107 memcpy(new->map
, old
->map
, sizeof(old
->map
[0]) * old
->num
);
110 /* Now append new entry. */
111 new->map
[new->num
].addr
= addr
;
112 new->map
[new->num
].event
= eventfd_ctx_fdget(fd
);
113 if (IS_ERR(new->map
[new->num
].event
)) {
114 int err
= PTR_ERR(new->map
[new->num
].event
);
121 * Now put new one in place: rcu_assign_pointer() is a fancy way of
122 * doing "lg->eventfds = new", but it uses memory barriers to make
123 * absolutely sure that the contents of "new" written above is nailed
124 * down before we actually do the assignment.
126 * We have to think about these kinds of things when we're operating on
127 * live data without locks.
129 rcu_assign_pointer(lg
->eventfds
, new);
132 * We're not in a big hurry. Wait until noone's looking at old
133 * version, then free it.
142 * Receiving notifications from the Guest is usually done by attaching a
143 * particular LHCALL_NOTIFY value to an event filedescriptor. The eventfd will
144 * become readable when the Guest does an LHCALL_NOTIFY with that value.
146 * This is really convenient for processing each virtqueue in a separate
149 static int attach_eventfd(struct lguest
*lg
, const unsigned long __user
*input
)
151 unsigned long addr
, fd
;
154 if (get_user(addr
, input
) != 0)
157 if (get_user(fd
, input
) != 0)
161 * Just make sure two callers don't add eventfds at once. We really
162 * only need to lock against callers adding to the same Guest, so using
163 * the Big Lguest Lock is overkill. But this is setup, not a fast path.
165 mutex_lock(&lguest_lock
);
166 err
= add_eventfd(lg
, addr
, fd
);
167 mutex_unlock(&lguest_lock
);
173 * Sending an interrupt is done by writing LHREQ_IRQ and an interrupt
174 * number to /dev/lguest.
176 static int user_send_irq(struct lg_cpu
*cpu
, const unsigned long __user
*input
)
180 if (get_user(irq
, input
) != 0)
182 if (irq
>= LGUEST_IRQS
)
186 * Next time the Guest runs, the core code will see if it can deliver
189 set_interrupt(cpu
, irq
);
194 * Once our Guest is initialized, the Launcher makes it run by reading
197 static ssize_t
read(struct file
*file
, char __user
*user
, size_t size
,loff_t
*o
)
199 struct lguest
*lg
= file
->private_data
;
201 unsigned int cpu_id
= *o
;
203 /* You must write LHREQ_INITIALIZE first! */
207 /* Watch out for arbitrary vcpu indexes! */
208 if (cpu_id
>= lg
->nr_cpus
)
211 cpu
= &lg
->cpus
[cpu_id
];
213 /* If you're not the task which owns the Guest, go away. */
214 if (current
!= cpu
->tsk
)
217 /* If the Guest is already dead, we indicate why */
221 /* lg->dead either contains an error code, or a string. */
222 if (IS_ERR(lg
->dead
))
223 return PTR_ERR(lg
->dead
);
225 /* We can only return as much as the buffer they read with. */
226 len
= min(size
, strlen(lg
->dead
)+1);
227 if (copy_to_user(user
, lg
->dead
, len
) != 0)
233 * If we returned from read() last time because the Guest sent I/O,
236 if (cpu
->pending_notify
)
237 cpu
->pending_notify
= 0;
239 /* Run the Guest until something interesting happens. */
240 return run_guest(cpu
, (unsigned long __user
*)user
);
244 * This actually initializes a CPU. For the moment, a Guest is only
245 * uniprocessor, so "id" is always 0.
247 static int lg_cpu_start(struct lg_cpu
*cpu
, unsigned id
, unsigned long start_ip
)
249 /* We have a limited number the number of CPUs in the lguest struct. */
250 if (id
>= ARRAY_SIZE(cpu
->lg
->cpus
))
253 /* Set up this CPU's id, and pointer back to the lguest struct. */
255 cpu
->lg
= container_of((cpu
- id
), struct lguest
, cpus
[0]);
258 /* Each CPU has a timer it can set. */
262 * We need a complete page for the Guest registers: they are accessible
263 * to the Guest and we can only grant it access to whole pages.
265 cpu
->regs_page
= get_zeroed_page(GFP_KERNEL
);
269 /* We actually put the registers at the bottom of the page. */
270 cpu
->regs
= (void *)cpu
->regs_page
+ PAGE_SIZE
- sizeof(*cpu
->regs
);
273 * Now we initialize the Guest's registers, handing it the start
276 lguest_arch_setup_regs(cpu
, start_ip
);
279 * We keep a pointer to the Launcher task (ie. current task) for when
280 * other Guests want to wake this one (eg. console input).
285 * We need to keep a pointer to the Launcher's memory map, because if
286 * the Launcher dies we need to clean it up. If we don't keep a
287 * reference, it is destroyed before close() is called.
289 cpu
->mm
= get_task_mm(cpu
->tsk
);
292 * We remember which CPU's pages this Guest used last, for optimization
293 * when the same Guest runs on the same CPU twice.
295 cpu
->last_pages
= NULL
;
297 /* No error == success. */
302 * The initialization write supplies 3 pointer sized (32 or 64 bit) values (in
303 * addition to the LHREQ_INITIALIZE value). These are:
305 * base: The start of the Guest-physical memory inside the Launcher memory.
307 * pfnlimit: The highest (Guest-physical) page number the Guest should be
308 * allowed to access. The Guest memory lives inside the Launcher, so it sets
309 * this to ensure the Guest can only reach its own memory.
311 * start: The first instruction to execute ("eip" in x86-speak).
313 static int initialize(struct file
*file
, const unsigned long __user
*input
)
315 /* "struct lguest" contains all we (the Host) know about a Guest. */
318 unsigned long args
[3];
321 * We grab the Big Lguest lock, which protects against multiple
322 * simultaneous initializations.
324 mutex_lock(&lguest_lock
);
325 /* You can't initialize twice! Close the device and start again... */
326 if (file
->private_data
) {
331 if (copy_from_user(args
, input
, sizeof(args
)) != 0) {
336 lg
= kzalloc(sizeof(*lg
), GFP_KERNEL
);
342 lg
->eventfds
= kmalloc(sizeof(*lg
->eventfds
), GFP_KERNEL
);
347 lg
->eventfds
->num
= 0;
349 /* Populate the easy fields of our "struct lguest" */
350 lg
->mem_base
= (void __user
*)args
[0];
351 lg
->pfn_limit
= args
[1];
353 /* This is the first cpu (cpu 0) and it will start booting at args[2] */
354 err
= lg_cpu_start(&lg
->cpus
[0], 0, args
[2]);
359 * Initialize the Guest's shadow page tables, using the toplevel
360 * address the Launcher gave us. This allocates memory, so can fail.
362 err
= init_guest_pagetable(lg
);
366 /* We keep our "struct lguest" in the file's private_data. */
367 file
->private_data
= lg
;
369 mutex_unlock(&lguest_lock
);
371 /* And because this is a write() call, we return the length used. */
375 /* FIXME: This should be in free_vcpu */
376 free_page(lg
->cpus
[0].regs_page
);
382 mutex_unlock(&lguest_lock
);
387 * The first operation the Launcher does must be a write. All writes
388 * start with an unsigned long number: for the first write this must be
389 * LHREQ_INITIALIZE to set up the Guest. After that the Launcher can use
390 * writes of other values to send interrupts or set up receipt of notifications.
392 * Note that we overload the "offset" in the /dev/lguest file to indicate what
393 * CPU number we're dealing with. Currently this is always 0 since we only
394 * support uniprocessor Guests, but you can see the beginnings of SMP support
397 static ssize_t
write(struct file
*file
, const char __user
*in
,
398 size_t size
, loff_t
*off
)
401 * Once the Guest is initialized, we hold the "struct lguest" in the
404 struct lguest
*lg
= file
->private_data
;
405 const unsigned long __user
*input
= (const unsigned long __user
*)in
;
407 struct lg_cpu
*uninitialized_var(cpu
);
408 unsigned int cpu_id
= *off
;
410 /* The first value tells us what this request is. */
411 if (get_user(req
, input
) != 0)
415 /* If you haven't initialized, you must do that first. */
416 if (req
!= LHREQ_INITIALIZE
) {
417 if (!lg
|| (cpu_id
>= lg
->nr_cpus
))
419 cpu
= &lg
->cpus
[cpu_id
];
421 /* Once the Guest is dead, you can only read() why it died. */
427 case LHREQ_INITIALIZE
:
428 return initialize(file
, input
);
430 return user_send_irq(cpu
, input
);
432 return attach_eventfd(lg
, input
);
439 * The final piece of interface code is the close() routine. It reverses
440 * everything done in initialize(). This is usually called because the
443 * Note that the close routine returns 0 or a negative error number: it can't
444 * really fail, but it can whine. I blame Sun for this wart, and K&R C for
445 * letting them do it.
447 static int close(struct inode
*inode
, struct file
*file
)
449 struct lguest
*lg
= file
->private_data
;
452 /* If we never successfully initialized, there's nothing to clean up */
457 * We need the big lock, to protect from inter-guest I/O and other
458 * Launchers initializing guests.
460 mutex_lock(&lguest_lock
);
462 /* Free up the shadow page tables for the Guest. */
463 free_guest_pagetable(lg
);
465 for (i
= 0; i
< lg
->nr_cpus
; i
++) {
466 /* Cancels the hrtimer set via LHCALL_SET_CLOCKEVENT. */
467 hrtimer_cancel(&lg
->cpus
[i
].hrt
);
468 /* We can free up the register page we allocated. */
469 free_page(lg
->cpus
[i
].regs_page
);
471 * Now all the memory cleanups are done, it's safe to release
472 * the Launcher's memory management structure.
474 mmput(lg
->cpus
[i
].mm
);
477 /* Release any eventfds they registered. */
478 for (i
= 0; i
< lg
->eventfds
->num
; i
++)
479 eventfd_ctx_put(lg
->eventfds
->map
[i
].event
);
483 * If lg->dead doesn't contain an error code it will be NULL or a
484 * kmalloc()ed string, either of which is ok to hand to kfree().
486 if (!IS_ERR(lg
->dead
))
488 /* Free the memory allocated to the lguest_struct */
490 /* Release lock and exit. */
491 mutex_unlock(&lguest_lock
);
497 * Welcome to our journey through the Launcher!
499 * The Launcher is the Host userspace program which sets up, runs and services
500 * the Guest. In fact, many comments in the Drivers which refer to "the Host"
501 * doing things are inaccurate: the Launcher does all the device handling for
502 * the Guest, but the Guest can't know that.
504 * Just to confuse you: to the Host kernel, the Launcher *is* the Guest and we
505 * shall see more of that later.
507 * We begin our understanding with the Host kernel interface which the Launcher
508 * uses: reading and writing a character device called /dev/lguest. All the
509 * work happens in the read(), write() and close() routines:
511 static const struct file_operations lguest_fops
= {
512 .owner
= THIS_MODULE
,
519 * This is a textbook example of a "misc" character device. Populate a "struct
520 * miscdevice" and register it with misc_register().
522 static struct miscdevice lguest_dev
= {
523 .minor
= MISC_DYNAMIC_MINOR
,
525 .fops
= &lguest_fops
,
528 int __init
lguest_device_init(void)
530 return misc_register(&lguest_dev
);
533 void __exit
lguest_device_remove(void)
535 misc_deregister(&lguest_dev
);