2 * Xen hypercall batching.
4 * Xen allows multiple hypercalls to be issued at once, using the
5 * multicall interface. This allows the cost of trapping into the
6 * hypervisor to be amortized over several calls.
8 * This file implements a simple interface for multicalls. There's a
9 * per-cpu buffer of outstanding multicalls. When you want to queue a
10 * multicall for issuing, you can allocate a multicall slot for the
11 * call and its arguments, along with storage for space which is
12 * pointed to by the arguments (for passing pointers to structures,
13 * etc). When the multicall is actually issued, all the space for the
14 * commands and allocated memory is freed for reuse.
16 * Multicalls are flushed whenever any of the buffers get full, or
17 * when explicitly requested. There's no way to get per-multicall
18 * return results back. It will BUG if any of the multicalls fail.
20 * Jeremy Fitzhardinge <jeremy@xensource.com>, XenSource Inc, 2007
22 #include <linux/percpu.h>
23 #include <linux/hardirq.h>
25 #include <asm/xen/hypercall.h>
27 #include "multicalls.h"
32 #define MC_ARGS (MC_BATCH * 16)
35 struct multicall_entry entries
[MC_BATCH
];
37 struct multicall_entry debug
[MC_BATCH
];
39 unsigned char args
[MC_ARGS
];
43 } callbacks
[MC_BATCH
];
44 unsigned mcidx
, argidx
, cbidx
;
47 static DEFINE_PER_CPU(struct mc_buffer
, mc_buffer
);
48 DEFINE_PER_CPU(unsigned long, xen_mc_irq_flags
);
50 void xen_mc_flush(void)
52 struct mc_buffer
*b
= &__get_cpu_var(mc_buffer
);
57 BUG_ON(preemptible());
59 /* Disable interrupts in case someone comes in and queues
60 something in the middle */
61 local_irq_save(flags
);
65 memcpy(b
->debug
, b
->entries
,
66 b
->mcidx
* sizeof(struct multicall_entry
));
69 if (HYPERVISOR_multicall(b
->entries
, b
->mcidx
) != 0)
71 for (i
= 0; i
< b
->mcidx
; i
++)
72 if (b
->entries
[i
].result
< 0)
77 printk(KERN_ERR
"%d multicall(s) failed: cpu %d\n",
78 ret
, smp_processor_id());
80 for (i
= 0; i
< b
->mcidx
; i
++) {
81 printk(" call %2d/%d: op=%lu arg=[%lx] result=%ld\n",
85 b
->entries
[i
].result
);
93 BUG_ON(b
->argidx
!= 0);
95 local_irq_restore(flags
);
97 for (i
= 0; i
< b
->cbidx
; i
++) {
98 struct callback
*cb
= &b
->callbacks
[i
];
107 struct multicall_space
__xen_mc_entry(size_t args
)
109 struct mc_buffer
*b
= &__get_cpu_var(mc_buffer
);
110 struct multicall_space ret
;
111 unsigned argidx
= roundup(b
->argidx
, sizeof(u64
));
113 BUG_ON(preemptible());
114 BUG_ON(b
->argidx
> MC_ARGS
);
116 if (b
->mcidx
== MC_BATCH
||
117 (argidx
+ args
) > MC_ARGS
) {
119 argidx
= roundup(b
->argidx
, sizeof(u64
));
122 ret
.mc
= &b
->entries
[b
->mcidx
];
124 ret
.args
= &b
->args
[argidx
];
125 b
->argidx
= argidx
+ args
;
127 BUG_ON(b
->argidx
> MC_ARGS
);
131 struct multicall_space
xen_mc_extend_args(unsigned long op
, size_t size
)
133 struct mc_buffer
*b
= &__get_cpu_var(mc_buffer
);
134 struct multicall_space ret
= { NULL
, NULL
};
136 BUG_ON(preemptible());
137 BUG_ON(b
->argidx
> MC_ARGS
);
142 if (b
->entries
[b
->mcidx
- 1].op
!= op
)
145 if ((b
->argidx
+ size
) > MC_ARGS
)
148 ret
.mc
= &b
->entries
[b
->mcidx
- 1];
149 ret
.args
= &b
->args
[b
->argidx
];
152 BUG_ON(b
->argidx
> MC_ARGS
);
156 void xen_mc_callback(void (*fn
)(void *), void *data
)
158 struct mc_buffer
*b
= &__get_cpu_var(mc_buffer
);
161 if (b
->cbidx
== MC_BATCH
)
164 cb
= &b
->callbacks
[b
->cbidx
++];