1 /******************************************************************************
3 * Driver for receiving and transferring machine check error infomation
5 * Copyright (c) 2012 Intel Corporation
6 * Author: Liu, Jinsong <jinsong.liu@intel.com>
7 * Author: Jiang, Yunhong <yunhong.jiang@intel.com>
8 * Author: Ke, Liping <liping.ke@intel.com>
10 * This program is free software; you can redistribute it and/or
11 * modify it under the terms of the GNU General Public License version 2
12 * as published by the Free Software Foundation; or, when distributed
13 * separately from the Linux kernel or incorporated into other
14 * software packages, subject to the following license:
16 * Permission is hereby granted, free of charge, to any person obtaining a copy
17 * of this source file (the "Software"), to deal in the Software without
18 * restriction, including without limitation the rights to use, copy, modify,
19 * merge, publish, distribute, sublicense, and/or sell copies of the Software,
20 * and to permit persons to whom the Software is furnished to do so, subject to
21 * the following conditions:
23 * The above copyright notice and this permission notice shall be included in
24 * all copies or substantial portions of the Software.
26 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
27 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
28 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
29 * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
30 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
31 * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
35 #define pr_fmt(fmt) "xen_mcelog: " fmt
37 #include <linux/init.h>
38 #include <linux/types.h>
39 #include <linux/kernel.h>
40 #include <linux/slab.h>
42 #include <linux/device.h>
43 #include <linux/miscdevice.h>
44 #include <linux/uaccess.h>
45 #include <linux/capability.h>
46 #include <linux/poll.h>
47 #include <linux/sched.h>
49 #include <xen/interface/xen.h>
50 #include <xen/events.h>
51 #include <xen/interface/vcpu.h>
53 #include <asm/xen/hypercall.h>
54 #include <asm/xen/hypervisor.h>
56 static struct mc_info g_mi
;
57 static struct mcinfo_logical_cpu
*g_physinfo
;
58 static uint32_t ncpus
;
60 static DEFINE_MUTEX(mcelog_lock
);
62 static struct xen_mce_log xen_mcelog
= {
63 .signature
= XEN_MCE_LOG_SIGNATURE
,
64 .len
= XEN_MCE_LOG_LEN
,
65 .recordlen
= sizeof(struct xen_mce
),
68 static DEFINE_SPINLOCK(xen_mce_chrdev_state_lock
);
69 static int xen_mce_chrdev_open_count
; /* #times opened */
70 static int xen_mce_chrdev_open_exclu
; /* already open exclusive? */
72 static DECLARE_WAIT_QUEUE_HEAD(xen_mce_chrdev_wait
);
74 static int xen_mce_chrdev_open(struct inode
*inode
, struct file
*file
)
76 spin_lock(&xen_mce_chrdev_state_lock
);
78 if (xen_mce_chrdev_open_exclu
||
79 (xen_mce_chrdev_open_count
&& (file
->f_flags
& O_EXCL
))) {
80 spin_unlock(&xen_mce_chrdev_state_lock
);
85 if (file
->f_flags
& O_EXCL
)
86 xen_mce_chrdev_open_exclu
= 1;
87 xen_mce_chrdev_open_count
++;
89 spin_unlock(&xen_mce_chrdev_state_lock
);
91 return nonseekable_open(inode
, file
);
94 static int xen_mce_chrdev_release(struct inode
*inode
, struct file
*file
)
96 spin_lock(&xen_mce_chrdev_state_lock
);
98 xen_mce_chrdev_open_count
--;
99 xen_mce_chrdev_open_exclu
= 0;
101 spin_unlock(&xen_mce_chrdev_state_lock
);
106 static ssize_t
xen_mce_chrdev_read(struct file
*filp
, char __user
*ubuf
,
107 size_t usize
, loff_t
*off
)
109 char __user
*buf
= ubuf
;
113 mutex_lock(&mcelog_lock
);
115 num
= xen_mcelog
.next
;
117 /* Only supports full reads right now */
119 if (*off
!= 0 || usize
< XEN_MCE_LOG_LEN
*sizeof(struct xen_mce
))
123 for (i
= 0; i
< num
; i
++) {
124 struct xen_mce
*m
= &xen_mcelog
.entry
[i
];
126 err
|= copy_to_user(buf
, m
, sizeof(*m
));
130 memset(xen_mcelog
.entry
, 0, num
* sizeof(struct xen_mce
));
137 mutex_unlock(&mcelog_lock
);
139 return err
? err
: buf
- ubuf
;
142 static unsigned int xen_mce_chrdev_poll(struct file
*file
, poll_table
*wait
)
144 poll_wait(file
, &xen_mce_chrdev_wait
, wait
);
147 return POLLIN
| POLLRDNORM
;
152 static long xen_mce_chrdev_ioctl(struct file
*f
, unsigned int cmd
,
155 int __user
*p
= (int __user
*)arg
;
157 if (!capable(CAP_SYS_ADMIN
))
161 case MCE_GET_RECORD_LEN
:
162 return put_user(sizeof(struct xen_mce
), p
);
163 case MCE_GET_LOG_LEN
:
164 return put_user(XEN_MCE_LOG_LEN
, p
);
165 case MCE_GETCLEAR_FLAGS
: {
169 flags
= xen_mcelog
.flags
;
170 } while (cmpxchg(&xen_mcelog
.flags
, flags
, 0) != flags
);
172 return put_user(flags
, p
);
179 static const struct file_operations xen_mce_chrdev_ops
= {
180 .open
= xen_mce_chrdev_open
,
181 .release
= xen_mce_chrdev_release
,
182 .read
= xen_mce_chrdev_read
,
183 .poll
= xen_mce_chrdev_poll
,
184 .unlocked_ioctl
= xen_mce_chrdev_ioctl
,
188 static struct miscdevice xen_mce_chrdev_device
= {
195 * Caller should hold the mcelog_lock
197 static void xen_mce_log(struct xen_mce
*mce
)
201 entry
= xen_mcelog
.next
;
204 * When the buffer fills up discard new entries.
205 * Assume that the earlier errors are the more
208 if (entry
>= XEN_MCE_LOG_LEN
) {
209 set_bit(XEN_MCE_OVERFLOW
,
210 (unsigned long *)&xen_mcelog
.flags
);
214 memcpy(xen_mcelog
.entry
+ entry
, mce
, sizeof(struct xen_mce
));
219 static int convert_log(struct mc_info
*mi
)
221 struct mcinfo_common
*mic
;
222 struct mcinfo_global
*mc_global
;
223 struct mcinfo_bank
*mc_bank
;
228 x86_mcinfo_lookup(&mic
, mi
, MC_TYPE_GLOBAL
);
229 if (unlikely(!mic
)) {
230 pr_warn("Failed to find global error info\n");
234 memset(&m
, 0, sizeof(struct xen_mce
));
236 mc_global
= (struct mcinfo_global
*)mic
;
237 m
.mcgstatus
= mc_global
->mc_gstatus
;
238 m
.apicid
= mc_global
->mc_apicid
;
240 for (i
= 0; i
< ncpus
; i
++)
241 if (g_physinfo
[i
].mc_apicid
== m
.apicid
)
243 if (unlikely(i
== ncpus
)) {
244 pr_warn("Failed to match cpu with apicid %d\n", m
.apicid
);
248 m
.socketid
= g_physinfo
[i
].mc_chipid
;
249 m
.cpu
= m
.extcpu
= g_physinfo
[i
].mc_cpunr
;
250 m
.cpuvendor
= (__u8
)g_physinfo
[i
].mc_vendor
;
251 m
.mcgcap
= g_physinfo
[i
].mc_msrvalues
[__MC_MSR_MCGCAP
].value
;
254 x86_mcinfo_lookup(&mic
, mi
, MC_TYPE_BANK
);
255 if (unlikely(!mic
)) {
256 pr_warn("Fail to find bank error info\n");
261 if ((!mic
) || (mic
->size
== 0) ||
262 (mic
->type
!= MC_TYPE_GLOBAL
&&
263 mic
->type
!= MC_TYPE_BANK
&&
264 mic
->type
!= MC_TYPE_EXTENDED
&&
265 mic
->type
!= MC_TYPE_RECOVERY
))
268 if (mic
->type
== MC_TYPE_BANK
) {
269 mc_bank
= (struct mcinfo_bank
*)mic
;
270 m
.misc
= mc_bank
->mc_misc
;
271 m
.status
= mc_bank
->mc_status
;
272 m
.addr
= mc_bank
->mc_addr
;
273 m
.tsc
= mc_bank
->mc_tsc
;
274 m
.bank
= mc_bank
->mc_bank
;
279 mic
= x86_mcinfo_next(mic
);
285 static int mc_queue_handle(uint32_t flags
)
290 mc_op
.cmd
= XEN_MC_fetch
;
291 mc_op
.interface_version
= XEN_MCA_INTERFACE_VERSION
;
292 set_xen_guest_handle(mc_op
.u
.mc_fetch
.data
, &g_mi
);
294 mc_op
.u
.mc_fetch
.flags
= flags
;
295 ret
= HYPERVISOR_mca(&mc_op
);
297 pr_err("Failed to fetch %surgent error log\n",
298 flags
== XEN_MC_URGENT
? "" : "non");
302 if (mc_op
.u
.mc_fetch
.flags
& XEN_MC_NODATA
||
303 mc_op
.u
.mc_fetch
.flags
& XEN_MC_FETCHFAILED
)
306 ret
= convert_log(&g_mi
);
308 pr_warn("Failed to convert this error log, continue acking it anyway\n");
310 mc_op
.u
.mc_fetch
.flags
= flags
| XEN_MC_ACK
;
311 ret
= HYPERVISOR_mca(&mc_op
);
313 pr_err("Failed to ack previous error log\n");
322 /* virq handler for machine check error info*/
323 static void xen_mce_work_fn(struct work_struct
*work
)
327 mutex_lock(&mcelog_lock
);
330 err
= mc_queue_handle(XEN_MC_URGENT
);
332 pr_err("Failed to handle urgent mc_info queue, continue handling nonurgent mc_info queue anyway\n");
334 /* nonurgent mc_info */
335 err
= mc_queue_handle(XEN_MC_NONURGENT
);
337 pr_err("Failed to handle nonurgent mc_info queue\n");
339 /* wake processes polling /dev/mcelog */
340 wake_up_interruptible(&xen_mce_chrdev_wait
);
342 mutex_unlock(&mcelog_lock
);
344 static DECLARE_WORK(xen_mce_work
, xen_mce_work_fn
);
346 static irqreturn_t
xen_mce_interrupt(int irq
, void *dev_id
)
348 schedule_work(&xen_mce_work
);
352 static int bind_virq_for_mce(void)
357 memset(&mc_op
, 0, sizeof(struct xen_mc
));
359 /* Fetch physical CPU Numbers */
360 mc_op
.cmd
= XEN_MC_physcpuinfo
;
361 mc_op
.interface_version
= XEN_MCA_INTERFACE_VERSION
;
362 set_xen_guest_handle(mc_op
.u
.mc_physcpuinfo
.info
, g_physinfo
);
363 ret
= HYPERVISOR_mca(&mc_op
);
365 pr_err("Failed to get CPU numbers\n");
369 /* Fetch each CPU Physical Info for later reference*/
370 ncpus
= mc_op
.u
.mc_physcpuinfo
.ncpus
;
371 g_physinfo
= kcalloc(ncpus
, sizeof(struct mcinfo_logical_cpu
),
375 set_xen_guest_handle(mc_op
.u
.mc_physcpuinfo
.info
, g_physinfo
);
376 ret
= HYPERVISOR_mca(&mc_op
);
378 pr_err("Failed to get CPU info\n");
383 ret
= bind_virq_to_irqhandler(VIRQ_MCA
, 0,
384 xen_mce_interrupt
, 0, "mce", NULL
);
386 pr_err("Failed to bind virq\n");
394 static int __init
xen_late_init_mcelog(void)
396 /* Only DOM0 is responsible for MCE logging */
397 if (xen_initial_domain()) {
398 /* register character device /dev/mcelog for xen mcelog */
399 if (misc_register(&xen_mce_chrdev_device
))
401 return bind_virq_for_mce();
406 device_initcall(xen_late_init_mcelog
);