| /****************************************************************************** |
| * mcelog.c |
| * Driver for receiving and transferring machine check error infomation |
| * |
| * Copyright (c) 2012 Intel Corporation |
| * Author: Liu, Jinsong <jinsong.liu@intel.com> |
| * Author: Jiang, Yunhong <yunhong.jiang@intel.com> |
| * Author: Ke, Liping <liping.ke@intel.com> |
| * |
| * This program is free software; you can redistribute it and/or |
| * modify it under the terms of the GNU General Public License version 2 |
| * as published by the Free Software Foundation; or, when distributed |
| * separately from the Linux kernel or incorporated into other |
| * software packages, subject to the following license: |
| * |
| * Permission is hereby granted, free of charge, to any person obtaining a copy |
| * of this source file (the "Software"), to deal in the Software without |
| * restriction, including without limitation the rights to use, copy, modify, |
| * merge, publish, distribute, sublicense, and/or sell copies of the Software, |
| * and to permit persons to whom the Software is furnished to do so, subject to |
| * the following conditions: |
| * |
| * The above copyright notice and this permission notice shall be included in |
| * all copies or substantial portions of the Software. |
| * |
| * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR |
| * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, |
| * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE |
| * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER |
| * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING |
| * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS |
| * IN THE SOFTWARE. |
| */ |
| |
| #define pr_fmt(fmt) "xen_mcelog: " fmt |
| |
| #include <linux/init.h> |
| #include <linux/types.h> |
| #include <linux/kernel.h> |
| #include <linux/slab.h> |
| #include <linux/fs.h> |
| #include <linux/device.h> |
| #include <linux/miscdevice.h> |
| #include <linux/uaccess.h> |
| #include <linux/capability.h> |
| #include <linux/poll.h> |
| #include <linux/sched.h> |
| |
| #include <xen/interface/xen.h> |
| #include <xen/events.h> |
| #include <xen/interface/vcpu.h> |
| #include <xen/xen.h> |
| #include <asm/xen/hypercall.h> |
| #include <asm/xen/hypervisor.h> |
| |
| static struct mc_info g_mi; |
| static struct mcinfo_logical_cpu *g_physinfo; |
| static uint32_t ncpus; |
| |
| static DEFINE_MUTEX(mcelog_lock); |
| |
| static struct xen_mce_log xen_mcelog = { |
| .signature = XEN_MCE_LOG_SIGNATURE, |
| .len = XEN_MCE_LOG_LEN, |
| .recordlen = sizeof(struct xen_mce), |
| }; |
| |
| static DEFINE_SPINLOCK(xen_mce_chrdev_state_lock); |
| static int xen_mce_chrdev_open_count; /* #times opened */ |
| static int xen_mce_chrdev_open_exclu; /* already open exclusive? */ |
| |
| static DECLARE_WAIT_QUEUE_HEAD(xen_mce_chrdev_wait); |
| |
| static int xen_mce_chrdev_open(struct inode *inode, struct file *file) |
| { |
| spin_lock(&xen_mce_chrdev_state_lock); |
| |
| if (xen_mce_chrdev_open_exclu || |
| (xen_mce_chrdev_open_count && (file->f_flags & O_EXCL))) { |
| spin_unlock(&xen_mce_chrdev_state_lock); |
| |
| return -EBUSY; |
| } |
| |
| if (file->f_flags & O_EXCL) |
| xen_mce_chrdev_open_exclu = 1; |
| xen_mce_chrdev_open_count++; |
| |
| spin_unlock(&xen_mce_chrdev_state_lock); |
| |
| return nonseekable_open(inode, file); |
| } |
| |
| static int xen_mce_chrdev_release(struct inode *inode, struct file *file) |
| { |
| spin_lock(&xen_mce_chrdev_state_lock); |
| |
| xen_mce_chrdev_open_count--; |
| xen_mce_chrdev_open_exclu = 0; |
| |
| spin_unlock(&xen_mce_chrdev_state_lock); |
| |
| return 0; |
| } |
| |
| static ssize_t xen_mce_chrdev_read(struct file *filp, char __user *ubuf, |
| size_t usize, loff_t *off) |
| { |
| char __user *buf = ubuf; |
| unsigned num; |
| int i, err; |
| |
| mutex_lock(&mcelog_lock); |
| |
| num = xen_mcelog.next; |
| |
| /* Only supports full reads right now */ |
| err = -EINVAL; |
| if (*off != 0 || usize < XEN_MCE_LOG_LEN*sizeof(struct xen_mce)) |
| goto out; |
| |
| err = 0; |
| for (i = 0; i < num; i++) { |
| struct xen_mce *m = &xen_mcelog.entry[i]; |
| |
| err |= copy_to_user(buf, m, sizeof(*m)); |
| buf += sizeof(*m); |
| } |
| |
| memset(xen_mcelog.entry, 0, num * sizeof(struct xen_mce)); |
| xen_mcelog.next = 0; |
| |
| if (err) |
| err = -EFAULT; |
| |
| out: |
| mutex_unlock(&mcelog_lock); |
| |
| return err ? err : buf - ubuf; |
| } |
| |
| static __poll_t xen_mce_chrdev_poll(struct file *file, poll_table *wait) |
| { |
| poll_wait(file, &xen_mce_chrdev_wait, wait); |
| |
| if (xen_mcelog.next) |
| return EPOLLIN | EPOLLRDNORM; |
| |
| return 0; |
| } |
| |
| static long xen_mce_chrdev_ioctl(struct file *f, unsigned int cmd, |
| unsigned long arg) |
| { |
| int __user *p = (int __user *)arg; |
| |
| if (!capable(CAP_SYS_ADMIN)) |
| return -EPERM; |
| |
| switch (cmd) { |
| case MCE_GET_RECORD_LEN: |
| return put_user(sizeof(struct xen_mce), p); |
| case MCE_GET_LOG_LEN: |
| return put_user(XEN_MCE_LOG_LEN, p); |
| case MCE_GETCLEAR_FLAGS: { |
| unsigned flags; |
| |
| do { |
| flags = xen_mcelog.flags; |
| } while (cmpxchg(&xen_mcelog.flags, flags, 0) != flags); |
| |
| return put_user(flags, p); |
| } |
| default: |
| return -ENOTTY; |
| } |
| } |
| |
| static const struct file_operations xen_mce_chrdev_ops = { |
| .open = xen_mce_chrdev_open, |
| .release = xen_mce_chrdev_release, |
| .read = xen_mce_chrdev_read, |
| .poll = xen_mce_chrdev_poll, |
| .unlocked_ioctl = xen_mce_chrdev_ioctl, |
| }; |
| |
| static struct miscdevice xen_mce_chrdev_device = { |
| MISC_MCELOG_MINOR, |
| "mcelog", |
| &xen_mce_chrdev_ops, |
| }; |
| |
| /* |
| * Caller should hold the mcelog_lock |
| */ |
| static void xen_mce_log(struct xen_mce *mce) |
| { |
| unsigned entry; |
| |
| entry = xen_mcelog.next; |
| |
| /* |
| * When the buffer fills up discard new entries. |
| * Assume that the earlier errors are the more |
| * interesting ones: |
| */ |
| if (entry >= XEN_MCE_LOG_LEN) { |
| set_bit(XEN_MCE_OVERFLOW, |
| (unsigned long *)&xen_mcelog.flags); |
| return; |
| } |
| |
| memcpy(xen_mcelog.entry + entry, mce, sizeof(struct xen_mce)); |
| |
| xen_mcelog.next++; |
| } |
| |
| static int convert_log(struct mc_info *mi) |
| { |
| struct mcinfo_common *mic; |
| struct mcinfo_global *mc_global; |
| struct mcinfo_bank *mc_bank; |
| struct xen_mce m; |
| unsigned int i, j; |
| |
| mic = NULL; |
| x86_mcinfo_lookup(&mic, mi, MC_TYPE_GLOBAL); |
| if (unlikely(!mic)) { |
| pr_warn("Failed to find global error info\n"); |
| return -ENODEV; |
| } |
| |
| memset(&m, 0, sizeof(struct xen_mce)); |
| |
| mc_global = (struct mcinfo_global *)mic; |
| m.mcgstatus = mc_global->mc_gstatus; |
| m.apicid = mc_global->mc_apicid; |
| |
| for (i = 0; i < ncpus; i++) |
| if (g_physinfo[i].mc_apicid == m.apicid) |
| break; |
| if (unlikely(i == ncpus)) { |
| pr_warn("Failed to match cpu with apicid %d\n", m.apicid); |
| return -ENODEV; |
| } |
| |
| m.socketid = g_physinfo[i].mc_chipid; |
| m.cpu = m.extcpu = g_physinfo[i].mc_cpunr; |
| m.cpuvendor = (__u8)g_physinfo[i].mc_vendor; |
| for (j = 0; j < g_physinfo[i].mc_nmsrvals; ++j) |
| switch (g_physinfo[i].mc_msrvalues[j].reg) { |
| case MSR_IA32_MCG_CAP: |
| m.mcgcap = g_physinfo[i].mc_msrvalues[j].value; |
| break; |
| |
| case MSR_PPIN: |
| case MSR_AMD_PPIN: |
| m.ppin = g_physinfo[i].mc_msrvalues[j].value; |
| break; |
| } |
| |
| mic = NULL; |
| x86_mcinfo_lookup(&mic, mi, MC_TYPE_BANK); |
| if (unlikely(!mic)) { |
| pr_warn("Fail to find bank error info\n"); |
| return -ENODEV; |
| } |
| |
| do { |
| if ((!mic) || (mic->size == 0) || |
| (mic->type != MC_TYPE_GLOBAL && |
| mic->type != MC_TYPE_BANK && |
| mic->type != MC_TYPE_EXTENDED && |
| mic->type != MC_TYPE_RECOVERY)) |
| break; |
| |
| if (mic->type == MC_TYPE_BANK) { |
| mc_bank = (struct mcinfo_bank *)mic; |
| m.misc = mc_bank->mc_misc; |
| m.status = mc_bank->mc_status; |
| m.addr = mc_bank->mc_addr; |
| m.tsc = mc_bank->mc_tsc; |
| m.bank = mc_bank->mc_bank; |
| m.finished = 1; |
| /*log this record*/ |
| xen_mce_log(&m); |
| } |
| mic = x86_mcinfo_next(mic); |
| } while (1); |
| |
| return 0; |
| } |
| |
| static int mc_queue_handle(uint32_t flags) |
| { |
| struct xen_mc mc_op; |
| int ret = 0; |
| |
| mc_op.cmd = XEN_MC_fetch; |
| set_xen_guest_handle(mc_op.u.mc_fetch.data, &g_mi); |
| do { |
| mc_op.u.mc_fetch.flags = flags; |
| ret = HYPERVISOR_mca(&mc_op); |
| if (ret) { |
| pr_err("Failed to fetch %surgent error log\n", |
| flags == XEN_MC_URGENT ? "" : "non"); |
| break; |
| } |
| |
| if (mc_op.u.mc_fetch.flags & XEN_MC_NODATA || |
| mc_op.u.mc_fetch.flags & XEN_MC_FETCHFAILED) |
| break; |
| else { |
| ret = convert_log(&g_mi); |
| if (ret) |
| pr_warn("Failed to convert this error log, continue acking it anyway\n"); |
| |
| mc_op.u.mc_fetch.flags = flags | XEN_MC_ACK; |
| ret = HYPERVISOR_mca(&mc_op); |
| if (ret) { |
| pr_err("Failed to ack previous error log\n"); |
| break; |
| } |
| } |
| } while (1); |
| |
| return ret; |
| } |
| |
| /* virq handler for machine check error info*/ |
| static void xen_mce_work_fn(struct work_struct *work) |
| { |
| int err; |
| |
| mutex_lock(&mcelog_lock); |
| |
| /* urgent mc_info */ |
| err = mc_queue_handle(XEN_MC_URGENT); |
| if (err) |
| pr_err("Failed to handle urgent mc_info queue, continue handling nonurgent mc_info queue anyway\n"); |
| |
| /* nonurgent mc_info */ |
| err = mc_queue_handle(XEN_MC_NONURGENT); |
| if (err) |
| pr_err("Failed to handle nonurgent mc_info queue\n"); |
| |
| /* wake processes polling /dev/mcelog */ |
| wake_up_interruptible(&xen_mce_chrdev_wait); |
| |
| mutex_unlock(&mcelog_lock); |
| } |
| static DECLARE_WORK(xen_mce_work, xen_mce_work_fn); |
| |
| static irqreturn_t xen_mce_interrupt(int irq, void *dev_id) |
| { |
| schedule_work(&xen_mce_work); |
| return IRQ_HANDLED; |
| } |
| |
| static int bind_virq_for_mce(void) |
| { |
| int ret; |
| struct xen_mc mc_op; |
| |
| memset(&mc_op, 0, sizeof(struct xen_mc)); |
| |
| /* Fetch physical CPU Numbers */ |
| mc_op.cmd = XEN_MC_physcpuinfo; |
| set_xen_guest_handle(mc_op.u.mc_physcpuinfo.info, g_physinfo); |
| ret = HYPERVISOR_mca(&mc_op); |
| if (ret) { |
| pr_err("Failed to get CPU numbers\n"); |
| return ret; |
| } |
| |
| /* Fetch each CPU Physical Info for later reference*/ |
| ncpus = mc_op.u.mc_physcpuinfo.ncpus; |
| g_physinfo = kcalloc(ncpus, sizeof(struct mcinfo_logical_cpu), |
| GFP_KERNEL); |
| if (!g_physinfo) |
| return -ENOMEM; |
| set_xen_guest_handle(mc_op.u.mc_physcpuinfo.info, g_physinfo); |
| ret = HYPERVISOR_mca(&mc_op); |
| if (ret) { |
| pr_err("Failed to get CPU info\n"); |
| kfree(g_physinfo); |
| return ret; |
| } |
| |
| ret = bind_virq_to_irqhandler(VIRQ_MCA, 0, |
| xen_mce_interrupt, 0, "mce", NULL); |
| if (ret < 0) { |
| pr_err("Failed to bind virq\n"); |
| kfree(g_physinfo); |
| return ret; |
| } |
| |
| return 0; |
| } |
| |
| static int __init xen_late_init_mcelog(void) |
| { |
| int ret; |
| |
| /* Only DOM0 is responsible for MCE logging */ |
| if (!xen_initial_domain()) |
| return -ENODEV; |
| |
| /* register character device /dev/mcelog for xen mcelog */ |
| ret = misc_register(&xen_mce_chrdev_device); |
| if (ret) |
| return ret; |
| |
| ret = bind_virq_for_mce(); |
| if (ret) |
| goto deregister; |
| |
| pr_info("/dev/mcelog registered by Xen\n"); |
| |
| return 0; |
| |
| deregister: |
| misc_deregister(&xen_mce_chrdev_device); |
| return ret; |
| } |
| device_initcall(xen_late_init_mcelog); |