| // SPDX-License-Identifier: GPL-2.0-only |
| /* |
| * VGIC: KVM DEVICE API |
| * |
| * Copyright (C) 2015 ARM Ltd. |
| * Author: Marc Zyngier <marc.zyngier@arm.com> |
| */ |
| #include <linux/kvm_host.h> |
| #include <kvm/arm_vgic.h> |
| #include <linux/uaccess.h> |
| #include <asm/kvm_mmu.h> |
| #include <asm/cputype.h> |
| #include "vgic.h" |
| |
| /* common helpers */ |
| |
| int vgic_check_ioaddr(struct kvm *kvm, phys_addr_t *ioaddr, |
| phys_addr_t addr, phys_addr_t alignment) |
| { |
| if (addr & ~kvm_phys_mask(kvm)) |
| return -E2BIG; |
| |
| if (!IS_ALIGNED(addr, alignment)) |
| return -EINVAL; |
| |
| if (!IS_VGIC_ADDR_UNDEF(*ioaddr)) |
| return -EEXIST; |
| |
| return 0; |
| } |
| |
| static int vgic_check_type(struct kvm *kvm, int type_needed) |
| { |
| if (kvm->arch.vgic.vgic_model != type_needed) |
| return -ENODEV; |
| else |
| return 0; |
| } |
| |
| /** |
| * kvm_vgic_addr - set or get vgic VM base addresses |
| * @kvm: pointer to the vm struct |
| * @type: the VGIC addr type, one of KVM_VGIC_V[23]_ADDR_TYPE_XXX |
| * @addr: pointer to address value |
| * @write: if true set the address in the VM address space, if false read the |
| * address |
| * |
| * Set or get the vgic base addresses for the distributor and the virtual CPU |
| * interface in the VM physical address space. These addresses are properties |
| * of the emulated core/SoC and therefore user space initially knows this |
| * information. |
| * Check them for sanity (alignment, double assignment). We can't check for |
| * overlapping regions in case of a virtual GICv3 here, since we don't know |
| * the number of VCPUs yet, so we defer this check to map_resources(). |
| */ |
| int kvm_vgic_addr(struct kvm *kvm, unsigned long type, u64 *addr, bool write) |
| { |
| int r = 0; |
| struct vgic_dist *vgic = &kvm->arch.vgic; |
| phys_addr_t *addr_ptr, alignment; |
| u64 undef_value = VGIC_ADDR_UNDEF; |
| |
| mutex_lock(&kvm->lock); |
| switch (type) { |
| case KVM_VGIC_V2_ADDR_TYPE_DIST: |
| r = vgic_check_type(kvm, KVM_DEV_TYPE_ARM_VGIC_V2); |
| addr_ptr = &vgic->vgic_dist_base; |
| alignment = SZ_4K; |
| break; |
| case KVM_VGIC_V2_ADDR_TYPE_CPU: |
| r = vgic_check_type(kvm, KVM_DEV_TYPE_ARM_VGIC_V2); |
| addr_ptr = &vgic->vgic_cpu_base; |
| alignment = SZ_4K; |
| break; |
| case KVM_VGIC_V3_ADDR_TYPE_DIST: |
| r = vgic_check_type(kvm, KVM_DEV_TYPE_ARM_VGIC_V3); |
| addr_ptr = &vgic->vgic_dist_base; |
| alignment = SZ_64K; |
| break; |
| case KVM_VGIC_V3_ADDR_TYPE_REDIST: { |
| struct vgic_redist_region *rdreg; |
| |
| r = vgic_check_type(kvm, KVM_DEV_TYPE_ARM_VGIC_V3); |
| if (r) |
| break; |
| if (write) { |
| r = vgic_v3_set_redist_base(kvm, 0, *addr, 0); |
| goto out; |
| } |
| rdreg = list_first_entry(&vgic->rd_regions, |
| struct vgic_redist_region, list); |
| if (!rdreg) |
| addr_ptr = &undef_value; |
| else |
| addr_ptr = &rdreg->base; |
| break; |
| } |
| case KVM_VGIC_V3_ADDR_TYPE_REDIST_REGION: |
| { |
| struct vgic_redist_region *rdreg; |
| u8 index; |
| |
| r = vgic_check_type(kvm, KVM_DEV_TYPE_ARM_VGIC_V3); |
| if (r) |
| break; |
| |
| index = *addr & KVM_VGIC_V3_RDIST_INDEX_MASK; |
| |
| if (write) { |
| gpa_t base = *addr & KVM_VGIC_V3_RDIST_BASE_MASK; |
| u32 count = (*addr & KVM_VGIC_V3_RDIST_COUNT_MASK) |
| >> KVM_VGIC_V3_RDIST_COUNT_SHIFT; |
| u8 flags = (*addr & KVM_VGIC_V3_RDIST_FLAGS_MASK) |
| >> KVM_VGIC_V3_RDIST_FLAGS_SHIFT; |
| |
| if (!count || flags) |
| r = -EINVAL; |
| else |
| r = vgic_v3_set_redist_base(kvm, index, |
| base, count); |
| goto out; |
| } |
| |
| rdreg = vgic_v3_rdist_region_from_index(kvm, index); |
| if (!rdreg) { |
| r = -ENOENT; |
| goto out; |
| } |
| |
| *addr = index; |
| *addr |= rdreg->base; |
| *addr |= (u64)rdreg->count << KVM_VGIC_V3_RDIST_COUNT_SHIFT; |
| goto out; |
| } |
| default: |
| r = -ENODEV; |
| } |
| |
| if (r) |
| goto out; |
| |
| if (write) { |
| r = vgic_check_ioaddr(kvm, addr_ptr, *addr, alignment); |
| if (!r) |
| *addr_ptr = *addr; |
| } else { |
| *addr = *addr_ptr; |
| } |
| |
| out: |
| mutex_unlock(&kvm->lock); |
| return r; |
| } |
| |
| static int vgic_set_common_attr(struct kvm_device *dev, |
| struct kvm_device_attr *attr) |
| { |
| int r; |
| |
| switch (attr->group) { |
| case KVM_DEV_ARM_VGIC_GRP_ADDR: { |
| u64 __user *uaddr = (u64 __user *)(long)attr->addr; |
| u64 addr; |
| unsigned long type = (unsigned long)attr->attr; |
| |
| if (copy_from_user(&addr, uaddr, sizeof(addr))) |
| return -EFAULT; |
| |
| r = kvm_vgic_addr(dev->kvm, type, &addr, true); |
| return (r == -ENODEV) ? -ENXIO : r; |
| } |
| case KVM_DEV_ARM_VGIC_GRP_NR_IRQS: { |
| u32 __user *uaddr = (u32 __user *)(long)attr->addr; |
| u32 val; |
| int ret = 0; |
| |
| if (get_user(val, uaddr)) |
| return -EFAULT; |
| |
| /* |
| * We require: |
| * - at least 32 SPIs on top of the 16 SGIs and 16 PPIs |
| * - at most 1024 interrupts |
| * - a multiple of 32 interrupts |
| */ |
| if (val < (VGIC_NR_PRIVATE_IRQS + 32) || |
| val > VGIC_MAX_RESERVED || |
| (val & 31)) |
| return -EINVAL; |
| |
| mutex_lock(&dev->kvm->lock); |
| |
| if (vgic_ready(dev->kvm) || dev->kvm->arch.vgic.nr_spis) |
| ret = -EBUSY; |
| else |
| dev->kvm->arch.vgic.nr_spis = |
| val - VGIC_NR_PRIVATE_IRQS; |
| |
| mutex_unlock(&dev->kvm->lock); |
| |
| return ret; |
| } |
| case KVM_DEV_ARM_VGIC_GRP_CTRL: { |
| switch (attr->attr) { |
| case KVM_DEV_ARM_VGIC_CTRL_INIT: |
| mutex_lock(&dev->kvm->lock); |
| r = vgic_init(dev->kvm); |
| mutex_unlock(&dev->kvm->lock); |
| return r; |
| } |
| break; |
| } |
| } |
| |
| return -ENXIO; |
| } |
| |
| static int vgic_get_common_attr(struct kvm_device *dev, |
| struct kvm_device_attr *attr) |
| { |
| int r = -ENXIO; |
| |
| switch (attr->group) { |
| case KVM_DEV_ARM_VGIC_GRP_ADDR: { |
| u64 __user *uaddr = (u64 __user *)(long)attr->addr; |
| u64 addr; |
| unsigned long type = (unsigned long)attr->attr; |
| |
| r = kvm_vgic_addr(dev->kvm, type, &addr, false); |
| if (r) |
| return (r == -ENODEV) ? -ENXIO : r; |
| |
| if (copy_to_user(uaddr, &addr, sizeof(addr))) |
| return -EFAULT; |
| break; |
| } |
| case KVM_DEV_ARM_VGIC_GRP_NR_IRQS: { |
| u32 __user *uaddr = (u32 __user *)(long)attr->addr; |
| |
| r = put_user(dev->kvm->arch.vgic.nr_spis + |
| VGIC_NR_PRIVATE_IRQS, uaddr); |
| break; |
| } |
| } |
| |
| return r; |
| } |
| |
| static int vgic_create(struct kvm_device *dev, u32 type) |
| { |
| return kvm_vgic_create(dev->kvm, type); |
| } |
| |
| static void vgic_destroy(struct kvm_device *dev) |
| { |
| kfree(dev); |
| } |
| |
| int kvm_register_vgic_device(unsigned long type) |
| { |
| int ret = -ENODEV; |
| |
| switch (type) { |
| case KVM_DEV_TYPE_ARM_VGIC_V2: |
| ret = kvm_register_device_ops(&kvm_arm_vgic_v2_ops, |
| KVM_DEV_TYPE_ARM_VGIC_V2); |
| break; |
| case KVM_DEV_TYPE_ARM_VGIC_V3: |
| ret = kvm_register_device_ops(&kvm_arm_vgic_v3_ops, |
| KVM_DEV_TYPE_ARM_VGIC_V3); |
| |
| if (ret) |
| break; |
| ret = kvm_vgic_register_its_device(); |
| break; |
| } |
| |
| return ret; |
| } |
| |
| int vgic_v2_parse_attr(struct kvm_device *dev, struct kvm_device_attr *attr, |
| struct vgic_reg_attr *reg_attr) |
| { |
| int cpuid; |
| |
| cpuid = (attr->attr & KVM_DEV_ARM_VGIC_CPUID_MASK) >> |
| KVM_DEV_ARM_VGIC_CPUID_SHIFT; |
| |
| if (cpuid >= atomic_read(&dev->kvm->online_vcpus)) |
| return -EINVAL; |
| |
| reg_attr->vcpu = kvm_get_vcpu(dev->kvm, cpuid); |
| reg_attr->addr = attr->attr & KVM_DEV_ARM_VGIC_OFFSET_MASK; |
| |
| return 0; |
| } |
| |
| /* unlocks vcpus from @vcpu_lock_idx and smaller */ |
| static void unlock_vcpus(struct kvm *kvm, int vcpu_lock_idx) |
| { |
| struct kvm_vcpu *tmp_vcpu; |
| |
| for (; vcpu_lock_idx >= 0; vcpu_lock_idx--) { |
| tmp_vcpu = kvm_get_vcpu(kvm, vcpu_lock_idx); |
| mutex_unlock(&tmp_vcpu->mutex); |
| } |
| } |
| |
| void unlock_all_vcpus(struct kvm *kvm) |
| { |
| unlock_vcpus(kvm, atomic_read(&kvm->online_vcpus) - 1); |
| } |
| |
| /* Returns true if all vcpus were locked, false otherwise */ |
| bool lock_all_vcpus(struct kvm *kvm) |
| { |
| struct kvm_vcpu *tmp_vcpu; |
| int c; |
| |
| /* |
| * Any time a vcpu is run, vcpu_load is called which tries to grab the |
| * vcpu->mutex. By grabbing the vcpu->mutex of all VCPUs we ensure |
| * that no other VCPUs are run and fiddle with the vgic state while we |
| * access it. |
| */ |
| kvm_for_each_vcpu(c, tmp_vcpu, kvm) { |
| if (!mutex_trylock(&tmp_vcpu->mutex)) { |
| unlock_vcpus(kvm, c - 1); |
| return false; |
| } |
| } |
| |
| return true; |
| } |
| |
| /** |
| * vgic_v2_attr_regs_access - allows user space to access VGIC v2 state |
| * |
| * @dev: kvm device handle |
| * @attr: kvm device attribute |
| * @reg: address the value is read or written |
| * @is_write: true if userspace is writing a register |
| */ |
| static int vgic_v2_attr_regs_access(struct kvm_device *dev, |
| struct kvm_device_attr *attr, |
| u32 *reg, bool is_write) |
| { |
| struct vgic_reg_attr reg_attr; |
| gpa_t addr; |
| struct kvm_vcpu *vcpu; |
| int ret; |
| |
| ret = vgic_v2_parse_attr(dev, attr, ®_attr); |
| if (ret) |
| return ret; |
| |
| vcpu = reg_attr.vcpu; |
| addr = reg_attr.addr; |
| |
| mutex_lock(&dev->kvm->lock); |
| |
| ret = vgic_init(dev->kvm); |
| if (ret) |
| goto out; |
| |
| if (!lock_all_vcpus(dev->kvm)) { |
| ret = -EBUSY; |
| goto out; |
| } |
| |
| switch (attr->group) { |
| case KVM_DEV_ARM_VGIC_GRP_CPU_REGS: |
| ret = vgic_v2_cpuif_uaccess(vcpu, is_write, addr, reg); |
| break; |
| case KVM_DEV_ARM_VGIC_GRP_DIST_REGS: |
| ret = vgic_v2_dist_uaccess(vcpu, is_write, addr, reg); |
| break; |
| default: |
| ret = -EINVAL; |
| break; |
| } |
| |
| unlock_all_vcpus(dev->kvm); |
| out: |
| mutex_unlock(&dev->kvm->lock); |
| return ret; |
| } |
| |
| static int vgic_v2_set_attr(struct kvm_device *dev, |
| struct kvm_device_attr *attr) |
| { |
| int ret; |
| |
| ret = vgic_set_common_attr(dev, attr); |
| if (ret != -ENXIO) |
| return ret; |
| |
| switch (attr->group) { |
| case KVM_DEV_ARM_VGIC_GRP_DIST_REGS: |
| case KVM_DEV_ARM_VGIC_GRP_CPU_REGS: { |
| u32 __user *uaddr = (u32 __user *)(long)attr->addr; |
| u32 reg; |
| |
| if (get_user(reg, uaddr)) |
| return -EFAULT; |
| |
| return vgic_v2_attr_regs_access(dev, attr, ®, true); |
| } |
| } |
| |
| return -ENXIO; |
| } |
| |
| static int vgic_v2_get_attr(struct kvm_device *dev, |
| struct kvm_device_attr *attr) |
| { |
| int ret; |
| |
| ret = vgic_get_common_attr(dev, attr); |
| if (ret != -ENXIO) |
| return ret; |
| |
| switch (attr->group) { |
| case KVM_DEV_ARM_VGIC_GRP_DIST_REGS: |
| case KVM_DEV_ARM_VGIC_GRP_CPU_REGS: { |
| u32 __user *uaddr = (u32 __user *)(long)attr->addr; |
| u32 reg = 0; |
| |
| ret = vgic_v2_attr_regs_access(dev, attr, ®, false); |
| if (ret) |
| return ret; |
| return put_user(reg, uaddr); |
| } |
| } |
| |
| return -ENXIO; |
| } |
| |
| static int vgic_v2_has_attr(struct kvm_device *dev, |
| struct kvm_device_attr *attr) |
| { |
| switch (attr->group) { |
| case KVM_DEV_ARM_VGIC_GRP_ADDR: |
| switch (attr->attr) { |
| case KVM_VGIC_V2_ADDR_TYPE_DIST: |
| case KVM_VGIC_V2_ADDR_TYPE_CPU: |
| return 0; |
| } |
| break; |
| case KVM_DEV_ARM_VGIC_GRP_DIST_REGS: |
| case KVM_DEV_ARM_VGIC_GRP_CPU_REGS: |
| return vgic_v2_has_attr_regs(dev, attr); |
| case KVM_DEV_ARM_VGIC_GRP_NR_IRQS: |
| return 0; |
| case KVM_DEV_ARM_VGIC_GRP_CTRL: |
| switch (attr->attr) { |
| case KVM_DEV_ARM_VGIC_CTRL_INIT: |
| return 0; |
| } |
| } |
| return -ENXIO; |
| } |
| |
| struct kvm_device_ops kvm_arm_vgic_v2_ops = { |
| .name = "kvm-arm-vgic-v2", |
| .create = vgic_create, |
| .destroy = vgic_destroy, |
| .set_attr = vgic_v2_set_attr, |
| .get_attr = vgic_v2_get_attr, |
| .has_attr = vgic_v2_has_attr, |
| }; |
| |
| int vgic_v3_parse_attr(struct kvm_device *dev, struct kvm_device_attr *attr, |
| struct vgic_reg_attr *reg_attr) |
| { |
| unsigned long vgic_mpidr, mpidr_reg; |
| |
| /* |
| * For KVM_DEV_ARM_VGIC_GRP_DIST_REGS group, |
| * attr might not hold MPIDR. Hence assume vcpu0. |
| */ |
| if (attr->group != KVM_DEV_ARM_VGIC_GRP_DIST_REGS) { |
| vgic_mpidr = (attr->attr & KVM_DEV_ARM_VGIC_V3_MPIDR_MASK) >> |
| KVM_DEV_ARM_VGIC_V3_MPIDR_SHIFT; |
| |
| mpidr_reg = VGIC_TO_MPIDR(vgic_mpidr); |
| reg_attr->vcpu = kvm_mpidr_to_vcpu(dev->kvm, mpidr_reg); |
| } else { |
| reg_attr->vcpu = kvm_get_vcpu(dev->kvm, 0); |
| } |
| |
| if (!reg_attr->vcpu) |
| return -EINVAL; |
| |
| reg_attr->addr = attr->attr & KVM_DEV_ARM_VGIC_OFFSET_MASK; |
| |
| return 0; |
| } |
| |
| /* |
| * vgic_v3_attr_regs_access - allows user space to access VGIC v3 state |
| * |
| * @dev: kvm device handle |
| * @attr: kvm device attribute |
| * @reg: address the value is read or written |
| * @is_write: true if userspace is writing a register |
| */ |
| static int vgic_v3_attr_regs_access(struct kvm_device *dev, |
| struct kvm_device_attr *attr, |
| u64 *reg, bool is_write) |
| { |
| struct vgic_reg_attr reg_attr; |
| gpa_t addr; |
| struct kvm_vcpu *vcpu; |
| int ret; |
| u32 tmp32; |
| |
| ret = vgic_v3_parse_attr(dev, attr, ®_attr); |
| if (ret) |
| return ret; |
| |
| vcpu = reg_attr.vcpu; |
| addr = reg_attr.addr; |
| |
| mutex_lock(&dev->kvm->lock); |
| |
| if (unlikely(!vgic_initialized(dev->kvm))) { |
| ret = -EBUSY; |
| goto out; |
| } |
| |
| if (!lock_all_vcpus(dev->kvm)) { |
| ret = -EBUSY; |
| goto out; |
| } |
| |
| switch (attr->group) { |
| case KVM_DEV_ARM_VGIC_GRP_DIST_REGS: |
| if (is_write) |
| tmp32 = *reg; |
| |
| ret = vgic_v3_dist_uaccess(vcpu, is_write, addr, &tmp32); |
| if (!is_write) |
| *reg = tmp32; |
| break; |
| case KVM_DEV_ARM_VGIC_GRP_REDIST_REGS: |
| if (is_write) |
| tmp32 = *reg; |
| |
| ret = vgic_v3_redist_uaccess(vcpu, is_write, addr, &tmp32); |
| if (!is_write) |
| *reg = tmp32; |
| break; |
| case KVM_DEV_ARM_VGIC_GRP_CPU_SYSREGS: { |
| u64 regid; |
| |
| regid = (attr->attr & KVM_DEV_ARM_VGIC_SYSREG_INSTR_MASK); |
| ret = vgic_v3_cpu_sysregs_uaccess(vcpu, is_write, |
| regid, reg); |
| break; |
| } |
| case KVM_DEV_ARM_VGIC_GRP_LEVEL_INFO: { |
| unsigned int info, intid; |
| |
| info = (attr->attr & KVM_DEV_ARM_VGIC_LINE_LEVEL_INFO_MASK) >> |
| KVM_DEV_ARM_VGIC_LINE_LEVEL_INFO_SHIFT; |
| if (info == VGIC_LEVEL_INFO_LINE_LEVEL) { |
| intid = attr->attr & |
| KVM_DEV_ARM_VGIC_LINE_LEVEL_INTID_MASK; |
| ret = vgic_v3_line_level_info_uaccess(vcpu, is_write, |
| intid, reg); |
| } else { |
| ret = -EINVAL; |
| } |
| break; |
| } |
| default: |
| ret = -EINVAL; |
| break; |
| } |
| |
| unlock_all_vcpus(dev->kvm); |
| out: |
| mutex_unlock(&dev->kvm->lock); |
| return ret; |
| } |
| |
| static int vgic_v3_set_attr(struct kvm_device *dev, |
| struct kvm_device_attr *attr) |
| { |
| int ret; |
| |
| ret = vgic_set_common_attr(dev, attr); |
| if (ret != -ENXIO) |
| return ret; |
| |
| switch (attr->group) { |
| case KVM_DEV_ARM_VGIC_GRP_DIST_REGS: |
| case KVM_DEV_ARM_VGIC_GRP_REDIST_REGS: { |
| u32 __user *uaddr = (u32 __user *)(long)attr->addr; |
| u32 tmp32; |
| u64 reg; |
| |
| if (get_user(tmp32, uaddr)) |
| return -EFAULT; |
| |
| reg = tmp32; |
| return vgic_v3_attr_regs_access(dev, attr, ®, true); |
| } |
| case KVM_DEV_ARM_VGIC_GRP_CPU_SYSREGS: { |
| u64 __user *uaddr = (u64 __user *)(long)attr->addr; |
| u64 reg; |
| |
| if (get_user(reg, uaddr)) |
| return -EFAULT; |
| |
| return vgic_v3_attr_regs_access(dev, attr, ®, true); |
| } |
| case KVM_DEV_ARM_VGIC_GRP_LEVEL_INFO: { |
| u32 __user *uaddr = (u32 __user *)(long)attr->addr; |
| u64 reg; |
| u32 tmp32; |
| |
| if (get_user(tmp32, uaddr)) |
| return -EFAULT; |
| |
| reg = tmp32; |
| return vgic_v3_attr_regs_access(dev, attr, ®, true); |
| } |
| case KVM_DEV_ARM_VGIC_GRP_CTRL: { |
| int ret; |
| |
| switch (attr->attr) { |
| case KVM_DEV_ARM_VGIC_SAVE_PENDING_TABLES: |
| mutex_lock(&dev->kvm->lock); |
| |
| if (!lock_all_vcpus(dev->kvm)) { |
| mutex_unlock(&dev->kvm->lock); |
| return -EBUSY; |
| } |
| ret = vgic_v3_save_pending_tables(dev->kvm); |
| unlock_all_vcpus(dev->kvm); |
| mutex_unlock(&dev->kvm->lock); |
| return ret; |
| } |
| break; |
| } |
| } |
| return -ENXIO; |
| } |
| |
| static int vgic_v3_get_attr(struct kvm_device *dev, |
| struct kvm_device_attr *attr) |
| { |
| int ret; |
| |
| ret = vgic_get_common_attr(dev, attr); |
| if (ret != -ENXIO) |
| return ret; |
| |
| switch (attr->group) { |
| case KVM_DEV_ARM_VGIC_GRP_DIST_REGS: |
| case KVM_DEV_ARM_VGIC_GRP_REDIST_REGS: { |
| u32 __user *uaddr = (u32 __user *)(long)attr->addr; |
| u64 reg; |
| u32 tmp32; |
| |
| ret = vgic_v3_attr_regs_access(dev, attr, ®, false); |
| if (ret) |
| return ret; |
| tmp32 = reg; |
| return put_user(tmp32, uaddr); |
| } |
| case KVM_DEV_ARM_VGIC_GRP_CPU_SYSREGS: { |
| u64 __user *uaddr = (u64 __user *)(long)attr->addr; |
| u64 reg; |
| |
| ret = vgic_v3_attr_regs_access(dev, attr, ®, false); |
| if (ret) |
| return ret; |
| return put_user(reg, uaddr); |
| } |
| case KVM_DEV_ARM_VGIC_GRP_LEVEL_INFO: { |
| u32 __user *uaddr = (u32 __user *)(long)attr->addr; |
| u64 reg; |
| u32 tmp32; |
| |
| ret = vgic_v3_attr_regs_access(dev, attr, ®, false); |
| if (ret) |
| return ret; |
| tmp32 = reg; |
| return put_user(tmp32, uaddr); |
| } |
| } |
| return -ENXIO; |
| } |
| |
| static int vgic_v3_has_attr(struct kvm_device *dev, |
| struct kvm_device_attr *attr) |
| { |
| switch (attr->group) { |
| case KVM_DEV_ARM_VGIC_GRP_ADDR: |
| switch (attr->attr) { |
| case KVM_VGIC_V3_ADDR_TYPE_DIST: |
| case KVM_VGIC_V3_ADDR_TYPE_REDIST: |
| case KVM_VGIC_V3_ADDR_TYPE_REDIST_REGION: |
| return 0; |
| } |
| break; |
| case KVM_DEV_ARM_VGIC_GRP_DIST_REGS: |
| case KVM_DEV_ARM_VGIC_GRP_REDIST_REGS: |
| case KVM_DEV_ARM_VGIC_GRP_CPU_SYSREGS: |
| return vgic_v3_has_attr_regs(dev, attr); |
| case KVM_DEV_ARM_VGIC_GRP_NR_IRQS: |
| return 0; |
| case KVM_DEV_ARM_VGIC_GRP_LEVEL_INFO: { |
| if (((attr->attr & KVM_DEV_ARM_VGIC_LINE_LEVEL_INFO_MASK) >> |
| KVM_DEV_ARM_VGIC_LINE_LEVEL_INFO_SHIFT) == |
| VGIC_LEVEL_INFO_LINE_LEVEL) |
| return 0; |
| break; |
| } |
| case KVM_DEV_ARM_VGIC_GRP_CTRL: |
| switch (attr->attr) { |
| case KVM_DEV_ARM_VGIC_CTRL_INIT: |
| return 0; |
| case KVM_DEV_ARM_VGIC_SAVE_PENDING_TABLES: |
| return 0; |
| } |
| } |
| return -ENXIO; |
| } |
| |
| struct kvm_device_ops kvm_arm_vgic_v3_ops = { |
| .name = "kvm-arm-vgic-v3", |
| .create = vgic_create, |
| .destroy = vgic_destroy, |
| .set_attr = vgic_v3_set_attr, |
| .get_attr = vgic_v3_get_attr, |
| .has_attr = vgic_v3_has_attr, |
| }; |