| // SPDX-License-Identifier: GPL-2.0 |
| /* |
| * s390 kvm PCI passthrough support |
| * |
| * Copyright IBM Corp. 2022 |
| * |
| * Author(s): Matthew Rosato <mjrosato@linux.ibm.com> |
| */ |
| |
| #include <linux/kvm_host.h> |
| #include <linux/pci.h> |
| #include <asm/pci.h> |
| #include <asm/pci_insn.h> |
| #include <asm/pci_io.h> |
| #include <asm/sclp.h> |
| #include "pci.h" |
| #include "kvm-s390.h" |
| |
| struct zpci_aift *aift; |
| |
| static inline int __set_irq_noiib(u16 ctl, u8 isc) |
| { |
| union zpci_sic_iib iib = {{0}}; |
| |
| return zpci_set_irq_ctrl(ctl, isc, &iib); |
| } |
| |
| void kvm_s390_pci_aen_exit(void) |
| { |
| unsigned long flags; |
| struct kvm_zdev **gait_kzdev; |
| |
| lockdep_assert_held(&aift->aift_lock); |
| |
| /* |
| * Contents of the aipb remain registered for the life of the host |
| * kernel, the information preserved in zpci_aipb and zpci_aif_sbv |
| * in case we insert the KVM module again later. Clear the AIFT |
| * information and free anything not registered with underlying |
| * firmware. |
| */ |
| spin_lock_irqsave(&aift->gait_lock, flags); |
| gait_kzdev = aift->kzdev; |
| aift->gait = NULL; |
| aift->sbv = NULL; |
| aift->kzdev = NULL; |
| spin_unlock_irqrestore(&aift->gait_lock, flags); |
| |
| kfree(gait_kzdev); |
| } |
| |
| static int zpci_setup_aipb(u8 nisc) |
| { |
| struct page *page; |
| int size, rc; |
| |
| zpci_aipb = kzalloc(sizeof(union zpci_sic_iib), GFP_KERNEL); |
| if (!zpci_aipb) |
| return -ENOMEM; |
| |
| aift->sbv = airq_iv_create(ZPCI_NR_DEVICES, AIRQ_IV_ALLOC, NULL); |
| if (!aift->sbv) { |
| rc = -ENOMEM; |
| goto free_aipb; |
| } |
| zpci_aif_sbv = aift->sbv; |
| size = get_order(PAGE_ALIGN(ZPCI_NR_DEVICES * |
| sizeof(struct zpci_gaite))); |
| page = alloc_pages(GFP_KERNEL | __GFP_ZERO, size); |
| if (!page) { |
| rc = -ENOMEM; |
| goto free_sbv; |
| } |
| aift->gait = (struct zpci_gaite *)page_to_virt(page); |
| |
| zpci_aipb->aipb.faisb = virt_to_phys(aift->sbv->vector); |
| zpci_aipb->aipb.gait = virt_to_phys(aift->gait); |
| zpci_aipb->aipb.afi = nisc; |
| zpci_aipb->aipb.faal = ZPCI_NR_DEVICES; |
| |
| /* Setup Adapter Event Notification Interpretation */ |
| if (zpci_set_irq_ctrl(SIC_SET_AENI_CONTROLS, 0, zpci_aipb)) { |
| rc = -EIO; |
| goto free_gait; |
| } |
| |
| return 0; |
| |
| free_gait: |
| free_pages((unsigned long)aift->gait, size); |
| free_sbv: |
| airq_iv_release(aift->sbv); |
| zpci_aif_sbv = NULL; |
| free_aipb: |
| kfree(zpci_aipb); |
| zpci_aipb = NULL; |
| |
| return rc; |
| } |
| |
| static int zpci_reset_aipb(u8 nisc) |
| { |
| /* |
| * AEN registration can only happen once per system boot. If |
| * an aipb already exists then AEN was already registered and |
| * we can re-use the aipb contents. This can only happen if |
| * the KVM module was removed and re-inserted. However, we must |
| * ensure that the same forwarding ISC is used as this is assigned |
| * during KVM module load. |
| */ |
| if (zpci_aipb->aipb.afi != nisc) |
| return -EINVAL; |
| |
| aift->sbv = zpci_aif_sbv; |
| aift->gait = phys_to_virt(zpci_aipb->aipb.gait); |
| |
| return 0; |
| } |
| |
| int kvm_s390_pci_aen_init(u8 nisc) |
| { |
| int rc = 0; |
| |
| /* If already enabled for AEN, bail out now */ |
| if (aift->gait || aift->sbv) |
| return -EPERM; |
| |
| mutex_lock(&aift->aift_lock); |
| aift->kzdev = kcalloc(ZPCI_NR_DEVICES, sizeof(struct kvm_zdev *), |
| GFP_KERNEL); |
| if (!aift->kzdev) { |
| rc = -ENOMEM; |
| goto unlock; |
| } |
| |
| if (!zpci_aipb) |
| rc = zpci_setup_aipb(nisc); |
| else |
| rc = zpci_reset_aipb(nisc); |
| if (rc) |
| goto free_zdev; |
| |
| /* Enable floating IRQs */ |
| if (__set_irq_noiib(SIC_IRQ_MODE_SINGLE, nisc)) { |
| rc = -EIO; |
| kvm_s390_pci_aen_exit(); |
| } |
| |
| goto unlock; |
| |
| free_zdev: |
| kfree(aift->kzdev); |
| unlock: |
| mutex_unlock(&aift->aift_lock); |
| return rc; |
| } |
| |
| /* Modify PCI: Register floating adapter interruption forwarding */ |
| static int kvm_zpci_set_airq(struct zpci_dev *zdev) |
| { |
| u64 req = ZPCI_CREATE_REQ(zdev->fh, 0, ZPCI_MOD_FC_REG_INT); |
| struct zpci_fib fib = {}; |
| u8 status; |
| |
| fib.fmt0.isc = zdev->kzdev->fib.fmt0.isc; |
| fib.fmt0.sum = 1; /* enable summary notifications */ |
| fib.fmt0.noi = airq_iv_end(zdev->aibv); |
| fib.fmt0.aibv = virt_to_phys(zdev->aibv->vector); |
| fib.fmt0.aibvo = 0; |
| fib.fmt0.aisb = virt_to_phys(aift->sbv->vector + (zdev->aisb / 64) * 8); |
| fib.fmt0.aisbo = zdev->aisb & 63; |
| fib.gd = zdev->gisa; |
| |
| return zpci_mod_fc(req, &fib, &status) ? -EIO : 0; |
| } |
| |
| /* Modify PCI: Unregister floating adapter interruption forwarding */ |
| static int kvm_zpci_clear_airq(struct zpci_dev *zdev) |
| { |
| u64 req = ZPCI_CREATE_REQ(zdev->fh, 0, ZPCI_MOD_FC_DEREG_INT); |
| struct zpci_fib fib = {}; |
| u8 cc, status; |
| |
| fib.gd = zdev->gisa; |
| |
| cc = zpci_mod_fc(req, &fib, &status); |
| if (cc == 3 || (cc == 1 && status == 24)) |
| /* Function already gone or IRQs already deregistered. */ |
| cc = 0; |
| |
| return cc ? -EIO : 0; |
| } |
| |
| static inline void unaccount_mem(unsigned long nr_pages) |
| { |
| struct user_struct *user = get_uid(current_user()); |
| |
| if (user) |
| atomic_long_sub(nr_pages, &user->locked_vm); |
| if (current->mm) |
| atomic64_sub(nr_pages, ¤t->mm->pinned_vm); |
| } |
| |
| static inline int account_mem(unsigned long nr_pages) |
| { |
| struct user_struct *user = get_uid(current_user()); |
| unsigned long page_limit, cur_pages, new_pages; |
| |
| page_limit = rlimit(RLIMIT_MEMLOCK) >> PAGE_SHIFT; |
| |
| do { |
| cur_pages = atomic_long_read(&user->locked_vm); |
| new_pages = cur_pages + nr_pages; |
| if (new_pages > page_limit) |
| return -ENOMEM; |
| } while (atomic_long_cmpxchg(&user->locked_vm, cur_pages, |
| new_pages) != cur_pages); |
| |
| atomic64_add(nr_pages, ¤t->mm->pinned_vm); |
| |
| return 0; |
| } |
| |
| static int kvm_s390_pci_aif_enable(struct zpci_dev *zdev, struct zpci_fib *fib, |
| bool assist) |
| { |
| struct page *pages[1], *aibv_page, *aisb_page = NULL; |
| unsigned int msi_vecs, idx; |
| struct zpci_gaite *gaite; |
| unsigned long hva, bit; |
| struct kvm *kvm; |
| phys_addr_t gaddr; |
| int rc = 0, gisc, npages, pcount = 0; |
| |
| /* |
| * Interrupt forwarding is only applicable if the device is already |
| * enabled for interpretation |
| */ |
| if (zdev->gisa == 0) |
| return -EINVAL; |
| |
| kvm = zdev->kzdev->kvm; |
| msi_vecs = min_t(unsigned int, fib->fmt0.noi, zdev->max_msi); |
| |
| /* Get the associated forwarding ISC - if invalid, return the error */ |
| gisc = kvm_s390_gisc_register(kvm, fib->fmt0.isc); |
| if (gisc < 0) |
| return gisc; |
| |
| /* Replace AIBV address */ |
| idx = srcu_read_lock(&kvm->srcu); |
| hva = gfn_to_hva(kvm, gpa_to_gfn((gpa_t)fib->fmt0.aibv)); |
| npages = pin_user_pages_fast(hva, 1, FOLL_WRITE | FOLL_LONGTERM, pages); |
| srcu_read_unlock(&kvm->srcu, idx); |
| if (npages < 1) { |
| rc = -EIO; |
| goto out; |
| } |
| aibv_page = pages[0]; |
| pcount++; |
| gaddr = page_to_phys(aibv_page) + (fib->fmt0.aibv & ~PAGE_MASK); |
| fib->fmt0.aibv = gaddr; |
| |
| /* Pin the guest AISB if one was specified */ |
| if (fib->fmt0.sum == 1) { |
| idx = srcu_read_lock(&kvm->srcu); |
| hva = gfn_to_hva(kvm, gpa_to_gfn((gpa_t)fib->fmt0.aisb)); |
| npages = pin_user_pages_fast(hva, 1, FOLL_WRITE | FOLL_LONGTERM, |
| pages); |
| srcu_read_unlock(&kvm->srcu, idx); |
| if (npages < 1) { |
| rc = -EIO; |
| goto unpin1; |
| } |
| aisb_page = pages[0]; |
| pcount++; |
| } |
| |
| /* Account for pinned pages, roll back on failure */ |
| if (account_mem(pcount)) |
| goto unpin2; |
| |
| /* AISB must be allocated before we can fill in GAITE */ |
| mutex_lock(&aift->aift_lock); |
| bit = airq_iv_alloc_bit(aift->sbv); |
| if (bit == -1UL) |
| goto unlock; |
| zdev->aisb = bit; /* store the summary bit number */ |
| zdev->aibv = airq_iv_create(msi_vecs, AIRQ_IV_DATA | |
| AIRQ_IV_BITLOCK | |
| AIRQ_IV_GUESTVEC, |
| phys_to_virt(fib->fmt0.aibv)); |
| |
| spin_lock_irq(&aift->gait_lock); |
| gaite = (struct zpci_gaite *)aift->gait + (zdev->aisb * |
| sizeof(struct zpci_gaite)); |
| |
| /* If assist not requested, host will get all alerts */ |
| if (assist) |
| gaite->gisa = (u32)virt_to_phys(&kvm->arch.sie_page2->gisa); |
| else |
| gaite->gisa = 0; |
| |
| gaite->gisc = fib->fmt0.isc; |
| gaite->count++; |
| gaite->aisbo = fib->fmt0.aisbo; |
| gaite->aisb = virt_to_phys(page_address(aisb_page) + (fib->fmt0.aisb & |
| ~PAGE_MASK)); |
| aift->kzdev[zdev->aisb] = zdev->kzdev; |
| spin_unlock_irq(&aift->gait_lock); |
| |
| /* Update guest FIB for re-issue */ |
| fib->fmt0.aisbo = zdev->aisb & 63; |
| fib->fmt0.aisb = virt_to_phys(aift->sbv->vector + (zdev->aisb / 64) * 8); |
| fib->fmt0.isc = gisc; |
| |
| /* Save some guest fib values in the host for later use */ |
| zdev->kzdev->fib.fmt0.isc = fib->fmt0.isc; |
| zdev->kzdev->fib.fmt0.aibv = fib->fmt0.aibv; |
| mutex_unlock(&aift->aift_lock); |
| |
| /* Issue the clp to setup the irq now */ |
| rc = kvm_zpci_set_airq(zdev); |
| return rc; |
| |
| unlock: |
| mutex_unlock(&aift->aift_lock); |
| unpin2: |
| if (fib->fmt0.sum == 1) |
| unpin_user_page(aisb_page); |
| unpin1: |
| unpin_user_page(aibv_page); |
| out: |
| return rc; |
| } |
| |
| static int kvm_s390_pci_aif_disable(struct zpci_dev *zdev, bool force) |
| { |
| struct kvm_zdev *kzdev = zdev->kzdev; |
| struct zpci_gaite *gaite; |
| struct page *vpage = NULL, *spage = NULL; |
| int rc, pcount = 0; |
| u8 isc; |
| |
| if (zdev->gisa == 0) |
| return -EINVAL; |
| |
| mutex_lock(&aift->aift_lock); |
| |
| /* |
| * If the clear fails due to an error, leave now unless we know this |
| * device is about to go away (force) -- In that case clear the GAITE |
| * regardless. |
| */ |
| rc = kvm_zpci_clear_airq(zdev); |
| if (rc && !force) |
| goto out; |
| |
| if (zdev->kzdev->fib.fmt0.aibv == 0) |
| goto out; |
| spin_lock_irq(&aift->gait_lock); |
| gaite = (struct zpci_gaite *)aift->gait + (zdev->aisb * |
| sizeof(struct zpci_gaite)); |
| isc = gaite->gisc; |
| gaite->count--; |
| if (gaite->count == 0) { |
| /* Release guest AIBV and AISB */ |
| vpage = phys_to_page(kzdev->fib.fmt0.aibv); |
| if (gaite->aisb != 0) |
| spage = phys_to_page(gaite->aisb); |
| /* Clear the GAIT entry */ |
| gaite->aisb = 0; |
| gaite->gisc = 0; |
| gaite->aisbo = 0; |
| gaite->gisa = 0; |
| aift->kzdev[zdev->aisb] = NULL; |
| /* Clear zdev info */ |
| airq_iv_free_bit(aift->sbv, zdev->aisb); |
| airq_iv_release(zdev->aibv); |
| zdev->aisb = 0; |
| zdev->aibv = NULL; |
| } |
| spin_unlock_irq(&aift->gait_lock); |
| kvm_s390_gisc_unregister(kzdev->kvm, isc); |
| kzdev->fib.fmt0.isc = 0; |
| kzdev->fib.fmt0.aibv = 0; |
| |
| if (vpage) { |
| unpin_user_page(vpage); |
| pcount++; |
| } |
| if (spage) { |
| unpin_user_page(spage); |
| pcount++; |
| } |
| if (pcount > 0) |
| unaccount_mem(pcount); |
| out: |
| mutex_unlock(&aift->aift_lock); |
| |
| return rc; |
| } |
| |
| static int kvm_s390_pci_dev_open(struct zpci_dev *zdev) |
| { |
| struct kvm_zdev *kzdev; |
| |
| kzdev = kzalloc(sizeof(struct kvm_zdev), GFP_KERNEL); |
| if (!kzdev) |
| return -ENOMEM; |
| |
| kzdev->zdev = zdev; |
| zdev->kzdev = kzdev; |
| |
| return 0; |
| } |
| |
| static void kvm_s390_pci_dev_release(struct zpci_dev *zdev) |
| { |
| struct kvm_zdev *kzdev; |
| |
| kzdev = zdev->kzdev; |
| WARN_ON(kzdev->zdev != zdev); |
| zdev->kzdev = NULL; |
| kfree(kzdev); |
| } |
| |
| |
| /* |
| * Register device with the specified KVM. If interpretation facilities are |
| * available, enable them and let userspace indicate whether or not they will |
| * be used (specify SHM bit to disable). |
| */ |
| static int kvm_s390_pci_register_kvm(void *opaque, struct kvm *kvm) |
| { |
| struct zpci_dev *zdev = opaque; |
| u8 status; |
| int rc; |
| |
| if (!zdev) |
| return -EINVAL; |
| |
| mutex_lock(&zdev->kzdev_lock); |
| |
| if (zdev->kzdev || zdev->gisa != 0 || !kvm) { |
| mutex_unlock(&zdev->kzdev_lock); |
| return -EINVAL; |
| } |
| |
| kvm_get_kvm(kvm); |
| |
| mutex_lock(&kvm->lock); |
| |
| rc = kvm_s390_pci_dev_open(zdev); |
| if (rc) |
| goto err; |
| |
| /* |
| * If interpretation facilities aren't available, add the device to |
| * the kzdev list but don't enable for interpretation. |
| */ |
| if (!kvm_s390_pci_interp_allowed()) |
| goto out; |
| |
| /* |
| * If this is the first request to use an interpreted device, make the |
| * necessary vcpu changes |
| */ |
| if (!kvm->arch.use_zpci_interp) |
| kvm_s390_vcpu_pci_enable_interp(kvm); |
| |
| if (zdev_enabled(zdev)) { |
| rc = zpci_disable_device(zdev); |
| if (rc) |
| goto err; |
| } |
| |
| /* |
| * Store information about the identity of the kvm guest allowed to |
| * access this device via interpretation to be used by host CLP |
| */ |
| zdev->gisa = (u32)virt_to_phys(&kvm->arch.sie_page2->gisa); |
| |
| rc = zpci_enable_device(zdev); |
| if (rc) |
| goto clear_gisa; |
| |
| /* Re-register the IOMMU that was already created */ |
| rc = zpci_register_ioat(zdev, 0, zdev->start_dma, zdev->end_dma, |
| virt_to_phys(zdev->dma_table), &status); |
| if (rc) |
| goto clear_gisa; |
| |
| out: |
| zdev->kzdev->kvm = kvm; |
| |
| spin_lock(&kvm->arch.kzdev_list_lock); |
| list_add_tail(&zdev->kzdev->entry, &kvm->arch.kzdev_list); |
| spin_unlock(&kvm->arch.kzdev_list_lock); |
| |
| mutex_unlock(&kvm->lock); |
| mutex_unlock(&zdev->kzdev_lock); |
| return 0; |
| |
| clear_gisa: |
| zdev->gisa = 0; |
| err: |
| if (zdev->kzdev) |
| kvm_s390_pci_dev_release(zdev); |
| mutex_unlock(&kvm->lock); |
| mutex_unlock(&zdev->kzdev_lock); |
| kvm_put_kvm(kvm); |
| return rc; |
| } |
| |
| static void kvm_s390_pci_unregister_kvm(void *opaque) |
| { |
| struct zpci_dev *zdev = opaque; |
| struct kvm *kvm; |
| u8 status; |
| |
| if (!zdev) |
| return; |
| |
| mutex_lock(&zdev->kzdev_lock); |
| |
| if (WARN_ON(!zdev->kzdev)) { |
| mutex_unlock(&zdev->kzdev_lock); |
| return; |
| } |
| |
| kvm = zdev->kzdev->kvm; |
| mutex_lock(&kvm->lock); |
| |
| /* |
| * A 0 gisa means interpretation was never enabled, just remove the |
| * device from the list. |
| */ |
| if (zdev->gisa == 0) |
| goto out; |
| |
| /* Forwarding must be turned off before interpretation */ |
| if (zdev->kzdev->fib.fmt0.aibv != 0) |
| kvm_s390_pci_aif_disable(zdev, true); |
| |
| /* Remove the host CLP guest designation */ |
| zdev->gisa = 0; |
| |
| if (zdev_enabled(zdev)) { |
| if (zpci_disable_device(zdev)) |
| goto out; |
| } |
| |
| if (zpci_enable_device(zdev)) |
| goto out; |
| |
| /* Re-register the IOMMU that was already created */ |
| zpci_register_ioat(zdev, 0, zdev->start_dma, zdev->end_dma, |
| virt_to_phys(zdev->dma_table), &status); |
| |
| out: |
| spin_lock(&kvm->arch.kzdev_list_lock); |
| list_del(&zdev->kzdev->entry); |
| spin_unlock(&kvm->arch.kzdev_list_lock); |
| kvm_s390_pci_dev_release(zdev); |
| |
| mutex_unlock(&kvm->lock); |
| mutex_unlock(&zdev->kzdev_lock); |
| |
| kvm_put_kvm(kvm); |
| } |
| |
| void kvm_s390_pci_init_list(struct kvm *kvm) |
| { |
| spin_lock_init(&kvm->arch.kzdev_list_lock); |
| INIT_LIST_HEAD(&kvm->arch.kzdev_list); |
| } |
| |
| void kvm_s390_pci_clear_list(struct kvm *kvm) |
| { |
| /* |
| * This list should already be empty, either via vfio device closures |
| * or kvm fd cleanup. |
| */ |
| spin_lock(&kvm->arch.kzdev_list_lock); |
| WARN_ON_ONCE(!list_empty(&kvm->arch.kzdev_list)); |
| spin_unlock(&kvm->arch.kzdev_list_lock); |
| } |
| |
| static struct zpci_dev *get_zdev_from_kvm_by_fh(struct kvm *kvm, u32 fh) |
| { |
| struct zpci_dev *zdev = NULL; |
| struct kvm_zdev *kzdev; |
| |
| spin_lock(&kvm->arch.kzdev_list_lock); |
| list_for_each_entry(kzdev, &kvm->arch.kzdev_list, entry) { |
| if (kzdev->zdev->fh == fh) { |
| zdev = kzdev->zdev; |
| break; |
| } |
| } |
| spin_unlock(&kvm->arch.kzdev_list_lock); |
| |
| return zdev; |
| } |
| |
| static int kvm_s390_pci_zpci_reg_aen(struct zpci_dev *zdev, |
| struct kvm_s390_zpci_op *args) |
| { |
| struct zpci_fib fib = {}; |
| bool hostflag; |
| |
| fib.fmt0.aibv = args->u.reg_aen.ibv; |
| fib.fmt0.isc = args->u.reg_aen.isc; |
| fib.fmt0.noi = args->u.reg_aen.noi; |
| if (args->u.reg_aen.sb != 0) { |
| fib.fmt0.aisb = args->u.reg_aen.sb; |
| fib.fmt0.aisbo = args->u.reg_aen.sbo; |
| fib.fmt0.sum = 1; |
| } else { |
| fib.fmt0.aisb = 0; |
| fib.fmt0.aisbo = 0; |
| fib.fmt0.sum = 0; |
| } |
| |
| hostflag = !(args->u.reg_aen.flags & KVM_S390_ZPCIOP_REGAEN_HOST); |
| return kvm_s390_pci_aif_enable(zdev, &fib, hostflag); |
| } |
| |
| int kvm_s390_pci_zpci_op(struct kvm *kvm, struct kvm_s390_zpci_op *args) |
| { |
| struct kvm_zdev *kzdev; |
| struct zpci_dev *zdev; |
| int r; |
| |
| zdev = get_zdev_from_kvm_by_fh(kvm, args->fh); |
| if (!zdev) |
| return -ENODEV; |
| |
| mutex_lock(&zdev->kzdev_lock); |
| mutex_lock(&kvm->lock); |
| |
| kzdev = zdev->kzdev; |
| if (!kzdev) { |
| r = -ENODEV; |
| goto out; |
| } |
| if (kzdev->kvm != kvm) { |
| r = -EPERM; |
| goto out; |
| } |
| |
| switch (args->op) { |
| case KVM_S390_ZPCIOP_REG_AEN: |
| /* Fail on unknown flags */ |
| if (args->u.reg_aen.flags & ~KVM_S390_ZPCIOP_REGAEN_HOST) { |
| r = -EINVAL; |
| break; |
| } |
| r = kvm_s390_pci_zpci_reg_aen(zdev, args); |
| break; |
| case KVM_S390_ZPCIOP_DEREG_AEN: |
| r = kvm_s390_pci_aif_disable(zdev, false); |
| break; |
| default: |
| r = -EINVAL; |
| } |
| |
| out: |
| mutex_unlock(&kvm->lock); |
| mutex_unlock(&zdev->kzdev_lock); |
| return r; |
| } |
| |
| int __init kvm_s390_pci_init(void) |
| { |
| zpci_kvm_hook.kvm_register = kvm_s390_pci_register_kvm; |
| zpci_kvm_hook.kvm_unregister = kvm_s390_pci_unregister_kvm; |
| |
| if (!kvm_s390_pci_interp_allowed()) |
| return 0; |
| |
| aift = kzalloc(sizeof(struct zpci_aift), GFP_KERNEL); |
| if (!aift) |
| return -ENOMEM; |
| |
| spin_lock_init(&aift->gait_lock); |
| mutex_init(&aift->aift_lock); |
| |
| return 0; |
| } |
| |
| void kvm_s390_pci_exit(void) |
| { |
| zpci_kvm_hook.kvm_register = NULL; |
| zpci_kvm_hook.kvm_unregister = NULL; |
| |
| if (!kvm_s390_pci_interp_allowed()) |
| return; |
| |
| mutex_destroy(&aift->aift_lock); |
| |
| kfree(aift); |
| } |