| // SPDX-License-Identifier: GPL-2.0-only |
| /* |
| * VFIO: IOMMU DMA mapping support for TCE on POWER |
| * |
| * Copyright (C) 2013 IBM Corp. All rights reserved. |
| * Author: Alexey Kardashevskiy <aik@ozlabs.ru> |
| * Copyright Gavin Shan, IBM Corporation 2014. |
| * |
| * Derived from original vfio_iommu_type1.c: |
| * Copyright (C) 2012 Red Hat, Inc. All rights reserved. |
| * Author: Alex Williamson <alex.williamson@redhat.com> |
| */ |
| |
| #include <linux/module.h> |
| #include <linux/pci.h> |
| #include <linux/slab.h> |
| #include <linux/uaccess.h> |
| #include <linux/err.h> |
| #include <linux/vfio.h> |
| #include <linux/vmalloc.h> |
| #include <linux/sched/mm.h> |
| #include <linux/sched/signal.h> |
| #include <linux/mm.h> |
| #include "vfio.h" |
| |
| #include <asm/iommu.h> |
| #include <asm/tce.h> |
| #include <asm/mmu_context.h> |
| |
| #define DRIVER_VERSION "0.1" |
| #define DRIVER_AUTHOR "aik@ozlabs.ru" |
| #define DRIVER_DESC "VFIO IOMMU SPAPR TCE" |
| |
| static void tce_iommu_detach_group(void *iommu_data, |
| struct iommu_group *iommu_group); |
| |
| /* |
| * VFIO IOMMU fd for SPAPR_TCE IOMMU implementation |
| * |
| * This code handles mapping and unmapping of user data buffers |
| * into DMA'ble space using the IOMMU |
| */ |
| |
| struct tce_iommu_group { |
| struct list_head next; |
| struct iommu_group *grp; |
| }; |
| |
| /* |
| * A container needs to remember which preregistered region it has |
| * referenced to do proper cleanup at the userspace process exit. |
| */ |
| struct tce_iommu_prereg { |
| struct list_head next; |
| struct mm_iommu_table_group_mem_t *mem; |
| }; |
| |
| /* |
| * The container descriptor supports only a single group per container. |
| * Required by the API as the container is not supplied with the IOMMU group |
| * at the moment of initialization. |
| */ |
| struct tce_container { |
| struct mutex lock; |
| bool enabled; |
| bool v2; |
| bool def_window_pending; |
| unsigned long locked_pages; |
| struct mm_struct *mm; |
| struct iommu_table *tables[IOMMU_TABLE_GROUP_MAX_TABLES]; |
| struct list_head group_list; |
| struct list_head prereg_list; |
| }; |
| |
| static long tce_iommu_mm_set(struct tce_container *container) |
| { |
| if (container->mm) { |
| if (container->mm == current->mm) |
| return 0; |
| return -EPERM; |
| } |
| BUG_ON(!current->mm); |
| container->mm = current->mm; |
| mmgrab(container->mm); |
| |
| return 0; |
| } |
| |
| static long tce_iommu_prereg_free(struct tce_container *container, |
| struct tce_iommu_prereg *tcemem) |
| { |
| long ret; |
| |
| ret = mm_iommu_put(container->mm, tcemem->mem); |
| if (ret) |
| return ret; |
| |
| list_del(&tcemem->next); |
| kfree(tcemem); |
| |
| return 0; |
| } |
| |
| static long tce_iommu_unregister_pages(struct tce_container *container, |
| __u64 vaddr, __u64 size) |
| { |
| struct mm_iommu_table_group_mem_t *mem; |
| struct tce_iommu_prereg *tcemem; |
| bool found = false; |
| long ret; |
| |
| if ((vaddr & ~PAGE_MASK) || (size & ~PAGE_MASK)) |
| return -EINVAL; |
| |
| mem = mm_iommu_get(container->mm, vaddr, size >> PAGE_SHIFT); |
| if (!mem) |
| return -ENOENT; |
| |
| list_for_each_entry(tcemem, &container->prereg_list, next) { |
| if (tcemem->mem == mem) { |
| found = true; |
| break; |
| } |
| } |
| |
| if (!found) |
| ret = -ENOENT; |
| else |
| ret = tce_iommu_prereg_free(container, tcemem); |
| |
| mm_iommu_put(container->mm, mem); |
| |
| return ret; |
| } |
| |
| static long tce_iommu_register_pages(struct tce_container *container, |
| __u64 vaddr, __u64 size) |
| { |
| long ret = 0; |
| struct mm_iommu_table_group_mem_t *mem = NULL; |
| struct tce_iommu_prereg *tcemem; |
| unsigned long entries = size >> PAGE_SHIFT; |
| |
| if ((vaddr & ~PAGE_MASK) || (size & ~PAGE_MASK) || |
| ((vaddr + size) < vaddr)) |
| return -EINVAL; |
| |
| mem = mm_iommu_get(container->mm, vaddr, entries); |
| if (mem) { |
| list_for_each_entry(tcemem, &container->prereg_list, next) { |
| if (tcemem->mem == mem) { |
| ret = -EBUSY; |
| goto put_exit; |
| } |
| } |
| } else { |
| ret = mm_iommu_new(container->mm, vaddr, entries, &mem); |
| if (ret) |
| return ret; |
| } |
| |
| tcemem = kzalloc(sizeof(*tcemem), GFP_KERNEL); |
| if (!tcemem) { |
| ret = -ENOMEM; |
| goto put_exit; |
| } |
| |
| tcemem->mem = mem; |
| list_add(&tcemem->next, &container->prereg_list); |
| |
| container->enabled = true; |
| |
| return 0; |
| |
| put_exit: |
| mm_iommu_put(container->mm, mem); |
| return ret; |
| } |
| |
| static bool tce_page_is_contained(struct mm_struct *mm, unsigned long hpa, |
| unsigned int it_page_shift) |
| { |
| struct page *page; |
| unsigned long size = 0; |
| |
| if (mm_iommu_is_devmem(mm, hpa, it_page_shift, &size)) |
| return size == (1UL << it_page_shift); |
| |
| page = pfn_to_page(hpa >> PAGE_SHIFT); |
| /* |
| * Check that the TCE table granularity is not bigger than the size of |
| * a page we just found. Otherwise the hardware can get access to |
| * a bigger memory chunk that it should. |
| */ |
| return page_shift(compound_head(page)) >= it_page_shift; |
| } |
| |
| static inline bool tce_groups_attached(struct tce_container *container) |
| { |
| return !list_empty(&container->group_list); |
| } |
| |
| static long tce_iommu_find_table(struct tce_container *container, |
| phys_addr_t ioba, struct iommu_table **ptbl) |
| { |
| long i; |
| |
| for (i = 0; i < IOMMU_TABLE_GROUP_MAX_TABLES; ++i) { |
| struct iommu_table *tbl = container->tables[i]; |
| |
| if (tbl) { |
| unsigned long entry = ioba >> tbl->it_page_shift; |
| unsigned long start = tbl->it_offset; |
| unsigned long end = start + tbl->it_size; |
| |
| if ((start <= entry) && (entry < end)) { |
| *ptbl = tbl; |
| return i; |
| } |
| } |
| } |
| |
| return -1; |
| } |
| |
| static int tce_iommu_find_free_table(struct tce_container *container) |
| { |
| int i; |
| |
| for (i = 0; i < IOMMU_TABLE_GROUP_MAX_TABLES; ++i) { |
| if (!container->tables[i]) |
| return i; |
| } |
| |
| return -ENOSPC; |
| } |
| |
| static int tce_iommu_enable(struct tce_container *container) |
| { |
| int ret = 0; |
| unsigned long locked; |
| struct iommu_table_group *table_group; |
| struct tce_iommu_group *tcegrp; |
| |
| if (container->enabled) |
| return -EBUSY; |
| |
| /* |
| * When userspace pages are mapped into the IOMMU, they are effectively |
| * locked memory, so, theoretically, we need to update the accounting |
| * of locked pages on each map and unmap. For powerpc, the map unmap |
| * paths can be very hot, though, and the accounting would kill |
| * performance, especially since it would be difficult to impossible |
| * to handle the accounting in real mode only. |
| * |
| * To address that, rather than precisely accounting every page, we |
| * instead account for a worst case on locked memory when the iommu is |
| * enabled and disabled. The worst case upper bound on locked memory |
| * is the size of the whole iommu window, which is usually relatively |
| * small (compared to total memory sizes) on POWER hardware. |
| * |
| * Also we don't have a nice way to fail on H_PUT_TCE due to ulimits, |
| * that would effectively kill the guest at random points, much better |
| * enforcing the limit based on the max that the guest can map. |
| * |
| * Unfortunately at the moment it counts whole tables, no matter how |
| * much memory the guest has. I.e. for 4GB guest and 4 IOMMU groups |
| * each with 2GB DMA window, 8GB will be counted here. The reason for |
| * this is that we cannot tell here the amount of RAM used by the guest |
| * as this information is only available from KVM and VFIO is |
| * KVM agnostic. |
| * |
| * So we do not allow enabling a container without a group attached |
| * as there is no way to know how much we should increment |
| * the locked_vm counter. |
| */ |
| if (!tce_groups_attached(container)) |
| return -ENODEV; |
| |
| tcegrp = list_first_entry(&container->group_list, |
| struct tce_iommu_group, next); |
| table_group = iommu_group_get_iommudata(tcegrp->grp); |
| if (!table_group) |
| return -ENODEV; |
| |
| if (!table_group->tce32_size) |
| return -EPERM; |
| |
| ret = tce_iommu_mm_set(container); |
| if (ret) |
| return ret; |
| |
| locked = table_group->tce32_size >> PAGE_SHIFT; |
| ret = account_locked_vm(container->mm, locked, true); |
| if (ret) |
| return ret; |
| |
| container->locked_pages = locked; |
| |
| container->enabled = true; |
| |
| return ret; |
| } |
| |
| static void tce_iommu_disable(struct tce_container *container) |
| { |
| if (!container->enabled) |
| return; |
| |
| container->enabled = false; |
| |
| BUG_ON(!container->mm); |
| account_locked_vm(container->mm, container->locked_pages, false); |
| } |
| |
| static void *tce_iommu_open(unsigned long arg) |
| { |
| struct tce_container *container; |
| |
| if ((arg != VFIO_SPAPR_TCE_IOMMU) && (arg != VFIO_SPAPR_TCE_v2_IOMMU)) { |
| pr_err("tce_vfio: Wrong IOMMU type\n"); |
| return ERR_PTR(-EINVAL); |
| } |
| |
| container = kzalloc(sizeof(*container), GFP_KERNEL); |
| if (!container) |
| return ERR_PTR(-ENOMEM); |
| |
| mutex_init(&container->lock); |
| INIT_LIST_HEAD_RCU(&container->group_list); |
| INIT_LIST_HEAD_RCU(&container->prereg_list); |
| |
| container->v2 = arg == VFIO_SPAPR_TCE_v2_IOMMU; |
| |
| return container; |
| } |
| |
| static int tce_iommu_clear(struct tce_container *container, |
| struct iommu_table *tbl, |
| unsigned long entry, unsigned long pages); |
| static void tce_iommu_free_table(struct tce_container *container, |
| struct iommu_table *tbl); |
| |
| static void tce_iommu_release(void *iommu_data) |
| { |
| struct tce_container *container = iommu_data; |
| struct tce_iommu_group *tcegrp; |
| struct tce_iommu_prereg *tcemem, *tmtmp; |
| long i; |
| |
| while (tce_groups_attached(container)) { |
| tcegrp = list_first_entry(&container->group_list, |
| struct tce_iommu_group, next); |
| tce_iommu_detach_group(iommu_data, tcegrp->grp); |
| } |
| |
| /* |
| * If VFIO created a table, it was not disposed |
| * by tce_iommu_detach_group() so do it now. |
| */ |
| for (i = 0; i < IOMMU_TABLE_GROUP_MAX_TABLES; ++i) { |
| struct iommu_table *tbl = container->tables[i]; |
| |
| if (!tbl) |
| continue; |
| |
| tce_iommu_free_table(container, tbl); |
| } |
| |
| list_for_each_entry_safe(tcemem, tmtmp, &container->prereg_list, next) |
| WARN_ON(tce_iommu_prereg_free(container, tcemem)); |
| |
| tce_iommu_disable(container); |
| if (container->mm) |
| mmdrop(container->mm); |
| mutex_destroy(&container->lock); |
| |
| kfree(container); |
| } |
| |
| static void tce_iommu_unuse_page(unsigned long hpa) |
| { |
| struct page *page; |
| |
| page = pfn_to_page(hpa >> PAGE_SHIFT); |
| unpin_user_page(page); |
| } |
| |
| static int tce_iommu_prereg_ua_to_hpa(struct tce_container *container, |
| unsigned long tce, unsigned long shift, |
| unsigned long *phpa, struct mm_iommu_table_group_mem_t **pmem) |
| { |
| long ret = 0; |
| struct mm_iommu_table_group_mem_t *mem; |
| |
| mem = mm_iommu_lookup(container->mm, tce, 1ULL << shift); |
| if (!mem) |
| return -EINVAL; |
| |
| ret = mm_iommu_ua_to_hpa(mem, tce, shift, phpa); |
| if (ret) |
| return -EINVAL; |
| |
| *pmem = mem; |
| |
| return 0; |
| } |
| |
| static void tce_iommu_unuse_page_v2(struct tce_container *container, |
| struct iommu_table *tbl, unsigned long entry) |
| { |
| struct mm_iommu_table_group_mem_t *mem = NULL; |
| int ret; |
| unsigned long hpa = 0; |
| __be64 *pua = IOMMU_TABLE_USERSPACE_ENTRY_RO(tbl, entry); |
| |
| if (!pua) |
| return; |
| |
| ret = tce_iommu_prereg_ua_to_hpa(container, be64_to_cpu(*pua), |
| tbl->it_page_shift, &hpa, &mem); |
| if (ret) |
| pr_debug("%s: tce %llx at #%lx was not cached, ret=%d\n", |
| __func__, be64_to_cpu(*pua), entry, ret); |
| if (mem) |
| mm_iommu_mapped_dec(mem); |
| |
| *pua = cpu_to_be64(0); |
| } |
| |
| static int tce_iommu_clear(struct tce_container *container, |
| struct iommu_table *tbl, |
| unsigned long entry, unsigned long pages) |
| { |
| unsigned long oldhpa; |
| long ret; |
| enum dma_data_direction direction; |
| unsigned long lastentry = entry + pages, firstentry = entry; |
| |
| for ( ; entry < lastentry; ++entry) { |
| if (tbl->it_indirect_levels && tbl->it_userspace) { |
| /* |
| * For multilevel tables, we can take a shortcut here |
| * and skip some TCEs as we know that the userspace |
| * addresses cache is a mirror of the real TCE table |
| * and if it is missing some indirect levels, then |
| * the hardware table does not have them allocated |
| * either and therefore does not require updating. |
| */ |
| __be64 *pua = IOMMU_TABLE_USERSPACE_ENTRY_RO(tbl, |
| entry); |
| if (!pua) { |
| /* align to level_size which is power of two */ |
| entry |= tbl->it_level_size - 1; |
| continue; |
| } |
| } |
| |
| cond_resched(); |
| |
| direction = DMA_NONE; |
| oldhpa = 0; |
| ret = iommu_tce_xchg_no_kill(container->mm, tbl, entry, &oldhpa, |
| &direction); |
| if (ret) |
| continue; |
| |
| if (direction == DMA_NONE) |
| continue; |
| |
| if (container->v2) { |
| tce_iommu_unuse_page_v2(container, tbl, entry); |
| continue; |
| } |
| |
| tce_iommu_unuse_page(oldhpa); |
| } |
| |
| iommu_tce_kill(tbl, firstentry, pages); |
| |
| return 0; |
| } |
| |
| static int tce_iommu_use_page(unsigned long tce, unsigned long *hpa) |
| { |
| struct page *page = NULL; |
| enum dma_data_direction direction = iommu_tce_direction(tce); |
| |
| if (pin_user_pages_fast(tce & PAGE_MASK, 1, |
| direction != DMA_TO_DEVICE ? FOLL_WRITE : 0, |
| &page) != 1) |
| return -EFAULT; |
| |
| *hpa = __pa((unsigned long) page_address(page)); |
| |
| return 0; |
| } |
| |
| static long tce_iommu_build(struct tce_container *container, |
| struct iommu_table *tbl, |
| unsigned long entry, unsigned long tce, unsigned long pages, |
| enum dma_data_direction direction) |
| { |
| long i, ret = 0; |
| unsigned long hpa; |
| enum dma_data_direction dirtmp; |
| |
| for (i = 0; i < pages; ++i) { |
| unsigned long offset = tce & IOMMU_PAGE_MASK(tbl) & ~PAGE_MASK; |
| |
| ret = tce_iommu_use_page(tce, &hpa); |
| if (ret) |
| break; |
| |
| if (!tce_page_is_contained(container->mm, hpa, |
| tbl->it_page_shift)) { |
| ret = -EPERM; |
| break; |
| } |
| |
| hpa |= offset; |
| dirtmp = direction; |
| ret = iommu_tce_xchg_no_kill(container->mm, tbl, entry + i, |
| &hpa, &dirtmp); |
| if (ret) { |
| tce_iommu_unuse_page(hpa); |
| pr_err("iommu_tce: %s failed ioba=%lx, tce=%lx, ret=%ld\n", |
| __func__, entry << tbl->it_page_shift, |
| tce, ret); |
| break; |
| } |
| |
| if (dirtmp != DMA_NONE) |
| tce_iommu_unuse_page(hpa); |
| |
| tce += IOMMU_PAGE_SIZE(tbl); |
| } |
| |
| if (ret) |
| tce_iommu_clear(container, tbl, entry, i); |
| else |
| iommu_tce_kill(tbl, entry, pages); |
| |
| return ret; |
| } |
| |
| static long tce_iommu_build_v2(struct tce_container *container, |
| struct iommu_table *tbl, |
| unsigned long entry, unsigned long tce, unsigned long pages, |
| enum dma_data_direction direction) |
| { |
| long i, ret = 0; |
| unsigned long hpa; |
| enum dma_data_direction dirtmp; |
| |
| for (i = 0; i < pages; ++i) { |
| struct mm_iommu_table_group_mem_t *mem = NULL; |
| __be64 *pua = IOMMU_TABLE_USERSPACE_ENTRY(tbl, entry + i); |
| |
| ret = tce_iommu_prereg_ua_to_hpa(container, |
| tce, tbl->it_page_shift, &hpa, &mem); |
| if (ret) |
| break; |
| |
| if (!tce_page_is_contained(container->mm, hpa, |
| tbl->it_page_shift)) { |
| ret = -EPERM; |
| break; |
| } |
| |
| /* Preserve offset within IOMMU page */ |
| hpa |= tce & IOMMU_PAGE_MASK(tbl) & ~PAGE_MASK; |
| dirtmp = direction; |
| |
| /* The registered region is being unregistered */ |
| if (mm_iommu_mapped_inc(mem)) |
| break; |
| |
| ret = iommu_tce_xchg_no_kill(container->mm, tbl, entry + i, |
| &hpa, &dirtmp); |
| if (ret) { |
| /* dirtmp cannot be DMA_NONE here */ |
| tce_iommu_unuse_page_v2(container, tbl, entry + i); |
| pr_err("iommu_tce: %s failed ioba=%lx, tce=%lx, ret=%ld\n", |
| __func__, entry << tbl->it_page_shift, |
| tce, ret); |
| break; |
| } |
| |
| if (dirtmp != DMA_NONE) |
| tce_iommu_unuse_page_v2(container, tbl, entry + i); |
| |
| *pua = cpu_to_be64(tce); |
| |
| tce += IOMMU_PAGE_SIZE(tbl); |
| } |
| |
| if (ret) |
| tce_iommu_clear(container, tbl, entry, i); |
| else |
| iommu_tce_kill(tbl, entry, pages); |
| |
| return ret; |
| } |
| |
| static long tce_iommu_create_table(struct tce_container *container, |
| struct iommu_table_group *table_group, |
| int num, |
| __u32 page_shift, |
| __u64 window_size, |
| __u32 levels, |
| struct iommu_table **ptbl) |
| { |
| long ret, table_size; |
| |
| table_size = table_group->ops->get_table_size(page_shift, window_size, |
| levels); |
| if (!table_size) |
| return -EINVAL; |
| |
| ret = account_locked_vm(container->mm, table_size >> PAGE_SHIFT, true); |
| if (ret) |
| return ret; |
| |
| ret = table_group->ops->create_table(table_group, num, |
| page_shift, window_size, levels, ptbl); |
| |
| WARN_ON(!ret && !(*ptbl)->it_ops->free); |
| WARN_ON(!ret && ((*ptbl)->it_allocated_size > table_size)); |
| |
| return ret; |
| } |
| |
| static void tce_iommu_free_table(struct tce_container *container, |
| struct iommu_table *tbl) |
| { |
| unsigned long pages = tbl->it_allocated_size >> PAGE_SHIFT; |
| |
| iommu_tce_table_put(tbl); |
| account_locked_vm(container->mm, pages, false); |
| } |
| |
| static long tce_iommu_create_window(struct tce_container *container, |
| __u32 page_shift, __u64 window_size, __u32 levels, |
| __u64 *start_addr) |
| { |
| struct tce_iommu_group *tcegrp; |
| struct iommu_table_group *table_group; |
| struct iommu_table *tbl = NULL; |
| long ret, num; |
| |
| num = tce_iommu_find_free_table(container); |
| if (num < 0) |
| return num; |
| |
| /* Get the first group for ops::create_table */ |
| tcegrp = list_first_entry(&container->group_list, |
| struct tce_iommu_group, next); |
| table_group = iommu_group_get_iommudata(tcegrp->grp); |
| if (!table_group) |
| return -EFAULT; |
| |
| if (!(table_group->pgsizes & (1ULL << page_shift))) |
| return -EINVAL; |
| |
| if (!table_group->ops->set_window || !table_group->ops->unset_window || |
| !table_group->ops->get_table_size || |
| !table_group->ops->create_table) |
| return -EPERM; |
| |
| /* Create TCE table */ |
| ret = tce_iommu_create_table(container, table_group, num, |
| page_shift, window_size, levels, &tbl); |
| if (ret) |
| return ret; |
| |
| BUG_ON(!tbl->it_ops->free); |
| |
| /* |
| * Program the table to every group. |
| * Groups have been tested for compatibility at the attach time. |
| */ |
| list_for_each_entry(tcegrp, &container->group_list, next) { |
| table_group = iommu_group_get_iommudata(tcegrp->grp); |
| |
| ret = table_group->ops->set_window(table_group, num, tbl); |
| if (ret) |
| goto unset_exit; |
| } |
| |
| container->tables[num] = tbl; |
| |
| /* Return start address assigned by platform in create_table() */ |
| *start_addr = tbl->it_offset << tbl->it_page_shift; |
| |
| return 0; |
| |
| unset_exit: |
| list_for_each_entry(tcegrp, &container->group_list, next) { |
| table_group = iommu_group_get_iommudata(tcegrp->grp); |
| table_group->ops->unset_window(table_group, num); |
| } |
| tce_iommu_free_table(container, tbl); |
| |
| return ret; |
| } |
| |
| static long tce_iommu_remove_window(struct tce_container *container, |
| __u64 start_addr) |
| { |
| struct iommu_table_group *table_group = NULL; |
| struct iommu_table *tbl; |
| struct tce_iommu_group *tcegrp; |
| int num; |
| |
| num = tce_iommu_find_table(container, start_addr, &tbl); |
| if (num < 0) |
| return -EINVAL; |
| |
| BUG_ON(!tbl->it_size); |
| |
| tce_iommu_clear(container, tbl, tbl->it_offset, tbl->it_size); |
| |
| /* Detach groups from IOMMUs */ |
| list_for_each_entry(tcegrp, &container->group_list, next) { |
| table_group = iommu_group_get_iommudata(tcegrp->grp); |
| |
| /* |
| * SPAPR TCE IOMMU exposes the default DMA window to |
| * the guest via dma32_window_start/size of |
| * VFIO_IOMMU_SPAPR_TCE_GET_INFO. Some platforms allow |
| * the userspace to remove this window, some do not so |
| * here we check for the platform capability. |
| */ |
| if (!table_group->ops || !table_group->ops->unset_window) |
| return -EPERM; |
| |
| table_group->ops->unset_window(table_group, num); |
| } |
| |
| /* Free table */ |
| tce_iommu_free_table(container, tbl); |
| container->tables[num] = NULL; |
| |
| return 0; |
| } |
| |
| static long tce_iommu_create_default_window(struct tce_container *container) |
| { |
| long ret; |
| __u64 start_addr = 0; |
| struct tce_iommu_group *tcegrp; |
| struct iommu_table_group *table_group; |
| |
| if (!container->def_window_pending) |
| return 0; |
| |
| if (!tce_groups_attached(container)) |
| return -ENODEV; |
| |
| tcegrp = list_first_entry(&container->group_list, |
| struct tce_iommu_group, next); |
| table_group = iommu_group_get_iommudata(tcegrp->grp); |
| if (!table_group) |
| return -ENODEV; |
| |
| ret = tce_iommu_create_window(container, IOMMU_PAGE_SHIFT_4K, |
| table_group->tce32_size, 1, &start_addr); |
| WARN_ON_ONCE(!ret && start_addr); |
| |
| if (!ret) |
| container->def_window_pending = false; |
| |
| return ret; |
| } |
| |
| static long vfio_spapr_ioctl_eeh_pe_op(struct iommu_group *group, |
| unsigned long arg) |
| { |
| struct eeh_pe *pe; |
| struct vfio_eeh_pe_op op; |
| unsigned long minsz; |
| |
| pe = eeh_iommu_group_to_pe(group); |
| if (!pe) |
| return -ENODEV; |
| |
| minsz = offsetofend(struct vfio_eeh_pe_op, op); |
| if (copy_from_user(&op, (void __user *)arg, minsz)) |
| return -EFAULT; |
| if (op.argsz < minsz || op.flags) |
| return -EINVAL; |
| |
| switch (op.op) { |
| case VFIO_EEH_PE_DISABLE: |
| return eeh_pe_set_option(pe, EEH_OPT_DISABLE); |
| case VFIO_EEH_PE_ENABLE: |
| return eeh_pe_set_option(pe, EEH_OPT_ENABLE); |
| case VFIO_EEH_PE_UNFREEZE_IO: |
| return eeh_pe_set_option(pe, EEH_OPT_THAW_MMIO); |
| case VFIO_EEH_PE_UNFREEZE_DMA: |
| return eeh_pe_set_option(pe, EEH_OPT_THAW_DMA); |
| case VFIO_EEH_PE_GET_STATE: |
| return eeh_pe_get_state(pe); |
| break; |
| case VFIO_EEH_PE_RESET_DEACTIVATE: |
| return eeh_pe_reset(pe, EEH_RESET_DEACTIVATE, true); |
| case VFIO_EEH_PE_RESET_HOT: |
| return eeh_pe_reset(pe, EEH_RESET_HOT, true); |
| case VFIO_EEH_PE_RESET_FUNDAMENTAL: |
| return eeh_pe_reset(pe, EEH_RESET_FUNDAMENTAL, true); |
| case VFIO_EEH_PE_CONFIGURE: |
| return eeh_pe_configure(pe); |
| case VFIO_EEH_PE_INJECT_ERR: |
| minsz = offsetofend(struct vfio_eeh_pe_op, err.mask); |
| if (op.argsz < minsz) |
| return -EINVAL; |
| if (copy_from_user(&op, (void __user *)arg, minsz)) |
| return -EFAULT; |
| |
| return eeh_pe_inject_err(pe, op.err.type, op.err.func, |
| op.err.addr, op.err.mask); |
| default: |
| return -EINVAL; |
| } |
| } |
| |
| static long tce_iommu_ioctl(void *iommu_data, |
| unsigned int cmd, unsigned long arg) |
| { |
| struct tce_container *container = iommu_data; |
| unsigned long minsz, ddwsz; |
| long ret; |
| |
| switch (cmd) { |
| case VFIO_CHECK_EXTENSION: |
| switch (arg) { |
| case VFIO_SPAPR_TCE_IOMMU: |
| case VFIO_SPAPR_TCE_v2_IOMMU: |
| return 1; |
| case VFIO_EEH: |
| return eeh_enabled(); |
| default: |
| return 0; |
| } |
| } |
| |
| /* |
| * Sanity check to prevent one userspace from manipulating |
| * another userspace mm. |
| */ |
| BUG_ON(!container); |
| if (container->mm && container->mm != current->mm) |
| return -EPERM; |
| |
| switch (cmd) { |
| case VFIO_IOMMU_SPAPR_TCE_GET_INFO: { |
| struct vfio_iommu_spapr_tce_info info; |
| struct tce_iommu_group *tcegrp; |
| struct iommu_table_group *table_group; |
| |
| if (!tce_groups_attached(container)) |
| return -ENXIO; |
| |
| tcegrp = list_first_entry(&container->group_list, |
| struct tce_iommu_group, next); |
| table_group = iommu_group_get_iommudata(tcegrp->grp); |
| |
| if (!table_group) |
| return -ENXIO; |
| |
| minsz = offsetofend(struct vfio_iommu_spapr_tce_info, |
| dma32_window_size); |
| |
| if (copy_from_user(&info, (void __user *)arg, minsz)) |
| return -EFAULT; |
| |
| if (info.argsz < minsz) |
| return -EINVAL; |
| |
| info.dma32_window_start = table_group->tce32_start; |
| info.dma32_window_size = table_group->tce32_size; |
| info.flags = 0; |
| memset(&info.ddw, 0, sizeof(info.ddw)); |
| |
| if (table_group->max_dynamic_windows_supported && |
| container->v2) { |
| info.flags |= VFIO_IOMMU_SPAPR_INFO_DDW; |
| info.ddw.pgsizes = table_group->pgsizes; |
| info.ddw.max_dynamic_windows_supported = |
| table_group->max_dynamic_windows_supported; |
| info.ddw.levels = table_group->max_levels; |
| } |
| |
| ddwsz = offsetofend(struct vfio_iommu_spapr_tce_info, ddw); |
| |
| if (info.argsz >= ddwsz) |
| minsz = ddwsz; |
| |
| if (copy_to_user((void __user *)arg, &info, minsz)) |
| return -EFAULT; |
| |
| return 0; |
| } |
| case VFIO_IOMMU_MAP_DMA: { |
| struct vfio_iommu_type1_dma_map param; |
| struct iommu_table *tbl = NULL; |
| long num; |
| enum dma_data_direction direction; |
| |
| if (!container->enabled) |
| return -EPERM; |
| |
| minsz = offsetofend(struct vfio_iommu_type1_dma_map, size); |
| |
| if (copy_from_user(¶m, (void __user *)arg, minsz)) |
| return -EFAULT; |
| |
| if (param.argsz < minsz) |
| return -EINVAL; |
| |
| if (param.flags & ~(VFIO_DMA_MAP_FLAG_READ | |
| VFIO_DMA_MAP_FLAG_WRITE)) |
| return -EINVAL; |
| |
| ret = tce_iommu_create_default_window(container); |
| if (ret) |
| return ret; |
| |
| num = tce_iommu_find_table(container, param.iova, &tbl); |
| if (num < 0) |
| return -ENXIO; |
| |
| if ((param.size & ~IOMMU_PAGE_MASK(tbl)) || |
| (param.vaddr & ~IOMMU_PAGE_MASK(tbl))) |
| return -EINVAL; |
| |
| /* iova is checked by the IOMMU API */ |
| if (param.flags & VFIO_DMA_MAP_FLAG_READ) { |
| if (param.flags & VFIO_DMA_MAP_FLAG_WRITE) |
| direction = DMA_BIDIRECTIONAL; |
| else |
| direction = DMA_TO_DEVICE; |
| } else { |
| if (param.flags & VFIO_DMA_MAP_FLAG_WRITE) |
| direction = DMA_FROM_DEVICE; |
| else |
| return -EINVAL; |
| } |
| |
| ret = iommu_tce_put_param_check(tbl, param.iova, param.vaddr); |
| if (ret) |
| return ret; |
| |
| if (container->v2) |
| ret = tce_iommu_build_v2(container, tbl, |
| param.iova >> tbl->it_page_shift, |
| param.vaddr, |
| param.size >> tbl->it_page_shift, |
| direction); |
| else |
| ret = tce_iommu_build(container, tbl, |
| param.iova >> tbl->it_page_shift, |
| param.vaddr, |
| param.size >> tbl->it_page_shift, |
| direction); |
| |
| iommu_flush_tce(tbl); |
| |
| return ret; |
| } |
| case VFIO_IOMMU_UNMAP_DMA: { |
| struct vfio_iommu_type1_dma_unmap param; |
| struct iommu_table *tbl = NULL; |
| long num; |
| |
| if (!container->enabled) |
| return -EPERM; |
| |
| minsz = offsetofend(struct vfio_iommu_type1_dma_unmap, |
| size); |
| |
| if (copy_from_user(¶m, (void __user *)arg, minsz)) |
| return -EFAULT; |
| |
| if (param.argsz < minsz) |
| return -EINVAL; |
| |
| /* No flag is supported now */ |
| if (param.flags) |
| return -EINVAL; |
| |
| ret = tce_iommu_create_default_window(container); |
| if (ret) |
| return ret; |
| |
| num = tce_iommu_find_table(container, param.iova, &tbl); |
| if (num < 0) |
| return -ENXIO; |
| |
| if (param.size & ~IOMMU_PAGE_MASK(tbl)) |
| return -EINVAL; |
| |
| ret = iommu_tce_clear_param_check(tbl, param.iova, 0, |
| param.size >> tbl->it_page_shift); |
| if (ret) |
| return ret; |
| |
| ret = tce_iommu_clear(container, tbl, |
| param.iova >> tbl->it_page_shift, |
| param.size >> tbl->it_page_shift); |
| iommu_flush_tce(tbl); |
| |
| return ret; |
| } |
| case VFIO_IOMMU_SPAPR_REGISTER_MEMORY: { |
| struct vfio_iommu_spapr_register_memory param; |
| |
| if (!container->v2) |
| break; |
| |
| minsz = offsetofend(struct vfio_iommu_spapr_register_memory, |
| size); |
| |
| ret = tce_iommu_mm_set(container); |
| if (ret) |
| return ret; |
| |
| if (copy_from_user(¶m, (void __user *)arg, minsz)) |
| return -EFAULT; |
| |
| if (param.argsz < minsz) |
| return -EINVAL; |
| |
| /* No flag is supported now */ |
| if (param.flags) |
| return -EINVAL; |
| |
| mutex_lock(&container->lock); |
| ret = tce_iommu_register_pages(container, param.vaddr, |
| param.size); |
| mutex_unlock(&container->lock); |
| |
| return ret; |
| } |
| case VFIO_IOMMU_SPAPR_UNREGISTER_MEMORY: { |
| struct vfio_iommu_spapr_register_memory param; |
| |
| if (!container->v2) |
| break; |
| |
| if (!container->mm) |
| return -EPERM; |
| |
| minsz = offsetofend(struct vfio_iommu_spapr_register_memory, |
| size); |
| |
| if (copy_from_user(¶m, (void __user *)arg, minsz)) |
| return -EFAULT; |
| |
| if (param.argsz < minsz) |
| return -EINVAL; |
| |
| /* No flag is supported now */ |
| if (param.flags) |
| return -EINVAL; |
| |
| mutex_lock(&container->lock); |
| ret = tce_iommu_unregister_pages(container, param.vaddr, |
| param.size); |
| mutex_unlock(&container->lock); |
| |
| return ret; |
| } |
| case VFIO_IOMMU_ENABLE: |
| if (container->v2) |
| break; |
| |
| mutex_lock(&container->lock); |
| ret = tce_iommu_enable(container); |
| mutex_unlock(&container->lock); |
| return ret; |
| |
| |
| case VFIO_IOMMU_DISABLE: |
| if (container->v2) |
| break; |
| |
| mutex_lock(&container->lock); |
| tce_iommu_disable(container); |
| mutex_unlock(&container->lock); |
| return 0; |
| |
| case VFIO_EEH_PE_OP: { |
| struct tce_iommu_group *tcegrp; |
| |
| ret = 0; |
| list_for_each_entry(tcegrp, &container->group_list, next) { |
| ret = vfio_spapr_ioctl_eeh_pe_op(tcegrp->grp, arg); |
| if (ret) |
| return ret; |
| } |
| return ret; |
| } |
| |
| case VFIO_IOMMU_SPAPR_TCE_CREATE: { |
| struct vfio_iommu_spapr_tce_create create; |
| |
| if (!container->v2) |
| break; |
| |
| ret = tce_iommu_mm_set(container); |
| if (ret) |
| return ret; |
| |
| if (!tce_groups_attached(container)) |
| return -ENXIO; |
| |
| minsz = offsetofend(struct vfio_iommu_spapr_tce_create, |
| start_addr); |
| |
| if (copy_from_user(&create, (void __user *)arg, minsz)) |
| return -EFAULT; |
| |
| if (create.argsz < minsz) |
| return -EINVAL; |
| |
| if (create.flags) |
| return -EINVAL; |
| |
| mutex_lock(&container->lock); |
| |
| ret = tce_iommu_create_default_window(container); |
| if (!ret) |
| ret = tce_iommu_create_window(container, |
| create.page_shift, |
| create.window_size, create.levels, |
| &create.start_addr); |
| |
| mutex_unlock(&container->lock); |
| |
| if (!ret && copy_to_user((void __user *)arg, &create, minsz)) |
| ret = -EFAULT; |
| |
| return ret; |
| } |
| case VFIO_IOMMU_SPAPR_TCE_REMOVE: { |
| struct vfio_iommu_spapr_tce_remove remove; |
| |
| if (!container->v2) |
| break; |
| |
| ret = tce_iommu_mm_set(container); |
| if (ret) |
| return ret; |
| |
| if (!tce_groups_attached(container)) |
| return -ENXIO; |
| |
| minsz = offsetofend(struct vfio_iommu_spapr_tce_remove, |
| start_addr); |
| |
| if (copy_from_user(&remove, (void __user *)arg, minsz)) |
| return -EFAULT; |
| |
| if (remove.argsz < minsz) |
| return -EINVAL; |
| |
| if (remove.flags) |
| return -EINVAL; |
| |
| if (container->def_window_pending && !remove.start_addr) { |
| container->def_window_pending = false; |
| return 0; |
| } |
| |
| mutex_lock(&container->lock); |
| |
| ret = tce_iommu_remove_window(container, remove.start_addr); |
| |
| mutex_unlock(&container->lock); |
| |
| return ret; |
| } |
| } |
| |
| return -ENOTTY; |
| } |
| |
| static void tce_iommu_release_ownership(struct tce_container *container, |
| struct iommu_table_group *table_group) |
| { |
| long i; |
| |
| if (!table_group->ops->unset_window) { |
| WARN_ON_ONCE(1); |
| return; |
| } |
| |
| for (i = 0; i < IOMMU_TABLE_GROUP_MAX_TABLES; ++i) { |
| if (container->tables[i]) { |
| tce_iommu_clear(container, container->tables[i], |
| container->tables[i]->it_offset, |
| container->tables[i]->it_size); |
| table_group->ops->unset_window(table_group, i); |
| } |
| } |
| } |
| |
| static long tce_iommu_take_ownership(struct tce_container *container, |
| struct iommu_table_group *table_group) |
| { |
| long i, ret = 0; |
| |
| /* Set all windows to the new group */ |
| for (i = 0; i < IOMMU_TABLE_GROUP_MAX_TABLES; ++i) { |
| struct iommu_table *tbl = container->tables[i]; |
| |
| if (!tbl) |
| continue; |
| |
| ret = table_group->ops->set_window(table_group, i, tbl); |
| if (ret) |
| goto release_exit; |
| } |
| |
| return 0; |
| |
| release_exit: |
| for (i = 0; i < IOMMU_TABLE_GROUP_MAX_TABLES; ++i) |
| table_group->ops->unset_window(table_group, i); |
| |
| return ret; |
| } |
| |
| static int tce_iommu_attach_group(void *iommu_data, |
| struct iommu_group *iommu_group, enum vfio_group_type type) |
| { |
| int ret = 0; |
| struct tce_container *container = iommu_data; |
| struct iommu_table_group *table_group; |
| struct tce_iommu_group *tcegrp = NULL; |
| |
| if (type == VFIO_EMULATED_IOMMU) |
| return -EINVAL; |
| |
| mutex_lock(&container->lock); |
| |
| /* pr_debug("tce_vfio: Attaching group #%u to iommu %p\n", |
| iommu_group_id(iommu_group), iommu_group); */ |
| table_group = iommu_group_get_iommudata(iommu_group); |
| if (!table_group) { |
| ret = -ENODEV; |
| goto unlock_exit; |
| } |
| |
| /* v2 requires full support of dynamic DMA windows */ |
| if (container->v2 && table_group->max_dynamic_windows_supported == 0) { |
| ret = -EINVAL; |
| goto unlock_exit; |
| } |
| |
| /* v1 reuses TCE tables and does not share them among PEs */ |
| if (!container->v2 && tce_groups_attached(container)) { |
| ret = -EBUSY; |
| goto unlock_exit; |
| } |
| |
| /* |
| * Check if new group has the same iommu_table_group_ops |
| * (i.e. compatible) |
| */ |
| list_for_each_entry(tcegrp, &container->group_list, next) { |
| struct iommu_table_group *table_group_tmp; |
| |
| if (tcegrp->grp == iommu_group) { |
| pr_warn("tce_vfio: Group %d is already attached\n", |
| iommu_group_id(iommu_group)); |
| ret = -EBUSY; |
| goto unlock_exit; |
| } |
| table_group_tmp = iommu_group_get_iommudata(tcegrp->grp); |
| if (table_group_tmp->ops->create_table != |
| table_group->ops->create_table) { |
| pr_warn("tce_vfio: Group %d is incompatible with group %d\n", |
| iommu_group_id(iommu_group), |
| iommu_group_id(tcegrp->grp)); |
| ret = -EPERM; |
| goto unlock_exit; |
| } |
| } |
| |
| tcegrp = kzalloc(sizeof(*tcegrp), GFP_KERNEL); |
| if (!tcegrp) { |
| ret = -ENOMEM; |
| goto unlock_exit; |
| } |
| |
| ret = tce_iommu_take_ownership(container, table_group); |
| if (!tce_groups_attached(container) && !container->tables[0]) |
| container->def_window_pending = true; |
| |
| if (!ret) { |
| tcegrp->grp = iommu_group; |
| list_add(&tcegrp->next, &container->group_list); |
| } |
| |
| if (ret && tcegrp) |
| kfree(tcegrp); |
| |
| unlock_exit: |
| mutex_unlock(&container->lock); |
| |
| return ret; |
| } |
| |
| static void tce_iommu_detach_group(void *iommu_data, |
| struct iommu_group *iommu_group) |
| { |
| struct tce_container *container = iommu_data; |
| struct iommu_table_group *table_group; |
| bool found = false; |
| struct tce_iommu_group *tcegrp; |
| |
| mutex_lock(&container->lock); |
| |
| list_for_each_entry(tcegrp, &container->group_list, next) { |
| if (tcegrp->grp == iommu_group) { |
| found = true; |
| break; |
| } |
| } |
| |
| if (!found) { |
| pr_warn("tce_vfio: detaching unattached group #%u\n", |
| iommu_group_id(iommu_group)); |
| goto unlock_exit; |
| } |
| |
| list_del(&tcegrp->next); |
| kfree(tcegrp); |
| |
| table_group = iommu_group_get_iommudata(iommu_group); |
| BUG_ON(!table_group); |
| |
| tce_iommu_release_ownership(container, table_group); |
| |
| unlock_exit: |
| mutex_unlock(&container->lock); |
| } |
| |
| static const struct vfio_iommu_driver_ops tce_iommu_driver_ops = { |
| .name = "iommu-vfio-powerpc", |
| .owner = THIS_MODULE, |
| .open = tce_iommu_open, |
| .release = tce_iommu_release, |
| .ioctl = tce_iommu_ioctl, |
| .attach_group = tce_iommu_attach_group, |
| .detach_group = tce_iommu_detach_group, |
| }; |
| |
| static int __init tce_iommu_init(void) |
| { |
| return vfio_register_iommu_driver(&tce_iommu_driver_ops); |
| } |
| |
| static void __exit tce_iommu_cleanup(void) |
| { |
| vfio_unregister_iommu_driver(&tce_iommu_driver_ops); |
| } |
| |
| module_init(tce_iommu_init); |
| module_exit(tce_iommu_cleanup); |
| |
| MODULE_VERSION(DRIVER_VERSION); |
| MODULE_LICENSE("GPL v2"); |
| MODULE_AUTHOR(DRIVER_AUTHOR); |
| MODULE_DESCRIPTION(DRIVER_DESC); |
| |