// SPDX-License-Identifier: GPL-2.0-only /* * VFIO: IOMMU DMA mapping support for TCE on POWER * * Copyright (C) 2013 IBM Corp. All rights reserved. * Author: Alexey Kardashevskiy <aik@ozlabs.ru> * Copyright Gavin Shan, IBM Corporation 2014. * * Derived from original vfio_iommu_type1.c: * Copyright (C) 2012 Red Hat, Inc. All rights reserved. * Author: Alex Williamson <alex.williamson@redhat.com> */ #include <linux/module.h> #include <linux/pci.h> #include <linux/slab.h> #include <linux/uaccess.h> #include <linux/err.h> #include <linux/vfio.h> #include <linux/vmalloc.h> #include <linux/sched/mm.h> #include <linux/sched/signal.h> #include <linux/mm.h> #include "vfio.h" #include <asm/iommu.h> #include <asm/tce.h> #include <asm/mmu_context.h> #define DRIVER_VERSION "0.1" #define DRIVER_AUTHOR "aik@ozlabs.ru" #define DRIVER_DESC "VFIO IOMMU SPAPR TCE" static void tce_iommu_detach_group(void *iommu_data, struct iommu_group *iommu_group); /* * VFIO IOMMU fd for SPAPR_TCE IOMMU implementation * * This code handles mapping and unmapping of user data buffers * into DMA'ble space using the IOMMU */ struct tce_iommu_group { struct list_head next; struct iommu_group *grp; }; /* * A container needs to remember which preregistered region it has * referenced to do proper cleanup at the userspace process exit. */ struct tce_iommu_prereg { struct list_head next; struct mm_iommu_table_group_mem_t *mem; }; /* * The container descriptor supports only a single group per container. * Required by the API as the container is not supplied with the IOMMU group * at the moment of initialization. */ struct tce_container { struct mutex lock; bool enabled; bool v2; bool def_window_pending; unsigned long locked_pages; struct mm_struct *mm; struct iommu_table *tables[IOMMU_TABLE_GROUP_MAX_TABLES]; struct list_head group_list; struct list_head prereg_list; }; static long tce_iommu_mm_set(struct tce_container *container) { if (container->mm) { if (container->mm == current->mm) return 0; return -EPERM; } BUG_ON(!current->mm); container->mm = current->mm; mmgrab(container->mm); return 0; } static long tce_iommu_prereg_free(struct tce_container *container, struct tce_iommu_prereg *tcemem) { long ret; ret = mm_iommu_put(container->mm, tcemem->mem); if (ret) return ret; list_del(&tcemem->next); kfree(tcemem); return 0; } static long tce_iommu_unregister_pages(struct tce_container *container, __u64 vaddr, __u64 size) { struct mm_iommu_table_group_mem_t *mem; struct tce_iommu_prereg *tcemem; bool found = false; long ret; if ((vaddr & ~PAGE_MASK) || (size & ~PAGE_MASK)) return -EINVAL; mem = mm_iommu_get(container->mm, vaddr, size >> PAGE_SHIFT); if (!mem) return -ENOENT; list_for_each_entry(tcemem, &container->prereg_list, next) { if (tcemem->mem == mem) { found = true; break; } } if (!found) ret = -ENOENT; else ret = tce_iommu_prereg_free(container, tcemem); mm_iommu_put(container->mm, mem); return ret; } static long tce_iommu_register_pages(struct tce_container *container, __u64 vaddr, __u64 size) { long ret = 0; struct mm_iommu_table_group_mem_t *mem = NULL; struct tce_iommu_prereg *tcemem; unsigned long entries = size >> PAGE_SHIFT; if ((vaddr & ~PAGE_MASK) || (size & ~PAGE_MASK) || ((vaddr + size) < vaddr)) return -EINVAL; mem = mm_iommu_get(container->mm, vaddr, entries); if (mem) { list_for_each_entry(tcemem, &container->prereg_list, next) { if (tcemem->mem == mem) { ret = -EBUSY; goto put_exit; } } } else { ret = mm_iommu_new(container->mm, vaddr, entries, &mem); if (ret) return ret; } tcemem = kzalloc(sizeof(*tcemem), GFP_KERNEL); if (!tcemem) { ret = -ENOMEM; goto put_exit; } tcemem->mem = mem; list_add(&tcemem->next, &container->prereg_list); container->enabled = true; return 0; put_exit: mm_iommu_put(container->mm, mem); return ret; } static bool tce_page_is_contained(struct mm_struct *mm, unsigned long hpa, unsigned int it_page_shift) { struct page *page; unsigned long size = 0; if (mm_iommu_is_devmem(mm, hpa, it_page_shift, &size)) return size == (1UL << it_page_shift); page = pfn_to_page(hpa >> PAGE_SHIFT); /* * Check that the TCE table granularity is not bigger than the size of * a page we just found. Otherwise the hardware can get access to * a bigger memory chunk that it should. */ return page_shift(compound_head(page)) >= it_page_shift; } static inline bool tce_groups_attached(struct tce_container *container) { return !list_empty(&container->group_list); } static long tce_iommu_find_table(struct tce_container *container, phys_addr_t ioba, struct iommu_table **ptbl) { long i; for (i = 0; i < IOMMU_TABLE_GROUP_MAX_TABLES; ++i) { struct iommu_table *tbl = container->tables[i]; if (tbl) { unsigned long entry = ioba >> tbl->it_page_shift; unsigned long start = tbl->it_offset; unsigned long end = start + tbl->it_size; if ((start <= entry) && (entry < end)) { *ptbl = tbl; return i; } } } return -1; } static int tce_iommu_find_free_table(struct tce_container *container) { int i; for (i = 0; i < IOMMU_TABLE_GROUP_MAX_TABLES; ++i) { if (!container->tables[i]) return i; } return -ENOSPC; } static int tce_iommu_enable(struct tce_container *container) { int ret = 0; unsigned long locked; struct iommu_table_group *table_group; struct tce_iommu_group *tcegrp; if (container->enabled) return -EBUSY; /* * When userspace pages are mapped into the IOMMU, they are effectively * locked memory, so, theoretically, we need to update the accounting * of locked pages on each map and unmap. For powerpc, the map unmap * paths can be very hot, though, and the accounting would kill * performance, especially since it would be difficult to impossible * to handle the accounting in real mode only. * * To address that, rather than precisely accounting every page, we * instead account for a worst case on locked memory when the iommu is * enabled and disabled. The worst case upper bound on locked memory * is the size of the whole iommu window, which is usually relatively * small (compared to total memory sizes) on POWER hardware. * * Also we don't have a nice way to fail on H_PUT_TCE due to ulimits, * that would effectively kill the guest at random points, much better * enforcing the limit based on the max that the guest can map. * * Unfortunately at the moment it counts whole tables, no matter how * much memory the guest has. I.e. for 4GB guest and 4 IOMMU groups * each with 2GB DMA window, 8GB will be counted here. The reason for * this is that we cannot tell here the amount of RAM used by the guest * as this information is only available from KVM and VFIO is * KVM agnostic. * * So we do not allow enabling a container without a group attached * as there is no way to know how much we should increment * the locked_vm counter. */ if (!tce_groups_attached(container)) return -ENODEV; tcegrp = list_first_entry(&container->group_list, struct tce_iommu_group, next); table_group = iommu_group_get_iommudata(tcegrp->grp); if (!table_group) return -ENODEV; if (!table_group->tce32_size) return -EPERM; ret = tce_iommu_mm_set(container); if (ret) return ret; locked = table_group->tce32_size >> PAGE_SHIFT; ret = account_locked_vm(container->mm, locked, true); if (ret) return ret; container->locked_pages = locked; container->enabled = true; return ret; } static void tce_iommu_disable(struct tce_container *container) { if (!container->enabled) return; container->enabled = false; BUG_ON(!container->mm); account_locked_vm(container->mm, container->locked_pages, false); } static void *tce_iommu_open(unsigned long arg) { struct tce_container *container; if ((arg != VFIO_SPAPR_TCE_IOMMU) && (arg != VFIO_SPAPR_TCE_v2_IOMMU)) { pr_err("tce_vfio: Wrong IOMMU type\n"); return ERR_PTR(-EINVAL); } container = kzalloc(sizeof(*container), GFP_KERNEL); if (!container) return ERR_PTR(-ENOMEM); mutex_init(&container->lock); INIT_LIST_HEAD_RCU(&container->group_list); INIT_LIST_HEAD_RCU(&container->prereg_list); container->v2 = arg == VFIO_SPAPR_TCE_v2_IOMMU; return container; } static int tce_iommu_clear(struct tce_container *container, struct iommu_table *tbl, unsigned long entry, unsigned long pages); static void tce_iommu_free_table(struct tce_container *container, struct iommu_table *tbl); static void tce_iommu_release(void *iommu_data) { struct tce_container *container = iommu_data; struct tce_iommu_group *tcegrp; struct tce_iommu_prereg *tcemem, *tmtmp; long i; while (tce_groups_attached(container)) { tcegrp = list_first_entry(&container->group_list, struct tce_iommu_group, next); tce_iommu_detach_group(iommu_data, tcegrp->grp); } /* * If VFIO created a table, it was not disposed * by tce_iommu_detach_group() so do it now. */ for (i = 0; i < IOMMU_TABLE_GROUP_MAX_TABLES; ++i) { struct iommu_table *tbl = container->tables[i]; if (!tbl) continue; tce_iommu_clear(container, tbl, tbl->it_offset, tbl->it_size); tce_iommu_free_table(container, tbl); } list_for_each_entry_safe(tcemem, tmtmp, &container->prereg_list, next) WARN_ON(tce_iommu_prereg_free(container, tcemem)); tce_iommu_disable(container); if (container->mm) mmdrop(container->mm); mutex_destroy(&container->lock); kfree(container); } static void tce_iommu_unuse_page(unsigned long hpa) { struct page *page; page = pfn_to_page(hpa >> PAGE_SHIFT); unpin_user_page(page); } static int tce_iommu_prereg_ua_to_hpa(struct tce_container *container, unsigned long tce, unsigned long shift, unsigned long *phpa, struct mm_iommu_table_group_mem_t **pmem) { long ret = 0; struct mm_iommu_table_group_mem_t *mem; mem = mm_iommu_lookup(container->mm, tce, 1ULL << shift); if (!mem) return -EINVAL; ret = mm_iommu_ua_to_hpa(mem, tce, shift, phpa); if (ret) return -EINVAL; *pmem = mem; return 0; } static void tce_iommu_unuse_page_v2(struct tce_container *container, struct iommu_table *tbl, unsigned long entry) { struct mm_iommu_table_group_mem_t *mem = NULL; int ret; unsigned long hpa = 0; __be64 *pua = IOMMU_TABLE_USERSPACE_ENTRY_RO(tbl, entry); if (!pua) return; ret = tce_iommu_prereg_ua_to_hpa(container, be64_to_cpu(*pua), tbl->it_page_shift, &hpa, &mem); if (ret) pr_debug("%s: tce %llx at #%lx was not cached, ret=%d\n", __func__, be64_to_cpu(*pua), entry, ret); if (mem) mm_iommu_mapped_dec(mem); *pua = cpu_to_be64(0); } static int tce_iommu_clear(struct tce_container *container, struct iommu_table *tbl, unsigned long entry, unsigned long pages) { unsigned long oldhpa; long ret; enum dma_data_direction direction; unsigned long lastentry = entry + pages, firstentry = entry; for ( ; entry < lastentry; ++entry) { if (tbl->it_indirect_levels && tbl->it_userspace) { /* * For multilevel tables, we can take a shortcut here * and skip some TCEs as we know that the userspace * addresses cache is a mirror of the real TCE table * and if it is missing some indirect levels, then * the hardware table does not have them allocated * either and therefore does not require updating. */ __be64 *pua = IOMMU_TABLE_USERSPACE_ENTRY_RO(tbl, entry); if (!pua) { /* align to level_size which is power of two */ entry |= tbl->it_level_size - 1; continue; } } cond_resched(); direction = DMA_NONE; oldhpa = 0; ret = iommu_tce_xchg_no_kill(container->mm, tbl, entry, &oldhpa, &direction); if (ret) continue; if (direction == DMA_NONE) continue; if (container->v2) { tce_iommu_unuse_page_v2(container, tbl, entry); continue; } tce_iommu_unuse_page(oldhpa); } iommu_tce_kill(tbl, firstentry, pages); return 0; } static int tce_iommu_use_page(unsigned long tce, unsigned long *hpa) { struct page *page = NULL; enum dma_data_direction direction = iommu_tce_direction(tce); if (pin_user_pages_fast(tce & PAGE_MASK, 1, direction != DMA_TO_DEVICE ? FOLL_WRITE : 0, &page) != 1) return -EFAULT; *hpa = __pa((unsigned long) page_address(page)); return 0; } static long tce_iommu_build(struct tce_container *container, struct iommu_table *tbl, unsigned long entry, unsigned long tce, unsigned long pages, enum dma_data_direction direction) { long i, ret = 0; unsigned long hpa; enum dma_data_direction dirtmp; for (i = 0; i < pages; ++i) { unsigned long offset = tce & IOMMU_PAGE_MASK(tbl) & ~PAGE_MASK; ret = tce_iommu_use_page(tce, &hpa); if (ret) break; if (!tce_page_is_contained(container->mm, hpa, tbl->it_page_shift)) { ret = -EPERM; break; } hpa |= offset; dirtmp = direction; ret = iommu_tce_xchg_no_kill(container->mm, tbl, entry + i, &hpa, &dirtmp); if (ret) { tce_iommu_unuse_page(hpa); pr_err("iommu_tce: %s failed ioba=%lx, tce=%lx, ret=%ld\n", __func__, entry << tbl->it_page_shift, tce, ret); break; } if (dirtmp != DMA_NONE) tce_iommu_unuse_page(hpa); tce += IOMMU_PAGE_SIZE(tbl); } if (ret) tce_iommu_clear(container, tbl, entry, i); else iommu_tce_kill(tbl, entry, pages); return ret; } static long tce_iommu_build_v2(struct tce_container *container, struct iommu_table *tbl, unsigned long entry, unsigned long tce, unsigned long pages, enum dma_data_direction direction) { long i, ret = 0; unsigned long hpa; enum dma_data_direction dirtmp; for (i = 0; i < pages; ++i) { struct mm_iommu_table_group_mem_t *mem = NULL; __be64 *pua = IOMMU_TABLE_USERSPACE_ENTRY(tbl, entry + i); ret = tce_iommu_prereg_ua_to_hpa(container, tce, tbl->it_page_shift, &hpa, &mem); if (ret) break; if (!tce_page_is_contained(container->mm, hpa, tbl->it_page_shift)) { ret = -EPERM; break; } /* Preserve offset within IOMMU page */ hpa |= tce & IOMMU_PAGE_MASK(tbl) & ~PAGE_MASK; dirtmp = direction; /* The registered region is being unregistered */ if (mm_iommu_mapped_inc(mem)) break; ret = iommu_tce_xchg_no_kill(container->mm, tbl, entry + i, &hpa, &dirtmp); if (ret) { /* dirtmp cannot be DMA_NONE here */ tce_iommu_unuse_page_v2(container, tbl, entry + i); pr_err("iommu_tce: %s failed ioba=%lx, tce=%lx, ret=%ld\n", __func__, entry << tbl->it_page_shift, tce, ret); break; } if (dirtmp != DMA_NONE) tce_iommu_unuse_page_v2(container, tbl, entry + i); *pua = cpu_to_be64(tce); tce += IOMMU_PAGE_SIZE(tbl); } if (ret) tce_iommu_clear(container, tbl, entry, i); else iommu_tce_kill(tbl, entry, pages); return ret; } static long tce_iommu_create_table(struct tce_container *container, struct iommu_table_group *table_group, int num, __u32 page_shift, __u64 window_size, __u32 levels, struct iommu_table **ptbl) { long ret, table_size; table_size = table_group->ops->get_table_size(page_shift, window_size, levels); if (!table_size) return -EINVAL; ret = account_locked_vm(container->mm, table_size >> PAGE_SHIFT, true); if (ret) return ret; ret = table_group->ops->create_table(table_group, num, page_shift, window_size, levels, ptbl); WARN_ON(!ret && !(*ptbl)->it_ops->free); WARN_ON(!ret && ((*ptbl)->it_allocated_size > table_size)); return ret; } static void tce_iommu_free_table(struct tce_container *container, struct iommu_table *tbl) { unsigned long pages = tbl->it_allocated_size >> PAGE_SHIFT; iommu_tce_table_put(tbl); account_locked_vm(container->mm, pages, false); } static long tce_iommu_create_window(struct tce_container *container, __u32 page_shift, __u64 window_size, __u32 levels, __u64 *start_addr) { struct tce_iommu_group *tcegrp; struct iommu_table_group *table_group; struct iommu_table *tbl = NULL; long ret, num; num = tce_iommu_find_free_table(container); if (num < 0) return num; /* Get the first group for ops::create_table */ tcegrp = list_first_entry(&container->group_list, struct tce_iommu_group, next); table_group = iommu_group_get_iommudata(tcegrp->grp); if (!table_group) return -EFAULT; if (!(table_group->pgsizes & (1ULL << page_shift))) return -EINVAL; if (!table_group->ops->set_window || !table_group->ops->unset_window || !table_group->ops->get_table_size || !table_group->ops->create_table) return -EPERM; /* Create TCE table */ ret = tce_iommu_create_table(container, table_group, num, page_shift, window_size, levels, &tbl); if (ret) return ret; BUG_ON(!tbl->it_ops->free); /* * Program the table to every group. * Groups have been tested for compatibility at the attach time. */ list_for_each_entry(tcegrp, &container->group_list, next) { table_group = iommu_group_get_iommudata(tcegrp->grp); ret = table_group->ops->set_window(table_group, num, tbl); if (ret) goto unset_exit; } container->tables[num] = tbl; /* Return start address assigned by platform in create_table() */ *start_addr = tbl->it_offset << tbl->it_page_shift; return 0; unset_exit: list_for_each_entry(tcegrp, &container->group_list, next) { table_group = iommu_group_get_iommudata(tcegrp->grp); table_group->ops->unset_window(table_group, num); } tce_iommu_free_table(container, tbl); return ret; } static long tce_iommu_remove_window(struct tce_container *container, __u64 start_addr) { struct iommu_table_group *table_group = NULL; struct iommu_table *tbl; struct tce_iommu_group *tcegrp; int num; num = tce_iommu_find_table(container, start_addr, &tbl); if (num < 0) return -EINVAL; BUG_ON(!tbl->it_size); /* Detach groups from IOMMUs */ list_for_each_entry(tcegrp, &container->group_list, next) { table_group = iommu_group_get_iommudata(tcegrp->grp); /* * SPAPR TCE IOMMU exposes the default DMA window to * the guest via dma32_window_start/size of * VFIO_IOMMU_SPAPR_TCE_GET_INFO. Some platforms allow * the userspace to remove this window, some do not so * here we check for the platform capability. */ if (!table_group->ops || !table_group->ops->unset_window) return -EPERM; table_group->ops->unset_window(table_group, num); } /* Free table */ tce_iommu_clear(container, tbl, tbl->it_offset, tbl->it_size); tce_iommu_free_table(container, tbl); container->tables[num] = NULL; return 0; } static long tce_iommu_create_default_window(struct tce_container *container) { long ret; __u64 start_addr = 0; struct tce_iommu_group *tcegrp; struct iommu_table_group *table_group; if (!container->def_window_pending) return 0; if (!tce_groups_attached(container)) return -ENODEV; tcegrp = list_first_entry(&container->group_list, struct tce_iommu_group, next); table_group = iommu_group_get_iommudata(tcegrp->grp); if (!table_group) return -ENODEV; ret = tce_iommu_create_window(container, IOMMU_PAGE_SHIFT_4K, table_group->tce32_size, 1, &start_addr); WARN_ON_ONCE(!ret && start_addr); if (!ret) container->def_window_pending = false; return ret; } static long vfio_spapr_ioctl_eeh_pe_op(struct iommu_group *group, unsigned long arg) { struct eeh_pe *pe; struct vfio_eeh_pe_op op; unsigned long minsz; pe = eeh_iommu_group_to_pe(group); if (!pe) return -ENODEV; minsz = offsetofend(struct vfio_eeh_pe_op, op); if (copy_from_user(&op, (void __user *)arg, minsz)) return -EFAULT; if (op.argsz < minsz || op.flags) return -EINVAL; switch (op.op) { case VFIO_EEH_PE_DISABLE: return eeh_pe_set_option(pe, EEH_OPT_DISABLE); case VFIO_EEH_PE_ENABLE: return eeh_pe_set_option(pe, EEH_OPT_ENABLE); case VFIO_EEH_PE_UNFREEZE_IO: return eeh_pe_set_option(pe, EEH_OPT_THAW_MMIO); case VFIO_EEH_PE_UNFREEZE_DMA: return eeh_pe_set_option(pe, EEH_OPT_THAW_DMA); case VFIO_EEH_PE_GET_STATE: return eeh_pe_get_state(pe); break; case VFIO_EEH_PE_RESET_DEACTIVATE: return eeh_pe_reset(pe, EEH_RESET_DEACTIVATE, true); case VFIO_EEH_PE_RESET_HOT: return eeh_pe_reset(pe, EEH_RESET_HOT, true); case VFIO_EEH_PE_RESET_FUNDAMENTAL: return eeh_pe_reset(pe, EEH_RESET_FUNDAMENTAL, true); case VFIO_EEH_PE_CONFIGURE: return eeh_pe_configure(pe); case VFIO_EEH_PE_INJECT_ERR: minsz = offsetofend(struct vfio_eeh_pe_op, err.mask); if (op.argsz < minsz) return -EINVAL; if (copy_from_user(&op, (void __user *)arg, minsz)) return -EFAULT; return eeh_pe_inject_err(pe, op.err.type, op.err.func, op.err.addr, op.err.mask); default: return -EINVAL; } } static long tce_iommu_ioctl(void *iommu_data, unsigned int cmd, unsigned long arg) { struct tce_container *container = iommu_data; unsigned long minsz, ddwsz; long ret; switch (cmd) { case VFIO_CHECK_EXTENSION: switch (arg) { case VFIO_SPAPR_TCE_IOMMU: case VFIO_SPAPR_TCE_v2_IOMMU: return 1; case VFIO_EEH: return eeh_enabled(); default: return 0; } } /* * Sanity check to prevent one userspace from manipulating * another userspace mm. */ BUG_ON(!container); if (container->mm && container->mm != current->mm) return -EPERM; switch (cmd) { case VFIO_IOMMU_SPAPR_TCE_GET_INFO: { struct vfio_iommu_spapr_tce_info info; struct tce_iommu_group *tcegrp; struct iommu_table_group *table_group; if (!tce_groups_attached(container)) return -ENXIO; tcegrp = list_first_entry(&container->group_list, struct tce_iommu_group, next); table_group = iommu_group_get_iommudata(tcegrp->grp); if (!table_group) return -ENXIO; minsz = offsetofend(struct vfio_iommu_spapr_tce_info, dma32_window_size); if (copy_from_user(&info, (void __user *)arg, minsz)) return -EFAULT; if (info.argsz < minsz) return -EINVAL; info.dma32_window_start = table_group->tce32_start; info.dma32_window_size = table_group->tce32_size; info.flags = 0; memset(&info.ddw, 0, sizeof(info.ddw)); if (table_group->max_dynamic_windows_supported && container->v2) { info.flags |= VFIO_IOMMU_SPAPR_INFO_DDW; info.ddw.pgsizes = table_group->pgsizes; info.ddw.max_dynamic_windows_supported = table_group->max_dynamic_windows_supported; info.ddw.levels = table_group->max_levels; } ddwsz = offsetofend(struct vfio_iommu_spapr_tce_info, ddw); if (info.argsz >= ddwsz) minsz = ddwsz; if (copy_to_user((void __user *)arg, &info, minsz)) return -EFAULT; return 0; } case VFIO_IOMMU_MAP_DMA: { struct vfio_iommu_type1_dma_map param; struct iommu_table *tbl = NULL; long num; enum dma_data_direction direction; if (!container->enabled) return -EPERM; minsz = offsetofend(struct vfio_iommu_type1_dma_map, size); if (copy_from_user(¶m, (void __user *)arg, minsz)) return -EFAULT; if (param.argsz < minsz) return -EINVAL; if (param.flags & ~(VFIO_DMA_MAP_FLAG_READ | VFIO_DMA_MAP_FLAG_WRITE)) return -EINVAL; ret = tce_iommu_create_default_window(container); if (ret) return ret; num = tce_iommu_find_table(container, param.iova, &tbl); if (num < 0) return -ENXIO; if ((param.size & ~IOMMU_PAGE_MASK(tbl)) || (param.vaddr & ~IOMMU_PAGE_MASK(tbl))) return -EINVAL; /* iova is checked by the IOMMU API */ if (param.flags & VFIO_DMA_MAP_FLAG_READ) { if (param.flags & VFIO_DMA_MAP_FLAG_WRITE) direction = DMA_BIDIRECTIONAL; else direction = DMA_TO_DEVICE; } else { if (param.flags & VFIO_DMA_MAP_FLAG_WRITE) direction = DMA_FROM_DEVICE; else return -EINVAL; } ret = iommu_tce_put_param_check(tbl, param.iova, param.vaddr); if (ret) return ret; if (container->v2) ret = tce_iommu_build_v2(container, tbl, param.iova >> tbl->it_page_shift, param.vaddr, param.size >> tbl->it_page_shift, direction); else ret = tce_iommu_build(container, tbl, param.iova >> tbl->it_page_shift, param.vaddr, param.size >> tbl->it_page_shift, direction); iommu_flush_tce(tbl); return ret; } case VFIO_IOMMU_UNMAP_DMA: { struct vfio_iommu_type1_dma_unmap param; struct iommu_table *tbl = NULL; long num; if (!container->enabled) return -EPERM; minsz = offsetofend(struct vfio_iommu_type1_dma_unmap, size); if (copy_from_user(¶m, (void __user *)arg, minsz)) return -EFAULT; if (param.argsz < minsz) return -EINVAL; /* No flag is supported now */ if (param.flags) return -EINVAL; ret = tce_iommu_create_default_window(container); if (ret) return ret; num = tce_iommu_find_table(container, param.iova, &tbl); if (num < 0) return -ENXIO; if (param.size & ~IOMMU_PAGE_MASK(tbl)) return -EINVAL; ret = iommu_tce_clear_param_check(tbl, param.iova, 0, param.size >> tbl->it_page_shift); if (ret) return ret; ret = tce_iommu_clear(container, tbl, param.iova >> tbl->it_page_shift, param.size >> tbl->it_page_shift); iommu_flush_tce(tbl); return ret; } case VFIO_IOMMU_SPAPR_REGISTER_MEMORY: { struct vfio_iommu_spapr_register_memory param; if (!container->v2) break; minsz = offsetofend(struct vfio_iommu_spapr_register_memory, size); ret = tce_iommu_mm_set(container); if (ret) return ret; if (copy_from_user(¶m, (void __user *)arg, minsz)) return -EFAULT; if (param.argsz < minsz) return -EINVAL; /* No flag is supported now */ if (param.flags) return -EINVAL; mutex_lock(&container->lock); ret = tce_iommu_register_pages(container, param.vaddr, param.size); mutex_unlock(&container->lock); return ret; } case VFIO_IOMMU_SPAPR_UNREGISTER_MEMORY: { struct vfio_iommu_spapr_register_memory param; if (!container->v2) break; if (!container->mm) return -EPERM; minsz = offsetofend(struct vfio_iommu_spapr_register_memory, size); if (copy_from_user(¶m, (void __user *)arg, minsz)) return -EFAULT; if (param.argsz < minsz) return -EINVAL; /* No flag is supported now */ if (param.flags) return -EINVAL; mutex_lock(&container->lock); ret = tce_iommu_unregister_pages(container, param.vaddr, param.size); mutex_unlock(&container->lock); return ret; } case VFIO_IOMMU_ENABLE: if (container->v2) break; mutex_lock(&container->lock); ret = tce_iommu_enable(container); mutex_unlock(&container->lock); return ret; case VFIO_IOMMU_DISABLE: if (container->v2) break; mutex_lock(&container->lock); tce_iommu_disable(container); mutex_unlock(&container->lock); return 0; case VFIO_EEH_PE_OP: { struct tce_iommu_group *tcegrp; ret = 0; list_for_each_entry(tcegrp, &container->group_list, next) { ret = vfio_spapr_ioctl_eeh_pe_op(tcegrp->grp, arg); if (ret) return ret; } return ret; } case VFIO_IOMMU_SPAPR_TCE_CREATE: { struct vfio_iommu_spapr_tce_create create; if (!container->v2) break; ret = tce_iommu_mm_set(container); if (ret) return ret; if (!tce_groups_attached(container)) return -ENXIO; minsz = offsetofend(struct vfio_iommu_spapr_tce_create, start_addr); if (copy_from_user(&create, (void __user *)arg, minsz)) return -EFAULT; if (create.argsz < minsz) return -EINVAL; if (create.flags) return -EINVAL; mutex_lock(&container->lock); ret = tce_iommu_create_default_window(container); if (!ret) ret = tce_iommu_create_window(container, create.page_shift, create.window_size, create.levels, &create.start_addr); mutex_unlock(&container->lock); if (!ret && copy_to_user((void __user *)arg, &create, minsz)) ret = -EFAULT; return ret; } case VFIO_IOMMU_SPAPR_TCE_REMOVE: { struct vfio_iommu_spapr_tce_remove remove; if (!container->v2) break; ret = tce_iommu_mm_set(container); if (ret) return ret; if (!tce_groups_attached(container)) return -ENXIO; minsz = offsetofend(struct vfio_iommu_spapr_tce_remove, start_addr); if (copy_from_user(&remove, (void __user *)arg, minsz)) return -EFAULT; if (remove.argsz < minsz) return -EINVAL; if (remove.flags) return -EINVAL; if (container->def_window_pending && !remove.start_addr) { container->def_window_pending = false; return 0; } mutex_lock(&container->lock); ret = tce_iommu_remove_window(container, remove.start_addr); mutex_unlock(&container->lock); return ret; } } return -ENOTTY; } static void tce_iommu_release_ownership(struct tce_container *container, struct iommu_table_group *table_group) { long i; if (!table_group->ops->unset_window) { WARN_ON_ONCE(1); return; } for (i = 0; i < IOMMU_TABLE_GROUP_MAX_TABLES; ++i) if (container->tables[i]) table_group->ops->unset_window(table_group, i); } static long tce_iommu_take_ownership(struct tce_container *container, struct iommu_table_group *table_group) { long i, ret = 0; /* Set all windows to the new group */ for (i = 0; i < IOMMU_TABLE_GROUP_MAX_TABLES; ++i) { struct iommu_table *tbl = container->tables[i]; if (!tbl) continue; ret = table_group->ops->set_window(table_group, i, tbl); if (ret) goto release_exit; } return 0; release_exit: for (i = 0; i < IOMMU_TABLE_GROUP_MAX_TABLES; ++i) table_group->ops->unset_window(table_group, i); return ret; } static int tce_iommu_attach_group(void *iommu_data, struct iommu_group *iommu_group, enum vfio_group_type type) { int ret = 0; struct tce_container *container = iommu_data; struct iommu_table_group *table_group; struct tce_iommu_group *tcegrp = NULL; if (type == VFIO_EMULATED_IOMMU) return -EINVAL; mutex_lock(&container->lock); /* pr_debug("tce_vfio: Attaching group #%u to iommu %p\n", iommu_group_id(iommu_group), iommu_group); */ table_group = iommu_group_get_iommudata(iommu_group); if (!table_group) { ret = -ENODEV; goto unlock_exit; } /* v2 requires full support of dynamic DMA windows */ if (container->v2 && table_group->max_dynamic_windows_supported == 0) { ret = -EINVAL; goto unlock_exit; } /* v1 reuses TCE tables and does not share them among PEs */ if (!container->v2 && tce_groups_attached(container)) { ret = -EBUSY; goto unlock_exit; } /* * Check if new group has the same iommu_table_group_ops * (i.e. compatible) */ list_for_each_entry(tcegrp, &container->group_list, next) { struct iommu_table_group *table_group_tmp; if (tcegrp->grp == iommu_group) { pr_warn("tce_vfio: Group %d is already attached\n", iommu_group_id(iommu_group)); ret = -EBUSY; goto unlock_exit; } table_group_tmp = iommu_group_get_iommudata(tcegrp->grp); if (table_group_tmp->ops->create_table != table_group->ops->create_table) { pr_warn("tce_vfio: Group %d is incompatible with group %d\n", iommu_group_id(iommu_group), iommu_group_id(tcegrp->grp)); ret = -EPERM; goto unlock_exit; } } tcegrp = kzalloc(sizeof(*tcegrp), GFP_KERNEL); if (!tcegrp) { ret = -ENOMEM; goto unlock_exit; } ret = tce_iommu_take_ownership(container, table_group); if (!tce_groups_attached(container) && !container->tables[0]) container->def_window_pending = true; if (!ret) { tcegrp->grp = iommu_group; list_add(&tcegrp->next, &container->group_list); } if (ret && tcegrp) kfree(tcegrp); unlock_exit: mutex_unlock(&container->lock); return ret; } static void tce_iommu_detach_group(void *iommu_data, struct iommu_group *iommu_group) { struct tce_container *container = iommu_data; struct iommu_table_group *table_group; bool found = false; struct tce_iommu_group *tcegrp; mutex_lock(&container->lock); list_for_each_entry(tcegrp, &container->group_list, next) { if (tcegrp->grp == iommu_group) { found = true; break; } } if (!found) { pr_warn("tce_vfio: detaching unattached group #%u\n", iommu_group_id(iommu_group)); goto unlock_exit; } list_del(&tcegrp->next); kfree(tcegrp); table_group = iommu_group_get_iommudata(iommu_group); BUG_ON(!table_group); tce_iommu_release_ownership(container, table_group); unlock_exit: mutex_unlock(&container->lock); } static const struct vfio_iommu_driver_ops tce_iommu_driver_ops = { .name = "iommu-vfio-powerpc", .owner = THIS_MODULE, .open = tce_iommu_open, .release = tce_iommu_release, .ioctl = tce_iommu_ioctl, .attach_group = tce_iommu_attach_group, .detach_group = tce_iommu_detach_group, }; static int __init tce_iommu_init(void) { return vfio_register_iommu_driver(&tce_iommu_driver_ops); } static void __exit tce_iommu_cleanup(void) { vfio_unregister_iommu_driver(&tce_iommu_driver_ops); } module_init(tce_iommu_init); module_exit(tce_iommu_cleanup); MODULE_VERSION(DRIVER_VERSION); MODULE_LICENSE("GPL v2"); MODULE_AUTHOR(DRIVER_AUTHOR); MODULE_DESCRIPTION(DRIVER_DESC);