// SPDX-License-Identifier: GPL-2.0 /* * Copyright (C) 2021 Western Digital Corporation or its affiliates. * Copyright (C) 2022 Ventana Micro Systems Inc. * * Authors: * Anup Patel <apatel@ventanamicro.com> */ #include <linux/bits.h> #include <linux/kvm_host.h> #include <linux/uaccess.h> #include <asm/kvm_aia_imsic.h> static void unlock_vcpus(struct kvm *kvm, int vcpu_lock_idx) { struct kvm_vcpu *tmp_vcpu; for (; vcpu_lock_idx >= 0; vcpu_lock_idx--) { tmp_vcpu = kvm_get_vcpu(kvm, vcpu_lock_idx); mutex_unlock(&tmp_vcpu->mutex); } } static void unlock_all_vcpus(struct kvm *kvm) { unlock_vcpus(kvm, atomic_read(&kvm->online_vcpus) - 1); } static bool lock_all_vcpus(struct kvm *kvm) { struct kvm_vcpu *tmp_vcpu; unsigned long c; kvm_for_each_vcpu(c, tmp_vcpu, kvm) { if (!mutex_trylock(&tmp_vcpu->mutex)) { unlock_vcpus(kvm, c - 1); return false; } } return true; } static int aia_create(struct kvm_device *dev, u32 type) { int ret; unsigned long i; struct kvm *kvm = dev->kvm; struct kvm_vcpu *vcpu; if (irqchip_in_kernel(kvm)) return -EEXIST; ret = -EBUSY; if (!lock_all_vcpus(kvm)) return ret; kvm_for_each_vcpu(i, vcpu, kvm) { if (vcpu->arch.ran_atleast_once) goto out_unlock; } ret = 0; kvm->arch.aia.in_kernel = true; out_unlock: unlock_all_vcpus(kvm); return ret; } static void aia_destroy(struct kvm_device *dev) { kfree(dev); } static int aia_config(struct kvm *kvm, unsigned long type, u32 *nr, bool write) { struct kvm_aia *aia = &kvm->arch.aia; /* Writes can only be done before irqchip is initialized */ if (write && kvm_riscv_aia_initialized(kvm)) return -EBUSY; switch (type) { case KVM_DEV_RISCV_AIA_CONFIG_MODE: if (write) { switch (*nr) { case KVM_DEV_RISCV_AIA_MODE_EMUL: break; case KVM_DEV_RISCV_AIA_MODE_HWACCEL: case KVM_DEV_RISCV_AIA_MODE_AUTO: /* * HW Acceleration and Auto modes only * supported on host with non-zero guest * external interrupts (i.e. non-zero * VS-level IMSIC pages). */ if (!kvm_riscv_aia_nr_hgei) return -EINVAL; break; default: return -EINVAL; } aia->mode = *nr; } else *nr = aia->mode; break; case KVM_DEV_RISCV_AIA_CONFIG_IDS: if (write) { if ((*nr < KVM_DEV_RISCV_AIA_IDS_MIN) || (*nr >= KVM_DEV_RISCV_AIA_IDS_MAX) || ((*nr & KVM_DEV_RISCV_AIA_IDS_MIN) != KVM_DEV_RISCV_AIA_IDS_MIN) || (kvm_riscv_aia_max_ids <= *nr)) return -EINVAL; aia->nr_ids = *nr; } else *nr = aia->nr_ids; break; case KVM_DEV_RISCV_AIA_CONFIG_SRCS: if (write) { if ((*nr >= KVM_DEV_RISCV_AIA_SRCS_MAX) || (*nr >= kvm_riscv_aia_max_ids)) return -EINVAL; aia->nr_sources = *nr; } else *nr = aia->nr_sources; break; case KVM_DEV_RISCV_AIA_CONFIG_GROUP_BITS: if (write) { if (*nr >= KVM_DEV_RISCV_AIA_GROUP_BITS_MAX) return -EINVAL; aia->nr_group_bits = *nr; } else *nr = aia->nr_group_bits; break; case KVM_DEV_RISCV_AIA_CONFIG_GROUP_SHIFT: if (write) { if ((*nr < KVM_DEV_RISCV_AIA_GROUP_SHIFT_MIN) || (*nr >= KVM_DEV_RISCV_AIA_GROUP_SHIFT_MAX)) return -EINVAL; aia->nr_group_shift = *nr; } else *nr = aia->nr_group_shift; break; case KVM_DEV_RISCV_AIA_CONFIG_HART_BITS: if (write) { if (*nr >= KVM_DEV_RISCV_AIA_HART_BITS_MAX) return -EINVAL; aia->nr_hart_bits = *nr; } else *nr = aia->nr_hart_bits; break; case KVM_DEV_RISCV_AIA_CONFIG_GUEST_BITS: if (write) { if (*nr >= KVM_DEV_RISCV_AIA_GUEST_BITS_MAX) return -EINVAL; aia->nr_guest_bits = *nr; } else *nr = aia->nr_guest_bits; break; default: return -ENXIO; } return 0; } static int aia_aplic_addr(struct kvm *kvm, u64 *addr, bool write) { struct kvm_aia *aia = &kvm->arch.aia; if (write) { /* Writes can only be done before irqchip is initialized */ if (kvm_riscv_aia_initialized(kvm)) return -EBUSY; if (*addr & (KVM_DEV_RISCV_APLIC_ALIGN - 1)) return -EINVAL; aia->aplic_addr = *addr; } else *addr = aia->aplic_addr; return 0; } static int aia_imsic_addr(struct kvm *kvm, u64 *addr, unsigned long vcpu_idx, bool write) { struct kvm_vcpu *vcpu; struct kvm_vcpu_aia *vcpu_aia; vcpu = kvm_get_vcpu(kvm, vcpu_idx); if (!vcpu) return -EINVAL; vcpu_aia = &vcpu->arch.aia_context; if (write) { /* Writes can only be done before irqchip is initialized */ if (kvm_riscv_aia_initialized(kvm)) return -EBUSY; if (*addr & (KVM_DEV_RISCV_IMSIC_ALIGN - 1)) return -EINVAL; } mutex_lock(&vcpu->mutex); if (write) vcpu_aia->imsic_addr = *addr; else *addr = vcpu_aia->imsic_addr; mutex_unlock(&vcpu->mutex); return 0; } static gpa_t aia_imsic_ppn(struct kvm_aia *aia, gpa_t addr) { u32 h, l; gpa_t mask = 0; h = aia->nr_hart_bits + aia->nr_guest_bits + IMSIC_MMIO_PAGE_SHIFT - 1; mask = GENMASK_ULL(h, 0); if (aia->nr_group_bits) { h = aia->nr_group_bits + aia->nr_group_shift - 1; l = aia->nr_group_shift; mask |= GENMASK_ULL(h, l); } return (addr & ~mask) >> IMSIC_MMIO_PAGE_SHIFT; } static u32 aia_imsic_hart_index(struct kvm_aia *aia, gpa_t addr) { u32 hart, group = 0; hart = (addr >> (aia->nr_guest_bits + IMSIC_MMIO_PAGE_SHIFT)) & GENMASK_ULL(aia->nr_hart_bits - 1, 0); if (aia->nr_group_bits) group = (addr >> aia->nr_group_shift) & GENMASK_ULL(aia->nr_group_bits - 1, 0); return (group << aia->nr_hart_bits) | hart; } static int aia_init(struct kvm *kvm) { int ret, i; unsigned long idx; struct kvm_vcpu *vcpu; struct kvm_vcpu_aia *vaia; struct kvm_aia *aia = &kvm->arch.aia; gpa_t base_ppn = KVM_RISCV_AIA_UNDEF_ADDR; /* Irqchip can be initialized only once */ if (kvm_riscv_aia_initialized(kvm)) return -EBUSY; /* We might be in the middle of creating a VCPU? */ if (kvm->created_vcpus != atomic_read(&kvm->online_vcpus)) return -EBUSY; /* Number of sources should be less than or equals number of IDs */ if (aia->nr_ids < aia->nr_sources) return -EINVAL; /* APLIC base is required for non-zero number of sources */ if (aia->nr_sources && aia->aplic_addr == KVM_RISCV_AIA_UNDEF_ADDR) return -EINVAL; /* Initialize APLIC */ ret = kvm_riscv_aia_aplic_init(kvm); if (ret) return ret; /* Iterate over each VCPU */ kvm_for_each_vcpu(idx, vcpu, kvm) { vaia = &vcpu->arch.aia_context; /* IMSIC base is required */ if (vaia->imsic_addr == KVM_RISCV_AIA_UNDEF_ADDR) { ret = -EINVAL; goto fail_cleanup_imsics; } /* All IMSICs should have matching base PPN */ if (base_ppn == KVM_RISCV_AIA_UNDEF_ADDR) base_ppn = aia_imsic_ppn(aia, vaia->imsic_addr); if (base_ppn != aia_imsic_ppn(aia, vaia->imsic_addr)) { ret = -EINVAL; goto fail_cleanup_imsics; } /* Update HART index of the IMSIC based on IMSIC base */ vaia->hart_index = aia_imsic_hart_index(aia, vaia->imsic_addr); /* Initialize IMSIC for this VCPU */ ret = kvm_riscv_vcpu_aia_imsic_init(vcpu); if (ret) goto fail_cleanup_imsics; } /* Set the initialized flag */ kvm->arch.aia.initialized = true; return 0; fail_cleanup_imsics: for (i = idx - 1; i >= 0; i--) { vcpu = kvm_get_vcpu(kvm, i); if (!vcpu) continue; kvm_riscv_vcpu_aia_imsic_cleanup(vcpu); } kvm_riscv_aia_aplic_cleanup(kvm); return ret; } static int aia_set_attr(struct kvm_device *dev, struct kvm_device_attr *attr) { u32 nr; u64 addr; int nr_vcpus, r = -ENXIO; unsigned long v, type = (unsigned long)attr->attr; void __user *uaddr = (void __user *)(long)attr->addr; switch (attr->group) { case KVM_DEV_RISCV_AIA_GRP_CONFIG: if (copy_from_user(&nr, uaddr, sizeof(nr))) return -EFAULT; mutex_lock(&dev->kvm->lock); r = aia_config(dev->kvm, type, &nr, true); mutex_unlock(&dev->kvm->lock); break; case KVM_DEV_RISCV_AIA_GRP_ADDR: if (copy_from_user(&addr, uaddr, sizeof(addr))) return -EFAULT; nr_vcpus = atomic_read(&dev->kvm->online_vcpus); mutex_lock(&dev->kvm->lock); if (type == KVM_DEV_RISCV_AIA_ADDR_APLIC) r = aia_aplic_addr(dev->kvm, &addr, true); else if (type < KVM_DEV_RISCV_AIA_ADDR_IMSIC(nr_vcpus)) r = aia_imsic_addr(dev->kvm, &addr, type - KVM_DEV_RISCV_AIA_ADDR_IMSIC(0), true); mutex_unlock(&dev->kvm->lock); break; case KVM_DEV_RISCV_AIA_GRP_CTRL: switch (type) { case KVM_DEV_RISCV_AIA_CTRL_INIT: mutex_lock(&dev->kvm->lock); r = aia_init(dev->kvm); mutex_unlock(&dev->kvm->lock); break; } break; case KVM_DEV_RISCV_AIA_GRP_APLIC: if (copy_from_user(&nr, uaddr, sizeof(nr))) return -EFAULT; mutex_lock(&dev->kvm->lock); r = kvm_riscv_aia_aplic_set_attr(dev->kvm, type, nr); mutex_unlock(&dev->kvm->lock); break; case KVM_DEV_RISCV_AIA_GRP_IMSIC: if (copy_from_user(&v, uaddr, sizeof(v))) return -EFAULT; mutex_lock(&dev->kvm->lock); r = kvm_riscv_aia_imsic_rw_attr(dev->kvm, type, true, &v); mutex_unlock(&dev->kvm->lock); break; } return r; } static int aia_get_attr(struct kvm_device *dev, struct kvm_device_attr *attr) { u32 nr; u64 addr; int nr_vcpus, r = -ENXIO; void __user *uaddr = (void __user *)(long)attr->addr; unsigned long v, type = (unsigned long)attr->attr; switch (attr->group) { case KVM_DEV_RISCV_AIA_GRP_CONFIG: if (copy_from_user(&nr, uaddr, sizeof(nr))) return -EFAULT; mutex_lock(&dev->kvm->lock); r = aia_config(dev->kvm, type, &nr, false); mutex_unlock(&dev->kvm->lock); if (r) return r; if (copy_to_user(uaddr, &nr, sizeof(nr))) return -EFAULT; break; case KVM_DEV_RISCV_AIA_GRP_ADDR: if (copy_from_user(&addr, uaddr, sizeof(addr))) return -EFAULT; nr_vcpus = atomic_read(&dev->kvm->online_vcpus); mutex_lock(&dev->kvm->lock); if (type == KVM_DEV_RISCV_AIA_ADDR_APLIC) r = aia_aplic_addr(dev->kvm, &addr, false); else if (type < KVM_DEV_RISCV_AIA_ADDR_IMSIC(nr_vcpus)) r = aia_imsic_addr(dev->kvm, &addr, type - KVM_DEV_RISCV_AIA_ADDR_IMSIC(0), false); mutex_unlock(&dev->kvm->lock); if (r) return r; if (copy_to_user(uaddr, &addr, sizeof(addr))) return -EFAULT; break; case KVM_DEV_RISCV_AIA_GRP_APLIC: if (copy_from_user(&nr, uaddr, sizeof(nr))) return -EFAULT; mutex_lock(&dev->kvm->lock); r = kvm_riscv_aia_aplic_get_attr(dev->kvm, type, &nr); mutex_unlock(&dev->kvm->lock); if (r) return r; if (copy_to_user(uaddr, &nr, sizeof(nr))) return -EFAULT; break; case KVM_DEV_RISCV_AIA_GRP_IMSIC: if (copy_from_user(&v, uaddr, sizeof(v))) return -EFAULT; mutex_lock(&dev->kvm->lock); r = kvm_riscv_aia_imsic_rw_attr(dev->kvm, type, false, &v); mutex_unlock(&dev->kvm->lock); if (r) return r; if (copy_to_user(uaddr, &v, sizeof(v))) return -EFAULT; break; } return r; } static int aia_has_attr(struct kvm_device *dev, struct kvm_device_attr *attr) { int nr_vcpus; switch (attr->group) { case KVM_DEV_RISCV_AIA_GRP_CONFIG: switch (attr->attr) { case KVM_DEV_RISCV_AIA_CONFIG_MODE: case KVM_DEV_RISCV_AIA_CONFIG_IDS: case KVM_DEV_RISCV_AIA_CONFIG_SRCS: case KVM_DEV_RISCV_AIA_CONFIG_GROUP_BITS: case KVM_DEV_RISCV_AIA_CONFIG_GROUP_SHIFT: case KVM_DEV_RISCV_AIA_CONFIG_HART_BITS: case KVM_DEV_RISCV_AIA_CONFIG_GUEST_BITS: return 0; } break; case KVM_DEV_RISCV_AIA_GRP_ADDR: nr_vcpus = atomic_read(&dev->kvm->online_vcpus); if (attr->attr == KVM_DEV_RISCV_AIA_ADDR_APLIC) return 0; else if (attr->attr < KVM_DEV_RISCV_AIA_ADDR_IMSIC(nr_vcpus)) return 0; break; case KVM_DEV_RISCV_AIA_GRP_CTRL: switch (attr->attr) { case KVM_DEV_RISCV_AIA_CTRL_INIT: return 0; } break; case KVM_DEV_RISCV_AIA_GRP_APLIC: return kvm_riscv_aia_aplic_has_attr(dev->kvm, attr->attr); case KVM_DEV_RISCV_AIA_GRP_IMSIC: return kvm_riscv_aia_imsic_has_attr(dev->kvm, attr->attr); } return -ENXIO; } struct kvm_device_ops kvm_riscv_aia_device_ops = { .name = "kvm-riscv-aia", .create = aia_create, .destroy = aia_destroy, .set_attr = aia_set_attr, .get_attr = aia_get_attr, .has_attr = aia_has_attr, }; int kvm_riscv_vcpu_aia_update(struct kvm_vcpu *vcpu) { /* Proceed only if AIA was initialized successfully */ if (!kvm_riscv_aia_initialized(vcpu->kvm)) return 1; /* Update the IMSIC HW state before entering guest mode */ return kvm_riscv_vcpu_aia_imsic_update(vcpu); } void kvm_riscv_vcpu_aia_reset(struct kvm_vcpu *vcpu) { struct kvm_vcpu_aia_csr *csr = &vcpu->arch.aia_context.guest_csr; struct kvm_vcpu_aia_csr *reset_csr = &vcpu->arch.aia_context.guest_reset_csr; if (!kvm_riscv_aia_available()) return; memcpy(csr, reset_csr, sizeof(*csr)); /* Proceed only if AIA was initialized successfully */ if (!kvm_riscv_aia_initialized(vcpu->kvm)) return; /* Reset the IMSIC context */ kvm_riscv_vcpu_aia_imsic_reset(vcpu); } int kvm_riscv_vcpu_aia_init(struct kvm_vcpu *vcpu) { struct kvm_vcpu_aia *vaia = &vcpu->arch.aia_context; if (!kvm_riscv_aia_available()) return 0; /* * We don't do any memory allocations over here because these * will be done after AIA device is initialized by the user-space. * * Refer, aia_init() implementation for more details. */ /* Initialize default values in AIA vcpu context */ vaia->imsic_addr = KVM_RISCV_AIA_UNDEF_ADDR; vaia->hart_index = vcpu->vcpu_idx; return 0; } void kvm_riscv_vcpu_aia_deinit(struct kvm_vcpu *vcpu) { /* Proceed only if AIA was initialized successfully */ if (!kvm_riscv_aia_initialized(vcpu->kvm)) return; /* Cleanup IMSIC context */ kvm_riscv_vcpu_aia_imsic_cleanup(vcpu); } int kvm_riscv_aia_inject_msi_by_id(struct kvm *kvm, u32 hart_index, u32 guest_index, u32 iid) { unsigned long idx; struct kvm_vcpu *vcpu; /* Proceed only if AIA was initialized successfully */ if (!kvm_riscv_aia_initialized(kvm)) return -EBUSY; /* Inject MSI to matching VCPU */ kvm_for_each_vcpu(idx, vcpu, kvm) { if (vcpu->arch.aia_context.hart_index == hart_index) return kvm_riscv_vcpu_aia_imsic_inject(vcpu, guest_index, 0, iid); } return 0; } int kvm_riscv_aia_inject_msi(struct kvm *kvm, struct kvm_msi *msi) { gpa_t tppn, ippn; unsigned long idx; struct kvm_vcpu *vcpu; u32 g, toff, iid = msi->data; struct kvm_aia *aia = &kvm->arch.aia; gpa_t target = (((gpa_t)msi->address_hi) << 32) | msi->address_lo; /* Proceed only if AIA was initialized successfully */ if (!kvm_riscv_aia_initialized(kvm)) return -EBUSY; /* Convert target address to target PPN */ tppn = target >> IMSIC_MMIO_PAGE_SHIFT; /* Extract and clear Guest ID from target PPN */ g = tppn & (BIT(aia->nr_guest_bits) - 1); tppn &= ~((gpa_t)(BIT(aia->nr_guest_bits) - 1)); /* Inject MSI to matching VCPU */ kvm_for_each_vcpu(idx, vcpu, kvm) { ippn = vcpu->arch.aia_context.imsic_addr >> IMSIC_MMIO_PAGE_SHIFT; if (ippn == tppn) { toff = target & (IMSIC_MMIO_PAGE_SZ - 1); return kvm_riscv_vcpu_aia_imsic_inject(vcpu, g, toff, iid); } } return 0; } int kvm_riscv_aia_inject_irq(struct kvm *kvm, unsigned int irq, bool level) { /* Proceed only if AIA was initialized successfully */ if (!kvm_riscv_aia_initialized(kvm)) return -EBUSY; /* Inject interrupt level change in APLIC */ return kvm_riscv_aia_aplic_inject(kvm, irq, level); } void kvm_riscv_aia_init_vm(struct kvm *kvm) { struct kvm_aia *aia = &kvm->arch.aia; if (!kvm_riscv_aia_available()) return; /* * We don't do any memory allocations over here because these * will be done after AIA device is initialized by the user-space. * * Refer, aia_init() implementation for more details. */ /* Initialize default values in AIA global context */ aia->mode = (kvm_riscv_aia_nr_hgei) ? KVM_DEV_RISCV_AIA_MODE_AUTO : KVM_DEV_RISCV_AIA_MODE_EMUL; aia->nr_ids = kvm_riscv_aia_max_ids - 1; aia->nr_sources = 0; aia->nr_group_bits = 0; aia->nr_group_shift = KVM_DEV_RISCV_AIA_GROUP_SHIFT_MIN; aia->nr_hart_bits = 0; aia->nr_guest_bits = 0; aia->aplic_addr = KVM_RISCV_AIA_UNDEF_ADDR; } void kvm_riscv_aia_destroy_vm(struct kvm *kvm) { /* Proceed only if AIA was initialized successfully */ if (!kvm_riscv_aia_initialized(kvm)) return; /* Cleanup APLIC context */ kvm_riscv_aia_aplic_cleanup(kvm); }