On Fri, Apr 23, 2021 at 03:21:45PM +0530, Vivek Gautam wrote: > SVA bind and unbind implementations will allow to prepare translation > context with CPU page tables that can be programmed into host iommu > hardware to realize shared address space utilization between the CPU > and virtualized devices using virtio-iommu. > > Signed-off-by: Vivek Gautam <vivek.gautam@xxxxxxx> > --- > drivers/iommu/virtio-iommu.c | 199 +++++++++++++++++++++++++++++- > include/uapi/linux/virtio_iommu.h | 2 + > 2 files changed, 199 insertions(+), 2 deletions(-) > > diff --git a/drivers/iommu/virtio-iommu.c b/drivers/iommu/virtio-iommu.c > index 250c137a211b..08f1294baeab 100644 > --- a/drivers/iommu/virtio-iommu.c > +++ b/drivers/iommu/virtio-iommu.c > @@ -14,6 +14,9 @@ > #include <linux/interval_tree.h> > #include <linux/iommu.h> > #include <linux/io-pgtable.h> > +#include <linux/mm.h> > +#include <linux/mmu_context.h> > +#include <linux/mmu_notifier.h> > #include <linux/module.h> > #include <linux/of_iommu.h> > #include <linux/of_platform.h> > @@ -28,6 +31,7 @@ > #include <uapi/linux/virtio_iommu.h> > #include "iommu-pasid-table.h" > #include "iommu-sva-lib.h" > +#include "io-pgtable-arm.h" Is this used here? > > #define MSI_IOVA_BASE 0x8000000 > #define MSI_IOVA_LENGTH 0x100000 > @@ -41,6 +45,7 @@ DEFINE_XARRAY_ALLOC1(viommu_asid_xa); > > static DEFINE_MUTEX(sva_lock); > static DEFINE_MUTEX(iopf_lock); > +static DEFINE_MUTEX(viommu_asid_lock); > > struct viommu_dev_pri_work { > struct work_struct work; > @@ -88,10 +93,22 @@ struct viommu_mapping { > struct viommu_mm { > int pasid; > u64 archid; > + struct viommu_sva_bond *bond; > struct io_pgtable_ops *ops; > struct viommu_domain *domain; > }; > > +struct viommu_sva_bond { > + struct iommu_sva sva; > + struct mm_struct *mm; > + struct iommu_psdtable_mmu_notifier *viommu_mn; > + struct list_head list; > + refcount_t refs; > +}; > + > +#define sva_to_viommu_bond(handle) \ > + container_of(handle, struct viommu_sva_bond, sva) > + > struct viommu_domain { > struct iommu_domain domain; > struct viommu_dev *viommu; > @@ -136,6 +153,7 @@ struct viommu_endpoint { > bool pri_supported; > bool sva_enabled; > bool iopf_enabled; > + struct list_head bonds; > }; > > struct viommu_ep_entry { > @@ -1423,14 +1441,15 @@ static int viommu_attach_pasid_table(struct viommu_endpoint *vdev, > > pst_cfg->iommu_dev = viommu->dev->parent; > > + mutex_lock(&viommu_asid_lock); > /* Prepare PASID tables info to allocate a new table */ > ret = viommu_prepare_pst(vdev, pst_cfg, fmt); > if (ret) > - return ret; > + goto err_out_unlock; > > ret = iommu_psdtable_alloc(tbl, pst_cfg); > if (ret) > - return ret; > + goto err_out_unlock; > > pst_cfg->iommu_dev = viommu->dev->parent; > pst_cfg->fmt = PASID_TABLE_ARM_SMMU_V3; > @@ -1452,6 +1471,7 @@ static int viommu_attach_pasid_table(struct viommu_endpoint *vdev, > if (ret) > goto err_free_ops; > } > + mutex_unlock(&viommu_asid_lock); > } else { > /* TODO: otherwise, check for compatibility with vdev. */ > return -ENOSYS; > @@ -1467,6 +1487,8 @@ static int viommu_attach_pasid_table(struct viommu_endpoint *vdev, > err_free_psdtable: > iommu_psdtable_free(tbl, &tbl->cfg); > > +err_out_unlock: > + mutex_unlock(&viommu_asid_lock); > return ret; > } > > @@ -1706,6 +1728,7 @@ static struct iommu_device *viommu_probe_device(struct device *dev) > vdev->dev = dev; > vdev->viommu = viommu; > INIT_LIST_HEAD(&vdev->resv_regions); > + INIT_LIST_HEAD(&vdev->bonds); > dev_iommu_priv_set(dev, vdev); > > if (viommu->probe_size) { > @@ -1755,6 +1778,175 @@ static int viommu_of_xlate(struct device *dev, struct of_phandle_args *args) > return iommu_fwspec_add_ids(dev, args->args, 1); > } > > +static u32 viommu_sva_get_pasid(struct iommu_sva *handle) > +{ > + struct viommu_sva_bond *bond = sva_to_viommu_bond(handle); > + > + return bond->mm->pasid; > +} > + > +static void viommu_mmu_notifier_free(struct mmu_notifier *mn) > +{ > + kfree(mn_to_pstiommu(mn)); > +} > + > +static struct mmu_notifier_ops viommu_mmu_notifier_ops = { > + .free_notifier = viommu_mmu_notifier_free, .invalidate_range and .release will be needed as well, to keep up to date with changes to the address space > +}; > + > +/* Allocate or get existing MMU notifier for this {domain, mm} pair */ > +static struct iommu_psdtable_mmu_notifier * > +viommu_mmu_notifier_get(struct viommu_domain *vdomain, struct mm_struct *mm, > + u32 asid_bits) > +{ > + int ret; > + struct iommu_psdtable_mmu_notifier *viommu_mn; > + struct iommu_pasid_table *tbl = vdomain->pasid_tbl; > + > + list_for_each_entry(viommu_mn, &tbl->mmu_notifiers, list) { > + if (viommu_mn->mn.mm == mm) { > + refcount_inc(&viommu_mn->refs); > + return viommu_mn; > + } > + } > + > + mutex_lock(&viommu_asid_lock); > + viommu_mn = iommu_psdtable_alloc_shared(tbl, mm, &viommu_asid_xa, > + asid_bits); > + mutex_unlock(&viommu_asid_lock); > + if (IS_ERR(viommu_mn)) > + return ERR_CAST(viommu_mn); > + > + refcount_set(&viommu_mn->refs, 1); > + viommu_mn->cookie = vdomain; > + viommu_mn->mn.ops = &viommu_mmu_notifier_ops; > + > + ret = mmu_notifier_register(&viommu_mn->mn, mm); > + if (ret) > + goto err_free_cd; > + > + ret = iommu_psdtable_write(tbl, &tbl->cfg, mm->pasid, > + viommu_mn->vendor.cd); > + if (ret) > + goto err_put_notifier; > + > + list_add(&viommu_mn->list, &tbl->mmu_notifiers); > + return viommu_mn; > + > +err_put_notifier: > + /* Frees viommu_mn */ > + mmu_notifier_put(&viommu_mn->mn); > +err_free_cd: > + iommu_psdtable_free_shared(tbl, &viommu_asid_xa, viommu_mn->vendor.cd); > + return ERR_PTR(ret); > +} > + > +static void > +viommu_mmu_notifier_put(struct iommu_psdtable_mmu_notifier *viommu_mn) > +{ > + struct mm_struct *mm = viommu_mn->mn.mm; > + struct viommu_domain *vdomain = viommu_mn->cookie; > + struct iommu_pasid_table *tbl = vdomain->pasid_tbl; > + u16 asid = viommu_mn->vendor.cd->asid; > + > + if (!refcount_dec_and_test(&viommu_mn->refs)) > + return; > + > + list_del(&viommu_mn->list); > + iommu_psdtable_write(tbl, &tbl->cfg, mm->pasid, NULL); > + > + /* > + * If we went through clear(), we've already invalidated, and no > + * new TLB entry can have been formed. > + */ > + if (!viommu_mn->cleared) > + iommu_psdtable_flush_tlb(tbl, vdomain, asid); > + > + /* Frees smmu_mn */ > + mmu_notifier_put(&viommu_mn->mn); > + iommu_psdtable_free_shared(tbl, &viommu_asid_xa, viommu_mn->vendor.cd); > +} > + > +static struct iommu_sva * > +__viommu_sva_bind(struct device *dev, struct mm_struct *mm) > +{ > + int ret; > + struct viommu_sva_bond *bond; > + struct iommu_domain *domain = iommu_get_domain_for_dev(dev); > + struct viommu_domain *vdomain = to_viommu_domain(domain); > + struct viommu_endpoint *vdev = dev_iommu_priv_get(dev); > + struct virtio_iommu_probe_table_format *desc = vdev->pgtf; > + > + if (!vdev || !vdev->sva_enabled) > + return ERR_PTR(-ENODEV); > + > + /* If bind() was already called for this {dev, mm} pair, reuse it. */ > + list_for_each_entry(bond, &vdev->bonds, list) { > + if (bond->mm == mm) { > + refcount_inc(&bond->refs); > + return &bond->sva; > + } > + } > + > + bond = kzalloc(sizeof(*bond), GFP_KERNEL); > + if (!bond) > + return ERR_PTR(-ENOMEM); > + > + /* Allocate a PASID for this mm if necessary */ > + ret = iommu_sva_alloc_pasid(mm, 1, (1U << vdev->pasid_bits) - 1); > + if (ret) > + goto err_free_bond; > + > + bond->mm = mm; > + bond->sva.dev = dev; > + refcount_set(&bond->refs, 1); > + > + bond->viommu_mn = viommu_mmu_notifier_get(vdomain, mm, desc->asid_bits); > + if (IS_ERR(bond->viommu_mn)) { > + ret = PTR_ERR(bond->viommu_mn); > + goto err_free_pasid; > + } > + > + list_add(&bond->list, &vdev->bonds); > + return &bond->sva; > + > +err_free_pasid: > + iommu_sva_free_pasid(mm); > +err_free_bond: > + kfree(bond); > + return ERR_PTR(ret); > +} > + > +/* closely follows arm_smmu_sva_bind() */ > +static struct iommu_sva *viommu_sva_bind(struct device *dev, > + struct mm_struct *mm, void *drvdata) > +{ > + struct iommu_sva *handle; > + > + mutex_lock(&sva_lock); > + handle = __viommu_sva_bind(dev, mm); > + mutex_unlock(&sva_lock); > + return handle; > +} > + > +void viommu_sva_unbind(struct iommu_sva *handle) > +{ > + struct viommu_sva_bond *bond = sva_to_viommu_bond(handle); > + struct viommu_endpoint *vdev = dev_iommu_priv_get(handle->dev); > + > + if (vdev->pri_supported) > + iopf_queue_flush_dev(handle->dev); > + > + mutex_lock(&sva_lock); > + if (refcount_dec_and_test(&bond->refs)) { > + list_del(&bond->list); > + viommu_mmu_notifier_put(bond->viommu_mn); > + iommu_sva_free_pasid(bond->mm); > + kfree(bond); > + } > + mutex_unlock(&sva_lock); > +} > + > static bool viommu_endpoint_iopf_supported(struct viommu_endpoint *vdev) > { > /* TODO: support Stall model later */ > @@ -1960,6 +2152,9 @@ static struct iommu_ops viommu_ops = { > .dev_feat_enabled = viommu_dev_feature_enabled, > .dev_enable_feat = viommu_dev_enable_feature, > .dev_disable_feat = viommu_dev_disable_feature, > + .sva_bind = viommu_sva_bind, > + .sva_unbind = viommu_sva_unbind, > + .sva_get_pasid = viommu_sva_get_pasid, > }; > > static int viommu_init_vqs(struct viommu_dev *viommu) > diff --git a/include/uapi/linux/virtio_iommu.h b/include/uapi/linux/virtio_iommu.h > index 88a3db493108..c12d9b6a7243 100644 > --- a/include/uapi/linux/virtio_iommu.h > +++ b/include/uapi/linux/virtio_iommu.h > @@ -122,6 +122,8 @@ struct virtio_iommu_req_attach_pst_arm { > #define VIRTIO_IOMMU_PGTF_ARM_HPD0 (1ULL << 41) > #define VIRTIO_IOMMU_PGTF_ARM_EPD1 (1 << 23) > > +#define VIRTIO_IOMMU_PGTF_ARM_IPS_SHIFT 32 > +#define VIRTIO_IOMMU_PGTF_ARM_IPS_MASK 0x7 Probably not the right place for this change Thanks, Jean > #define VIRTIO_IOMMU_PGTF_ARM_TG0_SHIFT 14 > #define VIRTIO_IOMMU_PGTF_ARM_TG0_MASK 0x3 > #define VIRTIO_IOMMU_PGTF_ARM_SH0_SHIFT 12 > -- > 2.17.1 > _______________________________________________ Virtualization mailing list Virtualization@xxxxxxxxxxxxxxxxxxxxxxxxxx https://lists.linuxfoundation.org/mailman/listinfo/virtualization