Re: [RFC PATCH 3/3] kvm: Add VFIO device for handling IOMMU cache coherency

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Fri, Sep 13, 2013 at 08:13:40AM -0600, Alex Williamson wrote:
> On Fri, 2013-09-13 at 15:39 +0300, Michael S. Tsirkin wrote:
> > On Thu, Sep 12, 2013 at 03:23:15PM -0600, Alex Williamson wrote:
> > > So far we've succeeded at making KVM and VFIO mostly unaware of each
> > > other, but there's any important point where that breaks down.  Intel
> > > VT-d hardware may or may not support snoop control.  When snoop
> > > control is available, intel-iommu promotes No-Snoop transactions on
> > > PCIe to be cache coherent.  That allows KVM to handle things like the
> > > x86 WBINVD opcode as a nop.  When the hardware does not support this,
> > > KVM must implement a hardware visible WBINVD for the guest.
> > > 
> > > We could simply let userspace tell KVM how to handle WBINVD, but it's
> > > privileged for a reason.  Allowing an arbitrary user to enable
> > > physical WBINVD gives them a more access to the hardware.  Previously,
> > > this has only been enabled for guests supporting legacy PCI device
> > > assignment.  In such cases it's necessary for proper guest execution.
> > > We therefore create a new KVM-VFIO virtual device.  The user can add
> > > and remove VFIO groups to this device via file descriptors.  KVM
> > > makes use of the VFIO external user interface to validate that the
> > > user has access to physical hardware and gets the coherency state of
> > > the IOMMU from VFIO.  This provides equivalent functionality to
> > > legacy KVM assignment, while keeping (nearly) all the bits isolated.
> > 
> > 
> > So how is the isolation handled then?
> 
> By isolation above, I'm only talking about the intrusive-ness of the
> code and "leaking" vfio knowledge into KVM.  Nothing to do with device
> isolation.
> 
> > How is this better than a ioctl to grant WBINVD to guest?
> > kvm char device can be opened by any user,
> > so any user can grant itself these priveledges.
> > What did I miss?
> 
> With this interface the caller must have physical access to one or more
> devices via vfio groups and each of those groups must be configured into
> one or more IOMMU domains.  Furthermore, at least one of the IOMMU
> domains must not include the IOMMU_CAP_CACHE_COHERENCY capability.  So
> it's actually quite a significantly higher hurdle than an ioctl open to
> anyone and the number of VMs on a given host capable of doing this is
> bound by the number of IOMMU groups.  We do not however verify that a
> vfio device is actually in use by the VM, but I don't think there's a
> way to do that from KVM and I'm not sure that it's important to do so.
> I believe having access to physical hardware is already a sufficient
> granting of privilege to enable things like WBINVD.  Thanks,
> 
> Alex

Fair enough, but how about revoking the priveledge?
For example, device can be removed by hotplug - does
priveledge remain?
Is this important at all?


> > > The one intrusion is the resulting flag indicating the coherency
> > > state.  For this RFC it's placed on the x86 kvm_arch struct, however
> > > I know POWER has interest in using the VFIO external user interface,
> > > and I'm hoping we can share a common KVM-VFIO device.  Perhaps they
> > > care about No-Snoop handling as well or the code can be #ifdef'd.
> > > 
> > > Signed-off-by: Alex Williamson <alex.williamson@xxxxxxxxxx>
> > > ---
> > >  Documentation/virtual/kvm/devices/vfio.txt |   22 +++
> > >  arch/x86/include/asm/kvm_host.h            |    1 
> > >  arch/x86/kvm/Makefile                      |    2 
> > >  arch/x86/kvm/vmx.c                         |    5 -
> > >  arch/x86/kvm/x86.c                         |    5 -
> > >  include/linux/kvm_host.h                   |    1 
> > >  include/uapi/linux/kvm.h                   |    4 
> > >  virt/kvm/kvm_main.c                        |    3 
> > >  virt/kvm/vfio.c                            |  237 ++++++++++++++++++++++++++++
> > >  9 files changed, 275 insertions(+), 5 deletions(-)
> > >  create mode 100644 Documentation/virtual/kvm/devices/vfio.txt
> > >  create mode 100644 virt/kvm/vfio.c
> > > 
> > > diff --git a/Documentation/virtual/kvm/devices/vfio.txt b/Documentation/virtual/kvm/devices/vfio.txt
> > > new file mode 100644
> > > index 0000000..831e6a6
> > > --- /dev/null
> > > +++ b/Documentation/virtual/kvm/devices/vfio.txt
> > > @@ -0,0 +1,22 @@
> > > +VFIO virtual device
> > > +===================
> > > +
> > > +Device types supported:
> > > +  KVM_DEV_TYPE_VFIO
> > > +
> > > +Only one VFIO instance may be created per VM.  The created device
> > > +tracks VFIO groups in use by the VM and features of those groups
> > > +important to the correctness and acceleration of the VM.  As groups
> > > +are enabled and disabled for use by the VM, KVM should be updated
> > > +about their presence.  When registered with KVM, a reference to the
> > > +VFIO-group is held by KVM.
> > > +
> > > +Groups:
> > > +  KVM_DEV_VFIO_ADD_GROUP
> > > +  KVM_DEV_VFIO_DEL_GROUP
> > > +
> > > +Each takes a int32_t file descriptor for kvm_device_attr.addr and
> > > +does not support any group device kvm_device_attr.attr.
> > > +
> > > +RFC - Should we use Group KVM_DEV_VFIO_GROUP with Attributes
> > > +      KVM_DEV_VFIO_GROUP_ADD & DEL?
> > > diff --git a/arch/x86/include/asm/kvm_host.h b/arch/x86/include/asm/kvm_host.h
> > > index c76ff74..5b9350d 100644
> > > --- a/arch/x86/include/asm/kvm_host.h
> > > +++ b/arch/x86/include/asm/kvm_host.h
> > > @@ -588,6 +588,7 @@ struct kvm_arch {
> > >  
> > >  	spinlock_t pvclock_gtod_sync_lock;
> > >  	bool use_master_clock;
> > > +	bool vfio_noncoherent;
> > >  	u64 master_kernel_ns;
> > >  	cycle_t master_cycle_now;
> > >  
> > > diff --git a/arch/x86/kvm/Makefile b/arch/x86/kvm/Makefile
> > > index bf4fb04..25d22b2 100644
> > > --- a/arch/x86/kvm/Makefile
> > > +++ b/arch/x86/kvm/Makefile
> > > @@ -9,7 +9,7 @@ KVM := ../../../virt/kvm
> > >  
> > >  kvm-y			+= $(KVM)/kvm_main.o $(KVM)/ioapic.o \
> > >  				$(KVM)/coalesced_mmio.o $(KVM)/irq_comm.o \
> > > -				$(KVM)/eventfd.o $(KVM)/irqchip.o
> > > +				$(KVM)/eventfd.o $(KVM)/irqchip.o $(KVM)/vfio.o
> > >  kvm-$(CONFIG_KVM_DEVICE_ASSIGNMENT)	+= $(KVM)/assigned-dev.o $(KVM)/iommu.o
> > >  kvm-$(CONFIG_KVM_ASYNC_PF)	+= $(KVM)/async_pf.o
> > >  
> > > diff --git a/arch/x86/kvm/vmx.c b/arch/x86/kvm/vmx.c
> > > index 1f1da43..94f7786 100644
> > > --- a/arch/x86/kvm/vmx.c
> > > +++ b/arch/x86/kvm/vmx.c
> > > @@ -7395,8 +7395,9 @@ static u64 vmx_get_mt_mask(struct kvm_vcpu *vcpu, gfn_t gfn, bool is_mmio)
> > >  	 */
> > >  	if (is_mmio)
> > >  		ret = MTRR_TYPE_UNCACHABLE << VMX_EPT_MT_EPTE_SHIFT;
> > > -	else if (vcpu->kvm->arch.iommu_domain &&
> > > -		!(vcpu->kvm->arch.iommu_flags & KVM_IOMMU_CACHE_COHERENCY))
> > > +	else if (vcpu->kvm->arch.vfio_noncoherent ||
> > > +		 vcpu->kvm->arch.iommu_domain &&
> > > +		 !(vcpu->kvm->arch.iommu_flags & KVM_IOMMU_CACHE_COHERENCY))
> > >  		ret = kvm_get_guest_memory_type(vcpu, gfn) <<
> > >  		      VMX_EPT_MT_EPTE_SHIFT;
> > >  	else
> > > diff --git a/arch/x86/kvm/x86.c b/arch/x86/kvm/x86.c
> > > index e5ca72a..406ba6f 100644
> > > --- a/arch/x86/kvm/x86.c
> > > +++ b/arch/x86/kvm/x86.c
> > > @@ -2715,8 +2715,9 @@ static void wbinvd_ipi(void *garbage)
> > >  
> > >  static bool need_emulate_wbinvd(struct kvm_vcpu *vcpu)
> > >  {
> > > -	return vcpu->kvm->arch.iommu_domain &&
> > > -		!(vcpu->kvm->arch.iommu_flags & KVM_IOMMU_CACHE_COHERENCY);
> > > +	return vcpu->kvm->arch.vfio_noncoherent ||
> > > +	       (vcpu->kvm->arch.iommu_domain &&
> > > +		!(vcpu->kvm->arch.iommu_flags & KVM_IOMMU_CACHE_COHERENCY));
> > >  }
> > >  
> > >  void kvm_arch_vcpu_load(struct kvm_vcpu *vcpu, int cpu)
> > > diff --git a/include/linux/kvm_host.h b/include/linux/kvm_host.h
> > > index ca645a0..615f0c3 100644
> > > --- a/include/linux/kvm_host.h
> > > +++ b/include/linux/kvm_host.h
> > > @@ -1065,6 +1065,7 @@ struct kvm_device *kvm_device_from_filp(struct file *filp);
> > >  
> > >  extern struct kvm_device_ops kvm_mpic_ops;
> > >  extern struct kvm_device_ops kvm_xics_ops;
> > > +extern struct kvm_device_ops kvm_vfio_ops;
> > >  
> > >  #ifdef CONFIG_HAVE_KVM_CPU_RELAX_INTERCEPT
> > >  
> > > diff --git a/include/uapi/linux/kvm.h b/include/uapi/linux/kvm.h
> > > index 99c2533..8869616 100644
> > > --- a/include/uapi/linux/kvm.h
> > > +++ b/include/uapi/linux/kvm.h
> > > @@ -843,6 +843,10 @@ struct kvm_device_attr {
> > >  #define KVM_DEV_TYPE_FSL_MPIC_20	1
> > >  #define KVM_DEV_TYPE_FSL_MPIC_42	2
> > >  #define KVM_DEV_TYPE_XICS		3
> > > +#define KVM_DEV_TYPE_VFIO		4
> > > +
> > > +#define KVM_DEV_VFIO_ADD_GROUP		1
> > > +#define KVM_DEV_VFIO_DEL_GROUP		2
> > >  
> > >  /*
> > >   * ioctls for VM fds
> > > diff --git a/virt/kvm/kvm_main.c b/virt/kvm/kvm_main.c
> > > index d9cad4d..1a20425 100644
> > > --- a/virt/kvm/kvm_main.c
> > > +++ b/virt/kvm/kvm_main.c
> > > @@ -2269,6 +2269,9 @@ static int kvm_ioctl_create_device(struct kvm *kvm,
> > >  		ops = &kvm_xics_ops;
> > >  		break;
> > >  #endif
> > > +	case KVM_DEV_TYPE_VFIO:
> > > +		ops = &kvm_vfio_ops;
> > > +		break;
> > >  	default:
> > >  		return -ENODEV;
> > >  	}
> > > diff --git a/virt/kvm/vfio.c b/virt/kvm/vfio.c
> > > new file mode 100644
> > > index 0000000..9a2faff
> > > --- /dev/null
> > > +++ b/virt/kvm/vfio.c
> > > @@ -0,0 +1,237 @@
> > > +/*
> > > + * VFIO bridge
> > > + *
> > > + * Copyright (C) 2013 Red Hat, Inc.  All rights reserved.
> > > + *     Author: Alex Williamson <alex.williamson@xxxxxxxxxx>
> > > + *
> > > + * This program is free software; you can redistribute it and/or modify
> > > + * it under the terms of the GNU General Public License version 2 as
> > > + * published by the Free Software Foundation.
> > > + */
> > > +
> > > +#include <linux/errno.h>
> > > +#include <linux/file.h>
> > > +#include <linux/kvm_host.h>
> > > +#include <linux/list.h>
> > > +#include <linux/module.h>
> > > +#include <linux/mutex.h>
> > > +#include <linux/slab.h>
> > > +#include <linux/vfio.h>
> > > +
> > > +struct kvm_vfio_group {
> > > +	struct list_head node;
> > > +	struct vfio_group *vfio_group;
> > > +};
> > > +
> > > +struct kvm_vfio {
> > > +	struct list_head group_list;
> > > +	struct mutex lock;
> > > +};
> > > +
> > > +static struct vfio_group *kvm_vfio_group_get_external_user(struct file *filep)
> > > +{
> > > +	struct vfio_group *vfio_group;
> > > +	struct vfio_group *(*fn)(struct file *);
> > > +
> > > +	fn = symbol_get(vfio_group_get_external_user);
> > > +	if (!fn)
> > > +		return ERR_PTR(-EINVAL);
> > > +
> > > +	vfio_group = fn(filep);
> > > +
> > > +	symbol_put(vfio_group_get_external_user);
> > > +
> > > +	return vfio_group;
> > > +}
> > > +
> > > +static void kvm_vfio_group_put_external_user(struct vfio_group *vfio_group)
> > > +{
> > > +	void (*fn)(struct vfio_group *);
> > > +
> > > +	fn = symbol_get(vfio_group_put_external_user);
> > > +	if (!fn)
> > > +		return;
> > > +
> > > +	fn(vfio_group);
> > > +
> > > +	symbol_put(vfio_group_put_external_user);
> > > +}
> > > +
> > > +static int kvm_vfio_external_user_check_extension(struct vfio_group *vfio_group,
> > > +						  unsigned long arg)
> > > +{
> > > +	int (*fn)(struct vfio_group *, unsigned long);
> > > +	int ret;
> > > +
> > > +	fn = symbol_get(vfio_external_user_check_extension);
> > > +	if (!fn)
> > > +		return -EINVAL;
> > > +
> > > +	ret = fn(vfio_group, arg);
> > > +
> > > +	symbol_put(vfio_group_put_external_user);
> > > +
> > > +	return ret;
> > > +}
> > > +
> > > +static void kvm_vfio_update_iommu_coherency(struct kvm_device *dev)
> > > +{
> > > +	struct kvm_vfio *kv = dev->private;
> > > +	bool coherent = true;
> > > +	struct kvm_vfio_group *kvg;
> > > +
> > > +	mutex_lock(&kv->lock);
> > > +
> > > +	list_for_each_entry(kvg, &kv->group_list, node) {
> > > +		if (!kvm_vfio_external_user_check_extension(kvg->vfio_group,
> > > +					VFIO_IOMMU_CAP_CACHE_COHERENCY)) {
> > > +			coherent = false;
> > > +			break;
> > > +		}
> > > +	}
> > > +
> > > +	mutex_unlock(&kv->lock);
> > > +
> > > +	dev->kvm->arch.vfio_noncoherent = !coherent;
> > > +}
> > > +
> > > +static int kvm_vfio_set_attr(struct kvm_device *dev,
> > > +			     struct kvm_device_attr *attr)
> > > +{
> > > +	struct kvm_vfio *kv = dev->private;
> > > +	struct fd f;
> > > +	struct vfio_group *vfio_group;
> > > +	struct kvm_vfio_group *kvg;
> > > +	int ret;
> > > +
> > > +	switch (attr->group) {
> > > +	case KVM_DEV_VFIO_ADD_GROUP:
> > > +		f = fdget(attr->addr);
> > > +		if (!f.file)
> > > +			return -EBADF;
> > > +
> > > +		vfio_group = kvm_vfio_group_get_external_user(f.file);
> > > +		fdput(f);
> > > +
> > > +		if (IS_ERR(vfio_group))
> > > +			return PTR_ERR(vfio_group);
> > > +
> > > +		mutex_lock(&kv->lock);
> > > +
> > > +		list_for_each_entry(kvg, &kv->group_list, node) {
> > > +			if (kvg->vfio_group == vfio_group) {
> > > +				mutex_unlock(&kv->lock);
> > > +				kvm_vfio_group_put_external_user(vfio_group);
> > > +				return -EEXIST;
> > > +			}
> > > +		}
> > > +
> > > +		kvg = kzalloc(sizeof(*kvg), GFP_KERNEL);
> > > +		if (!kvg) {
> > > +			mutex_unlock(&kv->lock);
> > > +			kvm_vfio_group_put_external_user(vfio_group);
> > > +			return -ENOMEM;
> > > +		}
> > > +
> > > +		list_add_tail(&kvg->node, &kv->group_list);
> > > +		kvg->vfio_group = vfio_group;
> > > +
> > > +		mutex_unlock(&kv->lock);
> > > +
> > > +		kvm_vfio_update_iommu_coherency(dev);
> > > +
> > > +		return 0;
> > > +
> > > +	case KVM_DEV_VFIO_DEL_GROUP:
> > > +		f = fdget(attr->addr);
> > > +		if (!f.file)
> > > +			return -EBADF;
> > > +
> > > +		vfio_group = kvm_vfio_group_get_external_user(f.file);
> > > +		fdput(f);
> > > +
> > > +		if (IS_ERR(vfio_group))
> > > +			return PTR_ERR(vfio_group);
> > > +
> > > +		ret = -ENOENT;
> > > +
> > > +		mutex_lock(&kv->lock);
> > > +
> > > +		list_for_each_entry(kvg, &kv->group_list, node) {
> > > +			if (kvg->vfio_group != vfio_group)
> > > +				continue;
> > > +
> > > +			list_del(&kvg->node);
> > > +			kvm_vfio_group_put_external_user(kvg->vfio_group);
> > > +			kfree(kvg);
> > > +			ret = 0;
> > > +			break;
> > > +		}
> > > +
> > > +		mutex_unlock(&kv->lock);
> > > +		kvm_vfio_group_put_external_user(vfio_group);
> > > +
> > > +		kvm_vfio_update_iommu_coherency(dev);
> > > +
> > > +		return ret;
> > > +	}
> > > +
> > > +	return -ENXIO;
> > > +}
> > > +
> > > +static int kvm_vfio_has_attr(struct kvm_device *dev,
> > > +				   struct kvm_device_attr *attr)
> > > +{
> > > +	switch (attr->group) {
> > > +	case KVM_DEV_VFIO_ADD_GROUP:
> > > +	case KVM_DEV_VFIO_DEL_GROUP:
> > > +		return 0;
> > > +	}
> > > +
> > > +	return -ENXIO;
> > > +}
> > > +
> > > +static void kvm_vfio_destroy(struct kvm_device *dev)
> > > +{
> > > +	struct kvm_vfio *kv = dev->private;
> > > +	struct kvm_vfio_group *kvg, *tmp;
> > > +
> > > +	list_for_each_entry_safe(kvg, tmp, &kv->group_list, node) {
> > > +		kvm_vfio_group_put_external_user(kvg->vfio_group);
> > > +		list_del(&kvg->node);
> > > +		kfree(kvg);
> > > +	}
> > > +
> > > +	dev->kvm->arch.vfio_noncoherent = false;
> > > +	kfree(kv);
> > > +}
> > > +
> > > +static int kvm_vfio_create(struct kvm_device *dev, u32 type)
> > > +{
> > > +	struct kvm_device *tmp;
> > > +	struct kvm_vfio *kv;
> > > +
> > > +	/* Only one VFIO "device" per VM */
> > > +	list_for_each_entry(tmp, &dev->kvm->devices, vm_node)
> > > +		if (tmp->ops == &kvm_vfio_ops)
> > > +			return -EBUSY;
> > > +
> > > +	kv = kzalloc(sizeof(*kv), GFP_KERNEL);
> > > +	if (!kv)
> > > +		return -ENOMEM;
> > > +
> > > +	INIT_LIST_HEAD(&kv->group_list);
> > > +	mutex_init(&kv->lock);
> > > +
> > > +	dev->private = kv;
> > > +
> > > +	return 0;
> > > +}
> > > +
> > > +struct kvm_device_ops kvm_vfio_ops = {
> > > +	.name = "kvm-vfio",
> > > +	.create = kvm_vfio_create,
> > > +	.destroy = kvm_vfio_destroy,
> > > +	.set_attr = kvm_vfio_set_attr,
> > > +	.has_attr = kvm_vfio_has_attr,
> > > +};
> 
> 
--
To unsubscribe from this list: send the line "unsubscribe kvm" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html




[Index of Archives]     [KVM ARM]     [KVM ia64]     [KVM ppc]     [Virtualization Tools]     [Spice Development]     [Libvirt]     [Libvirt Users]     [Linux USB Devel]     [Linux Audio Users]     [Yosemite Questions]     [Linux Kernel]     [Linux SCSI]     [XFree86]
  Powered by Linux