Re: [PATCH V2 vfio 3/9] virtio-pci: Introduce admin command sending function

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Sun, Oct 29, 2023 at 05:59:46PM +0200, Yishai Hadas wrote:
> From: Feng Liu <feliu@xxxxxxxxxx>
> 
> Add support for sending admin command through admin virtqueue interface.
> Abort any inflight admin commands once device reset completes.
> 
> To enforce the below statement from the specification [1], the admin
> queue is activated for the upper layer users only post of setting status
> to DRIVER_OK.
> 
> [1] The driver MUST NOT send any buffer available notifications to the
> device before setting DRIVER_OK.
> 
> Signed-off-by: Feng Liu <feliu@xxxxxxxxxx>
> Reviewed-by: Parav Pandit <parav@xxxxxxxxxx>
> Signed-off-by: Yishai Hadas <yishaih@xxxxxxxxxx>
> ---
>  drivers/virtio/virtio_pci_common.h |   3 +
>  drivers/virtio/virtio_pci_modern.c | 174 +++++++++++++++++++++++++++++
>  include/linux/virtio.h             |   8 ++
>  include/uapi/linux/virtio_pci.h    |  22 ++++
>  4 files changed, 207 insertions(+)
> 
> diff --git a/drivers/virtio/virtio_pci_common.h b/drivers/virtio/virtio_pci_common.h
> index e03af0966a4b..a21b9ba01a60 100644
> --- a/drivers/virtio/virtio_pci_common.h
> +++ b/drivers/virtio/virtio_pci_common.h
> @@ -44,9 +44,12 @@ struct virtio_pci_vq_info {
>  struct virtio_pci_admin_vq {
>  	/* Virtqueue info associated with this admin queue. */
>  	struct virtio_pci_vq_info info;
> +	struct completion flush_done;
> +	refcount_t refcount;

what exactly does this count?

>  	/* Name of the admin queue: avq.$index. */
>  	char name[10];
>  	u16 vq_index;
> +	bool abort;
>  };
>  
>  /* Our device structure */
> diff --git a/drivers/virtio/virtio_pci_modern.c b/drivers/virtio/virtio_pci_modern.c
> index 01c5ba346471..ccd7a4d9f57f 100644
> --- a/drivers/virtio/virtio_pci_modern.c
> +++ b/drivers/virtio/virtio_pci_modern.c
> @@ -36,6 +36,58 @@ static bool vp_is_avq(struct virtio_device *vdev, unsigned int index)
>  	return index == vp_dev->admin_vq.vq_index;
>  }
>  
> +static bool vp_modern_avq_get(struct virtio_pci_admin_vq *admin_vq)
> +{
> +	return refcount_inc_not_zero(&admin_vq->refcount);
> +}
> +
> +static void vp_modern_avq_put(struct virtio_pci_admin_vq *admin_vq)
> +{
> +	if (refcount_dec_and_test(&admin_vq->refcount))
> +		complete(&admin_vq->flush_done);
> +}
> +
> +static bool vp_modern_avq_is_abort(const struct virtio_pci_admin_vq *admin_vq)
> +{
> +	return READ_ONCE(admin_vq->abort);
> +}
> +
> +static void
> +vp_modern_avq_set_abort(struct virtio_pci_admin_vq *admin_vq, bool abort)
> +{
> +	/* Mark the AVQ to abort, so that inflight commands can be aborted. */
> +	WRITE_ONCE(admin_vq->abort, abort);
> +}
> +
> +static void vp_modern_avq_activate(struct virtio_device *vdev)
> +{
> +	struct virtio_pci_device *vp_dev = to_vp_device(vdev);
> +	struct virtio_pci_admin_vq *admin_vq = &vp_dev->admin_vq;
> +
> +	if (!virtio_has_feature(vdev, VIRTIO_F_ADMIN_VQ))
> +		return;
> +
> +	init_completion(&admin_vq->flush_done);
> +	refcount_set(&admin_vq->refcount, 1);
> +	vp_modern_avq_set_abort(admin_vq, false);
> +}
> +
> +static void vp_modern_avq_deactivate(struct virtio_device *vdev)
> +{
> +	struct virtio_pci_device *vp_dev = to_vp_device(vdev);
> +	struct virtio_pci_admin_vq *admin_vq = &vp_dev->admin_vq;
> +
> +	if (!virtio_has_feature(vdev, VIRTIO_F_ADMIN_VQ))
> +		return;
> +
> +	vp_modern_avq_set_abort(admin_vq, true);
> +	/* Balance with refcount_set() during vp_modern_avq_activate */
> +	vp_modern_avq_put(admin_vq);
> +
> +	/* Wait for all the inflight admin commands to be aborted */
> +	wait_for_completion(&vp_dev->admin_vq.flush_done);
> +}
> +
>  static void vp_transport_features(struct virtio_device *vdev, u64 features)
>  {
>  	struct virtio_pci_device *vp_dev = to_vp_device(vdev);
> @@ -172,6 +224,8 @@ static void vp_set_status(struct virtio_device *vdev, u8 status)
>  	/* We should never be setting status to 0. */
>  	BUG_ON(status == 0);
>  	vp_modern_set_status(&vp_dev->mdev, status);
> +	if (status & VIRTIO_CONFIG_S_DRIVER_OK)
> +		vp_modern_avq_activate(vdev);
>  }
>  
>  static void vp_reset(struct virtio_device *vdev)
> @@ -188,6 +242,9 @@ static void vp_reset(struct virtio_device *vdev)
>  	 */
>  	while (vp_modern_get_status(mdev))
>  		msleep(1);
> +
> +	vp_modern_avq_deactivate(vdev);
> +
>  	/* Flush pending VQ/configuration callbacks. */
>  	vp_synchronize_vectors(vdev);
>  }
> @@ -505,6 +562,121 @@ static bool vp_get_shm_region(struct virtio_device *vdev,
>  	return true;
>  }
>  
> +static int virtqueue_exec_admin_cmd(struct virtio_pci_admin_vq *admin_vq,
> +				    struct scatterlist **sgs,
> +				    unsigned int out_num,
> +				    unsigned int in_num,
> +				    void *data,
> +				    gfp_t gfp)
> +{
> +	struct virtqueue *vq;
> +	int ret, len;
> +
> +	if (!vp_modern_avq_get(admin_vq))
> +		return -EIO;
> +
> +	vq = admin_vq->info.vq;
> +
> +	ret = virtqueue_add_sgs(vq, sgs, out_num, in_num, data, gfp);
> +	if (ret < 0)
> +		goto out;
> +

virtqueue_add_sgs requires locking, it is not atomic.
Add some kind of mutex?
Will also help prevent vq overruns.

> +	if (unlikely(!virtqueue_kick(vq))) {
> +		ret = -EIO;
> +		goto out;
> +	}




> +
> +	while (!virtqueue_get_buf(vq, &len) &&
> +	       !virtqueue_is_broken(vq) &&
> +	       !vp_modern_avq_is_abort(admin_vq))
> +		cpu_relax();
> +
> +	if (vp_modern_avq_is_abort(admin_vq) || virtqueue_is_broken(vq)) {
> +		ret = -EIO;
> +		goto out;
> +	}
> +out:
> +	vp_modern_avq_put(admin_vq);
> +	return ret;
> +}
> +
> +#define VIRTIO_AVQ_SGS_MAX	4
> +
> +static int vp_modern_admin_cmd_exec(struct virtio_device *vdev,
> +				    struct virtio_admin_cmd *cmd)
> +{
> +	struct scatterlist *sgs[VIRTIO_AVQ_SGS_MAX], hdr, stat;
> +	struct virtio_pci_device *vp_dev = to_vp_device(vdev);
> +	struct virtio_admin_cmd_status *va_status;
> +	unsigned int out_num = 0, in_num = 0;
> +	struct virtio_admin_cmd_hdr *va_hdr;
> +	struct virtqueue *avq;
> +	u16 status;
> +	int ret;
> +
> +	avq = virtio_has_feature(vdev, VIRTIO_F_ADMIN_VQ) ?
> +		vp_dev->admin_vq.info.vq : NULL;
> +	if (!avq)
> +		return -EOPNOTSUPP;
> +
> +	va_status = kzalloc(sizeof(*va_status), GFP_KERNEL);
> +	if (!va_status)
> +		return -ENOMEM;
> +
> +	va_hdr = kzalloc(sizeof(*va_hdr), GFP_KERNEL);
> +	if (!va_hdr) {
> +		ret = -ENOMEM;
> +		goto err_alloc;
> +	}
> +
> +	va_hdr->opcode = cmd->opcode;
> +	va_hdr->group_type = cmd->group_type;
> +	va_hdr->group_member_id = cmd->group_member_id;
> +
> +	/* Add header */
> +	sg_init_one(&hdr, va_hdr, sizeof(*va_hdr));
> +	sgs[out_num] = &hdr;
> +	out_num++;
> +
> +	if (cmd->data_sg) {
> +		sgs[out_num] = cmd->data_sg;
> +		out_num++;
> +	}
> +
> +	/* Add return status */
> +	sg_init_one(&stat, va_status, sizeof(*va_status));
> +	sgs[out_num + in_num] = &stat;
> +	in_num++;
> +
> +	if (cmd->result_sg) {
> +		sgs[out_num + in_num] = cmd->result_sg;
> +		in_num++;
> +	}


BUG_ON here to check we did not overrun sgs?

> +
> +	ret = virtqueue_exec_admin_cmd(&vp_dev->admin_vq, sgs,
> +				       out_num, in_num,
> +				       sgs, GFP_KERNEL);
> +	if (ret) {
> +		dev_err(&vdev->dev,
> +			"Failed to execute command on admin vq: %d\n.", ret);
> +		goto err_cmd_exec;
> +	}
> +
> +	status = le16_to_cpu(va_status->status);
> +	if (status != VIRTIO_ADMIN_STATUS_OK) {
> +		dev_err(&vdev->dev,
> +			"admin command error: status(%#x) qualifier(%#x)\n",
> +			status, le16_to_cpu(va_status->status_qualifier));
> +		ret = -status;
> +	}
> +
> +err_cmd_exec:
> +	kfree(va_hdr);
> +err_alloc:
> +	kfree(va_status);
> +	return ret;
> +}
> +
>  static int vp_modern_create_avq(struct virtio_device *vdev)
>  {
>  	struct virtio_pci_device *vp_dev = to_vp_device(vdev);
> @@ -530,6 +702,7 @@ static int vp_modern_create_avq(struct virtio_device *vdev)
>  		return PTR_ERR(vq);
>  	}
>  
> +	refcount_set(&vp_dev->admin_vq.refcount, 0);
>  	vp_dev->admin_vq.info.vq = vq;
>  	vp_modern_set_queue_enable(&vp_dev->mdev, avq->info.vq->index, true);
>  	return 0;
> @@ -542,6 +715,7 @@ static void vp_modern_destroy_avq(struct virtio_device *vdev)
>  	if (!virtio_has_feature(vdev, VIRTIO_F_ADMIN_VQ))
>  		return;
>  
> +	WARN_ON(refcount_read(&vp_dev->admin_vq.refcount));
>  	vp_dev->del_vq(&vp_dev->admin_vq.info);
>  }
>  
> diff --git a/include/linux/virtio.h b/include/linux/virtio.h
> index 4cc614a38376..b0201747a263 100644
> --- a/include/linux/virtio.h
> +++ b/include/linux/virtio.h
> @@ -103,6 +103,14 @@ int virtqueue_resize(struct virtqueue *vq, u32 num,
>  int virtqueue_reset(struct virtqueue *vq,
>  		    void (*recycle)(struct virtqueue *vq, void *buf));
>  
> +struct virtio_admin_cmd {
> +	__le16 opcode;
> +	__le16 group_type;
> +	__le64 group_member_id;
> +	struct scatterlist *data_sg;
> +	struct scatterlist *result_sg;
> +};
> +
>  /**
>   * struct virtio_device - representation of a device using virtio
>   * @index: unique position on the virtio bus
> diff --git a/include/uapi/linux/virtio_pci.h b/include/uapi/linux/virtio_pci.h
> index f703afc7ad31..68eacc9676dc 100644
> --- a/include/uapi/linux/virtio_pci.h
> +++ b/include/uapi/linux/virtio_pci.h
> @@ -207,4 +207,26 @@ struct virtio_pci_cfg_cap {
>  
>  #endif /* VIRTIO_PCI_NO_MODERN */
>  
> +/* Admin command status. */
> +#define VIRTIO_ADMIN_STATUS_OK		0
> +
> +struct __packed virtio_admin_cmd_hdr {
> +	__le16 opcode;
> +	/*
> +	 * 1 - SR-IOV
> +	 * 2-65535 - reserved
> +	 */
> +	__le16 group_type;
> +	/* Unused, reserved for future extensions. */
> +	__u8 reserved1[12];
> +	__le64 group_member_id;
> +};
> +
> +struct __packed virtio_admin_cmd_status {
> +	__le16 status;
> +	__le16 status_qualifier;
> +	/* Unused, reserved for future extensions. */
> +	__u8 reserved2[4];
> +};
> +
>  #endif
> -- 
> 2.27.0

_______________________________________________
Virtualization mailing list
Virtualization@xxxxxxxxxxxxxxxxxxxxxxxxxx
https://lists.linuxfoundation.org/mailman/listinfo/virtualization



[Index of Archives]     [KVM Development]     [Libvirt Development]     [Libvirt Users]     [CentOS Virtualization]     [Netdev]     [Ethernet Bridging]     [Linux Wireless]     [Kernel Newbies]     [Security]     [Linux for Hams]     [Netfilter]     [Bugtraq]     [Yosemite Forum]     [MIPS Linux]     [ARM Linux]     [Linux RAID]     [Linux Admin]     [Samba]

  Powered by Linux