On Fri, Dec 11, 2015 at 01:45:29PM +0000, Alex Bennée wrote: > > + if (head == vq->num) { > > + if (unlikely(vhost_enable_notify(&vsock->dev, vq))) { > > + vhost_disable_notify(&vsock->dev, vq); > > + continue; > > Why are we doing this? If we enable something we then disable it? A > comment as to what is going on here would be useful. This is a standard optimization to avoid vmexits that other vhost devices and QEMU implement too. When the host begins pulling buffers off a virtqueue it first disables guest->host notifications. If the guest adds additional buffers while the host is processing, the notification (vmexit) is skipped. The host re-enables guest->host notifications when it finishes virtqueue processing. If the guest added buffers after vhost_get_vq_desc() but before vhost_enable_notify(), then vhost_enable_notify() returns true and the host must process the buffers (i.e. restart the loop). Failure to do so could result in deadlocks because the guest didn't notify and the host would be waiting for a notification. I will add comments to the code. > > + vhost_add_used(vq, head, pkt->len); /* TODO should this > > be sizeof(pkt->hdr) + pkt->len? */ > > TODO needs sorting our or removing. Will fix in the next revision. > > + /* Respect global tx buf limitation */ > > + mutex_lock(&vq->mutex); > > + while (pkt_len + vsock->total_tx_buf > > > VIRTIO_VSOCK_MAX_TX_BUF_SIZE) { > > I'm curious about the relationship between > VIRTIO_VSOCK_DEFAULT_RX_BUF_SIZE above and VIRTIO_VSOCK_MAX_TX_BUF_SIZE > just here. Why do we need to limit pkt_len to the smaller when really > all that matters is pkt_len + vsock->total_tx_buf > > VIRTIO_VSOCK_MAX_TX_BUF_SIZE? There are two separate issues: 1. The total amount of pending data. The idea is to stop queuing packets and make the caller wait until resources become available so that vhost_vsock.ko memory consumption is bounded. total_tx_buf len is an artificial limit that is lower than the actual virtqueue maximum data size. Otherwise we could just rely on the virtqueue to limit the size but it can be very large. 2. Splitting data into packets that fit into rx virtqueue buffers. The guest sets up the rx virtqueue with VIRTIO_VSOCK_DEFAULT_RX_BUF_SIZE buffers. Here, vhost_vsock.ko is assuming that the rx virtqueue buffers are always VIRTIO_VSOCK_DEFAULT_RX_BUF_SIZE bytes so it splits data along this boundary. This is ugly because the guest could choose a different buffer size and the host has VIRTIO_VSOCK_DEFAULT_RX_BUF_SIZE hardcoded. I'll look into eliminating this assumption. > > +static void vhost_vsock_handle_ctl_kick(struct vhost_work *work) > > +{ > > + struct vhost_virtqueue *vq = container_of(work, struct vhost_virtqueue, > > + poll.work); > > + struct vhost_vsock *vsock = container_of(vq->dev, struct vhost_vsock, > > + dev); > > + > > + pr_debug("%s vq=%p, vsock=%p\n", __func__, vq, vsock); > > +} > > This doesn't handle anything, it just prints debug stuff. Should this be > a NOP function? The control virtqueue is currently not used. In the next revision this function will be dropped. > > +static int vhost_vsock_set_features(struct vhost_vsock *vsock, u64 features) > > +{ > > + struct vhost_virtqueue *vq; > > + int i; > > + > > + if (features & ~VHOST_VSOCK_FEATURES) > > + return -EOPNOTSUPP; > > + > > + mutex_lock(&vsock->dev.mutex); > > + if ((features & (1 << VHOST_F_LOG_ALL)) && > > + !vhost_log_access_ok(&vsock->dev)) { > > + mutex_unlock(&vsock->dev.mutex); > > + return -EFAULT; > > + } > > + > > + for (i = 0; i < VSOCK_VQ_MAX; i++) { > > + vq = &vsock->vqs[i].vq; > > + mutex_lock(&vq->mutex); > > + vq->acked_features = features; > > Is this a user supplied flag? Should it be masked to valid values? That is already done above where VHOST_VSOCK_FEATURES is checked.
Attachment:
signature.asc
Description: PGP signature
_______________________________________________ Virtualization mailing list Virtualization@xxxxxxxxxxxxxxxxxxxxxxxxxx https://lists.linuxfoundation.org/mailman/listinfo/virtualization