Re: Re: [RFC v2] virtio-vsock: add description for datagram type

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Mon, May 17, 2021 at 4:02 AM Stefano Garzarella <sgarzare@xxxxxxxxxx> wrote:
>
> On Fri, May 14, 2021 at 11:55:29AM -0700, Jiang Wang . wrote:
> >On Fri, May 14, 2021 at 8:17 AM Stefano Garzarella <sgarzare@xxxxxxxxxx> wrote:
> >> On Thu, May 13, 2021 at 04:26:03PM -0700, Jiang Wang . wrote:
>
> [...]
>
> >> >I see. I will add some limit to dgram packets. Also, when the
> >> >virtqueues
> >> >are shared between stream and dgram, both of them need to grab a lock
> >> >before using the virtqueue, so one will not completely block another one.
> >>
> >> I'm not worried about the concurrent access that we definitely need to
> >> handle with a lock, but more about the uncontrolled packet sending that
> >> dgram might have, flooding the queues and preventing others from
> >> communicating.
> >
> >That is a valid concern. Let me explain how I would handle that if we
> >don't add two new virtqueues. For dgram, I also add a dgram_send_pkt_list,
> >which is similar to send_pkt_list for stream (and seqpacket). But there
> >is one difference. The dgram_send_pkt_list has a maximum size setting,
> >and keep tracking how many pkts are in the list. The track number
> >(dgram_send_pkt_list_size) is  increased when a packet is added
> >to the list and is decreased when a packet
> >is removed from the list and added to the virtqueue. In
> >virtio_transport_send_pkt, if the current
> >dgram_send_pkt_list_size is equal
> >to the maximum ( let's say 128), then it will not add to the
> >dgram_send_pkt_list and return an error to the application.
>
> For stream socket, we have the send_pkt_list and the send worker because
> the virtqueue can be full and the transmitter needs to wait available
> slots, because we can't discard packets.
>
> For dgram I think we don't need this, so we can avoid the
> dgram_send_pkt_list and directly enqueue packets in the virtqueue.
>
> If there are no slots available, we can simply drop the packet.
> In this way we don't have to put limits other than the available space
> in the virtqueue.

I considered this approach before. One question not clear to me is
whether we should call virtqueue_kick for every dgram packet. If I
am not wrong, each virtqueue_kick will lead to a vm_exit to the host.
If we call virtqueue_kick() for each dgram packet, the performance
might be bad when there are lots of packets. One idea is to set
a threshold and a timer to call virtqueue_kick(). When there are
lots of packets, we wait until a threshold. When there are few packets,
the timer will make sure those packets will be delivered not too late.

Any other ideas for virtqueue_kick? If the threshold plus timer is fine,
I can go this direction too.

> >
> >In this way, the number of pending dgram pkts to be sent is limited.
> >Then both stream and dgram sockets will compete to hold a lock
> >for the tx virtqueue. Depending on the linux scheduler, this competition
> >will be somewhat fair. As a result, dgram will not block stream completely.
> >It will compete and share the virtqueue with stream, but stream
> >can still send some pkts.
> >
> >Basically, the virtqueue becomes a first-come first-serve resource for
> >the stream and dgram. When both stream and dgram applications
> >have lots of data to send, dgram_send_pkt_list and send_pkt_list
> >will still be a limited size, and each will have some chance to send out
> >the data via virtqueue.  Does this address your concern?
> >
> >
> >> So having 2 dedicated queues could avoid a credit mechanism at all for
> >> connection-less sockets, and simply the receiver discards packets that
> >> it can't handle.
> >
> >With 2 dedicated queues, we still need some kind of accounting for
> >the dgram. Like the dgram_send_pkt_size I mentioned above.  Otherwise,
> >it will cause OOM. It is not a real credit mechanism, but code is similar
> >with or without 2 dedicated queues in my current prototype.
>
> I think in both cases we don't need an accounting, but we can drop
> packets if the virtqueue is full.
>
> It's still not clear to me where OOM may occur. Can you elaborate on
> this?

OOM depending on the implementation details. If we keep
dgram_send_pkt_list, and put no limitation,  it may increase indefinitely
and cause OOM. As long as we have some limitations or drop packets
quickly, then we should be fine.

> >
> >For receiver discarding packets part, could you explain more? I think
> >receiver discarding pkts code also is same with or without 2 dedicated
> >queues.
>
> Yep.
>
> > Both will use can_enqueue = virtio_transport_inc_rx_pkt(vvs, pkt);
> >to check if a packet should be discarded or not.
>
> I don't think we can use virtio_transport_inc_rx_pkt() for dgram.
> This is based on the credit mechanism (e.g. buf_alloc) that works when
> you have a connection oriented socket.
>
> With dgram you can have multiple transmitter for a single socket, so how
> we can exchange that information?
>
In my view, we don't need to exchange that information. Buf_alloc for
dgram is local to a receiving socket. The sending sockets do not know
about that. For receiving socket, it just checks if there is still buffer
available to decide to drop a packet or not. If multiple transmitter
send to a single socket, they will share the same buf_alloc, and packets
will be dropped randomly for those transmitters.

> If we reuse the VMCI implementation with skbuff, the sk_receive_skb()
> already checks if the sk_rcvbuf is full and discards the packet in this
> case, so we don't need an internal rx_queue.

OK.

> Maybe someday we should convert stream to skbuff too, it would make our
> lives easier.
>
Yeah.  I remember the original virtio vsock patch did use skb, but somehow
it  did not use skb anymore when merging to the upstream, not sure why.

Thanks,

Jiang

> Thanks,
> Stefano
>
_______________________________________________
Virtualization mailing list
Virtualization@xxxxxxxxxxxxxxxxxxxxxxxxxx
https://lists.linuxfoundation.org/mailman/listinfo/virtualization



[Index of Archives]     [KVM Development]     [Libvirt Development]     [Libvirt Users]     [CentOS Virtualization]     [Netdev]     [Ethernet Bridging]     [Linux Wireless]     [Kernel Newbies]     [Security]     [Linux for Hams]     [Netfilter]     [Bugtraq]     [Yosemite Forum]     [MIPS Linux]     [ARM Linux]     [Linux RAID]     [Linux Admin]     [Samba]

  Powered by Linux