On Sun, Apr 14, 2013 at 11:16:46AM -0400, Sasha Levin wrote: > On 04/14/2013 06:01 AM, Michael S. Tsirkin wrote: > > On Sat, Apr 13, 2013 at 05:23:41PM -0400, Sasha Levin wrote: > >> On 04/12/2013 07:36 AM, Rusty Russell wrote: > >>> Sasha Levin <sasha.levin@xxxxxxxxxx> writes: > >>>> On 04/11/2013 12:36 PM, Will Deacon wrote: > >>>>> Hello folks, > >>>>> > >>>>> Here's the latest round of ARM fixes and updates for kvmtool. Most of > >>>>> this is confined to the arm/ subdirectory, with the exception of a fix > >>>>> to the virtio-mmio vq definitions due to the multi-queue work from > >>>>> Sasha. I'm not terribly happy about that code though, since it seriously > >>>>> increases the memory footprint of the guest. > >>>>> > >>>>> Without multi-queue, we can boot Debian Wheezy to a prompt in 38MB. With > >>>>> the new changes, that increases to 170MB! Any chance we can try and tackle > >>>>> this regression please? I keep getting bitten by the OOM killer :( > >>>> > >>>> (cc Rusty, MST) > >>>> > >>>> The spec defines the operation of a virtio-net device with regards to multiple > >>>> queues as follows: > >>>> > >>>> """ > >>>> Device Initialization > >>>> > >>>> 1. The initialization routine should identify the receive and transmission > >>>> virtqueues, up to N+1 of each kind. If VIRTIO_NET_F_MQ feature > >>>> bit is negotiated, N=max_virtqueue_pairs-1, otherwise identify N=0. > >>>> > >>>> [...] > >>>> > >>>> 5. Only receiveq0, transmitq0 and controlq are used by default. To use more > >>>> queues driver must negotiate the VIRTIO_NET_F_MQ feature; initialize > >>>> up to max_virtqueue_pairs of each of transmit and receive queues; execute_ > >>>> VIRTIO_NET_CTRL_MQ_VQ_PAIRS_SET command specifying > >>>> the number of the transmit and receive queues that is going to be > >>>> used and wait until the device consumes the controlq buffer and acks this > >>>> command. > >>>> """ > >>>> > >>>> And kvmtool follows that to the letter: It will initialize the maximum amount of > >>>> queues it can support during initialization and will start using them only when > >>>> the device tells it it should use them. > >>>> > >>>> As Will has stated, this causes a memory issue since all the data structures that hold > >>>> all possible queues get initialized regardless of whether we actually need them or not, > >>>> which is quite troublesome for systems with small RAM. > >>>> > >>>> > >>>> Rusty, MST, would you be open to a spec and code change that would initialize the > >>>> RX/TX vqs on demand instead of on device initialization? Or is there an easier way > >>>> to work around this issue? > >>> > >>> I'm confused. kvmtool is using too much memory, or the guest? If > >>> kvmtool, the Device Initialization section above applies to the driver, > >>> not the device. If the guest, well, the language says "UP TO N+1". You > >>> want a small guest, don't use them all. Or any... > >>> > >>> What am I missing? > >> > >> It's in the guest - sorry. I was only trying to say that kvmtool doesn't do anything > >> odd with regards to initializing virtio-net. > >> > >> The thing is that there should be a difference between just allowing a larger number > >> of queues and actually using them (i.e. enabling them with ethtool). Right now I see > >> the kernel lose 130MB just by having kvmtool offer 8 queue pairs, without actually > >> using those queues. > >> > >> Yes, we can make it configurable in kvmtool (and I will make it so so the arm folks > >> could continue working with tiny guests) but does it make sense that you have to do > >> this configuration in *2* places? First in the hypervisor and then inside the guest? > >> > >> Memory usage should ideally depend on whether you are actually using multiple queues, > >> not on whether you just allow using those queues. > >> > >> > >> Thanks, > >> Sasha > > > > 8 queues eat up 130MB? Most of the memory is likely for the buffers? I > > think we could easily allocate these lazily as queues are enabled, > > without protocol changes. It's harder to clean them as there's no way to > > reset a specific queue, but maybe that' good enough for your purposes? > > > > Yup, this is how it looks in the guest right after booting: > > Without virtio-net mq: > > # free > total used free shared buffers cached > Mem: 918112 158052 760060 0 0 4308 > -/+ buffers/cache: 153744 764368 > > With queue pairs = 8: > > # free > total used free shared buffers cached > Mem: 918112 289168 628944 0 0 4244 > -/+ buffers/cache: 284924 633188 > > > Initializing them only when they're actually needed will do the trick here. Not initializing, adding the buffers. In the current spec, initialization is always done before DRIVER_OK. > We could also expose the facility to delete a single vq, and add a note to > the spec saying that if the amount of actual vq pairs was reduced below what > it was before, the now inactive queues become invalid and would need to be > re-initialized. It's not pretty but it would let both device and driver free > up those vqs. > > > Thanks, > Sasha If we want to do this, I would add a command to have the device complete all RX buffers of a given VQ with len=0. I wouldn't do this automatically when number of VQs changes since we had an idea to switch between MQ and SQ mode dynamically, which needs all buffers to stay in the VQ. -- MST -- To unsubscribe from this list: send the line "unsubscribe kvm" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html