On Tue, Nov 02, 2021 at 12:55:17PM +0100, David Hildenbrand wrote: > On 02.11.21 12:35, Michael S. Tsirkin wrote: > > On Tue, Nov 02, 2021 at 09:33:55AM +0100, David Hildenbrand wrote: > >> On 01.11.21 23:15, Michael S. Tsirkin wrote: > >>> On Wed, Oct 27, 2021 at 02:45:19PM +0200, David Hildenbrand wrote: > >>>> This is the follow-up of [1], dropping auto-detection and vhost-user > >>>> changes from the initial RFC. > >>>> > >>>> Based-on: 20211011175346.15499-1-david@xxxxxxxxxx > >>>> > >>>> A virtio-mem device is represented by a single large RAM memory region > >>>> backed by a single large mmap. > >>>> > >>>> Right now, we map that complete memory region into guest physical addres > >>>> space, resulting in a very large memory mapping, KVM memory slot, ... > >>>> although only a small amount of memory might actually be exposed to the VM. > >>>> > >>>> For example, when starting a VM with a 1 TiB virtio-mem device that only > >>>> exposes little device memory (e.g., 1 GiB) towards the VM initialliy, > >>>> in order to hotplug more memory later, we waste a lot of memory on metadata > >>>> for KVM memory slots (> 2 GiB!) and accompanied bitmaps. Although some > >>>> optimizations in KVM are being worked on to reduce this metadata overhead > >>>> on x86-64 in some cases, it remains a problem with nested VMs and there are > >>>> other reasons why we would want to reduce the total memory slot to a > >>>> reasonable minimum. > >>>> > >>>> We want to: > >>>> a) Reduce the metadata overhead, including bitmap sizes inside KVM but also > >>>> inside QEMU KVM code where possible. > >>>> b) Not always expose all device-memory to the VM, to reduce the attack > >>>> surface of malicious VMs without using userfaultfd. > >>> > >>> I'm confused by the mention of these security considerations, > >>> and I expect users will be just as confused. > >> > >> Malicious VMs wanting to consume more memory than desired is only > >> relevant when running untrusted VMs in some environments, and it can be > >> caught differently, for example, by carefully monitoring and limiting > >> the maximum memory consumption of a VM. We have the same issue already > >> when using virtio-balloon to logically unplug memory. For me, it's a > >> secondary concern ( optimizing a is much more important ). > >> > >> Some users showed interest in having QEMU disallow access to unplugged > >> memory, because coming up with a maximum memory consumption for a VM is > >> hard. This is one step into that direction without having to run with > >> uffd enabled all of the time. > > > > Sorry about missing the memo - is there a lot of overhead associated > > with uffd then? > > When used with huge/gigantic pages, we don't particularly care. > > For other memory backends, we'll have to route any population via the > uffd handler: guest accesses a 4k page -> place a 4k page from user > space. Instead of the kernel automatically placing a THP, we'd be > placing single 4k pages and have to hope the kernel will collapse them > into a THP later. How much value there is in a THP given it's not present? > khugepagd will only collapse into a THP if all affected page table > entries are present and don't map the zero page, though. > > So we'll most certainly use less THP for our VM and VM startup time > ("first memory access after plugging memory") can be slower. > > I have prototypes for it, with some optimizations (e.g., on 4k guest > access, populate the whole THP area), but we might not want to enable it > all of the time. (interaction with postcopy has to be fixed, but it's > not a fundamental issue) > > > Extending uffd-based protection for virtio-mem to other processes > (vhost-user), is a bit more complicated, and I am not 100% sure if it's > worth the trouble for now. memslots provide at least some high-level > protection for the important case of having a virtio-mem device to > eventually hotplug a lot of memory later. > > > > >> ("security is somewhat the wrong word. we won't be able to steal any > >> information from the hypervisor.) > > > > Right. Let's just spell it out. > > Further, removing memory still requires guest cooperation. > > Right. > > > -- > Thanks, > > David / dhildenb