On 02.11.21 12:35, Michael S. Tsirkin wrote: > On Tue, Nov 02, 2021 at 09:33:55AM +0100, David Hildenbrand wrote: >> On 01.11.21 23:15, Michael S. Tsirkin wrote: >>> On Wed, Oct 27, 2021 at 02:45:19PM +0200, David Hildenbrand wrote: >>>> This is the follow-up of [1], dropping auto-detection and vhost-user >>>> changes from the initial RFC. >>>> >>>> Based-on: 20211011175346.15499-1-david@xxxxxxxxxx >>>> >>>> A virtio-mem device is represented by a single large RAM memory region >>>> backed by a single large mmap. >>>> >>>> Right now, we map that complete memory region into guest physical addres >>>> space, resulting in a very large memory mapping, KVM memory slot, ... >>>> although only a small amount of memory might actually be exposed to the VM. >>>> >>>> For example, when starting a VM with a 1 TiB virtio-mem device that only >>>> exposes little device memory (e.g., 1 GiB) towards the VM initialliy, >>>> in order to hotplug more memory later, we waste a lot of memory on metadata >>>> for KVM memory slots (> 2 GiB!) and accompanied bitmaps. Although some >>>> optimizations in KVM are being worked on to reduce this metadata overhead >>>> on x86-64 in some cases, it remains a problem with nested VMs and there are >>>> other reasons why we would want to reduce the total memory slot to a >>>> reasonable minimum. >>>> >>>> We want to: >>>> a) Reduce the metadata overhead, including bitmap sizes inside KVM but also >>>> inside QEMU KVM code where possible. >>>> b) Not always expose all device-memory to the VM, to reduce the attack >>>> surface of malicious VMs without using userfaultfd. >>> >>> I'm confused by the mention of these security considerations, >>> and I expect users will be just as confused. >> >> Malicious VMs wanting to consume more memory than desired is only >> relevant when running untrusted VMs in some environments, and it can be >> caught differently, for example, by carefully monitoring and limiting >> the maximum memory consumption of a VM. We have the same issue already >> when using virtio-balloon to logically unplug memory. For me, it's a >> secondary concern ( optimizing a is much more important ). >> >> Some users showed interest in having QEMU disallow access to unplugged >> memory, because coming up with a maximum memory consumption for a VM is >> hard. This is one step into that direction without having to run with >> uffd enabled all of the time. > > Sorry about missing the memo - is there a lot of overhead associated > with uffd then? When used with huge/gigantic pages, we don't particularly care. For other memory backends, we'll have to route any population via the uffd handler: guest accesses a 4k page -> place a 4k page from user space. Instead of the kernel automatically placing a THP, we'd be placing single 4k pages and have to hope the kernel will collapse them into a THP later. khugepagd will only collapse into a THP if all affected page table entries are present and don't map the zero page, though. So we'll most certainly use less THP for our VM and VM startup time ("first memory access after plugging memory") can be slower. I have prototypes for it, with some optimizations (e.g., on 4k guest access, populate the whole THP area), but we might not want to enable it all of the time. (interaction with postcopy has to be fixed, but it's not a fundamental issue) Extending uffd-based protection for virtio-mem to other processes (vhost-user), is a bit more complicated, and I am not 100% sure if it's worth the trouble for now. memslots provide at least some high-level protection for the important case of having a virtio-mem device to eventually hotplug a lot of memory later. > >> ("security is somewhat the wrong word. we won't be able to steal any >> information from the hypervisor.) > > Right. Let's just spell it out. > Further, removing memory still requires guest cooperation. Right. -- Thanks, David / dhildenb