On Mon, Apr 25, 2022, Andy Lutomirski wrote: > > > On Mon, Apr 25, 2022, at 6:40 AM, Chao Peng wrote: > > On Sun, Apr 24, 2022 at 09:59:37AM -0700, Andy Lutomirski wrote: > >> > > >> > >> 2. Bind the memfile to a VM (or at least to a VM technology). Now it's in > >> the initial state appropriate for that VM. > >> > >> For TDX, this completely bypasses the cases where the data is prepopulated > >> and TDX can't handle it cleanly. I believe TDX can handle this cleanly, TDH.MEM.PAGE.ADD doesn't require that the source and destination have different HPAs. There's just no pressing need to support such behavior because userspace is highly motivated to keep the initial image small for performance reasons, i.e. burning a few extra pages while building the guest is a non-issue. > >> For SEV, it bypasses a situation in which data might be written to the > >> memory before we find out whether that data will be unreclaimable or > >> unmovable. > > > > This sounds a more strict rule to avoid semantics unclear. > > > > So userspace needs to know what excatly happens for a 'bind' operation. > > This is different when binds to different technologies. E.g. for SEV, it > > may imply after this call, the memfile can be accessed (through mmap or > > what ever) from userspace, while for current TDX this should be not allowed. > > I think this is actually a good thing. While SEV, TDX, pKVM, etc achieve > similar goals and have broadly similar ways of achieving them, they really > are different, and having userspace be aware of the differences seems okay to > me. I agree, _if_ the properties of the memory are enumerated in a technology-agnostic way. The underlying mechanisms are different, but conceptually the set of sane operations that userspace can perform/initiate are the same. E.g. TDX and SNP can support swap, they just don't because no one has requested Intel/AMD to provide that support (no use cases for oversubscribing confidential VMs). SNP does support page migration, and TDX can add that support without too much fuss. SEV "allows" the host to access guest private memory, but that doesn't mean it should be deliberately supported by the kernel. It's a bit of a moot point for SEV/SEV-ES, as the host doesn't get any kind of notification that the guest has "converted" a page, but the kernel shouldn't allow userspace to map memory that is _known_ to be private. > (Although I don't think that allowing userspace to mmap SEV shared pages is s/shared/private? > particularly wise -- it will result in faults or cache incoherence depending > on the variant of SEV in use.) > > > And I feel we still need a third flow/operation to indicate the > > completion of the initialization on the memfile before the guest's > > first-time launch. SEV needs to check previous mmap-ed areas are munmap-ed > > and prevent future userspace access. After this point, then the memfile > > becomes truely private fd. > > Even that is technology-dependent. For TDX, this operation doesn't really > exist. As above, I believe this is TDH.MEM.PAGE.ADD. > For SEV, I'm not sure (I haven't read the specs in nearly enough detail). QEMU+KVM does in-place conversion for SEV/SEV-ES via SNP_LAUNCH_UPDATE, AFAICT that's still allowed for SNP. > For pKVM, I guess it does exist and isn't quite the same as a > shared->private conversion. > > Maybe this could be generalized a bit as an operation "measure and make > private" that would be supported by the technologies for which it's useful. > > > >> Now I have a question, since I don't think anyone has really answered it: > >> how does this all work with SEV- or pKVM-like technologies in which > >> private and shared pages share the same address space? The current proposal is to have both a private fd and a shared hva for memslot that can be mapped private. A GPA is considered private by KVM if the memslot has a private fd and that corresponding page in the private fd is populated. KVM will always and only map the current flavor of shared/private based on that definition. If userspace punches a hole in the private fd, KVM will unmap any relevant private GPAs. If userspace populates a range in the private fd, KVM will unmap any relevant shared GPAs. > >> I sounds like you're proposing to have a big memfile that contains private > >> and shared pages and to use that same memfile as pages are converted back > >> and forth. IO and even real physical DMA could be done on that memfile. > >> Am I understanding correctly? > > > > For TDX case, and probably SEV as well, this memfile contains private memory > > only. But this design at least makes it possible for usage cases like > > pKVM which wants both private/shared memory in the same memfile and rely > > on other ways like mmap/munmap or mprotect to toggle private/shared instead > > of fallocate/hole punching. > > Hmm. Then we still need some way to get KVM to generate the correct SEV > pagetables. For TDX, there are private memslots and shared memslots, and > they can overlap. If they overlap and both contain valid pages at the same > address, then the results may not be what the guest-side ABI expects, but > everything will work. Absolutely not, KVM is not concurrently mapping both private and shared variants of a single GPA into the guest. The Shared bit is a glorified attribute/permission bit, that it's carved out of the GPA space is a hack to minimize Intel's hardware development cost. > So, when a single logical guest page transitions between shared and private, > no change to the memslots is needed. Hard no, KVM is not supporting different memslot semantics for TDX versus everything else. > For SEV, this is not the case: everything is in one set of pagetables, and > there isn't a natural way to resolve overlaps. > > If the memslot code becomes efficient enough, then the memslots could be > fragmented. No, because the only way to not artificially limit the amount of fragmentation is to turn the memslots into a tree structure, i.e. to make them effectively multi-level page tables as opposed to the single-level "tables" that they are today. > Or the memfile could support private and shared data in the same memslot. > And if pKVM does this, I don't see why SEV couldn't also do it and hopefully > reuse the same code. > > > > >> > >> If so, I think this makes sense, but I'm wondering if the actual memslot > >> setup should be different. For TDX, private memory lives in a logically > >> separate memslot space. For SEV and pKVM, it doesn't. I assume the API > >> can reflect this straightforwardly. Again, no. KVM is not going to give special treatment to TDX. > > I believe so. The flow should be similar but we do need pass different > > flags during the 'bind' to the backing store for different usages. That > > should be some new flags for pKVM but the callbacks (API here) between > > memfile_notifile and its consumers can be reused. > > And also some different flag in the operation that installs the fd as a memslot? No, memslots updates need to come directly from userspace. > >> And the corresponding TDX question: is the intent still that shared pages > >> aren't allowed at all in a TDX memfile? If so, that would be the most > >> direct mapping to what the hardware actually does. > > > > Exactly. TDX will still use fallocate/hole punching to turn on/off the > > private page. Once off, the traditional shared page will become > > effective in KVM. > > Works for me. > > For what it's worth, I still think it should be fine to land all the TDX > memfile bits upstream as long as we're confident that SEV, pKVM, etc can be > added on without issues. > > I think we can increase confidence in this by either getting one other > technology's maintainers to get far enough along in the design to be > confident and/or by having a pure-kernel-software implementation that serves > as a testbed. For the latter, maybe it could support two different models > with little overhead: > > Pure software "interleaved" model: pages are shared or private and a > hypercall converts them. The access mode is entirely determined by the state > programmed by hypercall. I think this is essentially what Vishal > implemented, but with the "HACK" replaced by something permanent and (if > they're not already in the series) appropriate access checks implemented to > actually protect the private memory. > > Pure software "separate" mode: one GPA bit is set aside as the shared vs > private bit. The normal memslots are restricted to the shared half of GPA > space. Private memslots use the private half. This works a lot like TDX. > This would be new code. We don't *really* need this for testing, since TDX > itself exercises the same programming model, but it would let people without > TDX hardware exercise the interesting bits of the memory management. No, KVM is not bifurcating the memslots into shared and private. Except for TDX, hardware can't support mapping both variants. That means KVM has to define some semantic for which memslot "wins", and that puts userspace/KVM back to square one in having punch a hole into a memslot in order to allow mapping the "loser" into the guest.