On Mon, 12 Apr 2021 13:43:41 +0100 Matthew Wilcox <willy@xxxxxxxxxxxxx> wrote: [...] > > For the rename case, how would you handle gup.c? > > At first, I'd turn it into > arch_make_folio_accessible(page_folio(page)); that's tricky. what happens if some pages in the folio cannot be made accessible? also, I assume you keep the semantic difference between get_page and pin_page? that's also very important for us > Eventually, gup.c needs to become folio-aware. I haven't spent too > much time thinking about it, but code written like this: > > page = pte_page(pte); > head = try_grab_compound_head(page, 1, flags); > if (!head) > goto pte_unmap; > if (unlikely(pte_val(pte) != pte_val(*ptep))) { > put_compound_head(head, 1, flags); > goto pte_unmap; > } > VM_BUG_ON_PAGE(compound_head(page) != head, page); > > is just crying out for use of folios. Also, some of the gup callers > would much prefer to work in terms of folios than individual struct > pages (imagine an RDMA adapter that wants to pin several gigabytes of > memory that's allocated using hugetlbfs for example). > > > Consider that arch_make_page_accessible deals (typically) with KVM > > guest pages. Once you bundle up the pages in folios, you can have > > different pages in the same folio with different properties. > > So what you're saying is that the host might allocate, eg a 1GB folio > for a guest, then the guest splits that up into smaller chunks (eg > 1MB), and would only want one of those small chunks accessible to the > hypervisor? qemu will allocate a big chunk of memory, and I/O would happen only on small chunks (depending on what the guest does). I don't know how swap and pagecache would behave in the folio scenario. Also consider that currently we need 4k hardware pages for protected guests (so folios would be ok, as long as they are backed by small pages) How and when are folios created actually? is there a way to prevent creation of multi-page folios? > > In case of failure, you could end up with a folio with some pages > > processed and some not processed. Would you stop at the first error? > > What would the state of the folio be? On s390x we use the PG_arch_1 > > bit to mark secure pages, how would that work with folios? > > > > and how are fault handlers affected by this folio conversion? would > > they still work on pages, or would that also work on folios? on > > s390x we use the arch_make_page_accessible function in some fault > > handlers. > > Folios can be mapped into userspace at an unaligned offset. So we > still have to work in pages, at least for now. We might have some > optimised path for aligned folios later. > > > a possible approach maybe would be to keep the _page variant, and > > add a _folio wrapper around it > > Yes, we can do that. It's what I'm currently doing for > flush_dcache_folio(). where would the page flags be stored? as I said, we really depend on that bit to be set correctly to prevent potentially disruptive I/O errors. It's ok if the bit overindicates protection (non-protected pages can be marked as protected), but protected pages must at all times have the bit set. the reason why this hook exists at all, is to prevent secure pages from being accidentally (or maliciously) fed into I/O > > for s390x the PG_arch_1 is very important to prevent protected pages > > from being fed to I/O, as in that case Very Bad Things™ would > > happen. > > > > sorry for the wall of questions, but I actually like your folio > > approach and I want to understand it better, so we can find a way to > > make everything work well together > > Great! > > > > PS: The prototype is in gfp.h. That's not really appropriate; > > > gfp.h is about allocating memory, and this call really has > > > nothing to do with memory allocation. I think mm.h is a better > > > place for it, if you can't find a better header file than that. > > > > I had put it there because arch_alloc_page and arch_free_page are > > also there, and the behaviour, from a kernel point of view, is > > similar (unaccessible/unallocated pages will trigger a fault). > > > > I actually do not have a preference regarding where the prototype > > lives, as long as everything works. If you think mm.h is more > > appropriate, go for it :) > > Heh, I see how you got there from the implementors point of view ;-) > I'll move it ...