On 17.05.22 10:30, Qi Zheng wrote: > > > On 2022/4/29 9:35 PM, Qi Zheng wrote: >> Hi, >> >> This patch series aims to try to free user PTE page table pages when no one is >> using it. >> >> The beginning of this story is that some malloc libraries(e.g. jemalloc or >> tcmalloc) usually allocate the amount of VAs by mmap() and do not unmap those >> VAs. They will use madvise(MADV_DONTNEED) to free physical memory if they want. >> But the page tables do not be freed by madvise(), so it can produce many >> page tables when the process touches an enormous virtual address space. >> >> The following figures are a memory usage snapshot of one process which actually >> happened on our server: >> >> VIRT: 55t >> RES: 590g >> VmPTE: 110g >> >> As we can see, the PTE page tables size is 110g, while the RES is 590g. In >> theory, the process only need 1.2g PTE page tables to map those physical >> memory. The reason why PTE page tables occupy a lot of memory is that >> madvise(MADV_DONTNEED) only empty the PTE and free physical memory but >> doesn't free the PTE page table pages. So we can free those empty PTE page >> tables to save memory. In the above cases, we can save memory about 108g(best >> case). And the larger the difference between the size of VIRT and RES, the >> more memory we save. >> >> In this patch series, we add a pte_ref field to the struct page of page table >> to track how many users of user PTE page table. Similar to the mechanism of page >> refcount, the user of PTE page table should hold a refcount to it before >> accessing. The user PTE page table page may be freed when the last refcount is >> dropped. >> >> Different from the idea of another patchset of mine before[1], the pte_ref >> becomes a struct percpu_ref type, and we switch it to atomic mode only in cases >> such as MADV_DONTNEED and MADV_FREE that may clear the user PTE page table >> entryies, and then release the user PTE page table page when checking that >> pte_ref is 0. The advantage of this is that there is basically no performance >> overhead in percpu mode, but it can also free the empty PTEs. In addition, the >> code implementation of this patchset is much simpler and more portable than the >> another patchset[1]. > > Hi David, > > I learned from the LWN article[1] that you led a session at the LSFMM on > the problems posed by the lack of page-table reclaim (And thank you very > much for mentioning some of my work in this direction). So I want to > know, what are the further plans of the community for this problem? Hi, yes, I talked about the involved challenges, especially, how malicious user space can trigger allocation of almost elusively page tables and essentially consume a lot of unmovable+unswappable memory and even store secrets in the page table structure. Empty PTE tables is one such case we care about, but there is more. Even with your approach, we can still end up with many page tables that are allocated on higher levels (e.g., PMD tables) or page tables that are not empty (especially, filled with the shared zeropage). Ideally, we'd have some mechanism that can reclaim also other reclaimable page tables (e.g., filled with shared zeropage). One idea was to add reclaimable page tables to the LRU list and to then scan+reclaim them on demand. There are multiple challenges involved, obviously. One is how to synchronize against concurrent page table walkers, another one is how to invalidate MMU notifiers from reclaim context. It would most probably involve storing required information in the memmap to be able to lock+synchronize. Having that said, adding infrastructure that might not be easy to extend to the more general case of reclaiming other reclaimable page tables on multiple levels (esp PMD tables) might not be what we want. OTOH, it gets the job done for once case we care about. It's really hard to tell what to do because reclaiming page tables and eventually handling malicious user space correctly is far from trivial :) I'll be on vacation until end of May, I'll come back to this mail once I'm back. -- Thanks, David / dhildenb