On Tue, 20 Jun 2023 00:51:19 -0700 (PDT) Hugh Dickins <hughd@xxxxxxxxxx> wrote: > Add s390-specific pte_free_defer(), to call pte_free() via call_rcu(). > pte_free_defer() will be called inside khugepaged's retract_page_tables() > loop, where allocating extra memory cannot be relied upon. This precedes > the generic version to avoid build breakage from incompatible pgtable_t. > > This version is more complicated than others: because s390 fits two 2K > page tables into one 4K page (so page->rcu_head must be shared between > both halves), and already uses page->lru (which page->rcu_head overlays) > to list any free halves; with clever management by page->_refcount bits. > > Build upon the existing management, adjusted to follow a new rule: that > a page is not linked to mm_context_t::pgtable_list while either half is > pending free, by either tlb_remove_table() or pte_free_defer(); but is > afterwards either relinked to the list (if other half is allocated), or > freed (if other half is free): by __tlb_remove_table() in both cases. > > This rule ensures that page->lru is no longer in use while page->rcu_head > may be needed for use by pte_free_defer(). And a fortuitous byproduct of > following this rule is that page_table_free() no longer needs its curious > two-step manipulation of _refcount - read commit c2c224932fd0 ("s390/mm: > fix 2KB pgtable release race") for what to think of there. But it does > not solve the problem that two halves may need rcu_head at the same time. > > For that, add HHead bits between s390's AAllocated and PPending bits in > the upper byte of page->_refcount: then the second pte_free_defer() can > see that rcu_head is already in use, and the RCU callee pte_free_half() > can see that it needs to make a further call_rcu() for that other half. > > page_table_alloc() set the page->pt_mm field, so __tlb_remove_table() > knows where to link the freed half while its other half is allocated. > But linking to the list needs mm->context.lock: and although AA bit set > guarantees that pt_mm must still be valid, it does not guarantee that mm > is still valid an instant later: so acquiring mm->context.lock would not > be safe. For now, use a static global mm_pgtable_list_lock instead: > then a soon-to-follow commit will split it per-mm as before (probably by > using a SLAB_TYPESAFE_BY_RCU structure for the list head and its lock); > and update the commentary on the pgtable_list. > > Signed-off-by: Hugh Dickins <hughd@xxxxxxxxxx> > --- > arch/s390/include/asm/pgalloc.h | 4 + > arch/s390/mm/pgalloc.c | 205 +++++++++++++++++++++++--------- > include/linux/mm_types.h | 2 +- > 3 files changed, 154 insertions(+), 57 deletions(-) As discussed in the other thread, we would rather go with less complexity, possibly switching to an approach w/o the list and fragment re-use in the future. For now, as a first step in that direction, we can try with not adding fragments back only for pte_free_defer(). Here is an adjusted version of your patch, copying most of your pte_free_defer() logic and also description, tested with LTP and all three of your patch series applied: Add s390-specific pte_free_defer(), to call pte_free() via call_rcu(). pte_free_defer() will be called inside khugepaged's retract_page_tables() loop, where allocating extra memory cannot be relied upon. This precedes the generic version to avoid build breakage from incompatible pgtable_t. This version is more complicated than others: because s390 fits two 2K page tables into one 4K page (so page->rcu_head must be shared between both halves), and already uses page->lru (which page->rcu_head overlays) to list any free halves; with clever management by page->_refcount bits. Build upon the existing management, adjusted to follow a new rule: that a page is never added back to the list in pte_free_defer(). It is only removed from the list, when currently listed because the other fragment is not allocated. This introduces some asymmetry compared to the other page table freeing paths, and in particular a list_del() for such pages must be avoided there. Use page->pt_frag_refcount to keep track of the list status, and check that before doing list_del() in any freeing path. Other paths would also not add back such pages to the list, if the other fragment happens to be freed in such a path at the same time, because they would observe cleared AA bits. This rule ensures that page->lru is no longer in use while page->rcu_head may be needed for use by pte_free_defer(). But it does not solve the problem that two halves may need rcu_head at the same time. For that, add HHead bits between s390's AAllocated and PPending bits in the upper byte of page->_refcount: then the second pte_free_defer() can see that rcu_head is already in use, and the RCU callee pte_free_half() can see that it needs to make a further call_rcu() for that other half. Not adding back unallocated fragments to the list in pte_free_defer() can result in wasting some amount of memory for pagetables, depending on how long the allocated fragment will stay in use. In practice, this effect is expected to be insignificant, and not justify a far more complex approach, which might allow to add the fragments back later in __tlb_remove_table(), where we might not have a stable mm any more. Signed-off-by: Gerald Schaefer <gerald.schaefer@xxxxxxxxxxxxx> --- arch/s390/include/asm/pgalloc.h | 4 + arch/s390/mm/pgalloc.c | 136 +++++++++++++++++++++++++++++++++++++--- 2 files changed, 132 insertions(+), 8 deletions(-) --- a/arch/s390/include/asm/pgalloc.h +++ b/arch/s390/include/asm/pgalloc.h @@ -143,6 +143,10 @@ static inline void pmd_populate(struct m #define pte_free_kernel(mm, pte) page_table_free(mm, (unsigned long *) pte) #define pte_free(mm, pte) page_table_free(mm, (unsigned long *) pte) +/* arch use pte_free_defer() implementation in arch/s390/mm/pgalloc.c */ +#define pte_free_defer pte_free_defer +void pte_free_defer(struct mm_struct *mm, pgtable_t pgtable); + void vmem_map_init(void); void *vmem_crst_alloc(unsigned long val); pte_t *vmem_pte_alloc(void); --- a/arch/s390/mm/pgalloc.c +++ b/arch/s390/mm/pgalloc.c @@ -185,11 +185,13 @@ void page_table_free_pgste(struct page * * The upper byte (bits 24-31) of the parent page _refcount is used * for tracking contained 2KB-pgtables and has the following format: * - * PP AA + * PPHHAA * 01234567 upper byte (bits 24-31) of struct page::_refcount - * || || - * || |+--- upper 2KB-pgtable is allocated - * || +---- lower 2KB-pgtable is allocated + * |||||| + * |||||+--- upper 2KB-pgtable is allocated + * ||||+---- lower 2KB-pgtable is allocated + * |||+----- upper 2KB-pgtable is pending free by page->rcu_head + * ||+------ lower 2KB-pgtable is pending free by page->rcu_head * |+------- upper 2KB-pgtable is pending for removal * +-------- lower 2KB-pgtable is pending for removal * @@ -229,6 +231,17 @@ void page_table_free_pgste(struct page * * logic described above. Both AA bits are set to 1 to denote a 4KB-pgtable * while the PP bits are never used, nor such a page is added to or removed * from mm_context_t::pgtable_list. + * + * The HH bits are used to prevent double use of page->rcu_head in + * pte_free_defer(), when both 2K pagetables inside a page happen to get + * freed by that path at the same time. + * + * pte_free_defer() also cannot add 2K fragments back to the list, because + * page->rcu_head overlays with page->lru. This introduces some asymmetry + * compared to the other pagetable freeing paths, and the missing list_add() + * in pte_free_defer() could result in incorrect list_del(). Therefore, track + * the the list status of a page with page->pt_frag_refcount, and check that + * before doing list_del() in any freeing path. */ unsigned long *page_table_alloc(struct mm_struct *mm) { @@ -262,6 +275,7 @@ unsigned long *page_table_alloc(struct m atomic_xor_bits(&page->_refcount, 0x01U << (bit + 24)); list_del(&page->lru); + atomic_set(&page->pt_frag_refcount, 0); } } spin_unlock_bh(&mm->context.lock); @@ -290,6 +304,7 @@ unsigned long *page_table_alloc(struct m memset64((u64 *)table, _PAGE_INVALID, 2 * PTRS_PER_PTE); spin_lock_bh(&mm->context.lock); list_add(&page->lru, &mm->context.pgtable_list); + atomic_set(&page->pt_frag_refcount, 1); spin_unlock_bh(&mm->context.lock); } return table; @@ -325,13 +340,24 @@ void page_table_free(struct mm_struct *m */ mask = atomic_xor_bits(&page->_refcount, 0x11U << (bit + 24)); mask >>= 24; - if (mask & 0x03U) + if (mask & 0x03U) { + /* + * Other half is allocated, add to list + */ list_add(&page->lru, &mm->context.pgtable_list); - else + atomic_set(&page->pt_frag_refcount, 1); + } else if (atomic_read(&page->pt_frag_refcount)) { + /* + * Other half is not allocated, and page is on the list, + * remove from list + */ list_del(&page->lru); + atomic_set(&page->pt_frag_refcount, 0); + } spin_unlock_bh(&mm->context.lock); mask = atomic_xor_bits(&page->_refcount, 0x10U << (bit + 24)); mask >>= 24; + /* Return if other half is allocated, or delayed release pending */ if (mask != 0x00U) return; half = 0x01U << bit; @@ -370,10 +396,22 @@ void page_table_free_rcu(struct mmu_gath */ mask = atomic_xor_bits(&page->_refcount, 0x11U << (bit + 24)); mask >>= 24; - if (mask & 0x03U) + if (mask & 0x03U) { + /* + * Other half is allocated, add to end of list, as this + * will not immediately be re-usable because it is marked + * for delayed release + */ list_add_tail(&page->lru, &mm->context.pgtable_list); - else + atomic_set(&page->pt_frag_refcount, 1); + } else if (atomic_read(&page->pt_frag_refcount)) { + /* + * Other half is not allocated, and page is on the list, + * remove from list + */ list_del(&page->lru); + atomic_set(&page->pt_frag_refcount, 0); + } spin_unlock_bh(&mm->context.lock); table = (unsigned long *) ((unsigned long) table | (0x01U << bit)); tlb_remove_table(tlb, table); @@ -407,6 +445,88 @@ void __tlb_remove_table(void *_table) __free_page(page); } +#ifdef CONFIG_TRANSPARENT_HUGEPAGE +static void pte_free_now0(struct rcu_head *head); +static void pte_free_now1(struct rcu_head *head); + +static void pte_free_pgste(struct rcu_head *head) +{ + unsigned long *table; + struct page *page; + + page = container_of(head, struct page, rcu_head); + table = (unsigned long *)page_to_virt(page); + table = (unsigned long *)((unsigned long)table | 0x03U); + __tlb_remove_table(table); +} + +static void pte_free_half(struct rcu_head *head, unsigned int bit) +{ + unsigned long *table; + struct page *page; + unsigned int mask; + + page = container_of(head, struct page, rcu_head); + mask = atomic_xor_bits(&page->_refcount, 0x04U << (bit + 24)); + + table = (unsigned long *)page_to_virt(page); + table += bit * PTRS_PER_PTE; + table = (unsigned long *)((unsigned long)table | (0x01U << bit)); + __tlb_remove_table(table); + + /* If pte_free_defer() of the other half came in, queue it now */ + if (mask & 0x0CU) + call_rcu(&page->rcu_head, bit ? pte_free_now0 : pte_free_now1); +} + +static void pte_free_now0(struct rcu_head *head) +{ + pte_free_half(head, 0); +} + +static void pte_free_now1(struct rcu_head *head) +{ + pte_free_half(head, 1); +} + +void pte_free_defer(struct mm_struct *mm, pgtable_t pgtable) +{ + unsigned int bit, mask; + struct page *page; + + page = virt_to_page(pgtable); + if (mm_alloc_pgste(mm)) { + /* + * TODO: Do we need gmap_unlink(mm, pgtable, addr), like in + * page_table_free_rcu()? + * If yes -> need addr parameter here, like in pte_free_tlb(). + */ + call_rcu(&page->rcu_head, pte_free_pgste); + return; +} + bit = ((unsigned long)pgtable & ~PAGE_MASK) / (PTRS_PER_PTE * sizeof(pte_t)); + + spin_lock_bh(&mm->context.lock); + mask = atomic_xor_bits(&page->_refcount, 0x15U << (bit + 24)); + mask >>= 24; + if ((mask & 0x03U) == 0x00U && atomic_read(&page->pt_frag_refcount)) { + /* + * Other half is not allocated, page is on the list, + * remove from list + */ + list_del(&page->lru); + atomic_set(&page->pt_frag_refcount, 0); + } + /* Page must not be on the list, so rcu_head can be used */ + BUG_ON(atomic_read(&page->pt_frag_refcount)); + spin_unlock_bh(&mm->context.lock); + + /* Do not relink on rcu_head if other half already linked on rcu_head */ + if ((mask & 0x0CU) != 0x0CU) + call_rcu(&page->rcu_head, bit ? pte_free_now1 : pte_free_now0); +} +#endif /* CONFIG_TRANSPARENT_HUGEPAGE */ + /* * Base infrastructure required to generate basic asces, region, segment, * and page tables that do not make use of enhanced features like EDAT1.