Patches 1-2: Introduces new arch-generic functions to use before and after lockless pagetable walks, instead of local_irq_*, and applies them to generic code. It makes lockless pagetable walks more explicit and improves documentation about it. Patches 3-9: Introduces a powerpc-specific version of the above functions with the option to not touch irq config. Then apply them to all powerpc code that do lockless pagetable walks. Patches 10-11: Introduces a percpu counting method to keep track of the lockless page table walks, then uses this info to reduce the waiting time on serialize_against_pte_lookup(). Use case: If a process (qemu) with a lot of CPUs (128) try to munmap() a large chunk of memory (496GB) mapped with THP, it takes an average of 275 seconds, which can cause a lot of problems to the load (in qemu case, the guest will lock for this time). Trying to find the source of this bug, I found out most of this time is spent on serialize_against_pte_lookup(). This function will take a lot of time in smp_call_function_many() if there is more than a couple CPUs running the user process. Since it has to happen to all THP mapped, it will take a very long time for large amounts of memory. By the docs, serialize_against_pte_lookup() is needed in order to avoid pmd_t to pte_t casting inside find_current_mm_pte(), or any lockless pagetable walk, to happen concurrently with THP splitting/collapsing. It does so by calling a do_nothing() on each CPU in mm->cpu_bitmap[], after interrupts are re-enabled. Since, interrupts are (usually) disabled during lockless pagetable walk, and serialize_against_pte_lookup will only return after interrupts are enabled, it is protected. Percpu count-based method: So, by what I could understand, if there is no lockless pagetable walk running on given cpu, there is no need to call serialize_against_pte_lookup() there. To reduce the cost of running serialize_against_pte_lookup(), I propose a percpu-counter that keeps track of how many lockless pagetable walks are currently running on each cpu, and if there is none, just skip smp_call_function_many() for that cpu. - Every percpu-counter can be changed only by it's own CPU - It makes use of the original memory barrier in the functions - Any counter can be read by any CPU Due to not locking nor using atomic variables, the impact on the lockless pagetable walk is intended to be minimum. The related functions are: begin_lockless_pgtbl_walk() Insert before starting any lockless pgtable walk end_lockless_pgtbl_walk() Insert after the end of any lockless pgtable walk (Mostly after the ptep is last used) Results: On my workload (qemu), I could see munmap's time reduction from 275 seconds to 430ms. Bonus: I documented some lockless pagetable walks in which it's not necessary to keep track, given they work on init_mm or guest pgd. Also fixed some misplaced local_irq_{restore, enable}. Changes since v5: Changed counting approach from atomic variables to percpu variables Counting method only affects powepc, arch-generic only toggle irqs Changed commit order, so the counting method is introduced at the end Removed config option, always enabled in powerpc Rebased on top of v5.5 Link: http://patchwork.ozlabs.org/project/linuxppc-dev/list/?series=133907 Changes since v4: Rebased on top of v5.4-rc1 Declared real generic functions instead of dummies start_lockless_pgtbl_walk renamed to begin_lockless_pgtbl_walk Interrupt {dis,en}able is now inside of {begin,end}_lockless_pgtbl_walk Power implementation has option to not {dis,en}able interrupt More documentation inside the funtions. Some irq masks variables renamed Removed some proxy mm_structs Few typos fixed Link: http://patchwork.ozlabs.org/project/linuxppc-dev/list/?series=133015 Changes since v3: Explain (comments) why some lockless pgtbl walks don't need local_irq_disable (real mode + MSR_EE=0) Explain (comments) places where counting method is not needed (guest pgd, which is not touched by THP) Fixes some misplaced local_irq_restore() Link: http://patchwork.ozlabs.org/project/linuxppc-dev/list/?series=132417 Changes since v2: Rebased to v5.3 Adds support on __get_user_pages_fast Adds usage decription to *_lockless_pgtbl_walk() Better style to dummy functions Link: http://patchwork.ozlabs.org/project/linuxppc-dev/list/?series=131839 Changes since v1: Isolated atomic operations in functions *_lockless_pgtbl_walk() Fixed behavior of decrementing before last ptep was used Link: http://patchwork.ozlabs.org/patch/1163093/ Special thanks for: Aneesh Kumar, Nick Piggin, Paul Mackerras, Michael Ellerman, Fabiano Rosas, Dipankar Sarma and Oliver O'Halloran. Leonardo Bras (11): asm-generic/pgtable: Adds generic functions to track lockless pgtable walks mm/gup: Use functions to track lockless pgtbl walks on gup_pgd_range powerpc/mm: Adds arch-specificic functions to track lockless pgtable walks powerpc/mce_power: Use functions to track lockless pgtbl walks powerpc/perf: Use functions to track lockless pgtbl walks powerpc/mm/book3s64/hash: Use functions to track lockless pgtbl walks powerpc/kvm/e500: Use functions to track lockless pgtbl walks powerpc/kvm/book3s_hv: Use functions to track lockless pgtbl walks powerpc/kvm/book3s_64: Use functions to track lockless pgtbl walks powerpc/mm: Adds counting method to track lockless pagetable walks powerpc/mm/book3s64/pgtable: Uses counting method to skip serializing arch/powerpc/include/asm/book3s/64/pgtable.h | 6 + arch/powerpc/kernel/mce_power.c | 6 +- arch/powerpc/kvm/book3s_64_mmu_hv.c | 6 +- arch/powerpc/kvm/book3s_64_mmu_radix.c | 34 +++++- arch/powerpc/kvm/book3s_64_vio_hv.c | 6 +- arch/powerpc/kvm/book3s_hv_nested.c | 22 +++- arch/powerpc/kvm/book3s_hv_rm_mmu.c | 28 +++-- arch/powerpc/kvm/e500_mmu_host.c | 9 +- arch/powerpc/mm/book3s64/hash_tlb.c | 6 +- arch/powerpc/mm/book3s64/hash_utils.c | 27 +++-- arch/powerpc/mm/book3s64/pgtable.c | 120 ++++++++++++++++++- arch/powerpc/perf/callchain.c | 6 +- include/asm-generic/pgtable.h | 51 ++++++++ mm/gup.c | 10 +- 14 files changed, 288 insertions(+), 49 deletions(-) -- 2.24.1