Some regression is found by LKP-tools(linux kernel performance) on this patch series tested on Intel 2s/4s Skylake platform. The regression result is sorted by the metric will-it-scale.per_process_ops. Branch:Laurent-Dufour/Speculative-page-faults/20171011-213456(V4 patch series) Commit id: base:9a4b4dd1d8700dd5771f11dd2c048e4363efb493 head:56a4a8962fb32555a42eefdc9a19eeedd3e8c2e6 Benchmark suite:will-it-scale Download link:https://github.com/antonblanchard/will-it-scale/tree/master/tests Metrics: will-it-scale.per_process_ops=processes/nr_cpu will-it-scale.per_thread_ops=threads/nr_cpu tbox:lkp-skl-4sp1(nr_cpu=192,memory=768G) kconfig:CONFIG_TRANSPARENT_HUGEPAGE is not set testcase base change head metric brk1 2251803 -18.1% 1843535 will-it-scale.per_process_ops 341101 -17.5% 281284 will-it-scale.per_thread_ops malloc1 48833 -9.2% 44343 will-it-scale.per_process_ops 31555 +2.9% 32473 will-it-scale.per_thread_ops page_fault3 913019 -8.5% 835203 will-it-scale.per_process_ops 233978 -18.1% 191593 will-it-scale.per_thread_ops mmap2 95892 -6.6% 89536 will-it-scale.per_process_ops 90180 -13.7% 77803 will-it-scale.per_thread_ops mmap1 109586 -4.7% 104414 will-it-scale.per_process_ops 104477 -12.4% 91484 will-it-scale.per_thread_ops sched_yield 4964649 -2.1% 4859927 will-it-scale.per_process_ops 4946759 -1.7% 4864924 will-it-scale.per_thread_ops write1 1345159 -1.3% 1327719 will-it-scale.per_process_ops 1228754 -2.2% 1201915 will-it-scale.per_thread_ops page_fault2 202519 -1.0% 200545 will-it-scale.per_process_ops 96573 -10.4% 86526 will-it-scale.per_thread_ops page_fault1 225608 -0.9% 223585 will-it-scale.per_process_ops 105945 +14.4% 121199 will-it-scale.per_thread_ops tbox:lkp-skl-4sp1(nr_cpu=192,memory=768G) kconfig:CONFIG_TRANSPARENT_HUGEPAGE_ALWAYS=y testcase base change head metric context_switch1 333780 -23.0% 256927 will-it-scale.per_process_ops brk1 2263539 -18.8% 1837462 will-it-scale.per_process_ops 325854 -15.7% 274752 will-it-scale.per_thread_ops malloc1 48746 -13.5% 42148 will-it-scale.per_process_ops mmap1 106860 -12.4% 93634 will-it-scale.per_process_ops 98082 -18.9% 79506 will-it-scale.per_thread_ops mmap2 92468 -11.3% 82059 will-it-scale.per_process_ops 80468 -8.9% 73343 will-it-scale.per_thread_ops page_fault3 900709 -9.1% 818851 will-it-scale.per_process_ops 229837 -18.3% 187769 will-it-scale.per_thread_ops write1 1327409 -1.7% 1305048 will-it-scale.per_process_ops 1215658 -1.6% 1196479 will-it-scale.per_thread_ops writeseek3 300639 -1.6% 295882 will-it-scale.per_process_ops 231118 -2.2% 225929 will-it-scale.per_thread_ops signal1 122011 -1.5% 120155 will-it-scale.per_process_ops futex1 5123778 -1.2% 5062087 will-it-scale.per_process_ops page_fault2 202321 -1.0% 200289 will-it-scale.per_process_ops 93073 -9.8% 83927 will-it-scale.per_thread_ops tbox:lkp-skl-2sp2(nr_cpu=112,memory=64G) kconfig:CONFIG_TRANSPARENT_HUGEPAGE_ALWAYS=y testcase base change head metric brk1 2177903 -20.0% 1742054 will-it-scale.per_process_ops 434558 -15.3% 367896 will-it-scale.per_thread_ops malloc1 64871 -10.3% 58174 will-it-scale.per_process_ops page_fault3 882435 -9.0% 802892 will-it-scale.per_process_ops 299176 -15.7% 252170 will-it-scale.per_thread_ops mmap2 124567 -8.3% 114214 will-it-scale.per_process_ops 110674 -12.1% 97272 will-it-scale.per_thread_ops mmap1 137205 -7.8% 126440 will-it-scale.per_process_ops 128973 -15.1% 109560 will-it-scale.per_thread_ops context_switch1 343790 -7.2% 319209 will-it-scale.per_process_ops page_fault2 161891 -2.1% 158458 will-it-scale.per_process_ops 123278 -5.4% 116629 will-it-scale.per_thread_ops malloc2 14354856 -1.8% 14096856 will-it-scale.per_process_ops read2 1204838 -1.7% 1183993 will-it-scale.per_process_ops futex1 5017718 -1.6% 4938677 will-it-scale.per_process_ops 1408250 -1.0% 1394022 will-it-scale.per_thread_ops writeseek3 399651 -1.4% 393935 will-it-scale.per_process_ops signal1 157952 -1.0% 156302 will-it-scale.per_process_ops On 2017年10月11日 21:52, Laurent Dufour wrote: > This patch enable the speculative page fault on the PowerPC > architecture. > > This will try a speculative page fault without holding the mmap_sem, > if it returns with VM_FAULT_RETRY, the mmap_sem is acquired and the > traditional page fault processing is done. > > Build on if CONFIG_SPF is defined (currently for BOOK3S_64 && SMP). > > Signed-off-by: Laurent Dufour <ldufour@xxxxxxxxxxxxxxxxxx> > --- > arch/powerpc/mm/fault.c | 17 +++++++++++++++++ > 1 file changed, 17 insertions(+) > > diff --git a/arch/powerpc/mm/fault.c b/arch/powerpc/mm/fault.c > index 4797d08581ce..c018c2554cc8 100644 > --- a/arch/powerpc/mm/fault.c > +++ b/arch/powerpc/mm/fault.c > @@ -442,6 +442,20 @@ static int __do_page_fault(struct pt_regs *regs, unsigned long address, > if (is_exec) > flags |= FAULT_FLAG_INSTRUCTION; > > +#ifdef CONFIG_SPF > + if (is_user) { > + /* let's try a speculative page fault without grabbing the > + * mmap_sem. > + */ > + fault = handle_speculative_fault(mm, address, flags); > + if (!(fault & VM_FAULT_RETRY)) { > + perf_sw_event(PERF_COUNT_SW_SPF, 1, > + regs, address); > + goto done; > + } > + } > +#endif /* CONFIG_SPF */ > + > /* When running in the kernel we expect faults to occur only to > * addresses in user space. All other faults represent errors in the > * kernel and should generate an OOPS. Unfortunately, in the case of an > @@ -526,6 +540,9 @@ static int __do_page_fault(struct pt_regs *regs, unsigned long address, > > up_read(¤t->mm->mmap_sem); > > +#ifdef CONFIG_SPF > +done: > +#endif > if (unlikely(fault & VM_FAULT_ERROR)) > return mm_fault_error(regs, address, fault); > > -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@xxxxxxxxx. For more info on Linux MM, see: http://www.linux-mm.org/ . Don't email: <a href=mailto:"dont@xxxxxxxxx"> email@xxxxxxxxx </a>