Hello, kernel test robot noticed a 3888.9% improvement of will-it-scale.per_process_ops on: commit: d4148aeab412432bf928f311eca8a2ba52bb05df ("mm, mmap: limit THP alignment of anonymous mappings to PMD-aligned sizes") https://git.kernel.org/cgit/linux/kernel/git/torvalds/linux.git master [test failed on linux-next/master 5b913f5d7d7fe0f567dea8605f21da6eaa1735fb] testcase: will-it-scale config: x86_64-rhel-8.3 compiler: gcc-12 test machine: 224 threads 4 sockets Intel(R) Xeon(R) Platinum 8380H CPU @ 2.90GHz (Cooper Lake) with 192G memory parameters: nr_task: 100% mode: process test: malloc1 cpufreq_governor: performance In addition to that, the commit also has significant impact on the following tests: +------------------+---------------------------------------------------------------------------------------------+ | testcase: change | stress-ng: stress-ng.radixsort.ops_per_sec 9.2% regression | | test machine | 224 threads 2 sockets Intel(R) Xeon(R) Platinum 8480CTDX (Sapphire Rapids) with 256G memory | | test parameters | cpufreq_governor=performance | | | nr_threads=100% | | | test=radixsort | | | testtime=60s | +------------------+---------------------------------------------------------------------------------------------+ Details are as below: --------------------------------------------------------------------------------------------------> The kernel config and materials to reproduce are available at: https://download.01.org/0day-ci/archive/20241107/202411072132.a8d2cf0f-oliver.sang@xxxxxxxxx ========================================================================================= compiler/cpufreq_governor/kconfig/mode/nr_task/rootfs/tbox_group/test/testcase: gcc-12/performance/x86_64-rhel-8.3/process/100%/debian-12-x86_64-20240206.cgz/lkp-cpl-4sp2/malloc1/will-it-scale commit: 15e8156713 ("mm: shrinker: avoid memleak in alloc_shrinker_info") d4148aeab4 ("mm, mmap: limit THP alignment of anonymous mappings to PMD-aligned sizes") 15e8156713cc3803 d4148aeab412432bf928f311eca ---------------- --------------------------- %stddev %change %stddev \ | \ 255351 +4.5% 266817 vmstat.system.in 1.00 -0.8 0.20 ± 7% mpstat.cpu.all.irq% 0.13 +1.1 1.28 ± 32% mpstat.cpu.all.soft% 0.49 ± 15% +1.3 1.77 ± 27% mpstat.cpu.all.usr% 4143 ± 5% +181.3% 11654 ± 7% perf-c2c.DRAM.remote 2500 ± 3% +2209.9% 57748 ± 15% perf-c2c.HITM.local 3810 ± 2% +1461.0% 59473 ± 15% perf-c2c.HITM.total 78888 +3883.3% 3142322 ± 22% will-it-scale.224.processes 351.67 +3888.9% 14027 ± 22% will-it-scale.per_process_ops 78888 +3883.3% 3142322 ± 22% will-it-scale.workload 702940 ± 15% +123.4% 1570646 ± 11% meminfo.Active 702881 ± 15% +123.5% 1570598 ± 11% meminfo.Active(anon) 3953778 +28.4% 5075703 ± 4% meminfo.Cached 382618 +117.7% 832966 ± 10% meminfo.SUnreclaim 913559 ± 5% +122.8% 2035523 ± 11% meminfo.Shmem 528864 +85.5% 980938 ± 8% meminfo.Slab 12139918 +3685.1% 4.595e+08 ± 22% numa-numastat.node0.local_node 12236476 +3655.6% 4.595e+08 ± 22% numa-numastat.node0.numa_hit 12182775 +3990.8% 4.984e+08 ± 23% numa-numastat.node1.local_node 12264802 +3963.8% 4.984e+08 ± 23% numa-numastat.node1.numa_hit 12222359 +3921.0% 4.915e+08 ± 23% numa-numastat.node2.local_node 12320084 +3889.7% 4.915e+08 ± 23% numa-numastat.node2.numa_hit 12317018 +3904.7% 4.933e+08 ± 22% numa-numastat.node3.local_node 12388608 +3881.9% 4.933e+08 ± 22% numa-numastat.node3.numa_hit 27453504 ± 11% +20.0% 32957345 ± 2% sched_debug.cfs_rq:/.avg_vruntime.min 1.78 ± 2% +12.5% 2.00 ± 4% sched_debug.cfs_rq:/.h_nr_running.max 27453504 ± 11% +20.0% 32957377 ± 2% sched_debug.cfs_rq:/.min_vruntime.min 384.34 ± 6% +28.2% 492.57 ± 4% sched_debug.cfs_rq:/.util_est.avg 113.34 ± 7% +87.5% 212.55 ± 11% sched_debug.cfs_rq:/.util_est.stddev 395.67 ± 6% -91.8% 32.63 ± 7% sched_debug.cpu.clock.stddev 0.00 ± 5% -85.8% 0.00 ± 26% sched_debug.cpu.next_balance.stddev 1.78 ± 2% +12.5% 2.00 ± 8% sched_debug.cpu.nr_running.max 5.99 ± 9% -16.9% 4.98 ± 8% sched_debug.cpu.nr_uninterruptible.stddev 369561 ± 5% -48.5% 190196 ± 65% numa-meminfo.node0.AnonPages.max 108019 ± 4% +102.2% 218399 ± 12% numa-meminfo.node0.SUnreclaim 148129 ± 16% +86.2% 275784 ± 14% numa-meminfo.node0.Slab 73527 ± 61% -53.5% 34174 ± 80% numa-meminfo.node1.Mapped 94870 ± 4% +122.7% 211285 ± 9% numa-meminfo.node1.SUnreclaim 136858 ± 15% +83.3% 250864 ± 11% numa-meminfo.node1.Slab 89356 ± 2% +124.1% 200203 ± 10% numa-meminfo.node2.SUnreclaim 118047 ± 16% +89.0% 223166 ± 15% numa-meminfo.node2.Slab 698747 ± 15% +117.2% 1517832 ± 12% numa-meminfo.node3.Active 698731 ± 15% +117.2% 1517816 ± 12% numa-meminfo.node3.Active(anon) 90353 ± 7% +122.1% 200636 ± 10% numa-meminfo.node3.SUnreclaim 902916 ± 5% +118.4% 1972406 ± 12% numa-meminfo.node3.Shmem 125802 ± 17% +81.8% 228694 ± 8% numa-meminfo.node3.Slab 175727 ± 15% +123.5% 392704 ± 11% proc-vmstat.nr_active_anon 988408 +28.4% 1268872 ± 4% proc-vmstat.nr_file_pages 228353 ± 5% +122.8% 508826 ± 11% proc-vmstat.nr_shmem 36558 +1.2% 36996 proc-vmstat.nr_slab_reclaimable 95650 +117.2% 207775 ± 10% proc-vmstat.nr_slab_unreclaimable 175727 ± 15% +123.5% 392704 ± 11% proc-vmstat.nr_zone_active_anon 61863 ± 20% +50.4% 93011 ± 17% proc-vmstat.numa_hint_faults 27849 ± 47% +108.0% 57912 ± 22% proc-vmstat.numa_hint_faults_local 49211744 +3847.8% 1.943e+09 ± 22% proc-vmstat.numa_hit 48863846 +3875.5% 1.943e+09 ± 22% proc-vmstat.numa_local 102439 ± 38% +399.0% 511155 ± 5% proc-vmstat.pgactivate 1.218e+10 -83.3% 2.036e+09 ± 23% proc-vmstat.pgalloc_normal 25163379 +3665.3% 9.475e+08 ± 22% proc-vmstat.pgfault 1.218e+10 -83.3% 2.035e+09 ± 23% proc-vmstat.pgfree 23741008 -100.0% 0.00 proc-vmstat.thp_fault_alloc 27004 ± 4% +102.3% 54625 ± 12% numa-vmstat.node0.nr_slab_unreclaimable 12236166 +3655.7% 4.595e+08 ± 22% numa-vmstat.node0.numa_hit 12139608 +3685.2% 4.595e+08 ± 22% numa-vmstat.node0.numa_local 18419 ± 61% -53.5% 8573 ± 80% numa-vmstat.node1.nr_mapped 23716 ± 4% +122.9% 52871 ± 9% numa-vmstat.node1.nr_slab_unreclaimable 12263982 +3964.1% 4.984e+08 ± 23% numa-vmstat.node1.numa_hit 12181955 +3991.1% 4.984e+08 ± 23% numa-vmstat.node1.numa_local 22339 ± 2% +124.8% 50226 ± 10% numa-vmstat.node2.nr_slab_unreclaimable 12319433 +3889.9% 4.915e+08 ± 23% numa-vmstat.node2.numa_hit 12221708 +3921.2% 4.915e+08 ± 23% numa-vmstat.node2.numa_local 174568 ± 15% +117.3% 379260 ± 12% numa-vmstat.node3.nr_active_anon 225581 ± 5% +118.6% 493053 ± 12% numa-vmstat.node3.nr_shmem 22588 ± 7% +122.8% 50325 ± 11% numa-vmstat.node3.nr_slab_unreclaimable 174566 ± 15% +117.3% 379259 ± 12% numa-vmstat.node3.nr_zone_active_anon 12386775 +3882.4% 4.933e+08 ± 22% numa-vmstat.node3.numa_hit 12315185 +3905.2% 4.933e+08 ± 22% numa-vmstat.node3.numa_local 20.80 -88.0% 2.51 ± 4% perf-stat.i.MPKI 1.314e+09 +1555.3% 2.175e+10 ± 18% perf-stat.i.branch-instructions 0.79 ± 3% -0.4 0.35 perf-stat.i.branch-miss-rate% 12157447 ± 2% +510.1% 74172493 ± 18% perf-stat.i.branch-misses 61.19 -29.2 31.97 ± 2% perf-stat.i.cache-miss-rate% 1.255e+08 +104.5% 2.567e+08 ± 19% perf-stat.i.cache-misses 2.043e+08 +292.8% 8.024e+08 ± 18% perf-stat.i.cache-references 139.74 -93.8% 8.69 ± 27% perf-stat.i.cpi 264.51 +8.5% 286.93 perf-stat.i.cpu-migrations 6664 -47.9% 3471 ± 28% perf-stat.i.cycles-between-cache-misses 6.183e+09 +1555.5% 1.024e+11 ± 18% perf-stat.i.instructions 0.01 ± 23% +808.8% 0.12 ± 18% perf-stat.i.ipc 82374 +3701.0% 3131041 ± 22% perf-stat.i.minor-faults 82375 +3701.0% 3131042 ± 22% perf-stat.i.page-faults 20.32 -91.8% 1.67 ± 70% perf-stat.overall.MPKI 0.92 ± 2% -0.7 0.23 ± 70% perf-stat.overall.branch-miss-rate% 61.44 -39.9 21.51 ± 70% perf-stat.overall.cache-miss-rate% 136.21 -95.6% 6.05 ± 80% perf-stat.overall.cpi 6703 -63.8% 2428 ± 80% perf-stat.overall.cycles-between-cache-misses 23815471 -71.7% 6746184 ± 71% perf-stat.overall.path-length 2.41 ± 6% -93.2% 0.16 ± 78% perf-sched.sch_delay.avg.ms.__cond_resched.__alloc_pages_noprof.alloc_pages_mpol_noprof.folio_alloc_mpol_noprof.vma_alloc_folio_noprof 0.60 ± 21% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.__alloc_pages_noprof.alloc_pages_mpol_noprof.pte_alloc_one.__do_huge_pmd_anonymous_page 0.74 ± 36% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.__anon_vma_prepare.__vmf_anon_prepare.do_huge_pmd_anonymous_page.__handle_mm_fault 2.00 ± 27% -76.6% 0.47 ±149% perf-sched.sch_delay.avg.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.mmap_region 0.35 ± 4% -78.5% 0.07 ± 10% perf-sched.sch_delay.avg.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity 1.99 ± 20% -77.4% 0.45 ± 93% perf-sched.sch_delay.avg.ms.__cond_resched.dput.step_into.link_path_walk.part 2.15 ± 14% -98.2% 0.04 ±196% perf-sched.sch_delay.avg.ms.__cond_resched.dput.terminate_walk.path_openat.do_filp_open 0.57 ± 7% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.folio_zero_user.__do_huge_pmd_anonymous_page.__handle_mm_fault.handle_mm_fault 0.84 ± 96% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.kmem_cache_alloc_noprof.__anon_vma_prepare.__vmf_anon_prepare.do_huge_pmd_anonymous_page 1.41 ± 39% -93.0% 0.10 ±223% perf-sched.sch_delay.avg.ms.__cond_resched.kmem_cache_alloc_noprof.getname_flags.part.0 1.52 ± 31% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.kmem_cache_alloc_noprof.security_file_alloc.init_file.alloc_empty_file 1.31 ± 17% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.mutex_lock.fifo_open.do_dentry_open.vfs_open 0.05 ± 4% -76.9% 0.01 ± 27% perf-sched.sch_delay.avg.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm 1.13 ± 33% -99.6% 0.00 ±223% perf-sched.sch_delay.avg.ms.__cond_resched.task_work_run.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown] 2.20 ± 19% -93.4% 0.14 ±223% perf-sched.sch_delay.avg.ms.__cond_resched.wp_page_copy.__handle_mm_fault.handle_mm_fault.do_user_addr_fault 1.65 ± 30% -70.6% 0.48 ± 65% perf-sched.sch_delay.avg.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep 0.50 ± 10% -89.7% 0.05 ± 50% perf-sched.sch_delay.avg.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64 0.03 ± 21% -63.7% 0.01 ± 53% perf-sched.sch_delay.avg.ms.ipmi_thread.kthread.ret_from_fork.ret_from_fork_asm 0.93 ± 11% -54.4% 0.42 ± 35% perf-sched.sch_delay.avg.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown].[unknown] 2.62 ± 33% -51.4% 1.27 ± 32% perf-sched.sch_delay.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown].[unknown] 0.81 ±133% -94.4% 0.05 ± 13% perf-sched.sch_delay.avg.ms.pipe_read.vfs_read.ksys_read.do_syscall_64 1.11 ± 12% -51.4% 0.54 ± 20% perf-sched.sch_delay.avg.ms.schedule_hrtimeout_range_clock.do_poll.constprop.0.do_sys_poll 0.49 ± 37% -77.7% 0.11 ± 52% perf-sched.sch_delay.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone 0.08 ± 3% -71.3% 0.02 ±107% perf-sched.sch_delay.avg.ms.schedule_timeout.kcompactd.kthread.ret_from_fork 0.05 ± 33% -80.9% 0.01 ± 11% perf-sched.sch_delay.avg.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread 0.05 ± 8% -81.4% 0.01 ± 41% perf-sched.sch_delay.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm 2.15 ± 13% -77.3% 0.49 ± 28% perf-sched.sch_delay.avg.ms.wait_for_partner.fifo_open.do_dentry_open.vfs_open 4.04 -57.7% 1.71 ± 97% perf-sched.sch_delay.max.ms.__cond_resched.__alloc_pages_noprof.alloc_pages_mpol_noprof.folio_alloc_mpol_noprof.vma_alloc_folio_noprof 4.09 ± 2% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.__alloc_pages_noprof.alloc_pages_mpol_noprof.pte_alloc_one.__do_huge_pmd_anonymous_page 1.19 ± 74% +593.4% 8.27 ± 62% perf-sched.sch_delay.max.ms.__cond_resched.__alloc_pages_noprof.alloc_pages_mpol_noprof.pte_alloc_one.__pte_alloc 0.84 ±173% +695.6% 6.69 ± 78% perf-sched.sch_delay.max.ms.__cond_resched.__anon_vma_prepare.__vmf_anon_prepare.do_anonymous_page.__handle_mm_fault 4.00 -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.__anon_vma_prepare.__vmf_anon_prepare.do_huge_pmd_anonymous_page.__handle_mm_fault 2.71 ± 21% -82.7% 0.47 ±149% perf-sched.sch_delay.max.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.mmap_region 7.66 ± 93% -91.2% 0.67 ±223% perf-sched.sch_delay.max.ms.__cond_resched.__kmalloc_cache_noprof.vmstat_start.seq_read_iter.proc_reg_read_iter 0.01 ±223% +9242.3% 0.81 ±179% perf-sched.sch_delay.max.ms.__cond_resched.__kmalloc_node_noprof.alloc_slab_obj_exts.allocate_slab.___slab_alloc 8.12 ±103% +186.4% 23.26 ± 73% perf-sched.sch_delay.max.ms.__cond_resched.__tlb_batch_free_encoded_pages.tlb_finish_mmu.vms_clear_ptes.vms_complete_munmap_vmas 2.34 ± 11% -75.7% 0.57 ± 75% perf-sched.sch_delay.max.ms.__cond_resched.dput.step_into.link_path_walk.part 2.77 ± 21% -96.7% 0.09 ±212% perf-sched.sch_delay.max.ms.__cond_resched.dput.terminate_walk.path_openat.do_filp_open 36.48 ± 69% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.folio_zero_user.__do_huge_pmd_anonymous_page.__handle_mm_fault.handle_mm_fault 0.37 ±223% +998.1% 4.04 ± 35% perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc_noprof.__anon_vma_prepare.__vmf_anon_prepare.do_anonymous_page 2.26 ± 78% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc_noprof.__anon_vma_prepare.__vmf_anon_prepare.do_huge_pmd_anonymous_page 1.83 ± 40% -94.6% 0.10 ±223% perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc_noprof.getname_flags.part.0 1.75 ± 25% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc_noprof.security_file_alloc.init_file.alloc_empty_file 1.51 ± 30% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.mutex_lock.fifo_open.do_dentry_open.vfs_open 0.13 ± 50% -86.1% 0.02 ± 46% perf-sched.sch_delay.max.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm 4.17 ± 5% -99.8% 0.01 ±223% perf-sched.sch_delay.max.ms.__cond_resched.task_work_run.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown] 4.02 +510.8% 24.57 ± 38% perf-sched.sch_delay.max.ms.__cond_resched.unmap_vmas.vms_clear_ptes.vms_complete_munmap_vmas.do_vmi_align_munmap 3.99 -92.2% 0.31 ±223% perf-sched.sch_delay.max.ms.__cond_resched.wp_page_copy.__handle_mm_fault.handle_mm_fault.do_user_addr_fault 4.11 ± 5% +33.0% 5.47 ± 27% perf-sched.sch_delay.max.ms.__cond_resched.zap_pmd_range.isra.0.unmap_page_range 4.54 ± 12% -54.1% 2.08 ± 65% perf-sched.sch_delay.max.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep 4.00 ± 25% -70.1% 1.20 ± 30% perf-sched.sch_delay.max.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64 0.03 ± 17% -60.8% 0.01 ± 64% perf-sched.sch_delay.max.ms.ipmi_thread.kthread.ret_from_fork.ret_from_fork_asm 4.08 ± 2% -61.1% 1.59 ±108% perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown] 389.98 ±121% -96.7% 13.00 ± 33% perf-sched.sch_delay.max.ms.pipe_read.vfs_read.ksys_read.do_syscall_64 9.71 ±119% -78.8% 2.06 ± 58% perf-sched.sch_delay.max.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone 8.60 ±210% -99.1% 0.08 ± 36% perf-sched.sch_delay.max.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread 4.05 ± 2% -34.0% 2.67 ± 42% perf-sched.sch_delay.max.ms.wait_for_partner.fifo_open.do_dentry_open.vfs_open 8.02 ± 5% -55.7% 3.55 ± 11% perf-sched.sch_delay.max.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm 0.47 ± 5% -13.9% 0.40 ± 13% perf-sched.total_sch_delay.average.ms 396.80 ± 9% -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.__cond_resched.__kmalloc_cache_noprof.vmstat_start.seq_read_iter.proc_reg_read_iter 0.17 ±223% +596.6% 1.17 ± 14% perf-sched.wait_and_delay.avg.ms.__cond_resched.__tlb_batch_free_encoded_pages.tlb_finish_mmu.vms_clear_ptes.vms_complete_munmap_vmas 1.15 ± 7% -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.__cond_resched.folio_zero_user.__do_huge_pmd_anonymous_page.__handle_mm_fault.handle_mm_fault 417.15 ± 5% +75.4% 731.48 ± 12% perf-sched.wait_and_delay.avg.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep 11.50 ± 25% -100.0% 0.00 perf-sched.wait_and_delay.count.__cond_resched.__kmalloc_cache_noprof.vmstat_start.seq_read_iter.proc_reg_read_iter 26.17 ±223% +15367.5% 4047 ± 49% perf-sched.wait_and_delay.count.__cond_resched.__tlb_batch_free_encoded_pages.tlb_finish_mmu.vms_clear_ptes.vms_complete_munmap_vmas 15379 ± 15% -100.0% 0.00 perf-sched.wait_and_delay.count.__cond_resched.folio_zero_user.__do_huge_pmd_anonymous_page.__handle_mm_fault.handle_mm_fault 30.33 ±223% +836.8% 284.17 ± 53% perf-sched.wait_and_delay.count.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown].[unknown] 1000 -100.0% 0.00 perf-sched.wait_and_delay.max.ms.__cond_resched.__kmalloc_cache_noprof.vmstat_start.seq_read_iter.proc_reg_read_iter 1.34 ±223% +3359.9% 46.51 ± 73% perf-sched.wait_and_delay.max.ms.__cond_resched.__tlb_batch_free_encoded_pages.tlb_finish_mmu.vms_clear_ptes.vms_complete_munmap_vmas 72.95 ± 69% -100.0% 0.00 perf-sched.wait_and_delay.max.ms.__cond_resched.folio_zero_user.__do_huge_pmd_anonymous_page.__handle_mm_fault.handle_mm_fault 13.11 ±122% -61.3% 5.08 perf-sched.wait_and_delay.max.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread 2.41 ± 6% -93.5% 0.16 ± 76% perf-sched.wait_time.avg.ms.__cond_resched.__alloc_pages_noprof.alloc_pages_mpol_noprof.folio_alloc_mpol_noprof.vma_alloc_folio_noprof 0.60 ± 21% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.__alloc_pages_noprof.alloc_pages_mpol_noprof.pte_alloc_one.__do_huge_pmd_anonymous_page 0.74 ± 36% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.__anon_vma_prepare.__vmf_anon_prepare.do_huge_pmd_anonymous_page.__handle_mm_fault 2.00 ± 27% -76.6% 0.47 ±149% perf-sched.wait_time.avg.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.mmap_region 394.43 ± 10% -99.8% 0.67 ±223% perf-sched.wait_time.avg.ms.__cond_resched.__kmalloc_cache_noprof.vmstat_start.seq_read_iter.proc_reg_read_iter 1.99 ± 20% -77.8% 0.44 ± 95% perf-sched.wait_time.avg.ms.__cond_resched.dput.step_into.link_path_walk.part 2.15 ± 14% -98.4% 0.03 ±193% perf-sched.wait_time.avg.ms.__cond_resched.dput.terminate_walk.path_openat.do_filp_open 0.57 ± 7% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.folio_zero_user.__do_huge_pmd_anonymous_page.__handle_mm_fault.handle_mm_fault 0.84 ± 96% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc_noprof.__anon_vma_prepare.__vmf_anon_prepare.do_huge_pmd_anonymous_page 1.41 ± 40% -93.0% 0.10 ±223% perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc_noprof.getname_flags.part.0 84.95 ±219% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc_noprof.security_file_alloc.init_file.alloc_empty_file 1.31 ± 17% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.mutex_lock.fifo_open.do_dentry_open.vfs_open 1.13 ± 33% -99.6% 0.00 ±223% perf-sched.wait_time.avg.ms.__cond_resched.task_work_run.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown] 2.11 ± 23% -93.1% 0.14 ±223% perf-sched.wait_time.avg.ms.__cond_resched.wp_page_copy.__handle_mm_fault.handle_mm_fault.do_user_addr_fault 415.50 ± 5% +75.9% 731.00 ± 13% perf-sched.wait_time.avg.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep 5.81 ± 41% -74.2% 1.50 ± 9% perf-sched.wait_time.avg.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64 0.93 ± 12% -54.5% 0.42 ± 35% perf-sched.wait_time.avg.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown].[unknown] 9.89 ±101% -87.1% 1.27 ± 32% perf-sched.wait_time.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown].[unknown] 13.22 ± 5% -77.3% 3.01 ± 14% perf-sched.wait_time.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone 2.13 ± 15% -77.6% 0.48 ± 28% perf-sched.wait_time.avg.ms.wait_for_partner.fifo_open.do_dentry_open.vfs_open 4.04 -57.7% 1.71 ± 97% perf-sched.wait_time.max.ms.__cond_resched.__alloc_pages_noprof.alloc_pages_mpol_noprof.folio_alloc_mpol_noprof.vma_alloc_folio_noprof 4.09 ± 2% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.__alloc_pages_noprof.alloc_pages_mpol_noprof.pte_alloc_one.__do_huge_pmd_anonymous_page 1.19 ± 74% +593.4% 8.27 ± 62% perf-sched.wait_time.max.ms.__cond_resched.__alloc_pages_noprof.alloc_pages_mpol_noprof.pte_alloc_one.__pte_alloc 0.84 ±173% +695.6% 6.69 ± 78% perf-sched.wait_time.max.ms.__cond_resched.__anon_vma_prepare.__vmf_anon_prepare.do_anonymous_page.__handle_mm_fault 4.00 -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.__anon_vma_prepare.__vmf_anon_prepare.do_huge_pmd_anonymous_page.__handle_mm_fault 2.71 ± 21% -82.7% 0.47 ±149% perf-sched.wait_time.max.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.mmap_region 1000 -99.9% 0.67 ±223% perf-sched.wait_time.max.ms.__cond_resched.__kmalloc_cache_noprof.vmstat_start.seq_read_iter.proc_reg_read_iter 8.12 ±103% +186.4% 23.26 ± 73% perf-sched.wait_time.max.ms.__cond_resched.__tlb_batch_free_encoded_pages.tlb_finish_mmu.vms_clear_ptes.vms_complete_munmap_vmas 2.34 ± 11% -75.7% 0.57 ± 75% perf-sched.wait_time.max.ms.__cond_resched.dput.step_into.link_path_walk.part 2.77 ± 21% -96.7% 0.09 ±212% perf-sched.wait_time.max.ms.__cond_resched.dput.terminate_walk.path_openat.do_filp_open 36.48 ± 69% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.folio_zero_user.__do_huge_pmd_anonymous_page.__handle_mm_fault.handle_mm_fault 0.37 ±223% +998.1% 4.04 ± 35% perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc_noprof.__anon_vma_prepare.__vmf_anon_prepare.do_anonymous_page 2.26 ± 78% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc_noprof.__anon_vma_prepare.__vmf_anon_prepare.do_huge_pmd_anonymous_page 1.83 ± 40% -94.6% 0.10 ±223% perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc_noprof.getname_flags.part.0 168.46 ±221% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc_noprof.security_file_alloc.init_file.alloc_empty_file 1.51 ± 30% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.mutex_lock.fifo_open.do_dentry_open.vfs_open 4.17 ± 5% -99.8% 0.01 ±223% perf-sched.wait_time.max.ms.__cond_resched.task_work_run.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown] 4.02 +510.8% 24.57 ± 38% perf-sched.wait_time.max.ms.__cond_resched.unmap_vmas.vms_clear_ptes.vms_complete_munmap_vmas.do_vmi_align_munmap 3.99 -92.2% 0.31 ±223% perf-sched.wait_time.max.ms.__cond_resched.wp_page_copy.__handle_mm_fault.handle_mm_fault.do_user_addr_fault 4.11 ± 5% +33.0% 5.47 ± 27% perf-sched.wait_time.max.ms.__cond_resched.zap_pmd_range.isra.0.unmap_page_range 250.47 ±134% -89.0% 27.62 ± 59% perf-sched.wait_time.max.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64 4.08 ± 2% -61.1% 1.59 ±108% perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown] 499.41 +278.7% 1891 ± 64% perf-sched.wait_time.max.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait 54.24 ± 46% -59.6% 21.93 ± 68% perf-sched.wait_time.max.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone 4.05 ± 2% -34.0% 2.67 ± 42% perf-sched.wait_time.max.ms.wait_for_partner.fifo_open.do_dentry_open.vfs_open 96.58 -96.6 0.00 perf-profile.calltrace.cycles-pp.folio_zero_user.__do_huge_pmd_anonymous_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault 96.37 -96.4 0.00 perf-profile.calltrace.cycles-pp.__do_huge_pmd_anonymous_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault 96.69 -90.4 6.27 ± 11% perf-profile.calltrace.cycles-pp.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault 96.73 -90.3 6.45 ± 11% perf-profile.calltrace.cycles-pp.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault 97.00 -90.2 6.79 ± 10% perf-profile.calltrace.cycles-pp.asm_exc_page_fault 96.77 -90.1 6.68 ± 10% perf-profile.calltrace.cycles-pp.do_user_addr_fault.exc_page_fault.asm_exc_page_fault 96.77 -90.1 6.69 ± 10% perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault 87.45 -87.4 0.00 perf-profile.calltrace.cycles-pp.clear_page_erms.folio_zero_user.__do_huge_pmd_anonymous_page.__handle_mm_fault.handle_mm_fault 0.00 +0.6 0.61 ± 11% perf-profile.calltrace.cycles-pp.lru_add.folio_batch_move_lru.lru_add_drain_cpu.lru_add_drain.vms_clear_ptes 0.00 +0.6 0.62 ± 12% perf-profile.calltrace.cycles-pp.__memcg_slab_post_alloc_hook.kmem_cache_alloc_noprof.__anon_vma_prepare.__vmf_anon_prepare.do_anonymous_page 0.00 +0.7 0.67 ± 10% perf-profile.calltrace.cycles-pp.kmem_cache_alloc_noprof.__anon_vma_prepare.__vmf_anon_prepare.do_anonymous_page.__handle_mm_fault 0.00 +0.7 0.69 ± 27% perf-profile.calltrace.cycles-pp.get_mem_cgroup_from_mm.__mem_cgroup_charge.alloc_anon_folio.do_anonymous_page.__handle_mm_fault 0.00 +0.8 0.82 ± 10% perf-profile.calltrace.cycles-pp.__lruvec_stat_mod_folio.folio_remove_rmap_ptes.zap_present_ptes.zap_pte_range.zap_pmd_range 0.00 +0.8 0.84 ± 13% perf-profile.calltrace.cycles-pp.__lruvec_stat_mod_folio.pte_alloc_one.__pte_alloc.do_anonymous_page.__handle_mm_fault 0.00 +0.8 0.84 ± 10% perf-profile.calltrace.cycles-pp.folio_remove_rmap_ptes.zap_present_ptes.zap_pte_range.zap_pmd_range.unmap_page_range 0.00 +0.8 0.85 ± 8% perf-profile.calltrace.cycles-pp.__lruvec_stat_mod_folio.___pte_free_tlb.free_pud_range.free_p4d_range.free_pgd_range 0.00 +0.9 0.86 ± 9% perf-profile.calltrace.cycles-pp.__anon_vma_prepare.__vmf_anon_prepare.do_anonymous_page.__handle_mm_fault.handle_mm_fault 0.00 +0.9 0.87 ± 8% perf-profile.calltrace.cycles-pp.___pte_free_tlb.free_pud_range.free_p4d_range.free_pgd_range.free_pgtables 0.00 +0.9 0.88 ± 9% perf-profile.calltrace.cycles-pp.__lruvec_stat_mod_folio.__folio_mod_stat.folio_add_new_anon_rmap.do_anonymous_page.__handle_mm_fault 0.00 +0.9 0.88 ± 10% perf-profile.calltrace.cycles-pp.__folio_mod_stat.folio_add_new_anon_rmap.do_anonymous_page.__handle_mm_fault.handle_mm_fault 0.00 +0.9 0.90 ± 9% perf-profile.calltrace.cycles-pp.zap_present_ptes.zap_pte_range.zap_pmd_range.unmap_page_range.unmap_vmas 0.00 +0.9 0.90 ± 9% perf-profile.calltrace.cycles-pp.__vmf_anon_prepare.do_anonymous_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault 0.00 +0.9 0.91 ± 10% perf-profile.calltrace.cycles-pp.folio_add_new_anon_rmap.do_anonymous_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault 0.00 +1.0 1.03 ± 9% perf-profile.calltrace.cycles-pp.free_pud_range.free_p4d_range.free_pgd_range.free_pgtables.vms_clear_ptes 0.00 +1.0 1.05 ± 9% perf-profile.calltrace.cycles-pp.free_p4d_range.free_pgd_range.free_pgtables.vms_clear_ptes.vms_complete_munmap_vmas 0.00 +1.1 1.07 ± 9% perf-profile.calltrace.cycles-pp.free_pgd_range.free_pgtables.vms_clear_ptes.vms_complete_munmap_vmas.do_vmi_align_munmap 0.00 +1.1 1.11 ± 75% perf-profile.calltrace.cycles-pp.uncharge_folio.__mem_cgroup_uncharge_folios.folios_put_refs.free_pages_and_swap_cache.__tlb_batch_free_encoded_pages 0.00 +1.3 1.29 ± 46% perf-profile.calltrace.cycles-pp.__mem_cgroup_charge.alloc_anon_folio.do_anonymous_page.__handle_mm_fault.handle_mm_fault 0.00 +1.4 1.38 ± 9% perf-profile.calltrace.cycles-pp.free_pgtables.vms_clear_ptes.vms_complete_munmap_vmas.do_vmi_align_munmap.do_vmi_munmap 0.00 +1.5 1.52 ± 28% perf-profile.calltrace.cycles-pp.__memcg_kmem_charge_page.__alloc_pages_noprof.alloc_pages_mpol_noprof.pte_alloc_one.__pte_alloc 0.00 +1.5 1.54 ± 34% perf-profile.calltrace.cycles-pp.alloc_anon_folio.do_anonymous_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault 0.00 +1.7 1.73 ± 22% perf-profile.calltrace.cycles-pp.__alloc_pages_noprof.alloc_pages_mpol_noprof.pte_alloc_one.__pte_alloc.do_anonymous_page 0.00 +1.7 1.74 ± 18% perf-profile.calltrace.cycles-pp.mmap_region.do_mmap.vm_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe 0.00 +1.8 1.75 ± 21% perf-profile.calltrace.cycles-pp.alloc_pages_mpol_noprof.pte_alloc_one.__pte_alloc.do_anonymous_page.__handle_mm_fault 0.00 +2.0 2.02 ± 14% perf-profile.calltrace.cycles-pp.zap_pte_range.zap_pmd_range.unmap_page_range.unmap_vmas.vms_clear_ptes 0.00 +2.1 2.10 ± 19% perf-profile.calltrace.cycles-pp.do_mmap.vm_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe.__mmap 0.00 +2.2 2.22 ± 19% perf-profile.calltrace.cycles-pp.vm_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe.__mmap 0.00 +2.3 2.28 ± 19% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.__mmap 0.00 +2.3 2.29 ± 19% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.__mmap 0.00 +2.5 2.46 ± 19% perf-profile.calltrace.cycles-pp.__mmap 0.00 +2.5 2.47 ± 16% perf-profile.calltrace.cycles-pp.zap_pmd_range.unmap_page_range.unmap_vmas.vms_clear_ptes.vms_complete_munmap_vmas 0.00 +2.5 2.54 ± 16% perf-profile.calltrace.cycles-pp.unmap_page_range.unmap_vmas.vms_clear_ptes.vms_complete_munmap_vmas.do_vmi_align_munmap 0.00 +2.6 2.58 ± 16% perf-profile.calltrace.cycles-pp.unmap_vmas.vms_clear_ptes.vms_complete_munmap_vmas.do_vmi_align_munmap.do_vmi_munmap 0.00 +2.6 2.63 ± 12% perf-profile.calltrace.cycles-pp.pte_alloc_one.__pte_alloc.do_anonymous_page.__handle_mm_fault.handle_mm_fault 0.00 +2.7 2.69 ± 12% perf-profile.calltrace.cycles-pp.__pte_alloc.do_anonymous_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault 0.00 +3.2 3.21 ± 38% perf-profile.calltrace.cycles-pp.page_counter_cancel.page_counter_uncharge.uncharge_batch.__mem_cgroup_uncharge_folios.folios_put_refs 0.00 +4.0 4.02 ± 37% perf-profile.calltrace.cycles-pp.page_counter_uncharge.uncharge_batch.__mem_cgroup_uncharge_folios.folios_put_refs.free_pages_and_swap_cache 0.00 +5.0 5.04 ± 23% perf-profile.calltrace.cycles-pp.uncharge_batch.__mem_cgroup_uncharge_folios.folios_put_refs.free_pages_and_swap_cache.__tlb_batch_free_encoded_pages 0.00 +6.2 6.16 ± 7% perf-profile.calltrace.cycles-pp.__mem_cgroup_uncharge_folios.folios_put_refs.free_pages_and_swap_cache.__tlb_batch_free_encoded_pages.tlb_finish_mmu 0.00 +6.2 6.16 ± 12% perf-profile.calltrace.cycles-pp.do_anonymous_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault 0.00 +36.4 36.44 ± 2% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.folio_lruvec_lock_irqsave.folio_batch_move_lru.lru_add_drain_cpu 0.00 +36.6 36.57 ± 2% perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.folio_lruvec_lock_irqsave.folio_batch_move_lru.lru_add_drain_cpu.lru_add_drain 0.00 +36.8 36.80 ± 3% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.folio_lruvec_lock_irqsave.__page_cache_release.folios_put_refs 0.00 +36.9 36.93 ± 3% perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.folio_lruvec_lock_irqsave.__page_cache_release.folios_put_refs.free_pages_and_swap_cache 0.00 +37.0 37.00 perf-profile.calltrace.cycles-pp.folio_lruvec_lock_irqsave.folio_batch_move_lru.lru_add_drain_cpu.lru_add_drain.vms_clear_ptes 0.00 +37.3 37.33 ± 3% perf-profile.calltrace.cycles-pp.folio_lruvec_lock_irqsave.__page_cache_release.folios_put_refs.free_pages_and_swap_cache.__tlb_batch_free_encoded_pages 0.00 +37.9 37.87 ± 3% perf-profile.calltrace.cycles-pp.__page_cache_release.folios_put_refs.free_pages_and_swap_cache.__tlb_batch_free_encoded_pages.tlb_finish_mmu 0.00 +38.4 38.42 perf-profile.calltrace.cycles-pp.folio_batch_move_lru.lru_add_drain_cpu.lru_add_drain.vms_clear_ptes.vms_complete_munmap_vmas 0.00 +38.4 38.45 perf-profile.calltrace.cycles-pp.lru_add_drain_cpu.lru_add_drain.vms_clear_ptes.vms_complete_munmap_vmas.do_vmi_align_munmap 0.00 +38.5 38.45 perf-profile.calltrace.cycles-pp.lru_add_drain.vms_clear_ptes.vms_complete_munmap_vmas.do_vmi_align_munmap.do_vmi_munmap 0.78 ± 2% +44.1 44.87 perf-profile.calltrace.cycles-pp.folios_put_refs.free_pages_and_swap_cache.__tlb_batch_free_encoded_pages.tlb_finish_mmu.vms_clear_ptes 0.79 ± 2% +44.2 44.99 perf-profile.calltrace.cycles-pp.free_pages_and_swap_cache.__tlb_batch_free_encoded_pages.tlb_finish_mmu.vms_clear_ptes.vms_complete_munmap_vmas 0.79 ± 2% +44.2 45.00 perf-profile.calltrace.cycles-pp.__tlb_batch_free_encoded_pages.tlb_finish_mmu.vms_clear_ptes.vms_complete_munmap_vmas.do_vmi_align_munmap 0.86 ± 2% +44.4 45.24 perf-profile.calltrace.cycles-pp.tlb_finish_mmu.vms_clear_ptes.vms_complete_munmap_vmas.do_vmi_align_munmap.do_vmi_munmap 1.23 ± 2% +86.5 87.70 perf-profile.calltrace.cycles-pp.vms_clear_ptes.vms_complete_munmap_vmas.do_vmi_align_munmap.do_vmi_munmap.__vm_munmap 1.25 +86.6 87.86 perf-profile.calltrace.cycles-pp.vms_complete_munmap_vmas.do_vmi_align_munmap.do_vmi_munmap.__vm_munmap.__x64_sys_munmap 1.44 +87.1 88.55 perf-profile.calltrace.cycles-pp.do_vmi_align_munmap.do_vmi_munmap.__vm_munmap.__x64_sys_munmap.do_syscall_64 1.48 +87.1 88.60 perf-profile.calltrace.cycles-pp.do_vmi_munmap.__vm_munmap.__x64_sys_munmap.do_syscall_64.entry_SYSCALL_64_after_hwframe 1.49 +87.2 88.66 perf-profile.calltrace.cycles-pp.__vm_munmap.__x64_sys_munmap.do_syscall_64.entry_SYSCALL_64_after_hwframe.__munmap 1.50 +87.2 88.66 perf-profile.calltrace.cycles-pp.__x64_sys_munmap.do_syscall_64.entry_SYSCALL_64_after_hwframe.__munmap 1.52 +87.2 88.72 perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.__munmap 1.52 +87.2 88.73 perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.__munmap 1.53 +87.4 88.90 perf-profile.calltrace.cycles-pp.__munmap 96.37 -96.4 0.00 perf-profile.children.cycles-pp.__do_huge_pmd_anonymous_page 95.62 -95.6 0.00 perf-profile.children.cycles-pp.folio_zero_user 96.71 -90.4 6.28 ± 11% perf-profile.children.cycles-pp.__handle_mm_fault 96.75 -90.3 6.46 ± 11% perf-profile.children.cycles-pp.handle_mm_fault 97.02 -90.2 6.83 ± 10% perf-profile.children.cycles-pp.asm_exc_page_fault 96.79 -90.1 6.68 ± 10% perf-profile.children.cycles-pp.do_user_addr_fault 96.80 -90.1 6.70 ± 10% perf-profile.children.cycles-pp.exc_page_fault 89.46 -89.3 0.20 ± 22% perf-profile.children.cycles-pp.clear_page_erms 4.15 -3.8 0.35 ± 23% perf-profile.children.cycles-pp.__cond_resched 0.81 ± 3% -0.6 0.23 ± 7% perf-profile.children.cycles-pp.__sysvec_apic_timer_interrupt 0.80 ± 3% -0.6 0.22 ± 7% perf-profile.children.cycles-pp.hrtimer_interrupt 0.73 ± 3% -0.5 0.18 ± 8% perf-profile.children.cycles-pp.__hrtimer_run_queues 0.60 ± 4% -0.5 0.14 ± 5% perf-profile.children.cycles-pp.tick_nohz_handler 0.83 ± 16% -0.4 0.39 ± 3% perf-profile.children.cycles-pp.__cmd_record 0.53 ± 3% -0.4 0.12 ± 6% perf-profile.children.cycles-pp.update_process_times 0.52 -0.4 0.12 ± 28% perf-profile.children.cycles-pp.free_unref_folios 0.44 ± 3% -0.3 0.12 ± 6% perf-profile.children.cycles-pp.cmd_record 0.44 ± 3% -0.3 0.12 ± 6% perf-profile.children.cycles-pp.record__mmap_read_evlist 0.44 ± 3% -0.3 0.12 ± 6% perf-profile.children.cycles-pp.handle_internal_command 0.44 ± 3% -0.3 0.12 ± 6% perf-profile.children.cycles-pp.main 0.44 ± 3% -0.3 0.12 ± 6% perf-profile.children.cycles-pp.run_builtin 0.43 ± 3% -0.3 0.11 ± 6% perf-profile.children.cycles-pp.perf_mmap__push 0.34 ± 3% -0.3 0.07 ± 17% perf-profile.children.cycles-pp.record__pushfn 0.34 ± 4% -0.3 0.07 ± 17% perf-profile.children.cycles-pp.write 0.34 ± 4% -0.3 0.07 ± 17% perf-profile.children.cycles-pp.writen 0.32 ± 3% -0.3 0.07 ± 15% perf-profile.children.cycles-pp.ksys_write 0.32 ± 3% -0.2 0.07 ± 16% perf-profile.children.cycles-pp.vfs_write 0.30 ± 3% -0.2 0.07 ± 16% perf-profile.children.cycles-pp.shmem_file_write_iter 0.30 ± 3% -0.2 0.06 ± 19% perf-profile.children.cycles-pp.generic_perform_write 0.29 ± 4% -0.2 0.09 ± 4% perf-profile.children.cycles-pp.sched_tick 0.38 ± 33% -0.1 0.24 ± 8% perf-profile.children.cycles-pp.process_simple 0.20 ± 5% -0.1 0.06 ± 8% perf-profile.children.cycles-pp.task_tick_fair 0.38 ± 33% -0.1 0.24 ± 8% perf-profile.children.cycles-pp.ordered_events__queue 0.38 ± 33% -0.1 0.24 ± 8% perf-profile.children.cycles-pp.queue_event 0.00 +0.1 0.10 ± 22% perf-profile.children.cycles-pp.sync_regs 0.00 +0.1 0.10 ± 59% perf-profile.children.cycles-pp.__count_memcg_events 0.00 +0.1 0.10 ± 24% perf-profile.children.cycles-pp.___perf_sw_event 0.00 +0.1 0.10 ± 26% perf-profile.children.cycles-pp.perf_event_mmap_output 0.00 +0.1 0.11 ± 24% perf-profile.children.cycles-pp.native_flush_tlb_local 0.00 +0.1 0.11 ± 8% perf-profile.children.cycles-pp.__put_anon_vma 0.00 +0.1 0.11 ± 17% perf-profile.children.cycles-pp.rmqueue 0.00 +0.1 0.12 ± 26% perf-profile.children.cycles-pp.find_mergeable_anon_vma 0.00 +0.1 0.12 ± 23% perf-profile.children.cycles-pp.flush_tlb_func 0.00 +0.1 0.13 ± 28% perf-profile.children.cycles-pp.__call_rcu_common 0.00 +0.1 0.13 ± 24% perf-profile.children.cycles-pp.__perf_sw_event 0.00 +0.1 0.13 ± 38% perf-profile.children.cycles-pp.obj_cgroup_uncharge_pages 0.05 +0.1 0.18 ± 23% perf-profile.children.cycles-pp.flush_tlb_mm_range 0.00 +0.1 0.14 ± 27% perf-profile.children.cycles-pp.___slab_alloc 0.00 +0.1 0.14 ± 23% perf-profile.children.cycles-pp.mas_preallocate 0.00 +0.1 0.15 ± 18% perf-profile.children.cycles-pp.mas_empty_area_rev 0.00 +0.2 0.16 ± 22% perf-profile.children.cycles-pp.clear_bhb_loop 0.02 ±141% +0.2 0.17 ± 25% perf-profile.children.cycles-pp.perf_iterate_sb 0.00 +0.2 0.16 ± 26% perf-profile.children.cycles-pp.mas_alloc_nodes 0.07 ± 5% +0.2 0.24 ± 25% perf-profile.children.cycles-pp.vms_gather_munmap_vmas 0.00 +0.2 0.17 ± 20% perf-profile.children.cycles-pp.obj_cgroup_charge 0.00 +0.2 0.18 ± 25% perf-profile.children.cycles-pp.mas_walk 0.05 +0.2 0.24 ± 24% perf-profile.children.cycles-pp.mas_find 0.06 ± 6% +0.2 0.27 ± 11% perf-profile.children.cycles-pp.unlink_anon_vmas 0.05 ± 7% +0.2 0.26 ± 7% perf-profile.children.cycles-pp.free_unref_page 0.07 ± 11% +0.2 0.28 ± 21% perf-profile.children.cycles-pp.perf_event_mmap_event 0.07 ± 9% +0.2 0.31 ± 22% perf-profile.children.cycles-pp.perf_event_mmap 0.00 +0.2 0.24 ± 20% perf-profile.children.cycles-pp.vm_unmapped_area 0.02 ± 99% +0.2 0.27 ± 20% perf-profile.children.cycles-pp.arch_get_unmapped_area_topdown 0.06 ± 6% +0.2 0.31 ± 20% perf-profile.children.cycles-pp.__get_unmapped_area 0.00 +0.3 0.26 ± 29% perf-profile.children.cycles-pp.rcu_cblist_dequeue 0.00 +0.3 0.26 ± 6% perf-profile.children.cycles-pp.free_pcppages_bulk 0.00 +0.3 0.26 ± 23% perf-profile.children.cycles-pp.mas_store_prealloc 0.00 +0.3 0.27 ± 7% perf-profile.children.cycles-pp.__put_partials 0.00 +0.3 0.28 ± 7% perf-profile.children.cycles-pp.free_unref_page_commit 0.00 +0.3 0.28 ± 19% perf-profile.children.cycles-pp.vm_area_free_rcu_cb 0.14 ± 2% +0.3 0.48 ± 25% perf-profile.children.cycles-pp.mas_store_gfp 0.05 ± 7% +0.3 0.40 ± 4% perf-profile.children.cycles-pp.__memcg_slab_free_hook 0.10 +0.4 0.47 ± 24% perf-profile.children.cycles-pp.mas_wr_node_store 0.00 +0.5 0.46 ± 27% perf-profile.children.cycles-pp.__slab_free 0.07 +0.5 0.55 ± 10% perf-profile.children.cycles-pp.vm_area_alloc 0.06 +0.5 0.61 ± 12% perf-profile.children.cycles-pp.lru_add 0.00 +0.7 0.70 ± 26% perf-profile.children.cycles-pp.get_mem_cgroup_from_mm 0.00 +0.7 0.72 ± 24% perf-profile.children.cycles-pp.__mod_memcg_state 0.12 ± 6% +0.8 0.88 ± 10% perf-profile.children.cycles-pp.__folio_mod_stat 0.13 ± 5% +0.8 0.91 ± 10% perf-profile.children.cycles-pp.folio_add_new_anon_rmap 0.04 ± 71% +0.8 0.83 ± 13% perf-profile.children.cycles-pp.mod_objcg_state 0.00 +0.8 0.83 ± 35% perf-profile.children.cycles-pp.propagate_protected_usage 0.00 +0.8 0.84 ± 10% perf-profile.children.cycles-pp.folio_remove_rmap_ptes 0.00 +0.9 0.86 ± 9% perf-profile.children.cycles-pp.__anon_vma_prepare 0.00 +0.9 0.88 ± 8% perf-profile.children.cycles-pp.___pte_free_tlb 0.01 ±223% +0.9 0.90 ± 9% perf-profile.children.cycles-pp.__vmf_anon_prepare 0.00 +0.9 0.90 ± 9% perf-profile.children.cycles-pp.zap_present_ptes 0.06 +1.0 1.02 ± 10% perf-profile.children.cycles-pp.__memcg_slab_post_alloc_hook 0.05 ± 7% +1.0 1.02 ± 33% perf-profile.children.cycles-pp.native_irq_return_iret 0.00 +1.0 1.04 ± 9% perf-profile.children.cycles-pp.free_pud_range 0.00 +1.1 1.05 ± 9% perf-profile.children.cycles-pp.free_p4d_range 0.00 +1.1 1.07 ± 9% perf-profile.children.cycles-pp.free_pgd_range 0.00 +1.1 1.11 ± 75% perf-profile.children.cycles-pp.uncharge_folio 0.12 ± 4% +1.1 1.24 ± 11% perf-profile.children.cycles-pp.kmem_cache_free 0.10 ± 4% +1.2 1.30 ± 46% perf-profile.children.cycles-pp.__mem_cgroup_charge 0.11 ± 4% +1.2 1.34 ± 9% perf-profile.children.cycles-pp.kmem_cache_alloc_noprof 0.16 ± 3% +1.3 1.41 ± 16% perf-profile.children.cycles-pp.handle_softirqs 0.13 +1.3 1.41 ± 16% perf-profile.children.cycles-pp.rcu_core 0.12 ± 4% +1.3 1.41 ± 16% perf-profile.children.cycles-pp.rcu_do_batch 0.10 ± 3% +1.3 1.39 ± 9% perf-profile.children.cycles-pp.free_pgtables 0.06 +1.5 1.53 ± 28% perf-profile.children.cycles-pp.__memcg_kmem_charge_page 0.28 ± 4% +1.5 1.75 ± 18% perf-profile.children.cycles-pp.mmap_region 0.00 +1.5 1.54 ± 34% perf-profile.children.cycles-pp.alloc_anon_folio 0.38 ± 2% +1.6 1.96 ± 17% perf-profile.children.cycles-pp.__alloc_pages_noprof 0.39 ± 2% +1.6 2.00 ± 16% perf-profile.children.cycles-pp.alloc_pages_mpol_noprof 0.34 ± 2% +1.8 2.11 ± 19% perf-profile.children.cycles-pp.do_mmap 0.36 ± 2% +1.9 2.24 ± 19% perf-profile.children.cycles-pp.vm_mmap_pgoff 0.00 +2.0 2.02 ± 14% perf-profile.children.cycles-pp.zap_pte_range 0.37 ± 3% +2.1 2.48 ± 19% perf-profile.children.cycles-pp.__mmap 0.22 ± 2% +2.3 2.50 ± 16% perf-profile.children.cycles-pp.zap_pmd_range 0.23 ± 3% +2.3 2.55 ± 16% perf-profile.children.cycles-pp.unmap_page_range 0.24 ± 3% +2.3 2.58 ± 16% perf-profile.children.cycles-pp.unmap_vmas 0.16 +2.5 2.63 ± 12% perf-profile.children.cycles-pp.pte_alloc_one 0.08 ± 12% +2.6 2.71 ± 32% perf-profile.children.cycles-pp.__mod_memcg_lruvec_state 0.00 +2.7 2.69 ± 12% perf-profile.children.cycles-pp.__pte_alloc 0.05 +3.2 3.22 ± 38% perf-profile.children.cycles-pp.page_counter_cancel 0.22 ± 5% +3.2 3.40 ± 10% perf-profile.children.cycles-pp.__lruvec_stat_mod_folio 0.05 +4.0 4.03 ± 37% perf-profile.children.cycles-pp.page_counter_uncharge 0.06 +5.0 5.04 ± 23% perf-profile.children.cycles-pp.uncharge_batch 0.06 ± 7% +6.1 6.16 ± 7% perf-profile.children.cycles-pp.__mem_cgroup_uncharge_folios 0.00 +6.2 6.16 ± 12% perf-profile.children.cycles-pp.do_anonymous_page 0.10 ± 5% +37.8 37.88 ± 3% perf-profile.children.cycles-pp.__page_cache_release 0.12 ± 3% +38.3 38.46 perf-profile.children.cycles-pp.folio_batch_move_lru 0.00 +38.5 38.46 perf-profile.children.cycles-pp.lru_add_drain_cpu 0.00 +38.5 38.47 perf-profile.children.cycles-pp.lru_add_drain 0.79 ± 2% +44.2 44.99 perf-profile.children.cycles-pp.free_pages_and_swap_cache 0.79 ± 2% +44.2 45.01 perf-profile.children.cycles-pp.__tlb_batch_free_encoded_pages 0.78 ± 2% +44.2 45.01 perf-profile.children.cycles-pp.folios_put_refs 0.86 ± 2% +44.4 45.24 perf-profile.children.cycles-pp.tlb_finish_mmu 0.12 ± 6% +73.4 73.50 ± 2% perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath 0.21 ± 4% +73.6 73.77 ± 2% perf-profile.children.cycles-pp._raw_spin_lock_irqsave 0.08 +74.3 74.35 ± 2% perf-profile.children.cycles-pp.folio_lruvec_lock_irqsave 1.23 ± 2% +86.5 87.71 perf-profile.children.cycles-pp.vms_clear_ptes 1.26 +86.6 87.87 perf-profile.children.cycles-pp.vms_complete_munmap_vmas 1.44 +87.1 88.55 perf-profile.children.cycles-pp.do_vmi_align_munmap 1.48 +87.1 88.60 perf-profile.children.cycles-pp.do_vmi_munmap 1.50 +87.2 88.66 perf-profile.children.cycles-pp.__x64_sys_munmap 1.49 +87.2 88.66 perf-profile.children.cycles-pp.__vm_munmap 1.53 +87.4 88.93 perf-profile.children.cycles-pp.__munmap 2.30 +88.8 91.10 perf-profile.children.cycles-pp.do_syscall_64 2.30 +88.8 91.12 perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe 88.52 -88.3 0.20 ± 22% perf-profile.self.cycles-pp.clear_page_erms 3.36 -3.2 0.20 ± 23% perf-profile.self.cycles-pp.__cond_resched 0.49 -0.5 0.03 ± 70% perf-profile.self.cycles-pp.free_unref_folios 0.37 ± 33% -0.1 0.24 ± 8% perf-profile.self.cycles-pp.queue_event 0.00 +0.1 0.09 ± 22% perf-profile.self.cycles-pp.___perf_sw_event 0.00 +0.1 0.10 ± 64% perf-profile.self.cycles-pp.__count_memcg_events 0.00 +0.1 0.10 ± 22% perf-profile.self.cycles-pp.sync_regs 0.00 +0.1 0.10 ± 23% perf-profile.self.cycles-pp.perf_event_mmap_output 0.00 +0.1 0.10 ± 47% perf-profile.self.cycles-pp.obj_cgroup_uncharge_pages 0.00 +0.1 0.11 ± 24% perf-profile.self.cycles-pp.native_flush_tlb_local 0.00 +0.1 0.11 ± 21% perf-profile.self.cycles-pp.kmem_cache_free 0.00 +0.1 0.11 ± 25% perf-profile.self.cycles-pp.__memcg_slab_free_hook 0.00 +0.1 0.12 ± 22% perf-profile.self.cycles-pp.__page_cache_release 0.00 +0.1 0.13 ± 24% perf-profile.self.cycles-pp.mmap_region 0.00 +0.1 0.13 ± 26% perf-profile.self.cycles-pp.obj_cgroup_charge 0.00 +0.2 0.15 ± 22% perf-profile.self.cycles-pp.__memcg_slab_post_alloc_hook 0.00 +0.2 0.16 ± 22% perf-profile.self.cycles-pp.clear_bhb_loop 0.00 +0.2 0.16 ± 25% perf-profile.self.cycles-pp.free_pud_range 0.00 +0.2 0.17 ± 24% perf-profile.self.cycles-pp.kmem_cache_alloc_noprof 0.00 +0.2 0.17 ± 23% perf-profile.self.cycles-pp.lru_add 0.00 +0.2 0.17 ± 24% perf-profile.self.cycles-pp.mas_walk 0.00 +0.2 0.18 ± 22% perf-profile.self.cycles-pp.folio_batch_move_lru 0.09 +0.2 0.28 ± 29% perf-profile.self.cycles-pp._raw_spin_lock_irqsave 0.06 ± 9% +0.2 0.26 ± 25% perf-profile.self.cycles-pp.mas_wr_node_store 0.00 +0.2 0.23 ± 21% perf-profile.self.cycles-pp.zap_pmd_range 0.00 +0.3 0.25 ± 28% perf-profile.self.cycles-pp.rcu_cblist_dequeue 0.00 +0.3 0.28 ± 21% perf-profile.self.cycles-pp.folios_put_refs 0.00 +0.4 0.44 ± 26% perf-profile.self.cycles-pp.__slab_free 0.00 +0.5 0.46 ± 33% perf-profile.self.cycles-pp.mod_objcg_state 0.00 +0.5 0.48 ± 96% perf-profile.self.cycles-pp.__mem_cgroup_charge 0.00 +0.6 0.59 ± 32% perf-profile.self.cycles-pp.uncharge_batch 0.00 +0.7 0.69 ± 24% perf-profile.self.cycles-pp.__mod_memcg_state 0.00 +0.7 0.69 ± 27% perf-profile.self.cycles-pp.get_mem_cgroup_from_mm 0.00 +0.8 0.81 ± 35% perf-profile.self.cycles-pp.propagate_protected_usage 0.00 +0.8 0.82 ± 36% perf-profile.self.cycles-pp.folio_lruvec_lock_irqsave 0.05 ± 7% +1.0 1.02 ± 33% perf-profile.self.cycles-pp.native_irq_return_iret 0.00 +1.0 1.02 ± 22% perf-profile.self.cycles-pp.zap_pte_range 0.00 +1.1 1.10 ± 76% perf-profile.self.cycles-pp.uncharge_folio 0.00 +1.1 1.14 ± 37% perf-profile.self.cycles-pp.__memcg_kmem_charge_page 0.08 ± 4% +1.7 1.78 ± 35% perf-profile.self.cycles-pp.__lruvec_stat_mod_folio 0.07 ± 16% +2.6 2.62 ± 33% perf-profile.self.cycles-pp.__mod_memcg_lruvec_state 0.05 +3.1 3.16 ± 38% perf-profile.self.cycles-pp.page_counter_cancel 0.12 ± 7% +73.4 73.50 ± 2% perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath *************************************************************************************************** lkp-spr-r02: 224 threads 2 sockets Intel(R) Xeon(R) Platinum 8480CTDX (Sapphire Rapids) with 256G memory ========================================================================================= compiler/cpufreq_governor/kconfig/nr_threads/rootfs/tbox_group/test/testcase/testtime: gcc-12/performance/x86_64-rhel-8.3/100%/debian-12-x86_64-20240206.cgz/lkp-spr-r02/radixsort/stress-ng/60s commit: 15e8156713 ("mm: shrinker: avoid memleak in alloc_shrinker_info") d4148aeab4 ("mm, mmap: limit THP alignment of anonymous mappings to PMD-aligned sizes") 15e8156713cc3803 d4148aeab412432bf928f311eca ---------------- --------------------------- %stddev %change %stddev \ | \ 18.98 ± 5% +30.5% 24.78 ± 14% sched_debug.cpu.clock.stddev 0.05 ± 7% +886.3% 0.50 ± 86% perf-sched.sch_delay.avg.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity 1.14 ± 74% -55.5% 0.51 ± 37% perf-sched.sch_delay.avg.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown] 265008 ± 13% -37.2% 166403 ± 22% meminfo.Active 264976 ± 13% -37.2% 166371 ± 22% meminfo.Active(anon) 1315990 -63.8% 475807 ± 3% meminfo.AnonHugePages 164713 -9.2% 149574 stress-ng.radixsort.ops 2745 -9.2% 2492 stress-ng.radixsort.ops_per_sec 78925 +289.9% 307739 stress-ng.time.minor_page_faults 327.88 ± 3% -67.1% 107.89 ± 3% numa-vmstat.node0.nr_anon_transparent_hugepages 63711 ± 11% -40.7% 37791 ± 25% numa-vmstat.node1.nr_active_anon 317.83 ± 3% -60.5% 125.70 ± 8% numa-vmstat.node1.nr_anon_transparent_hugepages 293623 ± 18% +35.1% 396637 ± 14% numa-vmstat.node1.nr_inactive_anon 63710 ± 11% -40.7% 37791 ± 25% numa-vmstat.node1.nr_zone_active_anon 293622 ± 18% +35.1% 396635 ± 14% numa-vmstat.node1.nr_zone_inactive_anon 668445 ± 3% -67.1% 220222 ± 3% numa-meminfo.node0.AnonHugePages 1317034 ± 13% -23.2% 1010970 ± 16% numa-meminfo.node0.AnonPages.max 254241 ± 11% -40.5% 151233 ± 26% numa-meminfo.node1.Active 254219 ± 11% -40.5% 151227 ± 26% numa-meminfo.node1.Active(anon) 647491 ± 3% -60.4% 256719 ± 8% numa-meminfo.node1.AnonHugePages 1170941 ± 18% +35.3% 1583801 ± 14% numa-meminfo.node1.Inactive 1170783 ± 18% +35.3% 1583760 ± 14% numa-meminfo.node1.Inactive(anon) 66923 ± 13% -36.6% 42442 ± 22% proc-vmstat.nr_active_anon 642.60 -63.8% 232.40 ± 3% proc-vmstat.nr_anon_transparent_hugepages 597472 ± 2% +5.6% 630776 proc-vmstat.nr_inactive_anon 66923 ± 13% -36.6% 42442 ± 22% proc-vmstat.nr_zone_active_anon 597472 ± 2% +5.6% 630776 proc-vmstat.nr_zone_inactive_anon 1188533 +19.2% 1416356 proc-vmstat.numa_hit 2502 -63.5% 913.67 proc-vmstat.numa_huge_pte_updates 956729 +23.2% 1178316 proc-vmstat.numa_local 1495630 ± 2% -54.0% 687360 ± 6% proc-vmstat.numa_pte_updates 820025 +26.8% 1040066 proc-vmstat.pgfault 6.20 +9.6% 6.79 perf-stat.i.MPKI 4.203e+10 -6.8% 3.916e+10 perf-stat.i.branch-instructions 6.15 +0.2 6.30 perf-stat.i.branch-miss-rate% 2.636e+09 -6.2% 2.474e+09 perf-stat.i.branch-misses 2.56 ± 2% +8.5% 2.77 perf-stat.i.cpi 2.444e+11 -6.9% 2.276e+11 perf-stat.i.instructions 0.40 -9.1% 0.36 perf-stat.i.ipc 2.20 ± 2% +90.0% 4.18 ± 2% perf-stat.i.metric.K/sec 10808 +45.4% 15713 ± 2% perf-stat.i.minor-faults 10809 +45.4% 15713 ± 2% perf-stat.i.page-faults 6.34 +7.9% 6.84 perf-stat.overall.MPKI 37.76 -0.6 37.12 perf-stat.overall.cache-miss-rate% 2.54 +9.9% 2.80 perf-stat.overall.cpi 401.20 +1.8% 408.57 perf-stat.overall.cycles-between-cache-misses 0.39 -9.0% 0.36 perf-stat.overall.ipc 4.134e+10 ± 2% -7.4% 3.829e+10 perf-stat.ps.branch-instructions 2.592e+09 -6.7% 2.419e+09 perf-stat.ps.branch-misses 2.404e+11 ± 2% -7.4% 2.226e+11 perf-stat.ps.instructions 10358 +36.2% 14109 ± 2% perf-stat.ps.minor-faults 10358 +36.2% 14109 ± 2% perf-stat.ps.page-faults 1.525e+13 -9.9% 1.374e+13 perf-stat.total.instructions 0.60 ± 10% -0.3 0.28 ±100% perf-profile.calltrace.cycles-pp.__perf_mmap__read_init.perf_mmap__read_init.perf_mmap__push.record__mmap_read_evlist.__cmd_record 0.71 ± 4% +0.1 0.77 ± 3% perf-profile.calltrace.cycles-pp.update_load_avg.task_tick_fair.sched_tick.update_process_times.tick_nohz_handler 1.25 ± 9% +0.3 1.54 ± 4% perf-profile.calltrace.cycles-pp.update_curr.task_tick_fair.sched_tick.update_process_times.tick_nohz_handler 0.26 ±100% +0.3 0.57 ± 3% perf-profile.calltrace.cycles-pp.__irq_exit_rcu.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt 2.64 ± 6% +0.4 3.09 ± 3% perf-profile.calltrace.cycles-pp.task_tick_fair.sched_tick.update_process_times.tick_nohz_handler.__hrtimer_run_queues 0.08 ±223% +0.5 0.55 ± 4% perf-profile.calltrace.cycles-pp.account_user_time.update_process_times.tick_nohz_handler.__hrtimer_run_queues.hrtimer_interrupt 5.00 ± 6% +0.7 5.75 ± 3% perf-profile.calltrace.cycles-pp.sched_tick.update_process_times.tick_nohz_handler.__hrtimer_run_queues.hrtimer_interrupt 7.66 ± 5% +1.1 8.72 ± 3% perf-profile.calltrace.cycles-pp.update_process_times.tick_nohz_handler.__hrtimer_run_queues.hrtimer_interrupt.__sysvec_apic_timer_interrupt 8.10 ± 5% +1.1 9.18 ± 3% perf-profile.calltrace.cycles-pp.tick_nohz_handler.__hrtimer_run_queues.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt 8.38 ± 5% +1.1 9.51 ± 3% perf-profile.calltrace.cycles-pp.__hrtimer_run_queues.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt 10.78 ± 5% +1.3 12.08 ± 3% perf-profile.calltrace.cycles-pp.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt 11.04 ± 5% +1.3 12.36 ± 3% perf-profile.calltrace.cycles-pp.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt 11.77 ± 5% +1.4 13.19 ± 2% perf-profile.calltrace.cycles-pp.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt 12.68 ± 5% +1.6 14.23 ± 2% perf-profile.calltrace.cycles-pp.asm_sysvec_apic_timer_interrupt 12.84 -0.4 12.47 ± 2% perf-profile.children.cycles-pp.strcmp@plt 0.65 ± 10% -0.2 0.50 ± 30% perf-profile.children.cycles-pp.__perf_mmap__read_init 0.12 ± 3% +0.0 0.14 ± 4% perf-profile.children.cycles-pp.sched_clock 0.05 ± 8% +0.0 0.08 ± 8% perf-profile.children.cycles-pp.rb_next 0.09 ± 7% +0.0 0.13 ± 10% perf-profile.children.cycles-pp.timerqueue_del 0.20 ± 4% +0.0 0.23 ± 4% perf-profile.children.cycles-pp.lapic_next_deadline 0.40 ± 7% +0.0 0.44 ± 5% perf-profile.children.cycles-pp.handle_softirqs 0.42 ± 4% +0.0 0.47 ± 5% perf-profile.children.cycles-pp.native_irq_return_iret 0.20 ± 9% +0.1 0.26 ± 6% perf-profile.children.cycles-pp.__cgroup_account_cputime_field 0.12 ± 4% +0.1 0.19 ± 10% perf-profile.children.cycles-pp._raw_spin_lock 0.75 ± 4% +0.1 0.82 ± 3% perf-profile.children.cycles-pp.update_load_avg 0.40 ± 4% +0.1 0.47 ± 2% perf-profile.children.cycles-pp.hrtimer_active 0.34 ± 9% +0.1 0.41 ± 5% perf-profile.children.cycles-pp.irqtime_account_irq 0.52 ± 5% +0.1 0.60 ± 3% perf-profile.children.cycles-pp.__irq_exit_rcu 0.12 ± 9% +0.1 0.20 ± 6% perf-profile.children.cycles-pp.perf_trace_sched_stat_runtime 0.36 ± 2% +0.1 0.44 ± 3% perf-profile.children.cycles-pp.task_mm_cid_work 0.50 ± 6% +0.1 0.58 ± 4% perf-profile.children.cycles-pp.account_user_time 0.38 ± 2% +0.1 0.48 ± 3% perf-profile.children.cycles-pp.task_work_run 0.46 ± 2% +0.1 0.56 ± 3% perf-profile.children.cycles-pp.irqentry_exit_to_user_mode 1.32 ± 9% +0.3 1.64 ± 4% perf-profile.children.cycles-pp.update_curr 2.76 ± 6% +0.5 3.25 ± 3% perf-profile.children.cycles-pp.task_tick_fair 5.24 ± 6% +0.8 6.04 ± 3% perf-profile.children.cycles-pp.sched_tick 7.99 ± 5% +1.2 9.14 ± 3% perf-profile.children.cycles-pp.update_process_times 8.45 ± 5% +1.2 9.63 ± 3% perf-profile.children.cycles-pp.tick_nohz_handler 8.75 ± 5% +1.2 9.99 ± 3% perf-profile.children.cycles-pp.__hrtimer_run_queues 11.23 ± 5% +1.4 12.66 ± 3% perf-profile.children.cycles-pp.hrtimer_interrupt 11.50 ± 5% +1.4 12.94 ± 3% perf-profile.children.cycles-pp.__sysvec_apic_timer_interrupt 12.26 ± 5% +1.6 13.82 ± 3% perf-profile.children.cycles-pp.sysvec_apic_timer_interrupt 13.22 ± 5% +1.7 14.92 ± 2% perf-profile.children.cycles-pp.asm_sysvec_apic_timer_interrupt 6.12 -0.2 5.92 ± 2% perf-profile.self.cycles-pp.strcmp@plt 0.09 ± 11% +0.0 0.11 ± 6% perf-profile.self.cycles-pp.__hrtimer_run_queues 0.10 ± 9% +0.0 0.12 ± 5% perf-profile.self.cycles-pp.task_tick_fair 0.05 ± 8% +0.0 0.08 ± 10% perf-profile.self.cycles-pp.rb_next 0.20 ± 4% +0.0 0.23 ± 3% perf-profile.self.cycles-pp.lapic_next_deadline 0.42 ± 4% +0.0 0.47 ± 5% perf-profile.self.cycles-pp.native_irq_return_iret 0.16 ± 14% +0.1 0.21 ± 4% perf-profile.self.cycles-pp.__cgroup_account_cputime_field 0.12 ± 4% +0.1 0.18 ± 10% perf-profile.self.cycles-pp._raw_spin_lock 0.24 ± 9% +0.1 0.30 ± 7% perf-profile.self.cycles-pp.irqtime_account_irq 0.40 ± 4% +0.1 0.46 ± 3% perf-profile.self.cycles-pp.hrtimer_active 0.10 ± 3% +0.1 0.18 ± 8% perf-profile.self.cycles-pp.hrtimer_interrupt 0.11 ± 11% +0.1 0.20 ± 6% perf-profile.self.cycles-pp.perf_trace_sched_stat_runtime 0.33 ± 3% +0.1 0.42 ± 4% perf-profile.self.cycles-pp.task_mm_cid_work 0.88 ± 11% +0.2 1.05 ± 5% perf-profile.self.cycles-pp.update_curr Disclaimer: Results have been estimated based on internal Intel analysis and are provided for informational purposes only. Any difference in system hardware or software design or configuration may affect actual performance. -- 0-DAY CI Kernel Test Service https://github.com/intel/lkp-tests/wiki