Hello, kernel test robot noticed a -43.6% regression of vm-scalability.throughput on: commit: 09f84589a8ed1162c0274a1b5b3dadb98ee3155e ("[PATCH v2 3/3] ACPI/NUMA: Remove architecture dependent remainings") url: https://github.com/intel-lab-lkp/linux/commits/Robert-Richter/x86-numa-Fix-SRAT-lookup-of-CFMWS-ranges-with-numa_fill_memblks/20240319-200345 base: https://git.kernel.org/cgit/linux/kernel/git/rafael/linux-pm.git linux-next patch link: https://lore.kernel.org/all/20240319120026.2246389-4-rrichter@xxxxxxx/ patch subject: [PATCH v2 3/3] ACPI/NUMA: Remove architecture dependent remainings testcase: vm-scalability test machine: 128 threads 2 sockets Intel(R) Xeon(R) Gold 6338 CPU @ 2.00GHz (Ice Lake) with 256G memory parameters: runtime: 300s size: 512G test: anon-cow-rand-hugetlb cpufreq_governor: performance In addition to that, the commit also has significant impact on the following tests: +------------------+-----------------------------------------------------------------------------------------------+ | testcase: change | phoronix-test-suite: phoronix-test-suite.npb.SP.B.total_mop_s -58.1% regression | | test machine | 96 threads 2 sockets Intel(R) Xeon(R) Gold 6252 CPU @ 2.10GHz (Cascade Lake) with 512G memory | | test parameters | cpufreq_governor=performance | | | option_a=SP.B [Scalar Penta-diagonal solver] | | | test=npb-1.4.5 | +------------------+-----------------------------------------------------------------------------------------------+ | testcase: change | will-it-scale: will-it-scale.per_process_ops -38.1% regression | | test machine | 104 threads 2 sockets (Skylake) with 192G memory | | test parameters | cpufreq_governor=performance | | | mode=process | | | nr_task=100% | | | test=mmap1 | +------------------+-----------------------------------------------------------------------------------------------+ | testcase: change | phoronix-test-suite: phoronix-test-suite.asmfish.0.nodes_second -23.9% regression | | test machine | 96 threads 2 sockets Intel(R) Xeon(R) Gold 6252 CPU @ 2.10GHz (Cascade Lake) with 512G memory | | test parameters | cpufreq_governor=performance | | | test=asmfish-1.1.2 | +------------------+-----------------------------------------------------------------------------------------------+ If you fix the issue in a separate patch/commit (i.e. not just a new version of the same patch/commit), kindly add following tags | Reported-by: kernel test robot <oliver.sang@xxxxxxxxx> | Closes: https://lore.kernel.org/oe-lkp/202403220943.96dde419-oliver.sang@xxxxxxxxx Details are as below: --------------------------------------------------------------------------------------------------> The kernel config and materials to reproduce are available at: https://download.01.org/0day-ci/archive/20240322/202403220943.96dde419-oliver.sang@xxxxxxxxx ========================================================================================= compiler/cpufreq_governor/kconfig/rootfs/runtime/size/tbox_group/test/testcase: gcc-12/performance/x86_64-rhel-8.3/debian-12-x86_64-20240206.cgz/300s/512G/lkp-icl-2sp2/anon-cow-rand-hugetlb/vm-scalability commit: 9ef3ad0f03 ("ACPI/NUMA: Print CXL Early Discovery Table (CEDT)") 09f84589a8 ("ACPI/NUMA: Remove architecture dependent remainings") 9ef3ad0f032818c5 09f84589a8ed1162c0274a1b5b3 ---------------- --------------------------- %stddev %change %stddev \ | \ 4.277e+09 ± 6% -70.5% 1.261e+09 ± 6% cpuidle..time 735569 ± 2% -34.8% 479905 ± 3% cpuidle..usage 243.64 ± 2% -24.9% 182.92 uptime.boot 9961 ± 4% -29.0% 7073 ± 4% uptime.idle 16.94 ± 3% -9.9 7.06 ± 8% mpstat.cpu.all.idle% 0.35 -0.1 0.29 mpstat.cpu.all.irq% 0.03 ± 3% -0.0 0.02 mpstat.cpu.all.soft% 12.05 -2.1 10.00 mpstat.cpu.all.sys% 70.63 +12.0 82.63 mpstat.cpu.all.usr% 17.78 ± 3% -52.2% 8.49 ± 6% vmstat.cpu.id 70.01 +16.5% 81.54 vmstat.cpu.us 107.54 +10.6% 118.93 vmstat.procs.r 2652 ± 2% -8.9% 2417 ± 5% vmstat.system.cs 141617 ± 3% -13.4% 122650 vmstat.system.in 573464 ± 8% -100.0% 0.00 numa-numastat.node0.local_node 632802 ± 4% -100.0% 0.00 numa-numastat.node0.numa_hit 59338 ± 51% -100.0% 0.00 numa-numastat.node0.other_node 677948 ± 7% -100.0% 0.00 numa-numastat.node1.local_node 751205 ± 4% -100.0% 0.00 numa-numastat.node1.numa_hit 73258 ± 41% -100.0% 0.00 numa-numastat.node1.other_node 95544 ± 27% -59.6% 38591 ± 5% meminfo.Active 95399 ± 27% -59.7% 38433 ± 5% meminfo.Active(anon) 10487673 ± 7% -28.6% 7487857 ± 6% meminfo.DirectMap2M 43001 +14.2% 49088 meminfo.HugePages_Surp 43001 +14.2% 49088 meminfo.HugePages_Total 88066074 +14.2% 1.005e+08 meminfo.Hugetlb 1.356e+08 -9.5% 1.228e+08 meminfo.MemAvailable 1.366e+08 -9.1% 1.241e+08 meminfo.MemFree 94133967 +13.0% 1.064e+08 meminfo.Memused 241580 ± 8% -26.3% 178118 meminfo.Shmem 267728 -81.4% 49921 meminfo.VmallocUsed 51764 ± 2% -37.7% 32258 vm-scalability.median 6.46 ± 19% -5.3 1.21 ± 20% vm-scalability.median_stddev% 2.86 ± 18% -2.7 0.11 ± 31% vm-scalability.stddev% 7315147 -43.6% 4122657 vm-scalability.throughput 193.69 ± 2% -32.1% 131.50 vm-scalability.time.elapsed_time 193.69 ± 2% -32.1% 131.50 vm-scalability.time.elapsed_time.max 138604 ± 4% -21.6% 108601 ± 7% vm-scalability.time.involuntary_context_switches 578101 -55.9% 254781 vm-scalability.time.minor_page_faults 10610 +12.8% 11965 vm-scalability.time.percent_of_cpu_this_job_got 2981 ± 3% -43.3% 1691 vm-scalability.time.system_time 17565 -20.0% 14044 vm-scalability.time.user_time 15370 ± 4% -51.6% 7439 ± 5% vm-scalability.time.voluntary_context_switches 1.159e+09 -55.6% 5.146e+08 vm-scalability.workload 495078 ± 11% -71.4% 141719 ± 6% sched_debug.cfs_rq:/.avg_vruntime.stddev 0.28 ± 21% -33.3% 0.19 ± 21% sched_debug.cfs_rq:/.h_nr_running.stddev 2570 ± 80% +109.3% 5381 sched_debug.cfs_rq:/.load.min 2.76 ± 52% +68.8% 4.67 sched_debug.cfs_rq:/.load_avg.min 495078 ± 11% -71.4% 141719 ± 6% sched_debug.cfs_rq:/.min_vruntime.stddev 0.23 ± 31% -53.6% 0.11 ± 15% sched_debug.cfs_rq:/.nr_running.stddev 242.06 ± 32% -50.0% 121.10 ± 9% sched_debug.cfs_rq:/.runnable_avg.stddev 235.93 ± 81% +137.9% 561.33 ± 19% sched_debug.cfs_rq:/.util_avg.min 226.75 ± 31% -52.3% 108.21 ± 13% sched_debug.cfs_rq:/.util_avg.stddev 192.74 ± 41% -73.6% 50.87 ± 50% sched_debug.cfs_rq:/.util_est.stddev 370662 ± 45% +53.6% 569380 ± 2% sched_debug.cpu.avg_idle.min 4569 ± 11% -23.5% 3494 sched_debug.cpu.curr->pid.avg 6312 ± 6% -26.8% 4619 sched_debug.cpu.curr->pid.max 37956 ± 30% -50.9% 18650 ± 70% sched_debug.cpu.max_idle_balance_cost.stddev 2949 ± 7% -22.4% 2287 ± 3% sched_debug.cpu.nr_switches.avg 821.53 ± 12% -31.9% 559.11 ± 6% sched_debug.cpu.nr_switches.min 0.25 ±101% -99.8% 0.00 ± 92% sched_debug.rt_rq:.rt_time.avg 32.43 ±101% -99.8% 0.06 ± 92% sched_debug.rt_rq:.rt_time.max 2.85 ±101% -99.8% 0.01 ± 92% sched_debug.rt_rq:.rt_time.stddev 507357 -55.6% 225492 proc-vmstat.htlb_buddy_alloc_success 23772 ± 27% -59.2% 9703 ± 5% proc-vmstat.nr_active_anon 3376161 -9.1% 3070574 proc-vmstat.nr_dirty_background_threshold 6760578 -9.1% 6148657 proc-vmstat.nr_dirty_threshold 823112 -1.9% 807289 proc-vmstat.nr_file_pages 34080722 -8.7% 31109176 proc-vmstat.nr_free_pages 25608 +3.2% 26432 proc-vmstat.nr_kernel_stack 2710 +3.5% 2804 proc-vmstat.nr_page_table_pages 60299 ± 8% -26.2% 44490 proc-vmstat.nr_shmem 28254 -1.1% 27945 proc-vmstat.nr_slab_reclaimable 66139 -1.8% 64958 proc-vmstat.nr_slab_unreclaimable 23772 ± 27% -59.2% 9703 ± 5% proc-vmstat.nr_zone_active_anon 33857 ± 16% -100.0% 0.00 proc-vmstat.numa_hint_faults 12661 ± 72% -100.0% 0.00 proc-vmstat.numa_hint_faults_local 1386858 -43.2% 787481 proc-vmstat.numa_hit 1254263 -37.2% 787477 proc-vmstat.numa_local 132597 -100.0% 0.00 proc-vmstat.numa_other 41301 ± 42% -100.0% 0.00 proc-vmstat.numa_pages_migrated 183630 ± 19% -100.0% 0.00 proc-vmstat.numa_pte_updates 2.608e+08 -55.5% 1.161e+08 proc-vmstat.pgalloc_normal 1247964 -55.1% 559863 proc-vmstat.pgfault 2.597e+08 -55.4% 1.159e+08 proc-vmstat.pgfree 41301 ± 42% -100.0% 0.00 proc-vmstat.pgmigrate_success 35726 ± 6% -55.9% 15762 ± 3% proc-vmstat.pgreuse 15.99 -14.2% 13.72 perf-stat.i.MPKI 1.556e+10 ± 3% -34.0% 1.027e+10 perf-stat.i.branch-instructions 0.10 ± 4% +0.1 0.18 ± 4% perf-stat.i.branch-miss-rate% 11581986 ± 3% +29.1% 14946939 perf-stat.i.branch-misses 8.579e+08 ± 2% -36.0% 5.495e+08 perf-stat.i.cache-misses 8.808e+08 ± 2% -35.7% 5.662e+08 perf-stat.i.cache-references 2591 ± 2% -12.0% 2281 ± 4% perf-stat.i.context-switches 4.97 +47.0% 7.31 perf-stat.i.cpi 2.814e+11 +11.0% 3.122e+11 perf-stat.i.cpu-cycles 218.48 -11.0% 194.53 perf-stat.i.cpu-migrations 337.74 +73.1% 584.61 perf-stat.i.cycles-between-cache-misses 6.664e+10 ± 3% -33.9% 4.408e+10 perf-stat.i.instructions 0.25 -36.7% 0.16 perf-stat.i.ipc 2.00 ± 12% -47.4% 1.05 ± 16% perf-stat.i.major-faults 6085 -37.6% 3798 perf-stat.i.minor-faults 6087 -37.6% 3799 perf-stat.i.page-faults 12.96 -3.2% 12.55 perf-stat.overall.MPKI 0.07 +0.1 0.15 perf-stat.overall.branch-miss-rate% 4.25 +67.1% 7.10 perf-stat.overall.cpi 327.67 +72.6% 565.42 perf-stat.overall.cycles-between-cache-misses 0.24 -40.2% 0.14 perf-stat.overall.ipc 10931 +3.1% 11273 perf-stat.overall.path-length 1.518e+10 ± 2% -33.1% 1.015e+10 perf-stat.ps.branch-instructions 10985194 ± 2% +35.0% 14830464 perf-stat.ps.branch-misses 8.425e+08 ± 2% -35.1% 5.466e+08 perf-stat.ps.cache-misses 8.655e+08 ± 2% -34.9% 5.635e+08 perf-stat.ps.cache-references 2574 -11.3% 2282 ± 4% perf-stat.ps.context-switches 2.759e+11 +12.0% 3.09e+11 perf-stat.ps.cpu-cycles 215.11 ± 2% -9.9% 193.81 perf-stat.ps.cpu-migrations 6.501e+10 ± 2% -33.0% 4.355e+10 perf-stat.ps.instructions 2.04 ± 11% -47.9% 1.06 ± 16% perf-stat.ps.major-faults 6066 -36.8% 3832 perf-stat.ps.minor-faults 6068 -36.8% 3833 perf-stat.ps.page-faults 1.267e+13 -54.2% 5.802e+12 perf-stat.total.instructions 792.04 ± 23% -100.0% 0.00 numa-vmstat.node0.nr_active_anon 48328 ± 63% -100.0% 0.00 numa-vmstat.node0.nr_anon_pages 229536 ± 90% -100.0% 0.00 numa-vmstat.node0.nr_file_pages 23714921 ± 2% -100.0% 0.00 numa-vmstat.node0.nr_free_pages 53088 ± 73% -100.0% 0.00 numa-vmstat.node0.nr_inactive_anon 12744 ± 7% -100.0% 0.00 numa-vmstat.node0.nr_kernel_stack 5388 ± 45% -100.0% 0.00 numa-vmstat.node0.nr_mapped 890.81 ± 13% -100.0% 0.00 numa-vmstat.node0.nr_page_table_pages 6425 ±158% -100.0% 0.00 numa-vmstat.node0.nr_shmem 10347 ± 33% -100.0% 0.00 numa-vmstat.node0.nr_slab_reclaimable 32990 ± 10% -100.0% 0.00 numa-vmstat.node0.nr_slab_unreclaimable 223791 ± 90% -100.0% 0.00 numa-vmstat.node0.nr_unevictable 792.05 ± 23% -100.0% 0.00 numa-vmstat.node0.nr_zone_active_anon 53088 ± 73% -100.0% 0.00 numa-vmstat.node0.nr_zone_inactive_anon 223791 ± 90% -100.0% 0.00 numa-vmstat.node0.nr_zone_unevictable 633548 ± 4% -100.0% 0.00 numa-vmstat.node0.numa_hit 574210 ± 8% -100.0% 0.00 numa-vmstat.node0.numa_local 59338 ± 51% -100.0% 0.00 numa-vmstat.node0.numa_other 23125 ± 27% -100.0% 0.00 numa-vmstat.node1.nr_active_anon 115486 ± 26% -100.0% 0.00 numa-vmstat.node1.nr_anon_pages 593673 ± 34% -100.0% 0.00 numa-vmstat.node1.nr_file_pages 10376888 ± 3% -100.0% 0.00 numa-vmstat.node1.nr_free_pages 146240 ± 26% -100.0% 0.00 numa-vmstat.node1.nr_inactive_anon 12863 ± 6% -100.0% 0.00 numa-vmstat.node1.nr_kernel_stack 13956 ± 5% -100.0% 0.00 numa-vmstat.node1.nr_mapped 1818 ± 6% -100.0% 0.00 numa-vmstat.node1.nr_page_table_pages 53973 ± 15% -100.0% 0.00 numa-vmstat.node1.nr_shmem 17907 ± 19% -100.0% 0.00 numa-vmstat.node1.nr_slab_reclaimable 33150 ± 10% -100.0% 0.00 numa-vmstat.node1.nr_slab_unreclaimable 539462 ± 37% -100.0% 0.00 numa-vmstat.node1.nr_unevictable 23125 ± 27% -100.0% 0.00 numa-vmstat.node1.nr_zone_active_anon 146241 ± 26% -100.0% 0.00 numa-vmstat.node1.nr_zone_inactive_anon 539462 ± 37% -100.0% 0.00 numa-vmstat.node1.nr_zone_unevictable 750435 ± 4% -100.0% 0.00 numa-vmstat.node1.numa_hit 677178 ± 7% -100.0% 0.00 numa-vmstat.node1.numa_local 73258 ± 41% -100.0% 0.00 numa-vmstat.node1.numa_other 3216 ± 24% -100.0% 0.00 numa-meminfo.node0.Active 3165 ± 23% -100.0% 0.00 numa-meminfo.node0.Active(anon) 193141 ± 63% -100.0% 0.00 numa-meminfo.node0.AnonPages 245306 ± 55% -100.0% 0.00 numa-meminfo.node0.AnonPages.max 918145 ± 90% -100.0% 0.00 numa-meminfo.node0.FilePages 16988 ± 6% -100.0% 0.00 numa-meminfo.node0.HugePages_Surp 16988 ± 6% -100.0% 0.00 numa-meminfo.node0.HugePages_Total 212488 ± 73% -100.0% 0.00 numa-meminfo.node0.Inactive 212188 ± 73% -100.0% 0.00 numa-meminfo.node0.Inactive(anon) 41389 ± 33% -100.0% 0.00 numa-meminfo.node0.KReclaimable 12743 ± 7% -100.0% 0.00 numa-meminfo.node0.KernelStack 21242 ± 45% -100.0% 0.00 numa-meminfo.node0.Mapped 94746960 ± 2% -100.0% 0.00 numa-meminfo.node0.MemFree 1.317e+08 -100.0% 0.00 numa-meminfo.node0.MemTotal 36938247 ± 5% -100.0% 0.00 numa-meminfo.node0.MemUsed 3559 ± 13% -100.0% 0.00 numa-meminfo.node0.PageTables 41389 ± 33% -100.0% 0.00 numa-meminfo.node0.SReclaimable 131966 ± 10% -100.0% 0.00 numa-meminfo.node0.SUnreclaim 25702 ±158% -100.0% 0.00 numa-meminfo.node0.Shmem 173356 ± 15% -100.0% 0.00 numa-meminfo.node0.Slab 895165 ± 90% -100.0% 0.00 numa-meminfo.node0.Unevictable 92044 ± 27% -100.0% 0.00 numa-meminfo.node1.Active 91950 ± 27% -100.0% 0.00 numa-meminfo.node1.Active(anon) 94963 ± 57% -100.0% 0.00 numa-meminfo.node1.AnonHugePages 462070 ± 26% -100.0% 0.00 numa-meminfo.node1.AnonPages 521211 ± 21% -100.0% 0.00 numa-meminfo.node1.AnonPages.max 2374512 ± 34% -100.0% 0.00 numa-meminfo.node1.FilePages 26160 ± 3% -100.0% 0.00 numa-meminfo.node1.HugePages_Surp 26160 ± 3% -100.0% 0.00 numa-meminfo.node1.HugePages_Total 586305 ± 26% -100.0% 0.00 numa-meminfo.node1.Inactive 585444 ± 26% -100.0% 0.00 numa-meminfo.node1.Inactive(anon) 71636 ± 19% -100.0% 0.00 numa-meminfo.node1.KReclaimable 12870 ± 6% -100.0% 0.00 numa-meminfo.node1.KernelStack 55637 ± 5% -100.0% 0.00 numa-meminfo.node1.Mapped 41532114 ± 3% -100.0% 0.00 numa-meminfo.node1.MemFree 99030534 -100.0% 0.00 numa-meminfo.node1.MemTotal 57498419 ± 2% -100.0% 0.00 numa-meminfo.node1.MemUsed 7297 ± 6% -100.0% 0.00 numa-meminfo.node1.PageTables 71636 ± 19% -100.0% 0.00 numa-meminfo.node1.SReclaimable 132602 ± 10% -100.0% 0.00 numa-meminfo.node1.SUnreclaim 215707 ± 15% -100.0% 0.00 numa-meminfo.node1.Shmem 204238 ± 12% -100.0% 0.00 numa-meminfo.node1.Slab 2157849 ± 37% -100.0% 0.00 numa-meminfo.node1.Unevictable 56.43 ± 18% -56.4 0.00 perf-profile.calltrace.cycles-pp.do_rw_once 40.90 ± 15% -40.9 0.00 perf-profile.calltrace.cycles-pp.lrand48_r@plt 19.35 ± 72% -19.4 0.00 perf-profile.calltrace.cycles-pp.do_access 17.15 ± 84% -17.2 0.00 perf-profile.calltrace.cycles-pp.asm_exc_page_fault.do_access 17.12 ± 84% -17.1 0.00 perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault.do_access 17.12 ± 84% -17.1 0.00 perf-profile.calltrace.cycles-pp.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.do_access 17.10 ± 84% -17.1 0.00 perf-profile.calltrace.cycles-pp.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.do_access 17.09 ± 84% -17.1 0.00 perf-profile.calltrace.cycles-pp.hugetlb_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault 13.57 ± 84% -13.6 0.00 perf-profile.calltrace.cycles-pp.hugetlb_wp.hugetlb_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault 13.30 ± 84% -13.3 0.00 perf-profile.calltrace.cycles-pp.copy_user_large_folio.hugetlb_wp.hugetlb_fault.handle_mm_fault.do_user_addr_fault 13.28 ± 84% -13.3 0.00 perf-profile.calltrace.cycles-pp.copy_subpage.copy_user_large_folio.hugetlb_wp.hugetlb_fault.handle_mm_fault 13.16 ± 84% -13.2 0.00 perf-profile.calltrace.cycles-pp.copy_mc_enhanced_fast_string.copy_subpage.copy_user_large_folio.hugetlb_wp.hugetlb_fault 0.00 +2.1 2.06 ± 9% perf-profile.calltrace.cycles-pp.__mmput.exit_mm.do_exit.do_group_exit.get_signal 0.00 +2.1 2.06 ± 9% perf-profile.calltrace.cycles-pp.exit_mm.do_exit.do_group_exit.get_signal.arch_do_signal_or_restart 0.00 +2.1 2.06 ± 9% perf-profile.calltrace.cycles-pp.put_files_struct.do_exit.do_group_exit.get_signal.arch_do_signal_or_restart 0.00 +3.3 3.35 ± 39% perf-profile.calltrace.cycles-pp.common_startup_64 0.00 +3.8 3.83 ± 43% perf-profile.calltrace.cycles-pp.__x64_sys_openat.do_syscall_64.entry_SYSCALL_64_after_hwframe 0.00 +3.8 3.83 ± 43% perf-profile.calltrace.cycles-pp.do_sys_openat2.__x64_sys_openat.do_syscall_64.entry_SYSCALL_64_after_hwframe 0.00 +4.3 4.32 ± 43% perf-profile.calltrace.cycles-pp.sw_perf_event_destroy._free_event.perf_event_release_kernel.perf_release.__fput 0.00 +4.4 4.37 ± 49% perf-profile.calltrace.cycles-pp.__x64_sys_exit_group.do_syscall_64.entry_SYSCALL_64_after_hwframe 0.00 +4.4 4.37 ± 49% perf-profile.calltrace.cycles-pp.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64.entry_SYSCALL_64_after_hwframe 0.00 +4.4 4.37 ± 49% perf-profile.calltrace.cycles-pp.do_group_exit.__x64_sys_exit_group.do_syscall_64.entry_SYSCALL_64_after_hwframe 0.00 +4.6 4.63 ± 41% perf-profile.calltrace.cycles-pp._free_event.perf_event_release_kernel.perf_release.__fput.task_work_run 0.00 +5.2 5.19 ± 25% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.write.writen.record__pushfn 0.00 +5.2 5.19 ± 25% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.write.writen.record__pushfn.perf_mmap__push 0.00 +5.2 5.19 ± 25% perf-profile.calltrace.cycles-pp.generic_perform_write.shmem_file_write_iter.vfs_write.ksys_write.do_syscall_64 0.00 +5.2 5.19 ± 25% perf-profile.calltrace.cycles-pp.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe.write.writen 0.00 +5.2 5.19 ± 25% perf-profile.calltrace.cycles-pp.record__pushfn.perf_mmap__push.record__mmap_read_evlist.__cmd_record.cmd_record 0.00 +5.2 5.19 ± 25% perf-profile.calltrace.cycles-pp.shmem_file_write_iter.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe 0.00 +5.2 5.19 ± 25% perf-profile.calltrace.cycles-pp.write.writen.record__pushfn.perf_mmap__push.record__mmap_read_evlist 0.00 +5.2 5.19 ± 25% perf-profile.calltrace.cycles-pp.writen.record__pushfn.perf_mmap__push.record__mmap_read_evlist.__cmd_record 0.00 +5.4 5.38 ± 34% perf-profile.calltrace.cycles-pp.exit_mmap.__mmput.exit_mm.do_exit.do_group_exit 0.00 +5.5 5.50 ± 28% perf-profile.calltrace.cycles-pp.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe.write 0.00 +6.2 6.18 ± 9% perf-profile.calltrace.cycles-pp.perf_mmap__push.record__mmap_read_evlist.__cmd_record.cmd_record.run_builtin 0.00 +6.2 6.18 ± 9% perf-profile.calltrace.cycles-pp.record__mmap_read_evlist.__cmd_record.cmd_record.run_builtin.main 0.00 +6.3 6.34 ± 50% perf-profile.calltrace.cycles-pp.do_filp_open.do_sys_openat2.__x64_sys_openat.do_syscall_64.entry_SYSCALL_64_after_hwframe 0.00 +6.3 6.34 ± 50% perf-profile.calltrace.cycles-pp.path_openat.do_filp_open.do_sys_openat2.__x64_sys_openat.do_syscall_64 0.00 +6.8 6.82 ± 41% perf-profile.calltrace.cycles-pp.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault 0.00 +6.9 6.94 ± 23% perf-profile.calltrace.cycles-pp.__cmd_record.cmd_record.run_builtin.main 0.00 +6.9 6.94 ± 23% perf-profile.calltrace.cycles-pp.cmd_record.run_builtin.main 0.00 +8.6 8.58 ± 19% perf-profile.calltrace.cycles-pp.main 0.00 +8.6 8.58 ± 19% perf-profile.calltrace.cycles-pp.run_builtin.main 0.00 +11.4 11.42 ± 30% perf-profile.calltrace.cycles-pp.vsnprintf.seq_printf.show_interrupts.seq_read_iter.proc_reg_read_iter 0.00 +12.2 12.18 ± 26% perf-profile.calltrace.cycles-pp.smp_call_function_single.event_function_call.perf_event_release_kernel.perf_release.__fput 0.00 +12.5 12.53 ± 26% perf-profile.calltrace.cycles-pp.event_function_call.perf_event_release_kernel.perf_release.__fput.task_work_run 0.00 +12.7 12.71 ± 17% perf-profile.calltrace.cycles-pp.seq_printf.show_interrupts.seq_read_iter.proc_reg_read_iter.vfs_read 0.00 +17.2 17.20 ± 17% perf-profile.calltrace.cycles-pp.proc_reg_read_iter.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe 0.00 +17.2 17.20 ± 17% perf-profile.calltrace.cycles-pp.seq_read_iter.proc_reg_read_iter.vfs_read.ksys_read.do_syscall_64 0.00 +17.2 17.20 ± 17% perf-profile.calltrace.cycles-pp.show_interrupts.seq_read_iter.proc_reg_read_iter.vfs_read.ksys_read 0.00 +17.8 17.82 ± 9% perf-profile.calltrace.cycles-pp.perf_event_release_kernel.perf_release.__fput.task_work_run.do_exit 0.00 +17.8 17.82 ± 9% perf-profile.calltrace.cycles-pp.perf_release.__fput.task_work_run.do_exit.do_group_exit 0.00 +21.6 21.56 ± 7% perf-profile.calltrace.cycles-pp.__fput.task_work_run.do_exit.do_group_exit.get_signal 0.00 +22.6 22.64 ± 9% perf-profile.calltrace.cycles-pp.task_work_run.do_exit.do_group_exit.get_signal.arch_do_signal_or_restart 0.00 +24.2 24.16 ± 13% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.read 0.00 +24.2 24.16 ± 13% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.read 0.00 +24.2 24.16 ± 13% perf-profile.calltrace.cycles-pp.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe.read 0.00 +24.2 24.16 ± 13% perf-profile.calltrace.cycles-pp.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe.read 0.00 +25.2 25.18 ± 17% perf-profile.calltrace.cycles-pp.read 0.00 +26.8 26.76 ± 9% perf-profile.calltrace.cycles-pp.arch_do_signal_or_restart.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe 0.00 +26.8 26.76 ± 9% perf-profile.calltrace.cycles-pp.do_exit.do_group_exit.get_signal.arch_do_signal_or_restart.syscall_exit_to_user_mode 0.00 +26.8 26.76 ± 9% perf-profile.calltrace.cycles-pp.do_group_exit.get_signal.arch_do_signal_or_restart.syscall_exit_to_user_mode.do_syscall_64 0.00 +26.8 26.76 ± 9% perf-profile.calltrace.cycles-pp.get_signal.arch_do_signal_or_restart.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe 0.00 +27.4 27.39 ± 8% perf-profile.calltrace.cycles-pp.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe 0.00 +41.1 41.08 ± 10% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe 0.00 +41.1 41.08 ± 10% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe 77.16 ± 17% -77.2 0.00 perf-profile.children.cycles-pp.do_rw_once 21.07 ± 14% -21.1 0.00 perf-profile.children.cycles-pp.lrand48_r@plt 19.67 ± 71% -19.7 0.00 perf-profile.children.cycles-pp.do_access 17.09 ± 84% -17.1 0.00 perf-profile.children.cycles-pp.hugetlb_fault 13.57 ± 84% -13.6 0.00 perf-profile.children.cycles-pp.hugetlb_wp 13.30 ± 84% -13.3 0.00 perf-profile.children.cycles-pp.copy_user_large_folio 13.28 ± 84% -13.3 0.00 perf-profile.children.cycles-pp.copy_subpage 13.27 ± 84% -12.9 0.38 ±223% perf-profile.children.cycles-pp.copy_mc_enhanced_fast_string 0.00 +2.1 2.06 ± 9% perf-profile.children.cycles-pp.put_files_struct 0.02 ±142% +3.3 3.35 ± 39% perf-profile.children.cycles-pp.common_startup_64 0.02 ±142% +3.3 3.35 ± 39% perf-profile.children.cycles-pp.cpu_startup_entry 0.02 ±142% +3.3 3.35 ± 39% perf-profile.children.cycles-pp.do_idle 0.00 +3.8 3.77 ± 51% perf-profile.children.cycles-pp.kmem_cache_free 0.00 +4.3 4.32 ± 43% perf-profile.children.cycles-pp.sw_perf_event_destroy 0.01 ±223% +4.4 4.37 ± 49% perf-profile.children.cycles-pp.__x64_sys_exit_group 0.00 +4.4 4.41 ± 39% perf-profile.children.cycles-pp.link_path_walk 0.00 +4.6 4.63 ± 41% perf-profile.children.cycles-pp._free_event 0.06 ± 85% +5.1 5.19 ± 25% perf-profile.children.cycles-pp.record__pushfn 0.06 ± 85% +5.1 5.19 ± 25% perf-profile.children.cycles-pp.writen 0.04 ±108% +5.2 5.19 ± 25% perf-profile.children.cycles-pp.generic_perform_write 0.04 ±108% +5.2 5.19 ± 25% perf-profile.children.cycles-pp.shmem_file_write_iter 0.01 ±223% +5.4 5.38 ± 34% perf-profile.children.cycles-pp.__mmput 0.01 ±223% +5.4 5.38 ± 34% perf-profile.children.cycles-pp.exit_mm 0.01 ±223% +5.4 5.38 ± 34% perf-profile.children.cycles-pp.exit_mmap 0.08 ± 98% +5.4 5.50 ± 28% perf-profile.children.cycles-pp.write 0.08 ±100% +5.4 5.50 ± 28% perf-profile.children.cycles-pp.ksys_write 0.08 ± 99% +5.4 5.50 ± 28% perf-profile.children.cycles-pp.vfs_write 0.07 ± 86% +6.1 6.18 ± 9% perf-profile.children.cycles-pp.record__mmap_read_evlist 0.06 ± 85% +6.1 6.18 ± 9% perf-profile.children.cycles-pp.perf_mmap__push 0.00 +6.3 6.34 ± 50% perf-profile.children.cycles-pp.do_filp_open 0.00 +6.3 6.34 ± 50% perf-profile.children.cycles-pp.path_openat 0.00 +6.5 6.50 ± 40% perf-profile.children.cycles-pp.__handle_mm_fault 0.00 +7.0 7.04 ± 54% perf-profile.children.cycles-pp.__x64_sys_openat 0.00 +7.0 7.04 ± 54% perf-profile.children.cycles-pp.do_sys_openat2 0.08 ± 82% +8.5 8.58 ± 19% perf-profile.children.cycles-pp.main 0.08 ± 82% +8.5 8.58 ± 19% perf-profile.children.cycles-pp.run_builtin 0.08 ± 85% +8.5 8.58 ± 19% perf-profile.children.cycles-pp.__cmd_record 0.08 ± 85% +8.5 8.58 ± 19% perf-profile.children.cycles-pp.cmd_record 0.00 +11.7 11.74 ± 30% perf-profile.children.cycles-pp.vsnprintf 0.00 +12.7 12.71 ± 17% perf-profile.children.cycles-pp.seq_printf 0.00 +13.8 13.81 ± 19% perf-profile.children.cycles-pp.smp_call_function_single 0.00 +14.2 14.17 ± 18% perf-profile.children.cycles-pp.event_function_call 0.00 +17.2 17.20 ± 17% perf-profile.children.cycles-pp.proc_reg_read_iter 0.00 +17.2 17.20 ± 17% perf-profile.children.cycles-pp.show_interrupts 0.00 +17.8 17.82 ± 9% perf-profile.children.cycles-pp.perf_event_release_kernel 0.00 +17.8 17.82 ± 9% perf-profile.children.cycles-pp.perf_release 0.00 +22.3 22.30 ± 11% perf-profile.children.cycles-pp.__fput 0.02 ±142% +22.9 22.96 ± 8% perf-profile.children.cycles-pp.task_work_run 0.00 +23.2 23.22 ± 10% perf-profile.children.cycles-pp.seq_read_iter 0.00 +24.5 24.47 ± 13% perf-profile.children.cycles-pp.ksys_read 0.00 +24.5 24.47 ± 13% perf-profile.children.cycles-pp.vfs_read 0.00 +25.2 25.18 ± 13% perf-profile.children.cycles-pp.read 0.00 +26.8 26.76 ± 9% perf-profile.children.cycles-pp.arch_do_signal_or_restart 0.00 +26.8 26.76 ± 9% perf-profile.children.cycles-pp.get_signal 0.00 +27.1 27.08 ± 8% perf-profile.children.cycles-pp.syscall_exit_to_user_mode 0.01 ±223% +31.1 31.13 ± 8% perf-profile.children.cycles-pp.do_exit 0.01 ±223% +31.1 31.13 ± 8% perf-profile.children.cycles-pp.do_group_exit 0.18 ± 66% +85.5 85.67 perf-profile.children.cycles-pp.do_syscall_64 0.18 ± 66% +85.8 86.02 perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe 76.28 ± 17% -76.3 0.00 perf-profile.self.cycles-pp.do_rw_once 13.17 ± 84% -12.8 0.38 ±223% perf-profile.self.cycles-pp.copy_mc_enhanced_fast_string 0.00 +3.7 3.73 ± 53% perf-profile.self.cycles-pp.show_interrupts 0.00 +10.8 10.84 ± 40% perf-profile.self.cycles-pp.smp_call_function_single *************************************************************************************************** lkp-csl-2sp7: 96 threads 2 sockets Intel(R) Xeon(R) Gold 6252 CPU @ 2.10GHz (Cascade Lake) with 512G memory ========================================================================================= compiler/cpufreq_governor/kconfig/option_a/rootfs/tbox_group/test/testcase: gcc-12/performance/x86_64-rhel-8.3/SP.B [Scalar Penta-diagonal solver]/debian-12-x86_64-phoronix/lkp-csl-2sp7/npb-1.4.5/phoronix-test-suite commit: 9ef3ad0f03 ("ACPI/NUMA: Print CXL Early Discovery Table (CEDT)") 09f84589a8 ("ACPI/NUMA: Remove architecture dependent remainings") 9ef3ad0f032818c5 09f84589a8ed1162c0274a1b5b3 ---------------- --------------------------- %stddev %change %stddev \ | \ 14.98 +1.9% 15.26 boot-time.dhcp 6.328e+09 +62.7% 1.03e+10 cpuidle..time 6773424 +61.9% 10963809 cpuidle..usage 136.23 +50.7% 205.27 uptime.boot 10617 +38.0% 14647 uptime.idle 231376 +16.8% 270293 meminfo.Active 74278 +37.5% 102161 meminfo.Active(anon) 276857 ± 6% +31.8% 364759 ± 3% meminfo.AnonHugePages 11829930 ± 3% -38.6% 7260501 ± 11% meminfo.DirectMap2M 261359 -83.6% 42734 meminfo.VmallocUsed 0.15 -0.1 0.08 ± 2% mpstat.cpu.all.iowait% 1.82 +0.8 2.62 mpstat.cpu.all.irq% 0.09 +0.1 0.14 mpstat.cpu.all.soft% 0.89 +0.3 1.14 mpstat.cpu.all.sys% 20.41 +5.9 26.33 mpstat.cpu.all.usr% 523084 ± 23% -100.0% 0.00 numa-numastat.node0.local_node 577073 ± 19% -100.0% 0.00 numa-numastat.node0.numa_hit 53991 ± 50% -100.0% 0.00 numa-numastat.node0.other_node 560471 ± 22% -100.0% 0.00 numa-numastat.node1.local_node 606158 ± 18% -100.0% 0.00 numa-numastat.node1.numa_hit 45685 ± 60% -100.0% 0.00 numa-numastat.node1.other_node 76.91 -8.8% 70.16 vmstat.cpu.id 19.78 +31.9% 26.08 vmstat.cpu.us 3610 -43.7% 2032 vmstat.io.bi 0.29 ± 48% -65.8% 0.10 ± 6% vmstat.procs.b 21.91 +29.1% 28.29 vmstat.procs.r 6078 ± 4% -27.6% 4398 ± 3% vmstat.system.cs 22219 -58.1% 9301 phoronix-test-suite.npb.SP.B.total_mop_s 84.27 +79.7% 151.42 phoronix-test-suite.time.elapsed_time 84.27 +79.7% 151.42 phoronix-test-suite.time.elapsed_time.max 9092 +61.6% 14688 phoronix-test-suite.time.involuntary_context_switches 1556010 -61.5% 598472 phoronix-test-suite.time.minor_page_faults 2019 +30.6% 2637 phoronix-test-suite.time.percent_of_cpu_this_job_got 64.18 +140.5% 154.34 phoronix-test-suite.time.system_time 1637 +134.4% 3839 phoronix-test-suite.time.user_time 18650 +37.3% 25601 proc-vmstat.nr_active_anon 39298 +6.9% 42026 proc-vmstat.nr_active_file 296059 +7.8% 319032 proc-vmstat.nr_anon_pages 644447 +1.9% 657001 proc-vmstat.nr_file_pages 7113787 +17.7% 8370599 proc-vmstat.nr_foll_pin_acquired 7113780 +17.7% 8370587 proc-vmstat.nr_foll_pin_released 468080 +4.9% 490832 proc-vmstat.nr_inactive_anon 21467 +1.3% 21743 proc-vmstat.nr_kernel_stack 40569 +3.0% 41805 proc-vmstat.nr_mapped 3552 +6.1% 3767 proc-vmstat.nr_page_table_pages 190633 +3.4% 197135 proc-vmstat.nr_shmem 21713 +2.3% 22201 proc-vmstat.nr_slab_reclaimable 18650 +37.3% 25601 proc-vmstat.nr_zone_active_anon 39298 +6.9% 42026 proc-vmstat.nr_zone_active_file 468080 +4.9% 490832 proc-vmstat.nr_zone_inactive_anon 978315 -100.0% 0.00 proc-vmstat.numa_hint_faults 976417 -100.0% 0.00 proc-vmstat.numa_hint_faults_local 1184528 -1.1% 1171481 proc-vmstat.numa_hit 1319 -100.0% 0.00 proc-vmstat.numa_huge_pte_updates 1084853 +8.0% 1171481 proc-vmstat.numa_local 99677 -100.0% 0.00 proc-vmstat.numa_other 3465 ± 63% -100.0% 0.00 proc-vmstat.numa_pages_migrated 1700329 -100.0% 0.00 proc-vmstat.numa_pte_updates 71755 +5.5% 75670 proc-vmstat.pgactivate 1444406 +0.9% 1458081 proc-vmstat.pgalloc_normal 2020069 -49.4% 1022408 proc-vmstat.pgfault 1268206 +3.8% 1316886 proc-vmstat.pgfree 3465 ± 63% -100.0% 0.00 proc-vmstat.pgmigrate_success 100000 ± 47% -100.0% 0.00 numa-meminfo.node0.Active 13486 ± 48% -100.0% 0.00 numa-meminfo.node0.Active(anon) 86514 ± 49% -100.0% 0.00 numa-meminfo.node0.Active(file) 143140 ± 29% -100.0% 0.00 numa-meminfo.node0.AnonHugePages 554668 ± 43% -100.0% 0.00 numa-meminfo.node0.AnonPages 720916 ± 33% -100.0% 0.00 numa-meminfo.node0.AnonPages.max 1320009 ± 59% -100.0% 0.00 numa-meminfo.node0.FilePages 1773835 ± 32% -100.0% 0.00 numa-meminfo.node0.Inactive 678653 ± 49% -100.0% 0.00 numa-meminfo.node0.Inactive(anon) 1095181 ± 64% -100.0% 0.00 numa-meminfo.node0.Inactive(file) 46789 ± 28% -100.0% 0.00 numa-meminfo.node0.KReclaimable 11095 ± 4% -100.0% 0.00 numa-meminfo.node0.KernelStack 101067 ± 28% -100.0% 0.00 numa-meminfo.node0.Mapped 2.611e+08 -100.0% 0.00 numa-meminfo.node0.MemFree 2.638e+08 -100.0% 0.00 numa-meminfo.node0.MemTotal 2732014 ± 23% -100.0% 0.00 numa-meminfo.node0.MemUsed 7763 ± 6% -100.0% 0.00 numa-meminfo.node0.PageTables 46789 ± 28% -100.0% 0.00 numa-meminfo.node0.SReclaimable 125988 ± 9% -100.0% 0.00 numa-meminfo.node0.SUnreclaim 138154 ±125% -100.0% 0.00 numa-meminfo.node0.Shmem 172777 ± 12% -100.0% 0.00 numa-meminfo.node0.Slab 3231 ± 5% -100.0% 0.00 numa-meminfo.node0.Unevictable 131809 ± 35% -100.0% 0.00 numa-meminfo.node1.Active 61136 ± 10% -100.0% 0.00 numa-meminfo.node1.Active(anon) 70672 ± 59% -100.0% 0.00 numa-meminfo.node1.Active(file) 133009 ± 33% -100.0% 0.00 numa-meminfo.node1.AnonHugePages 629614 ± 38% -100.0% 0.00 numa-meminfo.node1.AnonPages 819699 ± 29% -100.0% 0.00 numa-meminfo.node1.AnonPages.max 1257793 ± 62% -100.0% 0.00 numa-meminfo.node1.FilePages 1756533 ± 32% -100.0% 0.00 numa-meminfo.node1.Inactive 1193976 ± 28% -100.0% 0.00 numa-meminfo.node1.Inactive(anon) 562556 ±126% -100.0% 0.00 numa-meminfo.node1.Inactive(file) 40060 ± 33% -100.0% 0.00 numa-meminfo.node1.KReclaimable 10374 ± 4% -100.0% 0.00 numa-meminfo.node1.KernelStack 60225 ± 48% -100.0% 0.00 numa-meminfo.node1.Mapped 2.614e+08 -100.0% 0.00 numa-meminfo.node1.MemFree 2.642e+08 -100.0% 0.00 numa-meminfo.node1.MemTotal 2770706 ± 22% -100.0% 0.00 numa-meminfo.node1.MemUsed 6462 ± 8% -100.0% 0.00 numa-meminfo.node1.PageTables 40060 ± 33% -100.0% 0.00 numa-meminfo.node1.SReclaimable 116360 ± 10% -100.0% 0.00 numa-meminfo.node1.SUnreclaim 624395 ± 27% -100.0% 0.00 numa-meminfo.node1.Shmem 156420 ± 13% -100.0% 0.00 numa-meminfo.node1.Slab 0.01 ± 17% +56.1% 0.02 ± 16% perf-sched.sch_delay.avg.ms.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part 0.03 ± 51% +147.3% 0.08 ± 39% perf-sched.sch_delay.avg.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown] 0.02 ± 24% +132.8% 0.05 ± 52% perf-sched.sch_delay.avg.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep 0.01 ± 5% +110.7% 0.02 ± 13% perf-sched.sch_delay.avg.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64 0.02 ± 7% +207.9% 0.05 ± 87% perf-sched.sch_delay.avg.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64 0.02 ± 71% +159.7% 0.06 ± 37% perf-sched.sch_delay.avg.ms.futex_wait_queue.__futex_wait.futex_wait.do_futex 0.00 +4144.4% 0.13 ±213% perf-sched.sch_delay.avg.ms.pipe_read.vfs_read.ksys_read.do_syscall_64 0.01 ± 16% +135.3% 0.03 ± 33% perf-sched.sch_delay.avg.ms.schedule_hrtimeout_range_clock.do_select.core_sys_select.do_pselect.constprop 0.03 ± 37% +81.9% 0.05 ± 14% perf-sched.sch_delay.avg.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait 0.01 ± 5% +117.5% 0.02 ± 40% perf-sched.sch_delay.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone 0.02 ± 2% +74.2% 0.03 ± 5% perf-sched.sch_delay.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm 0.01 ± 18% +70.5% 0.01 ± 12% perf-sched.sch_delay.avg.ms.wait_for_partner.fifo_open.do_dentry_open.do_open 0.05 ± 56% +186.6% 0.14 ± 43% perf-sched.sch_delay.max.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown] 0.09 ± 52% +1385.3% 1.38 ±198% perf-sched.sch_delay.max.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64 0.06 ± 27% +2.6e+05% 167.11 ±223% perf-sched.sch_delay.max.ms.pipe_read.vfs_read.ksys_read.do_syscall_64 0.04 ± 37% +277.6% 0.13 ± 44% perf-sched.sch_delay.max.ms.schedule_hrtimeout_range_clock.do_select.core_sys_select.do_pselect.constprop 0.05 ± 22% +190.8% 0.14 ± 24% perf-sched.sch_delay.max.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait 0.05 ± 10% +3160.1% 1.61 ±197% perf-sched.sch_delay.max.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.do_epoll_pwait.part 0.04 ± 21% +222.5% 0.14 ± 78% perf-sched.sch_delay.max.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone 0.04 ± 33% +193.0% 0.12 ± 68% perf-sched.sch_delay.max.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread 8369 ± 7% -10.6% 7478 ± 4% perf-sched.total_wait_and_delay.count.ms 253.51 ± 3% +22.4% 310.30 ± 2% perf-sched.wait_and_delay.avg.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64 1.03 ± 5% +106.4% 2.13 ± 5% perf-sched.wait_and_delay.avg.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64 0.10 ± 5% +73.2% 0.18 ± 8% perf-sched.wait_and_delay.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown].[unknown] 1.51 ± 4% +62.7% 2.46 ± 8% perf-sched.wait_and_delay.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone 124.00 -48.8% 63.50 perf-sched.wait_and_delay.count.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64 123.00 -49.3% 62.33 perf-sched.wait_and_delay.count.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64 2125 -34.1% 1400 ± 27% perf-sched.wait_and_delay.count.pipe_read.vfs_read.ksys_read.do_syscall_64 86.00 -52.3% 41.00 perf-sched.wait_and_delay.count.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone 19.83 -49.6% 10.00 perf-sched.wait_and_delay.count.schedule_timeout.kcompactd.kthread.ret_from_fork 550.00 ± 2% -12.2% 482.67 ± 2% perf-sched.wait_and_delay.count.worker_thread.kthread.ret_from_fork.ret_from_fork_asm 21.99 ± 3% +26.6% 27.84 ± 7% perf-sched.wait_and_delay.max.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64 5.86 ± 6% +78.7% 10.47 ± 8% perf-sched.wait_and_delay.max.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone 253.50 ± 3% +22.4% 310.28 ± 2% perf-sched.wait_time.avg.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64 1.02 ± 5% +104.8% 2.08 ± 7% perf-sched.wait_time.avg.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64 0.10 ± 17% +66.2% 0.17 ± 11% perf-sched.wait_time.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown] 0.10 ± 5% +73.2% 0.18 ± 8% perf-sched.wait_time.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown].[unknown] 0.04 ±110% +332.7% 0.16 ± 57% perf-sched.wait_time.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown].[unknown] 1.50 ± 4% +62.3% 2.43 ± 8% perf-sched.wait_time.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone 0.00 ± 40% +750.0% 0.02 ±122% perf-sched.wait_time.avg.ms.wait_for_partner.fifo_open.do_dentry_open.do_open 21.97 ± 3% +26.6% 27.82 ± 7% perf-sched.wait_time.max.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64 0.35 ± 16% +48.7% 0.52 ± 7% perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown] 0.06 ±108% +379.9% 0.29 ± 50% perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown].[unknown] 5.84 ± 6% +78.8% 10.44 ± 8% perf-sched.wait_time.max.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone 0.03 ± 16% +851.5% 0.31 ±157% perf-sched.wait_time.max.ms.wait_for_partner.fifo_open.do_dentry_open.do_open 3375 ± 48% -100.0% 0.00 numa-vmstat.node0.nr_active_anon 21639 ± 49% -100.0% 0.00 numa-vmstat.node0.nr_active_file 138732 ± 43% -100.0% 0.00 numa-vmstat.node0.nr_anon_pages 69.78 ± 29% -100.0% 0.00 numa-vmstat.node0.nr_anon_transparent_hugepages 330019 ± 59% -100.0% 0.00 numa-vmstat.node0.nr_file_pages 3551956 -100.0% 0.00 numa-vmstat.node0.nr_foll_pin_acquired 3551953 -100.0% 0.00 numa-vmstat.node0.nr_foll_pin_released 65277218 -100.0% 0.00 numa-vmstat.node0.nr_free_pages 169733 ± 49% -100.0% 0.00 numa-vmstat.node0.nr_inactive_anon 273801 ± 64% -100.0% 0.00 numa-vmstat.node0.nr_inactive_file 11096 ± 4% -100.0% 0.00 numa-vmstat.node0.nr_kernel_stack 25457 ± 28% -100.0% 0.00 numa-vmstat.node0.nr_mapped 1937 ± 6% -100.0% 0.00 numa-vmstat.node0.nr_page_table_pages 34539 ±125% -100.0% 0.00 numa-vmstat.node0.nr_shmem 11695 ± 28% -100.0% 0.00 numa-vmstat.node0.nr_slab_reclaimable 31493 ± 9% -100.0% 0.00 numa-vmstat.node0.nr_slab_unreclaimable 807.98 ± 5% -100.0% 0.00 numa-vmstat.node0.nr_unevictable 3375 ± 48% -100.0% 0.00 numa-vmstat.node0.nr_zone_active_anon 21639 ± 49% -100.0% 0.00 numa-vmstat.node0.nr_zone_active_file 169735 ± 49% -100.0% 0.00 numa-vmstat.node0.nr_zone_inactive_anon 273801 ± 64% -100.0% 0.00 numa-vmstat.node0.nr_zone_inactive_file 807.98 ± 5% -100.0% 0.00 numa-vmstat.node0.nr_zone_unevictable 576852 ± 19% -100.0% 0.00 numa-vmstat.node0.numa_hit 522864 ± 24% -100.0% 0.00 numa-vmstat.node0.numa_local 53991 ± 50% -100.0% 0.00 numa-vmstat.node0.numa_other 15300 ± 10% -100.0% 0.00 numa-vmstat.node1.nr_active_anon 17669 ± 59% -100.0% 0.00 numa-vmstat.node1.nr_active_file 157621 ± 38% -100.0% 0.00 numa-vmstat.node1.nr_anon_pages 314384 ± 62% -100.0% 0.00 numa-vmstat.node1.nr_file_pages 3552396 -100.0% 0.00 numa-vmstat.node1.nr_foll_pin_acquired 3552390 -100.0% 0.00 numa-vmstat.node1.nr_foll_pin_released 65353661 -100.0% 0.00 numa-vmstat.node1.nr_free_pages 298619 ± 28% -100.0% 0.00 numa-vmstat.node1.nr_inactive_anon 140594 ±126% -100.0% 0.00 numa-vmstat.node1.nr_inactive_file 10373 ± 4% -100.0% 0.00 numa-vmstat.node1.nr_kernel_stack 15147 ± 48% -100.0% 0.00 numa-vmstat.node1.nr_mapped 1615 ± 8% -100.0% 0.00 numa-vmstat.node1.nr_page_table_pages 156079 ± 27% -100.0% 0.00 numa-vmstat.node1.nr_shmem 10015 ± 33% -100.0% 0.00 numa-vmstat.node1.nr_slab_reclaimable 29089 ± 10% -100.0% 0.00 numa-vmstat.node1.nr_slab_unreclaimable 15300 ± 10% -100.0% 0.00 numa-vmstat.node1.nr_zone_active_anon 17669 ± 59% -100.0% 0.00 numa-vmstat.node1.nr_zone_active_file 298620 ± 28% -100.0% 0.00 numa-vmstat.node1.nr_zone_inactive_anon 140594 ±126% -100.0% 0.00 numa-vmstat.node1.nr_zone_inactive_file 605836 ± 18% -100.0% 0.00 numa-vmstat.node1.numa_hit 560149 ± 22% -100.0% 0.00 numa-vmstat.node1.numa_local 45685 ± 60% -100.0% 0.00 numa-vmstat.node1.numa_other 9.24 -26.9% 6.75 perf-stat.i.MPKI 3.006e+09 ± 3% +30.6% 3.924e+09 perf-stat.i.branch-instructions 2.88 -0.7 2.19 ± 2% perf-stat.i.branch-miss-rate% 34648844 -23.3% 26559921 perf-stat.i.branch-misses 44.09 +8.1 52.19 perf-stat.i.cache-miss-rate% 4.032e+08 -44.5% 2.239e+08 perf-stat.i.cache-misses 6.104e+08 -42.7% 3.496e+08 perf-stat.i.cache-references 6150 ± 4% -27.2% 4479 ± 3% perf-stat.i.context-switches 1.98 ± 2% +40.0% 2.78 perf-stat.i.cpi 5.387e+10 +33.5% 7.189e+10 perf-stat.i.cpu-cycles 156.13 -9.6% 141.17 perf-stat.i.cpu-migrations 666.32 ± 3% -13.9% 574.01 perf-stat.i.cycles-between-cache-misses 1379778 -36.7% 873874 ± 2% perf-stat.i.dTLB-load-misses 1.241e+10 ± 2% -5.1% 1.178e+10 perf-stat.i.dTLB-loads 0.04 ± 4% -0.0 0.03 ± 3% perf-stat.i.dTLB-store-miss-rate% 976671 -45.5% 532272 perf-stat.i.dTLB-store-misses 4.839e+09 -12.5% 4.236e+09 perf-stat.i.dTLB-stores 41.13 ± 4% -3.8 37.28 perf-stat.i.iTLB-load-miss-rate% 1105884 -24.3% 837557 perf-stat.i.iTLB-load-misses 1971195 ± 11% -18.1% 1613463 perf-stat.i.iTLB-loads 26637 ± 2% +45.0% 38621 perf-stat.i.instructions-per-iTLB-miss 0.61 -26.0% 0.45 perf-stat.i.ipc 12.95 ± 3% -33.3% 8.64 ± 3% perf-stat.i.major-faults 0.56 +33.4% 0.75 perf-stat.i.metric.GHz 1256 -37.8% 781.77 perf-stat.i.metric.K/sec 217.05 ± 2% -2.8% 210.92 perf-stat.i.metric.M/sec 22076 -73.0% 5958 perf-stat.i.minor-faults 30.69 ± 2% +26.8 57.44 perf-stat.i.node-load-miss-rate% 5638356 +188.6% 16272394 perf-stat.i.node-load-misses 51095437 -68.4% 16149472 perf-stat.i.node-loads 24.17 ± 5% +33.5 57.66 perf-stat.i.node-store-miss-rate% 3634105 +207.6% 11177771 perf-stat.i.node-store-misses 33521254 -68.1% 10709700 perf-stat.i.node-stores 22089 -73.0% 5966 perf-stat.i.page-faults 13.73 -44.7% 7.60 perf-stat.overall.MPKI 1.12 ± 2% -0.5 0.65 perf-stat.overall.branch-miss-rate% 66.19 -1.9 64.29 perf-stat.overall.cache-miss-rate% 1.83 +33.2% 2.44 perf-stat.overall.cpi 133.26 +140.7% 320.76 perf-stat.overall.cycles-between-cache-misses 0.01 -0.0 0.01 ± 3% perf-stat.overall.dTLB-load-miss-rate% 0.02 -0.0 0.01 perf-stat.overall.dTLB-store-miss-rate% 27181 ± 2% +33.5% 36284 perf-stat.overall.instructions-per-iTLB-miss 0.55 -25.0% 0.41 perf-stat.overall.ipc 9.93 +40.3 50.21 perf-stat.overall.node-load-miss-rate% 9.78 +41.3 51.05 perf-stat.overall.node-store-miss-rate% 3.017e+09 ± 3% +31.3% 3.961e+09 perf-stat.ps.branch-instructions 33863841 -23.7% 25852545 perf-stat.ps.branch-misses 4.08e+08 -44.2% 2.279e+08 perf-stat.ps.cache-misses 6.164e+08 -42.5% 3.544e+08 perf-stat.ps.cache-references 5959 ± 4% -27.9% 4295 ± 4% perf-stat.ps.context-switches 5.437e+10 +34.4% 7.309e+10 perf-stat.ps.cpu-cycles 153.77 -10.2% 138.05 perf-stat.ps.cpu-migrations 1362804 -37.6% 851037 ± 2% perf-stat.ps.dTLB-load-misses 1.253e+10 ± 2% -4.7% 1.194e+10 perf-stat.ps.dTLB-loads 982670 -45.5% 535635 perf-stat.ps.dTLB-store-misses 4.886e+09 -12.1% 4.294e+09 perf-stat.ps.dTLB-stores 1093709 -24.4% 826521 perf-stat.ps.iTLB-load-misses 1967627 ± 10% -17.5% 1623220 perf-stat.ps.iTLB-loads 12.44 ± 3% -35.3% 8.05 ± 3% perf-stat.ps.major-faults 21849 -74.4% 5599 perf-stat.ps.minor-faults 5700455 +191.0% 16587479 perf-stat.ps.node-load-misses 51713905 -68.2% 16452569 perf-stat.ps.node-loads 3675605 +209.5% 11374853 perf-stat.ps.node-store-misses 33923436 -67.8% 10906955 perf-stat.ps.node-stores 21862 -74.3% 5607 perf-stat.ps.page-faults 2.516e+12 +82.7% 4.596e+12 perf-stat.total.instructions 7647 ± 15% +7433.9% 576144 ± 2% sched_debug.cfs_rq:/.avg_vruntime.avg 30894 ± 29% +3982.0% 1261110 ± 6% sched_debug.cfs_rq:/.avg_vruntime.max 2728 ± 33% +1541.3% 44777 ± 91% sched_debug.cfs_rq:/.avg_vruntime.min 4017 ± 12% +7295.9% 297155 ± 5% sched_debug.cfs_rq:/.avg_vruntime.stddev 0.07 ± 33% +400.4% 0.37 ± 4% sched_debug.cfs_rq:/.h_nr_running.avg 0.26 ± 14% +62.7% 0.42 ± 2% sched_debug.cfs_rq:/.h_nr_running.stddev 1477 ± 93% +658.6% 11208 ± 24% sched_debug.cfs_rq:/.load.avg 7647 ± 15% +7433.9% 576145 ± 2% sched_debug.cfs_rq:/.min_vruntime.avg 30894 ± 29% +3982.0% 1261110 ± 6% sched_debug.cfs_rq:/.min_vruntime.max 2728 ± 33% +1541.3% 44777 ± 91% sched_debug.cfs_rq:/.min_vruntime.min 4017 ± 12% +7295.9% 297155 ± 5% sched_debug.cfs_rq:/.min_vruntime.stddev 0.07 ± 33% +400.4% 0.37 ± 4% sched_debug.cfs_rq:/.nr_running.avg 0.26 ± 14% +62.7% 0.42 ± 2% sched_debug.cfs_rq:/.nr_running.stddev 250.19 ± 8% +65.6% 414.25 ± 4% sched_debug.cfs_rq:/.runnable_avg.avg 276.31 ± 5% +48.5% 410.18 sched_debug.cfs_rq:/.runnable_avg.stddev 249.62 ± 8% +65.5% 413.18 ± 4% sched_debug.cfs_rq:/.util_avg.avg 276.00 ± 5% +48.6% 410.18 sched_debug.cfs_rq:/.util_avg.stddev 23.11 ± 19% +497.9% 138.20 ± 5% sched_debug.cfs_rq:/.util_est.avg 845.17 ± 8% -16.7% 704.06 ± 9% sched_debug.cfs_rq:/.util_est.max 109.08 ± 8% +65.1% 180.12 ± 4% sched_debug.cfs_rq:/.util_est.stddev 6.61 ±128% +214.3% 20.79 ± 70% sched_debug.cfs_rq:/init.scope.avg_vruntime.min 6.61 ±128% +214.3% 20.79 ± 70% sched_debug.cfs_rq:/init.scope.min_vruntime.min 50610 +61.8% 81883 ± 34% sched_debug.cfs_rq:/init.scope.se->exec_start.avg 50661 +61.9% 82013 ± 34% sched_debug.cfs_rq:/init.scope.se->exec_start.max 50549 +61.5% 81627 ± 35% sched_debug.cfs_rq:/init.scope.se->exec_start.min 153.50 ±153% +494.9% 913.22 ± 62% sched_debug.cfs_rq:/init.scope.se->load.weight.min 6.80 ±133% +216.4% 21.51 ± 69% sched_debug.cfs_rq:/init.scope.se->sum_exec_runtime.min 6475 ± 32% +1515.3% 104591 ± 91% sched_debug.cfs_rq:/init.scope.se->vruntime.min 13999 ± 20% -66.0% 4760 ± 33% sched_debug.cfs_rq:/init.scope.tg_load_avg.avg 13999 ± 20% -66.0% 4760 ± 33% sched_debug.cfs_rq:/init.scope.tg_load_avg.max 13999 ± 20% -66.0% 4760 ± 33% sched_debug.cfs_rq:/init.scope.tg_load_avg.min 3518 ± 15% +18809.0% 665385 ± 2% sched_debug.cfs_rq:/system.slice.avg_vruntime.avg 23752 ± 49% +5906.4% 1426667 ± 7% sched_debug.cfs_rq:/system.slice.avg_vruntime.max 1472 ± 31% +3291.4% 49946 ± 90% sched_debug.cfs_rq:/system.slice.avg_vruntime.min 2920 ± 39% +11804.1% 347701 ± 7% sched_debug.cfs_rq:/system.slice.avg_vruntime.stddev 0.09 ± 34% +424.6% 0.46 ± 4% sched_debug.cfs_rq:/system.slice.h_nr_running.avg 0.28 ± 14% +46.3% 0.41 ± 2% sched_debug.cfs_rq:/system.slice.h_nr_running.stddev 2067 ± 52% +412.4% 10591 ± 6% sched_debug.cfs_rq:/system.slice.load.avg 10.36 ± 6% +16.9% 12.12 ± 4% sched_debug.cfs_rq:/system.slice.load_avg.avg 68.67 ± 21% -47.2% 36.28 ± 17% sched_debug.cfs_rq:/system.slice.load_avg.max 13.24 ± 10% -20.1% 10.58 ± 6% sched_debug.cfs_rq:/system.slice.load_avg.stddev 3518 ± 15% +18809.0% 665385 ± 2% sched_debug.cfs_rq:/system.slice.min_vruntime.avg 23752 ± 49% +5906.4% 1426667 ± 7% sched_debug.cfs_rq:/system.slice.min_vruntime.max 1472 ± 31% +3291.4% 49946 ± 90% sched_debug.cfs_rq:/system.slice.min_vruntime.min 2920 ± 39% +11804.1% 347701 ± 7% sched_debug.cfs_rq:/system.slice.min_vruntime.stddev 0.09 ± 34% +424.6% 0.46 ± 4% sched_debug.cfs_rq:/system.slice.nr_running.avg 0.28 ± 14% +46.3% 0.41 ± 2% sched_debug.cfs_rq:/system.slice.nr_running.stddev 287.37 ± 7% +75.3% 503.83 ± 5% sched_debug.cfs_rq:/system.slice.runnable_avg.avg 277.25 ± 5% +43.0% 396.44 sched_debug.cfs_rq:/system.slice.runnable_avg.stddev 6.50 ± 55% +91.9% 12.48 ± 8% sched_debug.cfs_rq:/system.slice.se->avg.load_avg.avg 287.37 ± 7% +75.3% 503.88 ± 5% sched_debug.cfs_rq:/system.slice.se->avg.runnable_avg.avg 277.25 ± 5% +43.0% 396.37 sched_debug.cfs_rq:/system.slice.se->avg.runnable_avg.stddev 287.35 ± 7% +75.3% 503.74 ± 5% sched_debug.cfs_rq:/system.slice.se->avg.util_avg.avg 277.24 ± 5% +43.0% 396.49 sched_debug.cfs_rq:/system.slice.se->avg.util_avg.stddev 50573 +116.7% 109592 sched_debug.cfs_rq:/system.slice.se->exec_start.avg 50741 +117.9% 110587 sched_debug.cfs_rq:/system.slice.se->exec_start.max 43876 +137.1% 104039 sched_debug.cfs_rq:/system.slice.se->exec_start.min 745.42 +42.9% 1065 ± 9% sched_debug.cfs_rq:/system.slice.se->exec_start.stddev 327582 ± 23% -63.1% 120920 ± 31% sched_debug.cfs_rq:/system.slice.se->load.weight.max 48399 ± 22% -58.8% 19928 ± 25% sched_debug.cfs_rq:/system.slice.se->load.weight.stddev 330.49 ± 22% +4461.4% 15074 sched_debug.cfs_rq:/system.slice.se->sum_exec_runtime.avg 10758 ± 44% +226.9% 35171 ± 7% sched_debug.cfs_rq:/system.slice.se->sum_exec_runtime.max 29.77 ± 30% +4193.1% 1277 ± 84% sched_debug.cfs_rq:/system.slice.se->sum_exec_runtime.min 1382 ± 35% +484.7% 8085 ± 7% sched_debug.cfs_rq:/system.slice.se->sum_exec_runtime.stddev 7861 ± 13% +7258.1% 578439 sched_debug.cfs_rq:/system.slice.se->vruntime.avg 29255 ± 32% +4127.3% 1236710 ± 6% sched_debug.cfs_rq:/system.slice.se->vruntime.max 2957 ± 30% +1554.1% 48917 ± 83% sched_debug.cfs_rq:/system.slice.se->vruntime.min 4044 ± 16% +7281.5% 298577 ± 6% sched_debug.cfs_rq:/system.slice.se->vruntime.stddev 302.45 ± 62% -77.0% 69.42 ± 88% sched_debug.cfs_rq:/system.slice.tg_load_avg.stddev 287.35 ± 7% +75.3% 503.74 ± 5% sched_debug.cfs_rq:/system.slice.util_avg.avg 277.24 ± 5% +43.0% 396.48 sched_debug.cfs_rq:/system.slice.util_avg.stddev 1.57 ±193% +1094.7% 18.73 ± 51% sched_debug.cfs_rq:/system.slice/containerd.service.avg_vruntime.min 0.86 ± 26% +364.2% 3.97 ± 16% sched_debug.cfs_rq:/system.slice/containerd.service.load_avg.avg 3.83 ± 55% +140.6% 9.22 ± 19% sched_debug.cfs_rq:/system.slice/containerd.service.load_avg.max 1.41 ± 48% +148.1% 3.51 ± 22% sched_debug.cfs_rq:/system.slice/containerd.service.load_avg.stddev 1.57 ±193% +1094.7% 18.73 ± 51% sched_debug.cfs_rq:/system.slice/containerd.service.min_vruntime.min 0.88 ± 28% +355.8% 4.01 ± 16% sched_debug.cfs_rq:/system.slice/containerd.service.runnable_avg.avg 4.00 ± 55% +130.6% 9.22 ± 19% sched_debug.cfs_rq:/system.slice/containerd.service.runnable_avg.max 1.47 ± 49% +138.4% 3.50 ± 22% sched_debug.cfs_rq:/system.slice/containerd.service.runnable_avg.stddev 0.26 ± 92% +390.2% 1.29 ± 20% sched_debug.cfs_rq:/system.slice/containerd.service.se->avg.load_avg.avg 0.85 ± 32% +368.2% 3.99 ± 16% sched_debug.cfs_rq:/system.slice/containerd.service.se->avg.runnable_avg.avg 3.83 ± 61% +137.7% 9.11 ± 18% sched_debug.cfs_rq:/system.slice/containerd.service.se->avg.runnable_avg.max 1.41 ± 53% +145.3% 3.46 ± 21% sched_debug.cfs_rq:/system.slice/containerd.service.se->avg.runnable_avg.stddev 0.85 ± 32% +365.3% 3.96 ± 17% sched_debug.cfs_rq:/system.slice/containerd.service.se->avg.util_avg.avg 3.83 ± 61% +136.2% 9.06 ± 19% sched_debug.cfs_rq:/system.slice/containerd.service.se->avg.util_avg.max 1.41 ± 53% +143.9% 3.44 ± 22% sched_debug.cfs_rq:/system.slice/containerd.service.se->avg.util_avg.stddev 50173 +118.7% 109726 sched_debug.cfs_rq:/system.slice/containerd.service.se->exec_start.avg 50702 +117.3% 110151 sched_debug.cfs_rq:/system.slice/containerd.service.se->exec_start.max 48149 ± 6% +125.9% 108746 sched_debug.cfs_rq:/system.slice/containerd.service.se->exec_start.min 738718 ± 15% -20.1% 590310 ± 6% sched_debug.cfs_rq:/system.slice/containerd.service.se->load.weight.max 2.11 ±154% +825.1% 19.48 ± 48% sched_debug.cfs_rq:/system.slice/containerd.service.se->sum_exec_runtime.min 3846 ± 24% +12555.9% 486815 ± 20% sched_debug.cfs_rq:/system.slice/containerd.service.se->vruntime.avg 8152 ± 52% +8857.9% 730249 ± 21% sched_debug.cfs_rq:/system.slice/containerd.service.se->vruntime.max 2008 ± 22% +14293.3% 289128 ± 44% sched_debug.cfs_rq:/system.slice/containerd.service.se->vruntime.min 2041 ± 71% +8227.2% 170022 ± 38% sched_debug.cfs_rq:/system.slice/containerd.service.se->vruntime.stddev 5.20 ± 39% +266.8% 19.08 ± 22% sched_debug.cfs_rq:/system.slice/containerd.service.tg_load_avg.avg 5.33 ± 38% +293.8% 21.00 ± 25% sched_debug.cfs_rq:/system.slice/containerd.service.tg_load_avg.max 5.00 ± 38% +255.6% 17.78 ± 22% sched_debug.cfs_rq:/system.slice/containerd.service.tg_load_avg.min 0.16 ±141% +651.8% 1.19 ± 70% sched_debug.cfs_rq:/system.slice/containerd.service.tg_load_avg.stddev 0.86 ± 26% +361.1% 3.95 ± 17% sched_debug.cfs_rq:/system.slice/containerd.service.tg_load_avg_contrib.avg 3.83 ± 55% +140.6% 9.22 ± 19% sched_debug.cfs_rq:/system.slice/containerd.service.tg_load_avg_contrib.max 1.41 ± 48% +146.1% 3.48 ± 22% sched_debug.cfs_rq:/system.slice/containerd.service.tg_load_avg_contrib.stddev 0.88 ± 28% +353.0% 3.99 ± 17% sched_debug.cfs_rq:/system.slice/containerd.service.util_avg.avg 4.00 ± 55% +129.2% 9.17 ± 20% sched_debug.cfs_rq:/system.slice/containerd.service.util_avg.max 1.47 ± 49% +137.0% 3.48 ± 22% sched_debug.cfs_rq:/system.slice/containerd.service.util_avg.stddev 50765 +119.4% 111370 sched_debug.cpu.clock.avg 50771 +119.4% 111382 sched_debug.cpu.clock.max 50759 +119.4% 111356 sched_debug.cpu.clock.min 3.27 ± 10% +126.1% 7.40 ± 21% sched_debug.cpu.clock.stddev 50601 +116.6% 109610 sched_debug.cpu.clock_task.avg 50745 +118.0% 110614 sched_debug.cpu.clock_task.max 43880 +134.7% 102986 sched_debug.cpu.clock_task.min 696.32 +51.9% 1057 ± 6% sched_debug.cpu.clock_task.stddev 165.30 ± 33% +792.2% 1474 ± 2% sched_debug.cpu.curr->pid.avg 2493 +85.2% 4618 sched_debug.cpu.curr->pid.max 587.01 ± 16% +230.2% 1938 sched_debug.cpu.curr->pid.stddev 0.00 ± 22% +76.0% 0.00 ± 7% sched_debug.cpu.next_balance.stddev 0.07 ± 28% +293.0% 0.29 ± 3% sched_debug.cpu.nr_running.avg 0.26 ± 12% +60.0% 0.42 ± 2% sched_debug.cpu.nr_running.stddev 1606 ± 3% +188.0% 4627 ± 3% sched_debug.cpu.nr_switches.avg 11893 ± 21% +400.8% 59557 ± 2% sched_debug.cpu.nr_switches.max 254.67 ± 13% +299.7% 1017 ± 17% sched_debug.cpu.nr_switches.min 2062 ± 11% +220.6% 6611 ± 3% sched_debug.cpu.nr_switches.stddev 0.14 ± 17% -76.4% 0.03 ± 14% sched_debug.cpu.nr_uninterruptible.avg 50761 +119.4% 111355 sched_debug.cpu_clk 50082 +121.0% 110688 sched_debug.ktime 51315 +118.1% 111904 sched_debug.sched_clk 19.88 ± 35% -19.9 0.00 perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe 19.88 ± 35% -19.9 0.00 perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe 19.20 ± 16% -19.2 0.00 perf-profile.calltrace.cycles-pp.write 17.84 ± 19% -17.8 0.00 perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.write 17.84 ± 19% -17.8 0.00 perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.write 17.84 ± 19% -17.8 0.00 perf-profile.calltrace.cycles-pp.generic_perform_write.shmem_file_write_iter.vfs_write.ksys_write.do_syscall_64 17.84 ± 19% -17.8 0.00 perf-profile.calltrace.cycles-pp.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe.write 17.84 ± 19% -17.8 0.00 perf-profile.calltrace.cycles-pp.shmem_file_write_iter.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe 17.84 ± 19% -17.8 0.00 perf-profile.calltrace.cycles-pp.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe.write 12.83 ± 44% -12.8 0.00 perf-profile.calltrace.cycles-pp.arch_do_signal_or_restart.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe 12.83 ± 44% -12.8 0.00 perf-profile.calltrace.cycles-pp.do_exit.do_group_exit.get_signal.arch_do_signal_or_restart.syscall_exit_to_user_mode 12.83 ± 44% -12.8 0.00 perf-profile.calltrace.cycles-pp.do_group_exit.get_signal.arch_do_signal_or_restart.syscall_exit_to_user_mode.do_syscall_64 12.83 ± 44% -12.8 0.00 perf-profile.calltrace.cycles-pp.get_signal.arch_do_signal_or_restart.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe 12.83 ± 44% -12.8 0.00 perf-profile.calltrace.cycles-pp.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe 11.79 ± 52% -11.8 0.00 perf-profile.calltrace.cycles-pp.task_work_run.do_exit.do_group_exit.get_signal.arch_do_signal_or_restart 10.96 ± 48% -11.0 0.00 perf-profile.calltrace.cycles-pp.__fput.task_work_run.do_exit.do_group_exit.get_signal 9.79 ± 55% -9.8 0.00 perf-profile.calltrace.cycles-pp.perf_event_release_kernel.perf_release.__fput.task_work_run.do_exit 9.79 ± 55% -9.8 0.00 perf-profile.calltrace.cycles-pp.perf_release.__fput.task_work_run.do_exit.do_group_exit 9.42 ± 39% -9.4 0.00 perf-profile.calltrace.cycles-pp.asm_exc_page_fault.fault_in_readable.fault_in_iov_iter_readable.generic_perform_write.shmem_file_write_iter 6.64 ± 37% -6.6 0.00 perf-profile.calltrace.cycles-pp.event_function_call.perf_event_release_kernel.perf_release.__fput.task_work_run 6.55 ± 65% -6.5 0.00 perf-profile.calltrace.cycles-pp.asm_exc_page_fault 6.41 ± 20% -6.4 0.00 perf-profile.calltrace.cycles-pp.fault_in_iov_iter_readable.generic_perform_write.shmem_file_write_iter.vfs_write.ksys_write 6.35 ± 42% -6.4 0.00 perf-profile.calltrace.cycles-pp.smp_call_function_single.event_function_call.perf_event_release_kernel.perf_release.__fput 6.07 ± 62% -6.1 0.00 perf-profile.calltrace.cycles-pp.do_user_addr_fault.exc_page_fault.asm_exc_page_fault 6.07 ± 62% -6.1 0.00 perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault 5.99 ± 73% -6.0 0.00 perf-profile.calltrace.cycles-pp.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault 5.54 ± 55% -5.5 0.00 perf-profile.calltrace.cycles-pp.copy_page_from_iter_atomic.generic_perform_write.shmem_file_write_iter.vfs_write.ksys_write 5.54 ± 55% -5.5 0.00 perf-profile.calltrace.cycles-pp.rep_movs_alternative.copy_page_from_iter_atomic.generic_perform_write.shmem_file_write_iter.vfs_write 5.03 ± 78% -5.0 0.00 perf-profile.calltrace.cycles-pp.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault 0.00 +1.0 0.96 ± 13% perf-profile.calltrace.cycles-pp.opal_progress 0.00 +1.8 1.83 perf-profile.calltrace.cycles-pp.tzetar_.z_solve_.adi_ 0.00 +2.2 2.16 ± 4% perf-profile.calltrace.cycles-pp.rep_movs_alternative._copy_to_iter.copy_page_to_iter.process_vm_rw_single_vec.process_vm_rw_core 0.00 +2.2 2.20 ± 4% perf-profile.calltrace.cycles-pp._copy_to_iter.copy_page_to_iter.process_vm_rw_single_vec.process_vm_rw_core.process_vm_rw 0.00 +2.2 2.20 ± 4% perf-profile.calltrace.cycles-pp.copy_page_to_iter.process_vm_rw_single_vec.process_vm_rw_core.process_vm_rw.__x64_sys_process_vm_readv 0.00 +2.5 2.50 ± 4% perf-profile.calltrace.cycles-pp.process_vm_rw_single_vec.process_vm_rw_core.process_vm_rw.__x64_sys_process_vm_readv.do_syscall_64 0.00 +2.7 2.70 ± 4% perf-profile.calltrace.cycles-pp.process_vm_rw_core.process_vm_rw.__x64_sys_process_vm_readv.do_syscall_64.entry_SYSCALL_64_after_hwframe 0.00 +2.7 2.73 ± 4% perf-profile.calltrace.cycles-pp.__x64_sys_process_vm_readv.do_syscall_64.entry_SYSCALL_64_after_hwframe.process_vm_readv 0.00 +2.7 2.73 ± 4% perf-profile.calltrace.cycles-pp.process_vm_rw.__x64_sys_process_vm_readv.do_syscall_64.entry_SYSCALL_64_after_hwframe.process_vm_readv 0.00 +2.7 2.74 ± 4% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.process_vm_readv 0.00 +2.7 2.75 ± 4% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.process_vm_readv 0.00 +2.8 2.82 ± 4% perf-profile.calltrace.cycles-pp.process_vm_readv 0.00 +3.1 3.09 ± 2% perf-profile.calltrace.cycles-pp.txinvr_.adi_ 0.00 +3.2 3.16 perf-profile.calltrace.cycles-pp.copy_faces_ 0.00 +3.4 3.40 ± 3% perf-profile.calltrace.cycles-pp.add_.MAIN__ 0.00 +3.9 3.91 ± 3% perf-profile.calltrace.cycles-pp.MAIN__ 0.00 +5.5 5.52 perf-profile.calltrace.cycles-pp.lhsx_.x_solve_.adi_ 0.00 +6.2 6.25 ± 3% perf-profile.calltrace.cycles-pp.lhsy_.y_solve_.adi_ 0.00 +7.4 7.38 perf-profile.calltrace.cycles-pp.lhsz_.z_solve_.adi_ 0.00 +9.1 9.14 perf-profile.calltrace.cycles-pp.compute_rhs_.adi_ 0.00 +12.1 12.09 ± 2% perf-profile.calltrace.cycles-pp.y_solve_.adi_ 0.00 +12.4 12.42 perf-profile.calltrace.cycles-pp.x_solve_.adi_ 0.00 +14.5 14.55 perf-profile.calltrace.cycles-pp.z_solve_.adi_ 0.00 +51.6 51.58 perf-profile.calltrace.cycles-pp.adi_ 51.56 ± 15% -48.2 3.36 ± 4% perf-profile.children.cycles-pp.do_syscall_64 51.56 ± 15% -48.2 3.36 ± 4% perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe 19.20 ± 16% -19.2 0.00 perf-profile.children.cycles-pp.write 17.84 ± 19% -17.8 0.00 perf-profile.children.cycles-pp.generic_perform_write 17.84 ± 19% -17.8 0.00 perf-profile.children.cycles-pp.ksys_write 17.84 ± 19% -17.8 0.00 perf-profile.children.cycles-pp.shmem_file_write_iter 17.84 ± 19% -17.8 0.00 perf-profile.children.cycles-pp.vfs_write 16.00 ± 34% -15.9 0.12 ± 15% perf-profile.children.cycles-pp.do_exit 16.00 ± 34% -15.9 0.12 ± 15% perf-profile.children.cycles-pp.do_group_exit 13.39 ± 40% -13.4 0.00 perf-profile.children.cycles-pp.syscall_exit_to_user_mode 12.83 ± 44% -12.8 0.00 perf-profile.children.cycles-pp.arch_do_signal_or_restart 12.83 ± 44% -12.8 0.00 perf-profile.children.cycles-pp.get_signal 12.81 ± 36% -12.8 0.00 perf-profile.children.cycles-pp.__fput 12.54 ± 46% -12.5 0.02 ± 99% perf-profile.children.cycles-pp.task_work_run 12.82 ± 42% -12.4 0.38 ± 35% perf-profile.children.cycles-pp.asm_exc_page_fault 10.45 ± 34% -10.4 0.00 perf-profile.children.cycles-pp.event_function_call 10.16 ± 37% -10.2 0.00 perf-profile.children.cycles-pp.smp_call_function_single 9.79 ± 55% -9.8 0.00 perf-profile.children.cycles-pp.perf_event_release_kernel 9.79 ± 55% -9.8 0.00 perf-profile.children.cycles-pp.perf_release 6.41 ± 20% -6.4 0.00 perf-profile.children.cycles-pp.fault_in_iov_iter_readable 6.41 ± 20% -6.4 0.00 perf-profile.children.cycles-pp.fault_in_readable 5.54 ± 55% -5.5 0.00 perf-profile.children.cycles-pp.copy_page_from_iter_atomic 4.88 ± 38% -4.8 0.13 ± 14% perf-profile.children.cycles-pp.__mmput 4.61 ± 31% -4.5 0.13 ± 14% perf-profile.children.cycles-pp.exit_mmap 4.38 ± 37% -4.3 0.13 ± 5% perf-profile.children.cycles-pp.__intel_pmu_enable_all 5.54 ± 55% -3.3 2.19 ± 4% perf-profile.children.cycles-pp.rep_movs_alternative 3.18 ± 26% -3.1 0.11 ± 15% perf-profile.children.cycles-pp.exit_mm 3.18 ± 26% -3.1 0.12 ± 15% perf-profile.children.cycles-pp.__x64_sys_exit_group 0.00 +0.1 0.05 perf-profile.children.cycles-pp.irqentry_exit_to_user_mode 0.00 +0.1 0.05 perf-profile.children.cycles-pp.irqtime_account_irq 0.00 +0.1 0.05 ± 7% perf-profile.children.cycles-pp.enqueue_hrtimer 0.00 +0.1 0.05 ± 7% perf-profile.children.cycles-pp.malloc 0.00 +0.1 0.05 ± 8% perf-profile.children.cycles-pp.__update_load_avg_se 0.00 +0.1 0.05 ± 8% perf-profile.children.cycles-pp.ktime_get_update_offsets_now 0.00 +0.1 0.06 ± 9% perf-profile.children.cycles-pp.run_rebalance_domains 0.00 +0.1 0.06 ± 9% perf-profile.children.cycles-pp.timerqueue_del 0.00 +0.1 0.06 ± 11% perf-profile.children.cycles-pp.__update_load_avg_cfs_rq 0.00 +0.1 0.06 ± 6% perf-profile.children.cycles-pp.clockevents_program_event 0.00 +0.1 0.06 ± 11% perf-profile.children.cycles-pp.find_get_task_by_vpid 0.00 +0.1 0.06 ± 11% perf-profile.children.cycles-pp.runtime.selectgo 0.00 +0.1 0.06 ± 13% perf-profile.children.cycles-pp.runtime.mcall 0.00 +0.1 0.06 ± 13% perf-profile.children.cycles-pp.runtime.park_m 0.00 +0.1 0.06 ± 13% perf-profile.children.cycles-pp.runtime.schedule 0.00 +0.1 0.06 ± 11% perf-profile.children.cycles-pp.try_grab_page 0.00 +0.1 0.06 ± 14% perf-profile.children.cycles-pp.github.com/containerd/containerd/gc/scheduler.(*gcScheduler).run 0.00 +0.1 0.06 ± 14% perf-profile.children.cycles-pp.github.com/containerd/containerd/gc/scheduler.init.0.func1.1 0.00 +0.1 0.07 ± 7% perf-profile.children.cycles-pp.rcu_pending 0.00 +0.1 0.07 ± 7% perf-profile.children.cycles-pp.tick_irq_enter 0.00 +0.1 0.07 ± 7% perf-profile.children.cycles-pp.yama_ptrace_access_check 0.00 +0.1 0.07 ± 5% perf-profile.children.cycles-pp.do_poll 0.00 +0.1 0.07 ± 13% perf-profile.children.cycles-pp.update_cfs_group 0.00 +0.1 0.07 ± 9% perf-profile.children.cycles-pp.account_user_time 0.00 +0.1 0.07 ± 9% perf-profile.children.cycles-pp.pinvr_ 0.00 +0.1 0.07 ± 6% perf-profile.children.cycles-pp.do_sys_poll 0.00 +0.1 0.08 ± 6% perf-profile.children.cycles-pp.__x64_sys_poll 0.00 +0.1 0.08 ± 6% perf-profile.children.cycles-pp.irq_enter_rcu 0.00 +0.1 0.08 ± 11% perf-profile.children.cycles-pp.ninvr_ 0.00 +0.1 0.09 ± 8% perf-profile.children.cycles-pp.__poll 0.00 +0.1 0.09 ± 8% perf-profile.children.cycles-pp.rcu_sched_clock_irq 0.00 +0.1 0.09 ± 10% perf-profile.children.cycles-pp.update_rq_clock_task 0.00 +0.1 0.09 ± 7% perf-profile.children.cycles-pp.rebalance_domains 0.00 +0.1 0.10 ± 18% perf-profile.children.cycles-pp.folio_lruvec_lock_irqsave 0.00 +0.1 0.10 ± 8% perf-profile.children.cycles-pp.security_ptrace_access_check 0.00 +0.1 0.10 ± 19% perf-profile.children.cycles-pp.runtime.goexit.abi0 0.00 +0.1 0.10 ± 10% perf-profile.children.cycles-pp.tick_nohz_next_event 0.00 +0.1 0.11 ± 7% perf-profile.children.cycles-pp.ptrace_may_access 0.00 +0.1 0.12 ± 7% perf-profile.children.cycles-pp.mca_btl_vader_poll_handle_frag 0.00 +0.1 0.12 ± 11% perf-profile.children.cycles-pp.follow_page_pte 0.00 +0.1 0.12 ± 15% perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath 0.00 +0.1 0.13 ± 5% perf-profile.children.cycles-pp.mm_access 0.00 +0.1 0.13 ± 16% perf-profile.children.cycles-pp._raw_spin_lock_irqsave 0.00 +0.1 0.13 ± 8% perf-profile.children.cycles-pp.tick_nohz_get_sleep_length 0.00 +0.1 0.15 ± 6% perf-profile.children.cycles-pp.memcpy@plt 0.00 +0.2 0.16 ± 11% perf-profile.children.cycles-pp.opal_progress@plt 0.00 +0.2 0.16 ± 7% perf-profile.children.cycles-pp.ompi_request_default_wait_all 0.00 +0.2 0.19 ± 7% perf-profile.children.cycles-pp.__do_softirq 0.00 +0.2 0.20 ± 12% perf-profile.children.cycles-pp.__irqentry_text_end 0.00 +0.2 0.20 ± 2% perf-profile.children.cycles-pp.update_load_avg 0.00 +0.2 0.21 ± 7% perf-profile.children.cycles-pp.verify_ 0.00 +0.2 0.22 ± 5% perf-profile.children.cycles-pp.menu_select 0.00 +0.2 0.22 ± 3% perf-profile.children.cycles-pp.irq_exit_rcu 0.00 +0.2 0.22 ± 7% perf-profile.children.cycles-pp.__get_user_pages 0.00 +0.2 0.23 ± 7% perf-profile.children.cycles-pp.__gup_longterm_locked 0.00 +0.2 0.24 ± 5% perf-profile.children.cycles-pp.pin_user_pages_remote 0.00 +0.5 0.46 ± 80% perf-profile.children.cycles-pp.arch_call_rest_init 0.00 +0.5 0.46 ± 80% perf-profile.children.cycles-pp.rest_init 0.00 +0.5 0.46 ± 80% perf-profile.children.cycles-pp.start_kernel 0.00 +0.5 0.46 ± 80% perf-profile.children.cycles-pp.x86_64_start_kernel 0.00 +0.5 0.46 ± 80% perf-profile.children.cycles-pp.x86_64_start_reservations 0.00 +0.5 0.46 ± 11% perf-profile.children.cycles-pp.ompi_coll_libnbc_progress 0.00 +1.1 1.06 ± 13% perf-profile.children.cycles-pp.opal_progress 0.00 +1.8 1.83 ± 2% perf-profile.children.cycles-pp.tzetar_ 0.00 +2.2 2.20 ± 4% perf-profile.children.cycles-pp._copy_to_iter 0.00 +2.2 2.20 ± 4% perf-profile.children.cycles-pp.copy_page_to_iter 0.00 +2.5 2.50 ± 4% perf-profile.children.cycles-pp.process_vm_rw_single_vec 0.00 +2.7 2.70 ± 4% perf-profile.children.cycles-pp.process_vm_rw_core 0.00 +2.7 2.73 ± 4% perf-profile.children.cycles-pp.__x64_sys_process_vm_readv 0.00 +2.7 2.73 ± 4% perf-profile.children.cycles-pp.process_vm_rw 0.00 +2.8 2.82 ± 4% perf-profile.children.cycles-pp.process_vm_readv 0.00 +3.1 3.10 ± 2% perf-profile.children.cycles-pp.txinvr_ 0.00 +3.2 3.20 perf-profile.children.cycles-pp.copy_faces_ 0.00 +3.4 3.41 ± 3% perf-profile.children.cycles-pp.add_ 0.00 +3.9 3.91 ± 3% perf-profile.children.cycles-pp.MAIN__ 0.00 +5.5 5.54 perf-profile.children.cycles-pp.lhsx_ 0.00 +6.3 6.27 ± 3% perf-profile.children.cycles-pp.lhsy_ 0.00 +7.4 7.40 perf-profile.children.cycles-pp.lhsz_ 0.00 +9.4 9.38 perf-profile.children.cycles-pp.compute_rhs_ 0.00 +12.1 12.10 ± 2% perf-profile.children.cycles-pp.y_solve_ 0.00 +12.4 12.44 perf-profile.children.cycles-pp.x_solve_ 0.00 +14.6 14.56 perf-profile.children.cycles-pp.z_solve_ 0.00 +51.6 51.58 perf-profile.children.cycles-pp.adi_ 6.35 ± 42% -6.4 0.00 perf-profile.self.cycles-pp.smp_call_function_single 4.38 ± 37% -4.3 0.13 ± 5% perf-profile.self.cycles-pp.__intel_pmu_enable_all 5.14 ± 44% -3.0 2.14 ± 4% perf-profile.self.cycles-pp.rep_movs_alternative 0.00 +0.1 0.06 ± 11% perf-profile.self.cycles-pp.__update_load_avg_cfs_rq 0.00 +0.1 0.06 ± 6% perf-profile.self.cycles-pp.try_grab_page 0.00 +0.1 0.06 perf-profile.self.cycles-pp.menu_select 0.00 +0.1 0.06 ± 19% perf-profile.self.cycles-pp.scheduler_tick 0.00 +0.1 0.06 ± 11% perf-profile.self.cycles-pp.update_cfs_group 0.00 +0.1 0.07 ± 14% perf-profile.self.cycles-pp.task_tick_fair 0.00 +0.1 0.07 ± 9% perf-profile.self.cycles-pp.pinvr_ 0.00 +0.1 0.07 ± 10% perf-profile.self.cycles-pp.ninvr_ 0.00 +0.1 0.08 ± 9% perf-profile.self.cycles-pp.memcpy@plt 0.00 +0.1 0.08 ± 18% perf-profile.self.cycles-pp.opal_progress@plt 0.00 +0.1 0.10 ± 6% perf-profile.self.cycles-pp.update_load_avg 0.00 +0.1 0.10 ± 6% perf-profile.self.cycles-pp.cpuidle_enter_state 0.00 +0.1 0.12 ± 8% perf-profile.self.cycles-pp.mca_btl_vader_poll_handle_frag 0.00 +0.1 0.12 ± 15% perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath 0.00 +0.2 0.16 ± 7% perf-profile.self.cycles-pp.ompi_request_default_wait_all 0.00 +0.2 0.20 ± 12% perf-profile.self.cycles-pp.__irqentry_text_end 0.00 +0.4 0.44 ± 11% perf-profile.self.cycles-pp.ompi_coll_libnbc_progress 0.00 +0.9 0.92 ± 12% perf-profile.self.cycles-pp.opal_progress 0.00 +1.8 1.79 ± 2% perf-profile.self.cycles-pp.tzetar_ 0.00 +3.0 3.04 ± 2% perf-profile.self.cycles-pp.txinvr_ 0.00 +3.1 3.15 perf-profile.self.cycles-pp.copy_faces_ 0.00 +3.4 3.36 ± 3% perf-profile.self.cycles-pp.add_ 0.00 +5.2 5.21 perf-profile.self.cycles-pp.z_solve_ 0.00 +5.4 5.39 perf-profile.self.cycles-pp.lhsx_ 0.00 +5.6 5.62 perf-profile.self.cycles-pp.y_solve_ 0.00 +6.1 6.09 ± 3% perf-profile.self.cycles-pp.lhsy_ 0.00 +6.7 6.66 perf-profile.self.cycles-pp.x_solve_ 0.00 +7.2 7.22 perf-profile.self.cycles-pp.lhsz_ 0.00 +9.1 9.11 perf-profile.self.cycles-pp.compute_rhs_ *************************************************************************************************** lkp-skl-fpga01: 104 threads 2 sockets (Skylake) with 192G memory ========================================================================================= compiler/cpufreq_governor/kconfig/mode/nr_task/rootfs/tbox_group/test/testcase: gcc-12/performance/x86_64-rhel-8.3/process/100%/debian-12-x86_64-20240206.cgz/lkp-skl-fpga01/mmap1/will-it-scale commit: 9ef3ad0f03 ("ACPI/NUMA: Print CXL Early Discovery Table (CEDT)") 09f84589a8 ("ACPI/NUMA: Remove architecture dependent remainings") 9ef3ad0f032818c5 09f84589a8ed1162c0274a1b5b3 ---------------- --------------------------- %stddev %change %stddev \ | \ 1.967e+08 ± 12% +104.1% 4.015e+08 ± 36% cpuidle..time 3475 +13.2% 3933 ± 9% uptime.idle 117.40 +15.4% 135.50 ± 5% vmstat.procs.r 7130 -31.4% 4890 vmstat.system.cs 7005637 -38.1% 4334264 will-it-scale.104.processes 67361 -38.1% 41675 will-it-scale.per_process_ops 7005637 -38.1% 4334264 will-it-scale.workload 0.45 -0.1 0.34 mpstat.cpu.all.irq% 21.80 +24.1 45.86 mpstat.cpu.all.soft% 67.95 -21.2 46.74 mpstat.cpu.all.sys% 9.30 -3.4 5.94 mpstat.cpu.all.usr% 10720 ± 6% -50.5% 5308 ± 7% perf-c2c.DRAM.local 669.83 ± 4% +478.1% 3872 ± 6% perf-c2c.DRAM.remote 6822 ± 5% -40.8% 4037 ± 7% perf-c2c.HITM.local 421.00 ± 5% +577.2% 2850 ± 6% perf-c2c.HITM.remote 1.49e+08 -100.0% 0.00 numa-numastat.node0.local_node 1.491e+08 -100.0% 0.00 numa-numastat.node0.numa_hit 59368 ± 59% -100.0% 0.00 numa-numastat.node0.other_node 1.567e+08 -100.0% 0.00 numa-numastat.node1.local_node 1.567e+08 -100.0% 0.00 numa-numastat.node1.numa_hit 48388 ± 72% -100.0% 0.00 numa-numastat.node1.other_node 10352298 ± 7% -55.6% 4591957 ± 15% meminfo.DirectMap2M 2206149 ± 3% -21.4% 1733730 ± 5% meminfo.Inactive 2203855 ± 3% -21.4% 1731430 ± 5% meminfo.Inactive(anon) 6859527 -10.6% 6129805 meminfo.Memused 49043 -12.2% 43044 meminfo.Percpu 495530 -22.7% 382930 ± 2% meminfo.SUnreclaim 1816899 ± 4% -22.8% 1403126 ± 5% meminfo.Shmem 599136 -18.9% 485960 meminfo.Slab 257232 -84.9% 38717 meminfo.VmallocUsed 1219270 -8.5% 1115918 proc-vmstat.nr_file_pages 550889 ± 3% -21.4% 432828 ± 5% proc-vmstat.nr_inactive_anon 21408 +1.5% 21731 proc-vmstat.nr_kernel_stack 3324 -4.1% 3187 proc-vmstat.nr_page_table_pages 454144 ± 4% -22.8% 350800 ± 5% proc-vmstat.nr_shmem 125759 -21.9% 98264 proc-vmstat.nr_slab_unreclaimable 550889 ± 3% -21.4% 432828 ± 5% proc-vmstat.nr_zone_inactive_anon 55488 ± 18% -100.0% 0.00 proc-vmstat.numa_hint_faults 40437 ± 29% -100.0% 0.00 proc-vmstat.numa_hint_faults_local 3.058e+08 -28.2% 2.195e+08 proc-vmstat.numa_hit 3.057e+08 -28.2% 2.195e+08 proc-vmstat.numa_local 107756 -100.0% 0.00 proc-vmstat.numa_other 18231 ± 96% -100.0% 0.00 proc-vmstat.numa_pages_migrated 182874 ± 22% -100.0% 0.00 proc-vmstat.numa_pte_updates 90886 +122.4% 202099 ± 73% proc-vmstat.pgactivate 5.952e+08 -28.2% 4.271e+08 proc-vmstat.pgalloc_normal 1042723 -39.2% 634241 proc-vmstat.pgfault 5.943e+08 -28.2% 4.265e+08 proc-vmstat.pgfree 18231 ± 96% -100.0% 0.00 proc-vmstat.pgmigrate_success 44763 ± 6% -27.1% 32626 proc-vmstat.pgreuse 2.02 +12.4% 2.27 perf-stat.i.MPKI 2.933e+10 -36.2% 1.871e+10 perf-stat.i.branch-instructions 0.82 -0.0 0.80 perf-stat.i.branch-miss-rate% 2.382e+08 -37.7% 1.485e+08 perf-stat.i.branch-misses 62.01 +12.4 74.38 perf-stat.i.cache-miss-rate% 3.092e+08 -28.9% 2.199e+08 perf-stat.i.cache-misses 4.978e+08 -40.9% 2.943e+08 perf-stat.i.cache-references 7111 -32.2% 4822 perf-stat.i.context-switches 1.83 +59.0% 2.91 perf-stat.i.cpi 217.02 +39.3% 302.23 perf-stat.i.cpu-migrations 905.82 +41.5% 1281 perf-stat.i.cycles-between-cache-misses 1.528e+11 -36.7% 9.67e+10 perf-stat.i.instructions 0.55 -36.9% 0.35 perf-stat.i.ipc 0.11 ± 29% -51.5% 0.05 ± 37% perf-stat.i.major-faults 3266 -40.8% 1932 perf-stat.i.minor-faults 3266 -40.8% 1932 perf-stat.i.page-faults 2.02 +12.4% 2.27 perf-stat.overall.MPKI 0.81 -0.0 0.79 perf-stat.overall.branch-miss-rate% 62.11 +12.6 74.69 perf-stat.overall.cache-miss-rate% 1.83 +58.9% 2.91 perf-stat.overall.cpi 905.85 +41.4% 1280 perf-stat.overall.cycles-between-cache-misses 0.55 -37.1% 0.34 perf-stat.overall.ipc 6581748 +3.8% 6830787 perf-stat.overall.path-length 2.923e+10 -36.2% 1.865e+10 perf-stat.ps.branch-instructions 2.373e+08 -37.7% 1.479e+08 perf-stat.ps.branch-misses 3.082e+08 -28.9% 2.192e+08 perf-stat.ps.cache-misses 4.962e+08 -40.9% 2.934e+08 perf-stat.ps.cache-references 7083 -32.2% 4802 perf-stat.ps.context-switches 215.95 +39.3% 300.81 perf-stat.ps.cpu-migrations 1.523e+11 -36.7% 9.636e+10 perf-stat.ps.instructions 0.11 ± 29% -51.1% 0.05 ± 37% perf-stat.ps.major-faults 3244 -41.2% 1908 perf-stat.ps.minor-faults 3245 -41.2% 1908 perf-stat.ps.page-faults 4.611e+13 -35.8% 2.961e+13 perf-stat.total.instructions 13971 ± 89% -100.0% 0.00 numa-vmstat.node0.nr_active_anon 97064 ± 52% -100.0% 0.00 numa-vmstat.node0.nr_anon_pages 551160 ± 47% -100.0% 0.00 numa-vmstat.node0.nr_file_pages 23600699 -100.0% 0.00 numa-vmstat.node0.nr_free_pages 97975 ± 52% -100.0% 0.00 numa-vmstat.node0.nr_inactive_anon 11699 ± 4% -100.0% 0.00 numa-vmstat.node0.nr_kernel_stack 14296 ± 11% -100.0% 0.00 numa-vmstat.node0.nr_mapped 1544 ± 6% -100.0% 0.00 numa-vmstat.node0.nr_page_table_pages 15651 ± 83% -100.0% 0.00 numa-vmstat.node0.nr_shmem 15308 ± 24% -100.0% 0.00 numa-vmstat.node0.nr_slab_reclaimable 67651 ± 3% -100.0% 0.00 numa-vmstat.node0.nr_slab_unreclaimable 536155 ± 47% -100.0% 0.00 numa-vmstat.node0.nr_unevictable 13971 ± 89% -100.0% 0.00 numa-vmstat.node0.nr_zone_active_anon 97974 ± 52% -100.0% 0.00 numa-vmstat.node0.nr_zone_inactive_anon 536155 ± 47% -100.0% 0.00 numa-vmstat.node0.nr_zone_unevictable 1.491e+08 -100.0% 0.00 numa-vmstat.node0.numa_hit 1.49e+08 -100.0% 0.00 numa-vmstat.node0.numa_local 59368 ± 59% -100.0% 0.00 numa-vmstat.node0.numa_other 67762 ± 21% -100.0% 0.00 numa-vmstat.node1.nr_active_anon 82222 ± 62% -100.0% 0.00 numa-vmstat.node1.nr_anon_pages 668044 ± 37% -100.0% 0.00 numa-vmstat.node1.nr_file_pages 23857326 -100.0% 0.00 numa-vmstat.node1.nr_free_pages 452882 ± 9% -100.0% 0.00 numa-vmstat.node1.nr_inactive_anon 9709 ± 5% -100.0% 0.00 numa-vmstat.node1.nr_kernel_stack 148558 ± 4% -100.0% 0.00 numa-vmstat.node1.nr_mapped 1779 ± 5% -100.0% 0.00 numa-vmstat.node1.nr_page_table_pages 438426 ± 5% -100.0% 0.00 numa-vmstat.node1.nr_shmem 10592 ± 35% -100.0% 0.00 numa-vmstat.node1.nr_slab_reclaimable 61563 ± 4% -100.0% 0.00 numa-vmstat.node1.nr_slab_unreclaimable 229017 ±111% -100.0% 0.00 numa-vmstat.node1.nr_unevictable 67762 ± 21% -100.0% 0.00 numa-vmstat.node1.nr_zone_active_anon 452882 ± 9% -100.0% 0.00 numa-vmstat.node1.nr_zone_inactive_anon 229017 ±111% -100.0% 0.00 numa-vmstat.node1.nr_zone_unevictable 1.567e+08 -100.0% 0.00 numa-vmstat.node1.numa_hit 1.566e+08 -100.0% 0.00 numa-vmstat.node1.numa_local 48388 ± 72% -100.0% 0.00 numa-vmstat.node1.numa_other 55900 ± 89% -100.0% 0.00 numa-meminfo.node0.Active 55801 ± 89% -100.0% 0.00 numa-meminfo.node0.Active(anon) 388254 ± 52% -100.0% 0.00 numa-meminfo.node0.AnonPages 440142 ± 48% -100.0% 0.00 numa-meminfo.node0.AnonPages.max 2204609 ± 47% -100.0% 0.00 numa-meminfo.node0.FilePages 392333 ± 52% -100.0% 0.00 numa-meminfo.node0.Inactive 391951 ± 52% -100.0% 0.00 numa-meminfo.node0.Inactive(anon) 61236 ± 24% -100.0% 0.00 numa-meminfo.node0.KReclaimable 11699 ± 4% -100.0% 0.00 numa-meminfo.node0.KernelStack 56750 ± 12% -100.0% 0.00 numa-meminfo.node0.Mapped 94405300 -100.0% 0.00 numa-meminfo.node0.MemFree 97679448 -100.0% 0.00 numa-meminfo.node0.MemTotal 3274147 ± 37% -100.0% 0.00 numa-meminfo.node0.MemUsed 6174 ± 6% -100.0% 0.00 numa-meminfo.node0.PageTables 61236 ± 24% -100.0% 0.00 numa-meminfo.node0.SReclaimable 268262 ± 4% -100.0% 0.00 numa-meminfo.node0.SUnreclaim 62576 ± 83% -100.0% 0.00 numa-meminfo.node0.Shmem 329498 ± 7% -100.0% 0.00 numa-meminfo.node0.Slab 2144623 ± 47% -100.0% 0.00 numa-meminfo.node0.Unevictable 271268 ± 21% -100.0% 0.00 numa-meminfo.node1.Active 270777 ± 21% -100.0% 0.00 numa-meminfo.node1.Active(anon) 328899 ± 62% -100.0% 0.00 numa-meminfo.node1.AnonPages 368362 ± 54% -100.0% 0.00 numa-meminfo.node1.AnonPages.max 2672717 ± 37% -100.0% 0.00 numa-meminfo.node1.FilePages 1814269 ± 9% -100.0% 0.00 numa-meminfo.node1.Inactive 1812357 ± 9% -100.0% 0.00 numa-meminfo.node1.Inactive(anon) 42376 ± 35% -100.0% 0.00 numa-meminfo.node1.KReclaimable 9709 ± 5% -100.0% 0.00 numa-meminfo.node1.KernelStack 594500 ± 4% -100.0% 0.00 numa-meminfo.node1.Mapped 95440412 -100.0% 0.00 numa-meminfo.node1.MemFree 99034620 -100.0% 0.00 numa-meminfo.node1.MemTotal 3594207 ± 32% -100.0% 0.00 numa-meminfo.node1.MemUsed 7121 ± 5% -100.0% 0.00 numa-meminfo.node1.PageTables 42376 ± 35% -100.0% 0.00 numa-meminfo.node1.SReclaimable 234056 ± 3% -100.0% 0.00 numa-meminfo.node1.SUnreclaim 1754246 ± 5% -100.0% 0.00 numa-meminfo.node1.Shmem 276432 ± 7% -100.0% 0.00 numa-meminfo.node1.Slab 916068 ±111% -100.0% 0.00 numa-meminfo.node1.Unevictable 12846621 -31.1% 8851378 sched_debug.cfs_rq:/.avg_vruntime.avg 13026668 -29.8% 9146405 sched_debug.cfs_rq:/.avg_vruntime.max 12010831 -34.2% 7905615 sched_debug.cfs_rq:/.avg_vruntime.min 172319 ± 4% +46.4% 252250 ± 4% sched_debug.cfs_rq:/.avg_vruntime.stddev 0.98 +15.9% 1.14 ± 4% sched_debug.cfs_rq:/.h_nr_running.avg 0.38 ± 4% +25.9% 0.48 ± 5% sched_debug.cfs_rq:/.h_nr_running.stddev 1482184 ± 18% +81.3% 2687007 ± 18% sched_debug.cfs_rq:/.left_deadline.avg 12984350 -29.9% 9098475 sched_debug.cfs_rq:/.left_deadline.max 1482165 ± 18% +81.3% 2686955 ± 18% sched_debug.cfs_rq:/.left_vruntime.avg 12984176 -29.9% 9098312 sched_debug.cfs_rq:/.left_vruntime.max 110018 ± 15% +151.7% 276928 ± 20% sched_debug.cfs_rq:/.load.avg 279353 ± 7% +39.9% 390900 ± 5% sched_debug.cfs_rq:/.load.stddev 153.45 ± 7% +165.5% 407.41 ± 5% sched_debug.cfs_rq:/.load_avg.avg 465.14 ± 19% +59.0% 739.61 ± 3% sched_debug.cfs_rq:/.load_avg.max 67.81 ± 5% +162.1% 177.72 ± 15% sched_debug.cfs_rq:/.load_avg.min 75.29 ± 10% +48.6% 111.88 ± 7% sched_debug.cfs_rq:/.load_avg.stddev 12846638 -31.1% 8851412 sched_debug.cfs_rq:/.min_vruntime.avg 13026687 -29.8% 9146424 sched_debug.cfs_rq:/.min_vruntime.max 12010851 -34.2% 7905665 sched_debug.cfs_rq:/.min_vruntime.min 172320 ± 4% +46.4% 252248 ± 4% sched_debug.cfs_rq:/.min_vruntime.stddev 0.95 +15.5% 1.10 ± 4% sched_debug.cfs_rq:/.nr_running.avg 0.32 ± 5% +31.7% 0.42 ± 4% sched_debug.cfs_rq:/.nr_running.stddev 1482165 ± 18% +81.3% 2686956 ± 18% sched_debug.cfs_rq:/.right_vruntime.avg 12984176 -29.9% 9098312 sched_debug.cfs_rq:/.right_vruntime.max 1047 +24.4% 1303 sched_debug.cfs_rq:/.runnable_avg.avg 2000 +20.3% 2405 ± 2% sched_debug.cfs_rq:/.runnable_avg.max 159.06 ± 4% +39.3% 221.64 ± 6% sched_debug.cfs_rq:/.runnable_avg.stddev 1151 ± 3% +15.3% 1328 ± 3% sched_debug.cfs_rq:/.util_avg.max 78.12 ± 15% +34.9% 105.36 ± 8% sched_debug.cfs_rq:/.util_avg.stddev 734.30 ± 4% -10.1% 659.87 ± 3% sched_debug.cfs_rq:/.util_est.avg 159.60 ± 9% +79.3% 286.17 ± 5% sched_debug.cfs_rq:/.util_est.stddev 18.37 ± 4% +22.8% 22.56 ± 4% sched_debug.cpu.clock.stddev 3171 ± 2% -18.0% 2599 ± 7% sched_debug.cpu.curr->pid.avg 7419 ± 2% -24.5% 5599 ± 3% sched_debug.cpu.curr->pid.max 1202 ± 6% +24.3% 1494 ± 4% sched_debug.cpu.curr->pid.stddev 0.00 ± 2% +12.6% 0.00 sched_debug.cpu.next_balance.stddev 0.98 +15.9% 1.14 ± 4% sched_debug.cpu.nr_running.avg 0.38 ± 4% +25.4% 0.47 ± 5% sched_debug.cpu.nr_running.stddev 11316 -28.4% 8107 sched_debug.cpu.nr_switches.avg 8294 -35.9% 5317 sched_debug.cpu.nr_switches.min 2590 ± 6% +20.3% 3115 ± 7% sched_debug.cpu.nr_switches.stddev 48.09 -14.2 33.91 perf-profile.calltrace.cycles-pp.__munmap 36.16 -13.8 22.35 perf-profile.calltrace.cycles-pp.__mmap 42.68 -12.0 30.66 perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.__munmap 30.76 -11.6 19.12 perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.__mmap 41.53 -11.6 29.96 perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.__munmap 40.88 -11.3 29.57 perf-profile.calltrace.cycles-pp.__x64_sys_munmap.do_syscall_64.entry_SYSCALL_64_after_hwframe.__munmap 40.76 -11.3 29.50 perf-profile.calltrace.cycles-pp.__vm_munmap.__x64_sys_munmap.do_syscall_64.entry_SYSCALL_64_after_hwframe.__munmap 29.62 -11.2 18.42 perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.__mmap 40.13 -11.0 29.13 perf-profile.calltrace.cycles-pp.do_vmi_munmap.__vm_munmap.__x64_sys_munmap.do_syscall_64.entry_SYSCALL_64_after_hwframe 28.85 -10.9 17.98 perf-profile.calltrace.cycles-pp.vm_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe.__mmap 39.44 -10.7 28.72 perf-profile.calltrace.cycles-pp.do_vmi_align_munmap.do_vmi_munmap.__vm_munmap.__x64_sys_munmap.do_syscall_64 27.54 -10.3 17.21 perf-profile.calltrace.cycles-pp.do_mmap.vm_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe.__mmap 23.80 -8.8 14.98 perf-profile.calltrace.cycles-pp.mmap_region.do_mmap.vm_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe 12.41 -5.1 7.34 perf-profile.calltrace.cycles-pp.unmap_region.do_vmi_align_munmap.do_vmi_munmap.__vm_munmap.__x64_sys_munmap 11.64 -4.5 7.10 perf-profile.calltrace.cycles-pp.mas_store_prealloc.mmap_region.do_mmap.vm_mmap_pgoff.do_syscall_64 23.99 -4.4 19.62 perf-profile.calltrace.cycles-pp.mas_store_gfp.do_vmi_align_munmap.do_vmi_munmap.__vm_munmap.__x64_sys_munmap 10.31 -4.1 6.25 perf-profile.calltrace.cycles-pp.mas_wr_bnode.mas_store_prealloc.mmap_region.do_mmap.vm_mmap_pgoff 9.12 -3.7 5.39 perf-profile.calltrace.cycles-pp.unmap_vmas.unmap_region.do_vmi_align_munmap.do_vmi_munmap.__vm_munmap 8.76 -3.6 5.17 perf-profile.calltrace.cycles-pp.unmap_page_range.unmap_vmas.unmap_region.do_vmi_align_munmap.do_vmi_munmap 8.65 -3.4 5.29 perf-profile.calltrace.cycles-pp.mas_split.mas_wr_bnode.mas_store_prealloc.mmap_region.do_mmap 18.03 -3.3 14.74 perf-profile.calltrace.cycles-pp.mas_wr_spanning_store.mas_store_gfp.do_vmi_align_munmap.do_vmi_munmap.__vm_munmap 7.98 -3.3 4.71 perf-profile.calltrace.cycles-pp.zap_pmd_range.unmap_page_range.unmap_vmas.unmap_region.do_vmi_align_munmap 6.44 -2.4 4.00 perf-profile.calltrace.cycles-pp.mas_spanning_rebalance.mas_wr_spanning_store.mas_store_gfp.do_vmi_align_munmap.do_vmi_munmap 3.32 -1.4 1.95 perf-profile.calltrace.cycles-pp.__cond_resched.zap_pmd_range.unmap_page_range.unmap_vmas.unmap_region 3.35 -1.4 2.00 perf-profile.calltrace.cycles-pp.get_unmapped_area.do_mmap.vm_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe 2.98 -1.2 1.78 perf-profile.calltrace.cycles-pp.thp_get_unmapped_area.get_unmapped_area.do_mmap.vm_mmap_pgoff.do_syscall_64 2.75 -1.1 1.64 perf-profile.calltrace.cycles-pp.arch_get_unmapped_area_topdown.thp_get_unmapped_area.get_unmapped_area.do_mmap.vm_mmap_pgoff 2.60 -1.0 1.55 perf-profile.calltrace.cycles-pp.perf_event_mmap.mmap_region.do_mmap.vm_mmap_pgoff.do_syscall_64 3.00 -1.0 1.96 perf-profile.calltrace.cycles-pp.vm_area_alloc.mmap_region.do_mmap.vm_mmap_pgoff.do_syscall_64 2.35 -1.0 1.37 perf-profile.calltrace.cycles-pp.mas_topiary_replace.mas_spanning_rebalance.mas_wr_spanning_store.mas_store_gfp.do_vmi_align_munmap 2.31 -0.9 1.37 perf-profile.calltrace.cycles-pp.vm_unmapped_area.arch_get_unmapped_area_topdown.thp_get_unmapped_area.get_unmapped_area.do_mmap 2.29 -0.9 1.37 perf-profile.calltrace.cycles-pp.perf_event_mmap_event.perf_event_mmap.mmap_region.do_mmap.vm_mmap_pgoff 2.16 -0.9 1.24 perf-profile.calltrace.cycles-pp.mas_topiary_replace.mas_split.mas_wr_bnode.mas_store_prealloc.mmap_region 2.54 -0.9 1.66 perf-profile.calltrace.cycles-pp.kmem_cache_alloc.vm_area_alloc.mmap_region.do_mmap.vm_mmap_pgoff 2.01 -0.8 1.19 perf-profile.calltrace.cycles-pp.free_pgd_range.unmap_region.do_vmi_align_munmap.do_vmi_munmap.__vm_munmap 2.04 -0.8 1.25 perf-profile.calltrace.cycles-pp.syscall_return_via_sysret.__mmap 2.05 -0.8 1.25 perf-profile.calltrace.cycles-pp.syscall_return_via_sysret.__munmap 5.22 -0.8 4.43 ± 3% perf-profile.calltrace.cycles-pp.mas_nomem.mas_store_gfp.do_vmi_align_munmap.do_vmi_munmap.__vm_munmap 1.83 -0.7 1.09 perf-profile.calltrace.cycles-pp.free_p4d_range.free_pgd_range.unmap_region.do_vmi_align_munmap.do_vmi_munmap 2.99 -0.7 2.25 perf-profile.calltrace.cycles-pp.mas_preallocate.mmap_region.do_mmap.vm_mmap_pgoff.do_syscall_64 5.12 -0.7 4.38 ± 3% perf-profile.calltrace.cycles-pp.mas_destroy.mas_nomem.mas_store_gfp.do_vmi_align_munmap.do_vmi_munmap 0.99 ± 19% -0.7 0.26 ±100% perf-profile.calltrace.cycles-pp.__do_softirq.irq_exit_rcu.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.__put_partials 1.60 -0.7 0.94 perf-profile.calltrace.cycles-pp.mast_fill_bnode.mas_split.mas_wr_bnode.mas_store_prealloc.mmap_region 1.58 -0.6 0.94 perf-profile.calltrace.cycles-pp.free_pud_range.free_p4d_range.free_pgd_range.unmap_region.do_vmi_align_munmap 1.65 -0.6 1.02 perf-profile.calltrace.cycles-pp.entry_SYSRETQ_unsafe_stack.__mmap 1.65 -0.6 1.03 perf-profile.calltrace.cycles-pp.entry_SYSRETQ_unsafe_stack.__munmap 1.53 -0.6 0.92 perf-profile.calltrace.cycles-pp.perf_iterate_sb.perf_event_mmap_event.perf_event_mmap.mmap_region.do_mmap 1.38 -0.6 0.82 perf-profile.calltrace.cycles-pp.mas_empty_area_rev.vm_unmapped_area.arch_get_unmapped_area_topdown.thp_get_unmapped_area.get_unmapped_area 2.42 -0.5 1.91 ± 2% perf-profile.calltrace.cycles-pp.mas_alloc_nodes.mas_preallocate.mmap_region.do_mmap.vm_mmap_pgoff 1.11 -0.5 0.64 perf-profile.calltrace.cycles-pp.mast_split_data.mas_split.mas_wr_bnode.mas_store_prealloc.mmap_region 0.88 -0.5 0.42 ± 44% perf-profile.calltrace.cycles-pp.mab_mas_cp.mast_split_data.mas_split.mas_wr_bnode.mas_store_prealloc 2.64 ± 2% -0.4 2.21 ± 3% perf-profile.calltrace.cycles-pp.kmem_cache_free_bulk.mas_destroy.mas_nomem.mas_store_gfp.do_vmi_align_munmap 1.04 -0.4 0.61 perf-profile.calltrace.cycles-pp.mas_rev_awalk.mas_empty_area_rev.vm_unmapped_area.arch_get_unmapped_area_topdown.thp_get_unmapped_area 1.07 -0.4 0.66 perf-profile.calltrace.cycles-pp.perf_event_mmap_output.perf_iterate_sb.perf_event_mmap_event.perf_event_mmap.mmap_region 0.94 -0.4 0.54 perf-profile.calltrace.cycles-pp.mas_store_b_node.mas_wr_bnode.mas_store_prealloc.mmap_region.do_mmap 0.86 -0.4 0.51 perf-profile.calltrace.cycles-pp.do_vmi_munmap.mmap_region.do_mmap.vm_mmap_pgoff.do_syscall_64 1.29 -0.3 0.97 perf-profile.calltrace.cycles-pp.mab_mas_cp.mas_spanning_rebalance.mas_wr_spanning_store.mas_store_gfp.do_vmi_align_munmap 0.97 -0.3 0.66 ± 2% perf-profile.calltrace.cycles-pp.mas_split_final_node.mas_split.mas_wr_bnode.mas_store_prealloc.mmap_region 0.55 -0.3 0.26 ±100% perf-profile.calltrace.cycles-pp.asm_sysvec_apic_timer_interrupt.__put_partials.kmem_cache_free.mas_destroy.mas_nomem 0.55 -0.3 0.26 ±100% perf-profile.calltrace.cycles-pp.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.__put_partials.kmem_cache_free.mas_destroy 0.55 -0.3 0.26 ±100% perf-profile.calltrace.cycles-pp.irq_exit_rcu.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.__put_partials.kmem_cache_free 0.85 -0.3 0.56 perf-profile.calltrace.cycles-pp.mas_push_data.mas_split.mas_wr_bnode.mas_store_prealloc.mmap_region 0.82 -0.3 0.56 ± 2% perf-profile.calltrace.cycles-pp.mas_destroy.mas_store_prealloc.mmap_region.do_mmap.vm_mmap_pgoff 0.69 -0.3 0.43 ± 44% perf-profile.calltrace.cycles-pp.mas_pop_node.mas_split.mas_wr_bnode.mas_store_prealloc.mmap_region 1.82 -0.3 1.56 ± 2% perf-profile.calltrace.cycles-pp.kmem_cache_alloc_bulk.mas_alloc_nodes.mas_preallocate.mmap_region.do_mmap 1.77 ± 2% -0.2 1.58 ± 4% perf-profile.calltrace.cycles-pp.kmem_cache_free.mas_destroy.mas_nomem.mas_store_gfp.do_vmi_align_munmap 1.52 ± 2% -0.1 1.39 ± 3% perf-profile.calltrace.cycles-pp.__kmem_cache_alloc_bulk.kmem_cache_alloc_bulk.mas_alloc_nodes.mas_preallocate.mmap_region 0.62 -0.0 0.59 ± 2% perf-profile.calltrace.cycles-pp.kmem_cache_free.vm_area_free_rcu_cb.rcu_do_batch.rcu_core.__do_softirq 0.92 +0.0 0.96 ± 2% perf-profile.calltrace.cycles-pp.allocate_slab.___slab_alloc.__kmem_cache_alloc_bulk.kmem_cache_alloc_bulk.mas_alloc_nodes 0.85 ± 4% +0.2 1.01 ± 2% perf-profile.calltrace.cycles-pp.rcu_cblist_dequeue.rcu_do_batch.rcu_core.__do_softirq.run_ksoftirqd 1.44 ± 2% +0.2 1.68 ± 2% perf-profile.calltrace.cycles-pp.__do_softirq.irq_exit_rcu.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.get_partial_node 1.44 ± 2% +0.2 1.69 ± 3% perf-profile.calltrace.cycles-pp.irq_exit_rcu.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.get_partial_node.___slab_alloc 1.44 ± 2% +0.3 1.70 ± 2% perf-profile.calltrace.cycles-pp.asm_sysvec_apic_timer_interrupt.get_partial_node.___slab_alloc.__kmem_cache_alloc_bulk.kmem_cache_alloc_bulk 1.44 ± 2% +0.3 1.70 ± 2% perf-profile.calltrace.cycles-pp.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.get_partial_node.___slab_alloc.__kmem_cache_alloc_bulk 1.24 ± 2% +0.3 1.59 ± 2% perf-profile.calltrace.cycles-pp.kmem_cache_free.rcu_do_batch.rcu_core.__do_softirq.irq_exit_rcu 8.54 +0.4 8.89 ± 2% perf-profile.calltrace.cycles-pp.mas_alloc_nodes.mas_wr_spanning_store.mas_store_gfp.do_vmi_align_munmap.do_vmi_munmap 8.10 +0.5 8.64 ± 2% perf-profile.calltrace.cycles-pp.kmem_cache_alloc_bulk.mas_alloc_nodes.mas_wr_spanning_store.mas_store_gfp.do_vmi_align_munmap 7.92 +0.6 8.53 ± 2% perf-profile.calltrace.cycles-pp.__kmem_cache_alloc_bulk.kmem_cache_alloc_bulk.mas_alloc_nodes.mas_wr_spanning_store.mas_store_gfp 7.30 +0.9 8.16 ± 2% perf-profile.calltrace.cycles-pp.___slab_alloc.__kmem_cache_alloc_bulk.kmem_cache_alloc_bulk.mas_alloc_nodes.mas_wr_spanning_store 1.85 +1.1 2.93 ± 4% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.__put_partials.kmem_cache_free.rcu_do_batch 1.88 +1.1 2.96 ± 4% perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.__put_partials.kmem_cache_free.rcu_do_batch.rcu_core 3.44 ± 2% +2.1 5.51 ± 4% perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.get_partial_node.___slab_alloc.__kmem_cache_alloc_bulk.kmem_cache_alloc_bulk 3.40 ± 2% +2.1 5.49 ± 4% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.get_partial_node.___slab_alloc.__kmem_cache_alloc_bulk 5.40 ± 2% +2.3 7.66 ± 3% perf-profile.calltrace.cycles-pp.get_partial_node.___slab_alloc.__kmem_cache_alloc_bulk.kmem_cache_alloc_bulk.mas_alloc_nodes 6.49 +3.1 9.58 ± 5% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.__slab_free.kmem_cache_free.rcu_do_batch 9.25 +3.1 12.33 ± 3% perf-profile.calltrace.cycles-pp.__slab_free.kmem_cache_free.rcu_do_batch.rcu_core.__do_softirq 6.61 +3.1 9.72 ± 5% perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.__slab_free.kmem_cache_free.rcu_do_batch.rcu_core 0.65 ± 8% +10.4 11.04 ± 3% perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.free_pcppages_bulk.free_unref_page_commit.free_unref_page.kmem_cache_free 0.70 ± 8% +10.4 11.12 ± 3% perf-profile.calltrace.cycles-pp.free_pcppages_bulk.free_unref_page_commit.free_unref_page.kmem_cache_free.rcu_do_batch 0.73 ± 7% +10.5 11.18 ± 3% perf-profile.calltrace.cycles-pp.free_unref_page.kmem_cache_free.rcu_do_batch.rcu_core.__do_softirq 0.71 ± 8% +10.5 11.16 ± 3% perf-profile.calltrace.cycles-pp.free_unref_page_commit.free_unref_page.kmem_cache_free.rcu_do_batch.rcu_core 0.84 ± 7% +13.3 14.15 ± 3% perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.free_pcppages_bulk.free_unref_page_commit.free_unref_page.__put_partials 0.91 ± 6% +13.4 14.26 ± 3% perf-profile.calltrace.cycles-pp.free_pcppages_bulk.free_unref_page_commit.free_unref_page.__put_partials.kmem_cache_free 0.92 ± 6% +13.4 14.32 ± 3% perf-profile.calltrace.cycles-pp.free_unref_page_commit.free_unref_page.__put_partials.kmem_cache_free.rcu_do_batch 0.94 ± 6% +13.4 14.34 ± 3% perf-profile.calltrace.cycles-pp.free_unref_page.__put_partials.kmem_cache_free.rcu_do_batch.rcu_core 2.95 +15.1 18.08 perf-profile.calltrace.cycles-pp.__put_partials.kmem_cache_free.rcu_do_batch.rcu_core.__do_softirq 1.45 ± 7% +23.6 25.09 ± 3% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.free_pcppages_bulk.free_unref_page_commit.free_unref_page 13.04 +28.0 40.99 perf-profile.calltrace.cycles-pp.kmem_cache_free.rcu_do_batch.rcu_core.__do_softirq.run_ksoftirqd 15.02 +28.1 43.09 perf-profile.calltrace.cycles-pp.kthread.ret_from_fork.ret_from_fork_asm 15.02 +28.1 43.09 perf-profile.calltrace.cycles-pp.ret_from_fork.ret_from_fork_asm 15.02 +28.1 43.09 perf-profile.calltrace.cycles-pp.ret_from_fork_asm 14.98 +28.1 43.06 perf-profile.calltrace.cycles-pp.rcu_core.__do_softirq.run_ksoftirqd.smpboot_thread_fn.kthread 14.96 +28.1 43.04 perf-profile.calltrace.cycles-pp.rcu_do_batch.rcu_core.__do_softirq.run_ksoftirqd.smpboot_thread_fn 14.98 +28.1 43.06 perf-profile.calltrace.cycles-pp.__do_softirq.run_ksoftirqd.smpboot_thread_fn.kthread.ret_from_fork 14.98 +28.1 43.06 perf-profile.calltrace.cycles-pp.run_ksoftirqd.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm 15.00 +28.1 43.08 perf-profile.calltrace.cycles-pp.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm 73.58 -23.7 49.89 perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe 71.25 -22.8 48.46 perf-profile.children.cycles-pp.do_syscall_64 48.23 -14.2 34.00 perf-profile.children.cycles-pp.__munmap 36.32 -13.9 22.44 perf-profile.children.cycles-pp.__mmap 41.03 -11.4 29.66 perf-profile.children.cycles-pp.do_vmi_munmap 40.90 -11.3 29.58 perf-profile.children.cycles-pp.__x64_sys_munmap 40.79 -11.3 29.52 perf-profile.children.cycles-pp.__vm_munmap 28.87 -10.9 18.00 perf-profile.children.cycles-pp.vm_mmap_pgoff 39.53 -10.8 28.78 perf-profile.children.cycles-pp.do_vmi_align_munmap 27.55 -10.3 17.22 perf-profile.children.cycles-pp.do_mmap 23.89 -8.9 15.04 perf-profile.children.cycles-pp.mmap_region 12.45 -5.1 7.37 perf-profile.children.cycles-pp.unmap_region 11.65 -4.5 7.11 perf-profile.children.cycles-pp.mas_store_prealloc 24.04 -4.4 19.64 perf-profile.children.cycles-pp.mas_store_gfp 10.32 -4.1 6.26 perf-profile.children.cycles-pp.mas_wr_bnode 10.22 -4.0 6.24 ± 2% perf-profile.children.cycles-pp.sysvec_apic_timer_interrupt 10.25 -4.0 6.28 perf-profile.children.cycles-pp.asm_sysvec_apic_timer_interrupt 9.84 -3.9 5.93 ± 2% perf-profile.children.cycles-pp.irq_exit_rcu 9.14 -3.7 5.39 perf-profile.children.cycles-pp.unmap_vmas 8.78 -3.6 5.18 perf-profile.children.cycles-pp.unmap_page_range 8.73 -3.4 5.34 perf-profile.children.cycles-pp.mas_split 8.20 -3.4 4.84 perf-profile.children.cycles-pp.zap_pmd_range 18.05 -3.3 14.75 perf-profile.children.cycles-pp.mas_wr_spanning_store 6.49 -2.4 4.04 perf-profile.children.cycles-pp.mas_spanning_rebalance 4.54 -1.9 2.62 perf-profile.children.cycles-pp.mas_topiary_replace 4.66 -1.9 2.75 perf-profile.children.cycles-pp.__cond_resched 4.13 -1.6 2.52 perf-profile.children.cycles-pp.syscall_return_via_sysret 3.37 -1.4 2.01 perf-profile.children.cycles-pp.get_unmapped_area 3.54 -1.4 2.19 perf-profile.children.cycles-pp.entry_SYSRETQ_unsafe_stack 1.34 ± 3% -1.3 0.07 ± 5% perf-profile.children.cycles-pp.get_any_partial 2.98 -1.2 1.78 perf-profile.children.cycles-pp.thp_get_unmapped_area 3.35 -1.2 2.16 perf-profile.children.cycles-pp.kmem_cache_alloc 2.77 -1.1 1.65 perf-profile.children.cycles-pp.arch_get_unmapped_area_topdown 2.60 -1.0 1.56 perf-profile.children.cycles-pp.perf_event_mmap 3.01 -1.0 1.96 perf-profile.children.cycles-pp.vm_area_alloc 5.97 -1.0 4.94 ± 3% perf-profile.children.cycles-pp.mas_destroy 2.32 -0.9 1.38 perf-profile.children.cycles-pp.vm_unmapped_area 2.33 -0.9 1.39 perf-profile.children.cycles-pp.perf_event_mmap_event 2.13 -0.9 1.26 perf-profile.children.cycles-pp.mas_find 2.71 -0.8 1.88 perf-profile.children.cycles-pp.mab_mas_cp 2.02 -0.8 1.20 perf-profile.children.cycles-pp.free_pgd_range 1.94 -0.8 1.14 perf-profile.children.cycles-pp.mas_mab_cp 5.24 -0.8 4.44 ± 3% perf-profile.children.cycles-pp.mas_nomem 1.66 -0.8 0.90 perf-profile.children.cycles-pp.__call_rcu_common 3.02 -0.8 2.26 perf-profile.children.cycles-pp.mas_preallocate 1.84 -0.7 1.10 perf-profile.children.cycles-pp.free_p4d_range 1.87 -0.7 1.13 perf-profile.children.cycles-pp.entry_SYSCALL_64 1.61 -0.7 0.95 perf-profile.children.cycles-pp.mast_fill_bnode 1.59 -0.6 0.94 perf-profile.children.cycles-pp.free_pud_range 1.54 -0.6 0.92 perf-profile.children.cycles-pp.perf_iterate_sb 1.41 -0.6 0.82 perf-profile.children.cycles-pp.mas_store_b_node 1.42 -0.6 0.84 perf-profile.children.cycles-pp.mas_empty_area_rev 1.24 -0.5 0.74 perf-profile.children.cycles-pp.rcu_all_qs 1.46 -0.5 0.98 perf-profile.children.cycles-pp.mas_pop_node 1.12 -0.5 0.65 perf-profile.children.cycles-pp.mast_split_data 1.46 -0.5 1.00 perf-profile.children.cycles-pp.vm_area_free_rcu_cb 2.66 ± 2% -0.4 2.22 ± 3% perf-profile.children.cycles-pp.kmem_cache_free_bulk 1.08 -0.4 0.64 perf-profile.children.cycles-pp.mas_walk 1.04 -0.4 0.62 perf-profile.children.cycles-pp.mas_rev_awalk 1.08 -0.4 0.66 perf-profile.children.cycles-pp.perf_event_mmap_output 0.95 -0.4 0.58 ± 2% perf-profile.children.cycles-pp.memcpy_orig 0.75 -0.3 0.43 perf-profile.children.cycles-pp.mtree_range_walk 0.98 -0.3 0.66 perf-profile.children.cycles-pp.mas_split_final_node 0.75 -0.3 0.44 perf-profile.children.cycles-pp.__memcg_slab_post_alloc_hook 0.86 -0.3 0.57 perf-profile.children.cycles-pp.mas_push_data 0.70 -0.3 0.41 perf-profile.children.cycles-pp.security_mmap_file 1.48 ± 3% -0.3 1.18 ± 2% perf-profile.children.cycles-pp.rcu_cblist_dequeue 0.66 -0.3 0.39 perf-profile.children.cycles-pp.mas_next_slot 0.70 -0.3 0.42 perf-profile.children.cycles-pp.__memcg_slab_free_hook 0.66 -0.3 0.38 perf-profile.children.cycles-pp.mas_prev_slot 0.64 -0.3 0.37 perf-profile.children.cycles-pp.mas_wr_store_entry 0.59 -0.2 0.34 perf-profile.children.cycles-pp.mas_leaf_max_gap 0.57 -0.2 0.34 ± 2% perf-profile.children.cycles-pp.mas_wr_walk_index 0.53 -0.2 0.30 perf-profile.children.cycles-pp.build_detached_freelist 0.50 -0.2 0.31 perf-profile.children.cycles-pp.entry_SYSCALL_64_safe_stack 0.44 -0.2 0.26 perf-profile.children.cycles-pp.mas_wr_walk 0.39 -0.2 0.22 ± 2% perf-profile.children.cycles-pp.mod_objcg_state 0.43 -0.2 0.26 perf-profile.children.cycles-pp.free_pgtables 0.44 -0.2 0.27 perf-profile.children.cycles-pp.rcu_segcblist_enqueue 0.36 -0.2 0.22 ± 2% perf-profile.children.cycles-pp.userfaultfd_unmap_complete 0.36 ± 2% -0.1 0.21 ± 2% perf-profile.children.cycles-pp.down_write_killable 0.36 -0.1 0.22 ± 2% perf-profile.children.cycles-pp.mas_next_sibling 0.31 -0.1 0.18 ± 2% perf-profile.children.cycles-pp.mas_ascend 0.30 -0.1 0.18 ± 2% perf-profile.children.cycles-pp.tlb_finish_mmu 0.32 ± 2% -0.1 0.20 ± 3% perf-profile.children.cycles-pp.down_write 0.29 ± 3% -0.1 0.18 ± 2% perf-profile.children.cycles-pp.up_write 0.31 ± 2% -0.1 0.20 ± 3% perf-profile.children.cycles-pp.obj_cgroup_charge 0.36 -0.1 0.25 perf-profile.children.cycles-pp.syscall_exit_to_user_mode 0.26 ± 2% -0.1 0.16 ± 2% perf-profile.children.cycles-pp.security_mmap_addr 0.24 ± 3% -0.1 0.15 ± 2% perf-profile.children.cycles-pp.security_vm_enough_memory_mm 0.23 ± 2% -0.1 0.14 ± 3% perf-profile.children.cycles-pp.vma_set_page_prot 0.22 ± 2% -0.1 0.12 ± 3% perf-profile.children.cycles-pp.mt_free_rcu 0.21 ± 2% -0.1 0.12 ± 3% perf-profile.children.cycles-pp.vm_get_page_prot 0.22 ± 2% -0.1 0.13 ± 3% perf-profile.children.cycles-pp.mas_prev_range 0.85 -0.1 0.78 ± 2% perf-profile.children.cycles-pp.shuffle_freelist 0.19 ± 3% -0.1 0.12 ± 3% perf-profile.children.cycles-pp.cap_vm_enough_memory 0.17 ± 2% -0.1 0.10 ± 4% perf-profile.children.cycles-pp.mas_prev 0.17 ± 2% -0.1 0.10 ± 4% perf-profile.children.cycles-pp.remove_vma 0.39 ± 2% -0.1 0.32 perf-profile.children.cycles-pp.__sysvec_apic_timer_interrupt 0.39 ± 2% -0.1 0.32 perf-profile.children.cycles-pp.hrtimer_interrupt 0.16 ± 3% -0.1 0.09 ± 5% perf-profile.children.cycles-pp.testcase 0.16 ± 3% -0.1 0.09 perf-profile.children.cycles-pp.percpu_counter_add_batch 0.16 ± 2% -0.1 0.10 ± 5% perf-profile.children.cycles-pp.sized_strscpy 0.16 ± 3% -0.1 0.09 perf-profile.children.cycles-pp.tlb_gather_mmu 0.16 ± 3% -0.1 0.09 ± 4% perf-profile.children.cycles-pp.strnlen 0.34 ± 3% -0.1 0.28 perf-profile.children.cycles-pp.__hrtimer_run_queues 0.14 ± 2% -0.1 0.08 ± 4% perf-profile.children.cycles-pp.lru_add_drain 0.28 ± 3% -0.1 0.22 ± 3% perf-profile.children.cycles-pp.tick_nohz_handler 0.13 ± 3% -0.1 0.08 ± 6% perf-profile.children.cycles-pp.ima_file_mmap 0.16 ± 3% -0.1 0.10 ± 4% perf-profile.children.cycles-pp.should_failslab 0.13 ± 3% -0.1 0.07 perf-profile.children.cycles-pp.syscall_exit_to_user_mode_prepare 0.17 ± 2% -0.1 0.11 ± 6% perf-profile.children.cycles-pp.current_obj_cgroup 0.13 ± 3% -0.1 0.07 ± 6% perf-profile.children.cycles-pp.mab_calc_split 0.11 ± 5% -0.1 0.06 perf-profile.children.cycles-pp.put_cpu_partial 0.12 -0.0 0.07 perf-profile.children.cycles-pp.mas_wr_store_setup 0.07 ± 6% -0.0 0.02 ± 99% perf-profile.children.cycles-pp.unmap_single_vma 0.25 ± 3% -0.0 0.20 ± 4% perf-profile.children.cycles-pp.update_process_times 0.11 -0.0 0.06 ± 7% perf-profile.children.cycles-pp.downgrade_write 0.12 ± 3% -0.0 0.07 ± 5% perf-profile.children.cycles-pp.mas_data_end 0.12 ± 4% -0.0 0.07 perf-profile.children.cycles-pp.mas_next 0.11 ± 4% -0.0 0.06 ± 6% perf-profile.children.cycles-pp.apparmor_mmap_file 0.10 ± 4% -0.0 0.06 ± 6% perf-profile.children.cycles-pp.cap_mmap_addr 0.11 ± 4% -0.0 0.07 perf-profile.children.cycles-pp.lru_add_drain_cpu 0.10 -0.0 0.06 perf-profile.children.cycles-pp.tlb_batch_pages_flush 0.10 ± 4% -0.0 0.06 ± 8% perf-profile.children.cycles-pp.mas_wr_end_piv 0.10 ± 5% -0.0 0.06 perf-profile.children.cycles-pp.may_expand_vm 0.08 -0.0 0.04 ± 44% perf-profile.children.cycles-pp.mas_node_count_gfp 0.09 ± 4% -0.0 0.05 perf-profile.children.cycles-pp.mas_next_range 0.10 ± 4% -0.0 0.06 perf-profile.children.cycles-pp.cap_mmap_file 0.09 ± 5% -0.0 0.06 ± 8% perf-profile.children.cycles-pp.__memcpy 0.09 ± 5% -0.0 0.06 ± 8% perf-profile.children.cycles-pp.__x64_sys_mmap 0.08 ± 5% -0.0 0.05 perf-profile.children.cycles-pp.mas_prev_setup 0.08 ± 5% -0.0 0.05 perf-profile.children.cycles-pp.refill_obj_stock 0.08 ± 5% -0.0 0.05 perf-profile.children.cycles-pp.__vm_enough_memory 0.10 ± 3% -0.0 0.07 ± 7% perf-profile.children.cycles-pp.kfree 0.08 ± 4% -0.0 0.05 perf-profile.children.cycles-pp.up_read 0.12 -0.0 0.10 ± 4% perf-profile.children.cycles-pp.discard_slab 0.16 ± 3% +0.0 0.20 ± 2% perf-profile.children.cycles-pp.__free_one_page 1.26 +0.1 1.32 ± 2% perf-profile.children.cycles-pp.allocate_slab 0.17 ± 2% +0.1 0.29 ± 6% perf-profile.children.cycles-pp.rmqueue 0.13 ± 2% +0.1 0.26 ± 7% perf-profile.children.cycles-pp.__rmqueue_pcplist 0.10 ± 4% +0.1 0.25 ± 6% perf-profile.children.cycles-pp.rmqueue_bulk 0.31 ± 2% +0.2 0.46 ± 4% perf-profile.children.cycles-pp.__alloc_pages 0.27 ± 2% +0.2 0.42 ± 3% perf-profile.children.cycles-pp.get_page_from_freelist 0.00 +0.5 0.46 ± 3% perf-profile.children.cycles-pp.perf_c2c__record 0.04 ± 44% +0.5 0.50 ± 2% perf-profile.children.cycles-pp.main 0.04 ± 44% +0.5 0.50 ± 2% perf-profile.children.cycles-pp.run_builtin 9.45 +0.5 9.92 ± 2% perf-profile.children.cycles-pp.__kmem_cache_alloc_bulk 0.02 ±141% +0.5 0.50 ± 2% perf-profile.children.cycles-pp.cmd_record 9.06 +0.8 9.81 ± 2% perf-profile.children.cycles-pp.___slab_alloc 6.60 ± 2% +1.1 7.69 ± 3% perf-profile.children.cycles-pp.get_partial_node 8.58 +14.3 22.89 perf-profile.children.cycles-pp.__put_partials 24.83 +24.2 49.03 perf-profile.children.cycles-pp.__do_softirq 24.81 +24.2 49.02 perf-profile.children.cycles-pp.rcu_core 24.80 +24.2 49.01 perf-profile.children.cycles-pp.rcu_do_batch 24.97 +24.3 49.31 perf-profile.children.cycles-pp.kmem_cache_free 2.44 ± 5% +24.8 27.19 ± 3% perf-profile.children.cycles-pp.free_pcppages_bulk 2.55 ± 5% +24.8 27.36 ± 3% perf-profile.children.cycles-pp.free_unref_page 2.49 ± 5% +24.8 27.32 ± 3% perf-profile.children.cycles-pp.free_unref_page_commit 15.02 +28.1 43.09 perf-profile.children.cycles-pp.kthread 15.02 +28.1 43.09 perf-profile.children.cycles-pp.ret_from_fork 15.02 +28.1 43.09 perf-profile.children.cycles-pp.ret_from_fork_asm 14.98 +28.1 43.06 perf-profile.children.cycles-pp.run_ksoftirqd 15.00 +28.1 43.08 perf-profile.children.cycles-pp.smpboot_thread_fn 22.12 +28.3 50.41 perf-profile.children.cycles-pp._raw_spin_lock_irqsave 21.74 +28.3 50.06 perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath 4.12 -1.6 2.52 perf-profile.self.cycles-pp.syscall_return_via_sysret 4.04 -1.6 2.47 perf-profile.self.cycles-pp.zap_pmd_range 4.82 -1.4 3.41 perf-profile.self.cycles-pp.__slab_free 3.50 -1.3 2.17 perf-profile.self.cycles-pp.entry_SYSRETQ_unsafe_stack 3.13 -1.2 1.88 perf-profile.self.cycles-pp.__cond_resched 2.93 -1.2 1.77 perf-profile.self.cycles-pp.mas_topiary_replace 2.38 -0.9 1.45 perf-profile.self.cycles-pp.entry_SYSCALL_64_after_hwframe 1.94 -0.8 1.16 perf-profile.self.cycles-pp.kmem_cache_free 1.64 -0.6 0.99 perf-profile.self.cycles-pp.entry_SYSCALL_64 1.51 -0.6 0.93 perf-profile.self.cycles-pp.mas_wr_spanning_store 2.01 -0.6 1.45 perf-profile.self.cycles-pp.mab_mas_cp 1.41 -0.6 0.85 perf-profile.self.cycles-pp.mas_mab_cp 1.45 -0.6 0.89 perf-profile.self.cycles-pp.free_pud_range 1.25 -0.5 0.79 ± 2% perf-profile.self.cycles-pp.kmem_cache_alloc 1.01 -0.4 0.56 perf-profile.self.cycles-pp.__call_rcu_common 1.14 -0.4 0.71 perf-profile.self.cycles-pp.mas_spanning_rebalance 1.30 -0.4 0.90 perf-profile.self.cycles-pp.mas_pop_node 0.94 -0.4 0.57 perf-profile.self.cycles-pp.mas_walk 0.97 -0.4 0.61 perf-profile.self.cycles-pp.perf_event_mmap_output 0.86 -0.3 0.52 perf-profile.self.cycles-pp.mas_rev_awalk 0.87 -0.3 0.54 perf-profile.self.cycles-pp.rcu_all_qs 0.85 -0.3 0.52 perf-profile.self.cycles-pp.do_vmi_align_munmap 0.84 -0.3 0.51 perf-profile.self.cycles-pp.mast_fill_bnode 0.83 -0.3 0.51 perf-profile.self.cycles-pp.mmap_region 0.78 -0.3 0.47 perf-profile.self.cycles-pp.__kmem_cache_alloc_bulk 0.84 -0.3 0.53 ± 2% perf-profile.self.cycles-pp.memcpy_orig 0.73 -0.3 0.44 perf-profile.self.cycles-pp.mas_split 1.45 ± 3% -0.3 1.17 ± 2% perf-profile.self.cycles-pp.rcu_cblist_dequeue 0.66 -0.3 0.40 perf-profile.self.cycles-pp.mtree_range_walk 0.60 -0.2 0.36 perf-profile.self.cycles-pp.mas_find 0.92 -0.2 0.69 perf-profile.self.cycles-pp.___slab_alloc 0.60 -0.2 0.36 perf-profile.self.cycles-pp.mas_prev_slot 0.59 -0.2 0.35 perf-profile.self.cycles-pp.mas_wr_bnode 0.59 -0.2 0.36 ± 2% perf-profile.self.cycles-pp.mas_next_slot 0.53 -0.2 0.32 perf-profile.self.cycles-pp.mas_leaf_max_gap 0.52 -0.2 0.32 perf-profile.self.cycles-pp.mas_wr_walk_index 0.48 -0.2 0.28 ± 2% perf-profile.self.cycles-pp.build_detached_freelist 0.50 -0.2 0.31 ± 2% perf-profile.self.cycles-pp.__memcg_slab_post_alloc_hook 0.46 -0.2 0.28 ± 2% perf-profile.self.cycles-pp.unmap_page_range 0.42 ± 2% -0.2 0.25 ± 2% perf-profile.self.cycles-pp.perf_iterate_sb 0.43 -0.2 0.27 ± 2% perf-profile.self.cycles-pp.rcu_segcblist_enqueue 0.40 -0.2 0.24 perf-profile.self.cycles-pp.mas_wr_walk 0.40 -0.2 0.25 ± 2% perf-profile.self.cycles-pp.do_syscall_64 0.38 -0.2 0.23 perf-profile.self.cycles-pp.kmem_cache_alloc_bulk 0.34 -0.1 0.20 ± 2% perf-profile.self.cycles-pp.mod_objcg_state 0.36 -0.1 0.22 ± 2% perf-profile.self.cycles-pp.vm_unmapped_area 0.38 -0.1 0.24 ± 2% perf-profile.self.cycles-pp.__memcg_slab_free_hook 0.36 -0.1 0.23 ± 2% perf-profile.self.cycles-pp.mas_store_gfp 0.36 -0.1 0.22 perf-profile.self.cycles-pp.perf_event_mmap_event 0.36 -0.1 0.23 perf-profile.self.cycles-pp.arch_get_unmapped_area_topdown 0.33 ± 2% -0.1 0.20 perf-profile.self.cycles-pp.mas_alloc_nodes 0.33 -0.1 0.20 perf-profile.self.cycles-pp.security_mmap_file 0.33 -0.1 0.20 perf-profile.self.cycles-pp.kmem_cache_free_bulk 0.44 -0.1 0.32 ± 2% perf-profile.self.cycles-pp.mas_push_data 0.34 -0.1 0.21 perf-profile.self.cycles-pp.mas_empty_area_rev 0.32 -0.1 0.19 perf-profile.self.cycles-pp.userfaultfd_unmap_complete 0.30 -0.1 0.18 ± 2% perf-profile.self.cycles-pp.mas_store_b_node 0.27 -0.1 0.15 perf-profile.self.cycles-pp.__mmap 0.30 ± 2% -0.1 0.19 ± 2% perf-profile.self.cycles-pp.mas_wr_store_entry 0.27 ± 2% -0.1 0.15 ± 3% perf-profile.self.cycles-pp.__munmap 0.22 ± 3% -0.1 0.11 ± 4% perf-profile.self.cycles-pp.get_partial_node 0.27 -0.1 0.16 ± 3% perf-profile.self.cycles-pp.mas_ascend 0.27 ± 2% -0.1 0.16 ± 3% perf-profile.self.cycles-pp.mas_preallocate 0.28 ± 2% -0.1 0.17 ± 2% perf-profile.self.cycles-pp.do_vmi_munmap 0.24 ± 2% -0.1 0.14 ± 3% perf-profile.self.cycles-pp.__vm_munmap 0.26 ± 2% -0.1 0.16 perf-profile.self.cycles-pp.do_mmap 0.26 ± 2% -0.1 0.16 perf-profile.self.cycles-pp.up_write 0.25 -0.1 0.15 perf-profile.self.cycles-pp.vm_area_alloc 0.24 -0.1 0.15 ± 2% perf-profile.self.cycles-pp.unmap_region 0.14 -0.1 0.05 perf-profile.self.cycles-pp.get_any_partial 0.23 -0.1 0.14 ± 2% perf-profile.self.cycles-pp.perf_event_mmap 0.21 ± 2% -0.1 0.12 ± 3% perf-profile.self.cycles-pp.mt_free_rcu 0.22 -0.1 0.13 ± 2% perf-profile.self.cycles-pp.mas_destroy 0.22 ± 2% -0.1 0.14 perf-profile.self.cycles-pp.free_p4d_range 0.22 ± 2% -0.1 0.13 ± 3% perf-profile.self.cycles-pp.mas_next_sibling 0.19 ± 3% -0.1 0.11 ± 3% perf-profile.self.cycles-pp.vm_get_page_prot 0.20 ± 2% -0.1 0.12 ± 3% perf-profile.self.cycles-pp.down_write_killable 0.21 ± 2% -0.1 0.14 ± 3% perf-profile.self.cycles-pp.obj_cgroup_charge 0.20 ± 2% -0.1 0.12 perf-profile.self.cycles-pp.mast_split_data 0.38 ± 2% -0.1 0.30 ± 2% perf-profile.self.cycles-pp.vm_area_free_rcu_cb 0.18 ± 3% -0.1 0.11 perf-profile.self.cycles-pp.free_pgtables 0.19 -0.1 0.12 ± 3% perf-profile.self.cycles-pp.thp_get_unmapped_area 0.19 -0.1 0.12 ± 3% perf-profile.self.cycles-pp.vm_mmap_pgoff 0.19 ± 2% -0.1 0.12 ± 3% perf-profile.self.cycles-pp.tlb_finish_mmu 0.17 ± 2% -0.1 0.10 ± 4% perf-profile.self.cycles-pp.cap_vm_enough_memory 0.17 ± 2% -0.1 0.11 ± 6% perf-profile.self.cycles-pp.mas_split_final_node 0.17 ± 2% -0.1 0.10 ± 4% perf-profile.self.cycles-pp.syscall_exit_to_user_mode 0.16 ± 2% -0.1 0.10 ± 3% perf-profile.self.cycles-pp.mas_store_prealloc 0.17 ± 2% -0.1 0.11 ± 4% perf-profile.self.cycles-pp.unmap_vmas 0.16 -0.1 0.10 ± 4% perf-profile.self.cycles-pp.free_pgd_range 0.20 ± 2% -0.1 0.13 ± 5% perf-profile.self.cycles-pp.down_write 0.17 ± 4% -0.1 0.11 ± 3% perf-profile.self.cycles-pp.__put_partials 0.15 ± 3% -0.1 0.09 ± 4% perf-profile.self.cycles-pp.rcu_do_batch 0.14 -0.1 0.08 ± 5% perf-profile.self.cycles-pp.security_mmap_addr 0.14 ± 3% -0.1 0.08 perf-profile.self.cycles-pp.tlb_gather_mmu 0.14 ± 3% -0.1 0.08 ± 5% perf-profile.self.cycles-pp.strnlen 0.13 ± 3% -0.1 0.08 perf-profile.self.cycles-pp.mas_prev_range 0.13 ± 5% -0.1 0.08 perf-profile.self.cycles-pp.percpu_counter_add_batch 0.11 ± 4% -0.1 0.06 ± 9% perf-profile.self.cycles-pp.put_cpu_partial 0.13 ± 2% -0.1 0.08 ± 4% perf-profile.self.cycles-pp.current_obj_cgroup 0.13 -0.1 0.08 perf-profile.self.cycles-pp.sized_strscpy 0.13 ± 2% -0.0 0.08 ± 4% perf-profile.self.cycles-pp.get_unmapped_area 0.12 ± 3% -0.0 0.07 perf-profile.self.cycles-pp.ima_file_mmap 0.12 -0.0 0.07 ± 5% perf-profile.self.cycles-pp.mas_prev 0.11 -0.0 0.06 ± 7% perf-profile.self.cycles-pp.syscall_exit_to_user_mode_prepare 0.12 ± 3% -0.0 0.08 ± 4% perf-profile.self.cycles-pp.entry_SYSCALL_64_safe_stack 0.10 ± 3% -0.0 0.06 ± 8% perf-profile.self.cycles-pp.downgrade_write 0.10 -0.0 0.06 perf-profile.self.cycles-pp.__x64_sys_munmap 0.10 -0.0 0.06 perf-profile.self.cycles-pp.lru_add_drain_cpu 0.10 ± 4% -0.0 0.06 ± 6% perf-profile.self.cycles-pp.apparmor_mmap_file 0.11 ± 3% -0.0 0.07 perf-profile.self.cycles-pp.mab_calc_split 0.10 ± 3% -0.0 0.06 ± 6% perf-profile.self.cycles-pp.vma_set_page_prot 0.07 -0.0 0.03 ± 70% perf-profile.self.cycles-pp.tlb_batch_pages_flush 0.09 ± 4% -0.0 0.05 ± 7% perf-profile.self.cycles-pp.may_expand_vm 0.08 ± 6% -0.0 0.04 ± 44% perf-profile.self.cycles-pp.kfree 0.09 ± 5% -0.0 0.05 ± 7% perf-profile.self.cycles-pp.cap_mmap_file 0.10 ± 4% -0.0 0.07 ± 5% perf-profile.self.cycles-pp.mas_data_end 0.38 -0.0 0.35 perf-profile.self.cycles-pp._raw_spin_lock_irqsave 0.09 -0.0 0.06 ± 8% perf-profile.self.cycles-pp.cap_mmap_addr 0.09 -0.0 0.06 ± 8% perf-profile.self.cycles-pp.mas_wr_store_setup 0.09 ± 4% -0.0 0.06 ± 6% perf-profile.self.cycles-pp.remove_vma 0.08 ± 5% -0.0 0.05 perf-profile.self.cycles-pp.refill_obj_stock 0.08 ± 4% -0.0 0.05 perf-profile.self.cycles-pp.mas_nomem 0.08 -0.0 0.05 perf-profile.self.cycles-pp.__x64_sys_mmap 0.08 -0.0 0.05 perf-profile.self.cycles-pp.mas_next 0.13 -0.0 0.10 ± 3% perf-profile.self.cycles-pp.setup_object 0.63 -0.0 0.61 perf-profile.self.cycles-pp.shuffle_freelist 0.12 -0.0 0.10 ± 4% perf-profile.self.cycles-pp.discard_slab 0.08 ± 4% +0.0 0.10 ± 4% perf-profile.self.cycles-pp.rmqueue_bulk 0.16 ± 3% +0.0 0.19 ± 2% perf-profile.self.cycles-pp.__free_one_page 0.00 +0.1 0.06 ± 9% perf-profile.self.cycles-pp.get_page_from_freelist 0.00 +0.1 0.06 ± 7% perf-profile.self.cycles-pp.free_unref_page_commit 21.74 +28.3 50.06 perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath *************************************************************************************************** lkp-csl-2sp7: 96 threads 2 sockets Intel(R) Xeon(R) Gold 6252 CPU @ 2.10GHz (Cascade Lake) with 512G memory ========================================================================================= compiler/cpufreq_governor/kconfig/rootfs/tbox_group/test/testcase: gcc-12/performance/x86_64-rhel-8.3/debian-12-x86_64-phoronix/lkp-csl-2sp7/asmfish-1.1.2/phoronix-test-suite commit: 9ef3ad0f03 ("ACPI/NUMA: Print CXL Early Discovery Table (CEDT)") 09f84589a8 ("ACPI/NUMA: Remove architecture dependent remainings") 9ef3ad0f032818c5 09f84589a8ed1162c0274a1b5b3 ---------------- --------------------------- %stddev %change %stddev \ | \ 14.98 +1.9% 15.27 boot-time.dhcp 93848406 -23.9% 71464814 phoronix-test-suite.asmfish.0.nodes_second 0.06 ± 61% -64.0% 0.02 ± 8% vmstat.procs.b 1183 ± 26% +689.3% 9337 ± 9% perf-c2c.DRAM.remote 112.50 ± 36% +2692.2% 3141 ± 10% perf-c2c.HITM.remote 11131904 ± 13% -33.9% 7360921 ± 7% meminfo.DirectMap2M 1251 ± 31% +56.6% 1959 ± 19% meminfo.Mlocked 4391 ± 8% +16.1% 5099 ± 7% meminfo.Unevictable 260856 -84.1% 41446 meminfo.VmallocUsed 0.05 ± 11% -0.0 0.01 ± 4% mpstat.cpu.all.iowait% 0.02 ± 3% -0.0 0.02 ± 6% mpstat.cpu.all.soft% 0.06 ± 6% -0.0 0.04 mpstat.cpu.all.sys% 31.83 ± 2% -17.1% 26.40 mpstat.max_utilization.seconds 986818 ± 45% -100.0% 0.00 numa-numastat.node0.local_node 1027333 ± 45% -100.0% 0.00 numa-numastat.node0.numa_hit 40474 ± 74% -100.0% 0.00 numa-numastat.node0.other_node 968499 ± 58% -100.0% 0.00 numa-numastat.node1.local_node 1032398 ± 53% -100.0% 0.00 numa-numastat.node1.numa_hit 63895 ± 44% -100.0% 0.00 numa-numastat.node1.other_node 313.19 ± 31% +56.4% 489.72 ± 19% proc-vmstat.nr_mlock 2722 -3.0% 2641 proc-vmstat.nr_page_table_pages 58851 -0.7% 58431 proc-vmstat.nr_slab_unreclaimable 1098 ± 8% +16.1% 1274 ± 7% proc-vmstat.nr_unevictable 1098 ± 8% +16.1% 1274 ± 7% proc-vmstat.nr_zone_unevictable 49674 ± 59% -100.0% 0.00 proc-vmstat.numa_hint_faults 36892 ± 57% -100.0% 0.00 proc-vmstat.numa_hint_faults_local 2061577 ± 48% -34.9% 1342268 proc-vmstat.numa_hit 14970 ± 76% -100.0% 0.00 proc-vmstat.numa_huge_pte_updates 1957163 ± 50% -31.4% 1342877 proc-vmstat.numa_local 104370 ± 3% -100.0% 0.00 proc-vmstat.numa_other 1784768 ± 75% -100.0% 0.00 proc-vmstat.numa_pages_migrated 7772796 ± 75% -100.0% 0.00 proc-vmstat.numa_pte_updates 5403170 ± 62% -54.3% 2470734 proc-vmstat.pgalloc_normal 2641844 ± 60% -45.6% 1438131 proc-vmstat.pgfault 5223988 ± 64% -56.2% 2290502 proc-vmstat.pgfree 1784768 ± 75% -100.0% 0.00 proc-vmstat.pgmigrate_success 98843 ± 61% -39.3% 60039 ± 3% proc-vmstat.pgreuse 3478 ± 75% -100.0% 0.00 proc-vmstat.thp_migration_success 137922 ± 38% -100.0% 0.00 numa-meminfo.node0.Active 39195 ± 31% -100.0% 0.00 numa-meminfo.node0.Active(anon) 98727 ± 43% -100.0% 0.00 numa-meminfo.node0.Active(file) 867525 ± 22% -100.0% 0.00 numa-meminfo.node0.AnonHugePages 1116586 ± 19% -100.0% 0.00 numa-meminfo.node0.AnonPages 1448118 ± 15% -100.0% 0.00 numa-meminfo.node0.AnonPages.max 1414495 ± 62% -100.0% 0.00 numa-meminfo.node0.FilePages 2393152 ± 28% -100.0% 0.00 numa-meminfo.node0.Inactive 1467103 ± 16% -100.0% 0.00 numa-meminfo.node0.Inactive(anon) 926048 ± 78% -100.0% 0.00 numa-meminfo.node0.Inactive(file) 42206 ± 33% -100.0% 0.00 numa-meminfo.node0.KReclaimable 10393 ± 8% -100.0% 0.00 numa-meminfo.node0.KernelStack 84153 ± 30% -100.0% 0.00 numa-meminfo.node0.Mapped 2.606e+08 -100.0% 0.00 numa-meminfo.node0.MemFree 2.638e+08 -100.0% 0.00 numa-meminfo.node0.MemTotal 3202155 ± 24% -100.0% 0.00 numa-meminfo.node0.MemUsed 5609 ± 19% -100.0% 0.00 numa-meminfo.node0.PageTables 42206 ± 33% -100.0% 0.00 numa-meminfo.node0.SReclaimable 110987 ± 21% -100.0% 0.00 numa-meminfo.node0.SUnreclaim 389372 ± 45% -100.0% 0.00 numa-meminfo.node0.Shmem 153193 ± 23% -100.0% 0.00 numa-meminfo.node0.Slab 3500 ± 14% -100.0% 0.00 numa-meminfo.node0.Unevictable 278067 ± 19% -100.0% 0.00 numa-meminfo.node1.Active 214635 ± 8% -100.0% 0.00 numa-meminfo.node1.Active(anon) 63432 ± 67% -100.0% 0.00 numa-meminfo.node1.Active(file) 680824 ± 23% -100.0% 0.00 numa-meminfo.node1.AnonHugePages 829328 ± 25% -100.0% 0.00 numa-meminfo.node1.AnonPages 1414486 ± 14% -100.0% 0.00 numa-meminfo.node1.AnonPages.max 1390641 ± 64% -100.0% 0.00 numa-meminfo.node1.FilePages 1942799 ± 35% -100.0% 0.00 numa-meminfo.node1.Inactive 1173590 ± 19% -100.0% 0.00 numa-meminfo.node1.Inactive(anon) 769209 ± 94% -100.0% 0.00 numa-meminfo.node1.Inactive(file) 42052 ± 34% -100.0% 0.00 numa-meminfo.node1.KReclaimable 10353 ± 7% -100.0% 0.00 numa-meminfo.node1.KernelStack 56786 ± 44% -100.0% 0.00 numa-meminfo.node1.Mapped 2.612e+08 -100.0% 0.00 numa-meminfo.node1.MemFree 2.642e+08 -100.0% 0.00 numa-meminfo.node1.MemTotal 3009654 ± 26% -100.0% 0.00 numa-meminfo.node1.MemUsed 5280 ± 19% -100.0% 0.00 numa-meminfo.node1.PageTables 42052 ± 34% -100.0% 0.00 numa-meminfo.node1.SReclaimable 124421 ± 19% -100.0% 0.00 numa-meminfo.node1.SUnreclaim 557220 ± 32% -100.0% 0.00 numa-meminfo.node1.Shmem 166473 ± 22% -100.0% 0.00 numa-meminfo.node1.Slab 9797 ± 31% -100.0% 0.00 numa-vmstat.node0.nr_active_anon 24680 ± 43% -100.0% 0.00 numa-vmstat.node0.nr_active_file 279210 ± 19% -100.0% 0.00 numa-vmstat.node0.nr_anon_pages 423.69 ± 22% -100.0% 0.00 numa-vmstat.node0.nr_anon_transparent_hugepages 353618 ± 62% -100.0% 0.00 numa-vmstat.node0.nr_file_pages 65159849 -100.0% 0.00 numa-vmstat.node0.nr_free_pages 366837 ± 16% -100.0% 0.00 numa-vmstat.node0.nr_inactive_anon 231510 ± 78% -100.0% 0.00 numa-vmstat.node0.nr_inactive_file 10394 ± 8% -100.0% 0.00 numa-vmstat.node0.nr_kernel_stack 21187 ± 29% -100.0% 0.00 numa-vmstat.node0.nr_mapped 1402 ± 19% -100.0% 0.00 numa-vmstat.node0.nr_page_table_pages 97341 ± 45% -100.0% 0.00 numa-vmstat.node0.nr_shmem 10551 ± 33% -100.0% 0.00 numa-vmstat.node0.nr_slab_reclaimable 27747 ± 21% -100.0% 0.00 numa-vmstat.node0.nr_slab_unreclaimable 875.15 ± 14% -100.0% 0.00 numa-vmstat.node0.nr_unevictable 9797 ± 31% -100.0% 0.00 numa-vmstat.node0.nr_zone_active_anon 24680 ± 43% -100.0% 0.00 numa-vmstat.node0.nr_zone_active_file 366837 ± 16% -100.0% 0.00 numa-vmstat.node0.nr_zone_inactive_anon 231510 ± 78% -100.0% 0.00 numa-vmstat.node0.nr_zone_inactive_file 875.15 ± 14% -100.0% 0.00 numa-vmstat.node0.nr_zone_unevictable 1027333 ± 45% -100.0% 0.00 numa-vmstat.node0.numa_hit 986816 ± 45% -100.0% 0.00 numa-vmstat.node0.numa_local 40474 ± 74% -100.0% 0.00 numa-vmstat.node0.numa_other 53643 ± 8% -100.0% 0.00 numa-vmstat.node1.nr_active_anon 15857 ± 67% -100.0% 0.00 numa-vmstat.node1.nr_active_file 207230 ± 25% -100.0% 0.00 numa-vmstat.node1.nr_anon_pages 332.22 ± 23% -100.0% 0.00 numa-vmstat.node1.nr_anon_transparent_hugepages 347657 ± 64% -100.0% 0.00 numa-vmstat.node1.nr_file_pages 65294238 -100.0% 0.00 numa-vmstat.node1.nr_free_pages 293307 ± 19% -100.0% 0.00 numa-vmstat.node1.nr_inactive_anon 192302 ± 94% -100.0% 0.00 numa-vmstat.node1.nr_inactive_file 10354 ± 7% -100.0% 0.00 numa-vmstat.node1.nr_kernel_stack 14315 ± 43% -100.0% 0.00 numa-vmstat.node1.nr_mapped 1320 ± 19% -100.0% 0.00 numa-vmstat.node1.nr_page_table_pages 139303 ± 32% -100.0% 0.00 numa-vmstat.node1.nr_shmem 10512 ± 34% -100.0% 0.00 numa-vmstat.node1.nr_slab_reclaimable 31105 ± 19% -100.0% 0.00 numa-vmstat.node1.nr_slab_unreclaimable 53643 ± 8% -100.0% 0.00 numa-vmstat.node1.nr_zone_active_anon 15857 ± 67% -100.0% 0.00 numa-vmstat.node1.nr_zone_active_file 293306 ± 19% -100.0% 0.00 numa-vmstat.node1.nr_zone_inactive_anon 192302 ± 94% -100.0% 0.00 numa-vmstat.node1.nr_zone_inactive_file 1032447 ± 53% -100.0% 0.00 numa-vmstat.node1.numa_hit 968548 ± 57% -100.0% 0.00 numa-vmstat.node1.numa_local 63895 ± 44% -100.0% 0.00 numa-vmstat.node1.numa_other 2.01 ± 5% -10.8% 1.80 ± 6% perf-sched.sch_delay.avg.ms.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part 0.02 ± 25% -31.8% 0.01 ± 9% perf-sched.sch_delay.avg.ms.do_nanosleep.hrtimer_nanosleep.__x64_sys_nanosleep.do_syscall_64 0.24 ± 20% -28.4% 0.17 ± 20% perf-sched.sch_delay.avg.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64 0.01 ± 8% -100.0% 0.00 perf-sched.sch_delay.avg.ms.io_schedule.migration_entry_wait_on_locked.__handle_mm_fault.handle_mm_fault 0.00 ± 33% -100.0% 0.00 perf-sched.sch_delay.avg.ms.io_schedule.migration_entry_wait_on_locked.migration_entry_wait.do_swap_page 0.02 ± 4% +12.1% 0.02 ± 3% perf-sched.sch_delay.avg.ms.irq_thread.kthread.ret_from_fork.ret_from_fork_asm 0.02 ± 2% +21.2% 0.02 ± 6% perf-sched.sch_delay.avg.ms.rcu_gp_kthread.kthread.ret_from_fork.ret_from_fork_asm 0.02 ± 50% -35.5% 0.01 ± 7% perf-sched.sch_delay.avg.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.do_epoll_pwait.part 0.02 ± 4% +21.3% 0.02 ± 4% perf-sched.sch_delay.avg.ms.schedule_timeout.kcompactd.kthread.ret_from_fork 0.02 ± 19% -28.6% 0.02 ± 17% perf-sched.sch_delay.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm 3.91 -21.9% 3.05 ± 23% perf-sched.sch_delay.max.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64 0.04 ± 13% -100.0% 0.00 perf-sched.sch_delay.max.ms.io_schedule.migration_entry_wait_on_locked.__handle_mm_fault.handle_mm_fault 0.02 ± 33% -100.0% 0.00 perf-sched.sch_delay.max.ms.io_schedule.migration_entry_wait_on_locked.migration_entry_wait.do_swap_page 0.32 ±208% -100.0% 0.00 perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown].[unknown] 0.01 ± 35% -100.0% 0.00 perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown].[unknown] 0.03 ± 45% -100.0% 0.00 perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi.[unknown].[unknown] 2.31 ± 31% -55.0% 1.04 ± 98% perf-sched.sch_delay.max.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm 65.46 ± 6% -15.5% 55.34 ± 2% perf-sched.total_wait_and_delay.average.ms 65.43 ± 6% -15.5% 55.31 ± 2% perf-sched.total_wait_time.average.ms 2.01 ± 5% -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part 3.39 ± 9% -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64 0.62 ± 7% -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.io_schedule.migration_entry_wait_on_locked.__handle_mm_fault.handle_mm_fault 0.15 ± 11% -18.2% 0.13 ± 3% perf-sched.wait_and_delay.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown].[unknown] 51.61 ± 7% -12.3% 45.26 ± 6% perf-sched.wait_and_delay.avg.ms.pipe_read.vfs_read.ksys_read.do_syscall_64 3.59 ± 16% -28.0% 2.58 ± 21% perf-sched.wait_and_delay.avg.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.do_epoll_pwait.part 2.44 ± 20% -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone 18.95 ± 4% +50.7% 28.56 ± 4% perf-sched.wait_and_delay.avg.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread 494.08 +11.4% 550.61 perf-sched.wait_and_delay.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm 37.67 ± 18% -50.1% 18.80 ± 6% perf-sched.wait_and_delay.count.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm 104.50 ± 5% -100.0% 0.00 perf-sched.wait_and_delay.count.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part 124.83 -48.7% 64.00 perf-sched.wait_and_delay.count.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64 114.67 ± 2% -100.0% 0.00 perf-sched.wait_and_delay.count.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64 749.83 ± 5% -100.0% 0.00 perf-sched.wait_and_delay.count.io_schedule.migration_entry_wait_on_locked.__handle_mm_fault.handle_mm_fault 1323 ± 6% -26.3% 975.20 ± 8% perf-sched.wait_and_delay.count.pipe_read.vfs_read.ksys_read.do_syscall_64 1427 ± 14% +50.1% 2143 ± 10% perf-sched.wait_and_delay.count.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.do_epoll_pwait.part 86.17 -100.0% 0.00 perf-sched.wait_and_delay.count.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone 19.83 -51.6% 9.60 ± 5% perf-sched.wait_and_delay.count.schedule_timeout.kcompactd.kthread.ret_from_fork 254.33 ± 5% -33.6% 168.80 ± 5% perf-sched.wait_and_delay.count.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread 854.17 -10.1% 768.00 perf-sched.wait_and_delay.count.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm 574.00 ± 7% -41.3% 336.80 ± 9% perf-sched.wait_and_delay.count.worker_thread.kthread.ret_from_fork.ret_from_fork_asm 3.93 -100.0% 0.00 perf-sched.wait_and_delay.max.ms.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part 47.29 ± 55% -100.0% 0.00 perf-sched.wait_and_delay.max.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64 2.88 ± 19% -100.0% 0.00 perf-sched.wait_and_delay.max.ms.io_schedule.migration_entry_wait_on_locked.__handle_mm_fault.handle_mm_fault 31.00 ± 94% -100.0% 0.00 perf-sched.wait_and_delay.max.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone 288.79 ±116% -98.7% 3.87 ± 6% perf-sched.wait_time.avg.ms.__cond_resched.__alloc_pages.alloc_pages_mpol.shmem_alloc_folio.shmem_alloc_and_add_folio 0.61 ± 7% -100.0% 0.00 perf-sched.wait_time.avg.ms.io_schedule.migration_entry_wait_on_locked.__handle_mm_fault.handle_mm_fault 0.07 ± 41% -100.0% 0.00 perf-sched.wait_time.avg.ms.io_schedule.migration_entry_wait_on_locked.migration_entry_wait.do_swap_page 0.15 ± 11% -18.1% 0.13 ± 3% perf-sched.wait_time.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown].[unknown] 51.58 ± 7% -12.7% 45.03 ± 7% perf-sched.wait_time.avg.ms.pipe_read.vfs_read.ksys_read.do_syscall_64 3.57 ± 16% -28.0% 2.57 ± 21% perf-sched.wait_time.avg.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.do_epoll_pwait.part 2.17 ± 18% +49.0% 3.23 ± 18% perf-sched.wait_time.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone 18.94 ± 4% +50.7% 28.54 ± 4% perf-sched.wait_time.avg.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread 494.06 +11.4% 550.60 perf-sched.wait_time.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm 2.87 ± 19% -100.0% 0.00 perf-sched.wait_time.max.ms.io_schedule.migration_entry_wait_on_locked.__handle_mm_fault.handle_mm_fault 0.74 ± 86% -100.0% 0.00 perf-sched.wait_time.max.ms.io_schedule.migration_entry_wait_on_locked.migration_entry_wait.do_swap_page 4.83 ± 14% -33.7% 3.21 ± 49% perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown] 0.86 +113.9% 1.84 perf-stat.i.MPKI 2.197e+10 -22.8% 1.697e+10 perf-stat.i.branch-instructions 8.68 -0.2 8.52 perf-stat.i.branch-miss-rate% 1.9e+09 -23.3% 1.458e+09 perf-stat.i.branch-misses 27.65 +23.9 51.52 perf-stat.i.cache-miss-rate% 1.621e+08 +77.1% 2.871e+08 perf-stat.i.cache-misses 5.87e+08 -7.6% 5.422e+08 perf-stat.i.cache-references 1.16 +38.1% 1.60 perf-stat.i.cpi 2.225e+11 +7.5% 2.393e+11 perf-stat.i.cpu-cycles 154.02 -16.1% 129.22 perf-stat.i.cpu-migrations 1391 -38.3% 857.88 perf-stat.i.cycles-between-cache-misses 0.06 ± 4% -0.0 0.05 ± 8% perf-stat.i.dTLB-load-miss-rate% 16519218 ± 3% -27.0% 12056970 ± 11% perf-stat.i.dTLB-load-misses 6.726e+10 -22.7% 5.201e+10 perf-stat.i.dTLB-loads 325201 -12.1% 285915 ± 3% perf-stat.i.dTLB-store-misses 2.069e+10 -22.8% 1.597e+10 perf-stat.i.dTLB-stores 1471789 -7.6% 1360644 ± 2% perf-stat.i.iTLB-load-misses 2.235e+11 -22.7% 1.728e+11 perf-stat.i.instructions 161257 -16.2% 135135 perf-stat.i.instructions-per-iTLB-miss 0.98 -26.7% 0.72 perf-stat.i.ipc 2.32 +7.5% 2.49 perf-stat.i.metric.GHz 316.36 +194.6% 931.87 perf-stat.i.metric.K/sec 1151 -22.6% 890.51 perf-stat.i.metric.M/sec 3133 ± 4% -46.7% 1669 perf-stat.i.minor-faults 27.70 +52.6 80.31 perf-stat.i.node-load-miss-rate% 5227534 ± 2% +993.2% 57147855 perf-stat.i.node-load-misses 15752155 -13.8% 13581362 perf-stat.i.node-loads 51.43 +41.3 92.70 perf-stat.i.node-store-miss-rate% 2278574 +614.2% 16272457 ± 2% perf-stat.i.node-store-misses 2080120 -48.2% 1078354 perf-stat.i.node-stores 3133 ± 4% -46.7% 1669 perf-stat.i.page-faults 0.73 +129.0% 1.66 perf-stat.overall.MPKI 27.63 +25.3 52.96 perf-stat.overall.cache-miss-rate% 1.00 +39.1% 1.38 perf-stat.overall.cpi 1372 -39.3% 833.36 perf-stat.overall.cycles-between-cache-misses 0.00 +0.0 0.00 ± 3% perf-stat.overall.dTLB-store-miss-rate% 151892 -16.2% 127248 ± 3% perf-stat.overall.instructions-per-iTLB-miss 1.00 -28.1% 0.72 perf-stat.overall.ipc 24.91 +55.9 80.79 perf-stat.overall.node-load-miss-rate% 52.27 +41.5 93.78 perf-stat.overall.node-store-miss-rate% 2.194e+10 -22.7% 1.696e+10 perf-stat.ps.branch-instructions 1.898e+09 -23.2% 1.457e+09 perf-stat.ps.branch-misses 1.619e+08 +77.2% 2.869e+08 perf-stat.ps.cache-misses 5.86e+08 -7.6% 5.417e+08 perf-stat.ps.cache-references 2.222e+11 +7.6% 2.391e+11 perf-stat.ps.cpu-cycles 153.79 -16.1% 128.97 perf-stat.ps.cpu-migrations 16497488 ± 3% -26.9% 12054157 ± 11% perf-stat.ps.dTLB-load-misses 6.717e+10 -22.6% 5.198e+10 perf-stat.ps.dTLB-loads 324755 -12.1% 285616 ± 3% perf-stat.ps.dTLB-store-misses 2.066e+10 -22.8% 1.596e+10 perf-stat.ps.dTLB-stores 1469530 -7.6% 1358463 ± 2% perf-stat.ps.iTLB-load-misses 2.232e+11 -22.6% 1.727e+11 perf-stat.ps.instructions 3129 ± 4% -46.8% 1664 perf-stat.ps.minor-faults 5220451 ± 2% +993.6% 57090593 perf-stat.ps.node-load-misses 15736071 -13.8% 13570457 perf-stat.ps.node-loads 2275264 +614.3% 16251080 ± 2% perf-stat.ps.node-store-misses 2077508 -48.1% 1077622 perf-stat.ps.node-stores 3129 ± 4% -46.8% 1664 perf-stat.ps.page-faults 2.84 ± 72% -2.5 0.34 ± 82% perf-profile.calltrace.cycles-pp.asm_exc_page_fault 2.78 ± 74% -2.4 0.33 ± 81% perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault 9.33 ± 6% -1.9 7.45 ± 8% perf-profile.calltrace.cycles-pp.ktime_get_update_offsets_now.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt 0.96 ± 12% -0.4 0.55 ± 54% perf-profile.calltrace.cycles-pp.rcu_core.__do_softirq.irq_exit_rcu.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt 1.38 ± 11% -0.3 1.12 ± 8% perf-profile.calltrace.cycles-pp.rcu_sched_clock_irq.update_process_times.tick_nohz_handler.__hrtimer_run_queues.hrtimer_interrupt 0.50 ± 46% +0.4 0.87 ± 8% perf-profile.calltrace.cycles-pp.kthread.ret_from_fork.ret_from_fork_asm 0.50 ± 46% +0.4 0.87 ± 8% perf-profile.calltrace.cycles-pp.ret_from_fork.ret_from_fork_asm 0.50 ± 46% +0.4 0.87 ± 8% perf-profile.calltrace.cycles-pp.ret_from_fork_asm 2.71 ± 5% +0.4 3.09 ± 3% perf-profile.calltrace.cycles-pp.__update_load_avg_se.update_load_avg.task_tick_fair.scheduler_tick.update_process_times 0.52 ± 46% +0.5 1.03 ± 28% perf-profile.calltrace.cycles-pp.hrtimer_forward.tick_nohz_handler.__hrtimer_run_queues.hrtimer_interrupt.__sysvec_apic_timer_interrupt 7.24 ± 6% +0.7 7.92 ± 5% perf-profile.calltrace.cycles-pp.update_load_avg.task_tick_fair.scheduler_tick.update_process_times.tick_nohz_handler 3.41 ± 12% +1.5 4.91 ± 13% perf-profile.calltrace.cycles-pp.ktime_get.tick_nohz_handler.__hrtimer_run_queues.hrtimer_interrupt.__sysvec_apic_timer_interrupt 18.04 ± 4% +1.8 19.80 ± 5% perf-profile.calltrace.cycles-pp.task_tick_fair.scheduler_tick.update_process_times.tick_nohz_handler.__hrtimer_run_queues 3.34 ± 61% -2.5 0.87 ± 9% perf-profile.children.cycles-pp.asm_exc_page_fault 3.19 ± 65% -2.4 0.75 ± 7% perf-profile.children.cycles-pp.exc_page_fault 3.18 ± 65% -2.4 0.74 ± 6% perf-profile.children.cycles-pp.do_user_addr_fault 3.09 ± 68% -2.4 0.68 ± 7% perf-profile.children.cycles-pp.handle_mm_fault 3.02 ± 69% -2.4 0.63 ± 8% perf-profile.children.cycles-pp.__handle_mm_fault 9.34 ± 6% -1.9 7.45 ± 8% perf-profile.children.cycles-pp.ktime_get_update_offsets_now 0.99 ± 12% -0.3 0.69 ± 24% perf-profile.children.cycles-pp.rcu_core 1.40 ± 13% -0.3 1.15 ± 7% perf-profile.children.cycles-pp.rcu_sched_clock_irq 0.46 ± 29% -0.2 0.24 ± 23% perf-profile.children.cycles-pp.check_cpu_stall 0.49 ± 16% -0.2 0.32 ± 28% perf-profile.children.cycles-pp.vm_mmap_pgoff 0.26 ± 33% -0.2 0.09 ± 57% perf-profile.children.cycles-pp._raw_spin_trylock 0.48 ± 16% -0.2 0.32 ± 27% perf-profile.children.cycles-pp.do_mmap 0.44 ± 20% -0.2 0.28 ± 20% perf-profile.children.cycles-pp.setlocale 0.36 ± 21% -0.1 0.21 ± 33% perf-profile.children.cycles-pp.ksys_mmap_pgoff 0.36 ± 19% -0.1 0.22 ± 22% perf-profile.children.cycles-pp.do_read_fault 0.19 ± 11% -0.1 0.09 ± 64% perf-profile.children.cycles-pp.kmem_cache_alloc 0.42 ± 17% -0.1 0.32 ± 24% perf-profile.children.cycles-pp.do_fault 0.28 ± 26% -0.1 0.18 ± 18% perf-profile.children.cycles-pp.avg_vruntime 0.36 ± 16% -0.1 0.27 ± 7% perf-profile.children.cycles-pp.exit_mm 0.28 ± 16% -0.1 0.19 ± 36% perf-profile.children.cycles-pp.__open64_nocancel 0.14 ± 18% -0.1 0.06 ± 30% perf-profile.children.cycles-pp.__mmap 0.12 ± 25% -0.1 0.04 ± 86% perf-profile.children.cycles-pp.__mem_cgroup_charge 0.15 ± 48% -0.1 0.08 ± 25% perf-profile.children.cycles-pp.__cond_resched 0.03 ±102% +0.1 0.12 ± 33% perf-profile.children.cycles-pp.task_mm_cid_work 0.26 ± 33% +0.1 0.36 ± 9% perf-profile.children.cycles-pp.worker_thread 0.05 ± 71% +0.1 0.15 ± 20% perf-profile.children.cycles-pp.apparmor_file_permission 0.05 ± 75% +0.1 0.16 ± 20% perf-profile.children.cycles-pp.security_file_permission 0.06 ± 74% +0.1 0.16 ± 20% perf-profile.children.cycles-pp.rw_verify_area 0.08 ± 14% +0.2 0.24 ± 30% perf-profile.children.cycles-pp.copy_mc_fragile 0.06 ± 87% +0.2 0.22 ± 32% perf-profile.children.cycles-pp.hpage_collapse_scan_pmd 0.06 ± 87% +0.2 0.22 ± 32% perf-profile.children.cycles-pp.khugepaged_scan_mm_slot 0.06 ± 92% +0.2 0.22 ± 32% perf-profile.children.cycles-pp.khugepaged 0.05 ± 91% +0.2 0.22 ± 32% perf-profile.children.cycles-pp.collapse_huge_page 0.25 ± 26% +0.2 0.45 ± 26% perf-profile.children.cycles-pp.update_rq_clock 0.28 ± 37% +0.2 0.50 ± 24% perf-profile.children.cycles-pp.runtime.epollwait.abi0 0.58 ± 15% +0.3 0.87 ± 8% perf-profile.children.cycles-pp.kthread 0.61 ± 15% +0.3 0.91 ± 6% perf-profile.children.cycles-pp.ret_from_fork_asm 0.60 ± 15% +0.3 0.90 ± 7% perf-profile.children.cycles-pp.ret_from_fork 1.22 ± 14% +0.3 1.52 ± 5% perf-profile.children.cycles-pp.sched_clock_cpu 0.20 ± 50% +0.4 0.55 ± 43% perf-profile.children.cycles-pp.copy_page_from_iter_atomic 0.71 ± 9% +0.4 1.11 ± 28% perf-profile.children.cycles-pp.hrtimer_forward 2.81 ± 6% +0.4 3.24 ± 3% perf-profile.children.cycles-pp.__update_load_avg_se 18.15 ± 4% +1.7 19.88 ± 5% perf-profile.children.cycles-pp.task_tick_fair 4.84 ± 9% +4.5 9.33 ± 56% perf-profile.children.cycles-pp.ktime_get 8.97 ± 7% -1.9 7.10 ± 8% perf-profile.self.cycles-pp.ktime_get_update_offsets_now 2.47 ± 12% -0.8 1.67 ± 41% perf-profile.self.cycles-pp.update_process_times 0.46 ± 29% -0.2 0.24 ± 23% perf-profile.self.cycles-pp.check_cpu_stall 0.26 ± 33% -0.2 0.09 ± 57% perf-profile.self.cycles-pp._raw_spin_trylock 0.02 ±142% +0.1 0.08 ± 22% perf-profile.self.cycles-pp.perf_poll 0.13 ± 28% +0.1 0.22 ± 19% perf-profile.self.cycles-pp.rebalance_domains 0.02 ±141% +0.1 0.12 ± 33% perf-profile.self.cycles-pp.task_mm_cid_work 0.08 ± 14% +0.2 0.24 ± 30% perf-profile.self.cycles-pp.copy_mc_fragile 0.70 ± 10% +0.4 1.09 ± 28% perf-profile.self.cycles-pp.hrtimer_forward 2.52 ± 9% +0.5 2.98 ± 5% perf-profile.self.cycles-pp.__update_load_avg_se 3.83 ± 12% +4.5 8.29 ± 62% perf-profile.self.cycles-pp.ktime_get 0.81 ± 9% +12.2% 0.91 ± 2% sched_debug.cfs_rq:/.h_nr_running.min 8252 ± 8% +15.1% 9494 ± 3% sched_debug.cfs_rq:/.load.min 7.77 ± 7% +14.8% 8.92 ± 3% sched_debug.cfs_rq:/.load_avg.min 0.81 ± 9% +12.2% 0.91 ± 2% sched_debug.cfs_rq:/.nr_running.min 825.73 ± 8% +12.7% 930.23 ± 3% sched_debug.cfs_rq:/.runnable_avg.min 798.93 ± 8% +12.3% 897.30 ± 4% sched_debug.cfs_rq:/.util_avg.min 0.81 ± 9% +12.2% 0.91 ± 2% sched_debug.cfs_rq:/system.slice.h_nr_running.min 8609 ± 9% +13.0% 9727 ± 3% sched_debug.cfs_rq:/system.slice.load.min 11.30 ± 7% -9.6% 10.22 sched_debug.cfs_rq:/system.slice.load_avg.avg 65.78 ± 30% -59.6% 26.56 ± 8% sched_debug.cfs_rq:/system.slice.load_avg.max 8.05 ± 8% +12.5% 9.05 ± 3% sched_debug.cfs_rq:/system.slice.load_avg.min 9.36 ± 37% -70.7% 2.75 ± 9% sched_debug.cfs_rq:/system.slice.load_avg.stddev 0.81 ± 9% +12.2% 0.91 ± 2% sched_debug.cfs_rq:/system.slice.nr_running.min 825.67 ± 8% +12.7% 930.23 ± 3% sched_debug.cfs_rq:/system.slice.runnable_avg.min 10.94 ± 42% -58.2% 4.58 ± 61% sched_debug.cfs_rq:/system.slice.se->avg.load_avg.stddev 825.67 ± 8% +12.7% 930.23 ± 3% sched_debug.cfs_rq:/system.slice.se->avg.runnable_avg.min 797.77 ± 8% +12.4% 896.95 ± 4% sched_debug.cfs_rq:/system.slice.se->avg.util_avg.min 122819 ± 40% -51.2% 59903 ± 44% sched_debug.cfs_rq:/system.slice.se->load.weight.max 25488 ± 50% -75.6% 6211 ± 42% sched_debug.cfs_rq:/system.slice.se->load.weight.stddev 8.06 ± 9% +12.4% 9.05 ± 3% sched_debug.cfs_rq:/system.slice.tg_load_avg_contrib.min 797.77 ± 8% +12.4% 896.95 ± 4% sched_debug.cfs_rq:/system.slice.util_avg.min 128.18 ± 22% +91.7% 245.75 ± 11% sched_debug.cfs_rq:/system.slice/containerd.service.avg_vruntime.avg 234.17 ± 25% +79.1% 419.30 ± 12% sched_debug.cfs_rq:/system.slice/containerd.service.avg_vruntime.max 43.50 ± 22% +120.5% 95.92 ± 28% sched_debug.cfs_rq:/system.slice/containerd.service.avg_vruntime.min 81.66 ± 29% +74.6% 142.59 ± 24% sched_debug.cfs_rq:/system.slice/containerd.service.avg_vruntime.stddev 2.80 ± 33% +81.1% 5.06 ± 33% sched_debug.cfs_rq:/system.slice/containerd.service.load_avg.avg 5.64 ± 29% +82.3% 10.28 ± 31% sched_debug.cfs_rq:/system.slice/containerd.service.load_avg.max 2.27 ± 34% +93.6% 4.40 ± 37% sched_debug.cfs_rq:/system.slice/containerd.service.load_avg.stddev 128.18 ± 22% +91.7% 245.75 ± 11% sched_debug.cfs_rq:/system.slice/containerd.service.min_vruntime.avg 234.17 ± 25% +79.1% 419.30 ± 12% sched_debug.cfs_rq:/system.slice/containerd.service.min_vruntime.max 43.50 ± 22% +120.5% 95.92 ± 28% sched_debug.cfs_rq:/system.slice/containerd.service.min_vruntime.min 81.66 ± 29% +74.6% 142.59 ± 24% sched_debug.cfs_rq:/system.slice/containerd.service.min_vruntime.stddev 2.79 ± 33% +81.1% 5.06 ± 33% sched_debug.cfs_rq:/system.slice/containerd.service.runnable_avg.avg 5.62 ± 29% +82.6% 10.27 ± 31% sched_debug.cfs_rq:/system.slice/containerd.service.runnable_avg.max 2.26 ± 33% +94.0% 4.39 ± 38% sched_debug.cfs_rq:/system.slice/containerd.service.runnable_avg.stddev 2.78 ± 33% +81.5% 5.05 ± 33% sched_debug.cfs_rq:/system.slice/containerd.service.se->avg.runnable_avg.avg 5.61 ± 29% +83.1% 10.27 ± 31% sched_debug.cfs_rq:/system.slice/containerd.service.se->avg.runnable_avg.max 2.26 ± 33% +94.6% 4.40 ± 37% sched_debug.cfs_rq:/system.slice/containerd.service.se->avg.runnable_avg.stddev 136.98 ± 21% +94.7% 266.75 ± 14% sched_debug.cfs_rq:/system.slice/containerd.service.se->sum_exec_runtime.avg 252.42 ± 24% +82.7% 461.23 ± 16% sched_debug.cfs_rq:/system.slice/containerd.service.se->sum_exec_runtime.max 45.53 ± 21% +119.8% 100.06 ± 28% sched_debug.cfs_rq:/system.slice/containerd.service.se->sum_exec_runtime.min 89.12 ± 27% +79.0% 159.55 ± 28% sched_debug.cfs_rq:/system.slice/containerd.service.se->sum_exec_runtime.stddev 8.81 ± 37% +4295.1% 387.34 ±118% sched_debug.cfs_rq:/system.slice/containerd.service.tg_load_avg.avg 9.12 ± 36% +4150.1% 387.69 ±118% sched_debug.cfs_rq:/system.slice/containerd.service.tg_load_avg.max 8.51 ± 37% +4444.4% 386.92 ±118% sched_debug.cfs_rq:/system.slice/containerd.service.tg_load_avg.min 2.80 ± 33% +81.1% 5.06 ± 33% sched_debug.cfs_rq:/system.slice/containerd.service.tg_load_avg_contrib.avg 5.64 ± 29% +82.3% 10.28 ± 31% sched_debug.cfs_rq:/system.slice/containerd.service.tg_load_avg_contrib.max 2.27 ± 34% +93.6% 4.40 ± 37% sched_debug.cfs_rq:/system.slice/containerd.service.tg_load_avg_contrib.stddev 0.17 ± 8% -60.1% 0.07 ±122% sched_debug.cfs_rq:/system.slice/lkp-bootstrap.service.h_nr_running.stddev 180165 ± 8% -60.0% 72004 ±122% sched_debug.cfs_rq:/system.slice/lkp-bootstrap.service.load.stddev 187.40 ± 15% -67.8% 60.33 ±122% sched_debug.cfs_rq:/system.slice/lkp-bootstrap.service.load_avg.stddev 0.17 ± 8% -60.1% 0.07 ±122% sched_debug.cfs_rq:/system.slice/lkp-bootstrap.service.nr_running.stddev 6.76 ± 69% -76.3% 1.60 ±129% sched_debug.cfs_rq:/system.slice/lkp-bootstrap.service.removed.load_avg.avg 147.10 ± 62% -79.3% 30.38 ±122% sched_debug.cfs_rq:/system.slice/lkp-bootstrap.service.removed.load_avg.max 28.32 ± 61% -76.8% 6.58 ±124% sched_debug.cfs_rq:/system.slice/lkp-bootstrap.service.removed.load_avg.stddev 2.89 ± 77% -75.8% 0.70 ±124% sched_debug.cfs_rq:/system.slice/lkp-bootstrap.service.removed.runnable_avg.avg 57.35 ± 50% -73.5% 15.17 ±122% sched_debug.cfs_rq:/system.slice/lkp-bootstrap.service.removed.runnable_avg.max 11.43 ± 64% -73.8% 2.99 ±122% sched_debug.cfs_rq:/system.slice/lkp-bootstrap.service.removed.runnable_avg.stddev 2.87 ± 77% -75.7% 0.70 ±124% sched_debug.cfs_rq:/system.slice/lkp-bootstrap.service.removed.util_avg.avg 55.94 ± 49% -72.9% 15.17 ±122% sched_debug.cfs_rq:/system.slice/lkp-bootstrap.service.removed.util_avg.max 11.28 ± 64% -73.5% 2.99 ±122% sched_debug.cfs_rq:/system.slice/lkp-bootstrap.service.removed.util_avg.stddev 69.45 ± 22% -63.6% 25.26 ±122% sched_debug.cfs_rq:/system.slice/lkp-bootstrap.service.runnable_avg.stddev 11.31 ± 7% -64.9% 3.97 ±122% sched_debug.cfs_rq:/system.slice/lkp-bootstrap.service.se->avg.load_avg.avg 60.48 ± 40% -83.8% 9.82 ±122% sched_debug.cfs_rq:/system.slice/lkp-bootstrap.service.se->avg.load_avg.max 9.81 ± 46% -89.7% 1.01 ±122% sched_debug.cfs_rq:/system.slice/lkp-bootstrap.service.se->avg.load_avg.stddev 69.41 ± 22% -63.6% 25.24 ±122% sched_debug.cfs_rq:/system.slice/lkp-bootstrap.service.se->avg.runnable_avg.stddev 43.67 ± 37% -73.2% 11.72 ±126% sched_debug.cfs_rq:/system.slice/lkp-bootstrap.service.se->avg.util_avg.stddev 25759 ± 39% -80.9% 4908 ±122% sched_debug.cfs_rq:/system.slice/lkp-bootstrap.service.se->load.weight.avg 102176 ± 47% -84.0% 16350 ±125% sched_debug.cfs_rq:/system.slice/lkp-bootstrap.service.se->load.weight.max 23115 ± 57% -92.6% 1718 ±126% sched_debug.cfs_rq:/system.slice/lkp-bootstrap.service.se->load.weight.stddev 850.39 ± 47% -74.6% 216.04 ±123% sched_debug.cfs_rq:/system.slice/lkp-bootstrap.service.tg_load_avg.stddev 188.15 ± 15% -68.1% 60.09 ±122% sched_debug.cfs_rq:/system.slice/lkp-bootstrap.service.tg_load_avg_contrib.stddev 43.66 ± 37% -73.2% 11.72 ±126% sched_debug.cfs_rq:/system.slice/lkp-bootstrap.service.util_avg.stddev 835849 +17.1% 979195 sched_debug.cpu.avg_idle.avg 390705 ± 8% +27.5% 498342 ± 8% sched_debug.cpu.avg_idle.min 112505 ± 5% -37.1% 70769 ± 6% sched_debug.cpu.avg_idle.stddev 13678 ± 52% -34.9% 8908 sched_debug.cpu.curr->pid.max 839.26 ± 43% -46.1% 452.62 ± 21% sched_debug.cpu.curr->pid.stddev 0.81 ± 9% +12.2% 0.91 ± 2% sched_debug.cpu.nr_running.min 15.22 ± 26% +51.7% 23.09 ± 11% sched_debug.cpu.nr_uninterruptible.max -17.17 +58.4% -27.19 sched_debug.cpu.nr_uninterruptible.min 5.71 ± 16% +31.9% 7.54 ± 10% sched_debug.cpu.nr_uninterruptible.stddev Disclaimer: Results have been estimated based on internal Intel analysis and are provided for informational purposes only. Any difference in system hardware or software design or configuration may affect actual performance. -- 0-DAY CI Kernel Test Service https://github.com/intel/lkp-tests/wiki