On Tue, Dec 20, 2022 at 08:48:53PM +0000, Liam Howlett wrote: > * Yang Shi <shy828301@xxxxxxxxx> [221220 13:04]: > > On Mon, Dec 19, 2022 at 3:30 AM kernel test robot <yujie.liu@xxxxxxxxx> wrote: > > > > > > Greetings, > > > > > > Please note that we reported a regression in will-it-scale malloc1 > > > benchmark on below commit > > > f35b5d7d676e ("mm: align larger anonymous mappings on THP boundaries") > > > at > > > https://lore.kernel.org/all/202210181535.7144dd15-yujie.liu@xxxxxxxxx/ > > > and Nathan reported a kbuild slowdown under clang toolchain at > > > https://lore.kernel.org/all/Y1DNQaoPWxE+rGce@dev-arch.thelio-3990X/ > > > That commit was finally reverted. > > > > > > When we tested the revert commit, the score in malloc1 benchmark > > > recovered, but we observed another regression in mmap1 benchmark. > > > > > > "Yin, Fengwei" helped to check and got below clues: > > > > > > 1. The regression is related with the VMA merge with prev/next > > > VMA when doing mmap. > > > > > > 2. Before the patch reverted, almost all the VMA for 128M mapping > > > can't be merged with prev/next VMA. So always create new VMA. > > > With the patch reverted, most VMA for 128 mapping can be merged. > > > > > > It looks like VMA merging introduce more latency comparing to > > > creating new VMA. > > > > > > 3. If force to create new VMA with patch reverted, the result of > > > mmap1_thread is restored. > > > > > > 4. The thp_get_unmapped_area() adds a padding to request mapping > > > length. The padding is 2M in general. I believe this padding > > > break VMA merging behavior. > > > > > > 5. No idea about why the difference of the two path (VMA merging > > > vs New VMA) is not shown in perf data > > > > IIRC thp_get_unmapped_area() has been behaving like that for years. > > The other change between the problematic commit and the revert commit, > > which might have an impact to VMA merging, is maple tree. Did you try to > > bisect further? > > > > There was also the work done to vma_merge(). Vlastimil (added to Cc) > tracked down an issue with mremap() quite recently [1], which sounds a > lot like what is happening here - especially with the padding. > > > > > BTW, is this similar to > > https://lore.kernel.org/linux-mm/20221219180857.u6opzhqqbbfxdj3h@revolver/T/#t > > ? > > Yes, it looks to be similar. I'm surprised the mmap1 benchmark was > altered with this commit, or am I reading this email incorrectly? We caught two mmap1 regressions on mailine, please see the data below: 830b3c68c1fb1 Linux 6.1 2085 2355 2088 76dcd734eca23 Linux 6.1-rc8 2093 2082 2094 2073 2304 2088 0ba09b1733878 Revert "mm: align larger anonymous mappings on THP boundaries" 2124 2286 2086 2114 2065 2081 23393c6461422 char: tpm: Protect tpm_pm_suspend with locks 2756 2711 2689 2696 2660 2665 b7b275e60bcd5 Linux 6.1-rc7 2670 2656 2720 2691 2667 ... 9abf2313adc1c Linux 6.1-rc1 2725 2717 2690 2691 2710 3b0e81a1cdc9a mmap: change zeroing of maple tree in __vma_adjust() 2736 2781 2748 524e00b36e8c5 mm: remove rb tree. 2747 2744 2747 0c563f1480435 proc: remove VMA rbtree use from nommu d0cf3dd47f0d5 damon: convert __damon_va_three_regions to use the VMA iterator 3499a13168da6 mm/mmap: use maple tree for unmapped_area{_topdown} 7fdbd37da5c6f mm/mmap: use the maple tree for find_vma_prev() instead of the rbtree f39af05949a42 mm: add VMA iterator d4af56c5c7c67 mm: start tracking VMAs with maple tree e15e06a839232 lib/test_maple_tree: add testing for maple tree 4638 4628 4502 9832fb87834e2 mm/demotion: expose memory tier details via sysfs 4625 4509 4548 4fe89d07dcc28 Linux 6.0 4385 4205 4348 4228 4504 The first regression was between v6.0 and v6.1-rc1. The score dropped from 4600 to 2700, and bisected to the patches switching from rb tree to maple tree. This was reported at https://lore.kernel.org/oe-lkp/202212191714.524e00b3-yujie.liu@xxxxxxxxx/ Thanks for the explanation that it is an expected regression as a trade off to benefit read performance. The second regression was between v6.1-rc7 and v6.1-rc8. The score dropped from 2700 to 2100, and bisected to this "Revert "mm: align larger anonymous mappings on THP boundaries"" commit. > The trace below does not seem to show what RedHad [2] found in its > testing. > > [1]. https://lore.kernel.org/all/20221216163227.24648-1-vbabka@xxxxxxx/T/#u > [2]. https://bugzilla.redhat.com/show_bug.cgi?id=2149636 Thanks for the info in [2], looks it is an overall regression between v6.0 and v6.1, not sure if it could be the result of adding up the two regressions above? -- Best Regards, Yujie > > > > > > > > Please check below report for details. > > > > > > > > > FYI, we noticed a -21.1% regression of will-it-scale.per_thread_ops due to commit: > > > > > > commit: 0ba09b1733878afe838fe35c310715fda3d46428 ("Revert "mm: align larger anonymous mappings on THP boundaries"") > > > https://git.kernel.org/cgit/linux/kernel/git/torvalds/linux.git master > > > > > > in testcase: will-it-scale > > > on test machine: 104 threads 2 sockets (Skylake) with 192G memory > > > with following parameters: > > > > > > nr_task: 50% > > > mode: thread > > > test: mmap1 > > > cpufreq_governor: performance > > > > > > test-description: Will It Scale takes a testcase and runs it from 1 through to n parallel copies to see if the testcase will scale. It builds both a process and threads based test in order to see any differences between the two. > > > test-url: https://github.com/antonblanchard/will-it-scale > > > > > > In addition to that, the commit also has significant impact on the following tests: > > > > > > +------------------+------------------------------------------------------------------------------------------------+ > > > | testcase: change | will-it-scale: will-it-scale.per_process_ops 1943.6% improvement | > > > | test machine | 128 threads 4 sockets Intel(R) Xeon(R) Gold 6338 CPU @ 2.00GHz (Ice Lake) with 256G memory | > > > | test parameters | cpufreq_governor=performance | > > > | | mode=process | > > > | | nr_task=50% | > > > | | test=malloc1 | > > > +------------------+------------------------------------------------------------------------------------------------+ > > > | testcase: change | unixbench: unixbench.score 2.6% improvement | > > > | test machine | 16 threads 1 sockets Intel(R) Xeon(R) E-2278G CPU @ 3.40GHz (Coffee Lake) with 32G memory | > > > | test parameters | cpufreq_governor=performance | > > > | | nr_task=30% | > > > | | runtime=300s | > > > | | test=shell8 | > > > +------------------+------------------------------------------------------------------------------------------------+ > > > | testcase: change | phoronix-test-suite: phoronix-test-suite.build-eigen.0.seconds 9.1% regression | > > > | test machine | 96 threads 2 sockets Intel(R) Xeon(R) Gold 6252 CPU @ 2.10GHz (Cascade Lake) with 512G memory | > > > | test parameters | cpufreq_governor=performance | > > > | | test=build-eigen-1.1.0 | > > > +------------------+------------------------------------------------------------------------------------------------+ > > > | testcase: change | will-it-scale: will-it-scale.per_process_ops 2882.9% improvement | > > > | test machine | 88 threads 2 sockets Intel(R) Xeon(R) Gold 6238M CPU @ 2.10GHz (Cascade Lake) with 128G memory | > > > | test parameters | cpufreq_governor=performance | > > > | | mode=process | > > > | | nr_task=100% | > > > | | test=malloc1 | > > > +------------------+------------------------------------------------------------------------------------------------+ > > > | testcase: change | will-it-scale: will-it-scale.per_process_ops 12.7% improvement | > > > | test machine | 88 threads 2 sockets Intel(R) Xeon(R) Gold 6238M CPU @ 2.10GHz (Cascade Lake) with 128G memory | > > > | test parameters | cpufreq_governor=performance | > > > | | mode=process | > > > | | nr_task=50% | > > > | | test=mmap1 | > > > +------------------+------------------------------------------------------------------------------------------------+ > > > | testcase: change | stress-ng: stress-ng.pthread.ops_per_sec 600.6% improvement | > > > | test machine | 128 threads 2 sockets Intel(R) Xeon(R) Platinum 8358 CPU @ 2.60GHz (Ice Lake) with 128G memory | > > > | test parameters | class=scheduler | > > > | | cpufreq_governor=performance | > > > | | nr_threads=100% | > > > | | sc_pid_max=4194304 | > > > | | test=pthread | > > > | | testtime=60s | > > > +------------------+------------------------------------------------------------------------------------------------+ > > > | testcase: change | will-it-scale: will-it-scale.per_process_ops 601.0% improvement | > > > | test machine | 104 threads 2 sockets (Skylake) with 192G memory | > > > | test parameters | cpufreq_governor=performance | > > > | | mode=process | > > > | | nr_task=50% | > > > | | test=malloc1 | > > > +------------------+------------------------------------------------------------------------------------------------+ > > > > > > > > > Details are as below: > > > > > > ========================================================================================= > > > compiler/cpufreq_governor/kconfig/mode/nr_task/rootfs/tbox_group/test/testcase: > > > gcc-11/performance/x86_64-rhel-8.3/thread/50%/debian-11.1-x86_64-20220510.cgz/lkp-skl-fpga01/mmap1/will-it-scale > > > > > > commit: > > > 23393c6461 ("char: tpm: Protect tpm_pm_suspend with locks") > > > 0ba09b1733 ("Revert "mm: align larger anonymous mappings on THP boundaries"") > > > > > > 23393c6461422df5 0ba09b1733878afe838fe35c310 > > > ---------------- --------------------------- > > > %stddev %change %stddev > > > \ | \ > > > 140227 -21.1% 110582 ą 3% will-it-scale.52.threads > > > 49.74 +0.1% 49.78 will-it-scale.52.threads_idle > > > 2696 -21.1% 2126 ą 3% will-it-scale.per_thread_ops > > > 301.30 -0.0% 301.26 will-it-scale.time.elapsed_time > > > 301.30 -0.0% 301.26 will-it-scale.time.elapsed_time.max > > > 3.67 ą 71% -22.7% 2.83 ą 47% will-it-scale.time.involuntary_context_switches > > > 0.67 ą165% -75.0% 0.17 ą223% will-it-scale.time.major_page_faults > > > 9772 -0.7% 9702 will-it-scale.time.maximum_resident_set_size > > > 7274 -0.3% 7254 will-it-scale.time.minor_page_faults > > > 4096 +0.0% 4096 will-it-scale.time.page_size > > > 0.04 ą 16% -4.0% 0.04 will-it-scale.time.system_time > > > 0.06 ą 24% -11.8% 0.05 ą 16% will-it-scale.time.user_time > > > 102.83 +1.9% 104.83 ą 2% will-it-scale.time.voluntary_context_switches > > > 140227 -21.1% 110582 ą 3% will-it-scale.workload > > > 1.582e+10 +0.1% 1.584e+10 cpuidle..time > > > 33034032 -0.0% 33021393 cpuidle..usage > > > 10.00 +0.0% 10.00 dmesg.bootstage:last > > > 172.34 +0.1% 172.58 dmesg.timestamp:last > > > 10.00 +0.0% 10.00 kmsg.bootstage:last > > > 172.34 +0.1% 172.58 kmsg.timestamp:last > > > 362.22 +0.0% 362.25 uptime.boot > > > 21363 +0.1% 21389 uptime.idle > > > 55.94 +0.2% 56.06 boot-time.boot > > > 38.10 +0.2% 38.19 boot-time.dhcp > > > 5283 +0.2% 5295 boot-time.idle > > > 1.11 -0.1% 1.11 boot-time.smp_boot > > > 50.14 +0.0 50.16 mpstat.cpu.all.idle% > > > 0.03 ą223% -0.0 0.00 ą223% mpstat.cpu.all.iowait% > > > 1.02 +0.0 1.03 mpstat.cpu.all.irq% > > > 0.03 ą 4% -0.0 0.02 mpstat.cpu.all.soft% > > > 48.59 +0.0 48.61 mpstat.cpu.all.sys% > > > 0.20 ą 2% -0.0 0.17 ą 4% mpstat.cpu.all.usr% > > > 0.00 -100.0% 0.00 numa-numastat.node0.interleave_hit > > > 328352 ą 15% -7.2% 304842 ą 20% numa-numastat.node0.local_node > > > 374230 ą 6% -4.2% 358578 ą 7% numa-numastat.node0.numa_hit > > > 45881 ą 75% +17.1% 53735 ą 69% numa-numastat.node0.other_node > > > 0.00 -100.0% 0.00 numa-numastat.node1.interleave_hit > > > 381812 ą 13% +5.9% 404461 ą 14% numa-numastat.node1.local_node > > > 430007 ą 5% +3.4% 444810 ą 5% numa-numastat.node1.numa_hit > > > 48195 ą 71% -16.3% 40348 ą 92% numa-numastat.node1.other_node > > > 301.30 -0.0% 301.26 time.elapsed_time > > > 301.30 -0.0% 301.26 time.elapsed_time.max > > > 3.67 ą 71% -22.7% 2.83 ą 47% time.involuntary_context_switches > > > 0.67 ą165% -75.0% 0.17 ą223% time.major_page_faults > > > 9772 -0.7% 9702 time.maximum_resident_set_size > > > 7274 -0.3% 7254 time.minor_page_faults > > > 4096 +0.0% 4096 time.page_size > > > 0.04 ą 16% -4.0% 0.04 time.system_time > > > 0.06 ą 24% -11.8% 0.05 ą 16% time.user_time > > > 102.83 +1.9% 104.83 ą 2% time.voluntary_context_switches > > > 50.00 +0.0% 50.00 vmstat.cpu.id > > > 49.00 +0.0% 49.00 vmstat.cpu.sy > > > 0.00 -100.0% 0.00 vmstat.cpu.us > > > 0.00 -100.0% 0.00 vmstat.cpu.wa > > > 12.50 ą100% -66.7% 4.17 ą223% vmstat.io.bi > > > 3.33 ą141% -55.0% 1.50 ą223% vmstat.io.bo > > > 6.00 ą 47% -16.7% 5.00 ą 44% vmstat.memory.buff > > > 4150651 -0.1% 4148516 vmstat.memory.cache > > > 1.912e+08 +0.1% 1.913e+08 vmstat.memory.free > > > 0.00 -100.0% 0.00 vmstat.procs.b > > > 50.50 -0.3% 50.33 vmstat.procs.r > > > 8274 ą 2% +1.2% 8371 ą 4% vmstat.system.cs > > > 211078 -0.1% 210826 vmstat.system.in > > > 1399 +0.0% 1399 turbostat.Avg_MHz > > > 50.12 +0.0 50.13 turbostat.Busy% > > > 2799 -0.0% 2798 turbostat.Bzy_MHz > > > 208677 ą 13% +1112.3% 2529776 ą194% turbostat.C1 > > > 0.03 ą 89% +0.3 0.36 ą203% turbostat.C1% > > > 27078371 ą 15% -22.0% 21125809 ą 51% turbostat.C1E > > > 37.41 ą 33% -9.4 28.04 ą 62% turbostat.C1E% > > > 5088326 ą 84% +63.1% 8298766 ą 77% turbostat.C6 > > > 12.59 ą 99% +9.1 21.69 ą 78% turbostat.C6% > > > 49.79 -0.1% 49.75 turbostat.CPU%c1 > > > 0.08 ą 71% +37.3% 0.12 ą 78% turbostat.CPU%c6 > > > 43.67 -0.4% 43.50 turbostat.CoreTmp > > > 0.03 +0.0% 0.03 turbostat.IPC > > > 64483530 -0.2% 64338768 turbostat.IRQ > > > 647657 ą 2% +63.2% 1057048 ą 98% turbostat.POLL > > > 0.01 +0.0 0.05 ą178% turbostat.POLL% > > > 0.01 ą223% +200.0% 0.04 ą147% turbostat.Pkg%pc2 > > > 0.01 ą223% +140.0% 0.02 ą165% turbostat.Pkg%pc6 > > > 44.17 +0.4% 44.33 turbostat.PkgTmp > > > 284.98 +0.1% 285.28 turbostat.PkgWatt > > > 26.78 +0.4% 26.89 turbostat.RAMWatt > > > 2095 +0.0% 2095 turbostat.TSC_MHz > > > 49585 ą 7% +1.1% 50139 ą 7% meminfo.Active > > > 49182 ą 7% +1.4% 49889 ą 7% meminfo.Active(anon) > > > 402.33 ą 99% -37.9% 250.00 ą123% meminfo.Active(file) > > > 290429 -33.7% 192619 meminfo.AnonHugePages > > > 419654 -25.9% 311054 meminfo.AnonPages > > > 6.00 ą 47% -16.7% 5.00 ą 44% meminfo.Buffers > > > 4026046 -0.1% 4023990 meminfo.Cached > > > 98360160 +0.0% 98360160 meminfo.CommitLimit > > > 4319751 +0.4% 4337801 meminfo.Committed_AS > > > 1.877e+08 -0.1% 1.875e+08 meminfo.DirectMap1G > > > 14383445 ą 12% +0.7% 14491306 ą 4% meminfo.DirectMap2M > > > 1042426 ą 9% +6.4% 1109328 ą 7% meminfo.DirectMap4k > > > 4.00 ą141% -50.0% 2.00 ą223% meminfo.Dirty > > > 2048 +0.0% 2048 meminfo.Hugepagesize > > > 434675 -26.3% 320518 meminfo.Inactive > > > 431330 -26.0% 319346 meminfo.Inactive(anon) > > > 3344 ą 95% -65.0% 1171 ą186% meminfo.Inactive(file) > > > 124528 -0.1% 124460 meminfo.KReclaimable > > > 18433 +0.7% 18559 meminfo.KernelStack > > > 40185 ą 2% -0.9% 39837 meminfo.Mapped > > > 1.903e+08 +0.1% 1.904e+08 meminfo.MemAvailable > > > 1.912e+08 +0.1% 1.913e+08 meminfo.MemFree > > > 1.967e+08 +0.0% 1.967e+08 meminfo.MemTotal > > > 5569412 -1.8% 5466754 meminfo.Memused > > > 4763 -5.7% 4489 meminfo.PageTables > > > 51956 +0.0% 51956 meminfo.Percpu > > > 124528 -0.1% 124460 meminfo.SReclaimable > > > 197128 +0.1% 197293 meminfo.SUnreclaim > > > 57535 ą 7% +0.8% 57986 ą 6% meminfo.Shmem > > > 321657 +0.0% 321754 meminfo.Slab > > > 3964769 -0.0% 3964586 meminfo.Unevictable > > > 3.436e+10 +0.0% 3.436e+10 meminfo.VmallocTotal > > > 280612 +0.1% 280841 meminfo.VmallocUsed > > > 6194619 -2.0% 6071944 meminfo.max_used_kB > > > 2626 ą 28% -7.7% 2423 ą 11% numa-meminfo.node0.Active > > > 2361 ą 20% -5.3% 2236 ą 10% numa-meminfo.node0.Active(anon) > > > 264.67 ą117% -29.5% 186.67 ą152% numa-meminfo.node0.Active(file) > > > 135041 ą 20% -22.4% 104774 ą 42% numa-meminfo.node0.AnonHugePages > > > 197759 ą 18% -20.4% 157470 ą 35% numa-meminfo.node0.AnonPages > > > 235746 ą 19% -11.8% 207988 ą 29% numa-meminfo.node0.AnonPages.max > > > 2.00 ą223% +0.0% 2.00 ą223% numa-meminfo.node0.Dirty > > > 1386137 ą123% +89.5% 2626100 ą 67% numa-meminfo.node0.FilePages > > > 202317 ą 19% -21.0% 159846 ą 36% numa-meminfo.node0.Inactive > > > 200223 ą 19% -20.7% 158765 ą 35% numa-meminfo.node0.Inactive(anon) > > > 2093 ą129% -48.4% 1080 ą200% numa-meminfo.node0.Inactive(file) > > > 46369 ą 57% +43.5% 66525 ą 41% numa-meminfo.node0.KReclaimable > > > 9395 ą 4% +4.6% 9822 ą 5% numa-meminfo.node0.KernelStack > > > 14343 ą101% +65.1% 23681 ą 58% numa-meminfo.node0.Mapped > > > 95532160 -1.3% 94306066 numa-meminfo.node0.MemFree > > > 97681544 +0.0% 97681544 numa-meminfo.node0.MemTotal > > > 2149382 ą 82% +57.0% 3375476 ą 53% numa-meminfo.node0.MemUsed > > > 2356 ą 21% -9.9% 2122 ą 9% numa-meminfo.node0.PageTables > > > 46369 ą 57% +43.5% 66525 ą 41% numa-meminfo.node0.SReclaimable > > > 109141 ą 6% +1.5% 110817 ą 7% numa-meminfo.node0.SUnreclaim > > > 4514 ą 34% -22.4% 3505 ą 30% numa-meminfo.node0.Shmem > > > 155511 ą 18% +14.0% 177344 ą 14% numa-meminfo.node0.Slab > > > 1379264 ą124% +90.1% 2621327 ą 67% numa-meminfo.node0.Unevictable > > > 46974 ą 8% +1.5% 47665 ą 7% numa-meminfo.node1.Active > > > 46837 ą 8% +1.6% 47601 ą 7% numa-meminfo.node1.Active(anon) > > > 137.33 ą219% -54.0% 63.17 ą 85% numa-meminfo.node1.Active(file) > > > 155559 ą 18% -43.5% 87865 ą 52% numa-meminfo.node1.AnonHugePages > > > 222077 ą 16% -30.8% 153725 ą 36% numa-meminfo.node1.AnonPages > > > 304080 ą 17% -27.5% 220544 ą 28% numa-meminfo.node1.AnonPages.max > > > 2.00 ą223% -100.0% 0.00 numa-meminfo.node1.Dirty > > > 2639873 ą 65% -47.0% 1397913 ą126% numa-meminfo.node1.FilePages > > > 232481 ą 17% -30.8% 160887 ą 34% numa-meminfo.node1.Inactive > > > 231228 ą 16% -30.5% 160796 ą 34% numa-meminfo.node1.Inactive(anon) > > > 1252 ą213% -92.8% 90.33 ą 96% numa-meminfo.node1.Inactive(file) > > > 78155 ą 34% -25.9% 57927 ą 47% numa-meminfo.node1.KReclaimable > > > 9041 ą 4% -3.3% 8740 ą 5% numa-meminfo.node1.KernelStack > > > 25795 ą 55% -37.5% 16118 ą 85% numa-meminfo.node1.Mapped > > > 95619356 +1.4% 96947357 numa-meminfo.node1.MemFree > > > 99038776 +0.0% 99038776 numa-meminfo.node1.MemTotal > > > 3419418 ą 52% -38.8% 2091417 ą 85% numa-meminfo.node1.MemUsed > > > 2405 ą 21% -1.5% 2369 ą 7% numa-meminfo.node1.PageTables > > > 78155 ą 34% -25.9% 57927 ą 47% numa-meminfo.node1.SReclaimable > > > 87984 ą 7% -1.7% 86475 ą 9% numa-meminfo.node1.SUnreclaim > > > 52978 ą 9% +2.9% 54500 ą 8% numa-meminfo.node1.Shmem > > > 166140 ą 16% -13.1% 144403 ą 17% numa-meminfo.node1.Slab > > > 2585504 ą 66% -48.0% 1343258 ą131% numa-meminfo.node1.Unevictable > > > 486.17 ą 9% +6.8% 519.17 ą 7% proc-vmstat.direct_map_level2_splits > > > 8.00 ą 22% +2.1% 8.17 ą 8% proc-vmstat.direct_map_level3_splits > > > 12303 ą 7% +1.3% 12461 ą 7% proc-vmstat.nr_active_anon > > > 100.50 ą 99% -37.8% 62.50 ą123% proc-vmstat.nr_active_file > > > 104906 -25.9% 77785 proc-vmstat.nr_anon_pages > > > 141.00 -33.6% 93.67 proc-vmstat.nr_anon_transparent_hugepages > > > 264.00 ą141% -54.3% 120.67 ą223% proc-vmstat.nr_dirtied > > > 1.00 ą141% -50.0% 0.50 ą223% proc-vmstat.nr_dirty > > > 4750146 +0.1% 4752612 proc-vmstat.nr_dirty_background_threshold > > > 9511907 +0.1% 9516846 proc-vmstat.nr_dirty_threshold > > > 1006517 -0.1% 1005995 proc-vmstat.nr_file_pages > > > 47787985 +0.1% 47813269 proc-vmstat.nr_free_pages > > > 107821 -25.9% 79869 proc-vmstat.nr_inactive_anon > > > 836.17 ą 95% -65.1% 292.17 ą186% proc-vmstat.nr_inactive_file > > > 18434 +0.7% 18563 proc-vmstat.nr_kernel_stack > > > 10033 ą 2% -1.1% 9924 proc-vmstat.nr_mapped > > > 1190 -5.7% 1122 proc-vmstat.nr_page_table_pages > > > 14387 ą 7% +0.7% 14493 ą 6% proc-vmstat.nr_shmem > > > 31131 -0.1% 31114 proc-vmstat.nr_slab_reclaimable > > > 49281 +0.1% 49323 proc-vmstat.nr_slab_unreclaimable > > > 991192 -0.0% 991146 proc-vmstat.nr_unevictable > > > 264.00 ą141% -54.3% 120.67 ą223% proc-vmstat.nr_written > > > 12303 ą 7% +1.3% 12461 ą 7% proc-vmstat.nr_zone_active_anon > > > 100.50 ą 99% -37.8% 62.50 ą123% proc-vmstat.nr_zone_active_file > > > 107821 -25.9% 79869 proc-vmstat.nr_zone_inactive_anon > > > 836.17 ą 95% -65.1% 292.17 ą186% proc-vmstat.nr_zone_inactive_file > > > 991192 -0.0% 991146 proc-vmstat.nr_zone_unevictable > > > 1.00 ą141% -50.0% 0.50 ą223% proc-vmstat.nr_zone_write_pending > > > 17990 ą 21% -17.6% 14820 ą 46% proc-vmstat.numa_hint_faults > > > 7847 ą 37% -41.5% 4588 ą 26% proc-vmstat.numa_hint_faults_local > > > 806662 +0.3% 809070 proc-vmstat.numa_hit > > > 488.50 ą 13% -73.4% 130.17 ą 22% proc-vmstat.numa_huge_pte_updates > > > 0.00 -100.0% 0.00 proc-vmstat.numa_interleave > > > 712588 -0.2% 711419 proc-vmstat.numa_local > > > 94077 +0.0% 94084 proc-vmstat.numa_other > > > 18894 ą 67% -3.1% 18303 ą 41% proc-vmstat.numa_pages_migrated > > > 337482 ą 10% -59.0% 138314 ą 10% proc-vmstat.numa_pte_updates > > > 61815 -1.6% 60823 proc-vmstat.pgactivate > > > 0.00 -100.0% 0.00 proc-vmstat.pgalloc_dma32 > > > 933601 -3.8% 898485 proc-vmstat.pgalloc_normal > > > 899579 -0.5% 895253 proc-vmstat.pgfault > > > 896972 -3.9% 861819 proc-vmstat.pgfree > > > 18894 ą 67% -3.1% 18303 ą 41% proc-vmstat.pgmigrate_success > > > 3845 ą100% -66.8% 1277 ą223% proc-vmstat.pgpgin > > > 1064 ą141% -54.3% 486.67 ą223% proc-vmstat.pgpgout > > > 40396 -0.6% 40172 proc-vmstat.pgreuse > > > 105.50 -9.2% 95.83 ą 5% proc-vmstat.thp_collapse_alloc > > > 57.00 -87.4% 7.17 ą 5% proc-vmstat.thp_deferred_split_page > > > 74.83 -72.4% 20.67 ą 4% proc-vmstat.thp_fault_alloc > > > 19.50 ą105% -15.4% 16.50 ą 71% proc-vmstat.thp_migration_success > > > 57.00 -87.4% 7.17 ą 5% proc-vmstat.thp_split_pmd > > > 0.00 -100.0% 0.00 proc-vmstat.thp_zero_page_alloc > > > 17.00 +0.0% 17.00 proc-vmstat.unevictable_pgs_culled > > > 589.83 ą 21% -5.2% 559.00 ą 10% numa-vmstat.node0.nr_active_anon > > > 66.00 ą117% -29.3% 46.67 ą152% numa-vmstat.node0.nr_active_file > > > 49406 ą 18% -20.3% 39355 ą 35% numa-vmstat.node0.nr_anon_pages > > > 65.17 ą 21% -22.0% 50.83 ą 42% numa-vmstat.node0.nr_anon_transparent_hugepages > > > 132.00 ą223% -8.6% 120.67 ą223% numa-vmstat.node0.nr_dirtied > > > 0.50 ą223% +0.0% 0.50 ą223% numa-vmstat.node0.nr_dirty > > > 346534 ą123% +89.5% 656525 ą 67% numa-vmstat.node0.nr_file_pages > > > 23883055 -1.3% 23576561 numa-vmstat.node0.nr_free_pages > > > 50051 ą 19% -20.7% 39679 ą 35% numa-vmstat.node0.nr_inactive_anon > > > 522.67 ą129% -48.4% 269.67 ą200% numa-vmstat.node0.nr_inactive_file > > > 0.00 -100.0% 0.00 numa-vmstat.node0.nr_isolated_anon > > > 9392 ą 4% +4.6% 9823 ą 5% numa-vmstat.node0.nr_kernel_stack > > > 3594 ą101% +64.8% 5922 ą 58% numa-vmstat.node0.nr_mapped > > > 587.83 ą 21% -9.8% 530.00 ą 9% numa-vmstat.node0.nr_page_table_pages > > > 1129 ą 34% -22.4% 876.67 ą 30% numa-vmstat.node0.nr_shmem > > > 11591 ą 57% +43.5% 16631 ą 41% numa-vmstat.node0.nr_slab_reclaimable > > > 27285 ą 6% +1.5% 27704 ą 7% numa-vmstat.node0.nr_slab_unreclaimable > > > 344815 ą124% +90.1% 655331 ą 67% numa-vmstat.node0.nr_unevictable > > > 132.00 ą223% -8.6% 120.67 ą223% numa-vmstat.node0.nr_written > > > 589.83 ą 21% -5.2% 559.00 ą 10% numa-vmstat.node0.nr_zone_active_anon > > > 66.00 ą117% -29.3% 46.67 ą152% numa-vmstat.node0.nr_zone_active_file > > > 50051 ą 19% -20.7% 39679 ą 35% numa-vmstat.node0.nr_zone_inactive_anon > > > 522.67 ą129% -48.4% 269.67 ą200% numa-vmstat.node0.nr_zone_inactive_file > > > 344815 ą124% +90.1% 655331 ą 67% numa-vmstat.node0.nr_zone_unevictable > > > 0.50 ą223% +0.0% 0.50 ą223% numa-vmstat.node0.nr_zone_write_pending > > > 374134 ą 6% -4.1% 358690 ą 7% numa-vmstat.node0.numa_hit > > > 0.00 -100.0% 0.00 numa-vmstat.node0.numa_interleave > > > 328256 ą 15% -7.1% 304955 ą 20% numa-vmstat.node0.numa_local > > > 45881 ą 75% +17.1% 53735 ą 69% numa-vmstat.node0.numa_other > > > 11706 ą 8% +1.7% 11901 ą 7% numa-vmstat.node1.nr_active_anon > > > 34.17 ą219% -54.1% 15.67 ą 84% numa-vmstat.node1.nr_active_file > > > 55500 ą 16% -30.8% 38424 ą 36% numa-vmstat.node1.nr_anon_pages > > > 75.50 ą 18% -43.7% 42.50 ą 53% numa-vmstat.node1.nr_anon_transparent_hugepages > > > 132.00 ą223% -100.0% 0.00 numa-vmstat.node1.nr_dirtied > > > 0.50 ą223% -100.0% 0.00 numa-vmstat.node1.nr_dirty > > > 659985 ą 65% -47.0% 349484 ą126% numa-vmstat.node1.nr_file_pages > > > 23904828 +1.4% 24236871 numa-vmstat.node1.nr_free_pages > > > 57826 ą 16% -30.5% 40197 ą 34% numa-vmstat.node1.nr_inactive_anon > > > 313.00 ą213% -92.9% 22.33 ą 96% numa-vmstat.node1.nr_inactive_file > > > 9043 ą 4% -3.3% 8740 ą 5% numa-vmstat.node1.nr_kernel_stack > > > 6467 ą 55% -37.6% 4038 ą 85% numa-vmstat.node1.nr_mapped > > > 601.50 ą 21% -1.6% 591.83 ą 7% numa-vmstat.node1.nr_page_table_pages > > > 13261 ą 9% +2.8% 13630 ą 8% numa-vmstat.node1.nr_shmem > > > 19538 ą 34% -25.9% 14481 ą 47% numa-vmstat.node1.nr_slab_reclaimable > > > 21995 ą 7% -1.7% 21618 ą 9% numa-vmstat.node1.nr_slab_unreclaimable > > > 646375 ą 66% -48.0% 335813 ą131% numa-vmstat.node1.nr_unevictable > > > 132.00 ą223% -100.0% 0.00 numa-vmstat.node1.nr_written > > > 11706 ą 8% +1.7% 11901 ą 7% numa-vmstat.node1.nr_zone_active_anon > > > 34.17 ą219% -54.1% 15.67 ą 84% numa-vmstat.node1.nr_zone_active_file > > > 57826 ą 16% -30.5% 40197 ą 34% numa-vmstat.node1.nr_zone_inactive_anon > > > 313.00 ą213% -92.9% 22.33 ą 96% numa-vmstat.node1.nr_zone_inactive_file > > > 646375 ą 66% -48.0% 335813 ą131% numa-vmstat.node1.nr_zone_unevictable > > > 0.50 ą223% -100.0% 0.00 numa-vmstat.node1.nr_zone_write_pending > > > 429997 ą 5% +3.5% 444962 ą 5% numa-vmstat.node1.numa_hit > > > 0.00 -100.0% 0.00 numa-vmstat.node1.numa_interleave > > > 381801 ą 13% +6.0% 404613 ą 14% numa-vmstat.node1.numa_local > > > 48195 ą 71% -16.3% 40348 ą 92% numa-vmstat.node1.numa_other > > > 2.47 ą 2% -2.0% 2.42 ą 5% perf-stat.i.MPKI > > > 3.282e+09 +0.7% 3.305e+09 perf-stat.i.branch-instructions > > > 0.41 -0.1 0.33 perf-stat.i.branch-miss-rate% > > > 13547319 -16.6% 11300609 perf-stat.i.branch-misses > > > 42.88 +0.7 43.53 perf-stat.i.cache-miss-rate% > > > 17114713 ą 3% +1.4% 17346470 ą 5% perf-stat.i.cache-misses > > > 40081707 ą 2% -0.0% 40073189 ą 5% perf-stat.i.cache-references > > > 8192 ą 2% +1.4% 8311 ą 4% perf-stat.i.context-switches > > > 8.84 -0.8% 8.77 perf-stat.i.cpi > > > 104007 +0.0% 104008 perf-stat.i.cpu-clock > > > 1.446e+11 +0.1% 1.447e+11 perf-stat.i.cpu-cycles > > > 140.10 -1.0% 138.76 perf-stat.i.cpu-migrations > > > 8487 ą 3% -0.9% 8412 ą 6% perf-stat.i.cycles-between-cache-misses > > > 0.01 ą 6% -0.0 0.01 perf-stat.i.dTLB-load-miss-rate% > > > 434358 ą 3% -16.9% 360889 perf-stat.i.dTLB-load-misses > > > 4.316e+09 +1.3% 4.373e+09 perf-stat.i.dTLB-loads > > > 0.00 ą 15% -0.0 0.00 ą 9% perf-stat.i.dTLB-store-miss-rate% > > > 10408 ą 11% -2.6% 10135 ą 8% perf-stat.i.dTLB-store-misses > > > 4.302e+08 +5.5% 4.539e+08 perf-stat.i.dTLB-stores > > > 16.21 ą 2% -2.5 13.73 ą 18% perf-stat.i.iTLB-load-miss-rate% > > > 394805 ą 5% -26.0% 292089 ą 8% perf-stat.i.iTLB-load-misses > > > 2041963 ą 3% -8.3% 1872405 ą 12% perf-stat.i.iTLB-loads > > > 1.638e+10 +1.0% 1.654e+10 perf-stat.i.instructions > > > 41729 ą 6% +37.4% 57323 ą 8% perf-stat.i.instructions-per-iTLB-miss > > > 0.11 +0.8% 0.11 perf-stat.i.ipc > > > 0.01 ą 55% -1.5% 0.01 ą 85% perf-stat.i.major-faults > > > 1.39 +0.1% 1.39 perf-stat.i.metric.GHz > > > 468.46 ą 2% -1.5% 461.59 ą 4% perf-stat.i.metric.K/sec > > > 77.18 +1.3% 78.18 perf-stat.i.metric.M/sec > > > 2473 -0.0% 2472 perf-stat.i.minor-faults > > > 89.67 -0.5 89.18 perf-stat.i.node-load-miss-rate% > > > 5070484 -10.3% 4547670 perf-stat.i.node-load-misses > > > 585336 ą 2% -5.5% 553260 ą 8% perf-stat.i.node-loads > > > 98.73 +0.2 98.91 perf-stat.i.node-store-miss-rate% > > > 935187 +2.2% 955923 ą 3% perf-stat.i.node-store-misses > > > 13301 ą 8% -12.6% 11631 ą 5% perf-stat.i.node-stores > > > 2473 -0.0% 2472 perf-stat.i.page-faults > > > 104007 +0.0% 104008 perf-stat.i.task-clock > > > 2.45 ą 2% -1.0% 2.42 ą 5% perf-stat.overall.MPKI > > > 0.41 -0.1 0.34 perf-stat.overall.branch-miss-rate% > > > 42.68 +0.6 43.26 perf-stat.overall.cache-miss-rate% > > > 8.83 -0.9% 8.75 perf-stat.overall.cpi > > > 8459 ą 3% -1.0% 8372 ą 6% perf-stat.overall.cycles-between-cache-misses > > > 0.01 ą 3% -0.0 0.01 perf-stat.overall.dTLB-load-miss-rate% > > > 0.00 ą 11% -0.0 0.00 ą 8% perf-stat.overall.dTLB-store-miss-rate% > > > 16.19 ą 2% -2.5 13.73 ą 18% perf-stat.overall.iTLB-load-miss-rate% > > > 41644 ą 6% +37.0% 57047 ą 8% perf-stat.overall.instructions-per-iTLB-miss > > > 0.11 +0.9% 0.11 perf-stat.overall.ipc > > > 89.65 -0.5 89.15 perf-stat.overall.node-load-miss-rate% > > > 98.59 +0.2 98.78 perf-stat.overall.node-store-miss-rate% > > > 35314961 +28.0% 45213422 ą 3% perf-stat.overall.path-length > > > 3.272e+09 +0.7% 3.295e+09 perf-stat.ps.branch-instructions > > > 13563215 -16.5% 11329031 perf-stat.ps.branch-misses > > > 17059170 ą 3% +1.3% 17288798 ą 5% perf-stat.ps.cache-misses > > > 39960738 ą 2% -0.0% 39951411 ą 5% perf-stat.ps.cache-references > > > 8205 ą 2% +1.4% 8320 ą 4% perf-stat.ps.context-switches > > > 103658 -0.0% 103657 perf-stat.ps.cpu-clock > > > 1.441e+11 +0.1% 1.442e+11 perf-stat.ps.cpu-cycles > > > 140.16 -1.0% 138.77 perf-stat.ps.cpu-migrations > > > 433133 ą 3% -16.9% 359910 perf-stat.ps.dTLB-load-misses > > > 4.302e+09 +1.3% 4.359e+09 perf-stat.ps.dTLB-loads > > > 10392 ą 11% -2.6% 10120 ą 8% perf-stat.ps.dTLB-store-misses > > > 4.29e+08 +5.5% 4.527e+08 perf-stat.ps.dTLB-stores > > > 393499 ą 5% -26.0% 291118 ą 8% perf-stat.ps.iTLB-load-misses > > > 2035052 ą 3% -8.3% 1866106 ą 12% perf-stat.ps.iTLB-loads > > > 1.633e+10 +1.0% 1.649e+10 perf-stat.ps.instructions > > > 0.01 ą 55% +0.1% 0.01 ą 85% perf-stat.ps.major-faults > > > 2466 +0.0% 2466 perf-stat.ps.minor-faults > > > 5053378 -10.3% 4532205 perf-stat.ps.node-load-misses > > > 583428 ą 2% -5.5% 551516 ą 8% perf-stat.ps.node-loads > > > 932227 +2.2% 952780 ą 3% perf-stat.ps.node-store-misses > > > 13342 ą 8% -12.1% 11729 ą 6% perf-stat.ps.node-stores > > > 2466 +0.0% 2466 perf-stat.ps.page-faults > > > 103658 -0.0% 103657 perf-stat.ps.task-clock > > > 4.952e+12 +0.9% 4.994e+12 perf-stat.total.instructions > > > 10.88 ą223% -100.0% 0.00 sched_debug.cfs_rq:/.MIN_vruntime.avg > > > 1132 ą223% -100.0% 0.00 sched_debug.cfs_rq:/.MIN_vruntime.max > > > 0.00 +0.0% 0.00 sched_debug.cfs_rq:/.MIN_vruntime.min > > > 110.47 ą223% -100.0% 0.00 sched_debug.cfs_rq:/.MIN_vruntime.stddev > > > 0.53 ą 4% +7.4% 0.57 ą 4% sched_debug.cfs_rq:/.h_nr_running.avg > > > 1.03 ą 7% -3.2% 1.00 sched_debug.cfs_rq:/.h_nr_running.max > > > 0.45 ą 2% -1.9% 0.44 ą 3% sched_debug.cfs_rq:/.h_nr_running.stddev > > > 11896 ą 12% -0.1% 11883 ą 13% sched_debug.cfs_rq:/.load.avg > > > 123097 ą123% -80.1% 24487 ą 18% sched_debug.cfs_rq:/.load.max > > > 19029 ą 74% -49.9% 9525 ą 13% sched_debug.cfs_rq:/.load.stddev > > > 22.63 ą 23% +1.4% 22.93 ą 16% sched_debug.cfs_rq:/.load_avg.avg > > > 530.85 ą 73% -13.1% 461.19 ą 43% sched_debug.cfs_rq:/.load_avg.max > > > 73.53 ą 46% -7.1% 68.30 ą 33% sched_debug.cfs_rq:/.load_avg.stddev > > > 10.88 ą223% -100.0% 0.00 sched_debug.cfs_rq:/.max_vruntime.avg > > > 1132 ą223% -100.0% 0.00 sched_debug.cfs_rq:/.max_vruntime.max > > > 0.00 +0.0% 0.00 sched_debug.cfs_rq:/.max_vruntime.min > > > 110.47 ą223% -100.0% 0.00 sched_debug.cfs_rq:/.max_vruntime.stddev > > > 3883756 ą 13% +12.7% 4377466 ą 4% sched_debug.cfs_rq:/.min_vruntime.avg > > > 6993455 ą 10% +6.5% 7445221 ą 2% sched_debug.cfs_rq:/.min_vruntime.max > > > 219925 ą 60% +43.7% 315970 ą 71% sched_debug.cfs_rq:/.min_vruntime.min > > > 2240239 ą 11% +14.0% 2554847 ą 14% sched_debug.cfs_rq:/.min_vruntime.stddev > > > 0.53 ą 5% +7.5% 0.57 ą 4% sched_debug.cfs_rq:/.nr_running.avg > > > 1.03 ą 7% -3.2% 1.00 sched_debug.cfs_rq:/.nr_running.max > > > 0.45 ą 2% -1.9% 0.44 ą 3% sched_debug.cfs_rq:/.nr_running.stddev > > > 6.96 ą 55% +26.9% 8.83 ą 45% sched_debug.cfs_rq:/.removed.load_avg.avg > > > 305.28 ą 32% +39.3% 425.39 ą 44% sched_debug.cfs_rq:/.removed.load_avg.max > > > 42.94 ą 36% +34.4% 57.70 ą 42% sched_debug.cfs_rq:/.removed.load_avg.stddev > > > 2.96 ą 58% +39.1% 4.12 ą 48% sched_debug.cfs_rq:/.removed.runnable_avg.avg > > > 150.06 ą 34% +44.0% 216.03 ą 45% sched_debug.cfs_rq:/.removed.runnable_avg.max > > > 19.33 ą 42% +42.6% 27.56 ą 45% sched_debug.cfs_rq:/.removed.runnable_avg.stddev > > > 2.96 ą 58% +39.1% 4.12 ą 48% sched_debug.cfs_rq:/.removed.util_avg.avg > > > 150.06 ą 34% +44.0% 216.03 ą 45% sched_debug.cfs_rq:/.removed.util_avg.max > > > 19.33 ą 42% +42.6% 27.56 ą 45% sched_debug.cfs_rq:/.removed.util_avg.stddev > > > 540.76 ą 6% +7.5% 581.25 ą 5% sched_debug.cfs_rq:/.runnable_avg.avg > > > 1060 ą 2% +2.5% 1087 ą 3% sched_debug.cfs_rq:/.runnable_avg.max > > > 442.07 ą 4% -0.1% 441.69 ą 5% sched_debug.cfs_rq:/.runnable_avg.stddev > > > 3123464 ą 14% +10.0% 3436745 ą 3% sched_debug.cfs_rq:/.spread0.avg > > > 6233151 ą 10% +4.4% 6504505 ą 3% sched_debug.cfs_rq:/.spread0.max > > > -540338 +15.6% -624739 sched_debug.cfs_rq:/.spread0.min > > > 2240217 ą 11% +14.0% 2554844 ą 14% sched_debug.cfs_rq:/.spread0.stddev > > > 540.71 ą 6% +7.5% 581.22 ą 5% sched_debug.cfs_rq:/.util_avg.avg > > > 1060 ą 2% +2.5% 1086 ą 3% sched_debug.cfs_rq:/.util_avg.max > > > 442.07 ą 4% -0.1% 441.67 ą 5% sched_debug.cfs_rq:/.util_avg.stddev > > > 454.69 ą 6% +7.0% 486.47 ą 8% sched_debug.cfs_rq:/.util_est_enqueued.avg > > > 1024 -0.0% 1023 sched_debug.cfs_rq:/.util_est_enqueued.max > > > 396.02 ą 2% -0.1% 395.79 sched_debug.cfs_rq:/.util_est_enqueued.stddev > > > 642171 ą 4% +16.6% 748912 ą 2% sched_debug.cpu.avg_idle.avg > > > 1051166 -1.2% 1038098 sched_debug.cpu.avg_idle.max > > > 2402 ą 5% +28.5% 3088 ą 9% sched_debug.cpu.avg_idle.min > > > 384501 ą 3% -12.3% 337306 ą 5% sched_debug.cpu.avg_idle.stddev > > > 198632 ą 7% +5.1% 208788 sched_debug.cpu.clock.avg > > > 198638 ą 7% +5.1% 208794 sched_debug.cpu.clock.max > > > 198626 ą 7% +5.1% 208783 sched_debug.cpu.clock.min > > > 3.25 +2.3% 3.32 ą 5% sched_debug.cpu.clock.stddev > > > 196832 ą 7% +5.1% 206882 sched_debug.cpu.clock_task.avg > > > 197235 ą 7% +5.1% 207282 sched_debug.cpu.clock_task.max > > > 181004 ą 7% +5.7% 191329 sched_debug.cpu.clock_task.min > > > 1575 ą 3% -1.8% 1546 sched_debug.cpu.clock_task.stddev > > > 2411 ą 4% +2.8% 2478 sched_debug.cpu.curr->pid.avg > > > 8665 ą 4% +3.1% 8935 sched_debug.cpu.curr->pid.max > > > 2522 ą 2% +1.0% 2548 sched_debug.cpu.curr->pid.stddev > > > 501318 -0.0% 501249 sched_debug.cpu.max_idle_balance_cost.avg > > > 528365 +0.5% 531236 ą 2% sched_debug.cpu.max_idle_balance_cost.max > > > 500000 +0.0% 500000 sched_debug.cpu.max_idle_balance_cost.min > > > 5157 ą 19% -4.2% 4941 ą 23% sched_debug.cpu.max_idle_balance_cost.stddev > > > 4294 +0.0% 4294 sched_debug.cpu.next_balance.avg > > > 4294 +0.0% 4294 sched_debug.cpu.next_balance.max > > > 4294 +0.0% 4294 sched_debug.cpu.next_balance.min > > > 0.00 ą 41% -40.0% 0.00 ą 13% sched_debug.cpu.next_balance.stddev > > > 0.44 ą 4% +2.4% 0.45 sched_debug.cpu.nr_running.avg > > > 1.00 +0.0% 1.00 sched_debug.cpu.nr_running.max > > > 0.47 +0.5% 0.47 sched_debug.cpu.nr_running.stddev > > > 14345 ą 8% +6.7% 15305 ą 4% sched_debug.cpu.nr_switches.avg > > > 30800 ą 8% +34.5% 41437 ą 10% sched_debug.cpu.nr_switches.max > > > 4563 ą 28% +5.7% 4822 ą 25% sched_debug.cpu.nr_switches.min > > > 5491 ą 8% +26.4% 6941 ą 10% sched_debug.cpu.nr_switches.stddev > > > 2.111e+09 ą 7% +1.5% 2.142e+09 ą 6% sched_debug.cpu.nr_uninterruptible.avg > > > 4.295e+09 +0.0% 4.295e+09 sched_debug.cpu.nr_uninterruptible.max > > > 2.14e+09 +0.1% 2.143e+09 sched_debug.cpu.nr_uninterruptible.stddev > > > 198627 ą 7% +5.1% 208783 sched_debug.cpu_clk > > > 996147 +0.0% 996147 sched_debug.dl_rq:.dl_bw->bw.avg > > > 996147 +0.0% 996147 sched_debug.dl_rq:.dl_bw->bw.max > > > 996147 +0.0% 996147 sched_debug.dl_rq:.dl_bw->bw.min > > > 4.295e+09 +0.0% 4.295e+09 sched_debug.jiffies > > > 198022 ą 7% +5.1% 208178 sched_debug.ktime > > > 950.00 +0.0% 950.00 sched_debug.rt_rq:.rt_runtime.avg > > > 950.00 +0.0% 950.00 sched_debug.rt_rq:.rt_runtime.max > > > 950.00 +0.0% 950.00 sched_debug.rt_rq:.rt_runtime.min > > > 199377 ą 7% +5.1% 209531 sched_debug.sched_clk > > > 1.00 +0.0% 1.00 sched_debug.sched_clock_stable() > > > 58611259 +0.0% 58611259 sched_debug.sysctl_sched.sysctl_sched_features > > > 0.75 +0.0% 0.75 sched_debug.sysctl_sched.sysctl_sched_idle_min_granularity > > > 24.00 +0.0% 24.00 sched_debug.sysctl_sched.sysctl_sched_latency > > > 3.00 +0.0% 3.00 sched_debug.sysctl_sched.sysctl_sched_min_granularity > > > 1.00 +0.0% 1.00 sched_debug.sysctl_sched.sysctl_sched_tunable_scaling > > > 4.00 +0.0% 4.00 sched_debug.sysctl_sched.sysctl_sched_wakeup_granularity > > > 20.90 ą 47% -6.4 14.49 ą100% perf-profile.calltrace.cycles-pp.mwait_idle_with_hints.intel_idle.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call > > > 20.90 ą 47% -6.4 14.49 ą100% perf-profile.calltrace.cycles-pp.intel_idle.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle > > > 0.48 ą 44% -0.5 0.00 perf-profile.calltrace.cycles-pp.do_mmap.vm_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe.__mmap > > > 29.41 ą 19% -0.2 29.23 ą 18% perf-profile.calltrace.cycles-pp.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle.cpu_startup_entry > > > 35.02 ą 8% -0.2 34.86 ą 7% perf-profile.calltrace.cycles-pp.__mmap > > > 34.95 ą 8% -0.1 34.81 ą 7% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.__mmap > > > 34.92 ą 8% -0.1 34.79 ą 7% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.__mmap > > > 34.87 ą 8% -0.1 34.74 ą 7% perf-profile.calltrace.cycles-pp.vm_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe.__mmap > > > 0.41 ą 74% -0.1 0.30 ą156% perf-profile.calltrace.cycles-pp.cpu_startup_entry.rest_init.arch_call_rest_init.start_kernel.secondary_startup_64_no_verify > > > 0.41 ą 74% -0.1 0.30 ą156% perf-profile.calltrace.cycles-pp.do_idle.cpu_startup_entry.rest_init.arch_call_rest_init.start_kernel > > > 0.41 ą 74% -0.1 0.30 ą156% perf-profile.calltrace.cycles-pp.cpuidle_idle_call.do_idle.cpu_startup_entry.rest_init.arch_call_rest_init > > > 0.41 ą 74% -0.1 0.30 ą156% perf-profile.calltrace.cycles-pp.cpuidle_enter.cpuidle_idle_call.do_idle.cpu_startup_entry.rest_init > > > 0.41 ą 74% -0.1 0.30 ą156% perf-profile.calltrace.cycles-pp.start_kernel.secondary_startup_64_no_verify > > > 0.41 ą 74% -0.1 0.30 ą156% perf-profile.calltrace.cycles-pp.arch_call_rest_init.start_kernel.secondary_startup_64_no_verify > > > 0.41 ą 74% -0.1 0.30 ą156% perf-profile.calltrace.cycles-pp.rest_init.arch_call_rest_init.start_kernel.secondary_startup_64_no_verify > > > 29.59 ą 19% -0.1 29.50 ą 17% perf-profile.calltrace.cycles-pp.secondary_startup_64_no_verify > > > 29.03 ą 19% -0.1 28.95 ą 17% perf-profile.calltrace.cycles-pp.cpu_startup_entry.start_secondary.secondary_startup_64_no_verify > > > 29.03 ą 19% -0.1 28.95 ą 17% perf-profile.calltrace.cycles-pp.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64_no_verify > > > 29.03 ą 19% -0.1 28.95 ą 17% perf-profile.calltrace.cycles-pp.start_secondary.secondary_startup_64_no_verify > > > 29.00 ą 19% -0.1 28.93 ą 17% perf-profile.calltrace.cycles-pp.cpuidle_idle_call.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64_no_verify > > > 29.00 ą 19% -0.1 28.93 ą 17% perf-profile.calltrace.cycles-pp.cpuidle_enter.cpuidle_idle_call.do_idle.cpu_startup_entry.start_secondary > > > 33.56 ą 8% -0.0 33.53 ą 7% perf-profile.calltrace.cycles-pp.osq_lock.rwsem_optimistic_spin.rwsem_down_write_slowpath.down_write_killable.vm_mmap_pgoff > > > 34.26 ą 8% -0.0 34.24 ą 7% perf-profile.calltrace.cycles-pp.down_write_killable.vm_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe.__mmap > > > 34.23 ą 8% -0.0 34.21 ą 7% perf-profile.calltrace.cycles-pp.rwsem_down_write_slowpath.down_write_killable.vm_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe > > > 34.19 ą 8% -0.0 34.18 ą 7% perf-profile.calltrace.cycles-pp.rwsem_optimistic_spin.rwsem_down_write_slowpath.down_write_killable.vm_mmap_pgoff.do_syscall_64 > > > 0.44 ą 44% +0.0 0.48 ą 44% perf-profile.calltrace.cycles-pp.rwsem_spin_on_owner.rwsem_optimistic_spin.rwsem_down_write_slowpath.down_write_killable.__vm_munmap > > > 0.45 ą 44% +0.0 0.48 ą 44% perf-profile.calltrace.cycles-pp.rwsem_spin_on_owner.rwsem_optimistic_spin.rwsem_down_write_slowpath.down_write_killable.vm_mmap_pgoff > > > 33.62 ą 8% +0.1 33.71 ą 7% perf-profile.calltrace.cycles-pp.osq_lock.rwsem_optimistic_spin.rwsem_down_write_slowpath.down_write_killable.__vm_munmap > > > 34.32 ą 8% +0.1 34.42 ą 7% perf-profile.calltrace.cycles-pp.down_write_killable.__vm_munmap.__x64_sys_munmap.do_syscall_64.entry_SYSCALL_64_after_hwframe > > > 34.29 ą 8% +0.1 34.39 ą 7% perf-profile.calltrace.cycles-pp.rwsem_down_write_slowpath.down_write_killable.__vm_munmap.__x64_sys_munmap.do_syscall_64 > > > 34.25 ą 8% +0.1 34.36 ą 7% perf-profile.calltrace.cycles-pp.rwsem_optimistic_spin.rwsem_down_write_slowpath.down_write_killable.__vm_munmap.__x64_sys_munmap > > > 35.11 ą 8% +0.2 35.31 ą 7% perf-profile.calltrace.cycles-pp.__munmap > > > 35.04 ą 8% +0.2 35.25 ą 7% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.__munmap > > > 35.02 ą 8% +0.2 35.24 ą 7% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.__munmap > > > 0.00 +0.2 0.22 ą223% perf-profile.calltrace.cycles-pp.intel_idle_irq.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle > > > 34.97 ą 8% +0.2 35.20 ą 7% perf-profile.calltrace.cycles-pp.__vm_munmap.__x64_sys_munmap.do_syscall_64.entry_SYSCALL_64_after_hwframe.__munmap > > > 34.97 ą 8% +0.2 35.20 ą 7% perf-profile.calltrace.cycles-pp.__x64_sys_munmap.do_syscall_64.entry_SYSCALL_64_after_hwframe.__munmap > > > 0.47 ą 44% +0.2 0.70 ą 7% perf-profile.calltrace.cycles-pp.do_mas_align_munmap.__vm_munmap.__x64_sys_munmap.do_syscall_64.entry_SYSCALL_64_after_hwframe > > > 0.00 +0.4 0.44 ą223% perf-profile.calltrace.cycles-pp.asm_sysvec_apic_timer_interrupt.mwait_idle_with_hints.intel_idle_irq.cpuidle_enter_state.cpuidle_enter > > > 8.27 ą 91% +6.2 14.46 ą 77% perf-profile.calltrace.cycles-pp.mwait_idle_with_hints.intel_idle_ibrs.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call > > > 8.27 ą 91% +6.2 14.46 ą 77% perf-profile.calltrace.cycles-pp.intel_idle_ibrs.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle > > > 21.09 ą 47% -6.5 14.62 ą 99% perf-profile.children.cycles-pp.intel_idle > > > 35.02 ą 8% -0.2 34.86 ą 7% perf-profile.children.cycles-pp.__mmap > > > 0.14 ą 9% -0.1 0.00 perf-profile.children.cycles-pp.thp_get_unmapped_area > > > 34.87 ą 8% -0.1 34.74 ą 7% perf-profile.children.cycles-pp.vm_mmap_pgoff > > > 0.55 ą 9% -0.1 0.46 ą 7% perf-profile.children.cycles-pp.do_mmap > > > 29.59 ą 19% -0.1 29.50 ą 17% perf-profile.children.cycles-pp.secondary_startup_64_no_verify > > > 29.59 ą 19% -0.1 29.50 ą 17% perf-profile.children.cycles-pp.cpu_startup_entry > > > 29.59 ą 19% -0.1 29.50 ą 17% perf-profile.children.cycles-pp.do_idle > > > 29.03 ą 19% -0.1 28.95 ą 17% perf-profile.children.cycles-pp.start_secondary > > > 29.56 ą 19% -0.1 29.49 ą 17% perf-profile.children.cycles-pp.cpuidle_idle_call > > > 29.56 ą 19% -0.1 29.48 ą 17% perf-profile.children.cycles-pp.cpuidle_enter > > > 29.56 ą 19% -0.1 29.48 ą 17% perf-profile.children.cycles-pp.cpuidle_enter_state > > > 29.52 ą 19% -0.1 29.45 ą 17% perf-profile.children.cycles-pp.mwait_idle_with_hints > > > 0.38 ą 9% -0.1 0.32 ą 6% perf-profile.children.cycles-pp.mmap_region > > > 0.05 ą 7% -0.1 0.00 perf-profile.children.cycles-pp.unmap_vmas > > > 0.11 ą 8% -0.1 0.06 ą 13% perf-profile.children.cycles-pp.unmap_region > > > 0.16 ą 10% -0.0 0.13 ą 9% perf-profile.children.cycles-pp.get_unmapped_area > > > 0.07 ą 7% -0.0 0.03 ą 70% perf-profile.children.cycles-pp.mas_find > > > 0.05 ą 44% -0.0 0.02 ą141% perf-profile.children.cycles-pp.mas_wr_node_store > > > 0.10 ą 10% -0.0 0.07 ą 14% perf-profile.children.cycles-pp.mas_spanning_rebalance > > > 0.14 ą 9% -0.0 0.11 ą 9% perf-profile.children.cycles-pp.arch_get_unmapped_area_topdown > > > 0.06 ą 11% -0.0 0.04 ą 72% perf-profile.children.cycles-pp.__schedule > > > 0.14 ą 10% -0.0 0.11 ą 9% perf-profile.children.cycles-pp.vm_unmapped_area > > > 0.07 ą 10% -0.0 0.04 ą 45% perf-profile.children.cycles-pp.do_mas_munmap > > > 0.02 ą 99% -0.0 0.00 perf-profile.children.cycles-pp.mas_next_entry > > > 0.04 ą 44% -0.0 0.02 ą141% perf-profile.children.cycles-pp.schedule > > > 0.06 ą 9% -0.0 0.04 ą 71% perf-profile.children.cycles-pp.mas_wr_modify > > > 0.10 ą 8% -0.0 0.08 ą 11% perf-profile.children.cycles-pp.mas_rev_awalk > > > 0.10 ą 12% -0.0 0.08 ą 16% perf-profile.children.cycles-pp.mas_wr_spanning_store > > > 0.06 ą 7% -0.0 0.04 ą 45% perf-profile.children.cycles-pp.mas_walk > > > 0.09 ą 11% -0.0 0.08 ą 16% perf-profile.children.cycles-pp.syscall_exit_to_user_mode > > > 0.02 ą141% -0.0 0.00 perf-profile.children.cycles-pp.perf_event_mmap > > > 0.02 ą141% -0.0 0.00 perf-profile.children.cycles-pp.unmap_page_range > > > 0.11 ą 26% -0.0 0.10 ą 10% perf-profile.children.cycles-pp.__get_user_nocheck_8 > > > 0.35 ą 19% -0.0 0.34 ą 11% perf-profile.children.cycles-pp.perf_tp_event > > > 0.11 ą 26% -0.0 0.10 ą 11% perf-profile.children.cycles-pp.perf_callchain_user > > > 0.34 ą 19% -0.0 0.33 ą 10% perf-profile.children.cycles-pp.__perf_event_overflow > > > 0.34 ą 19% -0.0 0.33 ą 10% perf-profile.children.cycles-pp.perf_event_output_forward > > > 0.31 ą 19% -0.0 0.30 ą 12% perf-profile.children.cycles-pp.perf_prepare_sample > > > 0.30 ą 19% -0.0 0.29 ą 10% perf-profile.children.cycles-pp.perf_callchain > > > 0.30 ą 19% -0.0 0.29 ą 10% perf-profile.children.cycles-pp.get_perf_callchain > > > 0.12 ą 9% -0.0 0.11 ą 9% perf-profile.children.cycles-pp.mas_empty_area_rev > > > 0.08 ą 7% -0.0 0.07 ą 8% perf-profile.children.cycles-pp.syscall_return_via_sysret > > > 0.01 ą223% -0.0 0.00 perf-profile.children.cycles-pp.mas_wr_bnode > > > 0.01 ą223% -0.0 0.00 perf-profile.children.cycles-pp.perf_event_mmap_event > > > 0.01 ą223% -0.0 0.00 perf-profile.children.cycles-pp.__entry_text_start > > > 0.33 ą 10% -0.0 0.32 ą 7% perf-profile.children.cycles-pp.mas_store_prealloc > > > 0.32 ą 20% -0.0 0.32 ą 10% perf-profile.children.cycles-pp.update_curr > > > 0.32 ą 19% -0.0 0.31 ą 11% perf-profile.children.cycles-pp.perf_trace_sched_stat_runtime > > > 0.56 ą 22% -0.0 0.56 ą 58% perf-profile.children.cycles-pp.start_kernel > > > 0.56 ą 22% -0.0 0.56 ą 58% perf-profile.children.cycles-pp.arch_call_rest_init > > > 0.56 ą 22% -0.0 0.56 ą 58% perf-profile.children.cycles-pp.rest_init > > > 0.07 ą 45% -0.0 0.07 ą 11% perf-profile.children.cycles-pp.native_irq_return_iret > > > 0.01 ą223% +0.0 0.01 ą223% perf-profile.children.cycles-pp.ktime_get_update_offsets_now > > > 0.06 ą 45% +0.0 0.06 ą 8% perf-profile.children.cycles-pp.asm_exc_page_fault > > > 0.18 ą 16% +0.0 0.18 ą 14% perf-profile.children.cycles-pp.perf_callchain_kernel > > > 0.12 ą 16% +0.0 0.12 ą 12% perf-profile.children.cycles-pp.unwind_next_frame > > > 0.36 ą 18% +0.0 0.37 ą 10% perf-profile.children.cycles-pp.task_tick_fair > > > 0.58 ą 14% +0.0 0.58 ą 10% perf-profile.children.cycles-pp.hrtimer_interrupt > > > 0.49 ą 14% +0.0 0.50 ą 11% perf-profile.children.cycles-pp.__hrtimer_run_queues > > > 0.05 ą 46% +0.0 0.05 ą 45% perf-profile.children.cycles-pp.__unwind_start > > > 0.45 ą 14% +0.0 0.46 ą 11% perf-profile.children.cycles-pp.tick_sched_handle > > > 0.46 ą 14% +0.0 0.46 ą 11% perf-profile.children.cycles-pp.tick_sched_timer > > > 0.45 ą 15% +0.0 0.45 ą 11% perf-profile.children.cycles-pp.update_process_times > > > 0.06 ą 11% +0.0 0.07 ą 12% perf-profile.children.cycles-pp.kmem_cache_free_bulk > > > 0.58 ą 14% +0.0 0.58 ą 10% perf-profile.children.cycles-pp.__sysvec_apic_timer_interrupt > > > 0.00 +0.0 0.01 ą223% perf-profile.children.cycles-pp.record__mmap_read_evlist > > > 0.00 +0.0 0.01 ą223% perf-profile.children.cycles-pp.perf_mmap__push > > > 0.00 +0.0 0.01 ą223% perf-profile.children.cycles-pp.record__pushfn > > > 0.00 +0.0 0.01 ą223% perf-profile.children.cycles-pp.ksys_write > > > 0.00 +0.0 0.01 ą223% perf-profile.children.cycles-pp.vfs_write > > > 0.00 +0.0 0.01 ą223% perf-profile.children.cycles-pp.__libc_write > > > 0.00 +0.0 0.01 ą223% perf-profile.children.cycles-pp.generic_file_write_iter > > > 0.00 +0.0 0.01 ą223% perf-profile.children.cycles-pp.__generic_file_write_iter > > > 0.00 +0.0 0.01 ą223% perf-profile.children.cycles-pp.generic_perform_write > > > 0.00 +0.0 0.01 ą223% perf-profile.children.cycles-pp.build_id__mark_dso_hit > > > 0.39 ą 17% +0.0 0.40 ą 10% perf-profile.children.cycles-pp.scheduler_tick > > > 0.00 +0.0 0.01 ą223% perf-profile.children.cycles-pp.clockevents_program_event > > > 0.05 ą 45% +0.0 0.06 ą 11% perf-profile.children.cycles-pp.mas_wr_store_entry > > > 0.60 ą 14% +0.0 0.61 ą 9% perf-profile.children.cycles-pp.sysvec_apic_timer_interrupt > > > 0.08 ą 8% +0.0 0.10 ą 12% perf-profile.children.cycles-pp.mas_destroy > > > 0.08 ą 9% +0.0 0.09 ą 21% perf-profile.children.cycles-pp.perf_session__deliver_event > > > 0.08 ą 12% +0.0 0.09 ą 33% perf-profile.children.cycles-pp.ordered_events__queue > > > 0.08 ą 11% +0.0 0.10 ą 22% perf-profile.children.cycles-pp.__ordered_events__flush > > > 0.08 ą 9% +0.0 0.10 ą 22% perf-profile.children.cycles-pp.perf_session__process_user_event > > > 0.06 ą 13% +0.0 0.08 ą 14% perf-profile.children.cycles-pp.kmem_cache_alloc > > > 0.07 ą 9% +0.0 0.09 ą 33% perf-profile.children.cycles-pp.queue_event > > > 0.08 ą 8% +0.0 0.10 ą 31% perf-profile.children.cycles-pp.process_simple > > > 0.00 +0.0 0.03 ą100% perf-profile.children.cycles-pp.evlist__parse_sample > > > 0.06 ą 6% +0.0 0.08 ą 8% perf-profile.children.cycles-pp.memset_erms > > > 0.22 ą 7% +0.0 0.26 ą 23% perf-profile.children.cycles-pp.__libc_start_main > > > 0.22 ą 7% +0.0 0.26 ą 23% perf-profile.children.cycles-pp.main > > > 0.22 ą 7% +0.0 0.26 ą 23% perf-profile.children.cycles-pp.run_builtin > > > 0.21 ą 9% +0.0 0.25 ą 23% perf-profile.children.cycles-pp.cmd_record > > > 0.21 ą 9% +0.0 0.25 ą 23% perf-profile.children.cycles-pp.__cmd_record > > > 0.20 ą 9% +0.0 0.24 ą 24% perf-profile.children.cycles-pp.cmd_sched > > > 0.17 ą 11% +0.0 0.21 ą 25% perf-profile.children.cycles-pp.reader__read_event > > > 0.17 ą 11% +0.0 0.21 ą 26% perf-profile.children.cycles-pp.record__finish_output > > > 0.17 ą 11% +0.0 0.21 ą 26% perf-profile.children.cycles-pp.perf_session__process_events > > > 0.00 +0.0 0.04 ą 45% perf-profile.children.cycles-pp.kmem_cache_free > > > 0.17 ą 7% +0.1 0.22 ą 8% perf-profile.children.cycles-pp.mas_alloc_nodes > > > 0.11 ą 9% +0.1 0.17 ą 6% perf-profile.children.cycles-pp.kmem_cache_alloc_bulk > > > 0.00 +0.1 0.06 ą 13% perf-profile.children.cycles-pp.vm_area_dup > > > 0.16 ą 8% +0.1 0.22 ą 6% perf-profile.children.cycles-pp.mas_preallocate > > > 67.20 ą 8% +0.1 67.28 ą 7% perf-profile.children.cycles-pp.osq_lock > > > 68.59 ą 8% +0.1 68.66 ą 7% perf-profile.children.cycles-pp.down_write_killable > > > 1.04 ą 8% +0.1 1.12 ą 7% perf-profile.children.cycles-pp.rwsem_spin_on_owner > > > 70.08 ą 8% +0.1 70.15 ą 7% perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe > > > 68.52 ą 8% +0.1 68.60 ą 7% perf-profile.children.cycles-pp.rwsem_down_write_slowpath > > > 70.03 ą 8% +0.1 70.11 ą 7% perf-profile.children.cycles-pp.do_syscall_64 > > > 68.46 ą 8% +0.1 68.55 ą 7% perf-profile.children.cycles-pp.rwsem_optimistic_spin > > > 0.55 ą 8% +0.2 0.71 ą 8% perf-profile.children.cycles-pp.do_mas_align_munmap > > > 35.12 ą 8% +0.2 35.31 ą 7% perf-profile.children.cycles-pp.__munmap > > > 0.00 +0.2 0.22 ą 7% perf-profile.children.cycles-pp.vma_expand > > > 0.00 +0.2 0.22 ą223% perf-profile.children.cycles-pp.intel_idle_irq > > > 34.98 ą 8% +0.2 35.20 ą 7% perf-profile.children.cycles-pp.__x64_sys_munmap > > > 34.97 ą 8% +0.2 35.20 ą 7% perf-profile.children.cycles-pp.__vm_munmap > > > 0.64 ą 13% +0.2 0.88 ą 55% perf-profile.children.cycles-pp.asm_sysvec_apic_timer_interrupt > > > 0.00 +0.3 0.30 ą 7% perf-profile.children.cycles-pp.__vma_adjust > > > 0.00 +0.4 0.36 ą 6% perf-profile.children.cycles-pp.__split_vma > > > 8.42 ą 91% +6.2 14.60 ą 77% perf-profile.children.cycles-pp.intel_idle_ibrs > > > 29.52 ą 19% -0.1 29.45 ą 17% perf-profile.self.cycles-pp.mwait_idle_with_hints > > > 0.18 ą 9% -0.1 0.12 ą 10% perf-profile.self.cycles-pp.rwsem_optimistic_spin > > > 0.04 ą 45% -0.0 0.00 perf-profile.self.cycles-pp.entry_SYSCALL_64_after_hwframe > > > 0.04 ą 44% -0.0 0.00 perf-profile.self.cycles-pp.mmap_region > > > 0.10 ą 5% -0.0 0.08 ą 9% perf-profile.self.cycles-pp.mas_rev_awalk > > > 0.06 ą 7% -0.0 0.04 ą 45% perf-profile.self.cycles-pp.mas_walk > > > 0.06 ą 11% -0.0 0.04 ą 45% perf-profile.self.cycles-pp.do_mas_align_munmap > > > 0.08 ą 8% -0.0 0.07 ą 14% perf-profile.self.cycles-pp.syscall_exit_to_user_mode > > > 0.08 ą 7% -0.0 0.07 ą 8% perf-profile.self.cycles-pp.syscall_return_via_sysret > > > 0.06 ą 13% -0.0 0.05 ą 7% perf-profile.self.cycles-pp.down_write_killable > > > 0.07 ą 45% -0.0 0.07 ą 11% perf-profile.self.cycles-pp.native_irq_return_iret > > > 0.05 ą 45% -0.0 0.05 ą 47% perf-profile.self.cycles-pp.unwind_next_frame > > > 0.00 +0.0 0.01 ą223% perf-profile.self.cycles-pp.ktime_get_update_offsets_now > > > 0.05 ą 45% +0.0 0.06 ą 11% perf-profile.self.cycles-pp.kmem_cache_free_bulk > > > 0.00 +0.0 0.02 ą141% perf-profile.self.cycles-pp.kmem_cache_free > > > 0.07 ą 8% +0.0 0.09 ą 33% perf-profile.self.cycles-pp.queue_event > > > 0.06 ą 8% +0.0 0.08 ą 8% perf-profile.self.cycles-pp.memset_erms > > > 0.04 ą 45% +0.0 0.08 ą 6% perf-profile.self.cycles-pp.kmem_cache_alloc_bulk > > > 66.61 ą 8% +0.1 66.68 ą 7% perf-profile.self.cycles-pp.osq_lock > > > 1.02 ą 8% +0.1 1.10 ą 7% perf-profile.self.cycles-pp.rwsem_spin_on_owner > > > > > > > > > > > > If you fix the issue, kindly add following tag > > > | Reported-by: kernel test robot <yujie.liu@xxxxxxxxx> > > > | Link: https://lore.kernel.org/oe-lkp/202212151657.5d11a672-yujie.liu@xxxxxxxxx > > > > > > > > > To reproduce: > > > > > > git clone https://github.com/intel/lkp-tests.git > > > cd lkp-tests > > > sudo bin/lkp install job.yaml # job file is attached in this email > > > bin/lkp split-job --compatible job.yaml # generate the yaml file for lkp run > > > sudo bin/lkp run generated-yaml-file > > > > > > # if come across any failure that blocks the test, > > > # please remove ~/.lkp and /lkp dir to run from a clean state. > > > > > > > > > Disclaimer: > > > Results have been estimated based on internal Intel analysis and are provided > > > for informational purposes only. Any difference in system hardware or software > > > design or configuration may affect actual performance. > > > > > > > > > -- > > > 0-DAY CI Kernel Test Service > > > https://01.org/lkp