Re: [linus:master] [mm, pcp] 6ccdcb6d3a: stress-ng.judy.ops_per_sec -4.7% regression

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



hi, Huang Ying,

On Thu, Nov 23, 2023 at 01:40:02PM +0800, Huang, Ying wrote:
> Hi,
> 
> Thanks for test!
> 
> kernel test robot <oliver.sang@xxxxxxxxx> writes:
> 
> > Hello,
> >
> > kernel test robot noticed a -4.7% regression of stress-ng.judy.ops_per_sec on:
> >
> >
> > commit: 6ccdcb6d3a741c4e005ca6ffd4a62ddf8b5bead3 ("mm, pcp: reduce detecting time of consecutive high order page freeing")
> > https://git.kernel.org/cgit/linux/kernel/git/torvalds/linux.git master
> >
> > testcase: stress-ng
> > test machine: 224 threads 2 sockets Intel(R) Xeon(R) Platinum 8480CTDX (Sapphire Rapids) with 512G memory
> > parameters:
> >
> > 	nr_threads: 100%
> > 	testtime: 60s
> > 	class: cpu-cache
> > 	test: judy
> > 	disk: 1SSD
> > 	cpufreq_governor: performance
> >
> >
> > In addition to that, the commit also has significant impact on the following tests:
> >
> > +------------------+-------------------------------------------------------------------------------------------------+
> > | testcase: change | lmbench3: lmbench3.TCP.socket.bandwidth.10MB.MB/sec 23.7% improvement                           |
> > | test machine     | 48 threads 2 sockets Intel(R) Xeon(R) CPU E5-2697 v2 @ 2.70GHz (Ivy Bridge-EP) with 112G memory |
> > | test parameters  | cpufreq_governor=performance                                                                    |
> > |                  | mode=development                                                                                |
> > |                  | nr_threads=100%                                                                                 |
> > |                  | test=TCP                                                                                        |
> > |                  | test_memory_size=50%                                                                            |
> > +------------------+-------------------------------------------------------------------------------------------------+
> > | testcase: change | stress-ng: stress-ng.file-ioctl.ops_per_sec -6.6% regression                                    |
> > | test machine     | 36 threads 1 sockets Intel(R) Core(TM) i9-9980XE CPU @ 3.00GHz (Skylake) with 32G memory        |
> > | test parameters  | class=filesystem                                                                                |
> > |                  | cpufreq_governor=performance                                                                    |
> > |                  | disk=1SSD                                                                                       |
> > |                  | fs=btrfs                                                                                        |
> > |                  | nr_threads=10%                                                                                  |
> > |                  | test=file-ioctl                                                                                 |
> > |                  | testtime=60s                                                                                    |
> > +------------------+-------------------------------------------------------------------------------------------------+
> 
> It's expected that this commit will benefit some workload (mainly
> network, inter-process communication related) and hurt some workload.
> But the  whole series should have no much regression.  Can you try the
> whole series for the regression test cases?  The series start from
> commit ca71fe1ad922 ("mm, pcp: avoid to drain PCP when process exit") to
> commit 6ccdcb6d3a74 ("mm, pcp: reduce detecting time of consecutive high
> order page freeing").

since:
* 6ccdcb6d3a741 mm, pcp: reduce detecting time of consecutive high order page freeing
* 57c0419c5f0ea mm, pcp: decrease PCP high if free pages < high watermark
* 51a755c56dc05 mm: tune PCP high automatically
* 90b41691b9881 mm: add framework for PCP high auto-tuning
* c0a242394cb98 mm, page_alloc: scale the number of pages that are batch allocated
* 52166607ecc98 mm: restrict the pcp batch scale factor to avoid too long latency
* 362d37a106dd3 mm, pcp: reduce lock contention for draining high-order pages
* 94a3bfe4073cd cacheinfo: calculate size of per-CPU data cache slice
* ca71fe1ad9221 mm, pcp: avoid to drain PCP when process exit
* 1f4f7f0f8845d mm/oom_killer: simplify OOM killer info dump helper

I tested 1f4f7f0f8845d vs 6ccdcb6d3a741.

for stress-ng.judy.ops_per_sec, there is a smaller regression (-2.0%):
(full comparison is attached as ncompare-judy)

=========================================================================================
class/compiler/cpufreq_governor/disk/kconfig/nr_threads/rootfs/tbox_group/test/testcase/testtime:
  cpu-cache/gcc-12/performance/1SSD/x86_64-rhel-8.3/100%/debian-11.1-x86_64-20220510.cgz/lkp-spr-2sp4/judy/stress-ng/60s

1f4f7f0f8845dbac 6ccdcb6d3a741c4e005ca6ffd4a
---------------- ---------------------------
         %stddev     %change         %stddev
             \          |                \
   6925490            -0.9%    6862477        stress-ng.judy.Judy_delete_operations_per_sec
  22515488            -0.4%   22420191        stress-ng.judy.Judy_find_operations_per_sec
   9036524            -3.9%    8685310 ±  3%  stress-ng.judy.Judy_insert_operations_per_sec
    171299            -2.0%     167905        stress-ng.judy.ops
      2853            -2.0%       2796        stress-ng.judy.ops_per_sec


for stress-ng.file-ioctl.ops_per_sec, there is a similar regression (-6.9%):
(full comparison is attached as ncompare-file-ioctl)

=========================================================================================
class/compiler/cpufreq_governor/disk/fs/kconfig/nr_threads/rootfs/tbox_group/test/testcase/testtime:
  filesystem/gcc-12/performance/1SSD/btrfs/x86_64-rhel-8.3/10%/debian-11.1-x86_64-20220510.cgz/lkp-skl-d08/file-ioctl/stress-ng/60s

1f4f7f0f8845dbac 6ccdcb6d3a741c4e005ca6ffd4a
---------------- ---------------------------
         %stddev     %change         %stddev
             \          |                \
    340971            -6.9%     317411        stress-ng.file-ioctl.ops
      5682            -6.9%       5290        stress-ng.file-ioctl.ops_per_sec

> 
> --
> Best Regards,
> Huang, Ying
> 
xsang@inn:~$ lkp ncompare -s commit=1f4f7f0f8845d/tbox_group=lkp-spr-2sp4/testcase=stress-ng -o commit=6ccdcb6d3a741  -a | grep -v interrupts | grep -v softirq
=========================================================================================
class/compiler/cpufreq_governor/disk/kconfig/nr_threads/rootfs/tbox_group/test/testcase/testtime:
  cpu-cache/gcc-12/performance/1SSD/x86_64-rhel-8.3/100%/debian-11.1-x86_64-20220510.cgz/lkp-spr-2sp4/judy/stress-ng/60s

commit:
  1f4f7f0f8845dbac40289cc3d50b81314c5a12b8
  6ccdcb6d3a741c4e005ca6ffd4a62ddf8b5bead3

1f4f7f0f8845dbac 6ccdcb6d3a741c4e005ca6ffd4a
---------------- ---------------------------
       fail:runs  %reproduction    fail:runs
           |             |             |
         10:10           0%          10:10    stress-ng.judy.Judy_delete_operations_per_sec.pass
         10:10           0%          10:10    stress-ng.judy.Judy_find_operations_per_sec.pass
         10:10           0%          10:10    stress-ng.judy.Judy_insert_operations_per_sec.pass
         10:10           0%          10:10    stress-ng.judy.pass
         %stddev     %change         %stddev
             \          |                \
    151.27 ± 56%      +6.6%     161.22 ± 52%  dmesg.timestamp:last
    124.42 ± 11%      +1.1%     125.79 ±  9%  kmsg.timestamp:last
 1.107e+09 ± 29%     +67.3%  1.852e+09 ± 75%  cpuidle..time
   1084691 ± 30%     +71.0%    1855097 ± 77%  cpuidle..usage
    125.81 ± 10%      +3.3%     129.98 ± 11%  uptime.boot
     13667 ± 21%      +6.8%      14591 ± 23%  uptime.idle
     53.76 ±  5%      +0.6%      54.06 ±  4%  boot-time.boot
     30.41            -0.2%      30.36        boot-time.dhcp
     11000 ±  5%      +0.6%      11066 ±  4%  boot-time.idle
     10.11 ± 18%     +35.2%      13.67 ± 51%  iostat.cpu.idle
      4.96 ±  2%     +25.9%       6.25 ± 17%  iostat.cpu.system
     84.92 ±  2%      -5.7%      80.08 ±  8%  iostat.cpu.user
      7.37 ± 27%      +3.8       11.16 ± 66%  mpstat.cpu.all.idle%
      1.57 ±  3%      -0.1        1.51 ±  5%  mpstat.cpu.all.irq%
      0.14 ±  3%      +0.4        0.49 ± 33%  mpstat.cpu.all.soft%
      3.38 ±  2%      +1.0        4.41 ± 22%  mpstat.cpu.all.sys%
     87.54 ±  2%      -5.1       82.43 ±  8%  mpstat.cpu.all.usr%
     17731 ± 33%      +7.7%      19098 ± 26%  perf-c2c.DRAM.local
    227.40 ± 39%     +19.3%     271.30 ± 26%  perf-c2c.DRAM.remote
    396.50 ± 33%     +21.0%     479.80 ± 29%  perf-c2c.HITM.local
     54.60 ± 51%     +12.1%      61.20 ± 35%  perf-c2c.HITM.remote
    451.10 ± 35%     +19.9%     541.00 ± 29%  perf-c2c.HITM.total
      0.00          -100.0%       0.00        numa-numastat.node0.interleave_hit
  42179114           -12.9%   36727757 ±  8%  numa-numastat.node0.local_node
  42322224           -12.9%   36881294 ±  8%  numa-numastat.node0.numa_hit
    141985 ± 51%      +8.1%     153548 ± 39%  numa-numastat.node0.other_node
      0.00          -100.0%       0.00        numa-numastat.node1.interleave_hit
  41820129           -12.8%   36449421 ±  9%  numa-numastat.node1.local_node
  41909301           -12.8%   36528507 ±  9%  numa-numastat.node1.numa_hit
     89842 ± 81%     -12.8%      78315 ± 77%  numa-numastat.node1.other_node
      9.94 ± 20%     +37.6%      13.68 ± 50%  vmstat.cpu.id
      4.91 ±  2%     +27.3%       6.25 ± 17%  vmstat.cpu.sy
     85.15 ±  2%      -6.0%      80.06 ±  8%  vmstat.cpu.us
      0.02 ±  2%      -4.2%       0.01 ±  8%  vmstat.io.bi
      4.00            +0.0%       4.00        vmstat.memory.buff
   3275886            +0.7%    3300023        vmstat.memory.cache
 5.199e+08            -0.1%  5.196e+08        vmstat.memory.free
    202.27 ±  2%      -3.2%     195.88 ±  7%  vmstat.procs.r
      9290            -4.0%       8915 ±  6%  vmstat.system.cs
    366738            -2.3%     358141 ±  2%  vmstat.system.in
     62.20            -0.0%      62.20        time.elapsed_time
     62.20            -0.0%      62.20        time.elapsed_time.max
    258360            -2.6%     251676        time.involuntary_context_switches
      0.30 ±152%    +166.7%       0.80 ±108%  time.major_page_faults
     10893 ±  7%     +23.0%      13397 ± 29%  time.maximum_resident_set_size
  82141065           -13.1%   71375734 ±  9%  time.minor_page_faults
      4096            +0.0%       4096        time.page_size
     21055            -1.2%      20804        time.percent_of_cpu_this_job_got
    480.34           +35.5%     650.87 ± 21%  time.system_time
     12618            -2.6%      12291        time.user_time
    563.40 ±  2%      -0.7%     559.50 ±  3%  time.voluntary_context_switches
      2608 ±  2%      -3.3%       2521 ±  7%  turbostat.Avg_MHz
     92.38 ±  2%      -3.8       88.53 ±  8%  turbostat.Busy%
      2839            +0.9%       2864        turbostat.Bzy_MHz
      1414 ± 18%     +32.0%       1866 ± 35%  turbostat.C1
    527910 ± 28%     +64.1%     866385 ± 73%  turbostat.C1E
      3.33 ± 26%      +1.7        5.05 ± 64%  turbostat.C1E%
    612375 ± 28%     +67.9%    1028102 ± 76%  turbostat.C6
      4.26 ± 24%      +2.1        6.39 ± 62%  turbostat.C6%
      7.50 ± 25%     +51.2%      11.34 ± 63%  turbostat.CPU%c1
      0.12 ± 17%      +1.6%       0.13 ± 21%  turbostat.CPU%c6
     59.70 ±  3%      -2.0%      58.50 ±  4%  turbostat.CoreTmp
      1.25            -2.6%       1.22        turbostat.IPC
  24279278            +2.3%   24843604 ±  5%  turbostat.IRQ
    180.02 ±  2%     -11.0      169.01 ±  8%  turbostat.PKG_%
      2789 ± 31%     +33.1%       3712 ± 47%  turbostat.POLL
     60.00 ±  3%      -2.3%      58.60 ±  4%  turbostat.PkgTmp
    672.54            -1.8%     660.52 ±  3%  turbostat.PkgWatt
     37.66            -2.4%      36.77 ±  2%  turbostat.RAMWatt
      1989            -0.0%       1988        turbostat.TSC_MHz
   6925490            -0.9%    6862477        stress-ng.judy.Judy_delete_operations_per_sec
  22515488            -0.4%   22420191        stress-ng.judy.Judy_find_operations_per_sec
   9036524            -3.9%    8685310 ±  3%  stress-ng.judy.Judy_insert_operations_per_sec
    171299            -2.0%     167905        stress-ng.judy.ops
      2853            -2.0%       2796        stress-ng.judy.ops_per_sec
     62.20            -0.0%      62.20        stress-ng.time.elapsed_time
     62.20            -0.0%      62.20        stress-ng.time.elapsed_time.max
    258360            -2.6%     251676        stress-ng.time.involuntary_context_switches
      0.30 ±152%    +166.7%       0.80 ±108%  stress-ng.time.major_page_faults
     10893 ±  7%     +23.0%      13397 ± 29%  stress-ng.time.maximum_resident_set_size
  82141065           -13.1%   71375734 ±  9%  stress-ng.time.minor_page_faults
      4096            +0.0%       4096        stress-ng.time.page_size
     21055            -1.2%      20804        stress-ng.time.percent_of_cpu_this_job_got
    480.34           +35.5%     650.87 ± 21%  stress-ng.time.system_time
     12618            -2.6%      12291        stress-ng.time.user_time
    563.40 ±  2%      -0.7%     559.50 ±  3%  stress-ng.time.voluntary_context_switches
     46768 ± 18%     +73.1%      80961 ± 56%  meminfo.Active
     46689 ± 18%     +73.2%      80881 ± 56%  meminfo.Active(anon)
     78.40 ±  6%      +2.0%      79.98        meminfo.Active(file)
    127245 ± 16%      +5.8%     134611 ± 14%  meminfo.AnonHugePages
   1451846            -2.1%    1422053 ±  6%  meminfo.AnonPages
      4.00            +0.0%       4.00        meminfo.Buffers
   3121206            +0.7%    3144477        meminfo.Cached
 2.637e+08            +0.0%  2.637e+08        meminfo.CommitLimit
   4368713            -2.2%    4271782 ±  5%  meminfo.Committed_AS
 5.245e+08            -0.1%   5.24e+08        meminfo.DirectMap1G
  11274240 ± 12%      +4.5%   11782144 ± 12%  meminfo.DirectMap2M
    198052 ±  5%      +8.3%     214436 ±  5%  meminfo.DirectMap4k
      2048            +0.0%       2048        meminfo.Hugepagesize
   1800847            -2.1%    1763898 ±  5%  meminfo.Inactive
   1800650            -2.1%    1763721 ±  5%  meminfo.Inactive(anon)
    197.33 ±  2%     -10.5%     176.70 ±  8%  meminfo.Inactive(file)
    148140            +0.2%     148422        meminfo.KReclaimable
     41102            -0.2%      41026        meminfo.KernelStack
    517263 ±  2%      +0.3%     518808 ±  2%  meminfo.Mapped
 5.178e+08            -0.1%  5.175e+08        meminfo.MemAvailable
 5.199e+08            -0.1%  5.196e+08        meminfo.MemFree
 5.274e+08            +0.0%  5.274e+08        meminfo.MemTotal
   7517161            +4.3%    7841550 ±  4%  meminfo.Memused
     10.95 ±  2%      -4.0%      10.51 ±  8%  meminfo.Mlocked
     25594 ±  3%      -1.9%      25112 ±  7%  meminfo.PageTables
    109277            +0.3%     109597        meminfo.Percpu
    148140            +0.2%     148422        meminfo.SReclaimable
    462907            +0.2%     463832        meminfo.SUnreclaim
    405085 ±  2%      +5.8%     428379 ±  8%  meminfo.Shmem
    611047            +0.2%     612255        meminfo.Slab
   2715849            -0.0%    2715845        meminfo.Unevictable
 1.374e+13            +0.0%  1.374e+13        meminfo.VmallocTotal
    292661            +0.0%     292707        meminfo.VmallocUsed
   7721333            +6.6%    8231493 ±  4%  meminfo.max_used_kB
      6639 ± 70%     +25.7%       8344 ±160%  numa-meminfo.node0.Active
      6609 ± 70%     +25.5%       8296 ±161%  numa-meminfo.node0.Active(anon)
     30.40 ±123%     +57.9%      48.00 ± 81%  numa-meminfo.node0.Active(file)
     65249 ± 56%     -24.1%      49545 ± 74%  numa-meminfo.node0.AnonHugePages
    698546 ± 10%     -10.1%     627683 ± 11%  numa-meminfo.node0.AnonPages
    783098 ±  9%      -4.6%     746765 ± 10%  numa-meminfo.node0.AnonPages.max
   1849056 ± 60%     -27.0%    1350657 ± 82%  numa-meminfo.node0.FilePages
    708083 ± 10%      -8.9%     645187 ± 13%  numa-meminfo.node0.Inactive
    708005 ± 10%      -8.9%     645081 ± 13%  numa-meminfo.node0.Inactive(anon)
     77.48 ±122%     +37.3%     106.35 ± 82%  numa-meminfo.node0.Inactive(file)
     82324 ± 32%      -8.0%      75702 ± 27%  numa-meminfo.node0.KReclaimable
     20512 ±  7%      +1.4%      20790 ±  7%  numa-meminfo.node0.KernelStack
    175959 ±  9%      -0.1%     175730 ±  7%  numa-meminfo.node0.Mapped
 2.594e+08            +0.2%  2.599e+08        numa-meminfo.node0.MemFree
 2.635e+08            +0.0%  2.635e+08        numa-meminfo.node0.MemTotal
   4076783 ± 28%     -10.2%    3660809 ± 31%  numa-meminfo.node0.MemUsed
      6.66 ± 81%      -2.3%       6.51 ± 81%  numa-meminfo.node0.Mlocked
      9901 ± 61%      +4.5%      10346 ± 49%  numa-meminfo.node0.PageTables
     82324 ± 32%      -8.0%      75702 ± 27%  numa-meminfo.node0.SReclaimable
    248888 ±  3%      +0.5%     250078 ±  4%  numa-meminfo.node0.SUnreclaim
     20568 ± 40%     +41.9%      29194 ± 98%  numa-meminfo.node0.Shmem
    331213 ±  9%      -1.6%     325780 ±  7%  numa-meminfo.node0.Slab
   1828379 ± 61%     -27.7%    1321308 ± 85%  numa-meminfo.node0.Unevictable
     40129 ± 22%     +83.1%      73478 ± 53%  numa-meminfo.node1.Active
     40081 ± 22%     +83.2%      73447 ± 53%  numa-meminfo.node1.Active(anon)
     48.00 ± 81%     -33.4%      31.98 ±122%  numa-meminfo.node1.Active(file)
     62048 ± 50%     +37.2%      85152 ± 49%  numa-meminfo.node1.AnonHugePages
    756214 ±  9%      +3.8%     784679 ± 12%  numa-meminfo.node1.AnonPages
    924235 ±  6%      +6.7%     986345 ±  7%  numa-meminfo.node1.AnonPages.max
   1271716 ± 88%     +41.1%    1793782 ± 63%  numa-meminfo.node1.FilePages
   1095798 ±  5%      +1.1%    1108364 ± 10%  numa-meminfo.node1.Inactive
   1095678 ±  5%      +1.2%    1108294 ± 10%  numa-meminfo.node1.Inactive(anon)
    119.85 ± 77%     -41.3%      70.35 ±118%  numa-meminfo.node1.Inactive(file)
     65822 ± 40%     +10.5%      72715 ± 29%  numa-meminfo.node1.KReclaimable
     20590 ±  7%      -1.7%      20233 ±  8%  numa-meminfo.node1.KernelStack
    341090 ±  4%      +0.0%     341144 ±  4%  numa-meminfo.node1.Mapped
 2.605e+08            -0.3%  2.597e+08        numa-meminfo.node1.MemFree
 2.639e+08            +0.0%  2.639e+08        numa-meminfo.node1.MemTotal
   3445580 ± 33%     +21.0%    4169560 ± 27%  numa-meminfo.node1.MemUsed
      4.33 ±122%      -8.0%       3.98 ±123%  numa-meminfo.node1.Mlocked
     15703 ± 37%      -6.1%      14739 ± 38%  numa-meminfo.node1.PageTables
     65822 ± 40%     +10.5%      72715 ± 29%  numa-meminfo.node1.SReclaimable
    214196 ±  4%      -0.2%     213692 ±  5%  numa-meminfo.node1.SUnreclaim
    384078 ±  3%      +3.9%     399143 ±  7%  numa-meminfo.node1.Shmem
    280018 ± 11%      +2.3%     286407 ±  9%  numa-meminfo.node1.Slab
    887469 ±126%     +57.1%    1394536 ± 80%  numa-meminfo.node1.Unevictable
      1655 ± 70%     +28.3%       2123 ±163%  numa-vmstat.node0.nr_active_anon
      7.60 ±123%     +57.9%      12.00 ± 81%  numa-vmstat.node0.nr_active_file
    173571 ± 10%      -9.6%     156867 ± 11%  numa-vmstat.node0.nr_anon_pages
     31.86 ± 57%     -24.0%      24.20 ± 74%  numa-vmstat.node0.nr_anon_transparent_hugepages
    462268 ± 60%     -27.0%     337669 ± 82%  numa-vmstat.node0.nr_file_pages
  64861249            +0.2%   64962956        numa-vmstat.node0.nr_free_pages
    176031 ±  9%      -8.5%     161129 ± 12%  numa-vmstat.node0.nr_inactive_anon
     19.37 ±122%     +37.3%      26.59 ± 82%  numa-vmstat.node0.nr_inactive_file
      0.00 ±300%      +1.6%       0.00 ±300%  numa-vmstat.node0.nr_isolated_anon
     20512 ±  7%      +1.4%      20790 ±  7%  numa-vmstat.node0.nr_kernel_stack
     43945 ±  9%      -0.2%      43864 ±  7%  numa-vmstat.node0.nr_mapped
      1.66 ± 81%      -2.5%       1.62 ± 81%  numa-vmstat.node0.nr_mlock
      2475 ± 61%      +4.5%       2585 ± 49%  numa-vmstat.node0.nr_page_table_pages
      5146 ± 40%     +41.9%       7303 ± 98%  numa-vmstat.node0.nr_shmem
     20581 ± 32%      -8.0%      18925 ± 27%  numa-vmstat.node0.nr_slab_reclaimable
     62224 ±  3%      +0.5%      62514 ±  4%  numa-vmstat.node0.nr_slab_unreclaimable
    457094 ± 61%     -27.7%     330327 ± 85%  numa-vmstat.node0.nr_unevictable
      1655 ± 70%     +28.3%       2123 ±163%  numa-vmstat.node0.nr_zone_active_anon
      7.60 ±123%     +57.9%      12.00 ± 81%  numa-vmstat.node0.nr_zone_active_file
    176029 ±  9%      -8.5%     161125 ± 12%  numa-vmstat.node0.nr_zone_inactive_anon
     19.37 ±122%     +37.3%      26.59 ± 82%  numa-vmstat.node0.nr_zone_inactive_file
    457094 ± 61%     -27.7%     330327 ± 85%  numa-vmstat.node0.nr_zone_unevictable
  42322270           -12.9%   36881741 ±  8%  numa-vmstat.node0.numa_hit
      0.00          -100.0%       0.00        numa-vmstat.node0.numa_interleave
  42179160           -12.9%   36728210 ±  8%  numa-vmstat.node0.numa_local
    141985 ± 51%      +8.1%     153548 ± 39%  numa-vmstat.node0.numa_other
     10088 ± 22%     +82.5%      18410 ± 53%  numa-vmstat.node1.nr_active_anon
     12.00 ± 81%     -33.4%       7.99 ±122%  numa-vmstat.node1.nr_active_file
    187803 ±  9%      +4.4%     196071 ± 12%  numa-vmstat.node1.nr_anon_pages
     30.30 ± 50%     +37.2%      41.58 ± 49%  numa-vmstat.node1.nr_anon_transparent_hugepages
    317887 ± 88%     +41.1%     448582 ± 63%  numa-vmstat.node1.nr_file_pages
  65116967            -0.3%   64935325        numa-vmstat.node1.nr_free_pages
    272583 ±  6%      +1.6%     276951 ± 10%  numa-vmstat.node1.nr_inactive_anon
     29.96 ± 77%     -41.3%      17.59 ±118%  numa-vmstat.node1.nr_inactive_file
      0.00       +5.9e+101%       0.59 ±300%  numa-vmstat.node1.nr_isolated_anon
     20596 ±  7%      -1.8%      20229 ±  8%  numa-vmstat.node1.nr_kernel_stack
     85156 ±  4%      +0.3%      85436 ±  4%  numa-vmstat.node1.nr_mapped
      1.08 ±122%      -8.0%       1.00 ±123%  numa-vmstat.node1.nr_mlock
      3929 ± 37%      -6.4%       3678 ± 38%  numa-vmstat.node1.nr_page_table_pages
     95977 ±  3%      +4.1%      99922 ±  7%  numa-vmstat.node1.nr_shmem
     16455 ± 40%     +10.5%      18180 ± 29%  numa-vmstat.node1.nr_slab_reclaimable
     53550 ±  4%      -0.2%      53423 ±  5%  numa-vmstat.node1.nr_slab_unreclaimable
    221867 ±126%     +57.1%     348634 ± 80%  numa-vmstat.node1.nr_unevictable
     10088 ± 22%     +82.5%      18410 ± 53%  numa-vmstat.node1.nr_zone_active_anon
     12.00 ± 81%     -33.4%       7.99 ±122%  numa-vmstat.node1.nr_zone_active_file
    272580 ±  6%      +1.6%     276963 ± 10%  numa-vmstat.node1.nr_zone_inactive_anon
     29.96 ± 77%     -41.3%      17.59 ±118%  numa-vmstat.node1.nr_zone_inactive_file
    221867 ±126%     +57.1%     348634 ± 80%  numa-vmstat.node1.nr_zone_unevictable
  41909012           -12.8%   36528578 ±  9%  numa-vmstat.node1.numa_hit
      0.00          -100.0%       0.00        numa-vmstat.node1.numa_interleave
  41819840           -12.8%   36449493 ±  9%  numa-vmstat.node1.numa_local
     89842 ± 81%     -12.8%      78314 ± 77%  numa-vmstat.node1.numa_other
     59.40 ±  9%     +14.0%      67.70 ±  9%  proc-vmstat.direct_map_level2_splits
      5.00 ± 28%     +12.0%       5.60 ± 27%  proc-vmstat.direct_map_level3_splits
     11737 ± 19%     +72.4%      20235 ± 56%  proc-vmstat.nr_active_anon
     19.60 ±  6%      +2.0%      19.99        proc-vmstat.nr_active_file
    362955            -2.2%     354917 ±  6%  proc-vmstat.nr_anon_pages
     62.13 ± 16%      +5.8%      65.73 ± 14%  proc-vmstat.nr_anon_transparent_hugepages
  12924190            -0.1%   12916030        proc-vmstat.nr_dirty_background_threshold
  25879981            -0.1%   25863641        proc-vmstat.nr_dirty_threshold
    780355            +0.7%     786173        proc-vmstat.nr_file_pages
   1.3e+08            -0.1%  1.299e+08        proc-vmstat.nr_free_pages
    450188            -2.2%     440379 ±  5%  proc-vmstat.nr_inactive_anon
     49.33 ±  2%     -10.5%      44.18 ±  8%  proc-vmstat.nr_inactive_file
      0.57 ±300%      -2.9%       0.56 ±299%  proc-vmstat.nr_isolated_anon
     41099            -0.2%      41022        proc-vmstat.nr_kernel_stack
    129179 ±  2%      +0.3%     129620 ±  2%  proc-vmstat.nr_mapped
      2.73 ±  2%      -4.2%       2.62 ±  8%  proc-vmstat.nr_mlock
      6394 ±  3%      -1.9%       6270 ±  7%  proc-vmstat.nr_page_table_pages
    101324 ±  2%      +5.7%     107147 ±  8%  proc-vmstat.nr_shmem
     37036            +0.2%      37105        proc-vmstat.nr_slab_reclaimable
    115732            +0.2%     115946        proc-vmstat.nr_slab_unreclaimable
    678962            -0.0%     678961        proc-vmstat.nr_unevictable
     11737 ± 19%     +72.4%      20235 ± 56%  proc-vmstat.nr_zone_active_anon
     19.60 ±  6%      +2.0%      19.99        proc-vmstat.nr_zone_active_file
    450188            -2.2%     440380 ±  5%  proc-vmstat.nr_zone_inactive_anon
     49.33 ±  2%     -10.5%      44.18 ±  8%  proc-vmstat.nr_zone_inactive_file
    678962            -0.0%     678961        proc-vmstat.nr_zone_unevictable
     59062 ±  5%      -2.1%      57815 ± 10%  proc-vmstat.numa_hint_faults
     55966 ±  7%      -3.9%      53755 ±  9%  proc-vmstat.numa_hint_faults_local
  84235288           -12.8%   73412158 ±  8%  proc-vmstat.numa_hit
     36.70 ±  4%      -1.9%      36.00        proc-vmstat.numa_huge_pte_updates
      0.00          -100.0%       0.00        proc-vmstat.numa_interleave
  84003235           -12.9%   73179535 ±  8%  proc-vmstat.numa_local
    231828            +0.0%     231864        proc-vmstat.numa_other
      4365 ± 37%      -0.1%       4362 ± 61%  proc-vmstat.numa_pages_migrated
    191149            +1.4%     193820 ±  4%  proc-vmstat.numa_pte_updates
     72087 ± 36%      -8.6%      65920 ± 67%  proc-vmstat.pgactivate
  85864265           -12.5%   75117942 ±  8%  proc-vmstat.pgalloc_normal
  82879163           -13.0%   72119383 ±  9%  proc-vmstat.pgfault
  85427394           -12.6%   74693640 ±  8%  proc-vmstat.pgfree
      4365 ± 37%      -0.1%       4362 ± 61%  proc-vmstat.pgmigrate_success
      0.00          -100.0%       0.00        proc-vmstat.pgpgin
     27273 ±  3%      +4.5%      28501 ± 11%  proc-vmstat.pgreuse
     46.70 ±  9%     +11.3%      52.00 ± 14%  proc-vmstat.thp_collapse_alloc
      0.20 ±200%    +200.0%       0.60 ± 81%  proc-vmstat.thp_deferred_split_page
     48.40            +0.6%      48.70        proc-vmstat.thp_fault_alloc
      4.80 ± 75%     -37.5%       3.00 ±100%  proc-vmstat.thp_migration_success
      0.20 ±200%    +200.0%       0.60 ± 81%  proc-vmstat.thp_split_pmd
      0.00          -100.0%       0.00        proc-vmstat.thp_zero_page_alloc
     35.00            +0.0%      35.00        proc-vmstat.unevictable_pgs_culled
      3.00            +0.0%       3.00        proc-vmstat.unevictable_pgs_mlocked
      3.00            +0.0%       3.00        proc-vmstat.unevictable_pgs_munlocked
      0.00          -100.0%       0.00        proc-vmstat.unevictable_pgs_rescued
   1563750 ±  2%      +3.0%    1610100 ± 12%  proc-vmstat.unevictable_pgs_scanned
      0.61 ±  3%      +0.4%       0.61        perf-stat.i.MPKI
 1.501e+11 ±  2%      -6.3%  1.407e+11 ±  8%  perf-stat.i.branch-instructions
      0.33 ±  3%      +0.0        0.33 ±  6%  perf-stat.i.branch-miss-rate%
 3.963e+08 ±  2%      -7.9%  3.649e+08 ±  9%  perf-stat.i.branch-misses
     65.13            -2.3       62.80 ±  7%  perf-stat.i.cache-miss-rate%
 7.382e+08 ±  2%      -6.3%  6.919e+08 ±  8%  perf-stat.i.cache-misses
 1.106e+09            -6.6%  1.033e+09 ±  8%  perf-stat.i.cache-references
      9256            -4.6%       8828 ±  6%  perf-stat.i.context-switches
      0.50 ±  7%     +33.3%       0.66 ± 43%  perf-stat.i.cpi
    224180            -0.0%     224171        perf-stat.i.cpu-clock
  5.93e+11 ±  2%      -3.8%  5.706e+11 ±  8%  perf-stat.i.cpu-cycles
    359.21 ±  2%      -6.1%     337.30 ±  3%  perf-stat.i.cpu-migrations
    810.86 ±  3%     +50.1%       1217 ± 61%  perf-stat.i.cycles-between-cache-misses
      0.01 ±134%      +0.1        0.13 ±176%  perf-stat.i.dTLB-load-miss-rate%
  12662250 ±  2%      +1.0%   12784614 ±  2%  perf-stat.i.dTLB-load-misses
 2.861e+11 ±  2%      -6.5%  2.677e+11 ±  8%  perf-stat.i.dTLB-loads
      0.01 ±  9%      +0.0        0.01 ± 43%  perf-stat.i.dTLB-store-miss-rate%
  13024682 ±  2%     -14.4%   11149515 ±  8%  perf-stat.i.dTLB-store-misses
 1.757e+11 ±  2%      -6.6%  1.641e+11 ±  8%  perf-stat.i.dTLB-stores
 1.275e+12 ±  2%      -6.4%  1.194e+12 ±  8%  perf-stat.i.instructions
      2.10            -6.3%       1.97 ±  7%  perf-stat.i.ipc
      0.20 ± 48%      +6.9%       0.21 ± 40%  perf-stat.i.major-faults
      1.26 ±  2%      -6.7%       1.18 ±  9%  perf-stat.i.metric.G/sec
      2.65 ±  2%      -3.8%       2.55 ±  8%  perf-stat.i.metric.GHz
    694.48            -3.0%     673.92 ±  2%  perf-stat.i.metric.K/sec
      1472            -6.2%       1380 ±  8%  perf-stat.i.metric.M/sec
   1305005 ±  2%     -17.5%    1076189 ±  9%  perf-stat.i.minor-faults
      4.59 ±  8%      +2.1        6.65 ± 51%  perf-stat.i.node-load-miss-rate%
   2077324 ±  2%      -2.2%    2032308 ± 12%  perf-stat.i.node-load-misses
 1.353e+08            -5.4%  1.281e+08 ±  8%  perf-stat.i.node-loads
   1305005 ±  2%     -17.5%    1076190 ±  9%  perf-stat.i.page-faults
    224180            -0.0%     224171        perf-stat.i.task-clock
      0.46 ± 50%     +25.1%       0.58        perf-stat.overall.MPKI
      0.21 ± 50%      +0.0        0.26        perf-stat.overall.branch-miss-rate%
     53.47 ± 50%     +13.6       67.06        perf-stat.overall.cache-miss-rate%
      0.37 ± 50%     +28.6%       0.48        perf-stat.overall.cpi
    643.14 ± 50%     +28.4%     826.00        perf-stat.overall.cycles-between-cache-misses
      0.00 ± 50%      +0.0        0.00 ± 10%  perf-stat.overall.dTLB-load-miss-rate%
      0.01 ± 50%      +0.0        0.01 ±  7%  perf-stat.overall.dTLB-store-miss-rate%
      1.72 ± 50%     +21.6%       2.09        perf-stat.overall.ipc
      1.19 ± 50%      +0.4        1.58 ± 12%  perf-stat.overall.node-load-miss-rate%
 1.173e+11 ± 50%     +17.9%  1.383e+11 ±  7%  perf-stat.ps.branch-instructions
 3.098e+08 ± 50%     +15.8%  3.586e+08 ±  8%  perf-stat.ps.branch-misses
 5.777e+08 ± 50%     +17.9%  6.809e+08 ±  7%  perf-stat.ps.cache-misses
 8.643e+08 ± 50%     +17.4%  1.015e+09 ±  7%  perf-stat.ps.cache-references
      7130 ± 50%     +19.9%       8552 ±  5%  perf-stat.ps.context-switches
    174005 ± 50%     +25.5%     218373        perf-stat.ps.cpu-clock
 4.644e+11 ± 50%     +21.0%  5.621e+11 ±  7%  perf-stat.ps.cpu-cycles
    272.51 ± 50%     +16.6%     317.85 ±  2%  perf-stat.ps.cpu-migrations
   9924949 ± 50%     +25.8%   12485559 ±  2%  perf-stat.ps.dTLB-load-misses
 2.237e+11 ± 50%     +17.6%  2.631e+11 ±  7%  perf-stat.ps.dTLB-loads
  10159493 ± 50%      +7.8%   10948561 ±  7%  perf-stat.ps.dTLB-store-misses
 1.374e+11 ± 50%     +17.4%  1.613e+11 ±  7%  perf-stat.ps.dTLB-stores
 9.964e+11 ± 50%     +17.7%  1.173e+12 ±  7%  perf-stat.ps.instructions
      0.16 ± 73%     +21.3%       0.20 ± 35%  perf-stat.ps.major-faults
   1016189 ± 50%      +3.9%    1056067 ±  9%  perf-stat.ps.minor-faults
   1612917 ± 50%     +25.8%    2029077 ± 12%  perf-stat.ps.node-load-misses
 1.067e+08 ± 50%     +18.9%  1.269e+08 ±  7%  perf-stat.ps.node-loads
   1016189 ± 50%      +3.9%    1056067 ±  9%  perf-stat.ps.page-faults
    174005 ± 50%     +25.5%     218373        perf-stat.ps.task-clock
 6.343e+13 ± 50%     +23.6%  7.843e+13        perf-stat.total.instructions
   7114718            -1.2%    7027229        sched_debug.cfs_rq:/.avg_vruntime.avg
   7180297            -1.0%    7105319        sched_debug.cfs_rq:/.avg_vruntime.max
   4657471            -3.9%    4477025 ±  2%  sched_debug.cfs_rq:/.avg_vruntime.min
    174545 ±  2%      +3.8%     181256 ±  3%  sched_debug.cfs_rq:/.avg_vruntime.stddev
      0.55            +2.0%       0.56        sched_debug.cfs_rq:/.h_nr_running.avg
      2.10 ± 20%      +9.5%       2.30 ± 17%  sched_debug.cfs_rq:/.h_nr_running.max
      0.50            +0.0%       0.50        sched_debug.cfs_rq:/.h_nr_running.min
      0.22 ±  8%     +12.7%       0.25 ± 10%  sched_debug.cfs_rq:/.h_nr_running.stddev
     43106 ± 33%    +312.8%     177940 ± 52%  sched_debug.cfs_rq:/.left_vruntime.avg
   6406425 ± 17%     +10.2%    7059947        sched_debug.cfs_rq:/.left_vruntime.max
      0.00            +0.0%       0.00        sched_debug.cfs_rq:/.left_vruntime.min
    490512 ± 23%    +111.4%    1037149 ± 28%  sched_debug.cfs_rq:/.left_vruntime.stddev
      9516 ± 26%    +105.3%      19537 ± 37%  sched_debug.cfs_rq:/.load.avg
   1009013 ± 48%     +15.6%    1166009 ± 43%  sched_debug.cfs_rq:/.load.max
      2280            -0.4%       2271        sched_debug.cfs_rq:/.load.min
     80417 ± 39%     +48.9%     119761 ± 28%  sched_debug.cfs_rq:/.load.stddev
    615.03 ± 25%     +70.0%       1045 ± 46%  sched_debug.cfs_rq:/.load_avg.avg
     43884 ±  3%      +0.4%      44060 ±  3%  sched_debug.cfs_rq:/.load_avg.max
      2.15 ± 10%      +0.0%       2.15 ± 14%  sched_debug.cfs_rq:/.load_avg.min
      4551 ± 15%     +29.7%       5903 ± 27%  sched_debug.cfs_rq:/.load_avg.stddev
   7114719            -1.2%    7027233        sched_debug.cfs_rq:/.min_vruntime.avg
   7180297            -1.0%    7105319        sched_debug.cfs_rq:/.min_vruntime.max
   4657590            -3.9%    4477122 ±  2%  sched_debug.cfs_rq:/.min_vruntime.min
    174538 ±  2%      +3.8%     181250 ±  3%  sched_debug.cfs_rq:/.min_vruntime.stddev
      0.54            +1.4%       0.54        sched_debug.cfs_rq:/.nr_running.avg
      1.80 ± 22%      +8.3%       1.95 ± 17%  sched_debug.cfs_rq:/.nr_running.max
      0.50            +0.0%       0.50        sched_debug.cfs_rq:/.nr_running.min
      0.17 ± 12%     +18.9%       0.21 ± 12%  sched_debug.cfs_rq:/.nr_running.stddev
     92.62 ±144%    +105.7%     190.49 ± 96%  sched_debug.cfs_rq:/.removed.load_avg.avg
     13672 ±147%     +96.3%      26833 ± 80%  sched_debug.cfs_rq:/.removed.load_avg.max
      1060 ±145%    +102.2%       2144 ± 84%  sched_debug.cfs_rq:/.removed.load_avg.stddev
      6.84 ± 48%      -1.3%       6.76 ± 45%  sched_debug.cfs_rq:/.removed.runnable_avg.avg
    265.30 ±  2%      -1.6%     261.00        sched_debug.cfs_rq:/.removed.runnable_avg.max
     39.85 ± 21%      -1.3%      39.33 ± 20%  sched_debug.cfs_rq:/.removed.runnable_avg.stddev
      6.84 ± 48%      -1.3%       6.76 ± 45%  sched_debug.cfs_rq:/.removed.util_avg.avg
    265.30 ±  2%      -1.6%     261.00        sched_debug.cfs_rq:/.removed.util_avg.max
     39.85 ± 21%      -1.3%      39.33 ± 20%  sched_debug.cfs_rq:/.removed.util_avg.stddev
     43106 ± 33%    +312.8%     177940 ± 52%  sched_debug.cfs_rq:/.right_vruntime.avg
   6406449 ± 17%     +10.2%    7059947        sched_debug.cfs_rq:/.right_vruntime.max
      0.00            +0.0%       0.00        sched_debug.cfs_rq:/.right_vruntime.min
    490516 ± 23%    +111.4%    1037152 ± 28%  sched_debug.cfs_rq:/.right_vruntime.stddev
    602.56            +1.7%     612.78        sched_debug.cfs_rq:/.runnable_avg.avg
      1824 ± 15%      +8.0%       1970 ± 13%  sched_debug.cfs_rq:/.runnable_avg.max
    511.80            -9.8%     461.45 ± 32%  sched_debug.cfs_rq:/.runnable_avg.min
    197.27 ±  7%      +6.7%     210.57 ±  8%  sched_debug.cfs_rq:/.runnable_avg.stddev
      0.50 ± 99%     +10.0%       0.55 ±124%  sched_debug.cfs_rq:/.spread.avg
    111.63 ± 99%      +7.6%     120.07 ±126%  sched_debug.cfs_rq:/.spread.max
      7.44 ± 99%      +7.7%       8.02 ±125%  sched_debug.cfs_rq:/.spread.stddev
    588.90            +0.3%     590.76        sched_debug.cfs_rq:/.util_avg.avg
      1267 ± 16%      +1.5%       1286 ± 17%  sched_debug.cfs_rq:/.util_avg.max
     32.70 ±290%    +205.4%      99.85 ±172%  sched_debug.cfs_rq:/.util_avg.min
    163.16 ±  4%      +0.8%     164.51 ±  9%  sched_debug.cfs_rq:/.util_avg.stddev
    261.48            +0.7%     263.22        sched_debug.cfs_rq:/.util_est_enqueued.avg
    924.20            +9.8%       1014 ± 13%  sched_debug.cfs_rq:/.util_est_enqueued.max
     16.40 ± 90%     +74.4%      28.60 ± 42%  sched_debug.cfs_rq:/.util_est_enqueued.min
     87.88 ±  8%      +5.1%      92.40 ± 10%  sched_debug.cfs_rq:/.util_est_enqueued.stddev
    895647            +2.7%     919389 ±  5%  sched_debug.cpu.avg_idle.avg
   1584773 ± 25%      -5.4%    1499839 ± 21%  sched_debug.cpu.avg_idle.max
    180026 ± 47%     -17.3%     148883 ± 47%  sched_debug.cpu.avg_idle.min
    139458 ± 12%     +17.5%     163799 ± 45%  sched_debug.cpu.avg_idle.stddev
     91837 ± 13%      +0.9%      92634 ± 10%  sched_debug.cpu.clock.avg
     91883 ± 13%      +0.9%      92682 ± 10%  sched_debug.cpu.clock.max
     91778 ± 13%      +0.9%      92573 ± 10%  sched_debug.cpu.clock.min
     28.81 ± 15%      +1.8%      29.34 ± 13%  sched_debug.cpu.clock.stddev
     91190 ± 13%      +0.9%      91995 ± 10%  sched_debug.cpu.clock_task.avg
     91362 ± 13%      +0.9%      92168 ± 10%  sched_debug.cpu.clock_task.max
     78321 ± 15%      +1.2%      79242 ± 12%  sched_debug.cpu.clock_task.min
    873.05            -1.0%     864.68        sched_debug.cpu.clock_task.stddev
      4249            -1.9%       4166        sched_debug.cpu.curr->pid.avg
      7337 ±  3%      +0.8%       7398 ±  2%  sched_debug.cpu.curr->pid.max
      1048 ± 97%     -90.8%      96.65 ± 46%  sched_debug.cpu.curr->pid.min
    836.33 ± 10%     +35.5%       1132 ± 14%  sched_debug.cpu.curr->pid.stddev
    502089            +1.2%     508197 ±  3%  sched_debug.cpu.max_idle_balance_cost.avg
    700999 ± 18%      -2.1%     685975 ± 15%  sched_debug.cpu.max_idle_balance_cost.max
    500000            +0.0%     500000        sched_debug.cpu.max_idle_balance_cost.min
     17056 ± 57%     +39.7%      23833 ±121%  sched_debug.cpu.max_idle_balance_cost.stddev
      4294            +0.0%       4294        sched_debug.cpu.next_balance.avg
      4294            +0.0%       4294        sched_debug.cpu.next_balance.max
      4294            +0.0%       4294        sched_debug.cpu.next_balance.min
      0.00 ± 57%     -20.1%       0.00 ± 36%  sched_debug.cpu.next_balance.stddev
      0.55            +1.9%       0.56 ±  2%  sched_debug.cpu.nr_running.avg
      2.45 ± 19%      +2.0%       2.50 ± 20%  sched_debug.cpu.nr_running.max
      0.50            +0.0%       0.50        sched_debug.cpu.nr_running.min
      0.24 ± 10%      +8.1%       0.26 ± 11%  sched_debug.cpu.nr_running.stddev
      2437            -0.6%       2423        sched_debug.cpu.nr_switches.avg
     46748 ± 21%      +2.2%      47779 ± 11%  sched_debug.cpu.nr_switches.max
      1195            -3.2%       1157 ±  2%  sched_debug.cpu.nr_switches.min
      3810 ± 11%      +1.6%       3869 ±  6%  sched_debug.cpu.nr_switches.stddev
      0.01 ± 56%     -54.2%       0.00 ±110%  sched_debug.cpu.nr_uninterruptible.avg
     42.85 ± 38%      -0.9%      42.45 ± 38%  sched_debug.cpu.nr_uninterruptible.max
    -17.50           +35.1%     -23.65        sched_debug.cpu.nr_uninterruptible.min
      5.78 ± 12%      +4.1%       6.02 ± 12%  sched_debug.cpu.nr_uninterruptible.stddev
     91781 ± 13%      +0.9%      92577 ± 10%  sched_debug.cpu_clk
    996147            +0.0%     996147        sched_debug.dl_rq:.dl_bw->bw.avg
    996147            +0.0%     996147        sched_debug.dl_rq:.dl_bw->bw.max
    996147            +0.0%     996147        sched_debug.dl_rq:.dl_bw->bw.min
 4.295e+09            +0.0%  4.295e+09        sched_debug.jiffies
     90574 ± 13%      +0.9%      91369 ± 10%  sched_debug.ktime
      0.00            +0.0%       0.00        sched_debug.rt_rq:.rt_nr_migratory.avg
      0.50            +0.0%       0.50        sched_debug.rt_rq:.rt_nr_migratory.max
      0.03            +0.0%       0.03        sched_debug.rt_rq:.rt_nr_migratory.stddev
      0.00            +0.0%       0.00        sched_debug.rt_rq:.rt_nr_running.avg
      0.50            +0.0%       0.50        sched_debug.rt_rq:.rt_nr_running.max
      0.03            +0.0%       0.03        sched_debug.rt_rq:.rt_nr_running.stddev
    950.00            +0.0%     950.00        sched_debug.rt_rq:.rt_runtime.avg
    950.00            +0.0%     950.00        sched_debug.rt_rq:.rt_runtime.max
    950.00            +0.0%     950.00        sched_debug.rt_rq:.rt_runtime.min
      0.00        +2.2e+98%       0.00 ±299%  sched_debug.rt_rq:.rt_throttled.avg
      0.00         +5e+100%       0.05 ±300%  sched_debug.rt_rq:.rt_throttled.max
      0.00        +3.3e+99%       0.00 ±299%  sched_debug.rt_rq:.rt_throttled.stddev
      0.91 ± 64%      -7.2%       0.84 ± 91%  sched_debug.rt_rq:.rt_time.avg
    203.22 ± 64%      -7.2%     188.67 ± 91%  sched_debug.rt_rq:.rt_time.max
     13.55 ± 64%      -7.2%      12.58 ± 91%  sched_debug.rt_rq:.rt_time.stddev
     92678 ± 12%      +0.9%      93471 ± 10%  sched_debug.sched_clk
      1.00            +0.0%       1.00        sched_debug.sched_clock_stable()
      3.00            +0.0%       3.00        sched_debug.sysctl_sched.sysctl_sched_base_slice
  25056823            +0.0%   25056823        sched_debug.sysctl_sched.sysctl_sched_features
      1.00            +0.0%       1.00        sched_debug.sysctl_sched.sysctl_sched_tunable_scaling
      0.13 ±200%      -0.1        0.00        perf-profile.calltrace.cycles-pp.tick_sched_handle.tick_sched_timer.__hrtimer_run_queues.hrtimer_interrupt.__sysvec_apic_timer_interrupt
      0.13 ±200%      -0.1        0.00        perf-profile.calltrace.cycles-pp.update_process_times.tick_sched_handle.tick_sched_timer.__hrtimer_run_queues.hrtimer_interrupt
      0.12 ±200%      -0.1        0.00        perf-profile.calltrace.cycles-pp.scheduler_tick.update_process_times.tick_sched_handle.tick_sched_timer.__hrtimer_run_queues
      0.10 ±200%      -0.1        0.05 ±300%  perf-profile.calltrace.cycles-pp.j__udyLAllocJBBJP
      0.05 ±299%      -0.1        0.00        perf-profile.calltrace.cycles-pp.get_mem_cgroup_from_mm.__mem_cgroup_charge.do_anonymous_page.__handle_mm_fault.handle_mm_fault
      0.14 ±199%      -0.0        0.10 ±200%  perf-profile.calltrace.cycles-pp.tick_sched_timer.__hrtimer_run_queues.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt
      0.12 ±200%      +0.0        0.16 ±152%  perf-profile.calltrace.cycles-pp.page_counter_uncharge.uncharge_batch.__mem_cgroup_uncharge_list.release_pages.tlb_batch_pages_flush
      0.13 ±200%      +0.1        0.22 ±122%  perf-profile.calltrace.cycles-pp.uncharge_batch.__mem_cgroup_uncharge_list.release_pages.tlb_batch_pages_flush.tlb_finish_mmu
      0.25 ±200%      +0.1        0.40 ± 65%  perf-profile.calltrace.cycles-pp.__hrtimer_run_queues.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt
      0.00            +0.2        0.19 ±152%  perf-profile.calltrace.cycles-pp.rmqueue.get_page_from_freelist.__alloc_pages.pte_alloc_one.__pte_alloc
      0.00            +0.2        0.19 ±152%  perf-profile.calltrace.cycles-pp.__rmqueue_pcplist.rmqueue.get_page_from_freelist.__alloc_pages.pte_alloc_one
      0.00            +0.2        0.19 ±153%  perf-profile.calltrace.cycles-pp.get_page_from_freelist.__alloc_pages.pte_alloc_one.__pte_alloc.do_anonymous_page
      0.00            +0.2        0.20 ±152%  perf-profile.calltrace.cycles-pp.__alloc_pages.pte_alloc_one.__pte_alloc.do_anonymous_page.__handle_mm_fault
      0.00            +0.2        0.20 ±152%  perf-profile.calltrace.cycles-pp.__pte_alloc.do_anonymous_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
      0.00            +0.2        0.20 ±152%  perf-profile.calltrace.cycles-pp.pte_alloc_one.__pte_alloc.do_anonymous_page.__handle_mm_fault.handle_mm_fault
      0.23 ±200%      +0.2        0.43 ±122%  perf-profile.calltrace.cycles-pp.JudyLDel@plt
      0.14 ±200%      +0.3        0.40 ± 66%  perf-profile.calltrace.cycles-pp.__mem_cgroup_uncharge_list.release_pages.tlb_batch_pages_flush.tlb_finish_mmu.unmap_region
      0.12 ±200%      +0.3        0.38 ± 65%  perf-profile.calltrace.cycles-pp.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.JudyLGet
      0.12 ±200%      +0.3        0.38 ± 65%  perf-profile.calltrace.cycles-pp.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.JudyLGet
      0.16 ±200%      +0.3        0.42 ± 66%  perf-profile.calltrace.cycles-pp.__mem_cgroup_charge.do_anonymous_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
      0.13 ±200%      +0.3        0.43 ± 65%  perf-profile.calltrace.cycles-pp.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.JudyLGet
      0.14 ±200%      +0.3        0.47 ± 65%  perf-profile.calltrace.cycles-pp.asm_sysvec_apic_timer_interrupt.JudyLGet
      0.17 ±200%      +0.4        0.53 ± 65%  perf-profile.calltrace.cycles-pp.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt
      0.17 ±200%      +0.4        0.54 ± 65%  perf-profile.calltrace.cycles-pp.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt
      0.16 ±200%      +0.4        0.53 ± 68%  perf-profile.calltrace.cycles-pp.ordered_events__queue.process_simple.reader__read_event.perf_session__process_events.record__finish_output
      0.16 ±200%      +0.4        0.53 ± 68%  perf-profile.calltrace.cycles-pp.queue_event.ordered_events__queue.process_simple.reader__read_event.perf_session__process_events
      0.17 ±200%      +0.4        0.54 ± 65%  perf-profile.calltrace.cycles-pp.cfree
      0.16 ±200%      +0.4        0.53 ± 68%  perf-profile.calltrace.cycles-pp.process_simple.reader__read_event.perf_session__process_events.record__finish_output.__cmd_record
      0.16 ±200%      +0.4        0.53 ± 68%  perf-profile.calltrace.cycles-pp.__cmd_record
      0.16 ±200%      +0.4        0.53 ± 68%  perf-profile.calltrace.cycles-pp.record__finish_output.__cmd_record
      0.16 ±200%      +0.4        0.53 ± 68%  perf-profile.calltrace.cycles-pp.perf_session__process_events.record__finish_output.__cmd_record
      0.16 ±200%      +0.4        0.53 ± 68%  perf-profile.calltrace.cycles-pp.reader__read_event.perf_session__process_events.record__finish_output.__cmd_record
      0.20 ±200%      +0.4        0.65 ± 65%  perf-profile.calltrace.cycles-pp.__split_vma.do_vmi_align_munmap.__do_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
      0.21 ±200%      +0.5        0.67 ± 65%  perf-profile.calltrace.cycles-pp.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt
      0.22 ±200%      +0.5        0.72 ± 65%  perf-profile.calltrace.cycles-pp.asm_sysvec_apic_timer_interrupt
      0.25 ±200%      +0.5        0.77 ± 65%  perf-profile.calltrace.cycles-pp.j__udyLCascade2
      0.24 ±200%      +0.5        0.76 ± 65%  perf-profile.calltrace.cycles-pp.j__udyLCreateBranchU
      0.27 ±200%      +0.6        0.87 ± 65%  perf-profile.calltrace.cycles-pp.JudyLGet@plt
      0.27 ±200%      +0.6        0.88 ± 65%  perf-profile.calltrace.cycles-pp.JudyLIns@plt
      0.00            +0.7        0.68 ± 92%  perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.free_pcppages_bulk.free_unref_page_commit.free_unref_page.rcu_do_batch
      0.34 ±200%      +0.8        1.11 ± 65%  perf-profile.calltrace.cycles-pp.malloc
      0.00            +0.8        0.82 ± 79%  perf-profile.calltrace.cycles-pp.free_pcppages_bulk.free_unref_page_commit.free_unref_page.rcu_do_batch.rcu_core
      0.00            +1.1        1.13 ± 84%  perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.free_pcppages_bulk.free_unref_page_commit.free_unref_page.__unfreeze_partials
      0.00            +1.2        1.24 ± 83%  perf-profile.calltrace.cycles-pp.free_pcppages_bulk.free_unref_page_commit.free_unref_page.__unfreeze_partials.rcu_do_batch
      0.00            +1.2        1.24 ± 82%  perf-profile.calltrace.cycles-pp.free_unref_page_commit.free_unref_page.__unfreeze_partials.rcu_do_batch.rcu_core
      0.73 ±200%      +1.6        2.29 ± 65%  perf-profile.calltrace.cycles-pp.JudyLDel
      0.00            +1.6        1.57 ±101%  perf-profile.calltrace.cycles-pp.__rmqueue_pcplist.rmqueue.get_page_from_freelist.__alloc_pages.__folio_alloc
      0.70 ±200%      +1.6        2.27 ± 65%  perf-profile.calltrace.cycles-pp.JudyLIns
      0.00            +1.6        1.58 ±100%  perf-profile.calltrace.cycles-pp.rmqueue.get_page_from_freelist.__alloc_pages.__folio_alloc.vma_alloc_folio
      0.00            +1.6        1.59 ±110%  perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.rmqueue_bulk.__rmqueue_pcplist.rmqueue.get_page_from_freelist
      0.00            +1.6        1.59 ±110%  perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.rmqueue_bulk.__rmqueue_pcplist.rmqueue
      0.00            +1.7        1.71 ±108%  perf-profile.calltrace.cycles-pp.rmqueue_bulk.__rmqueue_pcplist.rmqueue.get_page_from_freelist.__alloc_pages
      0.12 ±200%      +1.7        1.86 ± 92%  perf-profile.calltrace.cycles-pp.get_page_from_freelist.__alloc_pages.__folio_alloc.vma_alloc_folio.do_anonymous_page
      0.12 ±200%      +1.8        1.88 ± 92%  perf-profile.calltrace.cycles-pp.__alloc_pages.__folio_alloc.vma_alloc_folio.do_anonymous_page.__handle_mm_fault
      0.12 ±200%      +1.8        1.88 ± 92%  perf-profile.calltrace.cycles-pp.__folio_alloc.vma_alloc_folio.do_anonymous_page.__handle_mm_fault.handle_mm_fault
      0.13 ±200%      +1.8        1.90 ± 91%  perf-profile.calltrace.cycles-pp.vma_alloc_folio.do_anonymous_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
      0.00            +1.8        1.81 ± 87%  perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.free_pcppages_bulk.free_unref_page_commit.free_unref_page
      0.12 ±200%      +2.2        2.30 ± 73%  perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.folio_lruvec_lock_irqsave.release_pages.tlb_batch_pages_flush.tlb_finish_mmu
      0.13 ±200%      +2.2        2.32 ± 73%  perf-profile.calltrace.cycles-pp.folio_lruvec_lock_irqsave.release_pages.tlb_batch_pages_flush.tlb_finish_mmu.unmap_region
      0.00            +2.2        2.19 ± 74%  perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.folio_lruvec_lock_irqsave.release_pages.tlb_batch_pages_flush
      0.12 ±200%      +2.3        2.42 ± 77%  perf-profile.calltrace.cycles-pp.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      0.12 ±200%      +2.3        2.45 ± 77%  perf-profile.calltrace.cycles-pp.kthread.ret_from_fork.ret_from_fork_asm
      0.12 ±200%      +2.3        2.45 ± 77%  perf-profile.calltrace.cycles-pp.ret_from_fork_asm
      0.12 ±200%      +2.3        2.45 ± 77%  perf-profile.calltrace.cycles-pp.ret_from_fork.ret_from_fork_asm
      1.05 ±200%      +2.4        3.42 ± 65%  perf-profile.calltrace.cycles-pp.stress_judy
      0.33 ±200%      +2.6        2.90 ± 70%  perf-profile.calltrace.cycles-pp.release_pages.tlb_batch_pages_flush.tlb_finish_mmu.unmap_region.do_vmi_align_munmap
      0.34 ±200%      +2.6        2.92 ± 70%  perf-profile.calltrace.cycles-pp.tlb_batch_pages_flush.tlb_finish_mmu.unmap_region.do_vmi_align_munmap.__do_sys_brk
      0.46 ±200%      +2.6        3.10 ± 80%  perf-profile.calltrace.cycles-pp.do_anonymous_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault
      0.38 ±200%      +2.7        3.04 ± 69%  perf-profile.calltrace.cycles-pp.tlb_finish_mmu.unmap_region.do_vmi_align_munmap.__do_sys_brk.do_syscall_64
      0.48 ±200%      +2.7        3.15 ± 79%  perf-profile.calltrace.cycles-pp.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
      0.50 ±200%      +2.7        3.20 ± 79%  perf-profile.calltrace.cycles-pp.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
      0.53 ±200%      +2.8        3.30 ± 78%  perf-profile.calltrace.cycles-pp.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
      0.53 ±200%      +2.8        3.31 ± 78%  perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault
      0.57 ±200%      +2.9        3.43 ± 77%  perf-profile.calltrace.cycles-pp.asm_exc_page_fault
      0.49 ±200%      +2.9        3.36 ± 68%  perf-profile.calltrace.cycles-pp.unmap_region.do_vmi_align_munmap.__do_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
      0.78 ±200%      +3.5        4.26 ± 67%  perf-profile.calltrace.cycles-pp.do_vmi_align_munmap.__do_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe.brk
      0.81 ±200%      +3.6        4.36 ± 67%  perf-profile.calltrace.cycles-pp.__do_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe.brk
      0.82 ±200%      +3.6        4.38 ± 67%  perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.brk
      0.82 ±200%      +3.6        4.38 ± 67%  perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.brk
      0.83 ±200%      +3.6        4.40 ± 67%  perf-profile.calltrace.cycles-pp.brk
     13.33 ±200%     +29.8       43.13 ± 65%  perf-profile.calltrace.cycles-pp.JudyLGet
      0.02 ±200%      -0.0        0.00        perf-profile.children.cycles-pp.rcu_sched_clock_irq
      0.01 ±200%      -0.0        0.00        perf-profile.children.cycles-pp.rcu_pending
      0.01 ±201%      -0.0        0.00        perf-profile.children.cycles-pp.update_cfs_group
      0.01 ±200%      -0.0        0.00        perf-profile.children.cycles-pp.perf_rotate_context
      0.01 ±200%      -0.0        0.00        perf-profile.children.cycles-pp.check_cpu_stall
      0.01 ±200%      -0.0        0.00        perf-profile.children.cycles-pp.arch_scale_freq_tick
      0.01 ±300%      -0.0        0.00        perf-profile.children.cycles-pp.down_write
      0.01 ±200%      +0.0        0.01 ±200%  perf-profile.children.cycles-pp.__get_obj_cgroup_from_memcg
      0.01 ±300%      +0.0        0.01 ±299%  perf-profile.children.cycles-pp.mas_store_b_node
      0.01 ±300%      +0.0        0.01 ±299%  perf-profile.children.cycles-pp.cgroup_rstat_updated
      0.01 ±200%      +0.0        0.02 ±152%  perf-profile.children.cycles-pp.calc_global_load_tick
      0.02 ±200%      +0.0        0.02 ±122%  perf-profile.children.cycles-pp._compound_head
      0.00            +0.0        0.01 ±299%  perf-profile.children.cycles-pp.perf_mmap__read_head
      0.00            +0.0        0.01 ±300%  perf-profile.children.cycles-pp.__get_free_pages
      0.00            +0.0        0.01 ±300%  perf-profile.children.cycles-pp.__tlb_remove_page_size
      0.01 ±201%      +0.0        0.02 ±122%  perf-profile.children.cycles-pp.j__udyLAllocJLL3
      0.01 ±200%      +0.0        0.02 ±100%  perf-profile.children.cycles-pp.rcu_nocb_try_bypass
      0.00            +0.0        0.02 ±152%  perf-profile.children.cycles-pp.__list_add_valid_or_report
      0.02 ±200%      +0.0        0.04 ± 65%  perf-profile.children.cycles-pp.perf_trace_sched_stat_runtime
      0.00            +0.0        0.02 ±153%  perf-profile.children.cycles-pp.worker_thread
      0.00            +0.0        0.02 ±153%  perf-profile.children.cycles-pp.process_one_work
      0.01 ±200%      +0.0        0.03 ± 82%  perf-profile.children.cycles-pp.mas_find
      0.01 ±201%      +0.0        0.03 ± 81%  perf-profile.children.cycles-pp.mtree_range_walk
      0.03 ±200%      +0.0        0.05 ± 67%  perf-profile.children.cycles-pp.propagate_protected_usage
      0.01 ±200%      +0.0        0.04 ± 65%  perf-profile.children.cycles-pp.update_load_avg
      0.01 ±201%      +0.0        0.04 ± 65%  perf-profile.children.cycles-pp.mas_destroy
      0.01 ±201%      +0.0        0.04 ±100%  perf-profile.children.cycles-pp.tick_sched_do_timer
      0.01 ±200%      +0.0        0.04 ± 66%  perf-profile.children.cycles-pp.get_obj_cgroup_from_current
      0.01 ±200%      +0.0        0.04 ± 66%  perf-profile.children.cycles-pp.mem_cgroup_update_lru_size
      0.01 ±200%      +0.0        0.04 ± 66%  perf-profile.children.cycles-pp.__cond_resched
      0.01 ±200%      +0.0        0.04 ± 66%  perf-profile.children.cycles-pp.__mod_node_page_state
      0.01 ±200%      +0.0        0.04 ± 66%  perf-profile.children.cycles-pp.do_brk_flags
      0.01 ±200%      +0.0        0.04 ± 66%  perf-profile.children.cycles-pp.lock_vma_under_rcu
      0.02 ±200%      +0.0        0.05 ± 66%  perf-profile.children.cycles-pp.hrtimer_active
      0.02 ±200%      +0.0        0.04 ± 66%  perf-profile.children.cycles-pp.free_unref_page_list
      0.01 ±200%      +0.0        0.04 ± 66%  perf-profile.children.cycles-pp.unlink_anon_vmas
      0.02 ±200%      +0.0        0.05 ± 65%  perf-profile.children.cycles-pp.___perf_sw_event
      0.02 ±200%      +0.0        0.05 ± 66%  perf-profile.children.cycles-pp.__mod_lruvec_state
      0.02 ±200%      +0.0        0.05 ± 65%  perf-profile.children.cycles-pp.try_charge_memcg
      0.03 ±200%      +0.0        0.07 ± 65%  perf-profile.children.cycles-pp.update_curr
      0.02 ±200%      +0.0        0.05 ± 65%  perf-profile.children.cycles-pp.mas_walk
      0.01 ±200%      +0.0        0.05 ± 65%  perf-profile.children.cycles-pp.task_mm_cid_work
      0.00            +0.0        0.04 ± 66%  perf-profile.children.cycles-pp.note_gp_changes
      0.02 ±200%      +0.0        0.05 ± 65%  perf-profile.children.cycles-pp.free_pgtables
      0.02 ±200%      +0.0        0.05 ± 65%  perf-profile.children.cycles-pp.task_work_run
      0.02 ±200%      +0.0        0.06 ± 66%  perf-profile.children.cycles-pp.__perf_sw_event
      0.02 ±200%      +0.0        0.06 ± 66%  perf-profile.children.cycles-pp.j__udyLCascade3
      0.02 ±200%      +0.0        0.06 ± 65%  perf-profile.children.cycles-pp.j__udyLFreeJLL2
      0.02 ±200%      +0.0        0.06 ± 65%  perf-profile.children.cycles-pp.exit_to_user_mode_loop
      0.02 ±200%      +0.0        0.07 ± 67%  perf-profile.children.cycles-pp.lru_add_fn
      0.04 ±200%      +0.0        0.08 ± 66%  perf-profile.children.cycles-pp.mem_cgroup_commit_charge
      0.02 ±200%      +0.0        0.07 ± 65%  perf-profile.children.cycles-pp.j__udyLAllocJBBJP@plt
      0.02 ±200%      +0.0        0.06 ± 65%  perf-profile.children.cycles-pp.irqentry_exit_to_user_mode
      0.04 ±200%      +0.0        0.09 ± 66%  perf-profile.children.cycles-pp.__count_memcg_events
      0.02 ±200%      +0.0        0.07 ± 65%  perf-profile.children.cycles-pp.j__udyLLeaf2ToLeaf3
      0.02 ±200%      +0.0        0.07 ± 65%  perf-profile.children.cycles-pp.exit_to_user_mode_prepare
      0.02 ±200%      +0.0        0.07 ± 65%  perf-profile.children.cycles-pp.mas_commit_b_node
      0.02 ±200%      +0.0        0.07 ± 65%  perf-profile.children.cycles-pp.__call_rcu_common
      0.01 ±200%      +0.1        0.06 ± 68%  perf-profile.children.cycles-pp.generic_perform_write
      0.02 ±200%      +0.1        0.07 ± 65%  perf-profile.children.cycles-pp.j__udyLFreeJBBJP@plt
      0.03 ±200%      +0.1        0.08 ± 65%  perf-profile.children.cycles-pp.anon_vma_clone
      0.05 ±200%      +0.1        0.10 ± 66%  perf-profile.children.cycles-pp.__mod_memcg_lruvec_state
      0.01 ±200%      +0.1        0.07 ± 69%  perf-profile.children.cycles-pp.shmem_file_write_iter
      0.02 ±200%      +0.1        0.07 ± 70%  perf-profile.children.cycles-pp.vfs_write
      0.03 ±200%      +0.1        0.09 ± 65%  perf-profile.children.cycles-pp.mod_objcg_state
      0.03 ±200%      +0.1        0.08 ± 65%  perf-profile.children.cycles-pp.memcg_slab_post_alloc_hook
      0.02 ±200%      +0.1        0.07 ± 69%  perf-profile.children.cycles-pp.ksys_write
      0.00            +0.1        0.06 ± 74%  perf-profile.children.cycles-pp.allocate_slab
      0.05 ±200%      +0.1        0.10 ± 66%  perf-profile.children.cycles-pp.page_remove_rmap
      0.03 ±200%      +0.1        0.09 ± 65%  perf-profile.children.cycles-pp.j__udyLAllocJLL2
      0.03 ±200%      +0.1        0.09 ± 65%  perf-profile.children.cycles-pp.vm_area_free_rcu_cb
      0.02 ±200%      +0.1        0.08 ± 69%  perf-profile.children.cycles-pp.__libc_write
      0.04 ±200%      +0.1        0.10 ± 66%  perf-profile.children.cycles-pp.native_flush_tlb_one_user
      0.02 ±200%      +0.1        0.08 ± 70%  perf-profile.children.cycles-pp.record__pushfn
      0.02 ±200%      +0.1        0.08 ± 68%  perf-profile.children.cycles-pp.writen
      0.06 ±200%      +0.1        0.12 ± 66%  perf-profile.children.cycles-pp.folio_add_new_anon_rmap
      0.04 ±200%      +0.1        0.10 ± 66%  perf-profile.children.cycles-pp.flush_tlb_func
      0.02 ±200%      +0.1        0.08 ± 71%  perf-profile.children.cycles-pp.perf_mux_hrtimer_handler
      0.02 ±200%      +0.1        0.09 ± 68%  perf-profile.children.cycles-pp.___slab_alloc
      0.03 ±200%      +0.1        0.10 ± 67%  perf-profile.children.cycles-pp.ktime_get_update_offsets_now
      0.04 ±200%      +0.1        0.12 ± 66%  perf-profile.children.cycles-pp.flush_tlb_mm_range
      0.04 ±200%      +0.1        0.13 ± 65%  perf-profile.children.cycles-pp.sync_regs
      0.04 ±200%      +0.1        0.13 ± 65%  perf-profile.children.cycles-pp.rcu_cblist_dequeue
      0.04 ±200%      +0.1        0.13 ± 66%  perf-profile.children.cycles-pp.mas_wr_node_store
      0.03 ±200%      +0.1        0.12 ± 66%  perf-profile.children.cycles-pp.__kmem_cache_alloc_bulk
      0.04 ±200%      +0.1        0.13 ± 65%  perf-profile.children.cycles-pp.kmem_cache_free
      0.04 ±200%      +0.1        0.13 ± 65%  perf-profile.children.cycles-pp.mas_wr_bnode
      0.08 ±200%      +0.1        0.18 ± 65%  perf-profile.children.cycles-pp.task_tick_fair
      0.02 ±200%      +0.1        0.12 ± 69%  perf-profile.children.cycles-pp.perf_mmap__push
      0.03 ±199%      +0.1        0.13 ± 65%  perf-profile.children.cycles-pp.kmem_cache_alloc_bulk
      0.02 ±200%      +0.1        0.12 ± 69%  perf-profile.children.cycles-pp.record__mmap_read_evlist
      0.02 ±200%      +0.1        0.13 ± 69%  perf-profile.children.cycles-pp.cmd_record
      0.10 ±200%      +0.1        0.20 ± 66%  perf-profile.children.cycles-pp.__mod_lruvec_page_state
      0.05 ±200%      +0.1        0.16 ± 65%  perf-profile.children.cycles-pp.__slab_free
      0.05 ±200%      +0.1        0.16 ± 65%  perf-profile.children.cycles-pp.__pte_offset_map_lock
      0.05 ±200%      +0.1        0.16 ± 65%  perf-profile.children.cycles-pp.vm_area_dup
      0.03 ±200%      +0.1        0.14 ± 68%  perf-profile.children.cycles-pp.__libc_start_main
      0.03 ±200%      +0.1        0.14 ± 68%  perf-profile.children.cycles-pp.main
      0.03 ±200%      +0.1        0.14 ± 68%  perf-profile.children.cycles-pp.run_builtin
      0.00            +0.1        0.11 ± 68%  perf-profile.children.cycles-pp.__free_one_page
      0.05 ±200%      +0.1        0.16 ± 65%  perf-profile.children.cycles-pp.mas_wr_store_entry
      0.05 ±200%      +0.1        0.16 ± 65%  perf-profile.children.cycles-pp.free@plt
      0.05 ±200%      +0.1        0.16 ± 65%  perf-profile.children.cycles-pp.JudyFree
      0.05 ±200%      +0.1        0.16 ± 65%  perf-profile.children.cycles-pp.malloc@plt
      0.05 ±200%      +0.1        0.17 ± 65%  perf-profile.children.cycles-pp.JudyMalloc
      0.08 ±200%      +0.1        0.19 ± 67%  perf-profile.children.cycles-pp.zap_pte_range
      0.06 ±200%      +0.1        0.18 ± 65%  perf-profile.children.cycles-pp.mas_store_gfp
      0.06 ±200%      +0.1        0.18 ± 65%  perf-profile.children.cycles-pp.mas_store_prealloc
      0.08 ±200%      +0.1        0.20 ± 66%  perf-profile.children.cycles-pp.native_irq_return_iret
      0.06 ±200%      +0.1        0.18 ± 65%  perf-profile.children.cycles-pp.vma_complete
      0.08 ±200%      +0.1        0.20 ± 67%  perf-profile.children.cycles-pp.zap_pmd_range
      0.05 ±200%      +0.1        0.18 ± 68%  perf-profile.children.cycles-pp.folio_add_lru_vma
      0.06 ±200%      +0.1        0.19 ± 65%  perf-profile.children.cycles-pp.__intel_pmu_enable_all
      0.05 ±200%      +0.1        0.18 ± 68%  perf-profile.children.cycles-pp.folio_batch_move_lru
      0.08 ±200%      +0.1        0.22 ± 67%  perf-profile.children.cycles-pp.unmap_page_range
      0.05 ±200%      +0.1        0.18 ± 65%  perf-profile.children.cycles-pp.mas_alloc_nodes
      0.05 ±200%      +0.1        0.19 ± 65%  perf-profile.children.cycles-pp.mas_preallocate
      0.09 ±200%      +0.1        0.23 ± 67%  perf-profile.children.cycles-pp.unmap_vmas
      0.05 ±200%      +0.1        0.20 ± 65%  perf-profile.children.cycles-pp.__irq_exit_rcu
      0.07 ±200%      +0.2        0.22 ± 65%  perf-profile.children.cycles-pp.kmem_cache_alloc
      0.00            +0.2        0.16 ± 70%  perf-profile.children.cycles-pp.clockevents_program_event
      0.05 ±200%      +0.2        0.21 ± 66%  perf-profile.children.cycles-pp.__list_del_entry_valid_or_report
      0.08 ±200%      +0.2        0.24 ± 65%  perf-profile.children.cycles-pp.JudyFree@plt
      0.06 ±200%      +0.2        0.23 ± 66%  perf-profile.children.cycles-pp.clear_page_erms
      0.08 ±200%      +0.2        0.25 ± 65%  perf-profile.children.cycles-pp.JudyMalloc@plt
      0.10 ±200%      +0.2        0.29 ± 66%  perf-profile.children.cycles-pp.get_mem_cgroup_from_mm
      0.10 ±200%      +0.2        0.31 ± 65%  perf-profile.children.cycles-pp.perf_event_task_tick
      0.10 ±200%      +0.2        0.31 ± 65%  perf-profile.children.cycles-pp.perf_adjust_freq_unthr_context
      0.00            +0.2        0.20 ± 68%  perf-profile.children.cycles-pp.ktime_get
      0.12 ±200%      +0.2        0.33 ± 66%  perf-profile.children.cycles-pp.page_counter_uncharge
      0.13 ±200%      +0.2        0.37 ± 66%  perf-profile.children.cycles-pp.uncharge_batch
      0.14 ±200%      +0.3        0.40 ± 66%  perf-profile.children.cycles-pp.__mem_cgroup_uncharge_list
      0.16 ±200%      +0.3        0.43 ± 66%  perf-profile.children.cycles-pp.__mem_cgroup_charge
      0.00            +0.3        0.28 ± 94%  perf-profile.children.cycles-pp.__pte_alloc
      0.00            +0.3        0.28 ± 94%  perf-profile.children.cycles-pp.pte_alloc_one
      0.12 ±200%      +0.3        0.40 ± 65%  perf-profile.children.cycles-pp.j__udyLFreeJBBJP
      0.23 ±200%      +0.4        0.58 ± 65%  perf-profile.children.cycles-pp.scheduler_tick
      0.16 ±200%      +0.4        0.53 ± 65%  perf-profile.children.cycles-pp.j__udyLAllocJBBJP
      0.16 ±200%      +0.4        0.53 ± 68%  perf-profile.children.cycles-pp.ordered_events__queue
      0.16 ±200%      +0.4        0.53 ± 68%  perf-profile.children.cycles-pp.queue_event
      0.16 ±200%      +0.4        0.53 ± 68%  perf-profile.children.cycles-pp.process_simple
      0.16 ±200%      +0.4        0.53 ± 68%  perf-profile.children.cycles-pp.record__finish_output
      0.16 ±200%      +0.4        0.53 ± 68%  perf-profile.children.cycles-pp.perf_session__process_events
      0.16 ±200%      +0.4        0.53 ± 68%  perf-profile.children.cycles-pp.reader__read_event
      0.17 ±200%      +0.4        0.55 ± 65%  perf-profile.children.cycles-pp.JudyLDel@plt
      0.26 ±200%      +0.4        0.64 ± 65%  perf-profile.children.cycles-pp.update_process_times
      0.26 ±200%      +0.4        0.64 ± 65%  perf-profile.children.cycles-pp.tick_sched_handle
      0.27 ±200%      +0.4        0.70 ± 65%  perf-profile.children.cycles-pp.tick_sched_timer
      0.20 ±200%      +0.4        0.63 ± 65%  perf-profile.children.cycles-pp.cfree
      0.20 ±200%      +0.4        0.65 ± 65%  perf-profile.children.cycles-pp.__split_vma
      0.20 ±200%      +0.5        0.67 ± 65%  perf-profile.children.cycles-pp.JudyLIns@plt
      0.18 ±200%      +0.5        0.66 ± 67%  perf-profile.children.cycles-pp.__cmd_record
      0.29 ±200%      +0.5        0.80 ± 65%  perf-profile.children.cycles-pp.__hrtimer_run_queues
      0.25 ±200%      +0.5        0.77 ± 65%  perf-profile.children.cycles-pp.j__udyLCascade2
      0.24 ±199%      +0.5        0.77 ± 65%  perf-profile.children.cycles-pp.j__udyLCreateBranchU
      0.32 ±200%      +0.7        1.02 ± 65%  perf-profile.children.cycles-pp.JudyLGet@plt
      0.33 ±200%      +0.7        1.07 ± 65%  perf-profile.children.cycles-pp.hrtimer_interrupt
      0.33 ±200%      +0.7        1.08 ± 65%  perf-profile.children.cycles-pp.__sysvec_apic_timer_interrupt
      0.38 ±200%      +0.9        1.25 ± 65%  perf-profile.children.cycles-pp.malloc
      0.39 ±200%      +0.9        1.28 ± 65%  perf-profile.children.cycles-pp.sysvec_apic_timer_interrupt
      0.42 ±200%      +1.0        1.39 ± 65%  perf-profile.children.cycles-pp.asm_sysvec_apic_timer_interrupt
      0.00            +1.3        1.27 ± 82%  perf-profile.children.cycles-pp.__unfreeze_partials
      0.13 ±200%      +1.8        1.89 ± 91%  perf-profile.children.cycles-pp.__folio_alloc
      0.13 ±200%      +1.8        1.91 ± 91%  perf-profile.children.cycles-pp.vma_alloc_folio
      0.00            +1.8        1.84 ±100%  perf-profile.children.cycles-pp.rmqueue_bulk
      0.05 ±200%      +1.9        1.91 ± 98%  perf-profile.children.cycles-pp.rmqueue
      0.84 ±200%      +1.9        2.72 ± 65%  perf-profile.children.cycles-pp.JudyLIns
      0.00            +1.9        1.89 ± 99%  perf-profile.children.cycles-pp.__rmqueue_pcplist
      0.92 ±200%      +2.0        2.88 ± 65%  perf-profile.children.cycles-pp.JudyLDel
      0.12 ±200%      +2.1        2.18 ± 92%  perf-profile.children.cycles-pp.get_page_from_freelist
      0.13 ±200%      +2.1        2.21 ± 91%  perf-profile.children.cycles-pp.__alloc_pages
      0.01 ±300%      +2.1        2.09 ± 81%  perf-profile.children.cycles-pp.free_unref_page
      0.01 ±300%      +2.1        2.09 ± 81%  perf-profile.children.cycles-pp.free_pcppages_bulk
      0.00            +2.1        2.10 ± 81%  perf-profile.children.cycles-pp.free_unref_page_commit
      0.15 ±200%      +2.3        2.40 ± 73%  perf-profile.children.cycles-pp.folio_lruvec_lock_irqsave
      0.12 ±200%      +2.3        2.42 ± 77%  perf-profile.children.cycles-pp.smpboot_thread_fn
      0.12 ±200%      +2.3        2.45 ± 77%  perf-profile.children.cycles-pp.kthread
      0.12 ±200%      +2.3        2.45 ± 77%  perf-profile.children.cycles-pp.ret_from_fork_asm
      0.12 ±200%      +2.3        2.45 ± 77%  perf-profile.children.cycles-pp.ret_from_fork
      0.14 ±199%      +2.4        2.50 ± 76%  perf-profile.children.cycles-pp.rcu_do_batch
      0.16 ±200%      +2.4        2.56 ± 75%  perf-profile.children.cycles-pp.rcu_core
      0.34 ±200%      +2.6        2.92 ± 70%  perf-profile.children.cycles-pp.tlb_batch_pages_flush
      0.34 ±200%      +2.6        2.92 ± 70%  perf-profile.children.cycles-pp.release_pages
      0.46 ±200%      +2.6        3.11 ± 80%  perf-profile.children.cycles-pp.do_anonymous_page
      0.38 ±200%      +2.7        3.04 ± 69%  perf-profile.children.cycles-pp.tlb_finish_mmu
      0.48 ±200%      +2.7        3.16 ± 79%  perf-profile.children.cycles-pp.__handle_mm_fault
      0.50 ±200%      +2.7        3.21 ± 79%  perf-profile.children.cycles-pp.handle_mm_fault
      0.53 ±200%      +2.8        3.31 ± 78%  perf-profile.children.cycles-pp.do_user_addr_fault
      0.54 ±200%      +2.8        3.32 ± 78%  perf-profile.children.cycles-pp.exc_page_fault
      0.58 ±200%      +2.9        3.44 ± 77%  perf-profile.children.cycles-pp.asm_exc_page_fault
      0.49 ±200%      +2.9        3.36 ± 68%  perf-profile.children.cycles-pp.unmap_region
      1.44 ±200%      +3.2        4.66 ± 65%  perf-profile.children.cycles-pp.stress_judy
      0.78 ±200%      +3.5        4.27 ± 67%  perf-profile.children.cycles-pp.do_vmi_align_munmap
      0.81 ±200%      +3.6        4.37 ± 67%  perf-profile.children.cycles-pp.__do_sys_brk
      0.83 ±199%      +3.6        4.42 ± 67%  perf-profile.children.cycles-pp.brk
      0.84 ±200%      +3.6        4.48 ± 67%  perf-profile.children.cycles-pp.do_syscall_64
      0.84 ±200%      +3.6        4.49 ± 67%  perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe
      0.12 ±200%      +5.7        5.86 ± 83%  perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath
      0.16 ±200%      +5.8        6.00 ± 83%  perf-profile.children.cycles-pp._raw_spin_lock_irqsave
     13.47 ±200%     +30.1       43.58 ± 65%  perf-profile.children.cycles-pp.JudyLGet
      0.02 ±200%      -0.0        0.01 ±300%  perf-profile.self.cycles-pp.perf_trace_sched_stat_runtime
      0.01 ±201%      -0.0        0.00        perf-profile.self.cycles-pp.update_cfs_group
      0.01 ±200%      -0.0        0.00        perf-profile.self.cycles-pp.update_curr
      0.01 ±200%      -0.0        0.00        perf-profile.self.cycles-pp.memcg_slab_post_alloc_hook
      0.01 ±200%      -0.0        0.00        perf-profile.self.cycles-pp.check_cpu_stall
      0.01 ±200%      -0.0        0.00        perf-profile.self.cycles-pp.arch_scale_freq_tick
      0.01 ±300%      -0.0        0.00        perf-profile.self.cycles-pp.cgroup_rstat_updated
      0.01 ±200%      +0.0        0.01 ±200%  perf-profile.self.cycles-pp.__get_obj_cgroup_from_memcg
      0.01 ±200%      +0.0        0.02 ±152%  perf-profile.self.cycles-pp.calc_global_load_tick
      0.00            +0.0        0.01 ±299%  perf-profile.self.cycles-pp.rcu_nocb_try_bypass
      0.00            +0.0        0.01 ±300%  perf-profile.self.cycles-pp.zap_pte_range
      0.01 ±200%      +0.0        0.02 ±152%  perf-profile.self.cycles-pp.__handle_mm_fault
      0.01 ±200%      +0.0        0.02 ±122%  perf-profile.self.cycles-pp._compound_head
      0.01 ±201%      +0.0        0.02 ±122%  perf-profile.self.cycles-pp.mtree_range_walk
      0.01 ±201%      +0.0        0.02 ±123%  perf-profile.self.cycles-pp.try_charge_memcg
      0.01 ±200%      +0.0        0.03 ±100%  perf-profile.self.cycles-pp.__mod_node_page_state
      0.02 ±200%      +0.0        0.05 ± 68%  perf-profile.self.cycles-pp.propagate_protected_usage
      0.01 ±200%      +0.0        0.04 ± 66%  perf-profile.self.cycles-pp.mem_cgroup_update_lru_size
      0.01 ±300%      +0.0        0.03 ±101%  perf-profile.self.cycles-pp.tick_sched_do_timer
      0.02 ±200%      +0.0        0.04 ± 66%  perf-profile.self.cycles-pp.hrtimer_active
      0.01 ±200%      +0.0        0.04 ± 66%  perf-profile.self.cycles-pp.___perf_sw_event
      0.02 ±200%      +0.0        0.05 ± 66%  perf-profile.self.cycles-pp.__kmem_cache_alloc_bulk
      0.01 ±200%      +0.0        0.05 ± 66%  perf-profile.self.cycles-pp.task_mm_cid_work
      0.01 ±300%      +0.0        0.04 ± 66%  perf-profile.self.cycles-pp.get_page_from_freelist
      0.04 ±201%      +0.0        0.08 ± 66%  perf-profile.self.cycles-pp.__count_memcg_events
      0.02 ±200%      +0.0        0.05 ± 65%  perf-profile.self.cycles-pp.j__udyLFreeJLL2
      0.02 ±200%      +0.0        0.06 ± 66%  perf-profile.self.cycles-pp.release_pages
      0.02 ±200%      +0.0        0.06 ± 65%  perf-profile.self.cycles-pp.kmem_cache_alloc
      0.02 ±200%      +0.0        0.06 ± 65%  perf-profile.self.cycles-pp.j__udyLCascade3
      0.02 ±200%      +0.0        0.06 ± 66%  perf-profile.self.cycles-pp.j__udyLAllocJLL2
      0.05 ±200%      +0.0        0.10 ± 66%  perf-profile.self.cycles-pp.__mod_memcg_lruvec_state
      0.02 ±200%      +0.0        0.06 ± 66%  perf-profile.self.cycles-pp.mas_wr_node_store
      0.02 ±200%      +0.0        0.07 ± 65%  perf-profile.self.cycles-pp.j__udyLFreeJBBJP@plt
      0.02 ±200%      +0.0        0.07 ± 65%  perf-profile.self.cycles-pp.j__udyLAllocJBBJP@plt
      0.02 ±200%      +0.0        0.07 ± 65%  perf-profile.self.cycles-pp.mod_objcg_state
      0.02 ±200%      +0.0        0.07 ± 65%  perf-profile.self.cycles-pp.j__udyLLeaf2ToLeaf3
      0.02 ±200%      +0.1        0.08 ± 65%  perf-profile.self.cycles-pp.malloc@plt
      0.03 ±200%      +0.1        0.08 ± 65%  perf-profile.self.cycles-pp.free@plt
      0.03 ±200%      +0.1        0.08 ± 65%  perf-profile.self.cycles-pp.JudyFree@plt
      0.02 ±200%      +0.1        0.08 ± 65%  perf-profile.self.cycles-pp.JudyFree
      0.03 ±200%      +0.1        0.08 ± 65%  perf-profile.self.cycles-pp.kmem_cache_free
      0.02 ±200%      +0.1        0.08 ± 65%  perf-profile.self.cycles-pp.JudyMalloc
      0.05 ±200%      +0.1        0.11 ± 66%  perf-profile.self.cycles-pp.__mod_lruvec_page_state
      0.03 ±200%      +0.1        0.08 ± 65%  perf-profile.self.cycles-pp.JudyMalloc@plt
      0.04 ±200%      +0.1        0.10 ± 66%  perf-profile.self.cycles-pp.native_flush_tlb_one_user
      0.03 ±200%      +0.1        0.10 ± 67%  perf-profile.self.cycles-pp.ktime_get_update_offsets_now
      0.00            +0.1        0.07 ± 76%  perf-profile.self.cycles-pp.rmqueue_bulk
      0.05 ±200%      +0.1        0.13 ± 65%  perf-profile.self.cycles-pp.perf_adjust_freq_unthr_context
      0.04 ±200%      +0.1        0.13 ± 65%  perf-profile.self.cycles-pp.sync_regs
      0.04 ±200%      +0.1        0.13 ± 65%  perf-profile.self.cycles-pp.rcu_cblist_dequeue
      0.04 ±200%      +0.1        0.14 ± 65%  perf-profile.self.cycles-pp.__pte_offset_map_lock
      0.00            +0.1        0.09 ± 69%  perf-profile.self.cycles-pp.__free_one_page
      0.04 ±200%      +0.1        0.14 ± 65%  perf-profile.self.cycles-pp._raw_spin_lock_irqsave
      0.05 ±200%      +0.1        0.15 ± 65%  perf-profile.self.cycles-pp.__slab_free
      0.08 ±200%      +0.1        0.20 ± 66%  perf-profile.self.cycles-pp.native_irq_return_iret
      0.06 ±200%      +0.1        0.19 ± 65%  perf-profile.self.cycles-pp.JudyLDel@plt
      0.06 ±200%      +0.1        0.19 ± 65%  perf-profile.self.cycles-pp.__intel_pmu_enable_all
      0.07 ±200%      +0.2        0.23 ± 65%  perf-profile.self.cycles-pp.JudyLIns@plt
      0.05 ±200%      +0.2        0.21 ± 66%  perf-profile.self.cycles-pp.__list_del_entry_valid_or_report
      0.06 ±200%      +0.2        0.23 ± 66%  perf-profile.self.cycles-pp.clear_page_erms
      0.10 ±200%      +0.2        0.28 ± 66%  perf-profile.self.cycles-pp.get_mem_cgroup_from_mm
      0.09 ±200%      +0.2        0.29 ± 66%  perf-profile.self.cycles-pp.page_counter_uncharge
      0.00            +0.2        0.20 ± 68%  perf-profile.self.cycles-pp.ktime_get
      0.10 ±200%      +0.2        0.33 ± 65%  perf-profile.self.cycles-pp.j__udyLFreeJBBJP
      0.12 ±200%      +0.3        0.39 ± 65%  perf-profile.self.cycles-pp.JudyLGet@plt
      0.12 ±200%      +0.3        0.40 ± 65%  perf-profile.self.cycles-pp.j__udyLAllocJBBJP
      0.15 ±200%      +0.4        0.52 ± 68%  perf-profile.self.cycles-pp.queue_event
      0.17 ±200%      +0.4        0.54 ± 65%  perf-profile.self.cycles-pp.cfree
      0.25 ±200%      +0.5        0.76 ± 65%  perf-profile.self.cycles-pp.j__udyLCascade2
      0.23 ±200%      +0.5        0.75 ± 65%  perf-profile.self.cycles-pp.j__udyLCreateBranchU
      0.36 ±200%      +0.8        1.15 ± 65%  perf-profile.self.cycles-pp.malloc
      0.78 ±200%      +1.7        2.44 ± 65%  perf-profile.self.cycles-pp.JudyLDel
      0.76 ±200%      +1.7        2.48 ± 65%  perf-profile.self.cycles-pp.JudyLIns
      1.25 ±200%      +2.8        4.04 ± 65%  perf-profile.self.cycles-pp.stress_judy
      0.12 ±200%      +5.7        5.86 ± 83%  perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath
     13.20 ±200%     +29.5       42.71 ± 65%  perf-profile.self.cycles-pp.JudyLGet
      2886 ±  8%      +3.7%       2993 ± 12%  slabinfo.Acpi-State.active_objs
     56.60 ±  8%      +3.7%      58.70 ± 12%  slabinfo.Acpi-State.active_slabs
      2886 ±  8%      +3.7%       2993 ± 12%  slabinfo.Acpi-State.num_objs
     56.60 ±  8%      +3.7%      58.70 ± 12%  slabinfo.Acpi-State.num_slabs
     36.00            +0.0%      36.00        slabinfo.DCCP.active_objs
      2.00            +0.0%       2.00        slabinfo.DCCP.active_slabs
     36.00            +0.0%      36.00        slabinfo.DCCP.num_objs
      2.00            +0.0%       2.00        slabinfo.DCCP.num_slabs
     34.00            +0.0%      34.00        slabinfo.DCCPv6.active_objs
      2.00            +0.0%       2.00        slabinfo.DCCPv6.active_slabs
     34.00            +0.0%      34.00        slabinfo.DCCPv6.num_objs
      2.00            +0.0%       2.00        slabinfo.DCCPv6.num_slabs
    451.20 ±  2%      -0.7%     448.00        slabinfo.RAW.active_objs
     14.10 ±  2%      -0.7%      14.00        slabinfo.RAW.active_slabs
    451.20 ±  2%      -0.7%     448.00        slabinfo.RAW.num_objs
     14.10 ±  2%      -0.7%      14.00        slabinfo.RAW.num_slabs
    278.20 ±  4%      +0.0%     278.20 ±  4%  slabinfo.RAWv6.active_objs
     10.70 ±  4%      +0.0%      10.70 ±  4%  slabinfo.RAWv6.active_slabs
    278.20 ±  4%      +0.0%     278.20 ±  4%  slabinfo.RAWv6.num_objs
     10.70 ±  4%      +0.0%      10.70 ±  4%  slabinfo.RAWv6.num_slabs
     55.02 ±  7%      +3.5%      56.97        slabinfo.TCP.active_objs
      3.93 ±  7%      +3.5%       4.07        slabinfo.TCP.active_slabs
     55.02 ±  7%      +3.5%      56.97        slabinfo.TCP.num_objs
      3.93 ±  7%      +3.5%       4.07        slabinfo.TCP.num_slabs
     37.70 ± 10%      +3.4%      39.00        slabinfo.TCPv6.active_objs
      2.90 ± 10%      +3.4%       3.00        slabinfo.TCPv6.active_slabs
     37.70 ± 10%      +3.4%      39.00        slabinfo.TCPv6.num_objs
      2.90 ± 10%      +3.4%       3.00        slabinfo.TCPv6.num_slabs
    110.40 ± 10%      +6.5%     117.60 ±  6%  slabinfo.UDPv6.active_objs
      4.60 ± 10%      +6.5%       4.90 ±  6%  slabinfo.UDPv6.active_slabs
    110.40 ± 10%      +6.5%     117.60 ±  6%  slabinfo.UDPv6.num_objs
      4.60 ± 10%      +6.5%       4.90 ±  6%  slabinfo.UDPv6.num_slabs
      2115 ±  8%      -1.2%       2089 ± 10%  slabinfo.UNIX.active_objs
     70.51 ±  8%      -1.2%      69.65 ± 10%  slabinfo.UNIX.active_slabs
      2115 ±  8%      -1.2%       2089 ± 10%  slabinfo.UNIX.num_objs
     70.51 ±  8%      -1.2%      69.65 ± 10%  slabinfo.UNIX.num_slabs
     40729 ±  4%      +2.9%      41929 ±  3%  slabinfo.anon_vma.active_objs
      1047 ±  4%      +2.9%       1077 ±  3%  slabinfo.anon_vma.active_slabs
     40838 ±  4%      +2.9%      42028 ±  3%  slabinfo.anon_vma.num_objs
      1047 ±  4%      +2.9%       1077 ±  3%  slabinfo.anon_vma.num_slabs
     61114 ±  2%      +1.7%      62156 ±  2%  slabinfo.anon_vma_chain.active_objs
    959.75 ±  2%      +1.8%     976.69 ±  2%  slabinfo.anon_vma_chain.active_slabs
     61424 ±  2%      +1.8%      62508 ±  2%  slabinfo.anon_vma_chain.num_objs
    959.75 ±  2%      +1.8%     976.69 ±  2%  slabinfo.anon_vma_chain.num_slabs
     76.00 ± 15%     +10.5%      84.00 ± 17%  slabinfo.bdev_cache.active_objs
      3.80 ± 15%     +10.5%       4.20 ± 17%  slabinfo.bdev_cache.active_slabs
     76.00 ± 15%     +10.5%      84.00 ± 17%  slabinfo.bdev_cache.num_objs
      3.80 ± 15%     +10.5%       4.20 ± 17%  slabinfo.bdev_cache.num_slabs
    454.40 ±  4%      +1.4%     460.80 ±  5%  slabinfo.bio-120.active_objs
      7.10 ±  4%      +1.4%       7.20 ±  5%  slabinfo.bio-120.active_slabs
    454.40 ±  4%      +1.4%     460.80 ±  5%  slabinfo.bio-120.num_objs
      7.10 ±  4%      +1.4%       7.20 ±  5%  slabinfo.bio-120.num_slabs
    537.60 ± 22%      +3.1%     554.40 ±  8%  slabinfo.bio-184.active_objs
     12.80 ± 22%      +3.1%      13.20 ±  8%  slabinfo.bio-184.active_slabs
    537.60 ± 22%      +3.1%     554.40 ±  8%  slabinfo.bio-184.num_objs
     12.80 ± 22%      +3.1%      13.20 ±  8%  slabinfo.bio-184.num_slabs
    128.00            +0.0%     128.00        slabinfo.bio-248.active_objs
      2.00            +0.0%       2.00        slabinfo.bio-248.active_slabs
    128.00            +0.0%     128.00        slabinfo.bio-248.num_objs
      2.00            +0.0%       2.00        slabinfo.bio-248.num_slabs
     51.00            +0.0%      51.00        slabinfo.bio-296.active_objs
      1.00            +0.0%       1.00        slabinfo.bio-296.active_slabs
     51.00            +0.0%      51.00        slabinfo.bio-296.num_objs
      1.00            +0.0%       1.00        slabinfo.bio-296.num_slabs
    121.80 ± 10%      +3.4%     126.00        slabinfo.bio-360.active_objs
      2.90 ± 10%      +3.4%       3.00        slabinfo.bio-360.active_slabs
    121.80 ± 10%      +3.4%     126.00        slabinfo.bio-360.num_objs
      2.90 ± 10%      +3.4%       3.00        slabinfo.bio-360.num_slabs
     42.00            +0.0%      42.00        slabinfo.bio-376.active_objs
      1.00            +0.0%       1.00        slabinfo.bio-376.active_slabs
     42.00            +0.0%      42.00        slabinfo.bio-376.num_objs
      1.00            +0.0%       1.00        slabinfo.bio-376.num_slabs
     36.00            +0.0%      36.00        slabinfo.bio-432.active_objs
      1.00            +0.0%       1.00        slabinfo.bio-432.active_slabs
     36.00            +0.0%      36.00        slabinfo.bio-432.num_objs
      1.00            +0.0%       1.00        slabinfo.bio-432.num_slabs
    170.00            +0.0%     170.00        slabinfo.bio_post_read_ctx.active_objs
      2.00            +0.0%       2.00        slabinfo.bio_post_read_ctx.active_slabs
    170.00            +0.0%     170.00        slabinfo.bio_post_read_ctx.num_objs
      2.00            +0.0%       2.00        slabinfo.bio_post_read_ctx.num_slabs
     33.60 ± 14%      +9.5%      36.80 ± 19%  slabinfo.biovec-128.active_objs
      2.10 ± 14%      +9.5%       2.30 ± 19%  slabinfo.biovec-128.active_slabs
     33.60 ± 14%      +9.5%      36.80 ± 19%  slabinfo.biovec-128.num_objs
      2.10 ± 14%      +9.5%       2.30 ± 19%  slabinfo.biovec-128.num_slabs
    300.80 ±  9%      +7.4%     323.20 ± 12%  slabinfo.biovec-64.active_objs
      9.40 ±  9%      +7.4%      10.10 ± 12%  slabinfo.biovec-64.active_slabs
    300.80 ±  9%      +7.4%     323.20 ± 12%  slabinfo.biovec-64.num_objs
      9.40 ±  9%      +7.4%      10.10 ± 12%  slabinfo.biovec-64.num_slabs
     56.00            +0.0%      56.00        slabinfo.biovec-max.active_objs
      7.00            +0.0%       7.00        slabinfo.biovec-max.active_slabs
     56.00            +0.0%      56.00        slabinfo.biovec-max.num_objs
      7.00            +0.0%       7.00        slabinfo.biovec-max.num_slabs
    136.00            +0.0%     136.00        slabinfo.btrfs_extent_buffer.active_objs
      2.00            +0.0%       2.00        slabinfo.btrfs_extent_buffer.active_slabs
    136.00            +0.0%     136.00        slabinfo.btrfs_extent_buffer.num_objs
      2.00            +0.0%       2.00        slabinfo.btrfs_extent_buffer.num_slabs
      0.00       +7.8e+102%       7.80 ±200%  slabinfo.btrfs_free_space.active_objs
      0.00         +2e+101%       0.20 ±200%  slabinfo.btrfs_free_space.active_slabs
      0.00       +7.8e+102%       7.80 ±200%  slabinfo.btrfs_free_space.num_objs
      0.00         +2e+101%       0.20 ±200%  slabinfo.btrfs_free_space.num_slabs
    104.40 ± 13%     -11.1%      92.80 ± 12%  slabinfo.btrfs_inode.active_objs
      3.60 ± 13%     -11.1%       3.20 ± 12%  slabinfo.btrfs_inode.active_slabs
    104.40 ± 13%     -11.1%      92.80 ± 12%  slabinfo.btrfs_inode.num_objs
      3.60 ± 13%     -11.1%       3.20 ± 12%  slabinfo.btrfs_inode.num_slabs
    211.85 ±  5%      +1.7%     215.47        slabinfo.btrfs_path.active_objs
      5.88 ±  5%      +1.7%       5.99        slabinfo.btrfs_path.active_slabs
    211.85 ±  5%      +1.7%     215.47        slabinfo.btrfs_path.num_objs
      5.88 ±  5%      +1.7%       5.99        slabinfo.btrfs_path.num_slabs
    156.00           +15.0%     179.40 ± 10%  slabinfo.buffer_head.active_objs
      4.00           +15.0%       4.60 ± 10%  slabinfo.buffer_head.active_slabs
    156.00           +15.0%     179.40 ± 10%  slabinfo.buffer_head.num_objs
      4.00           +15.0%       4.60 ± 10%  slabinfo.buffer_head.num_slabs
     14481            +0.1%      14494        slabinfo.cred_jar.active_objs
    344.80            +0.1%     345.10        slabinfo.cred_jar.active_slabs
     14481            +0.1%      14494        slabinfo.cred_jar.num_objs
    344.80            +0.1%     345.10        slabinfo.cred_jar.num_slabs
     39.00            +0.0%      39.00        slabinfo.dax_cache.active_objs
      1.00            +0.0%       1.00        slabinfo.dax_cache.active_slabs
     39.00            +0.0%      39.00        slabinfo.dax_cache.num_objs
      1.00            +0.0%       1.00        slabinfo.dax_cache.num_slabs
    165719            +0.4%     166422        slabinfo.dentry.active_objs
      3967            +0.6%       3990        slabinfo.dentry.active_slabs
    166647            +0.6%     167586        slabinfo.dentry.num_objs
      3967            +0.6%       3990        slabinfo.dentry.num_slabs
     30.00            +0.0%      30.00        slabinfo.dmaengine-unmap-128.active_objs
      1.00            +0.0%       1.00        slabinfo.dmaengine-unmap-128.active_slabs
     30.00            +0.0%      30.00        slabinfo.dmaengine-unmap-128.num_objs
      1.00            +0.0%       1.00        slabinfo.dmaengine-unmap-128.num_slabs
     64.00            +0.0%      64.00        slabinfo.dmaengine-unmap-2.active_objs
      1.00            +0.0%       1.00        slabinfo.dmaengine-unmap-2.active_slabs
     64.00            +0.0%      64.00        slabinfo.dmaengine-unmap-2.num_objs
      1.00            +0.0%       1.00        slabinfo.dmaengine-unmap-2.num_slabs
     15.00            +0.0%      15.00        slabinfo.dmaengine-unmap-256.active_objs
      1.00            +0.0%       1.00        slabinfo.dmaengine-unmap-256.active_slabs
     15.00            +0.0%      15.00        slabinfo.dmaengine-unmap-256.num_objs
      1.00            +0.0%       1.00        slabinfo.dmaengine-unmap-256.num_slabs
     24640 ± 10%      -1.5%      24268 ±  6%  slabinfo.ep_head.active_objs
     96.25 ± 10%      -1.5%      94.80 ±  6%  slabinfo.ep_head.active_slabs
     24640 ± 10%      -1.5%      24268 ±  6%  slabinfo.ep_head.num_objs
     96.25 ± 10%      -1.5%      94.80 ±  6%  slabinfo.ep_head.num_slabs
    706.56 ±  7%      +1.7%     718.75 ±  9%  slabinfo.file_lock_cache.active_objs
     19.10 ±  7%      +1.7%      19.43 ±  9%  slabinfo.file_lock_cache.active_slabs
    706.56 ±  7%      +1.7%     718.75 ±  9%  slabinfo.file_lock_cache.num_objs
     19.10 ±  7%      +1.7%      19.43 ±  9%  slabinfo.file_lock_cache.num_slabs
      9531            +1.5%       9673 ±  2%  slabinfo.files_cache.active_objs
    207.21            +1.5%     210.29 ±  2%  slabinfo.files_cache.active_slabs
      9531            +1.5%       9673 ±  2%  slabinfo.files_cache.num_objs
    207.21            +1.5%     210.29 ±  2%  slabinfo.files_cache.num_slabs
     40884            +0.5%      41085        slabinfo.filp.active_objs
    672.40            +0.2%     674.07 ±  3%  slabinfo.filp.active_slabs
     43033            +0.2%      43140 ±  3%  slabinfo.filp.num_objs
    672.40            +0.2%     674.07 ±  3%  slabinfo.filp.num_slabs
      2316 ±  7%      +2.8%       2380 ±  8%  slabinfo.fsnotify_mark_connector.active_objs
     18.10 ±  7%      +2.8%      18.60 ±  8%  slabinfo.fsnotify_mark_connector.active_slabs
      2316 ±  7%      +2.8%       2380 ±  8%  slabinfo.fsnotify_mark_connector.num_objs
     18.10 ±  7%      +2.8%      18.60 ±  8%  slabinfo.fsnotify_mark_connector.num_slabs
      9300 ±  3%      +1.2%       9409 ±  5%  slabinfo.ftrace_event_field.active_objs
    127.40 ±  3%      +1.2%     128.90 ±  5%  slabinfo.ftrace_event_field.active_slabs
      9300 ±  3%      +1.2%       9409 ±  5%  slabinfo.ftrace_event_field.num_objs
    127.40 ±  3%      +1.2%     128.90 ±  5%  slabinfo.ftrace_event_field.num_slabs
     56.00            +0.0%      56.00        slabinfo.fuse_request.active_objs
      1.00            +0.0%       1.00        slabinfo.fuse_request.active_slabs
     56.00            +0.0%      56.00        slabinfo.fuse_request.num_objs
      1.00            +0.0%       1.00        slabinfo.fuse_request.num_slabs
     98.00            +0.0%      98.00        slabinfo.hugetlbfs_inode_cache.active_objs
      2.00            +0.0%       2.00        slabinfo.hugetlbfs_inode_cache.active_slabs
     98.00            +0.0%      98.00        slabinfo.hugetlbfs_inode_cache.num_objs
      2.00            +0.0%       2.00        slabinfo.hugetlbfs_inode_cache.num_slabs
    105824            +0.1%     105909        slabinfo.inode_cache.active_objs
      2077            +0.1%       2079        slabinfo.inode_cache.active_slabs
    105950            +0.1%     106056        slabinfo.inode_cache.num_objs
      2077            +0.1%       2079        slabinfo.inode_cache.num_slabs
    182.50 ± 20%     +16.0%     211.70 ± 10%  slabinfo.ip_fib_alias.active_objs
      2.50 ± 20%     +16.0%       2.90 ± 10%  slabinfo.ip_fib_alias.active_slabs
    182.50 ± 20%     +16.0%     211.70 ± 10%  slabinfo.ip_fib_alias.num_objs
      2.50 ± 20%     +16.0%       2.90 ± 10%  slabinfo.ip_fib_alias.num_slabs
    212.50 ± 20%     +16.0%     246.50 ± 10%  slabinfo.ip_fib_trie.active_objs
      2.50 ± 20%     +16.0%       2.90 ± 10%  slabinfo.ip_fib_trie.active_slabs
    212.50 ± 20%     +16.0%     246.50 ± 10%  slabinfo.ip_fib_trie.num_objs
      2.50 ± 20%     +16.0%       2.90 ± 10%  slabinfo.ip_fib_trie.num_slabs
    168800            +0.0%     168832        slabinfo.kernfs_node_cache.active_objs
      2637            +0.0%       2638        slabinfo.kernfs_node_cache.active_slabs
    168800            +0.0%     168832        slabinfo.kernfs_node_cache.num_objs
      2637            +0.0%       2638        slabinfo.kernfs_node_cache.num_slabs
     21600            -0.1%      21570        slabinfo.khugepaged_mm_slot.active_objs
    211.77            -0.1%     211.48        slabinfo.khugepaged_mm_slot.active_slabs
     21600            -0.1%      21570        slabinfo.khugepaged_mm_slot.num_objs
    211.77            -0.1%     211.48        slabinfo.khugepaged_mm_slot.num_slabs
     20061            +0.5%      20170        slabinfo.kmalloc-128.active_objs
    320.03            +0.0%     320.16        slabinfo.kmalloc-128.active_slabs
     20482            +0.0%      20490        slabinfo.kmalloc-128.num_objs
    320.03            +0.0%     320.16        slabinfo.kmalloc-128.num_slabs
     83404            -0.0%      83399        slabinfo.kmalloc-16.active_objs
    325.80            +0.1%     326.10        slabinfo.kmalloc-16.active_slabs
     83404            +0.1%      83481        slabinfo.kmalloc-16.num_objs
    325.80            +0.1%     326.10        slabinfo.kmalloc-16.num_slabs
     21409            +0.1%      21421        slabinfo.kmalloc-192.active_objs
    509.82            +0.1%     510.14        slabinfo.kmalloc-192.active_slabs
     21412            +0.1%      21425        slabinfo.kmalloc-192.num_objs
    509.82            +0.1%     510.14        slabinfo.kmalloc-192.num_slabs
     10382            -0.2%      10360        slabinfo.kmalloc-1k.active_objs
    325.71            -0.1%     325.33        slabinfo.kmalloc-1k.active_slabs
     10422            -0.1%      10410        slabinfo.kmalloc-1k.num_objs
    325.71            -0.1%     325.33        slabinfo.kmalloc-1k.num_slabs
     20195            +0.2%      20233        slabinfo.kmalloc-256.active_objs
    323.34 ±  2%      +0.2%     323.86        slabinfo.kmalloc-256.active_slabs
     20693 ±  2%      +0.2%      20727        slabinfo.kmalloc-256.num_objs
    323.34 ±  2%      +0.2%     323.86        slabinfo.kmalloc-256.num_slabs
      9041            -0.3%       9015        slabinfo.kmalloc-2k.active_objs
    566.68            -0.2%     565.41        slabinfo.kmalloc-2k.active_slabs
      9066            -0.2%       9046        slabinfo.kmalloc-2k.num_objs
    566.68            -0.2%     565.41        slabinfo.kmalloc-2k.num_slabs
    175334            -0.1%     175183        slabinfo.kmalloc-32.active_objs
      1370            -0.1%       1368        slabinfo.kmalloc-32.active_slabs
    175365            -0.1%     175221        slabinfo.kmalloc-32.num_objs
      1370            -0.1%       1368        slabinfo.kmalloc-32.num_slabs
      3202            -0.6%       3184        slabinfo.kmalloc-4k.active_objs
    400.65            -0.5%     398.60        slabinfo.kmalloc-4k.active_slabs
      3205            -0.5%       3188        slabinfo.kmalloc-4k.num_objs
    400.65            -0.5%     398.60        slabinfo.kmalloc-4k.num_slabs
     33467            +0.5%      33643        slabinfo.kmalloc-512.active_objs
    524.91            +0.4%     527.19        slabinfo.kmalloc-512.active_slabs
     33594            +0.4%      33740        slabinfo.kmalloc-512.num_objs
    524.91            +0.4%     527.19        slabinfo.kmalloc-512.num_slabs
    172450            -0.1%     172322        slabinfo.kmalloc-64.active_objs
      2694            -0.1%       2692        slabinfo.kmalloc-64.active_slabs
    172473            -0.1%     172339        slabinfo.kmalloc-64.num_objs
      2694            -0.1%       2692        slabinfo.kmalloc-64.num_slabs
    138463            +0.5%     139132        slabinfo.kmalloc-8.active_objs
    280.42 ±  2%      -0.1%     280.24        slabinfo.kmalloc-8.active_slabs
    143573 ±  2%      -0.1%     143480        slabinfo.kmalloc-8.num_objs
    280.42 ±  2%      -0.1%     280.24        slabinfo.kmalloc-8.num_slabs
      1429            -0.1%       1427        slabinfo.kmalloc-8k.active_objs
    357.53            -0.1%     357.17        slabinfo.kmalloc-8k.active_slabs
      1430            -0.1%       1428        slabinfo.kmalloc-8k.num_objs
    357.53            -0.1%     357.17        slabinfo.kmalloc-8k.num_slabs
     38336            +1.3%      38843 ±  2%  slabinfo.kmalloc-96.active_objs
    932.92            +1.5%     946.90 ±  2%  slabinfo.kmalloc-96.active_slabs
     39182            +1.5%      39769 ±  2%  slabinfo.kmalloc-96.num_objs
    932.92            +1.5%     946.90 ±  2%  slabinfo.kmalloc-96.num_slabs
      1542 ±  2%      -0.4%       1536 ±  5%  slabinfo.kmalloc-cg-128.active_objs
     24.10 ±  2%      -0.4%      24.00 ±  5%  slabinfo.kmalloc-cg-128.active_slabs
      1542 ±  2%      -0.4%       1536 ±  5%  slabinfo.kmalloc-cg-128.num_objs
     24.10 ±  2%      -0.4%      24.00 ±  5%  slabinfo.kmalloc-cg-128.num_slabs
      4325 ±  2%      -1.7%       4252 ±  4%  slabinfo.kmalloc-cg-16.active_objs
     16.90 ±  2%      -1.7%      16.61 ±  4%  slabinfo.kmalloc-cg-16.active_slabs
      4325 ±  2%      -1.7%       4252 ±  4%  slabinfo.kmalloc-cg-16.num_objs
     16.90 ±  2%      -1.7%      16.61 ±  4%  slabinfo.kmalloc-cg-16.num_slabs
      8798 ±  2%      +2.3%       9004        slabinfo.kmalloc-cg-192.active_objs
    209.49 ±  2%      +2.3%     214.40        slabinfo.kmalloc-cg-192.active_slabs
      8798 ±  2%      +2.3%       9004        slabinfo.kmalloc-cg-192.num_objs
    209.49 ±  2%      +2.3%     214.40        slabinfo.kmalloc-cg-192.num_slabs
      6641 ±  3%      +2.4%       6802 ±  3%  slabinfo.kmalloc-cg-1k.active_objs
    207.56 ±  3%      +2.4%     212.58 ±  3%  slabinfo.kmalloc-cg-1k.active_slabs
      6641 ±  3%      +2.4%       6802 ±  3%  slabinfo.kmalloc-cg-1k.num_objs
    207.56 ±  3%      +2.4%     212.58 ±  3%  slabinfo.kmalloc-cg-1k.num_slabs
      1472 ±  4%      -6.1%       1382 ±  5%  slabinfo.kmalloc-cg-256.active_objs
     23.00 ±  4%      -6.1%      21.60 ±  5%  slabinfo.kmalloc-cg-256.active_slabs
      1472 ±  4%      -6.1%       1382 ±  5%  slabinfo.kmalloc-cg-256.num_objs
     23.00 ±  4%      -6.1%      21.60 ±  5%  slabinfo.kmalloc-cg-256.num_slabs
      1018 ±  3%      +2.2%       1040 ±  4%  slabinfo.kmalloc-cg-2k.active_objs
     63.66 ±  3%      +2.2%      65.06 ±  4%  slabinfo.kmalloc-cg-2k.active_slabs
      1018 ±  3%      +2.2%       1040 ±  4%  slabinfo.kmalloc-cg-2k.num_objs
     63.66 ±  3%      +2.2%      65.06 ±  4%  slabinfo.kmalloc-cg-2k.num_slabs
     29379            +0.1%      29422        slabinfo.kmalloc-cg-32.active_objs
    229.52            +0.1%     229.87        slabinfo.kmalloc-cg-32.active_slabs
     29379            +0.1%      29422        slabinfo.kmalloc-cg-32.num_objs
    229.52            +0.1%     229.87        slabinfo.kmalloc-cg-32.num_slabs
      2697            -0.5%       2683        slabinfo.kmalloc-cg-4k.active_objs
    340.10            -0.6%     338.10        slabinfo.kmalloc-cg-4k.active_slabs
      2720            -0.6%       2704        slabinfo.kmalloc-cg-4k.num_objs
    340.10            -0.6%     338.10        slabinfo.kmalloc-cg-4k.num_slabs
     14431            -0.1%      14419        slabinfo.kmalloc-cg-512.active_objs
    225.50            -0.1%     225.30        slabinfo.kmalloc-cg-512.active_slabs
     14431            -0.1%      14419        slabinfo.kmalloc-cg-512.num_objs
    225.50            -0.1%     225.30        slabinfo.kmalloc-cg-512.num_slabs
      3134 ±  4%      -0.2%       3127 ±  7%  slabinfo.kmalloc-cg-64.active_objs
     48.97 ±  4%      -0.2%      48.87 ±  7%  slabinfo.kmalloc-cg-64.active_slabs
      3134 ±  4%      -0.2%       3127 ±  7%  slabinfo.kmalloc-cg-64.num_objs
     48.97 ±  4%      -0.2%      48.87 ±  7%  slabinfo.kmalloc-cg-64.num_slabs
    115595            +0.1%     115732        slabinfo.kmalloc-cg-8.active_objs
    225.77            +0.1%     226.04        slabinfo.kmalloc-cg-8.active_slabs
    115595            +0.1%     115732        slabinfo.kmalloc-cg-8.num_objs
    225.77            +0.1%     226.04        slabinfo.kmalloc-cg-8.num_slabs
     50.61 ±  3%      +0.3%      50.74 ±  3%  slabinfo.kmalloc-cg-8k.active_objs
     12.65 ±  3%      +0.3%      12.68 ±  3%  slabinfo.kmalloc-cg-8k.active_slabs
     50.61 ±  3%      +0.3%      50.74 ±  3%  slabinfo.kmalloc-cg-8k.num_objs
     12.65 ±  3%      +0.3%      12.68 ±  3%  slabinfo.kmalloc-cg-8k.num_slabs
      1931 ±  4%      +2.8%       1986 ±  7%  slabinfo.kmalloc-cg-96.active_objs
     46.00 ±  4%      +2.8%      47.30 ±  7%  slabinfo.kmalloc-cg-96.active_slabs
      1931 ±  4%      +2.8%       1986 ±  7%  slabinfo.kmalloc-cg-96.num_objs
     46.00 ±  4%      +2.8%      47.30 ±  7%  slabinfo.kmalloc-cg-96.num_slabs
    300.80 ± 13%      +2.1%     307.20 ±  8%  slabinfo.kmalloc-rcl-128.active_objs
      4.70 ± 13%      +2.1%       4.80 ±  8%  slabinfo.kmalloc-rcl-128.active_slabs
    300.80 ± 13%      +2.1%     307.20 ±  8%  slabinfo.kmalloc-rcl-128.num_objs
      4.70 ± 13%      +2.1%       4.80 ±  8%  slabinfo.kmalloc-rcl-128.num_slabs
     42.00            +0.0%      42.00        slabinfo.kmalloc-rcl-192.active_objs
      1.00            +0.0%       1.00        slabinfo.kmalloc-rcl-192.active_slabs
     42.00            +0.0%      42.00        slabinfo.kmalloc-rcl-192.num_objs
      1.00            +0.0%       1.00        slabinfo.kmalloc-rcl-192.num_slabs
      8078 ±  3%      -0.7%       8020 ±  4%  slabinfo.kmalloc-rcl-64.active_objs
    126.36 ±  3%      -0.7%     125.46 ±  3%  slabinfo.kmalloc-rcl-64.active_slabs
      8087 ±  3%      -0.7%       8029 ±  3%  slabinfo.kmalloc-rcl-64.num_objs
    126.36 ±  3%      -0.7%     125.46 ±  3%  slabinfo.kmalloc-rcl-64.num_slabs
      1247 ±  7%      +0.0%       1247 ±  7%  slabinfo.kmalloc-rcl-96.active_objs
     29.70 ±  7%      +0.0%      29.70 ±  7%  slabinfo.kmalloc-rcl-96.active_slabs
      1247 ±  7%      +0.0%       1247 ±  7%  slabinfo.kmalloc-rcl-96.num_objs
     29.70 ±  7%      +0.0%      29.70 ±  7%  slabinfo.kmalloc-rcl-96.num_slabs
      1113 ±  6%      -9.2%       1011 ± 10%  slabinfo.kmem_cache.active_objs
     17.40 ±  6%      -9.2%      15.80 ± 10%  slabinfo.kmem_cache.active_slabs
      1113 ±  6%      -9.2%       1011 ± 10%  slabinfo.kmem_cache.num_objs
     17.40 ±  6%      -9.2%      15.80 ± 10%  slabinfo.kmem_cache.num_slabs
      1297 ±  5%      -7.4%       1201 ±  8%  slabinfo.kmem_cache_node.active_objs
     20.30 ±  5%      -7.4%      18.80 ±  8%  slabinfo.kmem_cache_node.active_slabs
      1299 ±  5%      -7.4%       1203 ±  8%  slabinfo.kmem_cache_node.num_objs
     20.30 ±  5%      -7.4%      18.80 ±  8%  slabinfo.kmem_cache_node.num_slabs
     44655            -0.2%      44587        slabinfo.lsm_file_cache.active_objs
    263.38            -0.1%     263.16        slabinfo.lsm_file_cache.active_slabs
     44775            -0.1%      44737        slabinfo.lsm_file_cache.num_objs
    263.38            -0.1%     263.16        slabinfo.lsm_file_cache.num_slabs
     48621 ±  2%      +7.3%      52173 ±  4%  slabinfo.maple_node.active_objs
    772.36 ±  2%      +8.3%     836.19 ±  3%  slabinfo.maple_node.active_slabs
     49431 ±  2%      +8.3%      53516 ±  3%  slabinfo.maple_node.num_objs
    772.36 ±  2%      +8.3%     836.19 ±  3%  slabinfo.maple_node.num_slabs
      5734            +1.2%       5801        slabinfo.mm_struct.active_objs
    238.95            +1.2%     241.73        slabinfo.mm_struct.active_slabs
      5734            +1.2%       5801        slabinfo.mm_struct.num_objs
    238.95            +1.2%     241.73        slabinfo.mm_struct.num_slabs
      1116 ±  6%      -0.5%       1111 ±  5%  slabinfo.mnt_cache.active_objs
     21.90 ±  6%      -0.5%      21.80 ±  5%  slabinfo.mnt_cache.active_slabs
      1116 ±  6%      -0.5%       1111 ±  5%  slabinfo.mnt_cache.num_objs
     21.90 ±  6%      -0.5%      21.80 ±  5%  slabinfo.mnt_cache.num_slabs
     34.00            +0.0%      34.00        slabinfo.mqueue_inode_cache.active_objs
      1.00            +0.0%       1.00        slabinfo.mqueue_inode_cache.active_slabs
     34.00            +0.0%      34.00        slabinfo.mqueue_inode_cache.num_objs
      1.00            +0.0%       1.00        slabinfo.mqueue_inode_cache.num_slabs
      1792            +0.2%       1795        slabinfo.names_cache.active_objs
    224.00            +0.2%     224.40        slabinfo.names_cache.active_slabs
      1792            +0.2%       1795        slabinfo.names_cache.num_objs
    224.00            +0.2%     224.40        slabinfo.names_cache.num_slabs
      7.00            +0.0%       7.00        slabinfo.net_namespace.active_objs
      1.00            +0.0%       1.00        slabinfo.net_namespace.active_slabs
      7.00            +0.0%       7.00        slabinfo.net_namespace.num_objs
      1.00            +0.0%       1.00        slabinfo.net_namespace.num_slabs
     46.00            +0.0%      46.00        slabinfo.nfs_commit_data.active_objs
      1.00            +0.0%       1.00        slabinfo.nfs_commit_data.active_slabs
     46.00            +0.0%      46.00        slabinfo.nfs_commit_data.num_objs
      1.00            +0.0%       1.00        slabinfo.nfs_commit_data.num_slabs
     36.00            +0.0%      36.00        slabinfo.nfs_read_data.active_objs
      1.00            +0.0%       1.00        slabinfo.nfs_read_data.active_slabs
     36.00            +0.0%      36.00        slabinfo.nfs_read_data.num_objs
      1.00            +0.0%       1.00        slabinfo.nfs_read_data.num_slabs
    292.96 ±  8%      -0.9%     290.20 ±  8%  slabinfo.nsproxy.active_objs
      5.23 ±  8%      -0.9%       5.18 ±  8%  slabinfo.nsproxy.active_slabs
    292.96 ±  8%      -0.9%     290.20 ±  8%  slabinfo.nsproxy.num_objs
      5.23 ±  8%      -0.9%       5.18 ±  8%  slabinfo.nsproxy.num_slabs
    240.00            +0.0%     240.00        slabinfo.numa_policy.active_objs
      4.00            +0.0%       4.00        slabinfo.numa_policy.active_slabs
    240.00            +0.0%     240.00        slabinfo.numa_policy.num_objs
      4.00            +0.0%       4.00        slabinfo.numa_policy.num_slabs
     10075            -0.5%      10024        slabinfo.perf_event.active_objs
    404.18            -0.4%     402.52        slabinfo.perf_event.active_slabs
     10104            -0.4%      10062        slabinfo.perf_event.num_objs
    404.18            -0.4%     402.52        slabinfo.perf_event.num_slabs
     16201            -0.7%      16089        slabinfo.pid.active_objs
    253.20            -0.7%     251.40        slabinfo.pid.active_slabs
     16204            -0.7%      16089        slabinfo.pid.num_objs
    253.20            -0.7%     251.40        slabinfo.pid.num_slabs
     13300            -1.0%      13163        slabinfo.pool_workqueue.active_objs
    208.12            -1.1%     205.93        slabinfo.pool_workqueue.active_slabs
     13319            -1.1%      13179        slabinfo.pool_workqueue.num_objs
    208.12            -1.1%     205.93        slabinfo.pool_workqueue.num_slabs
      4792            -1.1%       4741        slabinfo.proc_dir_entry.active_objs
    114.10            -1.1%     112.90        slabinfo.proc_dir_entry.active_slabs
      4792            -1.1%       4741        slabinfo.proc_dir_entry.num_objs
    114.10            -1.1%     112.90        slabinfo.proc_dir_entry.num_slabs
     29409            +0.4%      29525        slabinfo.proc_inode_cache.active_objs
    639.60            +0.4%     642.26        slabinfo.proc_inode_cache.active_slabs
     29421            +0.4%      29544        slabinfo.proc_inode_cache.num_objs
    639.60            +0.4%     642.26        slabinfo.proc_inode_cache.num_slabs
     40983            +0.1%      41043        slabinfo.radix_tree_node.active_objs
    732.66            +0.2%     733.78        slabinfo.radix_tree_node.active_slabs
     41029            +0.2%      41091        slabinfo.radix_tree_node.num_objs
    732.66            +0.2%     733.78        slabinfo.radix_tree_node.num_slabs
    294.00 ± 18%      +8.3%     318.50 ± 18%  slabinfo.request_queue.active_objs
     10.20 ±  9%      +4.9%      10.70 ± 16%  slabinfo.request_queue.active_slabs
    357.00 ±  9%      +4.9%     374.50 ± 16%  slabinfo.request_queue.num_objs
     10.20 ±  9%      +4.9%      10.70 ± 16%  slabinfo.request_queue.num_slabs
     46.00            +0.0%      46.00        slabinfo.rpc_inode_cache.active_objs
      1.00            +0.0%       1.00        slabinfo.rpc_inode_cache.active_slabs
     46.00            +0.0%      46.00        slabinfo.rpc_inode_cache.num_objs
      1.00            +0.0%       1.00        slabinfo.rpc_inode_cache.num_slabs
      4812            +0.3%       4825        slabinfo.scsi_sense_cache.active_objs
     76.20            +0.3%      76.40        slabinfo.scsi_sense_cache.active_slabs
      4876            +0.3%       4889        slabinfo.scsi_sense_cache.num_objs
     76.20            +0.3%      76.40        slabinfo.scsi_sense_cache.num_slabs
     15979            -0.1%      15967        slabinfo.seq_file.active_objs
    234.99            -0.1%     234.81        slabinfo.seq_file.active_slabs
     15979            -0.1%      15967        slabinfo.seq_file.num_objs
    234.99            -0.1%     234.81        slabinfo.seq_file.num_slabs
    101821            +0.2%     101974        slabinfo.shared_policy_node.active_objs
      1197            +0.2%       1199        slabinfo.shared_policy_node.active_slabs
    101821            +0.2%     101974        slabinfo.shared_policy_node.num_objs
      1197            +0.2%       1199        slabinfo.shared_policy_node.num_slabs
      8779            +0.4%       8816 ±  3%  slabinfo.shmem_inode_cache.active_objs
    204.19            +0.4%     205.04 ±  3%  slabinfo.shmem_inode_cache.active_slabs
      8779            +0.4%       8816 ±  3%  slabinfo.shmem_inode_cache.num_objs
    204.19            +0.4%     205.04 ±  3%  slabinfo.shmem_inode_cache.num_slabs
      5473            +0.8%       5519        slabinfo.sighand_cache.active_objs
    365.02            +0.9%     368.23        slabinfo.sighand_cache.active_slabs
      5475            +0.9%       5523        slabinfo.sighand_cache.num_objs
    365.02            +0.9%     368.23        slabinfo.sighand_cache.num_slabs
      8990            +0.1%       8995 ±  2%  slabinfo.signal_cache.active_objs
    321.28            +0.1%     321.61 ±  2%  slabinfo.signal_cache.active_slabs
      8995            +0.1%       9005 ±  2%  slabinfo.signal_cache.num_objs
    321.28            +0.1%     321.61 ±  2%  slabinfo.signal_cache.num_slabs
     11923            +0.2%      11946        slabinfo.sigqueue.active_objs
    233.79            +0.2%     234.24        slabinfo.sigqueue.active_slabs
     11923            +0.2%      11946        slabinfo.sigqueue.num_objs
    233.79            +0.2%     234.24        slabinfo.sigqueue.num_slabs
    510.09 ±  8%      +3.2%     526.37 ±  6%  slabinfo.skbuff_ext_cache.active_objs
     12.15 ±  8%      +3.2%      12.53 ±  6%  slabinfo.skbuff_ext_cache.active_slabs
    510.09 ±  8%      +3.2%     526.37 ±  6%  slabinfo.skbuff_ext_cache.num_objs
     12.15 ±  8%      +3.2%      12.53 ±  6%  slabinfo.skbuff_ext_cache.num_slabs
     15648            -0.4%      15590        slabinfo.skbuff_head_cache.active_objs
    245.30            -0.4%     244.20        slabinfo.skbuff_head_cache.active_slabs
     15699            -0.4%      15628        slabinfo.skbuff_head_cache.num_objs
    245.30            -0.4%     244.20        slabinfo.skbuff_head_cache.num_slabs
      8653 ±  5%      -1.9%       8491 ±  5%  slabinfo.skbuff_small_head.active_objs
    169.68 ±  5%      -1.9%     166.50 ±  5%  slabinfo.skbuff_small_head.active_slabs
      8653 ±  5%      -1.9%       8491 ±  5%  slabinfo.skbuff_small_head.num_objs
    169.68 ±  5%      -1.9%     166.50 ±  5%  slabinfo.skbuff_small_head.num_slabs
      4352 ±  5%      -2.8%       4228 ±  5%  slabinfo.sock_inode_cache.active_objs
    111.59 ±  5%      -2.8%     108.42 ±  5%  slabinfo.sock_inode_cache.active_slabs
      4352 ±  5%      -2.8%       4228 ±  5%  slabinfo.sock_inode_cache.num_objs
    111.59 ±  5%      -2.8%     108.42 ±  5%  slabinfo.sock_inode_cache.num_slabs
      1634 ±  6%      +4.9%       1714 ±  5%  slabinfo.task_group.active_objs
     32.05 ±  6%      +4.9%      33.62 ±  5%  slabinfo.task_group.active_slabs
      1634 ±  6%      +4.9%       1714 ±  5%  slabinfo.task_group.num_objs
     32.05 ±  6%      +4.9%      33.62 ±  5%  slabinfo.task_group.num_slabs
      3787            -0.3%       3777        slabinfo.task_struct.active_objs
      3790            -0.3%       3779        slabinfo.task_struct.active_slabs
      3790            -0.3%       3779        slabinfo.task_struct.num_objs
      3790            -0.3%       3779        slabinfo.task_struct.num_slabs
    262.17 ±  8%      -1.9%     257.14 ±  9%  slabinfo.taskstats.active_objs
      7.09 ±  8%      -1.9%       6.95 ±  9%  slabinfo.taskstats.active_slabs
    262.17 ±  8%      -1.9%     257.14 ±  9%  slabinfo.taskstats.num_objs
      7.09 ±  8%      -1.9%       6.95 ±  9%  slabinfo.taskstats.num_slabs
      2350            -0.4%       2341        slabinfo.trace_event_file.active_objs
     51.10            -0.4%      50.90        slabinfo.trace_event_file.active_slabs
      2350            -0.4%       2341        slabinfo.trace_event_file.num_objs
     51.10            -0.4%      50.90        slabinfo.trace_event_file.num_slabs
      2394            +0.3%       2400        slabinfo.tracefs_inode_cache.active_objs
     47.89            +0.3%      48.02        slabinfo.tracefs_inode_cache.active_slabs
      2394            +0.3%       2400        slabinfo.tracefs_inode_cache.num_objs
     47.89            +0.3%      48.02        slabinfo.tracefs_inode_cache.num_slabs
     60.00            +0.0%      60.00        slabinfo.tw_sock_TCP.active_objs
      1.00            +0.0%       1.00        slabinfo.tw_sock_TCP.active_slabs
     60.00            +0.0%      60.00        slabinfo.tw_sock_TCP.num_objs
      1.00            +0.0%       1.00        slabinfo.tw_sock_TCP.num_slabs
     74.00            +0.0%      74.00        slabinfo.uts_namespace.active_objs
      2.00            +0.0%       2.00        slabinfo.uts_namespace.active_slabs
     74.00            +0.0%      74.00        slabinfo.uts_namespace.num_objs
      2.00            +0.0%       2.00        slabinfo.uts_namespace.num_slabs
     85556            -2.1%      83786 ±  4%  slabinfo.vm_area_struct.active_objs
      1952            -1.8%       1916 ±  4%  slabinfo.vm_area_struct.active_slabs
     85920            -1.8%      84334 ±  4%  slabinfo.vm_area_struct.num_objs
      1952            -1.8%       1916 ±  4%  slabinfo.vm_area_struct.num_slabs
     98458            -0.5%      97931 ±  2%  slabinfo.vma_lock.active_objs
    969.74            -0.4%     965.79 ±  2%  slabinfo.vma_lock.active_slabs
     98913            -0.4%      98510 ±  2%  slabinfo.vma_lock.num_objs
    969.74            -0.4%     965.79 ±  2%  slabinfo.vma_lock.num_slabs
    811570            +0.0%     811625        slabinfo.vmap_area.active_objs
     14497            +0.0%      14499        slabinfo.vmap_area.active_slabs
    811855            +0.0%     811960        slabinfo.vmap_area.num_objs
     14497            +0.0%      14499        slabinfo.vmap_area.num_slabs
    276.34            +0.1%     276.48        slabinfo.x86_lbr.active_objs
     13.95            +0.0%      13.96        slabinfo.x86_lbr.active_slabs
    293.03            +0.0%     293.07        slabinfo.x86_lbr.num_objs
     13.95            +0.0%      13.96        slabinfo.x86_lbr.num_slabs
      0.00       +1.4e+100%       0.01 ±146%  perf-sched.sch_delay.avg.ms.__cond_resched.__alloc_pages.__folio_alloc.vma_alloc_folio.do_anonymous_page
      0.00       +4.1e+101%       0.41 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.__alloc_pages.__pud_alloc.__handle_mm_fault.handle_mm_fault
      0.00          +7e+98%       0.00 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.__alloc_pages.pte_alloc_one.__pte_alloc.move_page_tables
      0.01 ±300%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.__cond_resched.__flush_work.isra.0.__lru_add_drain_all
      0.00 ±300%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.__cond_resched.__kmem_cache_alloc_node.__kmalloc.acpi_ex_allocate_name_string.acpi_ex_get_name_string
      0.00          +7e+98%       0.00 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.__kmem_cache_alloc_node.__kmalloc.load_elf_phdrs.load_elf_binary
      0.00          +8e+99%       0.01 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.__kmem_cache_alloc_node.__kmalloc_node.memcg_alloc_slab_cgroups.allocate_slab
      0.00 ±300%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.__cond_resched.__kmem_cache_alloc_node.kmalloc_node_trace.alloc_fair_sched_group.sched_create_group
      0.00 ±300%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.__cond_resched.__kmem_cache_alloc_node.kmalloc_trace.perf_event_mmap_event.perf_event_mmap
      0.00        +1.4e+99%       0.00 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.__kmem_cache_alloc_node.kmalloc_trace.vmstat_start.seq_read_iter
      0.00        +1.1e+99%       0.00 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.__put_anon_vma.unlink_anon_vmas.free_pgtables.exit_mmap
      0.00        +1.4e+99%       0.00 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.__vmalloc_area_node.__vmalloc_node_range.alloc_thread_stack_node.dup_task_struct
      0.00 ± 96%     +50.0%       0.00 ± 80%  perf-sched.sch_delay.avg.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
      0.18 ±292%     -97.9%       0.00 ±156%  perf-sched.sch_delay.avg.ms.__cond_resched.__wait_for_common.wait_for_completion_state.kernel_clone.__x64_sys_vfork
      0.00 ±300%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.__cond_resched.cgroup_css_set_fork.cgroup_can_fork.copy_process.kernel_clone
      0.00       +7.8e+100%       0.08 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.change_pmd_range.change_p4d_range.change_protection_range.mprotect_fixup
      0.00          +9e+98%       0.00 ±299%  perf-sched.sch_delay.avg.ms.__cond_resched.copy_strings.isra.0.do_execveat_common
      0.00       +3.7e+101%       0.37 ±299%  perf-sched.sch_delay.avg.ms.__cond_resched.dentry_kill.dput.step_into.link_path_walk
      0.00 ±300%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.__cond_resched.down_read.walk_component.path_lookupat.filename_lookup
      0.00        +1.1e+99%       0.00 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.down_write.anon_vma_clone.__split_vma.vma_modify
      0.05 ±298%    +730.6%       0.39 ±106%  perf-sched.sch_delay.avg.ms.__cond_resched.down_write_killable.exec_mmap.begin_new_exec.load_elf_binary
      0.00          +7e+98%       0.00 ±299%  perf-sched.sch_delay.avg.ms.__cond_resched.down_write_killable.setup_arg_pages.load_elf_binary.search_binary_handler
      0.00       +5.1e+100%       0.05 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64
      0.00       +2.5e+101%       0.25 ±298%  perf-sched.sch_delay.avg.ms.__cond_resched.dput.step_into.link_path_walk.part
      0.00       +2.7e+101%       0.27 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.dput.step_into.path_lookupat.filename_lookup
      0.00 ±300%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.__cond_resched.dput.terminate_walk.path_openat.do_filp_open
      0.00          +1e+99%       0.00 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.exit_mmap.__mmput.exec_mmap.begin_new_exec
      0.00 ±300%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.__cond_resched.filemap_read.__kernel_read.search_binary_handler.exec_binprm
      0.00          +8e+98%       0.00 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.khugepaged.kthread.ret_from_fork.ret_from_fork_asm
      0.00          +1e+99%       0.00 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.kmem_cache_alloc.__anon_vma_prepare.do_anonymous_page.__handle_mm_fault
      0.00        +2.7e+99%       0.00 ±299%  perf-sched.sch_delay.avg.ms.__cond_resched.kmem_cache_alloc.acpi_ut_create_generic_state.acpi_ds_result_push.acpi_ds_exec_end_op
      0.00 ±152%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.__cond_resched.kmem_cache_alloc.alloc_pid.copy_process.kernel_clone
      0.00       +2.5e+101%       0.25 ±299%  perf-sched.sch_delay.avg.ms.__cond_resched.kmem_cache_alloc.getname_flags.part.0
      0.05 ±300%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.__cond_resched.kmem_cache_alloc.mas_alloc_nodes.mas_preallocate.__split_vma
      0.00       +4.2e+100%       0.04 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.kmem_cache_alloc.mas_alloc_nodes.mas_preallocate.vma_link
      0.00 ±200%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.__cond_resched.kmem_cache_alloc.vm_area_alloc.alloc_bprm.do_execveat_common
      0.00        +1.3e+99%       0.00 ±299%  perf-sched.sch_delay.avg.ms.__cond_resched.kmem_cache_alloc.vm_area_dup.dup_mmap.dup_mm
      0.00        +1.6e+99%       0.00 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.mutex_lock.drm_property_free_blob.drm_gem_destroy_shadow_plane_state.drm_atomic_state_default_clear
      0.00       +5.4e+101%       0.54 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.mutex_lock_interruptible.devkmsg_read.vfs_read.ksys_read
      0.00       +9.1e+101%       0.91 ±130%  perf-sched.sch_delay.avg.ms.__cond_resched.mutex_lock_killable.pcpu_alloc.mm_init.alloc_bprm
      0.00        +7.9e+99%       0.01 ±133%  perf-sched.sch_delay.avg.ms.__cond_resched.mutex_lock_killable.pcpu_alloc.mm_init.dup_mm
      0.01 ±299%     -65.0%       0.00 ±223%  perf-sched.sch_delay.avg.ms.__cond_resched.process_one_work.worker_thread.kthread.ret_from_fork
      0.00        +2.1e+99%       0.00 ±213%  perf-sched.sch_delay.avg.ms.__cond_resched.refresh_cpu_vm_stats.vmstat_update.process_one_work.worker_thread
      0.30 ±299%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.__cond_resched.remove_vma.do_vmi_align_munmap.do_vmi_munmap.__vm_munmap
      0.00       +8.1e+101%       0.81 ±155%  perf-sched.sch_delay.avg.ms.__cond_resched.shmem_get_folio_gfp.shmem_read_folio_gfp.drm_gem_get_pages.drm_gem_shmem_get_pages
      0.01 ±  3%      -5.8%       0.01 ±  4%  perf-sched.sch_delay.avg.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      1.77 ±  5%      +4.7%       1.85 ± 10%  perf-sched.sch_delay.avg.ms.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part
      0.33 ±300%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.__cond_resched.task_work_run.exit_to_user_mode_loop.exit_to_user_mode_prepare.syscall_exit_to_user_mode
      0.00        +2.2e+99%       0.00 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.task_work_run.get_signal.arch_do_signal_or_restart.exit_to_user_mode_loop
      0.00       +5.9e+100%       0.06 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.tlb_batch_pages_flush.tlb_finish_mmu.exit_mmap.__mmput
      0.00       +3.1e+100%       0.03 ±201%  perf-sched.sch_delay.avg.ms.__cond_resched.unmap_vmas.unmap_region.constprop.0
      0.31 ±201%    +246.6%       1.06 ±228%  perf-sched.sch_delay.avg.ms.__cond_resched.wp_page_copy.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
      1.67 ±152%     +24.2%       2.08 ±122%  perf-sched.sch_delay.avg.ms.__cond_resched.ww_mutex_lock.drm_gem_vunmap_unlocked.drm_fbdev_generic_helper_fb_dirty.drm_fb_helper_damage_work
      1.06 ±200%     +40.8%       1.50 ±153%  perf-sched.sch_delay.avg.ms.__cond_resched.ww_mutex_lock.drm_gem_vunmap_unlocked.drm_gem_fb_vunmap.drm_atomic_helper_cleanup_planes
      0.00        +4.7e+99%       0.00 ±192%  perf-sched.sch_delay.avg.ms.__cond_resched.zap_pmd_range.isra.0.unmap_page_range
      0.69 ± 83%     -69.5%       0.21 ±172%  perf-sched.sch_delay.avg.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
      0.20 ±300%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.cleaner_kthread.kthread.ret_from_fork.ret_from_fork_asm
      0.84 ± 69%   +1807.1%      15.92 ±108%  perf-sched.sch_delay.avg.ms.devkmsg_read.vfs_read.ksys_read.do_syscall_64
      0.42 ±119%     -48.3%       0.22 ±121%  perf-sched.sch_delay.avg.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
      0.14 ± 54%    +235.6%       0.46 ± 31%  perf-sched.sch_delay.avg.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
      0.16 ± 32%     +37.1%       0.22 ± 20%  perf-sched.sch_delay.avg.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
      0.01 ±208%      +6.1%       0.01 ±123%  perf-sched.sch_delay.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_exc_page_fault
      0.02 ± 41%    +249.5%       0.07 ± 68%  perf-sched.sch_delay.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.syscall_exit_to_user_mode.do_syscall_64
      0.00 ±300%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.futex_wait_queue.futex_wait.do_futex.__x64_sys_futex
      0.20 ±300%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.io_schedule.folio_wait_bit_common.filemap_fault.__do_fault
      0.28 ±297%     -99.2%       0.00 ±300%  perf-sched.sch_delay.avg.ms.kthreadd.ret_from_fork.ret_from_fork_asm
      0.04 ± 40%     -28.4%       0.03 ± 28%  perf-sched.sch_delay.avg.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
      0.40 ± 23%     -27.9%       0.29 ± 40%  perf-sched.sch_delay.avg.ms.schedule_hrtimeout_range_clock.do_poll.constprop.0.do_sys_poll
      2.21 ± 49%      +9.3%       2.42 ± 42%  perf-sched.sch_delay.avg.ms.schedule_hrtimeout_range_clock.do_select.core_sys_select.kern_select
      0.25 ± 62%     +79.2%       0.46 ±168%  perf-sched.sch_delay.avg.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
      0.00 ±202%     -28.6%       0.00 ±300%  perf-sched.sch_delay.avg.ms.schedule_preempt_disabled.kthread.ret_from_fork.ret_from_fork_asm
      0.00 ±300%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.schedule_preempt_disabled.rwsem_down_read_slowpath.down_read_killable.__access_remote_vm
      2.67 ±203%     -30.4%       1.86 ±300%  perf-sched.sch_delay.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_killable.__kthread_create_on_node
      0.45 ± 23%     +14.3%       0.52 ± 16%  perf-sched.sch_delay.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
      0.02 ± 25%     -13.2%       0.02 ± 23%  perf-sched.sch_delay.avg.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
      0.02 ± 45%     +11.2%       0.03 ± 39%  perf-sched.sch_delay.avg.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
      0.00 ±299%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.schedule_timeout.transaction_kthread.kthread.ret_from_fork
      0.01           +77.7%       0.02 ± 32%  perf-sched.sch_delay.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      0.96 ± 77%   +1111.2%      11.68 ± 99%  perf-sched.sch_delay.avg.ms.syslog_print.do_syslog.kmsg_read.vfs_read
      0.12 ±102%    +410.5%       0.62 ± 19%  perf-sched.sch_delay.avg.ms.wait_for_partner.fifo_open.do_dentry_open.do_open
      0.02 ± 10%     +57.8%       0.04 ± 48%  perf-sched.sch_delay.avg.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
      0.00       +6.6e+101%       0.66 ±159%  perf-sched.sch_delay.max.ms.__cond_resched.__alloc_pages.__folio_alloc.vma_alloc_folio.do_anonymous_page
      0.00       +4.1e+101%       0.41 ±300%  perf-sched.sch_delay.max.ms.__cond_resched.__alloc_pages.__pud_alloc.__handle_mm_fault.handle_mm_fault
      0.00          +7e+98%       0.00 ±300%  perf-sched.sch_delay.max.ms.__cond_resched.__alloc_pages.pte_alloc_one.__pte_alloc.move_page_tables
      0.01 ±300%    -100.0%       0.00        perf-sched.sch_delay.max.ms.__cond_resched.__flush_work.isra.0.__lru_add_drain_all
      0.00 ±300%    -100.0%       0.00        perf-sched.sch_delay.max.ms.__cond_resched.__kmem_cache_alloc_node.__kmalloc.acpi_ex_allocate_name_string.acpi_ex_get_name_string
      0.00          +7e+98%       0.00 ±300%  perf-sched.sch_delay.max.ms.__cond_resched.__kmem_cache_alloc_node.__kmalloc.load_elf_phdrs.load_elf_binary
      0.00       +1.6e+100%       0.02 ±300%  perf-sched.sch_delay.max.ms.__cond_resched.__kmem_cache_alloc_node.__kmalloc_node.memcg_alloc_slab_cgroups.allocate_slab
      0.00 ±300%    -100.0%       0.00        perf-sched.sch_delay.max.ms.__cond_resched.__kmem_cache_alloc_node.kmalloc_node_trace.alloc_fair_sched_group.sched_create_group
      0.00 ±300%    -100.0%       0.00        perf-sched.sch_delay.max.ms.__cond_resched.__kmem_cache_alloc_node.kmalloc_trace.perf_event_mmap_event.perf_event_mmap
      0.00        +1.4e+99%       0.00 ±300%  perf-sched.sch_delay.max.ms.__cond_resched.__kmem_cache_alloc_node.kmalloc_trace.vmstat_start.seq_read_iter
      0.00        +1.1e+99%       0.00 ±300%  perf-sched.sch_delay.max.ms.__cond_resched.__put_anon_vma.unlink_anon_vmas.free_pgtables.exit_mmap
      0.00        +1.4e+99%       0.00 ±300%  perf-sched.sch_delay.max.ms.__cond_resched.__vmalloc_area_node.__vmalloc_node_range.alloc_thread_stack_node.dup_task_struct
      1.62 ± 97%     +65.5%       2.69 ± 82%  perf-sched.sch_delay.max.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
      0.18 ±292%     -97.9%       0.00 ±156%  perf-sched.sch_delay.max.ms.__cond_resched.__wait_for_common.wait_for_completion_state.kernel_clone.__x64_sys_vfork
      0.00 ±300%    -100.0%       0.00        perf-sched.sch_delay.max.ms.__cond_resched.cgroup_css_set_fork.cgroup_can_fork.copy_process.kernel_clone
      0.00       +2.3e+101%       0.23 ±300%  perf-sched.sch_delay.max.ms.__cond_resched.change_pmd_range.change_p4d_range.change_protection_range.mprotect_fixup
      0.00          +9e+98%       0.00 ±299%  perf-sched.sch_delay.max.ms.__cond_resched.copy_strings.isra.0.do_execveat_common
      0.00       +3.7e+101%       0.37 ±299%  perf-sched.sch_delay.max.ms.__cond_resched.dentry_kill.dput.step_into.link_path_walk
      0.00 ±300%    -100.0%       0.00        perf-sched.sch_delay.max.ms.__cond_resched.down_read.walk_component.path_lookupat.filename_lookup
      0.00        +1.1e+99%       0.00 ±300%  perf-sched.sch_delay.max.ms.__cond_resched.down_write.anon_vma_clone.__split_vma.vma_modify
      0.19 ±298%    +433.1%       1.00 ± 87%  perf-sched.sch_delay.max.ms.__cond_resched.down_write_killable.exec_mmap.begin_new_exec.load_elf_binary
      0.00          +7e+98%       0.00 ±299%  perf-sched.sch_delay.max.ms.__cond_resched.down_write_killable.setup_arg_pages.load_elf_binary.search_binary_handler
      0.00       +3.6e+101%       0.36 ±299%  perf-sched.sch_delay.max.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64
      0.00       +2.5e+101%       0.25 ±298%  perf-sched.sch_delay.max.ms.__cond_resched.dput.step_into.link_path_walk.part
      0.00       +2.7e+101%       0.27 ±300%  perf-sched.sch_delay.max.ms.__cond_resched.dput.step_into.path_lookupat.filename_lookup
      0.00 ±300%    -100.0%       0.00        perf-sched.sch_delay.max.ms.__cond_resched.dput.terminate_walk.path_openat.do_filp_open
      0.00          +1e+99%       0.00 ±300%  perf-sched.sch_delay.max.ms.__cond_resched.exit_mmap.__mmput.exec_mmap.begin_new_exec
      0.00 ±300%    -100.0%       0.00        perf-sched.sch_delay.max.ms.__cond_resched.filemap_read.__kernel_read.search_binary_handler.exec_binprm
      0.00        +1.6e+99%       0.00 ±300%  perf-sched.sch_delay.max.ms.__cond_resched.khugepaged.kthread.ret_from_fork.ret_from_fork_asm
      0.00          +1e+99%       0.00 ±300%  perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc.__anon_vma_prepare.do_anonymous_page.__handle_mm_fault
      0.00        +2.7e+99%       0.00 ±299%  perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc.acpi_ut_create_generic_state.acpi_ds_result_push.acpi_ds_exec_end_op
      0.00 ±152%    -100.0%       0.00        perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc.alloc_pid.copy_process.kernel_clone
      0.00       +2.5e+101%       0.25 ±299%  perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc.getname_flags.part.0
      0.26 ±300%    -100.0%       0.00        perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc.mas_alloc_nodes.mas_preallocate.__split_vma
      0.00       +4.2e+100%       0.04 ±300%  perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc.mas_alloc_nodes.mas_preallocate.vma_link
      0.00 ±200%    -100.0%       0.00        perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc.vm_area_alloc.alloc_bprm.do_execveat_common
      0.00        +1.3e+99%       0.00 ±299%  perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc.vm_area_dup.dup_mmap.dup_mm
      0.00        +1.6e+99%       0.00 ±300%  perf-sched.sch_delay.max.ms.__cond_resched.mutex_lock.drm_property_free_blob.drm_gem_destroy_shadow_plane_state.drm_atomic_state_default_clear
      0.00       +5.4e+101%       0.54 ±300%  perf-sched.sch_delay.max.ms.__cond_resched.mutex_lock_interruptible.devkmsg_read.vfs_read.ksys_read
      0.00       +1.2e+102%       1.24 ±129%  perf-sched.sch_delay.max.ms.__cond_resched.mutex_lock_killable.pcpu_alloc.mm_init.alloc_bprm
      0.00        +7.9e+99%       0.01 ±133%  perf-sched.sch_delay.max.ms.__cond_resched.mutex_lock_killable.pcpu_alloc.mm_init.dup_mm
      0.01 ±299%     -65.0%       0.00 ±223%  perf-sched.sch_delay.max.ms.__cond_resched.process_one_work.worker_thread.kthread.ret_from_fork
      0.00        +2.1e+99%       0.00 ±213%  perf-sched.sch_delay.max.ms.__cond_resched.refresh_cpu_vm_stats.vmstat_update.process_one_work.worker_thread
      0.30 ±299%    -100.0%       0.00        perf-sched.sch_delay.max.ms.__cond_resched.remove_vma.do_vmi_align_munmap.do_vmi_munmap.__vm_munmap
      0.00       +1.3e+102%       1.30 ±158%  perf-sched.sch_delay.max.ms.__cond_resched.shmem_get_folio_gfp.shmem_read_folio_gfp.drm_gem_get_pages.drm_gem_shmem_get_pages
      0.02 ± 22%      +4.5%       0.02 ± 23%  perf-sched.sch_delay.max.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      4.01 ±  3%      -0.8%       3.97 ±  2%  perf-sched.sch_delay.max.ms.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part
      0.33 ±300%    -100.0%       0.00        perf-sched.sch_delay.max.ms.__cond_resched.task_work_run.exit_to_user_mode_loop.exit_to_user_mode_prepare.syscall_exit_to_user_mode
      0.00        +2.2e+99%       0.00 ±300%  perf-sched.sch_delay.max.ms.__cond_resched.task_work_run.get_signal.arch_do_signal_or_restart.exit_to_user_mode_loop
      0.00       +1.2e+101%       0.12 ±300%  perf-sched.sch_delay.max.ms.__cond_resched.tlb_batch_pages_flush.tlb_finish_mmu.exit_mmap.__mmput
      0.00       +5.7e+101%       0.57 ±200%  perf-sched.sch_delay.max.ms.__cond_resched.unmap_vmas.unmap_region.constprop.0
      0.44 ±205%    +141.1%       1.06 ±228%  perf-sched.sch_delay.max.ms.__cond_resched.wp_page_copy.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
      1.67 ±152%     +24.2%       2.08 ±122%  perf-sched.sch_delay.max.ms.__cond_resched.ww_mutex_lock.drm_gem_vunmap_unlocked.drm_fbdev_generic_helper_fb_dirty.drm_fb_helper_damage_work
      1.06 ±200%     +40.8%       1.50 ±153%  perf-sched.sch_delay.max.ms.__cond_resched.ww_mutex_lock.drm_gem_vunmap_unlocked.drm_gem_fb_vunmap.drm_atomic_helper_cleanup_planes
      0.00       +3.7e+100%       0.04 ±191%  perf-sched.sch_delay.max.ms.__cond_resched.zap_pmd_range.isra.0.unmap_page_range
      3.10 ± 59%     -60.7%       1.22 ±161%  perf-sched.sch_delay.max.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
      0.20 ±300%    -100.0%       0.00        perf-sched.sch_delay.max.ms.cleaner_kthread.kthread.ret_from_fork.ret_from_fork_asm
      1.64 ± 70%    +909.5%      16.51 ±101%  perf-sched.sch_delay.max.ms.devkmsg_read.vfs_read.ksys_read.do_syscall_64
      2.08 ±108%     -37.6%       1.30 ±122%  perf-sched.sch_delay.max.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
      3.28 ± 34%     +29.2%       4.24 ± 13%  perf-sched.sch_delay.max.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
      3.32 ± 27%      -2.9%       3.22 ± 23%  perf-sched.sch_delay.max.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
      0.53 ±209%    +168.8%       1.43 ± 93%  perf-sched.sch_delay.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_exc_page_fault
      0.55 ±194%     +60.5%       0.88 ±201%  perf-sched.sch_delay.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt
      0.00 ±300%    -100.0%       0.00        perf-sched.sch_delay.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi
      2.77 ± 30%    +150.3%       6.93 ±142%  perf-sched.sch_delay.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.syscall_exit_to_user_mode.do_syscall_64
      0.00 ±300%    -100.0%       0.00        perf-sched.sch_delay.max.ms.futex_wait_queue.futex_wait.do_futex.__x64_sys_futex
      0.40 ±300%    -100.0%       0.00        perf-sched.sch_delay.max.ms.io_schedule.folio_wait_bit_common.filemap_fault.__do_fault
      0.28 ±297%     -99.2%       0.00 ±300%  perf-sched.sch_delay.max.ms.kthreadd.ret_from_fork.ret_from_fork_asm
      3.24 ± 30%     -10.1%       2.91 ± 47%  perf-sched.sch_delay.max.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
      3.63 ± 18%     -14.5%       3.10 ± 33%  perf-sched.sch_delay.max.ms.schedule_hrtimeout_range_clock.do_poll.constprop.0.do_sys_poll
      3.25 ± 30%     +14.0%       3.70 ± 27%  perf-sched.sch_delay.max.ms.schedule_hrtimeout_range_clock.do_select.core_sys_select.kern_select
      2.83 ± 57%     +81.2%       5.12 ±167%  perf-sched.sch_delay.max.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
      0.00 ±202%     -28.6%       0.00 ±300%  perf-sched.sch_delay.max.ms.schedule_preempt_disabled.kthread.ret_from_fork.ret_from_fork_asm
      0.00 ±300%    -100.0%       0.00        perf-sched.sch_delay.max.ms.schedule_preempt_disabled.rwsem_down_read_slowpath.down_read_killable.__access_remote_vm
      2.67 ±203%     -30.4%       1.86 ±300%  perf-sched.sch_delay.max.ms.schedule_timeout.__wait_for_common.wait_for_completion_killable.__kthread_create_on_node
      4.15 ± 10%     +17.7%       4.89 ± 23%  perf-sched.sch_delay.max.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
      0.08 ±106%     -26.8%       0.06 ±115%  perf-sched.sch_delay.max.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
      8.41 ±161%     +26.1%      10.61 ± 97%  perf-sched.sch_delay.max.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
      0.00 ±299%    -100.0%       0.00        perf-sched.sch_delay.max.ms.schedule_timeout.transaction_kthread.kthread.ret_from_fork
      4.38 ± 45%    +154.3%      11.15 ± 59%  perf-sched.sch_delay.max.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      1.90 ± 78%    +518.9%      11.77 ± 98%  perf-sched.sch_delay.max.ms.syslog_print.do_syslog.kmsg_read.vfs_read
      2.30 ± 46%     +56.5%       3.60 ±  7%  perf-sched.sch_delay.max.ms.wait_for_partner.fifo_open.do_dentry_open.do_open
      3.74 ± 39%     +32.3%       4.95 ± 55%  perf-sched.sch_delay.max.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
      0.01 ±  4%     +41.1%       0.02 ± 19%  perf-sched.total_sch_delay.average.ms
     13.07 ± 84%    +113.7%      27.94 ± 42%  perf-sched.total_sch_delay.max.ms
     73.81            +1.5%      74.92        perf-sched.total_wait_and_delay.average.ms
     42853            -1.5%      42224        perf-sched.total_wait_and_delay.count.ms
      4899            -0.2%       4892        perf-sched.total_wait_and_delay.max.ms
     73.80            +1.5%      74.90        perf-sched.total_wait_time.average.ms
      4898            -0.2%       4889        perf-sched.total_wait_time.max.ms
     40.85 ±299%    -100.0%       0.00        perf-sched.wait_and_delay.avg.ms.__cond_resched.__kmem_cache_alloc_node.__kmalloc.acpi_ex_allocate_name_string.acpi_ex_get_name_string
      0.00         +1e+104%     100.01 ±300%  perf-sched.wait_and_delay.avg.ms.__cond_resched.__kmem_cache_alloc_node.kmalloc_trace.vmstat_start.seq_read_iter
      6.73            -5.2%       6.37 ± 17%  perf-sched.wait_and_delay.avg.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
      0.00       +1.4e+103%      14.35 ±299%  perf-sched.wait_and_delay.avg.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64
      0.00         +5e+103%      50.14 ±300%  perf-sched.wait_and_delay.avg.ms.__cond_resched.kmem_cache_alloc.acpi_ut_create_generic_state.acpi_ds_result_push.acpi_ds_exec_end_op
      0.00         +1e+104%     100.02 ±300%  perf-sched.wait_and_delay.avg.ms.__cond_resched.kmem_cache_alloc.vm_area_dup.dup_mmap.dup_mm
      0.00       +2.1e+103%      20.74 ±299%  perf-sched.wait_and_delay.avg.ms.__cond_resched.mutex_lock.drm_property_free_blob.drm_gem_destroy_shadow_plane_state.drm_atomic_state_default_clear
      0.00         +3e+104%     300.08 ±152%  perf-sched.wait_and_delay.avg.ms.__cond_resched.mutex_lock_killable.pcpu_alloc.mm_init.dup_mm
    204.67 ±300%      +0.1%     204.80 ±200%  perf-sched.wait_and_delay.avg.ms.__cond_resched.process_one_work.worker_thread.kthread.ret_from_fork
      0.00         +2e+104%     204.67 ±200%  perf-sched.wait_and_delay.avg.ms.__cond_resched.refresh_cpu_vm_stats.vmstat_update.process_one_work.worker_thread
      0.00       +5.3e+103%      52.66 ±161%  perf-sched.wait_and_delay.avg.ms.__cond_resched.shmem_get_folio_gfp.shmem_read_folio_gfp.drm_gem_get_pages.drm_gem_shmem_get_pages
      0.00       +7.8e+102%       7.84 ±299%  perf-sched.wait_and_delay.avg.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      0.00         +1e+104%     100.06 ±300%  perf-sched.wait_and_delay.avg.ms.__cond_resched.task_work_run.get_signal.arch_do_signal_or_restart.exit_to_user_mode_loop
      0.00       +2.6e+103%      25.66 ±202%  perf-sched.wait_and_delay.avg.ms.__cond_resched.zap_pmd_range.isra.0.unmap_page_range
    271.66 ± 68%     -70.3%      80.71 ±155%  perf-sched.wait_and_delay.avg.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
    214.27 ± 86%     -29.6%     150.78 ±104%  perf-sched.wait_and_delay.avg.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
    235.93 ±  5%      +1.4%     239.24 ±  5%  perf-sched.wait_and_delay.avg.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
      0.00       +7.1e+101%       0.71 ±175%  perf-sched.wait_and_delay.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_exc_page_fault
      0.24 ±  6%    +311.6%       1.00 ± 46%  perf-sched.wait_and_delay.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt
      0.02 ±299%    +545.1%       0.11 ± 45%  perf-sched.wait_and_delay.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_call_function_single
      0.24 ±  6%    +338.2%       1.07 ± 53%  perf-sched.wait_and_delay.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi
     23.21 ± 50%      -6.7%      21.65 ± 34%  perf-sched.wait_and_delay.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.syscall_exit_to_user_mode.do_syscall_64
    169.92 ± 16%     -30.1%     118.73 ± 28%  perf-sched.wait_and_delay.avg.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
    552.06 ±  5%      +1.1%     558.36 ±  2%  perf-sched.wait_and_delay.avg.ms.schedule_hrtimeout_range_clock.do_poll.constprop.0.do_sys_poll
    370.29 ±  4%      +4.6%     387.38 ±  3%  perf-sched.wait_and_delay.avg.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
    319.32 ±299%    -100.0%       0.00        perf-sched.wait_and_delay.avg.ms.schedule_preempt_disabled.rwsem_down_read_slowpath.down_read_killable.__access_remote_vm
    452.79            +0.1%     453.34        perf-sched.wait_and_delay.avg.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
      3.65 ±  2%      -5.4%       3.45 ±  5%  perf-sched.wait_and_delay.avg.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
    121.38            +2.6%     124.58        perf-sched.wait_and_delay.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
    724.27 ±  2%      +1.6%     735.88        perf-sched.wait_and_delay.avg.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
      0.10 ±300%    -100.0%       0.00        perf-sched.wait_and_delay.count.__cond_resched.__kmem_cache_alloc_node.__kmalloc.acpi_ex_allocate_name_string.acpi_ex_get_name_string
      0.00         +1e+101%       0.10 ±300%  perf-sched.wait_and_delay.count.__cond_resched.__kmem_cache_alloc_node.kmalloc_trace.vmstat_start.seq_read_iter
      1344            +0.0%       1344        perf-sched.wait_and_delay.count.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
      0.00         +7e+101%       0.70 ±300%  perf-sched.wait_and_delay.count.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64
      0.00         +1e+101%       0.10 ±300%  perf-sched.wait_and_delay.count.__cond_resched.kmem_cache_alloc.acpi_ut_create_generic_state.acpi_ds_result_push.acpi_ds_exec_end_op
      0.00         +1e+101%       0.10 ±300%  perf-sched.wait_and_delay.count.__cond_resched.kmem_cache_alloc.vm_area_dup.dup_mmap.dup_mm
      0.00         +1e+101%       0.10 ±300%  perf-sched.wait_and_delay.count.__cond_resched.mutex_lock.drm_property_free_blob.drm_gem_destroy_shadow_plane_state.drm_atomic_state_default_clear
      0.00         +3e+101%       0.30 ±152%  perf-sched.wait_and_delay.count.__cond_resched.mutex_lock_killable.pcpu_alloc.mm_init.dup_mm
      0.10 ±300%    +100.0%       0.20 ±200%  perf-sched.wait_and_delay.count.__cond_resched.process_one_work.worker_thread.kthread.ret_from_fork
      0.00         +2e+101%       0.20 ±200%  perf-sched.wait_and_delay.count.__cond_resched.refresh_cpu_vm_stats.vmstat_update.process_one_work.worker_thread
      0.00         +4e+101%       0.40 ±165%  perf-sched.wait_and_delay.count.__cond_resched.shmem_get_folio_gfp.shmem_read_folio_gfp.drm_gem_get_pages.drm_gem_shmem_get_pages
      0.00       +1.3e+103%      12.90 ±299%  perf-sched.wait_and_delay.count.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      0.00         +1e+101%       0.10 ±300%  perf-sched.wait_and_delay.count.__cond_resched.task_work_run.get_signal.arch_do_signal_or_restart.exit_to_user_mode_loop
      0.00       +1.6e+102%       1.60 ±201%  perf-sched.wait_and_delay.count.__cond_resched.zap_pmd_range.isra.0.unmap_page_range
      8.00 ± 50%     -50.0%       4.00 ±122%  perf-sched.wait_and_delay.count.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
      4.40 ± 66%     -18.2%       3.60 ± 81%  perf-sched.wait_and_delay.count.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
    124.30            -0.2%     124.00        perf-sched.wait_and_delay.count.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
      0.00       +1.1e+104%     112.40 ±152%  perf-sched.wait_and_delay.count.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_exc_page_fault
     18268            -4.4%      17455 ±  3%  perf-sched.wait_and_delay.count.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt
     28.50 ±300%    +834.4%     266.30 ± 34%  perf-sched.wait_and_delay.count.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_call_function_single
      1595            +1.4%       1617 ±  3%  perf-sched.wait_and_delay.count.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi
    284.40 ± 37%     +95.5%     556.00 ± 36%  perf-sched.wait_and_delay.count.exit_to_user_mode_loop.exit_to_user_mode_prepare.syscall_exit_to_user_mode.do_syscall_64
    439.20 ± 15%     +50.7%     662.00 ± 30%  perf-sched.wait_and_delay.count.pipe_read.vfs_read.ksys_read.do_syscall_64
     21.30 ±  5%      -1.4%      21.00 ±  3%  perf-sched.wait_and_delay.count.schedule_hrtimeout_range_clock.do_poll.constprop.0.do_sys_poll
     13.20 ±  7%      -3.8%      12.70 ±  3%  perf-sched.wait_and_delay.count.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
      0.10 ±300%    -100.0%       0.00        perf-sched.wait_and_delay.count.schedule_preempt_disabled.rwsem_down_read_slowpath.down_read_killable.__access_remote_vm
     19.70 ±  3%      +1.0%      19.90        perf-sched.wait_and_delay.count.schedule_timeout.kcompactd.kthread.ret_from_fork
      1365            +6.3%       1451 ±  5%  perf-sched.wait_and_delay.count.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
     17282            -3.5%      16674 ±  2%  perf-sched.wait_and_delay.count.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      1249 ±  2%      -1.4%       1231        perf-sched.wait_and_delay.count.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
     40.85 ±299%    -100.0%       0.00        perf-sched.wait_and_delay.max.ms.__cond_resched.__kmem_cache_alloc_node.__kmalloc.acpi_ex_allocate_name_string.acpi_ex_get_name_string
      0.00         +1e+104%     100.01 ±300%  perf-sched.wait_and_delay.max.ms.__cond_resched.__kmem_cache_alloc_node.kmalloc_trace.vmstat_start.seq_read_iter
      4899            -8.0%       4506 ± 25%  perf-sched.wait_and_delay.max.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
      0.00         +1e+104%     100.35 ±300%  perf-sched.wait_and_delay.max.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64
      0.00         +5e+103%      50.14 ±300%  perf-sched.wait_and_delay.max.ms.__cond_resched.kmem_cache_alloc.acpi_ut_create_generic_state.acpi_ds_result_push.acpi_ds_exec_end_op
      0.00         +1e+104%     100.02 ±300%  perf-sched.wait_and_delay.max.ms.__cond_resched.kmem_cache_alloc.vm_area_dup.dup_mmap.dup_mm
      0.00       +2.1e+103%      20.74 ±299%  perf-sched.wait_and_delay.max.ms.__cond_resched.mutex_lock.drm_property_free_blob.drm_gem_destroy_shadow_plane_state.drm_atomic_state_default_clear
      0.00         +3e+104%     300.08 ±152%  perf-sched.wait_and_delay.max.ms.__cond_resched.mutex_lock_killable.pcpu_alloc.mm_init.dup_mm
    204.67 ±300%      +0.1%     204.80 ±200%  perf-sched.wait_and_delay.max.ms.__cond_resched.process_one_work.worker_thread.kthread.ret_from_fork
      0.00         +2e+104%     204.67 ±200%  perf-sched.wait_and_delay.max.ms.__cond_resched.refresh_cpu_vm_stats.vmstat_update.process_one_work.worker_thread
      0.00       +6.3e+103%      62.80 ±152%  perf-sched.wait_and_delay.max.ms.__cond_resched.shmem_get_folio_gfp.shmem_read_folio_gfp.drm_gem_get_pages.drm_gem_shmem_get_pages
      0.00       +9.3e+102%       9.33 ±300%  perf-sched.wait_and_delay.max.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      0.00         +1e+104%     100.06 ±300%  perf-sched.wait_and_delay.max.ms.__cond_resched.task_work_run.get_signal.arch_do_signal_or_restart.exit_to_user_mode_loop
      0.00         +2e+104%     200.44 ±200%  perf-sched.wait_and_delay.max.ms.__cond_resched.zap_pmd_range.isra.0.unmap_page_range
    804.48 ± 50%     -50.2%     400.68 ±122%  perf-sched.wait_and_delay.max.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
    702.01 ± 65%     -14.4%     600.63 ± 81%  perf-sched.wait_and_delay.max.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
      1323 ± 48%     +37.2%       1815 ± 54%  perf-sched.wait_and_delay.max.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
      0.00         +1e+104%     102.23 ±292%  perf-sched.wait_and_delay.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_exc_page_fault
    128.86 ±226%      +2.2%     131.68 ±219%  perf-sched.wait_and_delay.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt
      0.25 ±299%   +1670.4%       4.39 ± 60%  perf-sched.wait_and_delay.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_call_function_single
      5.66 ±156%     +54.5%       8.74 ± 54%  perf-sched.wait_and_delay.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi
    901.41 ± 33%     +53.9%       1387 ± 83%  perf-sched.wait_and_delay.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.syscall_exit_to_user_mode.do_syscall_64
      1051            +0.8%       1059        perf-sched.wait_and_delay.max.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
      1007            -0.1%       1006        perf-sched.wait_and_delay.max.ms.schedule_hrtimeout_range_clock.do_poll.constprop.0.do_sys_poll
    499.87            +0.0%     499.88        perf-sched.wait_and_delay.max.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
    319.32 ±299%    -100.0%       0.00        perf-sched.wait_and_delay.max.ms.schedule_preempt_disabled.rwsem_down_read_slowpath.down_read_killable.__access_remote_vm
    504.61            +0.1%     505.27        perf-sched.wait_and_delay.max.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
     20.67 ±122%     +10.5%      22.84 ± 81%  perf-sched.wait_and_delay.max.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
      1053           +10.8%       1167 ± 26%  perf-sched.wait_and_delay.max.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      2046            -0.0%       2046        perf-sched.wait_and_delay.max.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
      0.20 ± 25%    +330.8%       0.87 ± 40%  perf-sched.wait_time.avg.ms.__cond_resched.__alloc_pages.__folio_alloc.vma_alloc_folio.do_anonymous_page
      0.30 ±300%    +166.9%       0.80 ±228%  perf-sched.wait_time.avg.ms.__cond_resched.__alloc_pages.__folio_alloc.vma_alloc_folio.shmem_alloc_folio
      0.00 ±300%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.__alloc_pages.__folio_alloc.vma_alloc_folio.wp_page_copy
      0.00       +1.2e+102%       1.19 ±225%  perf-sched.wait_time.avg.ms.__cond_resched.__alloc_pages.__pud_alloc.alloc_new_pud.constprop
      0.40 ±300%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.__alloc_pages.pte_alloc_one.__do_fault.do_read_fault
      0.00       +4.7e+101%       0.47 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.__alloc_pages.pte_alloc_one.__pte_alloc.do_anonymous_page
      2.19 ±124%     -88.6%       0.25 ±299%  perf-sched.wait_time.avg.ms.__cond_resched.__do_fault.do_read_fault.do_fault.__handle_mm_fault
     40.85 ±300%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.__kmem_cache_alloc_node.__kmalloc.acpi_ex_allocate_name_string.acpi_ex_get_name_string
      0.00          +2e+99%       0.00 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.__kmem_cache_alloc_node.__kmalloc_node.alloc_cpumask_var_node.__sched_setaffinity
      0.00       +3.1e+101%       0.31 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.__kmem_cache_alloc_node.__kmalloc_node.memcg_alloc_slab_cgroups.allocate_slab
      0.00 ±300%    +308.3%       0.00 ±200%  perf-sched.wait_time.avg.ms.__cond_resched.__kmem_cache_alloc_node.kmalloc_trace.perf_event_mmap_event.perf_event_mmap
      0.00        +3.9e+99%       0.00 ±200%  perf-sched.wait_time.avg.ms.__cond_resched.__kmem_cache_alloc_node.kmalloc_trace.single_open.single_open_size
      0.00         +1e+104%     100.01 ±299%  perf-sched.wait_time.avg.ms.__cond_resched.__kmem_cache_alloc_node.kmalloc_trace.vmstat_start.seq_read_iter
      6.72            -5.2%       6.37 ± 17%  perf-sched.wait_time.avg.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
      0.84 ±132%     -63.0%       0.31 ±194%  perf-sched.wait_time.avg.ms.__cond_resched.__wait_for_common.wait_for_completion_state.kernel_clone.__x64_sys_vfork
      0.15 ±300%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.cgroup_css_set_fork.cgroup_can_fork.copy_process.kernel_clone
      0.00 ±300%    +327.3%       0.01 ±299%  perf-sched.wait_time.avg.ms.__cond_resched.change_pmd_range.change_p4d_range.change_protection_range.mprotect_fixup
      0.00 ±300%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.dentry_kill.dput.step_into.link_path_walk
      0.15 ±154%    +506.7%       0.92 ±112%  perf-sched.wait_time.avg.ms.__cond_resched.down_write.__split_vma.do_vmi_align_munmap.__do_sys_brk
      0.20 ± 41%    +161.6%       0.53 ± 98%  perf-sched.wait_time.avg.ms.__cond_resched.down_write.anon_vma_clone.__split_vma.do_vmi_align_munmap
      0.00        +2.2e+99%       0.00 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.down_write.anon_vma_clone.__split_vma.vma_modify
      0.00       +2.6e+101%       0.26 ±201%  perf-sched.wait_time.avg.ms.__cond_resched.down_write.do_brk_flags.__do_sys_brk.do_syscall_64
      0.01 ±204%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.down_write.free_pgtables.exit_mmap.__mmput
      0.00 ±300%   +1245.2%       0.04 ±299%  perf-sched.wait_time.avg.ms.__cond_resched.down_write.mmap_region.do_mmap.vm_mmap_pgoff
      0.05 ±194%    +126.6%       0.12 ±223%  perf-sched.wait_time.avg.ms.__cond_resched.down_write.unlink_anon_vmas.free_pgtables.unmap_region
      0.00 ±299%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.down_write.unlink_file_vma.free_pgtables.exit_mmap
      0.07 ±125%     -33.7%       0.05 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.down_write.vma_prepare.__split_vma.do_vmi_align_munmap
      0.02 ±300%    +723.8%       0.19 ±259%  perf-sched.wait_time.avg.ms.__cond_resched.down_write.vma_prepare.do_brk_flags.__do_sys_brk
      0.25 ± 76%    +104.2%       0.51 ±114%  perf-sched.wait_time.avg.ms.__cond_resched.down_write_killable.__do_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
      0.32 ±194%     -79.5%       0.07 ±248%  perf-sched.wait_time.avg.ms.__cond_resched.down_write_killable.exec_mmap.begin_new_exec.load_elf_binary
      0.00 ±299%  +12272.7%       0.41 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.down_write_killable.setup_arg_pages.load_elf_binary.search_binary_handler
      0.21 ±299%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.down_write_killable.vm_mmap_pgoff.elf_map.load_elf_binary
      0.00 ±300%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.down_write_killable.vm_mmap_pgoff.elf_map.load_elf_interp
      0.38 ±300%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.down_write_killable.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64
      0.08 ±198%  +17398.4%      14.54 ±294%  perf-sched.wait_time.avg.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64
      0.01 ±122%    +290.4%       0.04 ±282%  perf-sched.wait_time.avg.ms.__cond_resched.dput.step_into.link_path_walk.part
      0.58 ±206%     -99.7%       0.00 ±299%  perf-sched.wait_time.avg.ms.__cond_resched.dput.terminate_walk.path_openat.do_filp_open
      0.00 ±201%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.exit_mmap.__mmput.exit_mm.do_exit
      0.00 ±299%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.filemap_read.__kernel_read.search_binary_handler.exec_binprm
      0.00 ±200%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.filemap_read.vfs_read.ksys_read.do_syscall_64
      0.20 ±300%   +1341.7%       2.87 ±188%  perf-sched.wait_time.avg.ms.__cond_resched.generic_perform_write.shmem_file_write_iter.vfs_write.ksys_write
      0.00 ±300%   +2110.0%       0.02 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.khugepaged.kthread.ret_from_fork.ret_from_fork_asm
      0.00 ±300%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.__anon_vma_prepare.do_anonymous_page.__handle_mm_fault
      0.00       +3.9e+101%       0.39 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.__anon_vma_prepare.wp_page_copy.__handle_mm_fault
      0.00         +5e+103%      50.14 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.acpi_ut_create_generic_state.acpi_ds_result_push.acpi_ds_exec_end_op
      0.42 ±154%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.alloc_pid.copy_process.kernel_clone
      0.20 ± 30%    +241.4%       0.68 ±123%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.mas_alloc_nodes.mas_preallocate.__split_vma
      0.21 ±128%    +117.3%       0.46 ±137%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.mas_alloc_nodes.mas_preallocate.do_brk_flags
      0.16 ±300%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.mas_alloc_nodes.mas_preallocate.mmap_region
      0.00        +2.4e+99%       0.00 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.mas_alloc_nodes.mas_preallocate.vma_expand
      0.00 ±299%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.security_file_alloc.init_file.alloc_empty_file
      0.00 ±300%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.vm_area_alloc.__install_special_mapping.map_vdso
      0.32 ±300%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.vm_area_alloc.do_brk_flags.__do_sys_brk
      0.10 ±299%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.vm_area_alloc.mmap_region.do_mmap
      0.32 ± 65%     +94.9%       0.62 ±100%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.vm_area_dup.__split_vma.do_vmi_align_munmap
      0.00       +1.7e+101%       0.17 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.vm_area_dup.__split_vma.vma_modify
      0.00         +1e+104%     100.01 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.vm_area_dup.dup_mmap.dup_mm
      0.12 ± 94%     -11.6%       0.10 ±152%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc_bulk.mas_alloc_nodes.mas_preallocate.__split_vma
      0.00 ±300%  +24600.0%       0.42 ±250%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc_bulk.mas_alloc_nodes.mas_preallocate.do_brk_flags
      0.00       +2.1e+103%      20.73 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.mutex_lock.drm_property_free_blob.drm_gem_destroy_shadow_plane_state.drm_atomic_state_default_clear
      0.00 ±300%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.mutex_lock.futex_exit_release.exit_mm_release.exit_mm
      0.00 ±200%     -47.7%       0.00 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.mutex_lock.perf_event_ctx_lock_nested.constprop.0
      2.39 ± 81%     -18.7%       1.94 ±100%  perf-sched.wait_time.avg.ms.__cond_resched.mutex_lock.perf_poll.do_poll.constprop
      0.00       +7.7e+100%       0.08 ±299%  perf-sched.wait_time.avg.ms.__cond_resched.mutex_lock_killable.pcpu_alloc.mm_init.alloc_bprm
      0.00         +3e+104%     300.07 ±152%  perf-sched.wait_time.avg.ms.__cond_resched.mutex_lock_killable.pcpu_alloc.mm_init.dup_mm
    204.66 ±300%      +0.1%     204.80 ±200%  perf-sched.wait_time.avg.ms.__cond_resched.process_one_work.worker_thread.kthread.ret_from_fork
      0.00         +2e+104%     204.66 ±200%  perf-sched.wait_time.avg.ms.__cond_resched.refresh_cpu_vm_stats.vmstat_update.process_one_work.worker_thread
      0.20 ±122%    +125.5%       0.44 ±244%  perf-sched.wait_time.avg.ms.__cond_resched.remove_vma.do_vmi_align_munmap.__do_sys_brk.do_syscall_64
      0.00 ±300%  +1.3e+06%      57.48 ±143%  perf-sched.wait_time.avg.ms.__cond_resched.shmem_get_folio_gfp.shmem_read_folio_gfp.drm_gem_get_pages.drm_gem_shmem_get_pages
      0.15 ±300%    +362.8%       0.69 ±202%  perf-sched.wait_time.avg.ms.__cond_resched.shmem_inode_acct_blocks.shmem_alloc_and_add_folio.shmem_get_folio_gfp.shmem_write_begin
      0.00        +2.8e+99%       0.00 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.slab_pre_alloc_hook.constprop.0.kmem_cache_alloc_lru
     39.66 ± 14%     +56.8%      62.18 ± 16%  perf-sched.wait_time.avg.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      0.02 ±106%    +448.4%       0.08 ±146%  perf-sched.wait_time.avg.ms.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part
      0.00 ±300%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.switch_task_namespaces.do_exit.do_group_exit.__x64_sys_exit_group
      0.04 ± 89%     +47.0%       0.07 ±174%  perf-sched.wait_time.avg.ms.__cond_resched.task_work_run.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode
      0.00       +4.6e+100%       0.05 ±271%  perf-sched.wait_time.avg.ms.__cond_resched.task_work_run.exit_to_user_mode_loop.exit_to_user_mode_prepare.syscall_exit_to_user_mode
      0.00         +1e+104%     100.06 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.task_work_run.get_signal.arch_do_signal_or_restart.exit_to_user_mode_loop
      0.40 ±295%     -99.0%       0.00 ±209%  perf-sched.wait_time.avg.ms.__cond_resched.tlb_batch_pages_flush.tlb_finish_mmu.exit_mmap.__mmput
      0.28 ± 15%    +483.6%       1.63 ± 48%  perf-sched.wait_time.avg.ms.__cond_resched.tlb_batch_pages_flush.tlb_finish_mmu.unmap_region.constprop
      0.02 ±300%   +1253.6%       0.30 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.unmap_page_range.unmap_vmas.unmap_region.constprop
      0.28 ± 28%    +193.3%       0.81 ± 66%  perf-sched.wait_time.avg.ms.__cond_resched.unmap_vmas.unmap_region.constprop.0
      0.00 ±214%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.wp_page_copy.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
      5.09 ±300%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.ww_mutex_lock.drm_gem_vunmap_unlocked.drm_fbdev_generic_helper_fb_dirty.drm_fb_helper_damage_work
      0.00       +3.7e+100%       0.04 ±253%  perf-sched.wait_time.avg.ms.__cond_resched.ww_mutex_lock.drm_gem_vunmap_unlocked.drm_gem_fb_vunmap.drm_atomic_helper_cleanup_planes
      0.27 ± 29%   +9684.8%      26.29 ±195%  perf-sched.wait_time.avg.ms.__cond_resched.zap_pmd_range.isra.0.unmap_page_range
    271.34 ± 67%     -69.9%      81.76 ±152%  perf-sched.wait_time.avg.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
      2.37 ± 40%     -67.9%       0.76 ±154%  perf-sched.wait_time.avg.ms.devkmsg_read.vfs_read.ksys_read.do_syscall_64
    214.20 ± 86%     -29.4%     151.17 ±104%  perf-sched.wait_time.avg.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
    235.79 ±  5%      +1.3%     238.78 ±  5%  perf-sched.wait_time.avg.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
      2.94 ±  4%     +30.9%       3.84 ± 85%  perf-sched.wait_time.avg.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
      0.00       +2.6e+101%       0.26 ±300%  perf-sched.wait_time.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_common_interrupt
      0.26 ± 22%    +375.6%       1.22 ± 80%  perf-sched.wait_time.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_exc_page_fault
      0.24 ±  6%    +311.6%       1.00 ± 46%  perf-sched.wait_time.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt
      0.13 ± 18%      -9.5%       0.11 ± 29%  perf-sched.wait_time.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_call_function_single
      0.24 ±  6%    +338.2%       1.07 ± 53%  perf-sched.wait_time.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi
     27.55 ± 36%     -21.7%      21.58 ± 34%  perf-sched.wait_time.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.syscall_exit_to_user_mode.do_syscall_64
      0.73 ±300%    -100.0%       0.00        perf-sched.wait_time.avg.ms.futex_wait_queue.futex_wait.do_futex.__x64_sys_futex
      0.00 ±300%  +32366.7%       0.39 ±300%  perf-sched.wait_time.avg.ms.io_schedule.folio_wait_bit_common.filemap_fault.__do_fault
    169.88 ± 16%     -30.1%     118.71 ± 28%  perf-sched.wait_time.avg.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
    551.67 ±  5%      +1.2%     558.07 ±  2%  perf-sched.wait_time.avg.ms.schedule_hrtimeout_range_clock.do_poll.constprop.0.do_sys_poll
    370.03 ±  4%      +4.6%     386.92 ±  3%  perf-sched.wait_time.avg.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
    319.32 ±299%    -100.0%       0.00        perf-sched.wait_time.avg.ms.schedule_preempt_disabled.rwsem_down_read_slowpath.down_read_killable.__access_remote_vm
      0.01 ±122%      +6.5%       0.01 ±123%  perf-sched.wait_time.avg.ms.schedule_timeout.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr
      2.65 ± 10%     +17.7%       3.12 ± 11%  perf-sched.wait_time.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
    452.77            +0.1%     453.32        perf-sched.wait_time.avg.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
      0.18 ±300%    +102.3%       0.36 ±299%  perf-sched.wait_time.avg.ms.schedule_timeout.khugepaged_wait_work.khugepaged.kthread
      3.63            -5.5%       3.43 ±  5%  perf-sched.wait_time.avg.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
    121.37            +2.6%     124.56        perf-sched.wait_time.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      3.83 ± 19%     -71.1%       1.11 ±123%  perf-sched.wait_time.avg.ms.syslog_print.do_syslog.kmsg_read.vfs_read
      0.44 ± 43%     -21.7%       0.34 ± 40%  perf-sched.wait_time.avg.ms.wait_for_partner.fifo_open.do_dentry_open.do_open
    724.24 ±  2%      +1.6%     735.84        perf-sched.wait_time.avg.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
      0.83 ± 62%    +573.9%       5.62 ± 59%  perf-sched.wait_time.max.ms.__cond_resched.__alloc_pages.__folio_alloc.vma_alloc_folio.do_anonymous_page
      0.30 ±300%    +166.9%       0.80 ±228%  perf-sched.wait_time.max.ms.__cond_resched.__alloc_pages.__folio_alloc.vma_alloc_folio.shmem_alloc_folio
      0.00 ±300%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.__alloc_pages.__folio_alloc.vma_alloc_folio.wp_page_copy
      0.00       +1.2e+102%       1.19 ±225%  perf-sched.wait_time.max.ms.__cond_resched.__alloc_pages.__pud_alloc.alloc_new_pud.constprop
      0.40 ±300%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.__alloc_pages.pte_alloc_one.__do_fault.do_read_fault
      0.00       +8.8e+101%       0.88 ±300%  perf-sched.wait_time.max.ms.__cond_resched.__alloc_pages.pte_alloc_one.__pte_alloc.do_anonymous_page
      2.19 ±124%     -77.3%       0.50 ±300%  perf-sched.wait_time.max.ms.__cond_resched.__do_fault.do_read_fault.do_fault.__handle_mm_fault
     40.85 ±300%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.__kmem_cache_alloc_node.__kmalloc.acpi_ex_allocate_name_string.acpi_ex_get_name_string
      0.00          +2e+99%       0.00 ±300%  perf-sched.wait_time.max.ms.__cond_resched.__kmem_cache_alloc_node.__kmalloc_node.alloc_cpumask_var_node.__sched_setaffinity
      0.00       +5.5e+101%       0.55 ±300%  perf-sched.wait_time.max.ms.__cond_resched.__kmem_cache_alloc_node.__kmalloc_node.memcg_alloc_slab_cgroups.allocate_slab
      0.00 ±300%    +104.2%       0.00 ±200%  perf-sched.wait_time.max.ms.__cond_resched.__kmem_cache_alloc_node.kmalloc_trace.perf_event_mmap_event.perf_event_mmap
      0.00        +3.9e+99%       0.00 ±200%  perf-sched.wait_time.max.ms.__cond_resched.__kmem_cache_alloc_node.kmalloc_trace.single_open.single_open_size
      0.00         +1e+104%     100.01 ±299%  perf-sched.wait_time.max.ms.__cond_resched.__kmem_cache_alloc_node.kmalloc_trace.vmstat_start.seq_read_iter
      4898            -8.1%       4504 ± 25%  perf-sched.wait_time.max.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
      0.84 ±132%     -63.0%       0.31 ±194%  perf-sched.wait_time.max.ms.__cond_resched.__wait_for_common.wait_for_completion_state.kernel_clone.__x64_sys_vfork
      0.15 ±300%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.cgroup_css_set_fork.cgroup_can_fork.copy_process.kernel_clone
      0.00 ±300%   +1100.0%       0.03 ±300%  perf-sched.wait_time.max.ms.__cond_resched.change_pmd_range.change_p4d_range.change_protection_range.mprotect_fixup
      0.00 ±300%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.dentry_kill.dput.step_into.link_path_walk
      0.16 ±148%    +731.9%       1.33 ±146%  perf-sched.wait_time.max.ms.__cond_resched.down_write.__split_vma.do_vmi_align_munmap.__do_sys_brk
      0.52 ± 53%    +238.0%       1.75 ±108%  perf-sched.wait_time.max.ms.__cond_resched.down_write.anon_vma_clone.__split_vma.do_vmi_align_munmap
      0.00        +2.2e+99%       0.00 ±300%  perf-sched.wait_time.max.ms.__cond_resched.down_write.anon_vma_clone.__split_vma.vma_modify
      0.00       +2.9e+101%       0.29 ±178%  perf-sched.wait_time.max.ms.__cond_resched.down_write.do_brk_flags.__do_sys_brk.do_syscall_64
      0.01 ±212%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.down_write.free_pgtables.exit_mmap.__mmput
      0.00 ±300%   +1245.2%       0.04 ±299%  perf-sched.wait_time.max.ms.__cond_resched.down_write.mmap_region.do_mmap.vm_mmap_pgoff
      0.06 ±199%    +321.8%       0.24 ±258%  perf-sched.wait_time.max.ms.__cond_resched.down_write.unlink_anon_vmas.free_pgtables.unmap_region
      0.00 ±299%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.down_write.unlink_file_vma.free_pgtables.exit_mmap
      0.08 ±119%     -37.4%       0.05 ±300%  perf-sched.wait_time.max.ms.__cond_resched.down_write.vma_prepare.__split_vma.do_vmi_align_munmap
      0.02 ±300%    +763.4%       0.20 ±247%  perf-sched.wait_time.max.ms.__cond_resched.down_write.vma_prepare.do_brk_flags.__do_sys_brk
      0.48 ±115%     +96.0%       0.94 ± 98%  perf-sched.wait_time.max.ms.__cond_resched.down_write_killable.__do_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
      0.32 ±192%     -12.3%       0.28 ±276%  perf-sched.wait_time.max.ms.__cond_resched.down_write_killable.exec_mmap.begin_new_exec.load_elf_binary
      0.00 ±299%  +12272.7%       0.41 ±300%  perf-sched.wait_time.max.ms.__cond_resched.down_write_killable.setup_arg_pages.load_elf_binary.search_binary_handler
      0.21 ±299%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.down_write_killable.vm_mmap_pgoff.elf_map.load_elf_binary
      0.00 ±300%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.down_write_killable.vm_mmap_pgoff.elf_map.load_elf_interp
      0.38 ±300%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.down_write_killable.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64
      0.14 ±230%  +71629.0%     101.64 ±294%  perf-sched.wait_time.max.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64
      0.01 ±122%    +290.4%       0.04 ±282%  perf-sched.wait_time.max.ms.__cond_resched.dput.step_into.link_path_walk.part
      0.58 ±206%     -99.7%       0.00 ±299%  perf-sched.wait_time.max.ms.__cond_resched.dput.terminate_walk.path_openat.do_filp_open
      0.00 ±201%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.exit_mmap.__mmput.exit_mm.do_exit
      0.00 ±299%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.filemap_read.__kernel_read.search_binary_handler.exec_binprm
      0.00 ±200%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.filemap_read.vfs_read.ksys_read.do_syscall_64
      0.40 ±300%    +961.6%       4.23 ±219%  perf-sched.wait_time.max.ms.__cond_resched.generic_perform_write.shmem_file_write_iter.vfs_write.ksys_write
      0.00 ±300%   +4310.0%       0.04 ±300%  perf-sched.wait_time.max.ms.__cond_resched.khugepaged.kthread.ret_from_fork.ret_from_fork_asm
      0.00 ±300%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.__anon_vma_prepare.do_anonymous_page.__handle_mm_fault
      0.00       +3.9e+101%       0.39 ±300%  perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.__anon_vma_prepare.wp_page_copy.__handle_mm_fault
      0.00         +5e+103%      50.14 ±300%  perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.acpi_ut_create_generic_state.acpi_ds_result_push.acpi_ds_exec_end_op
      0.42 ±154%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.alloc_pid.copy_process.kernel_clone
      0.53 ±101%    +124.1%       1.19 ± 90%  perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.mas_alloc_nodes.mas_preallocate.__split_vma
      0.28 ±154%    +291.2%       1.11 ±157%  perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.mas_alloc_nodes.mas_preallocate.do_brk_flags
      0.16 ±300%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.mas_alloc_nodes.mas_preallocate.mmap_region
      0.00        +2.4e+99%       0.00 ±300%  perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.mas_alloc_nodes.mas_preallocate.vma_expand
      0.00 ±299%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.security_file_alloc.init_file.alloc_empty_file
      0.00 ±300%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.vm_area_alloc.__install_special_mapping.map_vdso
      0.32 ±300%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.vm_area_alloc.do_brk_flags.__do_sys_brk
      0.10 ±299%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.vm_area_alloc.mmap_region.do_mmap
      1.20 ± 83%     +66.0%       2.00 ±101%  perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.vm_area_dup.__split_vma.do_vmi_align_munmap
      0.00       +1.7e+101%       0.17 ±300%  perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.vm_area_dup.__split_vma.vma_modify
      0.00         +1e+104%     100.01 ±300%  perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.vm_area_dup.dup_mmap.dup_mm
      0.12 ± 92%     -16.6%       0.10 ±152%  perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc_bulk.mas_alloc_nodes.mas_preallocate.__split_vma
      0.00 ±300%  +38405.9%       0.65 ±244%  perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc_bulk.mas_alloc_nodes.mas_preallocate.do_brk_flags
      0.00       +2.1e+103%      20.73 ±300%  perf-sched.wait_time.max.ms.__cond_resched.mutex_lock.drm_property_free_blob.drm_gem_destroy_shadow_plane_state.drm_atomic_state_default_clear
      0.00 ±300%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.mutex_lock.futex_exit_release.exit_mm_release.exit_mm
      0.00 ±200%     -47.7%       0.00 ±300%  perf-sched.wait_time.max.ms.__cond_resched.mutex_lock.perf_event_ctx_lock_nested.constprop.0
      2.39 ± 81%     -18.7%       1.94 ±100%  perf-sched.wait_time.max.ms.__cond_resched.mutex_lock.perf_poll.do_poll.constprop
      0.00       +4.6e+101%       0.46 ±300%  perf-sched.wait_time.max.ms.__cond_resched.mutex_lock_killable.pcpu_alloc.mm_init.alloc_bprm
      0.00         +3e+104%     300.07 ±152%  perf-sched.wait_time.max.ms.__cond_resched.mutex_lock_killable.pcpu_alloc.mm_init.dup_mm
    204.66 ±300%      +0.1%     204.80 ±200%  perf-sched.wait_time.max.ms.__cond_resched.process_one_work.worker_thread.kthread.ret_from_fork
      0.00         +2e+104%     204.66 ±200%  perf-sched.wait_time.max.ms.__cond_resched.refresh_cpu_vm_stats.vmstat_update.process_one_work.worker_thread
      0.32 ±126%     +45.0%       0.46 ±237%  perf-sched.wait_time.max.ms.__cond_resched.remove_vma.do_vmi_align_munmap.__do_sys_brk.do_syscall_64
      0.00 ±300%  +1.6e+06%      73.17 ±128%  perf-sched.wait_time.max.ms.__cond_resched.shmem_get_folio_gfp.shmem_read_folio_gfp.drm_gem_get_pages.drm_gem_shmem_get_pages
      0.30 ±300%    +244.1%       1.02 ±220%  perf-sched.wait_time.max.ms.__cond_resched.shmem_inode_acct_blocks.shmem_alloc_and_add_folio.shmem_get_folio_gfp.shmem_write_begin
      0.00        +2.8e+99%       0.00 ±300%  perf-sched.wait_time.max.ms.__cond_resched.slab_pre_alloc_hook.constprop.0.kmem_cache_alloc_lru
     86.01 ±  6%      +5.5%      90.72 ±  4%  perf-sched.wait_time.max.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      1.75 ±105%     +80.5%       3.16 ±114%  perf-sched.wait_time.max.ms.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part
      0.00 ±300%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.switch_task_namespaces.do_exit.do_group_exit.__x64_sys_exit_group
      0.10 ±100%    +248.6%       0.35 ±258%  perf-sched.wait_time.max.ms.__cond_resched.task_work_run.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode
      0.00       +4.6e+100%       0.05 ±271%  perf-sched.wait_time.max.ms.__cond_resched.task_work_run.exit_to_user_mode_loop.exit_to_user_mode_prepare.syscall_exit_to_user_mode
      0.00         +1e+104%     100.06 ±300%  perf-sched.wait_time.max.ms.__cond_resched.task_work_run.get_signal.arch_do_signal_or_restart.exit_to_user_mode_loop
      0.40 ±295%     -98.7%       0.01 ±200%  perf-sched.wait_time.max.ms.__cond_resched.tlb_batch_pages_flush.tlb_finish_mmu.exit_mmap.__mmput
      2.03 ± 23%    +227.5%       6.66 ± 48%  perf-sched.wait_time.max.ms.__cond_resched.tlb_batch_pages_flush.tlb_finish_mmu.unmap_region.constprop
      0.02 ±300%   +1253.6%       0.30 ±300%  perf-sched.wait_time.max.ms.__cond_resched.unmap_page_range.unmap_vmas.unmap_region.constprop
      1.59 ± 53%    +153.9%       4.04 ± 57%  perf-sched.wait_time.max.ms.__cond_resched.unmap_vmas.unmap_region.constprop.0
      0.00 ±200%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.wp_page_copy.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
      5.09 ±300%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.ww_mutex_lock.drm_gem_vunmap_unlocked.drm_fbdev_generic_helper_fb_dirty.drm_fb_helper_damage_work
      0.00       +3.7e+100%       0.04 ±253%  perf-sched.wait_time.max.ms.__cond_resched.ww_mutex_lock.drm_gem_vunmap_unlocked.drm_gem_fb_vunmap.drm_atomic_helper_cleanup_planes
      1.18 ± 46%  +17108.0%     202.73 ±197%  perf-sched.wait_time.max.ms.__cond_resched.zap_pmd_range.isra.0.unmap_page_range
    802.26 ± 49%     -49.8%     402.51 ±120%  perf-sched.wait_time.max.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
      4.73 ± 40%     -67.8%       1.52 ±154%  perf-sched.wait_time.max.ms.devkmsg_read.vfs_read.ksys_read.do_syscall_64
    702.00 ± 65%     -14.3%     601.71 ± 81%  perf-sched.wait_time.max.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
      1321 ± 48%     +37.2%       1813 ± 54%  perf-sched.wait_time.max.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
      6.90 ±  9%   +1483.8%     109.24 ±272%  perf-sched.wait_time.max.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
      0.00       +2.6e+101%       0.26 ±300%  perf-sched.wait_time.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_common_interrupt
      2.33 ± 39%   +4436.5%     105.58 ±282%  perf-sched.wait_time.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_exc_page_fault
    128.51 ±225%      +2.5%     131.68 ±219%  perf-sched.wait_time.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt
      2.44 ± 12%     +89.7%       4.63 ± 50%  perf-sched.wait_time.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_call_function_single
      5.66 ±156%     +54.5%       8.74 ± 54%  perf-sched.wait_time.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi
      1000           +38.5%       1386 ± 83%  perf-sched.wait_time.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.syscall_exit_to_user_mode.do_syscall_64
      1.41 ±300%    -100.0%       0.00        perf-sched.wait_time.max.ms.futex_wait_queue.futex_wait.do_futex.__x64_sys_futex
      0.00 ±300%  +16133.3%       0.39 ±300%  perf-sched.wait_time.max.ms.io_schedule.folio_wait_bit_common.filemap_fault.__do_fault
      1051            +0.8%       1059        perf-sched.wait_time.max.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
      1004            -0.0%       1003        perf-sched.wait_time.max.ms.schedule_hrtimeout_range_clock.do_poll.constprop.0.do_sys_poll
    499.84            +0.0%     499.86        perf-sched.wait_time.max.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
    319.32 ±299%    -100.0%       0.00        perf-sched.wait_time.max.ms.schedule_preempt_disabled.rwsem_down_read_slowpath.down_read_killable.__access_remote_vm
      0.01 ±122%      +4.8%       0.01 ±123%  perf-sched.wait_time.max.ms.schedule_timeout.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr
      8.87 ±  6%     +50.9%      13.38 ± 50%  perf-sched.wait_time.max.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
    504.57            +0.1%     505.25        perf-sched.wait_time.max.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
      0.18 ±300%    +102.3%       0.36 ±299%  perf-sched.wait_time.max.ms.schedule_timeout.khugepaged_wait_work.khugepaged.kthread
     14.61 ± 99%      +0.9%      14.74 ± 73%  perf-sched.wait_time.max.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
      1053           +10.8%       1167 ± 26%  perf-sched.wait_time.max.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      7.67 ± 19%     -71.1%       2.22 ±123%  perf-sched.wait_time.max.ms.syslog_print.do_syslog.kmsg_read.vfs_read
      3.55 ± 12%     +13.4%       4.03 ± 29%  perf-sched.wait_time.max.ms.wait_for_partner.fifo_open.do_dentry_open.do_open
      2046            +0.0%       2046        perf-sched.wait_time.max.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm

xsang@inn:~$
xsang@inn:~$ lkp ncompare -s commit=1f4f7f0f8845d/tbox_group=lkp-skl-d08/testcase=stress-ng -o commit=6ccdcb6d3a741  -a | grep -v interrupts | grep -v softirq
=========================================================================================
class/compiler/cpufreq_governor/disk/fs/kconfig/nr_threads/rootfs/tbox_group/test/testcase/testtime:
  filesystem/gcc-12/performance/1SSD/btrfs/x86_64-rhel-8.3/10%/debian-11.1-x86_64-20220510.cgz/lkp-skl-d08/file-ioctl/stress-ng/60s

commit:
  1f4f7f0f8845dbac40289cc3d50b81314c5a12b8
  6ccdcb6d3a741c4e005ca6ffd4a62ddf8b5bead3

1f4f7f0f8845dbac 6ccdcb6d3a741c4e005ca6ffd4a
---------------- ---------------------------
         %stddev     %change         %stddev
             \          |                \
     10.99            +0.0%      10.99        kmsg.timestamp:EDAC_skx:ECC_is_disabled_on_imc
     75.12            +2.0%      76.60 ±  3%  kmsg.timestamp:last
     26.34 ±  2%      +5.6%      27.82 ± 10%  kmsg.timestamp:x86/PAT:bmc-watchdog:#map_pfn_expected_mapping_type_uncached-minus_for[mem#-#],got_write-back
     75.12            +2.0%      76.60 ±  3%  dmesg.timestamp:last
 2.143e+09            +5.1%  2.252e+09 ±  9%  cpuidle..time
   3124995            +3.6%    3237684 ±  7%  cpuidle..usage
     91.83            +5.0%      96.44 ±  7%  uptime.boot
      3043            +5.3%       3205 ±  8%  uptime.idle
     24.14            -0.2%      24.08        boot-time.boot
     10.14            -0.3%      10.11        boot-time.dhcp
    826.15            -0.3%     823.94        boot-time.idle
     12.62 ± 40%     +17.2%      14.80 ± 44%  perf-c2c.DRAM.local
    130.38 ± 10%     +26.7%     165.20 ± 16%  perf-c2c.HITM.local
    130.38 ± 10%     +26.7%     165.20 ± 16%  perf-c2c.HITM.total
     91.17            +0.2%      91.37        iostat.cpu.idle
      0.05           +39.9%       0.07 ± 38%  iostat.cpu.iowait
      8.11            -1.8%       7.96 ±  5%  iostat.cpu.system
      0.67 ±  3%     -11.3%       0.59 ± 14%  iostat.cpu.user
     90.99            +0.2       91.21        mpstat.cpu.all.idle%
      0.05            +0.0        0.07 ± 38%  mpstat.cpu.all.iowait%
      0.79 ±  3%      +0.1        0.87 ± 20%  mpstat.cpu.all.irq%
      0.04 ±  5%      +0.0        0.05 ± 20%  mpstat.cpu.all.soft%
      7.48            -0.2        7.23 ±  8%  mpstat.cpu.all.sys%
      0.65 ±  3%      -0.1        0.58 ± 14%  mpstat.cpu.all.usr%
     62.04            +0.0%      62.04        time.elapsed_time
     62.04            +0.0%      62.04        time.elapsed_time.max
      2112            +0.0%       2112        time.file_system_inputs
     96.00            +0.0%      96.00        time.file_system_outputs
      6184 ±  7%     -23.0%       4761 ±  5%  time.involuntary_context_switches
      0.62 ±210%     -52.0%       0.30 ±213%  time.major_page_faults
      6912            +0.0%       6912        time.maximum_resident_set_size
      9509            -0.1%       9496        time.minor_page_faults
      4096            +0.0%       4096        time.page_size
    286.00            +0.0%     286.00        time.percent_of_cpu_this_job_got
    172.84            +0.3%     173.27        time.system_time
      4.68            -8.0%       4.31        time.user_time
    123062 ±  5%      -8.5%     112627 ±  5%  time.voluntary_context_switches
     91.20            +0.2%      91.42        vmstat.cpu.id
      8.25            -2.1%       8.08 ±  5%  vmstat.cpu.sy
      0.49 ±  5%     -12.7%       0.43 ± 20%  vmstat.cpu.us
      0.05 ± 13%     +32.8%       0.07 ± 36%  vmstat.cpu.wa
     15.91            -3.6%      15.33 ±  8%  vmstat.io.bi
      6.65 ± 76%     +49.4%       9.93 ± 69%  vmstat.io.bo
      1034            +0.1%       1035        vmstat.memory.buff
   2954092            -0.0%    2953196        vmstat.memory.cache
  28985158            +0.0%   28997431        vmstat.memory.free
      0.02 ± 77%     +79.8%       0.03 ± 55%  vmstat.procs.b
      3.19 ±  2%      -3.3%       3.08 ±  8%  vmstat.procs.r
     33719            -4.5%      32188 ±  7%  vmstat.system.cs
     51824            -1.1%      51271 ±  2%  vmstat.system.in
    340971            -6.9%     317411        stress-ng.file-ioctl.ops
      5682            -6.9%       5290        stress-ng.file-ioctl.ops_per_sec
     62.04            +0.0%      62.04        stress-ng.time.elapsed_time
     62.04            +0.0%      62.04        stress-ng.time.elapsed_time.max
      2112            +0.0%       2112        stress-ng.time.file_system_inputs
     96.00            +0.0%      96.00        stress-ng.time.file_system_outputs
      6184 ±  7%     -23.0%       4761 ±  5%  stress-ng.time.involuntary_context_switches
      0.62 ±210%     -52.0%       0.30 ±213%  stress-ng.time.major_page_faults
      6912            +0.0%       6912        stress-ng.time.maximum_resident_set_size
      9509            -0.1%       9496        stress-ng.time.minor_page_faults
      4096            +0.0%       4096        stress-ng.time.page_size
    286.00            +0.0%     286.00        stress-ng.time.percent_of_cpu_this_job_got
    172.84            +0.3%     173.27        stress-ng.time.system_time
      4.68            -8.0%       4.31        stress-ng.time.user_time
    123062 ±  5%      -8.5%     112627 ±  5%  stress-ng.time.voluntary_context_switches
    357.50            -3.4%     345.30 ±  6%  turbostat.Avg_MHz
      9.56            -0.2        9.38 ±  3%  turbostat.Busy%
      3740            -1.8%       3672 ±  3%  turbostat.Bzy_MHz
     61848            -0.6%      61459        turbostat.C1
      0.14 ±  6%      +0.0        0.14 ± 10%  turbostat.C1%
    926012            -0.1%     924627        turbostat.C1E
      2.70            -0.1        2.58 ±  6%  turbostat.C1E%
   2097927            +5.4%    2211048 ± 10%  turbostat.C6
     87.94            +0.4       88.29        turbostat.C6%
     88.61            -1.3%      87.45 ±  3%  turbostat.CPU%c1
      1.83 ± 21%     +72.6%       3.16 ± 95%  turbostat.CPU%c6
     34.12            +4.9%      35.80 ±  3%  turbostat.CoreTmp
      0.43            -2.7%       0.42 ±  2%  turbostat.IPC
   3473046            +3.2%    3585079 ±  6%  turbostat.IRQ
     35417 ±  9%      +3.2%      36563 ±  9%  turbostat.POLL
      0.00 ±264%      -0.0        0.00        turbostat.POLL%
      1.58 ± 24%     +59.7%       2.53 ± 95%  turbostat.Pkg%pc2
     33.75 ±  3%      +3.4%      34.90 ±  2%  turbostat.PkgTmp
     87.13            -2.7%      84.80 ±  5%  turbostat.PkgWatt
      0.86            +1.5%       0.87        turbostat.RAMWatt
      3001            -0.0%       3000        turbostat.TSC_MHz
     69867            -0.7%      69398 ±  5%  meminfo.Active
     66909            -0.6%      66489 ±  5%  meminfo.Active(anon)
      2958            -1.7%       2908 ±  6%  meminfo.Active(file)
     55351 ±  9%     +11.3%      61587 ± 17%  meminfo.AnonHugePages
    297646            -0.3%     296691        meminfo.AnonPages
      1034            +0.1%       1035        meminfo.Buffers
   2874923            -0.0%    2873986        meminfo.Cached
  16281020            +0.0%   16281020        meminfo.CommitLimit
    566106            -1.1%     559898        meminfo.Committed_AS
  27525120 ±  3%      -0.2%   27472691 ±  3%  meminfo.DirectMap1G
   5554944 ± 18%      +1.0%    5609062 ± 16%  meminfo.DirectMap2M
    166176 ±  7%      -1.0%     164486 ± 11%  meminfo.DirectMap4k
     62.51 ± 16%     -10.6%      55.88        meminfo.Dirty
      2048            +0.0%       2048        meminfo.Hugepagesize
    307749            -0.5%     306284        meminfo.Inactive
    305087            -0.5%     303694        meminfo.Inactive(anon)
      2662 ±  2%      -2.7%       2589 ±  3%  meminfo.Inactive(file)
     78478            +0.1%      78579        meminfo.KReclaimable
      9050            +0.3%       9075        meminfo.KernelStack
     41083            -1.2%      40591 ±  2%  meminfo.Mapped
  28759154            +0.0%   28770515        meminfo.MemAvailable
  28981905            +0.0%   28993284        meminfo.MemFree
  32562044            +0.0%   32562044        meminfo.MemTotal
   3580138            -0.3%    3568759        meminfo.Memused
     10.86            -3.3%      10.50 ±  8%  meminfo.Mlocked
      5273            -1.0%       5220 ±  2%  meminfo.PageTables
     16259            -0.0%      16253        meminfo.Percpu
     78478            +0.1%      78579        meminfo.SReclaimable
     90009            +0.3%      90246        meminfo.SUnreclaim
     74602            -1.1%      73789 ±  4%  meminfo.Shmem
    168487            +0.2%     168825        meminfo.Slab
   2795729            -0.0%    2795724        meminfo.Unevictable
 3.436e+10            +0.0%  3.436e+10        meminfo.VmallocTotal
     20749            +0.2%      20783        meminfo.VmallocUsed
      0.77 ±200%      -6.4%       0.73 ±134%  meminfo.Writeback
   3758317            +0.1%    3762244        meminfo.max_used_kB
     70.38 ±  8%      -3.2%      68.10 ± 13%  proc-vmstat.direct_map_level2_splits
      2.75 ± 35%     -20.0%       2.20 ± 34%  proc-vmstat.direct_map_level3_splits
     16707            -0.5%      16622 ±  5%  proc-vmstat.nr_active_anon
    740.83            -1.9%     727.11 ±  6%  proc-vmstat.nr_active_file
     74446            -0.4%      74173        proc-vmstat.nr_anon_pages
     26.99 ±  9%     +11.4%      30.07 ± 17%  proc-vmstat.nr_anon_transparent_hugepages
    148.88 ± 65%     +53.4%     228.40 ± 86%  proc-vmstat.nr_dirtied
     15.63 ± 16%     -10.6%      13.97        proc-vmstat.nr_dirty
    717010            +0.0%     717280        proc-vmstat.nr_dirty_background_threshold
   1435774            +0.0%    1436316        proc-vmstat.nr_dirty_threshold
    718972            -0.0%     718748        proc-vmstat.nr_file_pages
      0.00         +1e+101%       0.10 ±300%  proc-vmstat.nr_foll_pin_acquired
      0.00         +1e+101%       0.10 ±300%  proc-vmstat.nr_foll_pin_released
   7245424            +0.0%    7248168        proc-vmstat.nr_free_pages
     76303            -0.5%      75925        proc-vmstat.nr_inactive_anon
    662.04 ±  3%      -3.0%     642.06 ±  3%  proc-vmstat.nr_inactive_file
      9051            +0.3%       9076        proc-vmstat.nr_kernel_stack
     10282            -1.3%      10149 ±  2%  proc-vmstat.nr_mapped
      2.72            -3.5%       2.62 ±  8%  proc-vmstat.nr_mlock
      1318            -1.0%       1305 ±  2%  proc-vmstat.nr_page_table_pages
     18634            -1.0%      18448 ±  4%  proc-vmstat.nr_shmem
     19618            +0.1%      19644        proc-vmstat.nr_slab_reclaimable
     22503            +0.3%      22561        proc-vmstat.nr_slab_unreclaimable
    698932            -0.0%     698931        proc-vmstat.nr_unevictable
      0.19 ±200%      -6.4%       0.18 ±134%  proc-vmstat.nr_writeback
    122.62 ± 66%     +55.4%     190.60 ± 76%  proc-vmstat.nr_written
     16707            -0.5%      16622 ±  5%  proc-vmstat.nr_zone_active_anon
    740.83            -1.9%     727.11 ±  6%  proc-vmstat.nr_zone_active_file
     76303            -0.5%      75925        proc-vmstat.nr_zone_inactive_anon
    662.04 ±  3%      -3.0%     642.06 ±  3%  proc-vmstat.nr_zone_inactive_file
    698932            -0.0%     698931        proc-vmstat.nr_zone_unevictable
     15.58 ± 15%      -9.5%      14.09 ±  2%  proc-vmstat.nr_zone_write_pending
  89530411            -7.1%   83208879        proc-vmstat.numa_hit
      0.00          -100.0%       0.00        proc-vmstat.numa_interleave
  89301961            -7.0%   83009577        proc-vmstat.numa_local
     26399            -2.7%      25675        proc-vmstat.pgactivate
      0.00          -100.0%       0.00        proc-vmstat.pgalloc_dma32
  93091135            -6.9%   86691043        proc-vmstat.pgalloc_normal
      0.00          -100.0%       0.00        proc-vmstat.pgdeactivate
    212409            +1.2%     215026 ±  3%  proc-vmstat.pgfault
  93049512            -6.9%   86647969        proc-vmstat.pgfree
      1056            +0.0%       1056        proc-vmstat.pgpgin
    440.00 ± 76%     +65.1%     726.40 ± 77%  proc-vmstat.pgpgout
      8496            +2.4%       8699 ±  4%  proc-vmstat.pgreuse
      0.00          -100.0%       0.00        proc-vmstat.pgrotated
     46.75 ±  4%      +9.1%      51.00 ± 11%  proc-vmstat.thp_collapse_alloc
      0.00          -100.0%       0.00        proc-vmstat.thp_deferred_split_page
      1.50 ± 74%     -46.7%       0.80 ± 93%  proc-vmstat.thp_fault_alloc
      0.00          -100.0%       0.00        proc-vmstat.thp_split_pmd
      0.00          -100.0%       0.00        proc-vmstat.thp_zero_page_alloc
     14.00            +0.0%      14.00        proc-vmstat.unevictable_pgs_culled
      3.00            +0.0%       3.00        proc-vmstat.unevictable_pgs_mlocked
      3.00            +0.0%       3.00        proc-vmstat.unevictable_pgs_munlocked
      0.13 ± 62%    +299.7%       0.51 ±132%  perf-stat.i.MPKI
 3.868e+09            -6.1%  3.632e+09 ±  8%  perf-stat.i.branch-instructions
      0.69 ± 17%      +0.6        1.26 ± 85%  perf-stat.i.branch-miss-rate%
  21656746 ±  2%      -7.3%   20079062 ±  9%  perf-stat.i.branch-misses
      1.38 ± 12%      +0.2        1.61 ± 23%  perf-stat.i.cache-miss-rate%
   1243295 ± 18%     +17.9%    1465860 ±  9%  perf-stat.i.cache-misses
 2.181e+08 ±  2%      +0.5%  2.192e+08 ±  7%  perf-stat.i.cache-references
     35135            -4.7%      33477 ±  8%  perf-stat.i.context-switches
      0.74 ± 12%     +67.7%       1.24 ± 71%  perf-stat.i.cpi
     36006            -0.0%      36004        perf-stat.i.cpu-clock
 1.247e+10            -3.6%  1.203e+10 ±  6%  perf-stat.i.cpu-cycles
     65.35            -1.7%      64.26 ±  4%  perf-stat.i.cpu-migrations
     24611 ± 31%     -29.1%      17449 ± 17%  perf-stat.i.cycles-between-cache-misses
      0.02 ± 68%      +0.1        0.08 ±143%  perf-stat.i.dTLB-load-miss-rate%
    311590 ±  2%      +4.9%     326828 ± 14%  perf-stat.i.dTLB-load-misses
 4.523e+09            -6.5%  4.227e+09 ±  8%  perf-stat.i.dTLB-loads
      0.00 ± 92%      +0.0        0.01 ±149%  perf-stat.i.dTLB-store-miss-rate%
      9906 ±  6%     +34.2%      13292 ± 45%  perf-stat.i.dTLB-store-misses
 2.698e+09            -6.0%  2.536e+09 ±  8%  perf-stat.i.dTLB-stores
     23.04 ±  4%      +1.5       24.52 ± 17%  perf-stat.i.iTLB-load-miss-rate%
    338747 ± 21%      -9.7%     305803 ± 14%  perf-stat.i.iTLB-load-misses
   1150637 ± 14%      -9.2%    1044691 ±  8%  perf-stat.i.iTLB-loads
 1.885e+10            -5.9%  1.775e+10 ±  8%  perf-stat.i.instructions
     57534 ± 14%      +1.4%      58352 ± 11%  perf-stat.i.instructions-per-iTLB-miss
      1.50            -5.5%       1.41 ±  8%  perf-stat.i.ipc
      0.28 ± 54%     +75.9%       0.49 ± 22%  perf-stat.i.major-faults
      0.35            -3.6%       0.33 ±  6%  perf-stat.i.metric.GHz
     68.15 ± 26%    +110.8%     143.65 ± 93%  perf-stat.i.metric.K/sec
    314.01            -6.2%     294.69 ±  8%  perf-stat.i.metric.M/sec
      1925            -3.0%       1867 ±  4%  perf-stat.i.minor-faults
    109386 ± 22%      -3.9%     105144 ±  7%  perf-stat.i.node-loads
    166651 ± 31%     +85.0%     308368 ±  4%  perf-stat.i.node-stores
      1925            -3.0%       1868 ±  4%  perf-stat.i.page-faults
     36006            -0.0%      36004        perf-stat.i.task-clock
      0.07 ± 19%     +26.7%       0.08 ± 16%  perf-stat.overall.MPKI
      0.56 ±  2%      -0.0        0.55 ±  8%  perf-stat.overall.branch-miss-rate%
      0.57 ± 20%      +0.1        0.67 ± 14%  perf-stat.overall.cache-miss-rate%
      0.66            +2.6%       0.68        perf-stat.overall.cpi
     10338 ± 15%     -19.7%       8301 ± 13%  perf-stat.overall.cycles-between-cache-misses
      0.01 ±  2%      +0.0        0.01 ± 24%  perf-stat.overall.dTLB-load-miss-rate%
      0.00 ±  7%      +0.0        0.00 ± 55%  perf-stat.overall.dTLB-store-miss-rate%
     22.61 ±  3%      +0.0       22.64 ± 10%  perf-stat.overall.iTLB-load-miss-rate%
     57455 ± 14%      +2.2%      58724 ± 11%  perf-stat.overall.instructions-per-iTLB-miss
      1.51            -2.5%       1.47        perf-stat.overall.ipc
 3.808e+09            -6.1%  3.578e+09 ±  8%  perf-stat.ps.branch-instructions
  21311538 ±  2%      -7.2%   19773267 ±  9%  perf-stat.ps.branch-misses
   1223635 ± 18%     +18.0%    1443696 ±  9%  perf-stat.ps.cache-misses
 2.148e+08 ±  2%      +0.5%  2.159e+08 ±  6%  perf-stat.ps.cache-references
     34590            -4.7%      32976 ±  7%  perf-stat.ps.context-switches
     35446            +0.1%      35466        perf-stat.ps.cpu-clock
 1.228e+10            -3.5%  1.185e+10 ±  6%  perf-stat.ps.cpu-cycles
     64.33            -1.6%      63.29 ±  3%  perf-stat.ps.cpu-migrations
    306740 ±  2%      +5.0%     321949 ± 14%  perf-stat.ps.dTLB-load-misses
 4.453e+09            -6.5%  4.164e+09 ±  8%  perf-stat.ps.dTLB-loads
      9745 ±  6%     +34.3%      13089 ± 45%  perf-stat.ps.dTLB-store-misses
 2.656e+09            -5.9%  2.498e+09 ±  8%  perf-stat.ps.dTLB-stores
    333519 ± 21%      -9.7%     301231 ± 14%  perf-stat.ps.iTLB-load-misses
   1132855 ± 14%      -9.2%    1029084 ±  8%  perf-stat.ps.iTLB-loads
 1.856e+10            -5.8%  1.748e+10 ±  8%  perf-stat.ps.instructions
      0.28 ± 54%     +76.0%       0.48 ± 22%  perf-stat.ps.major-faults
      1894            -2.9%       1838 ±  4%  perf-stat.ps.minor-faults
    107669 ± 22%      -3.8%     103537 ±  7%  perf-stat.ps.node-loads
    164076 ± 31%     +85.2%     303821 ±  4%  perf-stat.ps.node-stores
      1894            -2.9%       1839 ±  4%  perf-stat.ps.page-faults
     35446            +0.1%      35466        perf-stat.ps.task-clock
 1.201e+12            -2.2%  1.175e+12        perf-stat.total.instructions
      9668 ± 38%     +21.7%      11761        sched_debug.cfs_rq:/.avg_vruntime.avg
     73804 ± 37%     +22.9%      90724 ±  8%  sched_debug.cfs_rq:/.avg_vruntime.max
    761.77 ± 37%     +36.5%       1039 ± 20%  sched_debug.cfs_rq:/.avg_vruntime.min
     15647 ± 39%     +24.9%      19545 ±  4%  sched_debug.cfs_rq:/.avg_vruntime.stddev
      0.34 ± 22%     -18.3%       0.27 ±  6%  sched_debug.cfs_rq:/.h_nr_running.avg
      1.44 ± 47%     -20.0%       1.15 ± 19%  sched_debug.cfs_rq:/.h_nr_running.max
      0.48 ± 15%      -9.2%       0.43 ±  5%  sched_debug.cfs_rq:/.h_nr_running.stddev
     19.19 ±141%    +560.7%     126.76 ±184%  sched_debug.cfs_rq:/.left_vruntime.avg
    549.96 ±132%    +700.5%       4402 ±189%  sched_debug.cfs_rq:/.left_vruntime.max
      0.00            +0.0%       0.00        sched_debug.cfs_rq:/.left_vruntime.min
     94.06 ±133%    +671.5%     725.66 ±189%  sched_debug.cfs_rq:/.left_vruntime.stddev
     40431 ± 59%     -19.3%      32628 ± 34%  sched_debug.cfs_rq:/.load.avg
    664217 ±106%     -50.6%     327942 ± 70%  sched_debug.cfs_rq:/.load.max
    133230 ± 85%     -35.2%      86300 ± 54%  sched_debug.cfs_rq:/.load.stddev
      5030 ± 34%     -20.8%       3981 ±  3%  sched_debug.cfs_rq:/.load_avg.avg
     14574 ± 32%     -20.9%      11525 ±  3%  sched_debug.cfs_rq:/.load_avg.max
      1.06 ±197%     -10.6%       0.95 ±162%  sched_debug.cfs_rq:/.load_avg.min
      6600 ± 34%     -21.3%       5193 ±  2%  sched_debug.cfs_rq:/.load_avg.stddev
      9668 ± 38%     +21.7%      11761        sched_debug.cfs_rq:/.min_vruntime.avg
     73804 ± 37%     +22.9%      90727 ±  8%  sched_debug.cfs_rq:/.min_vruntime.max
    761.77 ± 37%     +36.5%       1039 ± 20%  sched_debug.cfs_rq:/.min_vruntime.min
     15647 ± 39%     +24.9%      19545 ±  4%  sched_debug.cfs_rq:/.min_vruntime.stddev
      0.34 ± 22%     -18.3%       0.27 ±  6%  sched_debug.cfs_rq:/.nr_running.avg
      1.44 ± 47%     -20.0%       1.15 ± 19%  sched_debug.cfs_rq:/.nr_running.max
      0.48 ± 15%      -9.2%       0.43 ±  5%  sched_debug.cfs_rq:/.nr_running.stddev
     24.80 ± 79%     +82.3%      45.20 ± 59%  sched_debug.cfs_rq:/.removed.load_avg.avg
    637.94 ± 53%     -11.9%     562.10 ± 49%  sched_debug.cfs_rq:/.removed.load_avg.max
    120.18 ± 63%     +20.1%     144.33 ± 49%  sched_debug.cfs_rq:/.removed.load_avg.stddev
      7.66 ± 68%    +140.4%      18.41 ± 64%  sched_debug.cfs_rq:/.removed.runnable_avg.avg
    270.06 ± 66%      -1.4%     266.15 ± 57%  sched_debug.cfs_rq:/.removed.runnable_avg.max
     44.38 ± 66%     +36.4%      60.54 ± 56%  sched_debug.cfs_rq:/.removed.runnable_avg.stddev
      7.65 ± 68%    +140.5%      18.41 ± 64%  sched_debug.cfs_rq:/.removed.util_avg.avg
    269.94 ± 66%      -1.4%     266.10 ± 57%  sched_debug.cfs_rq:/.removed.util_avg.max
     44.36 ± 66%     +36.5%      60.53 ± 56%  sched_debug.cfs_rq:/.removed.util_avg.stddev
     19.19 ±141%    +560.6%     126.76 ±184%  sched_debug.cfs_rq:/.right_vruntime.avg
    550.03 ±132%    +700.4%       4402 ±189%  sched_debug.cfs_rq:/.right_vruntime.max
      0.00            +0.0%       0.00        sched_debug.cfs_rq:/.right_vruntime.min
     94.07 ±133%    +671.4%     725.66 ±189%  sched_debug.cfs_rq:/.right_vruntime.stddev
    480.62 ± 26%     -15.5%     406.33 ±  3%  sched_debug.cfs_rq:/.runnable_avg.avg
      1462 ± 15%     -10.4%       1310 ±  9%  sched_debug.cfs_rq:/.runnable_avg.max
     16.62 ±140%     -39.5%      10.05 ±180%  sched_debug.cfs_rq:/.runnable_avg.min
    373.33 ± 16%     -11.4%     330.95 ±  7%  sched_debug.cfs_rq:/.runnable_avg.stddev
      0.00 ±228%    -100.0%       0.00        sched_debug.cfs_rq:/.spread.avg
      0.07 ±228%    -100.0%       0.00        sched_debug.cfs_rq:/.spread.max
      0.01 ±228%    -100.0%       0.00        sched_debug.cfs_rq:/.spread.stddev
    476.46 ± 26%     -15.4%     403.03 ±  2%  sched_debug.cfs_rq:/.util_avg.avg
      1454 ± 15%     -10.4%       1303 ±  9%  sched_debug.cfs_rq:/.util_avg.max
     15.75 ±142%     -42.9%       9.00 ±201%  sched_debug.cfs_rq:/.util_avg.min
    371.95 ± 16%     -11.4%     329.56 ±  8%  sched_debug.cfs_rq:/.util_avg.stddev
     83.12 ± 15%      -5.3%      78.72 ± 13%  sched_debug.cfs_rq:/.util_est_enqueued.avg
    836.69 ±  3%      -0.7%     830.70 ±  6%  sched_debug.cfs_rq:/.util_est_enqueued.max
    215.18 ± 12%      +1.9%     219.21 ±  5%  sched_debug.cfs_rq:/.util_est_enqueued.stddev
    577964 ± 14%      +5.2%     608099 ±  3%  sched_debug.cpu.avg_idle.avg
    985615 ±  2%      +0.2%     988020 ±  2%  sched_debug.cpu.avg_idle.max
     28127 ±134%     -81.3%       5250 ± 49%  sched_debug.cpu.avg_idle.min
    274888 ± 10%      +7.8%     296303 ±  4%  sched_debug.cpu.avg_idle.stddev
     48833 ± 26%     +18.4%      57815 ±  5%  sched_debug.cpu.clock.avg
     48834 ± 26%     +18.4%      57817 ±  5%  sched_debug.cpu.clock.max
     48832 ± 26%     +18.4%      57814 ±  5%  sched_debug.cpu.clock.min
      0.80 ±  6%      +3.2%       0.82 ±  9%  sched_debug.cpu.clock.stddev
     48639 ± 25%     +18.3%      57562 ±  5%  sched_debug.cpu.clock_task.avg
     48681 ± 25%     +18.3%      57610 ±  5%  sched_debug.cpu.clock_task.max
     48275 ± 26%     +18.5%      57194 ±  5%  sched_debug.cpu.clock_task.min
     74.28 ±  8%      +6.2%      78.89 ±  2%  sched_debug.cpu.clock_task.stddev
    972.91 ± 15%      -9.7%     878.68 ±  9%  sched_debug.cpu.curr->pid.avg
      3815 ±  7%      +4.4%       3984        sched_debug.cpu.curr->pid.max
      1478 ±  2%      -1.4%       1457        sched_debug.cpu.curr->pid.stddev
    500000            +0.0%     500000        sched_debug.cpu.max_idle_balance_cost.avg
    500000            +0.0%     500000        sched_debug.cpu.max_idle_balance_cost.max
    500000            +0.0%     500000        sched_debug.cpu.max_idle_balance_cost.min
      4294            +0.0%       4294        sched_debug.cpu.next_balance.avg
      4294            +0.0%       4294        sched_debug.cpu.next_balance.max
      4294            +0.0%       4294        sched_debug.cpu.next_balance.min
      0.00 ± 19%      -6.3%       0.00 ± 21%  sched_debug.cpu.next_balance.stddev
      0.32 ± 21%     -14.4%       0.28 ±  8%  sched_debug.cpu.nr_running.avg
      1.44 ± 47%     -20.0%       1.15 ± 19%  sched_debug.cpu.nr_running.max
      0.47 ± 15%      -8.9%       0.43 ±  4%  sched_debug.cpu.nr_running.stddev
     28149 ± 48%     +26.3%      35546        sched_debug.cpu.nr_switches.avg
    117924 ± 41%     +26.2%     148828 ± 16%  sched_debug.cpu.nr_switches.max
      1171 ± 52%     +24.6%       1459 ± 38%  sched_debug.cpu.nr_switches.min
     27983 ± 44%     +25.6%      35144 ± 15%  sched_debug.cpu.nr_switches.stddev
      0.00 ±264%     -20.0%       0.00 ±300%  sched_debug.cpu.nr_uninterruptible.avg
     18.56 ± 30%     +11.5%      20.70 ± 16%  sched_debug.cpu.nr_uninterruptible.max
    -12.56            -3.3%     -12.15        sched_debug.cpu.nr_uninterruptible.min
      6.74 ± 19%      +0.4%       6.77 ±  9%  sched_debug.cpu.nr_uninterruptible.stddev
     48832 ± 26%     +18.4%      57814 ±  5%  sched_debug.cpu_clk
    996147            +0.0%     996147        sched_debug.dl_rq:.dl_bw->bw.avg
    996147            +0.0%     996147        sched_debug.dl_rq:.dl_bw->bw.max
    996147            +0.0%     996147        sched_debug.dl_rq:.dl_bw->bw.min
 4.295e+09            +0.0%  4.295e+09        sched_debug.jiffies
     48753 ± 26%     +18.4%      57735 ±  5%  sched_debug.ktime
    950.00            +0.0%     950.00        sched_debug.rt_rq:.rt_runtime.avg
    950.00            +0.0%     950.00        sched_debug.rt_rq:.rt_runtime.max
    950.00            +0.0%     950.00        sched_debug.rt_rq:.rt_runtime.min
     48853 ± 26%     +18.4%      57835 ±  5%  sched_debug.sched_clk
      1.00            +0.0%       1.00        sched_debug.sched_clock_stable()
      3.00            +0.0%       3.00        sched_debug.sysctl_sched.sysctl_sched_base_slice
  25056823            +0.0%   25056823        sched_debug.sysctl_sched.sysctl_sched_features
      1.00            +0.0%       1.00        sched_debug.sysctl_sched.sysctl_sched_tunable_scaling
      0.00 ±264%    +416.0%       0.01 ±178%  perf-sched.sch_delay.avg.ms.__cond_resched.__alloc_pages.folio_alloc.do_read_cache_folio.vfs_dedupe_file_range_compare
      0.00          +5e+98%       0.00 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.__btrfs_run_delayed_refs.btrfs_run_delayed_refs.flush_space.btrfs_preempt_reclaim_metadata_space
      0.00        +1.5e+99%       0.00 ±161%  perf-sched.sch_delay.avg.ms.__cond_resched.__flush_work.isra.0.__lru_add_drain_all
      0.00 ±264%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.__cond_resched.__kmem_cache_alloc_node.kmalloc_trace.btrfs_alloc_block_rsv.btrfs_replace_file_extents
      0.00          +5e+98%       0.00 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.__kmem_cache_alloc_node.kmalloc_trace.btrfs_fallocate.vfs_fallocate
      0.00 ±264%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.__cond_resched.__wait_for_common.wait_for_completion_state.kernel_clone.__x64_sys_vfork
      0.00          +9e+98%       0.00 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.btrfs_alloc_path.btrfs_del_csums.cleanup_ref_head.__btrfs_run_delayed_refs
      0.00          +6e+98%       0.00 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.btrfs_search_slot.btrfs_del_csums.cleanup_ref_head.__btrfs_run_delayed_refs
      0.00 ±264%    +140.0%       0.00 ±152%  perf-sched.sch_delay.avg.ms.__cond_resched.btrfs_search_slot.btrfs_insert_empty_items.btrfs_insert_replace_extent.btrfs_replace_file_extents
      0.00 ±264%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.__cond_resched.cgroup_rstat_flush_locked.cgroup_rstat_flush.do_flush_stats.flush_memcg_stats_dwork
      0.00 ±264%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.__cond_resched.down_read.walk_component.link_path_walk.part
      0.00          +5e+98%       0.00 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.down_write.__btrfs_tree_lock.btrfs_lock_root_node.btrfs_search_slot
      0.02 ±259%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.__cond_resched.find_free_extent.btrfs_reserve_extent.__btrfs_prealloc_file_range.btrfs_prealloc_file_range
      0.17 ±262%     -99.8%       0.00 ±165%  perf-sched.sch_delay.avg.ms.__cond_resched.kmem_cache_alloc.alloc_extent_state.__clear_extent_bit.btrfs_clone_files
      0.00        +7.3e+99%       0.01 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.kmem_cache_alloc.alloc_extent_state.__clear_extent_bit.btrfs_do_readpage
      0.00 ±264%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.__cond_resched.kmem_cache_alloc.alloc_extent_state.__clear_extent_bit.btrfs_extent_same_range
      0.00       +1.3e+100%       0.01 ±297%  perf-sched.sch_delay.avg.ms.__cond_resched.kmem_cache_alloc.alloc_extent_state.__clear_extent_bit.btrfs_invalidate_folio
      0.00 ±264%     +60.0%       0.00 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.kmem_cache_alloc.alloc_extent_state.__clear_extent_bit.btrfs_replace_file_extents
      0.00 ±264%     -73.3%       0.00 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.kmem_cache_alloc.alloc_extent_state.__set_extent_bit.set_extent_bit
      0.00          +4e+98%       0.00 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.kmem_cache_alloc.btrfs_drop_extent_map_range.btrfs_drop_extents.btrfs_replace_file_extents
      0.00          +4e+98%       0.00 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.kmem_cache_alloc.start_transaction.btrfs_fileattr_set.vfs_fileattr_set
      0.00 ±264%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.__cond_resched.kmem_cache_alloc.start_transaction.btrfs_replace_file_extents.btrfs_punch_hole
      0.00          +6e+98%       0.00 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.mutex_lock.perf_poll.do_poll.constprop
      0.00          +5e+98%       0.00 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.rcu_gp_cleanup.rcu_gp_kthread.kthread.ret_from_fork
      0.00        +1.7e+99%       0.00 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.shrink_dentry_list.shrink_dcache_parent.d_invalidate.proc_invalidate_siblings_dcache
      0.00 ± 78%      +6.7%       0.01 ± 70%  perf-sched.sch_delay.avg.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      0.00          +8e+98%       0.00 ±156%  perf-sched.sch_delay.avg.ms.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part
      0.00 ±264%     -60.0%       0.00 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.task_work_run.exit_to_user_mode_loop.exit_to_user_mode_prepare.syscall_exit_to_user_mode
      0.01 ±264%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.__cond_resched.truncate_inode_pages_range.btrfs_clone_files.isra.0
      0.00          +2e+98%       0.00 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.wp_page_copy.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
      0.01 ± 17%      -4.0%       0.00 ± 15%  perf-sched.sch_delay.avg.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
      0.01 ± 44%      +1.7%       0.01 ± 39%  perf-sched.sch_delay.avg.ms.devkmsg_read.vfs_read.ksys_read.do_syscall_64
      0.00            +0.0%       0.00        perf-sched.sch_delay.avg.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
      0.00 ± 10%      -4.0%       0.00        perf-sched.sch_delay.avg.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
      0.01            +0.0%       0.01        perf-sched.sch_delay.avg.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
      0.00 ±264%    +140.0%       0.00 ±300%  perf-sched.sch_delay.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_exc_page_fault
      0.00          +2e+98%       0.00 ±200%  perf-sched.sch_delay.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt
      0.00 ±133%   +6324.6%       0.10 ±295%  perf-sched.sch_delay.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi
      0.08 ±260%     -99.6%       0.00 ±213%  perf-sched.sch_delay.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.syscall_exit_to_user_mode.do_syscall_64
      0.00          +1e+98%       0.00 ±300%  perf-sched.sch_delay.avg.ms.handle_reserve_ticket.__reserve_bytes.btrfs_reserve_data_bytes.btrfs_fallocate
      0.00 ±111%     +10.0%       0.00 ±103%  perf-sched.sch_delay.avg.ms.io_schedule.folio_wait_bit_common.filemap_fault.__do_fault
      0.00            +3.3%       0.00 ±  9%  perf-sched.sch_delay.avg.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
      0.00           +10.0%       0.00 ± 27%  perf-sched.sch_delay.avg.ms.schedule_hrtimeout_range_clock.do_poll.constprop.0.do_sys_poll
      0.01 ± 33%     -11.7%       0.01 ± 12%  perf-sched.sch_delay.avg.ms.schedule_hrtimeout_range_clock.do_select.core_sys_select.kern_select
      0.00 ± 10%     +14.6%       0.01 ± 18%  perf-sched.sch_delay.avg.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
      0.00 ±264%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.schedule_preempt_disabled.rwsem_down_read_slowpath.down_read.__btrfs_tree_read_lock
      0.00 ±158%     +30.9%       0.00 ±263%  perf-sched.sch_delay.avg.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.__btrfs_tree_lock
      0.00        +1.1e+99%       0.00 ±209%  perf-sched.sch_delay.avg.ms.schedule_timeout.__wait_for_common.__flush_work.isra.0
      0.00 ± 10%      -4.0%       0.00        perf-sched.sch_delay.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
      0.00 ± 14%      +4.2%       0.00 ± 18%  perf-sched.sch_delay.avg.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
      0.00        +1.1e+99%       0.00 ±124%  perf-sched.sch_delay.avg.ms.schedule_timeout.khugepaged_wait_work.khugepaged.kthread
      0.00 ± 29%      -8.1%       0.00 ±  9%  perf-sched.sch_delay.avg.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
      0.03 ± 46%      -1.7%       0.03 ± 42%  perf-sched.sch_delay.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      0.01 ± 30%     +10.9%       0.01 ± 40%  perf-sched.sch_delay.avg.ms.syslog_print.do_syslog.kmsg_read.vfs_read
      0.00 ± 15%     +17.6%       0.00 ± 36%  perf-sched.sch_delay.avg.ms.wait_for_partner.fifo_open.do_dentry_open.do_open
      0.03 ± 36%     +10.7%       0.04 ± 24%  perf-sched.sch_delay.avg.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
      0.16 ±261%    +395.6%       0.81 ±172%  perf-sched.sch_delay.max.ms.__cond_resched.__alloc_pages.folio_alloc.do_read_cache_folio.vfs_dedupe_file_range_compare
      0.00          +5e+98%       0.00 ±300%  perf-sched.sch_delay.max.ms.__cond_resched.__btrfs_run_delayed_refs.btrfs_run_delayed_refs.flush_space.btrfs_preempt_reclaim_metadata_space
      0.00        +1.5e+99%       0.00 ±161%  perf-sched.sch_delay.max.ms.__cond_resched.__flush_work.isra.0.__lru_add_drain_all
      0.00 ±264%    -100.0%       0.00        perf-sched.sch_delay.max.ms.__cond_resched.__kmem_cache_alloc_node.kmalloc_trace.btrfs_alloc_block_rsv.btrfs_replace_file_extents
      0.00          +5e+98%       0.00 ±300%  perf-sched.sch_delay.max.ms.__cond_resched.__kmem_cache_alloc_node.kmalloc_trace.btrfs_fallocate.vfs_fallocate
      0.01 ±  8%     +10.5%       0.01 ± 20%  perf-sched.sch_delay.max.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
      0.00 ±264%    -100.0%       0.00        perf-sched.sch_delay.max.ms.__cond_resched.__wait_for_common.wait_for_completion_state.kernel_clone.__x64_sys_vfork
      0.00          +9e+98%       0.00 ±300%  perf-sched.sch_delay.max.ms.__cond_resched.btrfs_alloc_path.btrfs_del_csums.cleanup_ref_head.__btrfs_run_delayed_refs
      0.00          +6e+98%       0.00 ±300%  perf-sched.sch_delay.max.ms.__cond_resched.btrfs_search_slot.btrfs_del_csums.cleanup_ref_head.__btrfs_run_delayed_refs
      0.00 ±130%     -20.0%       0.00 ±157%  perf-sched.sch_delay.max.ms.__cond_resched.btrfs_search_slot.btrfs_insert_empty_items.btrfs_insert_replace_extent.btrfs_replace_file_extents
      0.00 ±264%    -100.0%       0.00        perf-sched.sch_delay.max.ms.__cond_resched.cgroup_rstat_flush_locked.cgroup_rstat_flush.do_flush_stats.flush_memcg_stats_dwork
      0.00 ±264%    -100.0%       0.00        perf-sched.sch_delay.max.ms.__cond_resched.down_read.walk_component.link_path_walk.part
      0.00          +5e+98%       0.00 ±300%  perf-sched.sch_delay.max.ms.__cond_resched.down_write.__btrfs_tree_lock.btrfs_lock_root_node.btrfs_search_slot
      0.12 ±259%    -100.0%       0.00        perf-sched.sch_delay.max.ms.__cond_resched.find_free_extent.btrfs_reserve_extent.__btrfs_prealloc_file_range.btrfs_prealloc_file_range
      0.34 ±262%     -98.6%       0.00 ±133%  perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc.alloc_extent_state.__clear_extent_bit.btrfs_clone_files
      0.00 ± 86%  +10253.9%       0.43 ±297%  perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc.alloc_extent_state.__clear_extent_bit.btrfs_do_readpage
      0.00 ±264%    -100.0%       0.00        perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc.alloc_extent_state.__clear_extent_bit.btrfs_extent_same_range
      0.00 ±264%  +28300.0%       0.18 ±297%  perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc.alloc_extent_state.__clear_extent_bit.btrfs_invalidate_folio
      0.00 ±264%     -60.0%       0.00 ±300%  perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc.alloc_extent_state.__clear_extent_bit.btrfs_replace_file_extents
      0.01 ± 38%      +2.9%       0.01 ± 28%  perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc.alloc_extent_state.__set_extent_bit.lock_extent
      0.00 ±264%     -85.2%       0.00 ±300%  perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc.alloc_extent_state.__set_extent_bit.set_extent_bit
      0.00          +4e+98%       0.00 ±300%  perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc.btrfs_drop_extent_map_range.btrfs_drop_extents.btrfs_replace_file_extents
      0.00          +4e+98%       0.00 ±300%  perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc.start_transaction.btrfs_fileattr_set.vfs_fileattr_set
      0.00 ±264%    -100.0%       0.00        perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc.start_transaction.btrfs_replace_file_extents.btrfs_punch_hole
      0.00          +6e+98%       0.00 ±300%  perf-sched.sch_delay.max.ms.__cond_resched.mutex_lock.perf_poll.do_poll.constprop
      0.00          +5e+98%       0.00 ±300%  perf-sched.sch_delay.max.ms.__cond_resched.rcu_gp_cleanup.rcu_gp_kthread.kthread.ret_from_fork
      0.00        +1.7e+99%       0.00 ±300%  perf-sched.sch_delay.max.ms.__cond_resched.shrink_dentry_list.shrink_dcache_parent.d_invalidate.proc_invalidate_siblings_dcache
      0.01 ± 78%     +13.2%       0.01 ± 66%  perf-sched.sch_delay.max.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      0.00          +8e+98%       0.00 ±156%  perf-sched.sch_delay.max.ms.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part
      0.00 ±264%     -60.0%       0.00 ±300%  perf-sched.sch_delay.max.ms.__cond_resched.task_work_run.exit_to_user_mode_loop.exit_to_user_mode_prepare.syscall_exit_to_user_mode
      0.31 ±262%     -99.4%       0.00 ±127%  perf-sched.sch_delay.max.ms.__cond_resched.truncate_inode_pages_range.btrfs_clone_files.isra.0
      0.00 ±141%     -60.0%       0.00 ±300%  perf-sched.sch_delay.max.ms.__cond_resched.vfs_dedupe_file_range_compare.constprop.0.__generic_remap_file_range_prep
      0.00          +2e+98%       0.00 ±300%  perf-sched.sch_delay.max.ms.__cond_resched.wp_page_copy.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
      0.01 ± 27%      -6.0%       0.01 ± 36%  perf-sched.sch_delay.max.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
      0.01 ± 44%      +5.9%       0.01 ± 38%  perf-sched.sch_delay.max.ms.devkmsg_read.vfs_read.ksys_read.do_syscall_64
      0.21 ±201%    +167.9%       0.56 ±270%  perf-sched.sch_delay.max.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
      0.01 ± 11%      +5.6%       0.01 ± 36%  perf-sched.sch_delay.max.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
      0.01 ± 19%      +3.5%       0.01 ± 15%  perf-sched.sch_delay.max.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
      0.00 ±264%    +140.0%       0.00 ±300%  perf-sched.sch_delay.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_exc_page_fault
      0.00 ±264%    +513.3%       0.00 ±148%  perf-sched.sch_delay.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt
      0.00 ±152%   +7456.4%       0.21 ±296%  perf-sched.sch_delay.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi
      0.68 ±262%     -99.8%       0.00 ±200%  perf-sched.sch_delay.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.syscall_exit_to_user_mode.do_syscall_64
      0.00 ±130%      -7.7%       0.00 ±225%  perf-sched.sch_delay.max.ms.handle_reserve_ticket.__reserve_bytes.btrfs_reserve_data_bytes.btrfs_fallocate
      0.00 ±102%      -5.5%       0.00 ±103%  perf-sched.sch_delay.max.ms.io_schedule.folio_wait_bit_common.filemap_fault.__do_fault
      0.02 ± 23%     -23.3%       0.01 ± 29%  perf-sched.sch_delay.max.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
      0.01 ± 21%      +9.3%       0.01 ± 15%  perf-sched.sch_delay.max.ms.schedule_hrtimeout_range_clock.do_poll.constprop.0.do_sys_poll
      0.01 ± 48%     -20.0%       0.01 ± 17%  perf-sched.sch_delay.max.ms.schedule_hrtimeout_range_clock.do_select.core_sys_select.kern_select
      0.01 ± 39%      +4.9%       0.01 ± 28%  perf-sched.sch_delay.max.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
      1.64 ± 85%     -71.6%       0.47 ±128%  perf-sched.sch_delay.max.ms.schedule_preempt_disabled.rwsem_down_read_slowpath.down_read.__btrfs_tree_read_lock
      0.19 ±169%     -43.9%       0.11 ±259%  perf-sched.sch_delay.max.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.__btrfs_tree_lock
      0.00        +1.1e+99%       0.00 ±209%  perf-sched.sch_delay.max.ms.schedule_timeout.__wait_for_common.__flush_work.isra.0
      0.00 ± 35%     +29.7%       0.01 ± 40%  perf-sched.sch_delay.max.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
      0.00 ±  9%     +43.2%       0.01 ± 30%  perf-sched.sch_delay.max.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
      0.00        +1.1e+99%       0.00 ±124%  perf-sched.sch_delay.max.ms.schedule_timeout.khugepaged_wait_work.khugepaged.kthread
      0.51 ±260%     -98.1%       0.01 ± 20%  perf-sched.sch_delay.max.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
      3.09 ± 40%      -1.1%       3.06 ± 27%  perf-sched.sch_delay.max.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      0.01 ± 30%     +20.7%       0.01 ± 36%  perf-sched.sch_delay.max.ms.syslog_print.do_syslog.kmsg_read.vfs_read
      0.00 ± 85%     +75.1%       0.01 ±158%  perf-sched.sch_delay.max.ms.wait_for_partner.fifo_open.do_dentry_open.do_open
      2.39 ± 35%      +5.1%       2.51 ± 28%  perf-sched.sch_delay.max.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
      0.00            +0.0%       0.00        perf-sched.total_sch_delay.average.ms
      3.72 ± 30%      +0.8%       3.75 ± 17%  perf-sched.total_sch_delay.max.ms
      4.89 ±  3%      +4.9%       5.13 ±  3%  perf-sched.total_wait_and_delay.average.ms
     89766            -3.1%      86942        perf-sched.total_wait_and_delay.count.ms
      3132 ± 20%      +5.2%       3295 ± 18%  perf-sched.total_wait_and_delay.max.ms
      4.89 ±  3%      +4.9%       5.13 ±  3%  perf-sched.total_wait_time.average.ms
      3132 ± 20%      +5.2%       3295 ± 18%  perf-sched.total_wait_time.max.ms
      0.00       +1.5e+103%      15.38 ±300%  perf-sched.wait_and_delay.avg.ms.__cond_resched.__do_fault.do_read_fault.do_fault.__handle_mm_fault
     18.52 ±  5%      +1.5%      18.81 ±  8%  perf-sched.wait_and_delay.avg.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
      0.00       +2.3e+102%       2.33 ±300%  perf-sched.wait_and_delay.avg.ms.__cond_resched.btrfs_alloc_path.btrfs_del_csums.cleanup_ref_head.__btrfs_run_delayed_refs
      0.00         +1e+103%      10.44 ±300%  perf-sched.wait_and_delay.avg.ms.__cond_resched.btrfs_search_slot.btrfs_del_csums.cleanup_ref_head.__btrfs_run_delayed_refs
     18.90 ±264%    -100.0%       0.00        perf-sched.wait_and_delay.avg.ms.__cond_resched.cgroup_rstat_flush_locked.cgroup_rstat_flush.do_flush_stats.flush_memcg_stats_dwork
      0.00         +1e+104%     100.10 ±300%  perf-sched.wait_and_delay.avg.ms.__cond_resched.mutex_lock.perf_poll.do_poll.constprop
      0.00         +1e+104%     100.02 ±300%  perf-sched.wait_and_delay.avg.ms.__cond_resched.shrink_dentry_list.shrink_dcache_parent.d_invalidate.proc_invalidate_siblings_dcache
    250.06 ±172%     +54.6%     386.64 ±113%  perf-sched.wait_and_delay.avg.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
    799.87            -0.0%     799.87        perf-sched.wait_and_delay.avg.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
      0.12            +0.0%       0.12        perf-sched.wait_and_delay.avg.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
    283.11 ±  8%      +0.4%     284.37 ±  9%  perf-sched.wait_and_delay.avg.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
      0.00         +1e+104%     100.02 ±300%  perf-sched.wait_and_delay.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_exc_page_fault
    286.77 ±  7%      -1.3%     283.11 ±  6%  perf-sched.wait_and_delay.avg.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
    298.84            +0.4%     299.90 ±  3%  perf-sched.wait_and_delay.avg.ms.schedule_hrtimeout_range_clock.do_poll.constprop.0.do_sys_poll
    420.93 ±  7%      -8.3%     386.20 ±  5%  perf-sched.wait_and_delay.avg.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
      0.01 ±  9%      +7.9%       0.01 ±  6%  perf-sched.wait_and_delay.avg.ms.schedule_preempt_disabled.rwsem_down_read_slowpath.down_read.__btrfs_tree_read_lock
    453.61            -0.2%     452.50        perf-sched.wait_and_delay.avg.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
      4.01 ±  8%      -0.5%       3.99 ±  8%  perf-sched.wait_and_delay.avg.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
    312.30 ±  5%      +2.2%     319.28 ±  8%  perf-sched.wait_and_delay.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
    496.24 ±  3%      +2.0%     506.30 ±  6%  perf-sched.wait_and_delay.avg.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
      0.00         +1e+101%       0.10 ±300%  perf-sched.wait_and_delay.count.__cond_resched.__do_fault.do_read_fault.do_fault.__handle_mm_fault
    200.50 ±  8%      +0.5%     201.50 ±  8%  perf-sched.wait_and_delay.count.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
      0.00         +1e+101%       0.10 ±300%  perf-sched.wait_and_delay.count.__cond_resched.btrfs_alloc_path.btrfs_del_csums.cleanup_ref_head.__btrfs_run_delayed_refs
      0.00         +1e+101%       0.10 ±300%  perf-sched.wait_and_delay.count.__cond_resched.btrfs_search_slot.btrfs_del_csums.cleanup_ref_head.__btrfs_run_delayed_refs
      0.12 ±264%    -100.0%       0.00        perf-sched.wait_and_delay.count.__cond_resched.cgroup_rstat_flush_locked.cgroup_rstat_flush.do_flush_stats.flush_memcg_stats_dwork
      0.00         +1e+101%       0.10 ±300%  perf-sched.wait_and_delay.count.__cond_resched.mutex_lock.perf_poll.do_poll.constprop
      0.00         +1e+101%       0.10 ±300%  perf-sched.wait_and_delay.count.__cond_resched.shrink_dentry_list.shrink_dcache_parent.d_invalidate.proc_invalidate_siblings_dcache
      0.50 ±141%    +160.0%       1.30 ±142%  perf-sched.wait_and_delay.count.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
     10.00            +0.0%      10.00        perf-sched.wait_and_delay.count.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
     74939            +0.0%      74972        perf-sched.wait_and_delay.count.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
     64.00            +0.0%      64.00        perf-sched.wait_and_delay.count.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
      0.00         +1e+101%       0.10 ±300%  perf-sched.wait_and_delay.count.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_exc_page_fault
    165.88 ±  7%      +1.2%     167.80 ±  5%  perf-sched.wait_and_delay.count.pipe_read.vfs_read.ksys_read.do_syscall_64
     56.50            -0.9%      56.00 ±  2%  perf-sched.wait_and_delay.count.schedule_hrtimeout_range_clock.do_poll.constprop.0.do_sys_poll
     11.75 ±  7%      +6.4%      12.50 ±  4%  perf-sched.wait_and_delay.count.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
     11447 ±  6%     -24.2%       8673 ±  6%  perf-sched.wait_and_delay.count.schedule_preempt_disabled.rwsem_down_read_slowpath.down_read.__btrfs_tree_read_lock
     10.00            -2.0%       9.80 ±  4%  perf-sched.wait_and_delay.count.schedule_timeout.kcompactd.kthread.ret_from_fork
      1253 ±  8%      +0.5%       1259 ±  8%  perf-sched.wait_and_delay.count.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
    674.00 ±  7%      -3.4%     651.10 ±  4%  perf-sched.wait_and_delay.count.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
    225.25 ±  5%      +5.3%     237.20 ±  4%  perf-sched.wait_and_delay.count.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
      0.00       +1.5e+103%      15.38 ±300%  perf-sched.wait_and_delay.max.ms.__cond_resched.__do_fault.do_read_fault.do_fault.__handle_mm_fault
      1000            +0.0%       1000        perf-sched.wait_and_delay.max.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
      0.00       +2.3e+102%       2.33 ±300%  perf-sched.wait_and_delay.max.ms.__cond_resched.btrfs_alloc_path.btrfs_del_csums.cleanup_ref_head.__btrfs_run_delayed_refs
      0.00         +1e+103%      10.44 ±300%  perf-sched.wait_and_delay.max.ms.__cond_resched.btrfs_search_slot.btrfs_del_csums.cleanup_ref_head.__btrfs_run_delayed_refs
     18.90 ±264%    -100.0%       0.00        perf-sched.wait_and_delay.max.ms.__cond_resched.cgroup_rstat_flush_locked.cgroup_rstat_flush.do_flush_stats.flush_memcg_stats_dwork
      0.00         +1e+104%     100.10 ±300%  perf-sched.wait_and_delay.max.ms.__cond_resched.mutex_lock.perf_poll.do_poll.constprop
      0.00         +1e+104%     100.02 ±300%  perf-sched.wait_and_delay.max.ms.__cond_resched.shrink_dentry_list.shrink_dcache_parent.d_invalidate.proc_invalidate_siblings_dcache
    251.12 ±171%    +179.6%     702.09 ±128%  perf-sched.wait_and_delay.max.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
    999.87            -0.0%     999.86        perf-sched.wait_and_delay.max.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
      1000            +0.0%       1000        perf-sched.wait_and_delay.max.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
      1868 ± 49%     -19.8%       1498 ± 44%  perf-sched.wait_and_delay.max.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
      0.00         +1e+104%     100.02 ±300%  perf-sched.wait_and_delay.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_exc_page_fault
      1008            -0.0%       1008        perf-sched.wait_and_delay.max.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
      1000            -0.0%       1000        perf-sched.wait_and_delay.max.ms.schedule_hrtimeout_range_clock.do_poll.constprop.0.do_sys_poll
    531.24 ±  7%      -5.9%     499.99        perf-sched.wait_and_delay.max.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
      2.63 ± 60%     -53.1%       1.23 ± 67%  perf-sched.wait_and_delay.max.ms.schedule_preempt_disabled.rwsem_down_read_slowpath.down_read.__btrfs_tree_read_lock
    504.63            +0.0%     504.71        perf-sched.wait_and_delay.max.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
      5.01            +0.1%       5.01        perf-sched.wait_and_delay.max.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
      2794 ± 23%      +7.7%       3011 ± 18%  perf-sched.wait_and_delay.max.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      2561 ± 18%     +14.4%       2930 ± 33%  perf-sched.wait_and_delay.max.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
      0.00 ±182%     +31.8%       0.00 ±109%  perf-sched.wait_time.avg.ms.__cond_resched.__alloc_pages.__kmalloc_large_node.__kmalloc_node.kvmalloc_node
      0.01 ± 43%      -1.9%       0.01 ± 49%  perf-sched.wait_time.avg.ms.__cond_resched.__alloc_pages.folio_alloc.do_read_cache_folio.vfs_dedupe_file_range_compare
      0.00          +8e+98%       0.00 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.__btrfs_run_delayed_refs.btrfs_run_delayed_refs.flush_space.btrfs_preempt_reclaim_metadata_space
      0.00 ±174%     -64.4%       0.00 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.__clear_extent_bit.btrfs_replace_file_extents.btrfs_punch_hole.btrfs_fallocate
      0.00       +1.5e+103%      15.38 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.__do_fault.do_read_fault.do_fault.__handle_mm_fault
      0.00 ±174%     -20.0%       0.00 ±201%  perf-sched.wait_time.avg.ms.__cond_resched.__kmem_cache_alloc_node.__kmalloc_node_track_caller.memdup_user.do_vfs_ioctl
      0.01 ±101%     -20.0%       0.01 ± 49%  perf-sched.wait_time.avg.ms.__cond_resched.__kmem_cache_alloc_node.kmalloc_trace.btrfs_alloc_block_rsv.btrfs_replace_file_extents
      0.00 ±264%    +508.0%       0.00 ±191%  perf-sched.wait_time.avg.ms.__cond_resched.__kmem_cache_alloc_node.kmalloc_trace.btrfs_fallocate.vfs_fallocate
     18.52 ±  5%      +1.5%      18.81 ±  8%  perf-sched.wait_time.avg.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
      0.08 ±264%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.__wait_for_common.wait_for_completion_state.kernel_clone.__x64_sys_vfork
      0.00 ±264%     -52.9%       0.00 ±200%  perf-sched.wait_time.avg.ms.__cond_resched.btrfs_alloc_path.btrfs_clone.btrfs_clone_files.isra
      0.00        +3.8e+99%       0.00 ± 52%  perf-sched.wait_time.avg.ms.__cond_resched.btrfs_alloc_path.btrfs_clone.btrfs_extent_same_range.btrfs_remap_file_range
      0.00       +2.3e+102%       2.33 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.btrfs_alloc_path.btrfs_del_csums.cleanup_ref_head.__btrfs_run_delayed_refs
      0.00 ±264%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.btrfs_alloc_path.btrfs_get_extent.find_first_non_hole.btrfs_punch_hole
      0.00 ±174%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.btrfs_alloc_path.btrfs_punch_hole.btrfs_fallocate.vfs_fallocate
      0.00        +2.4e+99%       0.00 ±112%  perf-sched.wait_time.avg.ms.__cond_resched.btrfs_alloc_path.insert_prealloc_file_extent.__btrfs_prealloc_file_range.btrfs_prealloc_file_range
      0.00 ±264%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.btrfs_search_slot.btrfs_clone.btrfs_clone_files.isra
      0.00          +6e+98%       0.00 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.btrfs_search_slot.btrfs_clone.btrfs_extent_same_range.btrfs_remap_file_range
      0.00         +1e+103%      10.44 ±299%  perf-sched.wait_time.avg.ms.__cond_resched.btrfs_search_slot.btrfs_del_csums.cleanup_ref_head.__btrfs_run_delayed_refs
      0.01 ± 53%    +269.2%       0.04 ±240%  perf-sched.wait_time.avg.ms.__cond_resched.btrfs_search_slot.btrfs_insert_empty_items.btrfs_insert_replace_extent.btrfs_replace_file_extents
      0.01 ±106%     -75.2%       0.00 ±152%  perf-sched.wait_time.avg.ms.__cond_resched.btrfs_search_slot.btrfs_lookup_file_extent.btrfs_drop_extents.btrfs_replace_file_extents
      0.00 ±264%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.btrfs_search_slot.btrfs_lookup_file_extent.btrfs_get_extent.find_first_non_hole
     18.90 ±264%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.cgroup_rstat_flush_locked.cgroup_rstat_flush.do_flush_stats.flush_memcg_stats_dwork
      0.00 ±105%     -24.4%       0.00 ±162%  perf-sched.wait_time.avg.ms.__cond_resched.down_read.__btrfs_tree_read_lock.btrfs_read_lock_root_node.btrfs_search_slot
      0.00 ±174%      -2.2%       0.00 ±201%  perf-sched.wait_time.avg.ms.__cond_resched.down_read.find_free_extent.btrfs_reserve_extent.__btrfs_prealloc_file_range
      0.00 ±264%      +0.0%       0.00 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.down_read.unmap_mapping_range.truncate_pagecache_range.btrfs_punch_hole_lock_range
      0.00 ± 85%     +45.2%       0.00 ± 95%  perf-sched.wait_time.avg.ms.__cond_resched.down_write.__btrfs_tree_lock.btrfs_lock_root_node.btrfs_search_slot
      0.00          +5e+98%       0.00 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.down_write.btrfs_inode_lock.btrfs_punch_hole.btrfs_fallocate
      0.00 ±264%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.down_write.btrfs_inode_lock.btrfs_remap_file_range.do_clone_file_range
      0.00 ±264%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.down_write.btrfs_remap_file_range.do_clone_file_range.vfs_clone_file_range
      0.00 ±174%     -70.9%       0.00 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.down_write.lock_two_inodes.btrfs_remap_file_range.do_clone_file_range
      0.00 ±264%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.down_write.lock_two_inodes.btrfs_remap_file_range.vfs_dedupe_file_range_one
      0.00          +4e+98%       0.00 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.down_write.vfs_fileattr_set.do_vfs_ioctl.__x64_sys_ioctl
      0.03 ±210%     -80.2%       0.01 ±  8%  perf-sched.wait_time.avg.ms.__cond_resched.find_free_extent.btrfs_reserve_extent.__btrfs_prealloc_file_range.btrfs_prealloc_file_range
      0.00 ± 41%     -15.4%       0.00 ± 76%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.alloc_extent_map.__btrfs_prealloc_file_range.btrfs_prealloc_file_range
      0.00 ±196%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.alloc_extent_map.btrfs_get_extent.btrfs_do_readpage
      0.01 ± 83%     -18.6%       0.01 ± 34%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.alloc_extent_state.__clear_extent_bit.btrfs_clone_files
      0.01 ±  8%     +10.5%       0.01 ± 12%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.alloc_extent_state.__clear_extent_bit.btrfs_do_readpage
      0.00 ±102%     -45.9%       0.00 ±101%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.alloc_extent_state.__clear_extent_bit.btrfs_extent_same_range
      0.00 ±105%     +11.1%       0.00 ±101%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.alloc_extent_state.__clear_extent_bit.btrfs_fallocate
      0.01 ± 42%     +14.7%       0.01 ± 77%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.alloc_extent_state.__clear_extent_bit.btrfs_invalidate_folio
      0.00 ±264%     -55.6%       0.00 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.alloc_extent_state.__clear_extent_bit.btrfs_punch_hole
      0.01 ± 73%     -33.7%       0.01 ±105%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.alloc_extent_state.__clear_extent_bit.btrfs_replace_file_extents
      0.01 ±  9%     +39.5%       0.01 ± 69%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.alloc_extent_state.__set_extent_bit.lock_extent
      0.01 ± 27%     +26.8%       0.01 ± 60%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.alloc_extent_state.__set_extent_bit.set_extent_bit
      0.00 ±264%     +12.0%       0.00 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.btrfs_add_delayed_data_ref.btrfs_alloc_reserved_file_extent.btrfs_insert_replace_extent
      0.00 ± 79%     +12.0%       0.00 ± 83%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.btrfs_add_delayed_data_ref.btrfs_drop_extents.btrfs_replace_file_extents
      0.00 ±100%     -41.8%       0.00 ±153%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.btrfs_add_delayed_data_ref.btrfs_insert_replace_extent.btrfs_replace_file_extents
      0.00        +3.7e+99%       0.00 ±265%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.btrfs_clone.btrfs_clone_files.isra
      0.00 ±100%     +84.0%       0.00 ± 57%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.btrfs_drop_extent_map_range.btrfs_drop_extents.btrfs_replace_file_extents
      0.00 ±264%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.btrfs_drop_extent_map_range.btrfs_replace_extent_map_range.__btrfs_prealloc_file_range
      0.00 ±264%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.btrfs_get_extent.btrfs_do_readpage.btrfs_read_folio
      0.00          +5e+98%       0.00 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.btrfs_punch_hole.btrfs_fallocate.vfs_fallocate
      0.00 ±264%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.insert_prealloc_file_extent.__btrfs_prealloc_file_range.btrfs_prealloc_file_range
      0.00 ± 83%     -40.7%       0.00 ±122%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.start_transaction.btrfs_dirty_inode.file_modified
      0.00 ±173%    +520.0%       0.01 ±230%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.start_transaction.btrfs_fileattr_set.vfs_fileattr_set
      0.00 ±133%     +42.9%       0.00 ±100%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.start_transaction.btrfs_replace_file_extents.btrfs_clone
      0.00 ±264%     -27.3%       0.00 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.start_transaction.btrfs_replace_file_extents.btrfs_punch_hole
      0.00 ±173%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.start_transaction.btrfs_replace_file_extents.insert_prealloc_file_extent
      0.00 ±264%     +20.0%       0.00 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.mnt_want_write_file.do_vfs_ioctl.__x64_sys_ioctl.do_syscall_64
      0.00          +1e+99%       0.00 ±204%  perf-sched.wait_time.avg.ms.__cond_resched.mnt_want_write_file.vfs_dedupe_file_range_one.vfs_dedupe_file_range.do_vfs_ioctl
      0.00 ±136%     -20.0%       0.00 ±153%  perf-sched.wait_time.avg.ms.__cond_resched.mutex_lock.__btrfs_release_delayed_node.part.0
      0.00 ±264%      +0.0%       0.00 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.mutex_lock.btrfs_delayed_update_inode.btrfs_update_inode.__btrfs_prealloc_file_range
      0.00 ±264%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.mutex_lock.btrfs_delayed_update_inode.btrfs_update_inode.btrfs_dirty_inode
      0.00 ±264%    +380.0%       0.00 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.mutex_lock.btrfs_delayed_update_inode.btrfs_update_inode.btrfs_fileattr_set
      0.00 ±264%    +252.0%       0.00 ±245%  perf-sched.wait_time.avg.ms.__cond_resched.mutex_lock.btrfs_delayed_update_inode.btrfs_update_inode.btrfs_punch_hole
      0.00 ±173%    +212.0%       0.00 ± 51%  perf-sched.wait_time.avg.ms.__cond_resched.mutex_lock.btrfs_delayed_update_inode.btrfs_update_inode.clone_finish_inode_update
      0.00         +1e+104%     100.10 ±299%  perf-sched.wait_time.avg.ms.__cond_resched.mutex_lock.perf_poll.do_poll.constprop
      0.00         +4e+101%       0.40 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.rcu_gp_cleanup.rcu_gp_kthread.kthread.ret_from_fork
      0.00         +1e+104%     100.02 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.shrink_dentry_list.shrink_dcache_parent.d_invalidate.proc_invalidate_siblings_dcache
    250.06 ±172%     +54.6%     386.64 ±113%  perf-sched.wait_time.avg.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      0.00          +4e+98%       0.00 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.start_transaction.btrfs_replace_file_extents.btrfs_clone.btrfs_extent_same_range
      0.00        +5.8e+99%       0.01 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.task_work_run.exit_to_user_mode_loop.exit_to_user_mode_prepare.syscall_exit_to_user_mode
      0.01 ± 22%     +21.7%       0.01 ± 67%  perf-sched.wait_time.avg.ms.__cond_resched.truncate_inode_pages_range.btrfs_clone_files.isra.0
      0.00        +1.4e+99%       0.00 ±153%  perf-sched.wait_time.avg.ms.__cond_resched.truncate_inode_pages_range.btrfs_punch_hole_lock_range.btrfs_punch_hole.btrfs_fallocate
      0.00 ± 43%     -12.5%       0.00 ± 69%  perf-sched.wait_time.avg.ms.__cond_resched.vfs_clone_file_range.ioctl_file_clone.do_vfs_ioctl.__x64_sys_ioctl
      0.01 ± 93%     -31.9%       0.01 ± 17%  perf-sched.wait_time.avg.ms.__cond_resched.vfs_dedupe_file_range_compare.constprop.0.__generic_remap_file_range_prep
      0.00 ± 39%     -38.3%       0.00 ± 82%  perf-sched.wait_time.avg.ms.__cond_resched.vfs_fallocate.ioctl_preallocate.__x64_sys_ioctl.do_syscall_64
    799.87            -0.0%     799.86        perf-sched.wait_time.avg.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
      0.11 ±173%    +141.9%       0.27 ± 82%  perf-sched.wait_time.avg.ms.devkmsg_read.vfs_read.ksys_read.do_syscall_64
      0.11            +0.0%       0.11        perf-sched.wait_time.avg.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
    283.11 ±  8%      +0.4%     284.37 ±  9%  perf-sched.wait_time.avg.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
      0.28 ±  2%      +4.4%       0.29 ±  2%  perf-sched.wait_time.avg.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
      0.00         +1e+104%     100.02 ±300%  perf-sched.wait_time.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_exc_page_fault
      0.00 ± 10%     +21.1%       0.01 ± 29%  perf-sched.wait_time.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt
      0.08 ±205%     -69.0%       0.03 ± 77%  perf-sched.wait_time.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi
      0.02 ±155%     -57.8%       0.01 ± 47%  perf-sched.wait_time.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.syscall_exit_to_user_mode.do_syscall_64
      0.01 ± 12%      +1.8%       0.01 ± 26%  perf-sched.wait_time.avg.ms.handle_reserve_ticket.__reserve_bytes.btrfs_reserve_data_bytes.btrfs_fallocate
      0.00 ±185%     -46.7%       0.00 ±200%  perf-sched.wait_time.avg.ms.io_schedule.folio_wait_bit_common.filemap_fault.__do_fault
    286.76 ±  7%      -1.3%     283.11 ±  6%  perf-sched.wait_time.avg.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
    298.84            +0.4%     299.90 ±  3%  perf-sched.wait_time.avg.ms.schedule_hrtimeout_range_clock.do_poll.constprop.0.do_sys_poll
    420.93 ±  7%      -8.3%     386.19 ±  5%  perf-sched.wait_time.avg.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
      0.01 ±  6%      +3.4%       0.01 ±  8%  perf-sched.wait_time.avg.ms.schedule_preempt_disabled.rwsem_down_read_slowpath.down_read.__btrfs_tree_read_lock
      0.01 ±  9%    +106.5%       0.01 ±117%  perf-sched.wait_time.avg.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.__btrfs_tree_lock
      0.00        +1.5e+99%       0.00 ±155%  perf-sched.wait_time.avg.ms.schedule_timeout.__wait_for_common.__flush_work.isra.0
      0.00 ±188%     -70.5%       0.00 ±300%  perf-sched.wait_time.avg.ms.schedule_timeout.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr
      0.42            +1.3%       0.42        perf-sched.wait_time.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
    453.61            -0.2%     452.50        perf-sched.wait_time.avg.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
      0.00         +5e+101%       0.50 ±202%  perf-sched.wait_time.avg.ms.schedule_timeout.khugepaged_wait_work.khugepaged.kthread
      4.01 ±  8%      -0.5%       3.99 ±  8%  perf-sched.wait_time.avg.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
    312.27 ±  5%      +2.2%     319.25 ±  8%  perf-sched.wait_time.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      0.11 ±173%    +141.6%       0.27 ± 82%  perf-sched.wait_time.avg.ms.syslog_print.do_syslog.kmsg_read.vfs_read
      0.00          +1e+98%       0.00 ±300%  perf-sched.wait_time.avg.ms.wait_for_partner.fifo_open.do_dentry_open.do_open
    496.21 ±  3%      +2.0%     506.26 ±  6%  perf-sched.wait_time.avg.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
      0.00 ±178%     +24.4%       0.00 ±109%  perf-sched.wait_time.max.ms.__cond_resched.__alloc_pages.__kmalloc_large_node.__kmalloc_node.kvmalloc_node
      0.12 ±183%     -29.5%       0.08 ±218%  perf-sched.wait_time.max.ms.__cond_resched.__alloc_pages.folio_alloc.do_read_cache_folio.vfs_dedupe_file_range_compare
      0.00          +8e+98%       0.00 ±300%  perf-sched.wait_time.max.ms.__cond_resched.__btrfs_run_delayed_refs.btrfs_run_delayed_refs.flush_space.btrfs_preempt_reclaim_metadata_space
      0.00 ±174%     -64.4%       0.00 ±300%  perf-sched.wait_time.max.ms.__cond_resched.__clear_extent_bit.btrfs_replace_file_extents.btrfs_punch_hole.btrfs_fallocate
      0.00       +1.5e+103%      15.38 ±300%  perf-sched.wait_time.max.ms.__cond_resched.__do_fault.do_read_fault.do_fault.__handle_mm_fault
      0.00 ±174%     -20.0%       0.00 ±201%  perf-sched.wait_time.max.ms.__cond_resched.__kmem_cache_alloc_node.__kmalloc_node_track_caller.memdup_user.do_vfs_ioctl
      0.01 ±125%     -32.6%       0.01 ± 59%  perf-sched.wait_time.max.ms.__cond_resched.__kmem_cache_alloc_node.kmalloc_trace.btrfs_alloc_block_rsv.btrfs_replace_file_extents
      0.00 ±264%    +508.0%       0.00 ±191%  perf-sched.wait_time.max.ms.__cond_resched.__kmem_cache_alloc_node.kmalloc_trace.btrfs_fallocate.vfs_fallocate
      1000            +0.0%       1000        perf-sched.wait_time.max.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
      0.08 ±264%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.__wait_for_common.wait_for_completion_state.kernel_clone.__x64_sys_vfork
      0.00 ±264%     -52.9%       0.00 ±200%  perf-sched.wait_time.max.ms.__cond_resched.btrfs_alloc_path.btrfs_clone.btrfs_clone_files.isra
      0.00        +3.9e+99%       0.00 ± 51%  perf-sched.wait_time.max.ms.__cond_resched.btrfs_alloc_path.btrfs_clone.btrfs_extent_same_range.btrfs_remap_file_range
      0.00       +2.3e+102%       2.33 ±300%  perf-sched.wait_time.max.ms.__cond_resched.btrfs_alloc_path.btrfs_del_csums.cleanup_ref_head.__btrfs_run_delayed_refs
      0.00 ±264%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.btrfs_alloc_path.btrfs_get_extent.find_first_non_hole.btrfs_punch_hole
      0.00 ±174%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.btrfs_alloc_path.btrfs_punch_hole.btrfs_fallocate.vfs_fallocate
      0.00        +2.6e+99%       0.00 ±120%  perf-sched.wait_time.max.ms.__cond_resched.btrfs_alloc_path.insert_prealloc_file_extent.__btrfs_prealloc_file_range.btrfs_prealloc_file_range
      0.00 ±264%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.btrfs_search_slot.btrfs_clone.btrfs_clone_files.isra
      0.00          +6e+98%       0.00 ±300%  perf-sched.wait_time.max.ms.__cond_resched.btrfs_search_slot.btrfs_clone.btrfs_extent_same_range.btrfs_remap_file_range
      0.00         +1e+103%      10.44 ±299%  perf-sched.wait_time.max.ms.__cond_resched.btrfs_search_slot.btrfs_del_csums.cleanup_ref_head.__btrfs_run_delayed_refs
      0.03 ± 85%    +731.0%       0.27 ±273%  perf-sched.wait_time.max.ms.__cond_resched.btrfs_search_slot.btrfs_insert_empty_items.btrfs_insert_replace_extent.btrfs_replace_file_extents
      0.01 ± 99%     -77.1%       0.00 ±152%  perf-sched.wait_time.max.ms.__cond_resched.btrfs_search_slot.btrfs_lookup_file_extent.btrfs_drop_extents.btrfs_replace_file_extents
      0.00 ±264%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.btrfs_search_slot.btrfs_lookup_file_extent.btrfs_get_extent.find_first_non_hole
     18.90 ±264%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.cgroup_rstat_flush_locked.cgroup_rstat_flush.do_flush_stats.flush_memcg_stats_dwork
      0.00 ±105%     -15.6%       0.00 ±170%  perf-sched.wait_time.max.ms.__cond_resched.down_read.__btrfs_tree_read_lock.btrfs_read_lock_root_node.btrfs_search_slot
      0.00 ±174%      -2.2%       0.00 ±201%  perf-sched.wait_time.max.ms.__cond_resched.down_read.find_free_extent.btrfs_reserve_extent.__btrfs_prealloc_file_range
      0.00 ±264%      +0.0%       0.00 ±300%  perf-sched.wait_time.max.ms.__cond_resched.down_read.unmap_mapping_range.truncate_pagecache_range.btrfs_punch_hole_lock_range
      0.00 ± 89%     +30.7%       0.00 ± 95%  perf-sched.wait_time.max.ms.__cond_resched.down_write.__btrfs_tree_lock.btrfs_lock_root_node.btrfs_search_slot
      0.00          +5e+98%       0.00 ±300%  perf-sched.wait_time.max.ms.__cond_resched.down_write.btrfs_inode_lock.btrfs_punch_hole.btrfs_fallocate
      0.00 ±264%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.down_write.btrfs_inode_lock.btrfs_remap_file_range.do_clone_file_range
      0.00 ±264%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.down_write.btrfs_remap_file_range.do_clone_file_range.vfs_clone_file_range
      0.00 ±174%     -70.9%       0.00 ±300%  perf-sched.wait_time.max.ms.__cond_resched.down_write.lock_two_inodes.btrfs_remap_file_range.do_clone_file_range
      0.00 ±264%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.down_write.lock_two_inodes.btrfs_remap_file_range.vfs_dedupe_file_range_one
      0.00          +4e+98%       0.00 ±300%  perf-sched.wait_time.max.ms.__cond_resched.down_write.vfs_fileattr_set.do_vfs_ioctl.__x64_sys_ioctl
      0.18 ±251%     -96.5%       0.01 ± 31%  perf-sched.wait_time.max.ms.__cond_resched.find_free_extent.btrfs_reserve_extent.__btrfs_prealloc_file_range.btrfs_prealloc_file_range
      0.00 ± 41%     -13.1%       0.00 ± 76%  perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.alloc_extent_map.__btrfs_prealloc_file_range.btrfs_prealloc_file_range
      0.00 ±196%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.alloc_extent_map.btrfs_get_extent.btrfs_do_readpage
      0.01 ±114%     +58.0%       0.02 ± 82%  perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.alloc_extent_state.__clear_extent_bit.btrfs_clone_files
      0.02 ± 71%      +7.8%       0.03 ± 61%  perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.alloc_extent_state.__clear_extent_bit.btrfs_do_readpage
      0.01 ±122%     -56.2%       0.00 ±101%  perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.alloc_extent_state.__clear_extent_bit.btrfs_extent_same_range
      0.00 ±105%     +11.1%       0.00 ±101%  perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.alloc_extent_state.__clear_extent_bit.btrfs_fallocate
      0.02 ±136%     +68.4%       0.04 ±226%  perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.alloc_extent_state.__clear_extent_bit.btrfs_invalidate_folio
      0.00 ±264%     -55.6%       0.00 ±300%  perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.alloc_extent_state.__clear_extent_bit.btrfs_punch_hole
      0.01 ± 96%     -53.5%       0.01 ±102%  perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.alloc_extent_state.__clear_extent_bit.btrfs_replace_file_extents
      0.04 ± 44%    +210.9%       0.11 ±198%  perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.alloc_extent_state.__set_extent_bit.lock_extent
      0.02 ± 72%      +6.8%       0.02 ± 97%  perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.alloc_extent_state.__set_extent_bit.set_extent_bit
      0.00 ±264%     +12.0%       0.00 ±300%  perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.btrfs_add_delayed_data_ref.btrfs_alloc_reserved_file_extent.btrfs_insert_replace_extent
      0.00 ± 79%     +24.8%       0.00 ± 91%  perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.btrfs_add_delayed_data_ref.btrfs_drop_extents.btrfs_replace_file_extents
      0.00 ±100%     -41.8%       0.00 ±153%  perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.btrfs_add_delayed_data_ref.btrfs_insert_replace_extent.btrfs_replace_file_extents
      0.00        +3.7e+99%       0.00 ±265%  perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.btrfs_clone.btrfs_clone_files.isra
      0.00 ±100%     +84.0%       0.00 ± 57%  perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.btrfs_drop_extent_map_range.btrfs_drop_extents.btrfs_replace_file_extents
      0.00 ±264%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.btrfs_drop_extent_map_range.btrfs_replace_extent_map_range.__btrfs_prealloc_file_range
      0.00 ±264%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.btrfs_get_extent.btrfs_do_readpage.btrfs_read_folio
      0.00          +5e+98%       0.00 ±300%  perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.btrfs_punch_hole.btrfs_fallocate.vfs_fallocate
      0.00 ±264%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.insert_prealloc_file_extent.__btrfs_prealloc_file_range.btrfs_prealloc_file_range
      0.00 ± 94%     -46.7%       0.00 ±122%  perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.start_transaction.btrfs_dirty_inode.file_modified
      0.00 ±173%    +520.0%       0.01 ±230%  perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.start_transaction.btrfs_fileattr_set.vfs_fileattr_set
      0.00 ±133%     +42.9%       0.00 ±100%  perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.start_transaction.btrfs_replace_file_extents.btrfs_clone
      0.00 ±264%     -27.3%       0.00 ±300%  perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.start_transaction.btrfs_replace_file_extents.btrfs_punch_hole
      0.00 ±174%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.start_transaction.btrfs_replace_file_extents.insert_prealloc_file_extent
      0.00 ±264%     +20.0%       0.00 ±300%  perf-sched.wait_time.max.ms.__cond_resched.mnt_want_write_file.do_vfs_ioctl.__x64_sys_ioctl.do_syscall_64
      0.00          +1e+99%       0.00 ±204%  perf-sched.wait_time.max.ms.__cond_resched.mnt_want_write_file.vfs_dedupe_file_range_one.vfs_dedupe_file_range.do_vfs_ioctl
      0.00 ±136%     -20.0%       0.00 ±153%  perf-sched.wait_time.max.ms.__cond_resched.mutex_lock.__btrfs_release_delayed_node.part.0
      0.00 ±264%     +20.0%       0.00 ±300%  perf-sched.wait_time.max.ms.__cond_resched.mutex_lock.btrfs_delayed_update_inode.btrfs_update_inode.__btrfs_prealloc_file_range
      0.00 ±264%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.mutex_lock.btrfs_delayed_update_inode.btrfs_update_inode.btrfs_dirty_inode
      0.00 ±264%    +380.0%       0.00 ±300%  perf-sched.wait_time.max.ms.__cond_resched.mutex_lock.btrfs_delayed_update_inode.btrfs_update_inode.btrfs_fileattr_set
      0.00 ±264%    +252.0%       0.00 ±245%  perf-sched.wait_time.max.ms.__cond_resched.mutex_lock.btrfs_delayed_update_inode.btrfs_update_inode.btrfs_punch_hole
      0.00 ±173%    +220.0%       0.00 ± 52%  perf-sched.wait_time.max.ms.__cond_resched.mutex_lock.btrfs_delayed_update_inode.btrfs_update_inode.clone_finish_inode_update
      0.00         +1e+104%     100.10 ±299%  perf-sched.wait_time.max.ms.__cond_resched.mutex_lock.perf_poll.do_poll.constprop
      0.00         +4e+101%       0.40 ±300%  perf-sched.wait_time.max.ms.__cond_resched.rcu_gp_cleanup.rcu_gp_kthread.kthread.ret_from_fork
      0.00         +1e+104%     100.02 ±300%  perf-sched.wait_time.max.ms.__cond_resched.shrink_dentry_list.shrink_dcache_parent.d_invalidate.proc_invalidate_siblings_dcache
    251.12 ±171%    +179.6%     702.09 ±128%  perf-sched.wait_time.max.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      0.00          +4e+98%       0.00 ±300%  perf-sched.wait_time.max.ms.__cond_resched.start_transaction.btrfs_replace_file_extents.btrfs_clone.btrfs_extent_same_range
      0.00        +5.8e+99%       0.01 ±300%  perf-sched.wait_time.max.ms.__cond_resched.task_work_run.exit_to_user_mode_loop.exit_to_user_mode_prepare.syscall_exit_to_user_mode
      0.03 ± 80%    +209.9%       0.08 ±208%  perf-sched.wait_time.max.ms.__cond_resched.truncate_inode_pages_range.btrfs_clone_files.isra.0
      0.00        +1.4e+99%       0.00 ±153%  perf-sched.wait_time.max.ms.__cond_resched.truncate_inode_pages_range.btrfs_punch_hole_lock_range.btrfs_punch_hole.btrfs_fallocate
      0.00 ± 43%      -2.5%       0.00 ± 77%  perf-sched.wait_time.max.ms.__cond_resched.vfs_clone_file_range.ioctl_file_clone.do_vfs_ioctl.__x64_sys_ioctl
      0.08 ±209%     -77.4%       0.02 ±111%  perf-sched.wait_time.max.ms.__cond_resched.vfs_dedupe_file_range_compare.constprop.0.__generic_remap_file_range_prep
      0.00 ± 40%     -37.8%       0.00 ± 82%  perf-sched.wait_time.max.ms.__cond_resched.vfs_fallocate.ioctl_preallocate.__x64_sys_ioctl.do_syscall_64
    999.86            -0.0%     999.86        perf-sched.wait_time.max.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
      0.22 ±173%    +141.8%       0.53 ± 82%  perf-sched.wait_time.max.ms.devkmsg_read.vfs_read.ksys_read.do_syscall_64
      1000            +0.0%       1000        perf-sched.wait_time.max.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
      1868 ± 49%     -19.8%       1498 ± 44%  perf-sched.wait_time.max.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
      0.86 ±  3%      +1.6%       0.87 ±  3%  perf-sched.wait_time.max.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
      0.00         +1e+104%     100.02 ±300%  perf-sched.wait_time.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_exc_page_fault
      0.01 ± 14%    +127.8%       0.01 ±127%  perf-sched.wait_time.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt
      0.09 ±198%     -58.3%       0.04 ± 70%  perf-sched.wait_time.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi
      0.08 ±214%     -83.2%       0.01 ±104%  perf-sched.wait_time.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.syscall_exit_to_user_mode.do_syscall_64
      0.02 ± 81%      -3.8%       0.02 ± 91%  perf-sched.wait_time.max.ms.handle_reserve_ticket.__reserve_bytes.btrfs_reserve_data_bytes.btrfs_fallocate
      0.00 ±174%     -28.9%       0.00 ±200%  perf-sched.wait_time.max.ms.io_schedule.folio_wait_bit_common.filemap_fault.__do_fault
      1008            -0.0%       1008        perf-sched.wait_time.max.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
      1000            -0.0%       1000        perf-sched.wait_time.max.ms.schedule_hrtimeout_range_clock.do_poll.constprop.0.do_sys_poll
    531.24 ±  7%      -5.9%     499.98        perf-sched.wait_time.max.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
      1.34 ± 92%     -23.4%       1.02 ± 82%  perf-sched.wait_time.max.ms.schedule_preempt_disabled.rwsem_down_read_slowpath.down_read.__btrfs_tree_read_lock
      0.04 ± 54%    +806.7%       0.39 ±206%  perf-sched.wait_time.max.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.__btrfs_tree_lock
      0.00        +2.1e+99%       0.00 ±157%  perf-sched.wait_time.max.ms.schedule_timeout.__wait_for_common.__flush_work.isra.0
      0.00 ±188%     -70.5%       0.00 ±300%  perf-sched.wait_time.max.ms.schedule_timeout.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr
      0.93            +0.4%       0.94        perf-sched.wait_time.max.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
    504.62            +0.0%     504.70        perf-sched.wait_time.max.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
      0.00         +5e+101%       0.50 ±202%  perf-sched.wait_time.max.ms.schedule_timeout.khugepaged_wait_work.khugepaged.kthread
      5.00            +0.0%       5.00        perf-sched.wait_time.max.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
      2794 ± 23%      +7.7%       3011 ± 18%  perf-sched.wait_time.max.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      0.22 ±173%    +141.7%       0.54 ± 82%  perf-sched.wait_time.max.ms.syslog_print.do_syslog.kmsg_read.vfs_read
      0.00 ±264%    +100.0%       0.00 ±200%  perf-sched.wait_time.max.ms.wait_for_partner.fifo_open.do_dentry_open.do_open
      2561 ± 18%     +14.4%       2930 ± 33%  perf-sched.wait_time.max.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
    860.62 ± 13%      -2.8%     836.40 ± 18%  slabinfo.Acpi-State.active_objs
     16.88 ± 13%      -2.8%      16.40 ± 18%  slabinfo.Acpi-State.active_slabs
    860.62 ± 13%      -2.8%     836.40 ± 18%  slabinfo.Acpi-State.num_objs
     16.88 ± 13%      -2.8%      16.40 ± 18%  slabinfo.Acpi-State.num_slabs
     36.00            +0.0%      36.00        slabinfo.DCCP.active_objs
      2.00            +0.0%       2.00        slabinfo.DCCP.active_slabs
     36.00            +0.0%      36.00        slabinfo.DCCP.num_objs
      2.00            +0.0%       2.00        slabinfo.DCCP.num_slabs
     34.00            +0.0%      34.00        slabinfo.DCCPv6.active_objs
      2.00            +0.0%       2.00        slabinfo.DCCPv6.active_slabs
     34.00            +0.0%      34.00        slabinfo.DCCPv6.num_objs
      2.00            +0.0%       2.00        slabinfo.DCCPv6.num_slabs
     96.00            +0.0%      96.00        slabinfo.RAW.active_objs
      3.00            +0.0%       3.00        slabinfo.RAW.active_slabs
     96.00            +0.0%      96.00        slabinfo.RAW.num_objs
      3.00            +0.0%       3.00        slabinfo.RAW.num_slabs
     78.00            +6.7%      83.20 ± 12%  slabinfo.RAWv6.active_objs
      3.00            +6.7%       3.20 ± 12%  slabinfo.RAWv6.active_slabs
     78.00            +6.7%      83.20 ± 12%  slabinfo.RAWv6.num_objs
      3.00            +6.7%       3.20 ± 12%  slabinfo.RAWv6.num_slabs
     54.83 ±  8%      +3.6%      56.80        slabinfo.TCP.active_objs
      3.92 ±  8%      +3.6%       4.06        slabinfo.TCP.active_slabs
     54.83 ±  8%      +3.6%      56.80        slabinfo.TCP.num_objs
      3.92 ±  8%      +3.6%       4.06        slabinfo.TCP.num_slabs
     37.38 ± 11%      -2.6%      36.40 ± 14%  slabinfo.TCPv6.active_objs
      2.88 ± 11%      -2.6%       2.80 ± 14%  slabinfo.TCPv6.active_slabs
     37.38 ± 11%      -2.6%      36.40 ± 14%  slabinfo.TCPv6.num_objs
      2.88 ± 11%      -2.6%       2.80 ± 14%  slabinfo.TCPv6.num_slabs
    105.00 ± 11%      -1.7%     103.20 ± 14%  slabinfo.UDPv6.active_objs
      4.38 ± 11%      -1.7%       4.30 ± 14%  slabinfo.UDPv6.active_slabs
    105.00 ± 11%      -1.7%     103.20 ± 14%  slabinfo.UDPv6.num_objs
      4.38 ± 11%      -1.7%       4.30 ± 14%  slabinfo.UDPv6.num_slabs
    986.19 ±  4%      +2.2%       1008 ±  4%  slabinfo.UNIX.active_objs
     32.87 ±  4%      +2.2%      33.60 ±  4%  slabinfo.UNIX.active_slabs
    986.19 ±  4%      +2.2%       1008 ±  4%  slabinfo.UNIX.num_objs
     32.87 ±  4%      +2.2%      33.60 ±  4%  slabinfo.UNIX.num_slabs
      6255 ±  3%      -0.8%       6207 ±  5%  slabinfo.anon_vma.active_objs
    163.36 ±  3%      -1.2%     161.37 ±  5%  slabinfo.anon_vma.active_slabs
      6371 ±  3%      -1.2%       6293 ±  5%  slabinfo.anon_vma.num_objs
    163.36 ±  3%      -1.2%     161.37 ±  5%  slabinfo.anon_vma.num_slabs
      7503            +2.4%       7681 ±  4%  slabinfo.anon_vma_chain.active_objs
    119.97            +1.6%     121.86 ±  4%  slabinfo.anon_vma_chain.active_slabs
      7678            +1.6%       7798 ±  4%  slabinfo.anon_vma_chain.num_objs
    119.97            +1.6%     121.86 ±  4%  slabinfo.anon_vma_chain.num_slabs
     70.00 ± 14%     +14.3%      80.00 ± 15%  slabinfo.bdev_cache.active_objs
      3.50 ± 14%     +14.3%       4.00 ± 15%  slabinfo.bdev_cache.active_slabs
     70.00 ± 14%     +14.3%      80.00 ± 15%  slabinfo.bdev_cache.num_objs
      3.50 ± 14%     +14.3%       4.00 ± 15%  slabinfo.bdev_cache.num_slabs
     60.55 ± 16%     +15.3%      69.81 ± 10%  slabinfo.bio-120.active_objs
      1.89 ± 16%     +15.3%       2.18 ± 10%  slabinfo.bio-120.active_slabs
     60.55 ± 16%     +15.3%      69.81 ± 10%  slabinfo.bio-120.num_objs
      1.89 ± 16%     +15.3%       2.18 ± 10%  slabinfo.bio-120.num_slabs
    990.14 ±  8%      +7.9%       1068 ±  6%  slabinfo.bio-184.active_objs
     23.57 ±  8%      +7.9%      25.45 ±  6%  slabinfo.bio-184.active_slabs
    990.14 ±  8%      +7.9%       1068 ±  6%  slabinfo.bio-184.num_objs
     23.57 ±  8%      +7.9%      25.45 ±  6%  slabinfo.bio-184.num_slabs
     64.00            +0.0%      64.00        slabinfo.bio-248.active_objs
      2.00            +0.0%       2.00        slabinfo.bio-248.active_slabs
     64.00            +0.0%      64.00        slabinfo.bio-248.num_objs
      2.00            +0.0%       2.00        slabinfo.bio-248.num_slabs
     51.00            +0.0%      51.00        slabinfo.bio-296.active_objs
      1.00            +0.0%       1.00        slabinfo.bio-296.active_slabs
     51.00            +0.0%      51.00        slabinfo.bio-296.num_objs
      1.00            +0.0%       1.00        slabinfo.bio-296.num_slabs
    176.79 ±  8%      +7.2%     189.60 ± 10%  slabinfo.bio-360.active_objs
      4.21 ±  8%      +7.2%       4.51 ± 10%  slabinfo.bio-360.active_slabs
    176.79 ±  8%      +7.2%     189.60 ± 10%  slabinfo.bio-360.num_objs
      4.21 ±  8%      +7.2%       4.51 ± 10%  slabinfo.bio-360.num_slabs
     42.00            +0.0%      42.00        slabinfo.bio-376.active_objs
      1.00            +0.0%       1.00        slabinfo.bio-376.active_slabs
     42.00            +0.0%      42.00        slabinfo.bio-376.num_objs
      1.00            +0.0%       1.00        slabinfo.bio-376.num_slabs
     36.00            +0.0%      36.00        slabinfo.bio-432.active_objs
      1.00            +0.0%       1.00        slabinfo.bio-432.active_slabs
     36.00            +0.0%      36.00        slabinfo.bio-432.num_objs
      1.00            +0.0%       1.00        slabinfo.bio-432.num_slabs
    170.00            +0.0%     170.00        slabinfo.bio_post_read_ctx.active_objs
      2.00            +0.0%       2.00        slabinfo.bio_post_read_ctx.active_slabs
    170.00            +0.0%     170.00        slabinfo.bio_post_read_ctx.num_objs
      2.00            +0.0%       2.00        slabinfo.bio_post_read_ctx.num_slabs
     52.00 ± 29%      -7.7%      48.00 ± 53%  slabinfo.biovec-128.active_objs
      3.25 ± 29%      -7.7%       3.00 ± 53%  slabinfo.biovec-128.active_slabs
     52.00 ± 29%      -7.7%      48.00 ± 53%  slabinfo.biovec-128.num_objs
      3.25 ± 29%      -7.7%       3.00 ± 53%  slabinfo.biovec-128.num_slabs
    423.27 ± 11%      -0.3%     421.86 ± 10%  slabinfo.biovec-64.active_objs
     13.23 ± 11%      -0.3%      13.18 ± 10%  slabinfo.biovec-64.active_slabs
    423.27 ± 11%      -0.3%     421.86 ± 10%  slabinfo.biovec-64.num_objs
     13.23 ± 11%      -0.3%      13.18 ± 10%  slabinfo.biovec-64.num_slabs
     56.14            +2.3%      57.45 ±  3%  slabinfo.biovec-max.active_objs
      7.02            +2.3%       7.18 ±  3%  slabinfo.biovec-max.active_slabs
     56.14            +2.3%      57.45 ±  3%  slabinfo.biovec-max.num_objs
      7.02            +2.3%       7.18 ±  3%  slabinfo.biovec-max.num_slabs
      5847 ±129%     +16.9%       6837 ±122%  slabinfo.btrfs_delayed_ref_head.active_objs
    195.09 ±129%     +16.9%     228.11 ±122%  slabinfo.btrfs_delayed_ref_head.active_slabs
      5852 ±129%     +16.9%       6843 ±122%  slabinfo.btrfs_delayed_ref_head.num_objs
    195.09 ±129%     +16.9%     228.11 ±122%  slabinfo.btrfs_delayed_ref_head.num_slabs
    217.62 ± 11%      +5.8%     230.16 ± 11%  slabinfo.btrfs_extent_buffer.active_objs
      6.40 ± 11%      +5.8%       6.77 ± 11%  slabinfo.btrfs_extent_buffer.active_slabs
    217.62 ± 11%      +5.8%     230.16 ± 11%  slabinfo.btrfs_extent_buffer.num_objs
      6.40 ± 11%      +5.8%       6.77 ± 11%  slabinfo.btrfs_extent_buffer.num_slabs
    376.75 ± 12%      +6.1%     399.82 ± 14%  slabinfo.btrfs_free_space.active_objs
      9.66 ± 12%      +6.1%      10.25 ± 14%  slabinfo.btrfs_free_space.active_slabs
    376.75 ± 12%      +6.1%     399.82 ± 14%  slabinfo.btrfs_free_space.num_objs
      9.66 ± 12%      +6.1%      10.25 ± 14%  slabinfo.btrfs_free_space.num_slabs
    212.92 ± 11%      -0.6%     211.70 ±  9%  slabinfo.btrfs_inode.active_objs
      7.34 ± 11%      -0.6%       7.30 ±  9%  slabinfo.btrfs_inode.active_slabs
    212.92 ± 11%      -0.6%     211.70 ±  9%  slabinfo.btrfs_inode.num_objs
      7.34 ± 11%      -0.6%       7.30 ±  9%  slabinfo.btrfs_inode.num_slabs
      0.65 ±174%    +958.1%       6.90 ±126%  slabinfo.btrfs_ordered_extent.active_objs
      0.02 ±174%    +958.1%       0.18 ±126%  slabinfo.btrfs_ordered_extent.active_slabs
      0.65 ±174%    +958.1%       6.90 ±126%  slabinfo.btrfs_ordered_extent.num_objs
      0.02 ±174%    +958.1%       0.18 ±126%  slabinfo.btrfs_ordered_extent.num_slabs
      1128 ±  3%      -0.8%       1119 ±  4%  slabinfo.btrfs_path.active_objs
     31.36 ±  3%      -0.8%      31.09 ±  4%  slabinfo.btrfs_path.active_slabs
      1128 ±  3%      -0.8%       1119 ±  4%  slabinfo.btrfs_path.num_objs
     31.36 ±  3%      -0.8%      31.09 ±  4%  slabinfo.btrfs_path.num_slabs
      1553 ±  3%      +0.3%       1557 ±  2%  slabinfo.buffer_head.active_objs
     39.83 ±  3%      +0.3%      39.94 ±  2%  slabinfo.buffer_head.active_slabs
      1553 ±  3%      +0.3%       1557 ±  2%  slabinfo.buffer_head.num_objs
     39.83 ±  3%      +0.3%      39.94 ±  2%  slabinfo.buffer_head.num_slabs
      2441 ±  4%      -1.9%       2393 ±  4%  slabinfo.cred_jar.active_objs
     58.13 ±  4%      -1.9%      57.00 ±  4%  slabinfo.cred_jar.active_slabs
      2441 ±  4%      -1.9%       2393 ±  4%  slabinfo.cred_jar.num_objs
     58.13 ±  4%      -1.9%      57.00 ±  4%  slabinfo.cred_jar.num_slabs
     39.00            +0.0%      39.00        slabinfo.dax_cache.active_objs
      1.00            +0.0%       1.00        slabinfo.dax_cache.active_slabs
     39.00            +0.0%      39.00        slabinfo.dax_cache.num_objs
      1.00            +0.0%       1.00        slabinfo.dax_cache.num_slabs
     75745            -0.0%      75742        slabinfo.dentry.active_objs
      1827            -0.1%       1825        slabinfo.dentry.active_slabs
     76744            -0.1%      76680        slabinfo.dentry.num_objs
      1827            -0.1%       1825        slabinfo.dentry.num_slabs
     30.00            +0.0%      30.00        slabinfo.dmaengine-unmap-128.active_objs
      1.00            +0.0%       1.00        slabinfo.dmaengine-unmap-128.active_slabs
     30.00            +0.0%      30.00        slabinfo.dmaengine-unmap-128.num_objs
      1.00            +0.0%       1.00        slabinfo.dmaengine-unmap-128.num_slabs
     64.00            +0.0%      64.00        slabinfo.dmaengine-unmap-2.active_objs
      1.00            +0.0%       1.00        slabinfo.dmaengine-unmap-2.active_slabs
     64.00            +0.0%      64.00        slabinfo.dmaengine-unmap-2.num_objs
      1.00            +0.0%       1.00        slabinfo.dmaengine-unmap-2.num_slabs
     15.00            +0.0%      15.00        slabinfo.dmaengine-unmap-256.active_objs
      1.00            +0.0%       1.00        slabinfo.dmaengine-unmap-256.active_slabs
     15.00            +0.0%      15.00        slabinfo.dmaengine-unmap-256.num_objs
      1.00            +0.0%       1.00        slabinfo.dmaengine-unmap-256.num_slabs
      7872 ±  6%      +0.8%       7936 ±  5%  slabinfo.ep_head.active_objs
     30.75 ±  6%      +0.8%      31.00 ±  5%  slabinfo.ep_head.active_slabs
      7872 ±  6%      +0.8%       7936 ±  5%  slabinfo.ep_head.num_objs
     30.75 ±  6%      +0.8%      31.00 ±  5%  slabinfo.ep_head.num_slabs
    693.75 ± 14%      -2.8%     674.48 ± 14%  slabinfo.file_lock_cache.active_objs
     18.75 ± 14%      -2.8%      18.23 ± 14%  slabinfo.file_lock_cache.active_slabs
    693.75 ± 14%      -2.8%     674.48 ± 14%  slabinfo.file_lock_cache.num_objs
     18.75 ± 14%      -2.8%      18.23 ± 14%  slabinfo.file_lock_cache.num_slabs
      1684            -0.3%       1679        slabinfo.files_cache.active_objs
     36.62            -0.3%      36.50        slabinfo.files_cache.active_slabs
      1684            -0.3%       1679        slabinfo.files_cache.num_objs
     36.62            -0.3%      36.50        slabinfo.files_cache.num_slabs
      5932 ±  2%      +0.5%       5964        slabinfo.filp.active_objs
    282.47 ±  2%      +0.4%     283.61 ±  2%  slabinfo.filp.active_slabs
      9039 ±  2%      +0.4%       9075 ±  2%  slabinfo.filp.num_objs
    282.47 ±  2%      +0.4%     283.61 ±  2%  slabinfo.filp.num_slabs
      2371 ±  8%      +7.7%       2553 ± 11%  slabinfo.fsnotify_mark_connector.active_objs
     18.53 ±  8%      +7.7%      19.95 ± 11%  slabinfo.fsnotify_mark_connector.active_slabs
      2371 ±  8%      +7.7%       2553 ± 11%  slabinfo.fsnotify_mark_connector.num_objs
     18.53 ±  8%      +7.7%      19.95 ± 11%  slabinfo.fsnotify_mark_connector.num_slabs
      6433            -0.4%       6409        slabinfo.ftrace_event_field.active_objs
     88.12            -0.4%      87.80        slabinfo.ftrace_event_field.active_slabs
      6433            -0.4%       6409        slabinfo.ftrace_event_field.num_objs
     88.12            -0.4%      87.80        slabinfo.ftrace_event_field.num_slabs
    942.93 ±  3%      -1.7%     926.82 ±  3%  slabinfo.fuse_request.active_objs
     33.68 ±  3%      -1.7%      33.10 ±  3%  slabinfo.fuse_request.active_slabs
    942.93 ±  3%      -1.7%     926.82 ±  3%  slabinfo.fuse_request.num_objs
     33.68 ±  3%      -1.7%      33.10 ±  3%  slabinfo.fuse_request.num_slabs
     98.00            +0.0%      98.00        slabinfo.hugetlbfs_inode_cache.active_objs
      2.00            +0.0%       2.00        slabinfo.hugetlbfs_inode_cache.active_slabs
     98.00            +0.0%      98.00        slabinfo.hugetlbfs_inode_cache.num_objs
      2.00            +0.0%       2.00        slabinfo.hugetlbfs_inode_cache.num_slabs
     64153            +0.1%      64202        slabinfo.inode_cache.active_objs
      1260            +0.1%       1262        slabinfo.inode_cache.active_slabs
     64287            +0.1%      64367        slabinfo.inode_cache.num_objs
      1260            +0.1%       1262        slabinfo.inode_cache.num_slabs
    191.62 ± 26%      -8.6%     175.20 ± 20%  slabinfo.ip_fib_alias.active_objs
      2.62 ± 26%      -8.6%       2.40 ± 20%  slabinfo.ip_fib_alias.active_slabs
    191.62 ± 26%      -8.6%     175.20 ± 20%  slabinfo.ip_fib_alias.num_objs
      2.62 ± 26%      -8.6%       2.40 ± 20%  slabinfo.ip_fib_alias.num_slabs
    223.12 ± 26%      -8.6%     204.00 ± 20%  slabinfo.ip_fib_trie.active_objs
      2.62 ± 26%      -8.6%       2.40 ± 20%  slabinfo.ip_fib_trie.active_slabs
    223.12 ± 26%      -8.6%     204.00 ± 20%  slabinfo.ip_fib_trie.num_objs
      2.62 ± 26%      -8.6%       2.40 ± 20%  slabinfo.ip_fib_trie.num_slabs
     47676            -0.1%      47638        slabinfo.kernfs_node_cache.active_objs
      1489            -0.1%       1488        slabinfo.kernfs_node_cache.active_slabs
     47676            -0.1%      47638        slabinfo.kernfs_node_cache.num_objs
      1489            -0.1%       1488        slabinfo.kernfs_node_cache.num_slabs
      2852 ±  5%      -1.8%       2800 ±  9%  slabinfo.khugepaged_mm_slot.active_objs
     27.96 ±  5%      -1.8%      27.46 ±  9%  slabinfo.khugepaged_mm_slot.active_slabs
      2852 ±  5%      -1.8%       2800 ±  9%  slabinfo.khugepaged_mm_slot.num_objs
     27.96 ±  5%      -1.8%      27.46 ±  9%  slabinfo.khugepaged_mm_slot.num_slabs
      2123            +0.0%       2124        slabinfo.kmalloc-128.active_objs
     66.37            +0.0%      66.39        slabinfo.kmalloc-128.active_slabs
      2123            +0.0%       2124        slabinfo.kmalloc-128.num_objs
     66.37            +0.0%      66.39        slabinfo.kmalloc-128.num_slabs
     18176            -0.7%      18048        slabinfo.kmalloc-16.active_objs
     71.00            -0.7%      70.50        slabinfo.kmalloc-16.active_slabs
     18176            -0.7%      18048        slabinfo.kmalloc-16.num_objs
     71.00            -0.7%      70.50        slabinfo.kmalloc-16.num_slabs
      2713 ±  4%      -1.0%       2685 ±  2%  slabinfo.kmalloc-192.active_objs
     64.60 ±  4%      -1.0%      63.95 ±  2%  slabinfo.kmalloc-192.active_slabs
      2713 ±  4%      -1.0%       2685 ±  2%  slabinfo.kmalloc-192.num_objs
     64.60 ±  4%      -1.0%      63.95 ±  2%  slabinfo.kmalloc-192.num_slabs
      2439            -0.4%       2428        slabinfo.kmalloc-1k.active_objs
     76.23            -0.4%      75.89        slabinfo.kmalloc-1k.active_slabs
      2439            -0.4%       2428        slabinfo.kmalloc-1k.num_objs
     76.23            -0.4%      75.89        slabinfo.kmalloc-1k.num_slabs
      4185 ±  2%      +2.5%       4291        slabinfo.kmalloc-256.active_objs
    131.05 ±  2%      +2.7%     134.64        slabinfo.kmalloc-256.active_slabs
      4193 ±  2%      +2.7%       4308        slabinfo.kmalloc-256.num_objs
    131.05 ±  2%      +2.7%     134.64        slabinfo.kmalloc-256.num_slabs
      1734 ±  2%      -0.5%       1725 ±  2%  slabinfo.kmalloc-2k.active_objs
    111.17 ±  2%      -0.8%     110.32 ±  2%  slabinfo.kmalloc-2k.active_slabs
      1778 ±  2%      -0.8%       1765 ±  2%  slabinfo.kmalloc-2k.num_objs
    111.17 ±  2%      -0.8%     110.32 ±  2%  slabinfo.kmalloc-2k.num_slabs
     16923            +1.2%      17130        slabinfo.kmalloc-32.active_objs
    132.22            +1.3%     133.91        slabinfo.kmalloc-32.active_slabs
     16923            +1.3%      17140        slabinfo.kmalloc-32.num_objs
    132.22            +1.3%     133.91        slabinfo.kmalloc-32.num_slabs
    649.65            -0.5%     646.70        slabinfo.kmalloc-4k.active_objs
     82.06            -0.3%      81.81        slabinfo.kmalloc-4k.active_slabs
    656.49            -0.3%     654.49        slabinfo.kmalloc-4k.num_objs
     82.06            -0.3%      81.81        slabinfo.kmalloc-4k.num_slabs
      5183 ±  2%      +0.9%       5231 ±  2%  slabinfo.kmalloc-512.active_objs
    162.48 ±  2%      +0.9%     163.93 ±  2%  slabinfo.kmalloc-512.active_slabs
      5199 ±  2%      +0.9%       5245 ±  2%  slabinfo.kmalloc-512.num_objs
    162.48 ±  2%      +0.9%     163.93 ±  2%  slabinfo.kmalloc-512.num_slabs
     19017            -0.1%      18996        slabinfo.kmalloc-64.active_objs
    297.15            -0.1%     296.82        slabinfo.kmalloc-64.active_slabs
     19017            -0.1%      18996        slabinfo.kmalloc-64.num_objs
    297.15            -0.1%     296.82        slabinfo.kmalloc-64.num_slabs
     33657            -0.2%      33580        slabinfo.kmalloc-8.active_objs
     65.75            -0.2%      65.60        slabinfo.kmalloc-8.active_slabs
     33664            -0.2%      33587        slabinfo.kmalloc-8.num_objs
     65.75            -0.2%      65.60        slabinfo.kmalloc-8.num_slabs
    197.38            -0.5%     196.45        slabinfo.kmalloc-8k.active_objs
     50.72            -0.1%      50.65        slabinfo.kmalloc-8k.active_slabs
    202.86            -0.1%     202.60        slabinfo.kmalloc-8k.num_objs
     50.72            -0.1%      50.65        slabinfo.kmalloc-8k.num_slabs
     20981            -0.6%      20845        slabinfo.kmalloc-96.active_objs
    517.73            -0.8%     513.39        slabinfo.kmalloc-96.active_slabs
     21744            -0.8%      21562        slabinfo.kmalloc-96.num_objs
    517.73            -0.8%     513.39        slabinfo.kmalloc-96.num_slabs
    447.63 ± 10%     +10.6%     495.16 ±  6%  slabinfo.kmalloc-cg-128.active_objs
     13.99 ± 10%     +10.6%      15.47 ±  6%  slabinfo.kmalloc-cg-128.active_slabs
    447.63 ± 10%     +10.6%     495.16 ±  6%  slabinfo.kmalloc-cg-128.num_objs
     13.99 ± 10%     +10.6%      15.47 ±  6%  slabinfo.kmalloc-cg-128.num_slabs
      3257 ± 10%      +8.9%       3547 ±  8%  slabinfo.kmalloc-cg-16.active_objs
     12.73 ± 10%      +8.9%      13.86 ±  8%  slabinfo.kmalloc-cg-16.active_slabs
      3257 ± 10%      +8.9%       3547 ±  8%  slabinfo.kmalloc-cg-16.num_objs
     12.73 ± 10%      +8.9%      13.86 ±  8%  slabinfo.kmalloc-cg-16.num_slabs
      1638            +0.5%       1646        slabinfo.kmalloc-cg-192.active_objs
     39.00            +0.5%      39.20        slabinfo.kmalloc-cg-192.active_slabs
      1638            +0.5%       1646        slabinfo.kmalloc-cg-192.num_objs
     39.00            +0.5%      39.20        slabinfo.kmalloc-cg-192.num_slabs
      1568 ±  3%      +3.5%       1622 ±  5%  slabinfo.kmalloc-cg-1k.active_objs
     49.00 ±  3%      +3.5%      50.70 ±  5%  slabinfo.kmalloc-cg-1k.active_slabs
      1568 ±  3%      +3.5%       1622 ±  5%  slabinfo.kmalloc-cg-1k.num_objs
     49.00 ±  3%      +3.5%      50.70 ±  5%  slabinfo.kmalloc-cg-1k.num_slabs
    280.00 ±  9%      -0.6%     278.40 ±  8%  slabinfo.kmalloc-cg-256.active_objs
      8.75 ±  9%      -0.6%       8.70 ±  8%  slabinfo.kmalloc-cg-256.active_slabs
    280.00 ±  9%      -0.6%     278.40 ±  8%  slabinfo.kmalloc-cg-256.num_objs
      8.75 ±  9%      -0.6%       8.70 ±  8%  slabinfo.kmalloc-cg-256.num_slabs
    530.09 ±  7%      -0.7%     526.12 ±  6%  slabinfo.kmalloc-cg-2k.active_objs
     33.13 ±  7%      -0.7%      32.88 ±  6%  slabinfo.kmalloc-cg-2k.active_slabs
    530.09 ±  7%      -0.7%     526.12 ±  6%  slabinfo.kmalloc-cg-2k.num_objs
     33.13 ±  7%      -0.7%      32.88 ±  6%  slabinfo.kmalloc-cg-2k.num_slabs
      5010 ±  2%      -0.4%       4991        slabinfo.kmalloc-cg-32.active_objs
     39.14 ±  2%      -0.4%      39.00        slabinfo.kmalloc-cg-32.active_slabs
      5010 ±  2%      -0.4%       4991        slabinfo.kmalloc-cg-32.num_objs
     39.14 ±  2%      -0.4%      39.00        slabinfo.kmalloc-cg-32.num_slabs
    498.48            -1.8%     489.34 ±  3%  slabinfo.kmalloc-cg-4k.active_objs
     65.83            -1.4%      64.93 ±  2%  slabinfo.kmalloc-cg-4k.active_slabs
    526.61            -1.4%     519.45 ±  2%  slabinfo.kmalloc-cg-4k.num_objs
     65.83            -1.4%      64.93 ±  2%  slabinfo.kmalloc-cg-4k.num_slabs
      1188            +0.5%       1193        slabinfo.kmalloc-cg-512.active_objs
     37.12            +0.5%      37.30        slabinfo.kmalloc-cg-512.active_slabs
      1188            +0.5%       1193        slabinfo.kmalloc-cg-512.num_objs
     37.12            +0.5%      37.30        slabinfo.kmalloc-cg-512.num_slabs
      1572 ±  8%      +1.5%       1596 ±  7%  slabinfo.kmalloc-cg-64.active_objs
     24.58 ±  8%      +1.5%      24.94 ±  7%  slabinfo.kmalloc-cg-64.active_slabs
      1572 ±  8%      +1.5%       1596 ±  7%  slabinfo.kmalloc-cg-64.num_objs
     24.58 ±  8%      +1.5%      24.94 ±  7%  slabinfo.kmalloc-cg-64.num_slabs
     18432            +0.0%      18432        slabinfo.kmalloc-cg-8.active_objs
     36.00            +0.0%      36.00        slabinfo.kmalloc-cg-8.active_slabs
     18432            +0.0%      18432        slabinfo.kmalloc-cg-8.num_objs
     36.00            +0.0%      36.00        slabinfo.kmalloc-cg-8.num_slabs
     59.11 ±  6%      -4.1%      56.71 ±  7%  slabinfo.kmalloc-cg-8k.active_objs
     14.78 ±  6%      -4.1%      14.18 ±  7%  slabinfo.kmalloc-cg-8k.active_slabs
     59.11 ±  6%      -4.1%      56.71 ±  7%  slabinfo.kmalloc-cg-8k.num_objs
     14.78 ±  6%      -4.1%      14.18 ±  7%  slabinfo.kmalloc-cg-8k.num_slabs
    969.83 ± 11%      -0.1%     969.09 ±  6%  slabinfo.kmalloc-cg-96.active_objs
     23.09 ± 11%      -0.1%      23.07 ±  6%  slabinfo.kmalloc-cg-96.active_slabs
    969.83 ± 11%      -0.1%     969.09 ±  6%  slabinfo.kmalloc-cg-96.num_objs
     23.09 ± 11%      -0.1%      23.07 ±  6%  slabinfo.kmalloc-cg-96.num_slabs
     88.00 ± 24%      +9.1%      96.00 ± 21%  slabinfo.kmalloc-rcl-128.active_objs
      2.75 ± 24%      +9.1%       3.00 ± 21%  slabinfo.kmalloc-rcl-128.active_slabs
     88.00 ± 24%      +9.1%      96.00 ± 21%  slabinfo.kmalloc-rcl-128.num_objs
      2.75 ± 24%      +9.1%       3.00 ± 21%  slabinfo.kmalloc-rcl-128.num_slabs
     42.00            +0.0%      42.00        slabinfo.kmalloc-rcl-192.active_objs
      1.00            +0.0%       1.00        slabinfo.kmalloc-rcl-192.active_slabs
     42.00            +0.0%      42.00        slabinfo.kmalloc-rcl-192.num_objs
      1.00            +0.0%       1.00        slabinfo.kmalloc-rcl-192.num_slabs
      6310 ±  2%      +1.7%       6416 ±  2%  slabinfo.kmalloc-rcl-64.active_objs
    100.99 ±  2%      +0.4%     101.41 ±  2%  slabinfo.kmalloc-rcl-64.active_slabs
      6463 ±  2%      +0.4%       6490 ±  2%  slabinfo.kmalloc-rcl-64.num_objs
    100.99 ±  2%      +0.4%     101.41 ±  2%  slabinfo.kmalloc-rcl-64.num_slabs
      1386 ±  7%      -0.9%       1373 ±  7%  slabinfo.kmalloc-rcl-96.active_objs
     33.00 ±  7%      -0.9%      32.70 ±  7%  slabinfo.kmalloc-rcl-96.active_slabs
      1386 ±  7%      -0.9%       1373 ±  7%  slabinfo.kmalloc-rcl-96.num_objs
     33.00 ±  7%      -0.9%      32.70 ±  7%  slabinfo.kmalloc-rcl-96.num_slabs
    416.00 ±  9%      -2.3%     406.40 ±  9%  slabinfo.kmem_cache.active_objs
     13.00 ±  9%      -2.3%      12.70 ±  9%  slabinfo.kmem_cache.active_slabs
    416.00 ±  9%      -2.3%     406.40 ±  9%  slabinfo.kmem_cache.num_objs
     13.00 ±  9%      -2.3%      12.70 ±  9%  slabinfo.kmem_cache.num_slabs
    632.00 ± 11%      -1.8%     620.80 ± 13%  slabinfo.kmem_cache_node.active_objs
      9.88 ± 11%      -1.8%       9.70 ± 13%  slabinfo.kmem_cache_node.active_slabs
    632.00 ± 11%      -1.8%     620.80 ± 13%  slabinfo.kmem_cache_node.num_objs
      9.88 ± 11%      -1.8%       9.70 ± 13%  slabinfo.kmem_cache_node.num_slabs
      9711 ± 77%      +4.6%      10157 ± 81%  slabinfo.kvm_async_pf.active_objs
    324.09 ± 77%      +4.5%     338.80 ± 81%  slabinfo.kvm_async_pf.active_slabs
      9722 ± 77%      +4.5%      10163 ± 81%  slabinfo.kvm_async_pf.num_objs
    324.09 ± 77%      +4.5%     338.80 ± 81%  slabinfo.kvm_async_pf.num_slabs
      7974 ±  5%      -0.8%       7909 ±  4%  slabinfo.lsm_file_cache.active_objs
     47.18 ±  4%      -1.0%      46.72 ±  4%  slabinfo.lsm_file_cache.active_slabs
      8021 ±  4%      -1.0%       7941 ±  4%  slabinfo.lsm_file_cache.num_objs
     47.18 ±  4%      -1.0%      46.72 ±  4%  slabinfo.lsm_file_cache.num_slabs
      4530            -1.2%       4476 ±  3%  slabinfo.maple_node.active_objs
    164.60 ±  2%      -2.0%     161.37 ±  5%  slabinfo.maple_node.active_slabs
      5267 ±  2%      -2.0%       5163 ±  5%  slabinfo.maple_node.num_objs
    164.60 ±  2%      -2.0%     161.37 ±  5%  slabinfo.maple_node.num_slabs
    903.00            +0.5%     907.20        slabinfo.mm_struct.active_objs
     37.62            +0.5%      37.80        slabinfo.mm_struct.active_slabs
    903.00            +0.5%     907.20        slabinfo.mm_struct.num_objs
     37.62            +0.5%      37.80        slabinfo.mm_struct.num_slabs
    930.36 ±  6%      +4.0%     967.46 ±  8%  slabinfo.mnt_cache.active_objs
     18.24 ±  6%      +4.0%      18.97 ±  8%  slabinfo.mnt_cache.active_slabs
    930.36 ±  6%      +4.0%     967.46 ±  8%  slabinfo.mnt_cache.num_objs
     18.24 ±  6%      +4.0%      18.97 ±  8%  slabinfo.mnt_cache.num_slabs
     34.00            +0.0%      34.00        slabinfo.mqueue_inode_cache.active_objs
      1.00            +0.0%       1.00        slabinfo.mqueue_inode_cache.active_slabs
     34.00            +0.0%      34.00        slabinfo.mqueue_inode_cache.num_objs
      1.00            +0.0%       1.00        slabinfo.mqueue_inode_cache.num_slabs
    290.00            -0.7%     288.00        slabinfo.names_cache.active_objs
     36.25            -0.7%      36.00        slabinfo.names_cache.active_slabs
    290.00            -0.7%     288.00        slabinfo.names_cache.num_objs
     36.25            -0.7%      36.00        slabinfo.names_cache.num_slabs
      7.00            +0.0%       7.00        slabinfo.net_namespace.active_objs
      1.00            +0.0%       1.00        slabinfo.net_namespace.active_slabs
      7.00            +0.0%       7.00        slabinfo.net_namespace.num_objs
      1.00            +0.0%       1.00        slabinfo.net_namespace.num_slabs
     46.00            +0.0%      46.00        slabinfo.nfs_commit_data.active_objs
      1.00            +0.0%       1.00        slabinfo.nfs_commit_data.active_slabs
     46.00            +0.0%      46.00        slabinfo.nfs_commit_data.num_objs
      1.00            +0.0%       1.00        slabinfo.nfs_commit_data.num_slabs
     36.00            +0.0%      36.00        slabinfo.nfs_read_data.active_objs
      1.00            +0.0%       1.00        slabinfo.nfs_read_data.active_slabs
     36.00            +0.0%      36.00        slabinfo.nfs_read_data.num_objs
      1.00            +0.0%       1.00        slabinfo.nfs_read_data.num_slabs
    338.26 ±  9%      -5.3%     320.40 ± 13%  slabinfo.nsproxy.active_objs
      6.04 ±  9%      -5.3%       5.72 ± 13%  slabinfo.nsproxy.active_slabs
    338.26 ±  9%      -5.3%     320.40 ± 13%  slabinfo.nsproxy.num_objs
      6.04 ±  9%      -5.3%       5.72 ± 13%  slabinfo.nsproxy.num_slabs
     60.00            +0.0%      60.00        slabinfo.numa_policy.active_objs
      2.00            +0.0%       2.00        slabinfo.numa_policy.active_slabs
     60.00            +0.0%      60.00        slabinfo.numa_policy.num_objs
      2.00            +0.0%       2.00        slabinfo.numa_policy.num_slabs
      1489            +0.0%       1489        slabinfo.perf_event.active_objs
     62.34 ±  2%      +0.1%      62.38        slabinfo.perf_event.active_slabs
      1558 ±  2%      +0.1%       1559        slabinfo.perf_event.num_objs
     62.34 ±  2%      +0.1%      62.38        slabinfo.perf_event.num_slabs
      1822 ±  3%      +0.4%       1829 ±  2%  slabinfo.pid.active_objs
     56.95 ±  3%      +0.4%      57.16 ±  2%  slabinfo.pid.active_slabs
      1822 ±  3%      +0.4%       1829 ±  2%  slabinfo.pid.num_objs
     56.95 ±  3%      +0.4%      57.16 ±  2%  slabinfo.pid.num_slabs
      2692 ±  2%      +0.7%       2711        slabinfo.pool_workqueue.active_objs
     84.15 ±  2%      +0.7%      84.75        slabinfo.pool_workqueue.active_slabs
      2692 ±  2%      +0.7%       2711        slabinfo.pool_workqueue.num_objs
     84.15 ±  2%      +0.7%      84.75        slabinfo.pool_workqueue.num_slabs
      1013 ±  2%      +4.9%       1062 ±  5%  slabinfo.proc_dir_entry.active_objs
     24.12 ±  2%      +4.9%      25.30 ±  5%  slabinfo.proc_dir_entry.active_slabs
      1013 ±  2%      +4.9%       1062 ±  5%  slabinfo.proc_dir_entry.num_objs
     24.12 ±  2%      +4.9%      25.30 ±  5%  slabinfo.proc_dir_entry.num_slabs
      5273            +0.8%       5313        slabinfo.proc_inode_cache.active_objs
    114.97            +0.8%     115.94        slabinfo.proc_inode_cache.active_slabs
      5288            +0.8%       5333        slabinfo.proc_inode_cache.num_objs
    114.97            +0.8%     115.94        slabinfo.proc_inode_cache.num_slabs
     29682            -0.0%      29682        slabinfo.radix_tree_node.active_objs
      1062            +0.0%       1062        slabinfo.radix_tree_node.active_slabs
     29749            +0.0%      29752        slabinfo.radix_tree_node.num_objs
      1062            +0.0%       1062        slabinfo.radix_tree_node.num_slabs
     87.50 ± 20%      +4.0%      91.00 ± 18%  slabinfo.request_queue.active_objs
      2.50 ± 20%      +4.0%       2.60 ± 18%  slabinfo.request_queue.active_slabs
     87.50 ± 20%      +4.0%      91.00 ± 18%  slabinfo.request_queue.num_objs
      2.50 ± 20%      +4.0%       2.60 ± 18%  slabinfo.request_queue.num_slabs
     46.00            +0.0%      46.00        slabinfo.rpc_inode_cache.active_objs
      1.00            +0.0%       1.00        slabinfo.rpc_inode_cache.active_slabs
     46.00            +0.0%      46.00        slabinfo.rpc_inode_cache.num_objs
      1.00            +0.0%       1.00        slabinfo.rpc_inode_cache.num_slabs
    352.00            +0.0%     352.00        slabinfo.scsi_sense_cache.active_objs
     11.00            +0.0%      11.00        slabinfo.scsi_sense_cache.active_slabs
    352.00            +0.0%     352.00        slabinfo.scsi_sense_cache.num_objs
     11.00            +0.0%      11.00        slabinfo.scsi_sense_cache.num_slabs
      1407            -0.7%       1397        slabinfo.seq_file.active_objs
     41.40            -0.7%      41.11        slabinfo.seq_file.active_slabs
      1407            -0.7%       1397        slabinfo.seq_file.num_objs
     41.40            -0.7%      41.11        slabinfo.seq_file.num_slabs
      8021            -0.2%       8007        slabinfo.shared_policy_node.active_objs
     94.38            -0.2%      94.20        slabinfo.shared_policy_node.active_slabs
      8021            -0.2%       8007        slabinfo.shared_policy_node.num_objs
     94.38            -0.2%      94.20        slabinfo.shared_policy_node.num_slabs
      2346 ±  2%      -2.4%       2290 ±  3%  slabinfo.shmem_inode_cache.active_objs
     54.58 ±  2%      -2.4%      53.26 ±  3%  slabinfo.shmem_inode_cache.active_slabs
      2346 ±  2%      -2.4%       2290 ±  3%  slabinfo.shmem_inode_cache.num_objs
     54.58 ±  2%      -2.4%      53.26 ±  3%  slabinfo.shmem_inode_cache.num_slabs
      1053            -0.9%       1043        slabinfo.sighand_cache.active_objs
     70.23            -0.9%      69.58        slabinfo.sighand_cache.active_slabs
      1053            -0.9%       1043        slabinfo.sighand_cache.num_objs
     70.23            -0.9%      69.58        slabinfo.sighand_cache.num_slabs
      1711 ±  3%      -1.0%       1694 ±  3%  slabinfo.signal_cache.active_objs
     61.14 ±  3%      -0.9%      60.60 ±  3%  slabinfo.signal_cache.active_slabs
      1711 ±  3%      -0.9%       1696 ±  3%  slabinfo.signal_cache.num_objs
     61.14 ±  3%      -0.9%      60.60 ±  3%  slabinfo.signal_cache.num_slabs
      2441 ±  2%      +1.1%       2468 ±  3%  slabinfo.sigqueue.active_objs
     47.88 ±  2%      +1.1%      48.40 ±  3%  slabinfo.sigqueue.active_slabs
      2441 ±  2%      +1.1%       2468 ±  3%  slabinfo.sigqueue.num_objs
     47.88 ±  2%      +1.1%      48.40 ±  3%  slabinfo.sigqueue.num_slabs
    326.54 ±  8%      +2.0%     333.10 ± 14%  slabinfo.skbuff_ext_cache.active_objs
      7.77 ±  8%      +2.0%       7.93 ± 14%  slabinfo.skbuff_ext_cache.active_slabs
    326.54 ±  8%      +2.0%     333.10 ± 14%  slabinfo.skbuff_ext_cache.num_objs
      7.77 ±  8%      +2.0%       7.93 ± 14%  slabinfo.skbuff_ext_cache.num_slabs
      2389 ±  7%      +5.9%       2531 ±  7%  slabinfo.skbuff_head_cache.active_objs
     76.38 ±  7%      +4.2%      79.60 ±  7%  slabinfo.skbuff_head_cache.active_slabs
      2444 ±  7%      +4.2%       2547 ±  7%  slabinfo.skbuff_head_cache.num_objs
     76.38 ±  7%      +4.2%      79.60 ±  7%  slabinfo.skbuff_head_cache.num_slabs
      2454 ±  4%      +1.4%       2488 ±  7%  slabinfo.skbuff_small_head.active_objs
     49.62 ±  3%      +1.8%      50.50 ±  6%  slabinfo.skbuff_small_head.active_slabs
      2530 ±  3%      +1.8%       2575 ±  6%  slabinfo.skbuff_small_head.num_objs
     49.62 ±  3%      +1.8%      50.50 ±  6%  slabinfo.skbuff_small_head.num_slabs
      1506 ±  2%      +1.2%       1524 ±  2%  slabinfo.sock_inode_cache.active_objs
     38.62 ±  2%      +1.2%      39.10 ±  2%  slabinfo.sock_inode_cache.active_slabs
      1506 ±  2%      +1.2%       1524 ±  2%  slabinfo.sock_inode_cache.num_objs
     38.62 ±  2%      +1.2%      39.10 ±  2%  slabinfo.sock_inode_cache.num_slabs
      1045 ± 10%      +1.6%       1062 ±  8%  slabinfo.task_group.active_objs
     20.49 ± 10%      +1.6%      20.83 ±  8%  slabinfo.task_group.active_slabs
      1045 ± 10%      +1.6%       1062 ±  8%  slabinfo.task_group.num_objs
     20.49 ± 10%      +1.6%      20.83 ±  8%  slabinfo.task_group.num_slabs
    825.90            +1.0%     833.92        slabinfo.task_struct.active_objs
    826.76            +1.0%     834.85        slabinfo.task_struct.active_slabs
    826.76            +1.0%     834.85        slabinfo.task_struct.num_objs
    826.76            +1.0%     834.85        slabinfo.task_struct.num_slabs
    293.53 ±  6%      -5.6%     277.11 ±  9%  slabinfo.taskstats.active_objs
      7.93 ±  6%      -5.6%       7.49 ±  9%  slabinfo.taskstats.active_slabs
    293.53 ±  6%      -5.6%     277.11 ±  9%  slabinfo.taskstats.num_objs
      7.93 ±  6%      -5.6%       7.49 ±  9%  slabinfo.taskstats.num_slabs
      2294            +0.1%       2295        slabinfo.trace_event_file.active_objs
     49.88            +0.1%      49.90        slabinfo.trace_event_file.active_slabs
      2294            +0.1%       2295        slabinfo.trace_event_file.num_objs
     49.88            +0.1%      49.90        slabinfo.trace_event_file.num_slabs
    788.76 ±  4%      -0.2%     787.16 ±  6%  slabinfo.tracefs_inode_cache.active_objs
     15.78 ±  4%      -0.2%      15.74 ±  6%  slabinfo.tracefs_inode_cache.active_slabs
    788.76 ±  4%      -0.2%     787.16 ±  6%  slabinfo.tracefs_inode_cache.num_objs
     15.78 ±  4%      -0.2%      15.74 ±  6%  slabinfo.tracefs_inode_cache.num_slabs
     30.00            +0.0%      30.00        slabinfo.tw_sock_TCP.active_objs
      1.00            +0.0%       1.00        slabinfo.tw_sock_TCP.active_slabs
     30.00            +0.0%      30.00        slabinfo.tw_sock_TCP.num_objs
      1.00            +0.0%       1.00        slabinfo.tw_sock_TCP.num_slabs
     74.00            -5.0%      70.30 ± 15%  slabinfo.uts_namespace.active_objs
      2.00            -5.0%       1.90 ± 15%  slabinfo.uts_namespace.active_slabs
     74.00            -5.0%      70.30 ± 15%  slabinfo.uts_namespace.num_objs
      2.00            -5.0%       1.90 ± 15%  slabinfo.uts_namespace.num_slabs
      8533 ±  2%      -0.8%       8462        slabinfo.vm_area_struct.active_objs
    197.16 ±  2%      -0.8%     195.49        slabinfo.vm_area_struct.active_slabs
      8674 ±  2%      -0.8%       8601        slabinfo.vm_area_struct.num_objs
    197.16 ±  2%      -0.8%     195.49        slabinfo.vm_area_struct.num_slabs
     11310 ±  2%      +0.5%      11367 ±  3%  slabinfo.vma_lock.active_objs
    113.38 ±  2%      +0.0%     113.38 ±  3%  slabinfo.vma_lock.active_slabs
     11564 ±  2%      +0.0%      11565 ±  3%  slabinfo.vma_lock.num_objs
    113.38 ±  2%      +0.0%     113.38 ±  3%  slabinfo.vma_lock.num_slabs
     50524            +0.1%      50562        slabinfo.vmap_area.active_objs
    903.07            +0.0%     903.44        slabinfo.vmap_area.active_slabs
     50571            +0.0%      50592        slabinfo.vmap_area.num_objs
    903.07            +0.0%     903.44        slabinfo.vmap_area.num_slabs
    143.69            +0.2%     144.05        slabinfo.xfrm_dst_cache.active_objs
      2.82            +0.2%       2.82        slabinfo.xfrm_dst_cache.active_slabs
    143.69            +0.2%     144.05        slabinfo.xfrm_dst_cache.num_objs
      2.82            +0.2%       2.82        slabinfo.xfrm_dst_cache.num_slabs
     20.08            -1.5       18.61        perf-profile.calltrace.cycles-pp.truncate_inode_pages_range.btrfs_clone_files.btrfs_remap_file_range.do_clone_file_range.vfs_clone_file_range
     11.79 ±  2%      -0.9       10.85        perf-profile.calltrace.cycles-pp.truncate_cleanup_folio.truncate_inode_pages_range.btrfs_clone_files.btrfs_remap_file_range.do_clone_file_range
     11.45 ±  2%      -0.9       10.53        perf-profile.calltrace.cycles-pp.btrfs_invalidate_folio.truncate_cleanup_folio.truncate_inode_pages_range.btrfs_clone_files.btrfs_remap_file_range
     16.01 ±  2%      -0.8       15.17 ±  3%  perf-profile.calltrace.cycles-pp.filemap_read_folio.do_read_cache_folio.vfs_dedupe_file_range_compare.__generic_remap_file_range_prep.generic_remap_file_range_prep
     15.67 ±  2%      -0.8       14.87 ±  3%  perf-profile.calltrace.cycles-pp.btrfs_read_folio.filemap_read_folio.do_read_cache_folio.vfs_dedupe_file_range_compare.__generic_remap_file_range_prep
      9.93 ±  2%      -0.6        9.38 ±  2%  perf-profile.calltrace.cycles-pp.memcmp.vfs_dedupe_file_range_compare.__generic_remap_file_range_prep.generic_remap_file_range_prep.btrfs_remap_file_range
     12.06 ±  2%      -0.6       11.51 ±  3%  perf-profile.calltrace.cycles-pp.btrfs_do_readpage.btrfs_read_folio.filemap_read_folio.do_read_cache_folio.vfs_dedupe_file_range_compare
      4.16 ±  3%      -0.4        3.76 ±  3%  perf-profile.calltrace.cycles-pp.__clear_extent_bit.btrfs_invalidate_folio.truncate_cleanup_folio.truncate_inode_pages_range.btrfs_clone_files
     27.14 ±  4%      -0.4       26.75 ±  6%  perf-profile.calltrace.cycles-pp.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64_no_verify
     27.15 ±  4%      -0.4       26.77 ±  6%  perf-profile.calltrace.cycles-pp.cpu_startup_entry.start_secondary.secondary_startup_64_no_verify
     27.15 ±  4%      -0.4       26.77 ±  6%  perf-profile.calltrace.cycles-pp.start_secondary.secondary_startup_64_no_verify
     26.86 ±  4%      -0.4       26.51 ±  6%  perf-profile.calltrace.cycles-pp.cpuidle_idle_call.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64_no_verify
     26.47 ±  4%      -0.3       26.13 ±  6%  perf-profile.calltrace.cycles-pp.cpuidle_enter.cpuidle_idle_call.do_idle.cpu_startup_entry.start_secondary
      8.63 ±  2%      -0.3        8.29 ±  2%  perf-profile.calltrace.cycles-pp.filemap_add_folio.do_read_cache_folio.vfs_dedupe_file_range_compare.__generic_remap_file_range_prep.generic_remap_file_range_prep
      0.33 ± 77%      -0.3        0.07 ±300%  perf-profile.calltrace.cycles-pp.alloc_pages.folio_alloc.do_read_cache_folio.vfs_dedupe_file_range_compare.__generic_remap_file_range_prep
      3.44 ±  4%      -0.3        3.18 ±  3%  perf-profile.calltrace.cycles-pp.btrfs_lock_and_flush_ordered_range.btrfs_read_folio.filemap_read_folio.do_read_cache_folio.vfs_dedupe_file_range_compare
     23.72 ±  4%      -0.3       23.46 ±  5%  perf-profile.calltrace.cycles-pp.intel_idle_ibrs.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle
      3.23 ±  3%      -0.3        2.98 ±  3%  perf-profile.calltrace.cycles-pp.delete_from_page_cache_batch.truncate_inode_pages_range.btrfs_clone_files.btrfs_remap_file_range.do_clone_file_range
      0.57 ±  3%      -0.2        0.32 ± 81%  perf-profile.calltrace.cycles-pp.kmem_cache_alloc.alloc_extent_state.__clear_extent_bit.try_release_extent_mapping.btrfs_invalidate_folio
     27.67 ±  4%      -0.2       27.44 ±  6%  perf-profile.calltrace.cycles-pp.secondary_startup_64_no_verify
      0.61 ±  4%      -0.2        0.38 ± 65%  perf-profile.calltrace.cycles-pp.find_free_extent.btrfs_reserve_extent.__btrfs_prealloc_file_range.btrfs_prealloc_file_range.btrfs_fallocate
      0.62 ±  9%      -0.2        0.40 ± 66%  perf-profile.calltrace.cycles-pp.clock_nanosleep
      0.26 ±100%      -0.2        0.05 ±300%  perf-profile.calltrace.cycles-pp.folio_mark_accessed.do_read_cache_folio.vfs_dedupe_file_range_compare.__generic_remap_file_range_prep.generic_remap_file_range_prep
      3.78 ±  2%      -0.2        3.58 ±  3%  perf-profile.calltrace.cycles-pp.__folio_batch_release.truncate_inode_pages_range.btrfs_clone_files.btrfs_remap_file_range.do_clone_file_range
      3.68 ±  2%      -0.2        3.48 ±  3%  perf-profile.calltrace.cycles-pp.release_pages.__folio_batch_release.truncate_inode_pages_range.btrfs_clone_files.btrfs_remap_file_range
      5.42 ±  2%      -0.2        5.22 ±  2%  perf-profile.calltrace.cycles-pp.__filemap_add_folio.filemap_add_folio.do_read_cache_folio.vfs_dedupe_file_range_compare.__generic_remap_file_range_prep
      2.47 ±  2%      -0.2        2.28 ±  5%  perf-profile.calltrace.cycles-pp.lock_extent.btrfs_invalidate_folio.truncate_cleanup_folio.truncate_inode_pages_range.btrfs_clone_files
      0.62 ±  7%      -0.2        0.43 ± 67%  perf-profile.calltrace.cycles-pp.intel_idle.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle
      0.66 ±  9%      -0.2        0.48 ± 51%  perf-profile.calltrace.cycles-pp.set_page_extent_mapped.btrfs_do_readpage.btrfs_read_folio.filemap_read_folio.do_read_cache_folio
      0.62 ±  4%      -0.2        0.45 ± 50%  perf-profile.calltrace.cycles-pp.btrfs_reserve_extent.__btrfs_prealloc_file_range.btrfs_prealloc_file_range.btrfs_fallocate.vfs_fallocate
      1.50 ±  5%      -0.2        1.32 ±  5%  perf-profile.calltrace.cycles-pp.clear_state_bit.__clear_extent_bit.btrfs_invalidate_folio.truncate_cleanup_folio.truncate_inode_pages_range
      2.50 ±  4%      -0.2        2.33 ±  4%  perf-profile.calltrace.cycles-pp.ioctl_preallocate.__x64_sys_ioctl.do_syscall_64.entry_SYSCALL_64_after_hwframe.ioctl
      2.37 ±  4%      -0.2        2.20 ±  4%  perf-profile.calltrace.cycles-pp.lock_extent.btrfs_lock_and_flush_ordered_range.btrfs_read_folio.filemap_read_folio.do_read_cache_folio
      2.28            -0.2        2.11 ±  5%  perf-profile.calltrace.cycles-pp.__set_extent_bit.lock_extent.btrfs_invalidate_folio.truncate_cleanup_folio.truncate_inode_pages_range
      2.46 ±  4%      -0.2        2.29 ±  3%  perf-profile.calltrace.cycles-pp.vfs_fallocate.ioctl_preallocate.__x64_sys_ioctl.do_syscall_64.entry_SYSCALL_64_after_hwframe
      2.20 ±  5%      -0.2        2.04 ±  5%  perf-profile.calltrace.cycles-pp.__set_extent_bit.lock_extent.btrfs_lock_and_flush_ordered_range.btrfs_read_folio.filemap_read_folio
      2.37 ±  3%      -0.2        2.21 ±  4%  perf-profile.calltrace.cycles-pp.btrfs_fallocate.vfs_fallocate.ioctl_preallocate.__x64_sys_ioctl.do_syscall_64
      5.28            -0.1        5.14 ±  2%  perf-profile.calltrace.cycles-pp.memset_orig.btrfs_do_readpage.btrfs_read_folio.filemap_read_folio.do_read_cache_folio
      2.86 ±  3%      -0.1        2.72 ±  3%  perf-profile.calltrace.cycles-pp.folio_add_lru.filemap_add_folio.do_read_cache_folio.vfs_dedupe_file_range_compare.__generic_remap_file_range_prep
      3.71 ±  2%      -0.1        3.57 ±  4%  perf-profile.calltrace.cycles-pp.__clear_extent_bit.btrfs_do_readpage.btrfs_read_folio.filemap_read_folio.do_read_cache_folio
      2.59 ±  4%      -0.1        2.46 ±  3%  perf-profile.calltrace.cycles-pp.folio_batch_move_lru.folio_add_lru.filemap_add_folio.do_read_cache_folio.vfs_dedupe_file_range_compare
      0.13 ±173%      -0.1        0.00        perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.clock_nanosleep
      1.14 ±  2%      -0.1        1.03 ±  3%  perf-profile.calltrace.cycles-pp.alloc_extent_state.__clear_extent_bit.btrfs_invalidate_folio.truncate_cleanup_folio.truncate_inode_pages_range
      0.72 ±  5%      -0.1        0.61 ±  3%  perf-profile.calltrace.cycles-pp.__mod_lruvec_page_state.__filemap_add_folio.filemap_add_folio.do_read_cache_folio.vfs_dedupe_file_range_compare
      0.95 ±  6%      -0.1        0.84 ±  7%  perf-profile.calltrace.cycles-pp.btrfs_get_extent.btrfs_do_readpage.btrfs_read_folio.filemap_read_folio.do_read_cache_folio
      1.19 ±  3%      -0.1        1.09 ±  5%  perf-profile.calltrace.cycles-pp.btrfs_prealloc_file_range.btrfs_fallocate.vfs_fallocate.ioctl_preallocate.__x64_sys_ioctl
      1.19 ±  3%      -0.1        1.09 ±  5%  perf-profile.calltrace.cycles-pp.__btrfs_prealloc_file_range.btrfs_prealloc_file_range.btrfs_fallocate.vfs_fallocate.ioctl_preallocate
      1.52 ±  5%      -0.1        1.42 ±  4%  perf-profile.calltrace.cycles-pp.xas_store.delete_from_page_cache_batch.truncate_inode_pages_range.btrfs_clone_files.btrfs_remap_file_range
      2.27 ±  5%      -0.1        2.17 ±  3%  perf-profile.calltrace.cycles-pp.try_release_extent_mapping.btrfs_invalidate_folio.truncate_cleanup_folio.truncate_inode_pages_range.btrfs_clone_files
     26.65 ±  4%      -0.1       26.56 ±  6%  perf-profile.calltrace.cycles-pp.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle.cpu_startup_entry
      1.88 ±  6%      -0.1        1.79 ±  4%  perf-profile.calltrace.cycles-pp.__filemap_get_folio.do_read_cache_folio.vfs_dedupe_file_range_compare.__generic_remap_file_range_prep.generic_remap_file_range_prep
      0.41 ± 57%      -0.1        0.32 ± 81%  perf-profile.calltrace.cycles-pp.xa_get_order.__filemap_add_folio.filemap_add_folio.do_read_cache_folio.vfs_dedupe_file_range_compare
      0.84 ±  3%      -0.1        0.76 ±  3%  perf-profile.calltrace.cycles-pp.kmem_cache_alloc.alloc_extent_state.__clear_extent_bit.btrfs_invalidate_folio.truncate_cleanup_folio
      1.75 ±  5%      -0.1        1.67 ±  3%  perf-profile.calltrace.cycles-pp.__clear_extent_bit.try_release_extent_mapping.btrfs_invalidate_folio.truncate_cleanup_folio.truncate_inode_pages_range
      1.49 ±  5%      -0.1        1.40 ±  4%  perf-profile.calltrace.cycles-pp.lru_add_fn.folio_batch_move_lru.folio_add_lru.filemap_add_folio.do_read_cache_folio
      0.72 ±  5%      -0.1        0.63 ±  6%  perf-profile.calltrace.cycles-pp.clear_page_extent_mapped.btrfs_invalidate_folio.truncate_cleanup_folio.truncate_inode_pages_range.btrfs_clone_files
      0.87 ±  3%      -0.1        0.79 ±  5%  perf-profile.calltrace.cycles-pp.filemap_unaccount_folio.delete_from_page_cache_batch.truncate_inode_pages_range.btrfs_clone_files.btrfs_remap_file_range
      0.98 ±  5%      -0.1        0.90 ±  6%  perf-profile.calltrace.cycles-pp.alloc_extent_state.__set_extent_bit.lock_extent.btrfs_lock_and_flush_ordered_range.btrfs_read_folio
      0.65 ±  6%      -0.1        0.57 ±  7%  perf-profile.calltrace.cycles-pp.__mod_lruvec_page_state.filemap_unaccount_folio.delete_from_page_cache_batch.truncate_inode_pages_range.btrfs_clone_files
      1.60 ±  7%      -0.1        1.53 ±  4%  perf-profile.calltrace.cycles-pp.filemap_get_entry.__filemap_get_folio.do_read_cache_folio.vfs_dedupe_file_range_compare.__generic_remap_file_range_prep
      0.07 ±264%      -0.1        0.00        perf-profile.calltrace.cycles-pp.end_page_read.btrfs_do_readpage.btrfs_read_folio.filemap_read_folio.do_read_cache_folio
      1.06 ±  2%      -0.1        1.00 ±  6%  perf-profile.calltrace.cycles-pp.alloc_extent_state.__set_extent_bit.lock_extent.btrfs_invalidate_folio.truncate_cleanup_folio
      0.06 ±264%      -0.1        0.00        perf-profile.calltrace.cycles-pp.__wake_up.clear_state_bit.__clear_extent_bit.btrfs_invalidate_folio.truncate_cleanup_folio
      1.62 ±  3%      -0.1        1.56 ±  5%  perf-profile.calltrace.cycles-pp.clear_state_bit.__clear_extent_bit.btrfs_do_readpage.btrfs_read_folio.filemap_read_folio
      0.06 ±264%      -0.1        0.00        perf-profile.calltrace.cycles-pp.__wake_up.clear_state_bit.__clear_extent_bit.btrfs_do_readpage.btrfs_read_folio
      0.06 ±264%      -0.1        0.00        perf-profile.calltrace.cycles-pp.folio_unlock.vfs_dedupe_file_range_compare.__generic_remap_file_range_prep.generic_remap_file_range_prep.btrfs_remap_file_range
      0.68 ± 16%      -0.1        0.62 ±  8%  perf-profile.calltrace.cycles-pp.xas_load.filemap_get_entry.__filemap_get_folio.do_read_cache_folio.vfs_dedupe_file_range_compare
      0.67 ±  8%      -0.1        0.62 ±  7%  perf-profile.calltrace.cycles-pp.btrfs_replace_file_extents.btrfs_clone.btrfs_extent_same_range.btrfs_remap_file_range.vfs_dedupe_file_range_one
      0.79 ±  3%      -0.1        0.74 ±  6%  perf-profile.calltrace.cycles-pp.alloc_extent_state.__clear_extent_bit.try_release_extent_mapping.btrfs_invalidate_folio.truncate_cleanup_folio
      0.72 ±  7%      -0.0        0.67 ±  6%  perf-profile.calltrace.cycles-pp.kmem_cache_alloc.alloc_extent_state.__set_extent_bit.lock_extent.btrfs_lock_and_flush_ordered_range
      0.78 ±  4%      -0.0        0.74 ±  5%  perf-profile.calltrace.cycles-pp.kmem_cache_alloc.alloc_extent_state.__set_extent_bit.lock_extent.btrfs_invalidate_folio
      0.74 ±  5%      -0.0        0.70 ±  6%  perf-profile.calltrace.cycles-pp.syscall_return_via_sysret.ioctl.stress_run
      0.65 ±  4%      -0.0        0.61 ±  2%  perf-profile.calltrace.cycles-pp.btrfs_lookup_ordered_range.btrfs_lock_and_flush_ordered_range.btrfs_read_folio.filemap_read_folio.do_read_cache_folio
      0.67 ±  4%      -0.0        0.63 ±  7%  perf-profile.calltrace.cycles-pp.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe.ioctl.stress_run
      0.20 ±129%      -0.0        0.16 ±152%  perf-profile.calltrace.cycles-pp.btrfs_insert_replace_extent.btrfs_replace_file_extents.btrfs_clone.btrfs_clone_files.btrfs_remap_file_range
     41.84 ±  2%      -0.0       41.80 ±  2%  perf-profile.calltrace.cycles-pp.generic_remap_file_range_prep.btrfs_remap_file_range.vfs_dedupe_file_range_one.vfs_dedupe_file_range.do_vfs_ioctl
     41.84 ±  2%      -0.0       41.80 ±  2%  perf-profile.calltrace.cycles-pp.__generic_remap_file_range_prep.generic_remap_file_range_prep.btrfs_remap_file_range.vfs_dedupe_file_range_one.vfs_dedupe_file_range
     41.78 ±  2%      -0.0       41.75 ±  2%  perf-profile.calltrace.cycles-pp.vfs_dedupe_file_range_compare.__generic_remap_file_range_prep.generic_remap_file_range_prep.btrfs_remap_file_range.vfs_dedupe_file_range_one
      0.93 ±  4%      -0.0        0.90 ±  5%  perf-profile.calltrace.cycles-pp.find_lock_entries.truncate_inode_pages_range.btrfs_clone_files.btrfs_remap_file_range.do_clone_file_range
      1.23 ±  4%      -0.0        1.20 ±  2%  perf-profile.calltrace.cycles-pp.free_unref_page_list.release_pages.__folio_batch_release.truncate_inode_pages_range.btrfs_clone_files
      1.12 ±  3%      -0.0        1.10 ±  4%  perf-profile.calltrace.cycles-pp.xas_store.__filemap_add_folio.filemap_add_folio.do_read_cache_folio.vfs_dedupe_file_range_compare
      1.26 ±  4%      -0.0        1.23 ±  4%  perf-profile.calltrace.cycles-pp.btrfs_replace_file_extents.btrfs_clone.btrfs_clone_files.btrfs_remap_file_range.do_clone_file_range
      0.90 ±  3%      -0.0        0.90 ±  7%  perf-profile.calltrace.cycles-pp.alloc_extent_state.__clear_extent_bit.btrfs_do_readpage.btrfs_read_folio.filemap_read_folio
      1.27 ±  2%      -0.0        1.27 ±  6%  perf-profile.calltrace.cycles-pp.__mem_cgroup_charge.__filemap_add_folio.filemap_add_folio.do_read_cache_folio.vfs_dedupe_file_range_compare
      0.65 ±  6%      +0.0        0.65 ±  5%  perf-profile.calltrace.cycles-pp.btrfs_punch_hole.btrfs_fallocate.vfs_fallocate.ioctl_preallocate.__x64_sys_ioctl
      1.12 ±  6%      +0.0        1.14 ± 10%  perf-profile.calltrace.cycles-pp.__hrtimer_run_queues.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt
      0.66 ±  3%      +0.0        0.67 ±  6%  perf-profile.calltrace.cycles-pp.kmem_cache_alloc.alloc_extent_state.__clear_extent_bit.btrfs_do_readpage.btrfs_read_folio
      1.39 ±  7%      +0.0        1.42 ± 10%  perf-profile.calltrace.cycles-pp.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.cpuidle_enter_state
      0.13 ±173%      +0.0        0.16 ±152%  perf-profile.calltrace.cycles-pp.perf_adjust_freq_unthr_context.perf_event_task_tick.scheduler_tick.update_process_times.tick_sched_handle
      1.40 ±  7%      +0.0        1.44 ± 10%  perf-profile.calltrace.cycles-pp.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.cpuidle_enter_state.cpuidle_enter
      0.60 ±  7%      +0.0        0.64 ±  9%  perf-profile.calltrace.cycles-pp.scheduler_tick.update_process_times.tick_sched_handle.tick_sched_timer.__hrtimer_run_queues
      0.76 ±  4%      +0.0        0.80 ±  9%  perf-profile.calltrace.cycles-pp.tick_sched_timer.__hrtimer_run_queues.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt
     42.89 ±  2%      +0.0       42.94 ±  2%  perf-profile.calltrace.cycles-pp.btrfs_remap_file_range.vfs_dedupe_file_range_one.vfs_dedupe_file_range.do_vfs_ioctl.__x64_sys_ioctl
      0.68 ±  5%      +0.0        0.73 ±  9%  perf-profile.calltrace.cycles-pp.update_process_times.tick_sched_handle.tick_sched_timer.__hrtimer_run_queues.hrtimer_interrupt
      0.69 ±  5%      +0.1        0.74 ±  9%  perf-profile.calltrace.cycles-pp.tick_sched_handle.tick_sched_timer.__hrtimer_run_queues.hrtimer_interrupt.__sysvec_apic_timer_interrupt
      2.02 ±  7%      +0.1        2.07 ±  9%  perf-profile.calltrace.cycles-pp.asm_sysvec_apic_timer_interrupt.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle
      0.00            +0.1        0.05 ±300%  perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.rmqueue_bulk.__rmqueue_pcplist.rmqueue
      0.00            +0.1        0.05 ±300%  perf-profile.calltrace.cycles-pp.mem_cgroup_commit_charge.__mem_cgroup_charge.__filemap_add_folio.filemap_add_folio.do_read_cache_folio
     42.93 ±  2%      +0.1       42.98 ±  2%  perf-profile.calltrace.cycles-pp.vfs_dedupe_file_range.do_vfs_ioctl.__x64_sys_ioctl.do_syscall_64.entry_SYSCALL_64_after_hwframe
     42.92 ±  2%      +0.1       42.97 ±  2%  perf-profile.calltrace.cycles-pp.vfs_dedupe_file_range_one.vfs_dedupe_file_range.do_vfs_ioctl.__x64_sys_ioctl.do_syscall_64
      0.49 ± 38%      +0.1        0.55 ±  5%  perf-profile.calltrace.cycles-pp.btrfs_drop_extents.btrfs_replace_file_extents.btrfs_clone.btrfs_clone_files.btrfs_remap_file_range
      1.85 ±  9%      +0.1        1.91 ±  9%  perf-profile.calltrace.cycles-pp.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call
      0.13 ±173%      +0.1        0.21 ±122%  perf-profile.calltrace.cycles-pp.perf_event_task_tick.scheduler_tick.update_process_times.tick_sched_handle.tick_sched_timer
      1.00 ±  7%      +0.1        1.08 ±  6%  perf-profile.calltrace.cycles-pp.btrfs_extent_same_range.btrfs_remap_file_range.vfs_dedupe_file_range_one.vfs_dedupe_file_range.do_vfs_ioctl
      0.92 ±  7%      +0.1        1.01 ±  6%  perf-profile.calltrace.cycles-pp.btrfs_clone.btrfs_extent_same_range.btrfs_remap_file_range.vfs_dedupe_file_range_one.vfs_dedupe_file_range
      0.06 ±264%      +0.1        0.16 ±152%  perf-profile.calltrace.cycles-pp.__mem_cgroup_uncharge_list.release_pages.__folio_batch_release.truncate_inode_pages_range.btrfs_clone_files
      0.00            +0.1        0.10 ±200%  perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.rmqueue_bulk.__rmqueue_pcplist.rmqueue.get_page_from_freelist
      0.34 ±195%      +0.2        0.58 ±114%  perf-profile.calltrace.cycles-pp.cpuidle_enter.cpuidle_idle_call.do_idle.cpu_startup_entry.rest_init
      0.34 ±195%      +0.3        0.59 ±114%  perf-profile.calltrace.cycles-pp.cpuidle_idle_call.do_idle.cpu_startup_entry.rest_init.arch_call_rest_init
      0.34 ±195%      +0.3        0.60 ±115%  perf-profile.calltrace.cycles-pp.cpu_startup_entry.rest_init.arch_call_rest_init.start_kernel.x86_64_start_reservations
      0.34 ±195%      +0.3        0.60 ±115%  perf-profile.calltrace.cycles-pp.do_idle.cpu_startup_entry.rest_init.arch_call_rest_init.start_kernel
      0.34 ±195%      +0.3        0.60 ±115%  perf-profile.calltrace.cycles-pp.x86_64_start_kernel.secondary_startup_64_no_verify
      0.34 ±195%      +0.3        0.60 ±115%  perf-profile.calltrace.cycles-pp.x86_64_start_reservations.x86_64_start_kernel.secondary_startup_64_no_verify
      0.34 ±195%      +0.3        0.60 ±115%  perf-profile.calltrace.cycles-pp.start_kernel.x86_64_start_reservations.x86_64_start_kernel.secondary_startup_64_no_verify
      0.34 ±195%      +0.3        0.60 ±115%  perf-profile.calltrace.cycles-pp.arch_call_rest_init.start_kernel.x86_64_start_reservations.x86_64_start_kernel.secondary_startup_64_no_verify
      0.34 ±195%      +0.3        0.60 ±115%  perf-profile.calltrace.cycles-pp.rest_init.arch_call_rest_init.start_kernel.x86_64_start_reservations.x86_64_start_kernel
     70.99            +0.3       71.33 ±  2%  perf-profile.calltrace.cycles-pp.stress_run
     70.59            +0.4       70.96 ±  2%  perf-profile.calltrace.cycles-pp.ioctl.stress_run
     69.39            +0.5       69.86 ±  2%  perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.ioctl.stress_run
     68.99            +0.5       69.46 ±  2%  perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.ioctl.stress_run
     68.25            +0.5       68.76 ±  2%  perf-profile.calltrace.cycles-pp.__x64_sys_ioctl.do_syscall_64.entry_SYSCALL_64_after_hwframe.ioctl.stress_run
     22.27            +0.6       22.89 ±  2%  perf-profile.calltrace.cycles-pp.btrfs_remap_file_range.do_clone_file_range.vfs_clone_file_range.ioctl_file_clone.do_vfs_ioctl
     22.35            +0.6       22.97 ±  2%  perf-profile.calltrace.cycles-pp.ioctl_file_clone.do_vfs_ioctl.__x64_sys_ioctl.do_syscall_64.entry_SYSCALL_64_after_hwframe
     22.32            +0.6       22.94 ±  2%  perf-profile.calltrace.cycles-pp.do_clone_file_range.vfs_clone_file_range.ioctl_file_clone.do_vfs_ioctl.__x64_sys_ioctl
     21.97            +0.6       22.59 ±  2%  perf-profile.calltrace.cycles-pp.btrfs_clone_files.btrfs_remap_file_range.do_clone_file_range.vfs_clone_file_range.ioctl_file_clone
     22.34            +0.6       22.96 ±  2%  perf-profile.calltrace.cycles-pp.vfs_clone_file_range.ioctl_file_clone.do_vfs_ioctl.__x64_sys_ioctl.do_syscall_64
     29.94 ±  2%      +0.6       30.58 ±  3%  perf-profile.calltrace.cycles-pp.do_read_cache_folio.vfs_dedupe_file_range_compare.__generic_remap_file_range_prep.generic_remap_file_range_prep.btrfs_remap_file_range
     65.60            +0.7       66.27 ±  2%  perf-profile.calltrace.cycles-pp.do_vfs_ioctl.__x64_sys_ioctl.do_syscall_64.entry_SYSCALL_64_after_hwframe.ioctl
      0.00            +0.8        0.85 ±  8%  perf-profile.calltrace.cycles-pp.__list_del_entry_valid_or_report.rmqueue_bulk.__rmqueue_pcplist.rmqueue.get_page_from_freelist
      0.00            +1.5        1.51 ±  6%  perf-profile.calltrace.cycles-pp.__free_one_page.free_pcppages_bulk.free_unref_page_commit.free_unref_page.btrfs_clone
      0.89 ±  7%      +1.9        2.80 ±  5%  perf-profile.calltrace.cycles-pp.rmqueue.get_page_from_freelist.__alloc_pages.folio_alloc.do_read_cache_folio
      2.57 ±  3%      +2.0        4.54 ±  5%  perf-profile.calltrace.cycles-pp.folio_alloc.do_read_cache_folio.vfs_dedupe_file_range_compare.__generic_remap_file_range_prep.generic_remap_file_range_prep
      1.19 ±  6%      +2.0        3.17 ±  5%  perf-profile.calltrace.cycles-pp.get_page_from_freelist.__alloc_pages.folio_alloc.do_read_cache_folio.vfs_dedupe_file_range_compare
      1.91 ±  3%      +2.0        3.90 ±  4%  perf-profile.calltrace.cycles-pp.__alloc_pages.folio_alloc.do_read_cache_folio.vfs_dedupe_file_range_compare.__generic_remap_file_range_prep
      0.00            +2.0        1.99 ±  5%  perf-profile.calltrace.cycles-pp.rmqueue_bulk.__rmqueue_pcplist.rmqueue.get_page_from_freelist.__alloc_pages
      1.67 ±  4%      +2.1        3.76 ±  3%  perf-profile.calltrace.cycles-pp.btrfs_clone.btrfs_clone_files.btrfs_remap_file_range.do_clone_file_range.vfs_clone_file_range
      0.00            +2.1        2.09 ±  4%  perf-profile.calltrace.cycles-pp.free_pcppages_bulk.free_unref_page_commit.free_unref_page.btrfs_clone.btrfs_clone_files
      0.00            +2.1        2.12 ±  4%  perf-profile.calltrace.cycles-pp.free_unref_page_commit.free_unref_page.btrfs_clone.btrfs_clone_files.btrfs_remap_file_range
      0.00            +2.1        2.14 ±  4%  perf-profile.calltrace.cycles-pp.free_unref_page.btrfs_clone.btrfs_clone_files.btrfs_remap_file_range.do_clone_file_range
      0.00            +2.3        2.29 ±  6%  perf-profile.calltrace.cycles-pp.__rmqueue_pcplist.rmqueue.get_page_from_freelist.__alloc_pages.folio_alloc
     20.14            -1.5       18.68        perf-profile.children.cycles-pp.truncate_inode_pages_range
     11.79 ±  2%      -0.9       10.85        perf-profile.children.cycles-pp.truncate_cleanup_folio
     11.57 ±  2%      -0.9       10.64        perf-profile.children.cycles-pp.btrfs_invalidate_folio
     16.02 ±  2%      -0.8       15.18 ±  3%  perf-profile.children.cycles-pp.filemap_read_folio
     15.69 ±  2%      -0.8       14.89 ±  3%  perf-profile.children.cycles-pp.btrfs_read_folio
     10.00 ±  2%      -0.7        9.34 ±  2%  perf-profile.children.cycles-pp.__clear_extent_bit
      9.99 ±  2%      -0.6        9.43 ±  2%  perf-profile.children.cycles-pp.memcmp
     12.08 ±  2%      -0.5       11.53 ±  3%  perf-profile.children.cycles-pp.btrfs_do_readpage
     27.15 ±  4%      -0.4       26.77 ±  6%  perf-profile.children.cycles-pp.start_secondary
      5.00 ±  2%      -0.4        4.63 ±  3%  perf-profile.children.cycles-pp.lock_extent
      4.82 ±  2%      -0.4        4.46 ±  3%  perf-profile.children.cycles-pp.__set_extent_bit
      5.19 ±  2%      -0.3        4.84 ±  3%  perf-profile.children.cycles-pp.alloc_extent_state
      8.65 ±  2%      -0.3        8.32 ±  2%  perf-profile.children.cycles-pp.filemap_add_folio
      3.27 ±  3%      -0.3        3.01 ±  3%  perf-profile.children.cycles-pp.delete_from_page_cache_batch
      4.23 ±  2%      -0.3        3.97 ±  3%  perf-profile.children.cycles-pp.kmem_cache_alloc
      3.45 ±  4%      -0.3        3.20 ±  3%  perf-profile.children.cycles-pp.btrfs_lock_and_flush_ordered_range
      3.40 ±  2%      -0.2        3.15 ±  4%  perf-profile.children.cycles-pp.clear_state_bit
     27.67 ±  4%      -0.2       27.44 ±  6%  perf-profile.children.cycles-pp.secondary_startup_64_no_verify
     27.67 ±  4%      -0.2       27.44 ±  6%  perf-profile.children.cycles-pp.cpu_startup_entry
     27.66 ±  4%      -0.2       27.43 ±  6%  perf-profile.children.cycles-pp.do_idle
     23.81 ±  4%      -0.2       23.59 ±  5%  perf-profile.children.cycles-pp.intel_idle_ibrs
      3.97 ±  2%      -0.2        3.76 ±  3%  perf-profile.children.cycles-pp.release_pages
      5.54 ±  2%      -0.2        5.33 ±  2%  perf-profile.children.cycles-pp.__filemap_add_folio
     27.38 ±  4%      -0.2       27.18 ±  6%  perf-profile.children.cycles-pp.cpuidle_idle_call
      3.78 ±  2%      -0.2        3.58 ±  3%  perf-profile.children.cycles-pp.__folio_batch_release
     26.97 ±  4%      -0.2       26.78 ±  6%  perf-profile.children.cycles-pp.cpuidle_enter
     26.97 ±  4%      -0.2       26.78 ±  6%  perf-profile.children.cycles-pp.cpuidle_enter_state
      1.42 ±  3%      -0.2        1.24 ±  3%  perf-profile.children.cycles-pp.__mod_lruvec_page_state
      2.50 ±  4%      -0.2        2.33 ±  4%  perf-profile.children.cycles-pp.ioctl_preallocate
      0.86 ±  6%      -0.2        0.69 ±  7%  perf-profile.children.cycles-pp.__mod_memcg_lruvec_state
      2.46 ±  4%      -0.2        2.30 ±  3%  perf-profile.children.cycles-pp.vfs_fallocate
      2.38 ±  3%      -0.2        2.21 ±  4%  perf-profile.children.cycles-pp.btrfs_fallocate
      5.31            -0.1        5.16 ±  2%  perf-profile.children.cycles-pp.memset_orig
      2.88 ±  3%      -0.1        2.74 ±  3%  perf-profile.children.cycles-pp.folio_add_lru
      2.67 ±  4%      -0.1        2.53 ±  3%  perf-profile.children.cycles-pp.folio_batch_move_lru
      1.26 ±  2%      -0.1        1.13 ±  3%  perf-profile.children.cycles-pp.free_extent_state
      2.70 ±  4%      -0.1        2.58 ±  3%  perf-profile.children.cycles-pp.xas_store
      1.10 ±  5%      -0.1        0.99 ±  7%  perf-profile.children.cycles-pp.btrfs_get_extent
      1.19 ±  3%      -0.1        1.09 ±  5%  perf-profile.children.cycles-pp.btrfs_prealloc_file_range
      1.19 ±  3%      -0.1        1.09 ±  5%  perf-profile.children.cycles-pp.__btrfs_prealloc_file_range
      2.74 ±  2%      -0.1        2.64 ±  3%  perf-profile.children.cycles-pp.btrfs_replace_file_extents
      0.96 ±  3%      -0.1        0.87 ±  5%  perf-profile.children.cycles-pp.__wake_up
      2.28 ±  5%      -0.1        2.18 ±  2%  perf-profile.children.cycles-pp.try_release_extent_mapping
      1.93 ±  5%      -0.1        1.83 ±  3%  perf-profile.children.cycles-pp.__filemap_get_folio
      1.56 ±  5%      -0.1        1.46 ±  3%  perf-profile.children.cycles-pp.lru_add_fn
      0.62 ±  9%      -0.1        0.53 ± 17%  perf-profile.children.cycles-pp.clock_nanosleep
      1.33 ±  9%      -0.1        1.24 ±  4%  perf-profile.children.cycles-pp.xas_load
      0.72 ±  5%      -0.1        0.63 ±  6%  perf-profile.children.cycles-pp.clear_page_extent_mapped
      0.66 ±  9%      -0.1        0.58 ± 13%  perf-profile.children.cycles-pp.set_page_extent_mapped
      0.61 ±  4%      -0.1        0.53 ±  7%  perf-profile.children.cycles-pp.find_free_extent
      0.62 ±  4%      -0.1        0.54 ±  7%  perf-profile.children.cycles-pp.btrfs_reserve_extent
      0.88 ±  3%      -0.1        0.81 ±  4%  perf-profile.children.cycles-pp.filemap_unaccount_folio
      0.63 ±  4%      -0.1        0.55 ±  7%  perf-profile.children.cycles-pp.__cond_resched
      1.63 ±  7%      -0.1        1.56 ±  4%  perf-profile.children.cycles-pp.filemap_get_entry
      0.46 ±  8%      -0.1        0.39 ±  8%  perf-profile.children.cycles-pp.end_page_read
      1.59 ±  4%      -0.1        1.52 ±  5%  perf-profile.children.cycles-pp.kmem_cache_free
      0.89 ±  4%      -0.1        0.83 ±  5%  perf-profile.children.cycles-pp.syscall_exit_to_user_mode
      0.97 ±  3%      -0.1        0.92 ±  3%  perf-profile.children.cycles-pp.syscall_return_via_sysret
      0.64 ±  9%      -0.1        0.58 ±  7%  perf-profile.children.cycles-pp.cache_state_if_flags
      0.47 ±  6%      -0.1        0.41 ±  8%  perf-profile.children.cycles-pp.__entry_text_start
      1.01 ±  3%      -0.1        0.96 ±  6%  perf-profile.children.cycles-pp.btrfs_insert_replace_extent
      0.81 ±  4%      -0.1        0.76 ±  5%  perf-profile.children.cycles-pp.folio_unlock
      0.33 ±  4%      -0.0        0.28 ± 12%  perf-profile.children.cycles-pp.try_charge_memcg
      0.27 ± 11%      -0.0        0.22 ±  9%  perf-profile.children.cycles-pp._raw_read_lock
      0.33 ± 11%      -0.0        0.28 ± 15%  perf-profile.children.cycles-pp.__x64_sys_clock_nanosleep
      0.68 ±  4%      -0.0        0.63 ±  2%  perf-profile.children.cycles-pp.btrfs_lookup_ordered_range
      0.31 ± 17%      -0.0        0.26 ± 16%  perf-profile.children.cycles-pp.cgroup_rstat_updated
      0.30 ± 13%      -0.0        0.26 ± 14%  perf-profile.children.cycles-pp.common_nsleep
     41.92 ±  2%      -0.0       41.88 ±  2%  perf-profile.children.cycles-pp.__generic_remap_file_range_prep
      2.01 ±  3%      -0.0        1.96 ±  4%  perf-profile.children.cycles-pp._raw_spin_lock
      0.43 ± 10%      -0.0        0.38 ± 13%  perf-profile.children.cycles-pp.__schedule
     41.96 ±  2%      -0.0       41.92 ±  2%  perf-profile.children.cycles-pp.generic_remap_file_range_prep
      0.42 ±  6%      -0.0        0.38 ±  4%  perf-profile.children.cycles-pp.memcg_slab_post_alloc_hook
      0.81 ±  7%      -0.0        0.77 ±  6%  perf-profile.children.cycles-pp.__mod_lruvec_state
      0.30 ± 14%      -0.0        0.26 ± 14%  perf-profile.children.cycles-pp.hrtimer_nanosleep
      0.63 ±  7%      -0.0        0.59 ± 18%  perf-profile.children.cycles-pp.intel_idle
      0.29 ±  5%      -0.0        0.25 ±  9%  perf-profile.children.cycles-pp.btrfs_update_inode
      0.59 ±  7%      -0.0        0.55 ±  7%  perf-profile.children.cycles-pp.folio_lruvec_lock_irqsave
     41.82 ±  2%      -0.0       41.78 ±  2%  perf-profile.children.cycles-pp.vfs_dedupe_file_range_compare
      0.22 ± 12%      -0.0        0.18 ± 11%  perf-profile.children.cycles-pp.btrfs_page_clear_checked
      0.71 ±  8%      -0.0        0.68 ±  3%  perf-profile.children.cycles-pp._raw_spin_lock_irq
      0.28 ±  9%      -0.0        0.25 ± 11%  perf-profile.children.cycles-pp.schedule
      0.79 ± 12%      -0.0        0.76 ±  7%  perf-profile.children.cycles-pp.xas_descend
      0.47 ±  3%      -0.0        0.44 ±  6%  perf-profile.children.cycles-pp.insert_prealloc_file_extent
      0.35 ± 11%      -0.0        0.32 ± 10%  perf-profile.children.cycles-pp.__init_waitqueue_head
      0.34 ± 10%      -0.0        0.31 ±  6%  perf-profile.children.cycles-pp.btrfs_lookup_first_ordered_range
      0.53 ±  6%      -0.0        0.50 ± 12%  perf-profile.children.cycles-pp.alloc_pages
      0.28 ± 13%      -0.0        0.24 ± 16%  perf-profile.children.cycles-pp.do_nanosleep
      0.31 ± 11%      -0.0        0.28 ±  8%  perf-profile.children.cycles-pp.xas_clear_mark
      0.12 ± 11%      -0.0        0.09 ± 41%  perf-profile.children.cycles-pp.activate_task
      0.23 ±  8%      -0.0        0.20 ± 10%  perf-profile.children.cycles-pp.filemap_free_folio
      0.96 ±  4%      -0.0        0.93 ±  5%  perf-profile.children.cycles-pp.find_lock_entries
      0.28 ± 11%      -0.0        0.25 ± 12%  perf-profile.children.cycles-pp.__lookup_extent_mapping
      0.03 ±102%      -0.0        0.00        perf-profile.children.cycles-pp.btrfs_free_path
      0.28 ± 11%      -0.0        0.25 ± 11%  perf-profile.children.cycles-pp.btrfs_page_set_uptodate
      0.28 ± 11%      -0.0        0.25 ± 11%  perf-profile.children.cycles-pp.btrfs_wait_ordered_range
      0.22 ±  9%      -0.0        0.19 ± 10%  perf-profile.children.cycles-pp.rcu_all_qs
      0.31 ± 10%      -0.0        0.28 ± 10%  perf-profile.children.cycles-pp.btrfs_del_items
      0.16 ±  7%      -0.0        0.13 ± 14%  perf-profile.children.cycles-pp.__tree_search
      0.37 ±  6%      -0.0        0.34 ± 12%  perf-profile.children.cycles-pp.setup_items_for_insert
      0.22 ±  9%      -0.0        0.19 ±  8%  perf-profile.children.cycles-pp.btrfs_delayed_update_inode
      0.15 ± 11%      -0.0        0.12 ± 15%  perf-profile.children.cycles-pp.read_tsc
      0.25 ± 12%      -0.0        0.23 ±  6%  perf-profile.children.cycles-pp._raw_spin_trylock
      0.19 ±  8%      -0.0        0.16 ±  8%  perf-profile.children.cycles-pp.clone_finish_inode_update
      0.20 ±  9%      -0.0        0.18 ± 16%  perf-profile.children.cycles-pp.try_to_wake_up
      0.11 ± 12%      -0.0        0.09 ± 41%  perf-profile.children.cycles-pp.enqueue_task_fair
      0.11 ± 13%      -0.0        0.09 ± 22%  perf-profile.children.cycles-pp.update_load_avg
      0.06 ± 13%      -0.0        0.04 ± 66%  perf-profile.children.cycles-pp.__do_sys_newfstat
      0.04 ± 58%      -0.0        0.02 ±153%  perf-profile.children.cycles-pp.xas_nomem
      0.10 ± 11%      -0.0        0.08 ± 41%  perf-profile.children.cycles-pp.enqueue_entity
      0.44 ±  6%      -0.0        0.42 ±  2%  perf-profile.children.cycles-pp.free_unref_page_prepare
      0.11 ± 13%      -0.0        0.09 ± 18%  perf-profile.children.cycles-pp.btrfs_put_block_group
      0.10 ± 18%      -0.0        0.08 ± 18%  perf-profile.children.cycles-pp.entry_SYSCALL_64_safe_stack
      0.26 ±  3%      -0.0        0.23 ±  7%  perf-profile.children.cycles-pp.xas_find
      0.48 ±  7%      -0.0        0.45 ±  6%  perf-profile.children.cycles-pp.xas_find_conflict
      0.14 ± 15%      -0.0        0.12 ± 28%  perf-profile.children.cycles-pp.hrtimer_wakeup
      0.14 ± 15%      -0.0        0.12 ± 18%  perf-profile.children.cycles-pp.vfs_fileattr_set
      0.54 ±  6%      -0.0        0.52 ±  6%  perf-profile.children.cycles-pp.xa_get_order
      0.32 ±  7%      -0.0        0.30 ±  9%  perf-profile.children.cycles-pp.btrfs_is_subpage
      0.02 ±100%      -0.0        0.01 ±300%  perf-profile.children.cycles-pp.__update_load_avg_cfs_rq
      0.26 ±  7%      -0.0        0.24 ±  8%  perf-profile.children.cycles-pp.get_pfnblock_flags_mask
      0.06 ± 63%      -0.0        0.04 ±111%  perf-profile.children.cycles-pp.cleanup_ref_head
      0.15 ± 12%      -0.0        0.13 ± 12%  perf-profile.children.cycles-pp.page_counter_uncharge
      0.13 ± 11%      -0.0        0.11 ± 28%  perf-profile.children.cycles-pp.ttwu_do_activate
      0.37 ± 11%      -0.0        0.35 ±  8%  perf-profile.children.cycles-pp.start_transaction
      0.62 ±  8%      -0.0        0.60 ±  8%  perf-profile.children.cycles-pp.__mod_node_page_state
      0.52 ±  6%      -0.0        0.50 ±  8%  perf-profile.children.cycles-pp.btrfs_insert_empty_items
      0.14 ± 15%      -0.0        0.12 ± 17%  perf-profile.children.cycles-pp.intel_idle_irq
      0.06 ± 18%      -0.0        0.04 ± 66%  perf-profile.children.cycles-pp.wake_up_q
      0.51 ±  6%      -0.0        0.49 ±  7%  perf-profile.children.cycles-pp.folio_mark_accessed
      0.08 ± 30%      -0.0        0.07 ± 74%  perf-profile.children.cycles-pp.process_one_work
      0.07 ± 64%      -0.0        0.05 ± 96%  perf-profile.children.cycles-pp.btrfs_preempt_reclaim_metadata_space
      0.07 ± 64%      -0.0        0.05 ± 96%  perf-profile.children.cycles-pp.flush_space
      0.07 ± 64%      -0.0        0.05 ± 96%  perf-profile.children.cycles-pp.btrfs_run_delayed_refs
      0.07 ± 64%      -0.0        0.05 ± 96%  perf-profile.children.cycles-pp.__btrfs_run_delayed_refs
      0.03 ±101%      -0.0        0.01 ±206%  perf-profile.children.cycles-pp.tick_sched_do_timer
      1.24 ±  4%      -0.0        1.22 ±  3%  perf-profile.children.cycles-pp.free_unref_page_list
      0.36 ±  9%      -0.0        0.35 ±  7%  perf-profile.children.cycles-pp.set_state_bits
      0.15 ±  8%      -0.0        0.13 ± 11%  perf-profile.children.cycles-pp.__fxstat64
      0.05 ± 38%      -0.0        0.04 ± 65%  perf-profile.children.cycles-pp.mutex_unlock
      0.36 ±  7%      -0.0        0.34 ±  7%  perf-profile.children.cycles-pp.should_failslab
      0.10 ± 16%      -0.0        0.08 ± 10%  perf-profile.children.cycles-pp.memmove_extent_buffer
      0.39 ± 10%      -0.0        0.37 ±  3%  perf-profile.children.cycles-pp.test_range_bit
      0.06 ± 15%      -0.0        0.05 ± 53%  perf-profile.children.cycles-pp.btrfs_get_64
      0.12 ± 15%      -0.0        0.10 ±  9%  perf-profile.children.cycles-pp.file_modified
      0.12 ± 14%      -0.0        0.10 ± 16%  perf-profile.children.cycles-pp.btrfs_get_block_group
      1.20 ±  5%      -0.0        1.18 ±  5%  perf-profile.children.cycles-pp.btrfs_drop_extents
      0.19 ± 11%      -0.0        0.17 ±  9%  perf-profile.children.cycles-pp.btrfs_read_lock_root_node
      0.03 ±100%      -0.0        0.01 ±201%  perf-profile.children.cycles-pp.free_extent_buffer
      1.52 ±  6%      -0.0        1.51 ±  7%  perf-profile.children.cycles-pp.__hrtimer_run_queues
      0.12 ± 13%      -0.0        0.10 ± 13%  perf-profile.children.cycles-pp.dequeue_task_fair
      0.13 ± 11%      -0.0        0.12 ± 13%  perf-profile.children.cycles-pp._raw_spin_unlock_irqrestore
      0.11 ± 18%      -0.0        0.10 ± 15%  perf-profile.children.cycles-pp.__btrfs_tree_lock
      0.02 ±131%      -0.0        0.01 ±201%  perf-profile.children.cycles-pp.btrfs_lookup_first_ordered_extent
      0.09 ± 22%      -0.0        0.08 ± 62%  perf-profile.children.cycles-pp.worker_thread
      0.18 ±  9%      -0.0        0.16 ±  8%  perf-profile.children.cycles-pp.__write_extent_buffer
      0.07 ± 17%      -0.0        0.06 ± 37%  perf-profile.children.cycles-pp.btrfs_duplicate_item
      0.26 ± 10%      -0.0        0.24 ±  6%  perf-profile.children.cycles-pp.add_delayed_ref_head
      0.18 ± 18%      -0.0        0.17 ± 14%  perf-profile.children.cycles-pp.llseek
      0.14 ± 21%      -0.0        0.12 ± 22%  perf-profile.children.cycles-pp.tick_nohz_get_sleep_length
      0.07 ± 14%      -0.0        0.05 ± 35%  perf-profile.children.cycles-pp.memcpy_extent_buffer
      0.07 ± 21%      -0.0        0.06 ± 51%  perf-profile.children.cycles-pp.hrtimer_start_range_ns
      0.56 ±  5%      -0.0        0.55 ±  5%  perf-profile.children.cycles-pp.btrfs_search_slot
      0.14 ± 13%      -0.0        0.12 ± 11%  perf-profile.children.cycles-pp.__btrfs_qgroup_release_data
      0.02 ±129%      -0.0        0.01 ±300%  perf-profile.children.cycles-pp.policy_nodemask
      0.02 ±129%      -0.0        0.01 ±300%  perf-profile.children.cycles-pp.set_next_entity
      0.12 ±  9%      -0.0        0.11 ± 14%  perf-profile.children.cycles-pp.__btrfs_release_delayed_node
      0.01 ±173%      -0.0        0.00        perf-profile.children.cycles-pp.sched_ttwu_pending
      0.01 ±173%      -0.0        0.00        perf-profile.children.cycles-pp.inode_needs_update_time
      0.02 ±130%      -0.0        0.01 ±200%  perf-profile.children.cycles-pp.set_extent_buffer_dirty
      0.52 ±  3%      -0.0        0.50 ±  6%  perf-profile.children.cycles-pp.__mem_cgroup_uncharge_list
      0.17 ± 16%      -0.0        0.16 ± 22%  perf-profile.children.cycles-pp.perf_mux_hrtimer_handler
      0.09 ± 10%      -0.0        0.08 ± 20%  perf-profile.children.cycles-pp.btrfs_set_token_32
      0.05 ± 40%      -0.0        0.04 ± 67%  perf-profile.children.cycles-pp.work_busy
      0.18 ± 13%      -0.0        0.17 ± 16%  perf-profile.children.cycles-pp.btrfs_drop_extent_map_range
      0.14 ± 15%      -0.0        0.13 ± 13%  perf-profile.children.cycles-pp.filemap_alloc_folio
      0.19 ±  9%      -0.0        0.18 ± 13%  perf-profile.children.cycles-pp.btrfs_set_delalloc_extent
      0.11 ± 17%      -0.0        0.10 ± 20%  perf-profile.children.cycles-pp.btrfs_fileattr_set
      0.12 ± 18%      -0.0        0.11 ± 11%  perf-profile.children.cycles-pp.should_fail_alloc_page
      0.11 ± 13%      -0.0        0.10 ± 20%  perf-profile.children.cycles-pp.down_read
      0.22 ±  5%      -0.0        0.21 ±  7%  perf-profile.children.cycles-pp.uncharge_batch
      0.12 ± 18%      -0.0        0.11 ± 14%  perf-profile.children.cycles-pp.btrfs_lock_root_node
      0.02 ±129%      -0.0        0.01 ±200%  perf-profile.children.cycles-pp.__hrtimer_next_event_base
      0.20 ±  8%      -0.0        0.19 ± 14%  perf-profile.children.cycles-pp.policy_node
      0.14 ± 10%      -0.0        0.13 ± 11%  perf-profile.children.cycles-pp.btrfs_alloc_reserved_file_extent
      0.14 ±  9%      -0.0        0.13 ±  8%  perf-profile.children.cycles-pp.kmem_cache_alloc_lru
      0.07 ± 16%      -0.0        0.06 ± 18%  perf-profile.children.cycles-pp.schedule_preempt_disabled
      0.06 ± 13%      -0.0        0.05 ± 36%  perf-profile.children.cycles-pp.update_rq_clock_task
      0.08 ± 13%      -0.0        0.08 ± 17%  perf-profile.children.cycles-pp.rwsem_down_read_slowpath
      0.08 ± 13%      -0.0        0.07 ± 53%  perf-profile.children.cycles-pp.perf_rotate_context
      0.02 ±129%      -0.0        0.01 ±200%  perf-profile.children.cycles-pp.read
      0.11 ± 11%      -0.0        0.10 ± 19%  perf-profile.children.cycles-pp.__btrfs_tree_read_lock
      0.08 ± 14%      -0.0        0.07 ± 16%  perf-profile.children.cycles-pp.__filemap_fdatawrite_range
      0.29 ±  9%      -0.0        0.28 ±  9%  perf-profile.children.cycles-pp.mem_cgroup_update_lru_size
      0.40 ±  9%      -0.0        0.39 ±  7%  perf-profile.children.cycles-pp.xas_start
      0.16 ± 16%      -0.0        0.15 ± 18%  perf-profile.children.cycles-pp.schedule_idle
      0.08 ± 27%      -0.0        0.07 ± 42%  perf-profile.children.cycles-pp.tick_nohz_next_event
      0.08 ± 19%      -0.0        0.07 ± 42%  perf-profile.children.cycles-pp.sched_clock
      0.10 ± 18%      -0.0        0.10 ± 12%  perf-profile.children.cycles-pp.dequeue_entity
      0.06 ± 17%      -0.0        0.05 ± 55%  perf-profile.children.cycles-pp.update_sg_lb_stats
      0.07 ± 23%      -0.0        0.06 ± 18%  perf-profile.children.cycles-pp.btrfs_fdatawrite_range
      0.07 ± 16%      -0.0        0.06 ± 40%  perf-profile.children.cycles-pp.rwsem_wake
      0.01 ±173%      -0.0        0.01 ±299%  perf-profile.children.cycles-pp.__switch_to_asm
      0.01 ±264%      -0.0        0.00        perf-profile.children.cycles-pp.update_rq_clock
      0.01 ±264%      -0.0        0.00        perf-profile.children.cycles-pp.rep_movs_alternative
      0.11 ± 20%      -0.0        0.10 ± 17%  perf-profile.children.cycles-pp.rebalance_domains
      0.11 ± 23%      -0.0        0.10 ± 24%  perf-profile.children.cycles-pp.pick_next_task_fair
      0.23 ±  9%      -0.0        0.22 ±  6%  perf-profile.children.cycles-pp.btrfs_lookup_file_extent
      0.12 ± 20%      -0.0        0.12 ± 38%  perf-profile.children.cycles-pp.ret_from_fork_asm
      0.06 ± 39%      -0.0        0.05 ± 54%  perf-profile.children.cycles-pp.task_tick_fair
      0.40 ±  5%      -0.0        0.40 ±  6%  perf-profile.children.cycles-pp.wait_subpage_spinlock
      0.12 ± 21%      -0.0        0.12 ± 38%  perf-profile.children.cycles-pp.ret_from_fork
      0.03 ±100%      -0.0        0.02 ±123%  perf-profile.children.cycles-pp.hrtimer_next_event_without
      0.12 ± 20%      -0.0        0.12 ± 40%  perf-profile.children.cycles-pp.kthread
      0.07 ± 19%      -0.0        0.07 ± 22%  perf-profile.children.cycles-pp.entry_SYSRETQ_unsafe_stack
      0.07 ± 19%      -0.0        0.06 ± 12%  perf-profile.children.cycles-pp.slab_pre_alloc_hook
      0.01 ±264%      -0.0        0.00        perf-profile.children.cycles-pp.btrfs_del_csums
      0.01 ±264%      -0.0        0.00        perf-profile.children.cycles-pp.ksys_read
      0.01 ±264%      -0.0        0.00        perf-profile.children.cycles-pp.btrfs_get_8
      0.01 ±264%      -0.0        0.00        perf-profile.children.cycles-pp.__switch_to
      0.01 ±264%      -0.0        0.00        perf-profile.children.cycles-pp.hrtimer_update_next_event
      0.01 ±264%      -0.0        0.00        perf-profile.children.cycles-pp.assert_eb_page_uptodate
      0.01 ±264%      -0.0        0.00        perf-profile.children.cycles-pp.__slab_free
      0.18 ± 14%      -0.0        0.17 ± 28%  perf-profile.children.cycles-pp.btrfs_clear_delalloc_extent
      0.15 ±  8%      -0.0        0.14 ±  6%  perf-profile.children.cycles-pp.xas_alloc
      0.12 ±  5%      -0.0        0.12 ± 10%  perf-profile.children.cycles-pp.__filemap_fdatawait_range
      0.08 ± 19%      -0.0        0.08 ± 22%  perf-profile.children.cycles-pp.find_busiest_group
      0.10 ±  7%      -0.0        0.09 ± 20%  perf-profile.children.cycles-pp._find_first_bit
      0.04 ± 78%      -0.0        0.03 ±100%  perf-profile.children.cycles-pp.read_cache_folio
      0.09 ± 14%      -0.0        0.09 ± 16%  perf-profile.children.cycles-pp.perf_pmu_nop_void
      0.03 ±100%      -0.0        0.02 ±123%  perf-profile.children.cycles-pp.__kmem_cache_alloc_node
      0.03 ± 78%      -0.0        0.03 ±100%  perf-profile.children.cycles-pp.read_extent_buffer
      0.09 ± 16%      -0.0        0.08 ± 16%  perf-profile.children.cycles-pp.rwsem_optimistic_spin
      0.04 ± 58%      -0.0        0.04 ± 82%  perf-profile.children.cycles-pp.xas_find_marked
      0.05 ± 60%      -0.0        0.04 ± 50%  perf-profile.children.cycles-pp.btrfs_update_root_times
      0.11 ± 13%      -0.0        0.11 ± 15%  perf-profile.children.cycles-pp.filemap_fdatawait_range
      0.17 ±  7%      -0.0        0.16 ± 13%  perf-profile.children.cycles-pp.__intel_pmu_enable_all
      0.21 ±  7%      -0.0        0.20 ±  6%  perf-profile.children.cycles-pp.rb_next
      0.08 ± 20%      -0.0        0.08 ± 25%  perf-profile.children.cycles-pp.update_sd_lb_stats
      1.85 ±  7%      -0.0        1.85 ±  7%  perf-profile.children.cycles-pp.__sysvec_apic_timer_interrupt
      0.09 ± 14%      -0.0        0.08 ± 14%  perf-profile.children.cycles-pp.__memmove
      0.07 ± 12%      -0.0        0.06 ± 17%  perf-profile.children.cycles-pp.btrfs_alloc_block_rsv
      0.06 ± 20%      -0.0        0.06 ± 10%  perf-profile.children.cycles-pp.lru_add_drain_cpu
      0.08 ± 17%      -0.0        0.08 ± 19%  perf-profile.children.cycles-pp.leaf_space_used
      0.08 ± 13%      -0.0        0.08 ± 16%  perf-profile.children.cycles-pp.filemap_get_folios_tag
      0.02 ±129%      -0.0        0.02 ±153%  perf-profile.children.cycles-pp.__flush_smp_call_function_queue
      0.09 ± 15%      -0.0        0.08 ± 13%  perf-profile.children.cycles-pp.rcu_do_batch
      0.01 ±174%      -0.0        0.01 ±200%  perf-profile.children.cycles-pp.mutex_lock
      0.04 ± 79%      -0.0        0.04 ± 67%  perf-profile.children.cycles-pp._copy_from_user
      0.10 ± 28%      -0.0        0.10 ± 11%  perf-profile.children.cycles-pp.memcg_check_events
      0.02 ±129%      -0.0        0.02 ±153%  perf-profile.children.cycles-pp.filemap_write_and_wait_range
      0.16 ± 18%      -0.0        0.16 ± 15%  perf-profile.children.cycles-pp.btrfs_release_path
      0.09 ± 16%      -0.0        0.08 ± 13%  perf-profile.children.cycles-pp.rwsem_down_write_slowpath
      1.29 ±  2%      -0.0        1.29 ±  6%  perf-profile.children.cycles-pp.__mem_cgroup_charge
      1.84 ±  7%      -0.0        1.83 ±  7%  perf-profile.children.cycles-pp.hrtimer_interrupt
      0.13 ± 10%      -0.0        0.13 ± 13%  perf-profile.children.cycles-pp.btrfs_get_token_32
      0.08 ± 20%      -0.0        0.08 ± 40%  perf-profile.children.cycles-pp.sched_clock_cpu
      0.11 ± 19%      -0.0        0.11 ± 21%  perf-profile.children.cycles-pp.load_balance
      0.27 ±  6%      -0.0        0.27 ±  8%  perf-profile.children.cycles-pp.uncharge_folio
      0.08 ± 16%      -0.0        0.08 ± 21%  perf-profile.children.cycles-pp.page_counter_try_charge
      0.01 ±173%      -0.0        0.01 ±201%  perf-profile.children.cycles-pp.enqueue_hrtimer
      0.08 ± 17%      -0.0        0.08 ±  7%  perf-profile.children.cycles-pp.btrfs_root_node
      0.01 ±264%      -0.0        0.01 ±299%  perf-profile.children.cycles-pp.folio_wait_writeback
      0.01 ±264%      -0.0        0.01 ±300%  perf-profile.children.cycles-pp.btrfs_free_block_rsv
      0.12 ± 19%      -0.0        0.12 ± 14%  perf-profile.children.cycles-pp.up_write
      0.29 ± 16%      -0.0        0.29 ± 14%  perf-profile.children.cycles-pp.menu_select
      0.61 ±  8%      -0.0        0.61 ±  9%  perf-profile.children.cycles-pp.perf_event_task_tick
      0.06 ± 40%      -0.0        0.06 ± 38%  perf-profile.children.cycles-pp.btrfs_leaf_free_space
      0.08 ± 20%      -0.0        0.08 ± 15%  perf-profile.children.cycles-pp.apparmor_file_permission
      0.01 ±264%      -0.0        0.01 ±300%  perf-profile.children.cycles-pp.prepare_task_switch
      0.04 ± 58%      +0.0        0.04 ± 68%  perf-profile.children.cycles-pp.btrfs_dirty_inode
      0.02 ±173%      +0.0        0.02 ±152%  perf-profile.children.cycles-pp.kmalloc_trace
      0.09 ± 15%      +0.0        0.09 ± 18%  perf-profile.children.cycles-pp.find_first_non_hole
      0.15 ± 10%      +0.0        0.15 ±  9%  perf-profile.children.cycles-pp.down_write
      0.01 ±264%      +0.0        0.01 ±300%  perf-profile.children.cycles-pp.timekeeping_max_deferment
      0.04 ± 79%      +0.0        0.04 ± 65%  perf-profile.children.cycles-pp.join_transaction
      0.46 ±  5%      +0.0        0.47 ±  5%  perf-profile.children.cycles-pp.workingset_update_node
      0.09 ± 20%      +0.0        0.09 ± 16%  perf-profile.children.cycles-pp.security_file_permission
      0.19 ± 10%      +0.0        0.19 ± 12%  perf-profile.children.cycles-pp.__reserve_bytes
      0.09 ± 22%      +0.0        0.09 ± 16%  perf-profile.children.cycles-pp.set_extent_bit
      0.05 ± 61%      +0.0        0.05 ± 51%  perf-profile.children.cycles-pp.__libc_start_main
      0.05 ± 61%      +0.0        0.05 ± 51%  perf-profile.children.cycles-pp.main
      0.05 ± 61%      +0.0        0.05 ± 51%  perf-profile.children.cycles-pp.run_builtin
      0.11 ± 17%      +0.0        0.11 ± 11%  perf-profile.children.cycles-pp.rcu_core
      0.65 ±  5%      +0.0        0.65 ±  5%  perf-profile.children.cycles-pp.btrfs_punch_hole
      0.08 ± 15%      +0.0        0.09 ±  8%  perf-profile.children.cycles-pp.btrfs_double_extent_lock
      0.18 ± 12%      +0.0        0.19 ± 12%  perf-profile.children.cycles-pp.btrfs_reserve_metadata_bytes
      0.09 ± 21%      +0.0        0.09 ± 15%  perf-profile.children.cycles-pp.native_irq_return_iret
      0.10 ± 15%      +0.0        0.11 ±  5%  perf-profile.children.cycles-pp.__wake_up_common
      0.10 ± 14%      +0.0        0.11 ± 13%  perf-profile.children.cycles-pp.btrfs_get_32
      0.02 ±129%      +0.0        0.02 ±123%  perf-profile.children.cycles-pp.vfs_fstat
      0.01 ±174%      +0.0        0.02 ±153%  perf-profile.children.cycles-pp.exit_to_user_mode_prepare
      0.01 ±174%      +0.0        0.02 ±153%  perf-profile.children.cycles-pp.newidle_balance
      0.10 ± 10%      +0.0        0.11 ± 13%  perf-profile.children.cycles-pp.rb_erase
      0.20 ± 11%      +0.0        0.20 ±  8%  perf-profile.children.cycles-pp.clockevents_program_event
      0.09 ± 20%      +0.0        0.09 ± 23%  perf-profile.children.cycles-pp.btrfs_bin_search
      0.01 ±173%      +0.0        0.02 ±153%  perf-profile.children.cycles-pp.update_irq_load_avg
      0.01 ±173%      +0.0        0.02 ±153%  perf-profile.children.cycles-pp.security_capable
      0.01 ±173%      +0.0        0.02 ±153%  perf-profile.children.cycles-pp.asm_exc_page_fault
      0.09 ± 19%      +0.0        0.09 ± 17%  perf-profile.children.cycles-pp.xas_expand
      0.01 ±264%      +0.0        0.01 ±200%  perf-profile.children.cycles-pp.stress_file_ioctl
      0.37 ± 15%      +0.0        0.38 ± 10%  perf-profile.children.cycles-pp.__irq_exit_rcu
      0.08 ± 13%      +0.0        0.08 ± 19%  perf-profile.children.cycles-pp.btrfs_punch_hole_lock_range
      0.09 ± 24%      +0.0        0.10 ± 15%  perf-profile.children.cycles-pp.__fget_light
      0.06 ± 40%      +0.0        0.06 ± 57%  perf-profile.children.cycles-pp.ktime_get_update_offsets_now
      0.60 ±  8%      +0.0        0.60 ±  9%  perf-profile.children.cycles-pp.perf_adjust_freq_unthr_context
      0.00            +0.0        0.01 ±299%  perf-profile.children.cycles-pp.finish_task_switch
      0.00            +0.0        0.01 ±299%  perf-profile.children.cycles-pp.add_extent_mapping
      0.00            +0.0        0.01 ±299%  perf-profile.children.cycles-pp.btrfs_trans_release_metadata
      0.00            +0.0        0.01 ±299%  perf-profile.children.cycles-pp.folio_mapping
      0.00            +0.0        0.01 ±299%  perf-profile.children.cycles-pp.vfs_write
      0.00            +0.0        0.01 ±299%  perf-profile.children.cycles-pp.ksys_write
      0.00            +0.0        0.01 ±300%  perf-profile.children.cycles-pp.cpuidle_governor_latency_req
      0.00            +0.0        0.01 ±300%  perf-profile.children.cycles-pp.execve
      0.00            +0.0        0.01 ±300%  perf-profile.children.cycles-pp.__x64_sys_execve
      0.00            +0.0        0.01 ±300%  perf-profile.children.cycles-pp.do_execveat_common
      0.00            +0.0        0.01 ±300%  perf-profile.children.cycles-pp.cmd_record
      0.00            +0.0        0.01 ±300%  perf-profile.children.cycles-pp.__cmd_record
      0.27 ± 30%      +0.0        0.28 ± 39%  perf-profile.children.cycles-pp.ktime_get
      0.02 ±129%      +0.0        0.03 ±100%  perf-profile.children.cycles-pp.capable
      0.01 ±174%      +0.0        0.02 ±123%  perf-profile.children.cycles-pp.update_existing_head_ref
      0.09 ± 20%      +0.0        0.09 ± 24%  perf-profile.children.cycles-pp.native_sched_clock
      0.15 ± 16%      +0.0        0.16 ± 22%  perf-profile.children.cycles-pp.__count_memcg_events
      0.07 ± 22%      +0.0        0.08 ± 17%  perf-profile.children.cycles-pp.btrfs_block_rsv_release
      0.01 ±264%      +0.0        0.02 ±152%  perf-profile.children.cycles-pp.local_clock_noinstr
      0.28 ±  7%      +0.0        0.29 ±  6%  perf-profile.children.cycles-pp.xas_create
      0.00            +0.0        0.01 ±200%  perf-profile.children.cycles-pp.btrfs_inode_lock
      0.00            +0.0        0.01 ±200%  perf-profile.children.cycles-pp.need_preemptive_reclaim
      0.51 ±  7%      +0.0        0.52 ±  2%  perf-profile.children.cycles-pp.btrfs_add_delayed_data_ref
      0.00            +0.0        0.01 ±201%  perf-profile.children.cycles-pp.tick_nohz_irq_exit
      0.02 ±129%      +0.0        0.03 ±102%  perf-profile.children.cycles-pp.poll_idle
      0.07 ± 17%      +0.0        0.09 ± 20%  perf-profile.children.cycles-pp.__btrfs_end_transaction
      0.23 ±  9%      +0.0        0.24 ± 13%  perf-profile.children.cycles-pp.free_extent_map
      0.06 ± 44%      +0.0        0.07 ± 18%  perf-profile.children.cycles-pp.update_blocked_averages
      0.06 ± 45%      +0.0        0.08 ± 42%  perf-profile.children.cycles-pp.arch_scale_freq_tick
      0.07 ± 15%      +0.0        0.09 ± 21%  perf-profile.children.cycles-pp.irqtime_account_irq
      0.04 ± 60%      +0.0        0.06 ± 35%  perf-profile.children.cycles-pp.rb_insert_color
      0.13 ± 21%      +0.0        0.15 ±  7%  perf-profile.children.cycles-pp.lapic_next_deadline
      0.01 ±173%      +0.0        0.03 ±104%  perf-profile.children.cycles-pp.btrfs_find_space_for_alloc
      0.02 ±131%      +0.0        0.04 ± 66%  perf-profile.children.cycles-pp.btrfs_reduce_alloc_profile
      0.06 ± 43%      +0.0        0.08 ± 18%  perf-profile.children.cycles-pp.run_rebalance_domains
      0.05 ± 65%      +0.0        0.06 ± 12%  perf-profile.children.cycles-pp.insert_delayed_ref
      1.07 ±  5%      +0.0        1.09 ±  7%  perf-profile.children.cycles-pp.tick_sched_timer
      0.38 ± 14%      +0.0        0.40 ±  9%  perf-profile.children.cycles-pp.get_mem_cgroup_from_mm
      0.84 ±  5%      +0.0        0.86 ±  7%  perf-profile.children.cycles-pp.scheduler_tick
      0.01 ±264%      +0.0        0.03 ±100%  perf-profile.children.cycles-pp.syscall_enter_from_user_mode
      0.00            +0.0        0.02 ±122%  perf-profile.children.cycles-pp.___slab_alloc
      2.45 ±  9%      +0.0        2.47 ±  7%  perf-profile.children.cycles-pp.sysvec_apic_timer_interrupt
      0.02 ±131%      +0.0        0.04 ± 51%  perf-profile.children.cycles-pp.calc_available_free_space
      0.01 ±174%      +0.0        0.04 ± 66%  perf-profile.children.cycles-pp.update_curr
      0.02 ±134%      +0.0        0.05 ± 52%  perf-profile.children.cycles-pp.btrfs_can_overcommit
      2.69 ±  7%      +0.0        2.71 ±  7%  perf-profile.children.cycles-pp.asm_sysvec_apic_timer_interrupt
      0.98 ±  5%      +0.0        1.01 ±  7%  perf-profile.children.cycles-pp.tick_sched_handle
      0.42 ±  7%      +0.0        0.45 ±  9%  perf-profile.children.cycles-pp.mem_cgroup_commit_charge
      0.13 ± 74%      +0.0        0.16 ± 73%  perf-profile.children.cycles-pp.irq_enter_rcu
      0.04 ± 80%      +0.0        0.07 ± 18%  perf-profile.children.cycles-pp.rcu_sched_clock_irq
      0.96 ±  5%      +0.0        1.00 ±  7%  perf-profile.children.cycles-pp.update_process_times
      0.01 ±264%      +0.0        0.04 ± 67%  perf-profile.children.cycles-pp.rcu_pending
      0.11 ± 86%      +0.0        0.15 ± 77%  perf-profile.children.cycles-pp.tick_irq_enter
     42.93 ±  2%      +0.1       42.98 ±  2%  perf-profile.children.cycles-pp.vfs_dedupe_file_range
     42.92 ±  2%      +0.1       42.97 ±  2%  perf-profile.children.cycles-pp.vfs_dedupe_file_range_one
      1.00 ±  7%      +0.1        1.08 ±  5%  perf-profile.children.cycles-pp.btrfs_extent_same_range
      0.29 ± 16%      +0.1        0.39 ±  6%  perf-profile.children.cycles-pp.__mod_zone_page_state
      0.52 ±113%      +0.2        0.67 ± 96%  perf-profile.children.cycles-pp.x86_64_start_kernel
      0.52 ±113%      +0.2        0.67 ± 96%  perf-profile.children.cycles-pp.x86_64_start_reservations
      0.52 ±113%      +0.2        0.67 ± 96%  perf-profile.children.cycles-pp.start_kernel
      0.52 ±113%      +0.2        0.67 ± 96%  perf-profile.children.cycles-pp.arch_call_rest_init
      0.52 ±113%      +0.2        0.67 ± 96%  perf-profile.children.cycles-pp.rest_init
      0.26 ±  8%      +0.2        0.41 ±  7%  perf-profile.children.cycles-pp.__list_add_valid_or_report
      0.08 ± 21%      +0.2        0.25 ±  8%  perf-profile.children.cycles-pp.__kmalloc_node
      0.08 ± 22%      +0.2        0.24 ±  8%  perf-profile.children.cycles-pp.__kmalloc_large_node
      0.08 ± 21%      +0.2        0.25 ±  6%  perf-profile.children.cycles-pp.kvmalloc_node
     70.99            +0.3       71.33 ±  2%  perf-profile.children.cycles-pp.stress_run
     70.65            +0.4       71.02 ±  2%  perf-profile.children.cycles-pp.ioctl
     70.37            +0.4       70.77 ±  2%  perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe
     69.83            +0.4       70.25 ±  2%  perf-profile.children.cycles-pp.do_syscall_64
     68.25            +0.5       68.76 ±  2%  perf-profile.children.cycles-pp.__x64_sys_ioctl
      1.13 ±  3%      +0.6        1.69 ±  5%  perf-profile.children.cycles-pp._raw_spin_lock_irqsave
      0.31 ± 11%      +0.6        0.90 ±  6%  perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath
     22.35            +0.6       22.97 ±  2%  perf-profile.children.cycles-pp.ioctl_file_clone
     21.97            +0.6       22.59 ±  2%  perf-profile.children.cycles-pp.btrfs_clone_files
     22.34            +0.6       22.96 ±  2%  perf-profile.children.cycles-pp.vfs_clone_file_range
     22.32            +0.6       22.94 ±  2%  perf-profile.children.cycles-pp.do_clone_file_range
     30.09 ±  2%      +0.6       30.73 ±  3%  perf-profile.children.cycles-pp.do_read_cache_folio
     65.16            +0.7       65.82 ±  2%  perf-profile.children.cycles-pp.btrfs_remap_file_range
     65.61            +0.7       66.28 ±  2%  perf-profile.children.cycles-pp.do_vfs_ioctl
      0.52 ±  4%      +1.0        1.48 ±  6%  perf-profile.children.cycles-pp.__list_del_entry_valid_or_report
      0.00            +1.5        1.54 ±  6%  perf-profile.children.cycles-pp.__free_one_page
      2.58 ±  3%      +2.0        4.55 ±  5%  perf-profile.children.cycles-pp.folio_alloc
      0.93 ±  6%      +2.1        3.02 ±  5%  perf-profile.children.cycles-pp.rmqueue
      0.00            +2.1        2.10 ±  4%  perf-profile.children.cycles-pp.free_pcppages_bulk
      0.01 ±264%      +2.2        2.17 ±  4%  perf-profile.children.cycles-pp.free_unref_page
      2.03 ±  3%      +2.2        4.19 ±  4%  perf-profile.children.cycles-pp.__alloc_pages
      1.25 ±  6%      +2.2        3.42 ±  5%  perf-profile.children.cycles-pp.get_page_from_freelist
      0.37 ± 10%      +2.2        2.54 ±  4%  perf-profile.children.cycles-pp.free_unref_page_commit
      2.60 ±  4%      +2.2        4.77 ±  3%  perf-profile.children.cycles-pp.btrfs_clone
      0.00            +2.2        2.20 ±  6%  perf-profile.children.cycles-pp.rmqueue_bulk
      0.00            +2.5        2.49 ±  6%  perf-profile.children.cycles-pp.__rmqueue_pcplist
      9.91 ±  2%      -0.5        9.38 ±  2%  perf-profile.self.cycles-pp.memcmp
     23.80 ±  4%      -0.2       23.59 ±  5%  perf-profile.self.cycles-pp.intel_idle_ibrs
      3.24 ±  2%      -0.2        3.05 ±  3%  perf-profile.self.cycles-pp.kmem_cache_alloc
      1.16 ±  3%      -0.1        1.02 ±  4%  perf-profile.self.cycles-pp.__set_extent_bit
      5.27            -0.1        5.12 ±  2%  perf-profile.self.cycles-pp.memset_orig
      0.63 ±  6%      -0.1        0.50 ±  7%  perf-profile.self.cycles-pp.__mod_memcg_lruvec_state
      1.19 ±  2%      -0.1        1.06 ±  3%  perf-profile.self.cycles-pp.free_extent_state
      1.46 ±  4%      -0.1        1.36 ±  4%  perf-profile.self.cycles-pp.xas_store
      0.92 ±  3%      -0.1        0.84 ±  6%  perf-profile.self.cycles-pp.alloc_extent_state
      1.45 ±  3%      -0.1        1.37 ±  4%  perf-profile.self.cycles-pp.__clear_extent_bit
      0.58 ±  7%      -0.1        0.51 ± 11%  perf-profile.self.cycles-pp.set_page_extent_mapped
      1.15 ±  4%      -0.1        1.08 ±  9%  perf-profile.self.cycles-pp.vfs_dedupe_file_range_compare
      0.62 ±  5%      -0.1        0.55 ±  6%  perf-profile.self.cycles-pp.clear_page_extent_mapped
      0.84 ±  4%      -0.1        0.78 ±  5%  perf-profile.self.cycles-pp.syscall_exit_to_user_mode
      0.88 ±  4%      -0.1        0.82 ±  4%  perf-profile.self.cycles-pp.btrfs_invalidate_folio
      0.94 ±  8%      -0.1        0.88 ±  7%  perf-profile.self.cycles-pp.clear_state_bit
      0.82 ±  7%      -0.1        0.76 ±  4%  perf-profile.self.cycles-pp.lru_add_fn
      0.97 ±  3%      -0.1        0.92 ±  3%  perf-profile.self.cycles-pp.syscall_return_via_sysret
      0.60 ±  9%      -0.1        0.54 ±  7%  perf-profile.self.cycles-pp.cache_state_if_flags
      1.51 ±  4%      -0.1        1.46 ±  5%  perf-profile.self.cycles-pp.kmem_cache_free
      0.38 ±  7%      -0.1        0.33 ±  9%  perf-profile.self.cycles-pp.__cond_resched
      0.42 ±  7%      -0.1        0.37 ±  9%  perf-profile.self.cycles-pp.__entry_text_start
      0.79 ±  4%      -0.1        0.74 ±  5%  perf-profile.self.cycles-pp.folio_unlock
      1.11 ±  4%      -0.0        1.06 ±  4%  perf-profile.self.cycles-pp.release_pages
      0.27 ± 11%      -0.0        0.22 ±  9%  perf-profile.self.cycles-pp._raw_read_lock
      0.41 ±  5%      -0.0        0.36 ±  5%  perf-profile.self.cycles-pp.memcg_slab_post_alloc_hook
      0.25 ±  7%      -0.0        0.20 ± 12%  perf-profile.self.cycles-pp.try_charge_memcg
      0.33 ±  4%      -0.0        0.29 ± 13%  perf-profile.self.cycles-pp.filemap_read_folio
      0.53 ±  2%      -0.0        0.48 ±  7%  perf-profile.self.cycles-pp.delete_from_page_cache_batch
      0.23 ±  8%      -0.0        0.19 ± 12%  perf-profile.self.cycles-pp.xas_find_conflict
      1.92 ±  4%      -0.0        1.88 ±  5%  perf-profile.self.cycles-pp._raw_spin_lock
      0.63 ±  7%      -0.0        0.59 ± 18%  perf-profile.self.cycles-pp.intel_idle
      0.49 ±  7%      -0.0        0.46 ±  9%  perf-profile.self.cycles-pp.xas_load
      0.50 ±  8%      -0.0        0.47 ±  8%  perf-profile.self.cycles-pp.__mod_lruvec_page_state
      0.28 ± 18%      -0.0        0.24 ± 14%  perf-profile.self.cycles-pp.cgroup_rstat_updated
      0.15 ± 16%      -0.0        0.12 ± 22%  perf-profile.self.cycles-pp.end_page_read
      0.21 ± 11%      -0.0        0.18 ± 11%  perf-profile.self.cycles-pp.btrfs_page_clear_checked
      0.53 ±  5%      -0.0        0.50 ±  6%  perf-profile.self.cycles-pp.folio_batch_move_lru
      0.29 ± 11%      -0.0        0.26 ±  8%  perf-profile.self.cycles-pp.xas_clear_mark
      0.40 ±  7%      -0.0        0.37 ±  5%  perf-profile.self.cycles-pp.btrfs_lock_and_flush_ordered_range
      0.34 ±  9%      -0.0        0.31 ±  6%  perf-profile.self.cycles-pp.btrfs_lookup_ordered_range
      0.28 ± 11%      -0.0        0.25 ± 12%  perf-profile.self.cycles-pp.__lookup_extent_mapping
      0.41 ±  7%      -0.0        0.38 ±  8%  perf-profile.self.cycles-pp.free_unref_page_list
      0.42 ±  7%      -0.0        0.39 ±  9%  perf-profile.self.cycles-pp.do_read_cache_folio
      0.20 ± 11%      -0.0        0.17 ±  9%  perf-profile.self.cycles-pp.__mod_lruvec_state
      0.23 ±  8%      -0.0        0.20 ± 10%  perf-profile.self.cycles-pp.filemap_free_folio
      0.24 ± 13%      -0.0        0.21 ±  6%  perf-profile.self.cycles-pp._raw_spin_trylock
      0.67 ±  8%      -0.0        0.64 ±  4%  perf-profile.self.cycles-pp._raw_spin_lock_irq
      0.03 ± 78%      -0.0        0.01 ±300%  perf-profile.self.cycles-pp.cpuidle_idle_call
      0.28 ± 11%      -0.0        0.25 ± 11%  perf-profile.self.cycles-pp.btrfs_page_set_uptodate
      0.17 ±  6%      -0.0        0.14 ± 11%  perf-profile.self.cycles-pp.xas_find
      0.17 ±  5%      -0.0        0.14 ± 15%  perf-profile.self.cycles-pp.rcu_all_qs
      0.29 ± 12%      -0.0        0.27 ± 10%  perf-profile.self.cycles-pp.__init_waitqueue_head
      0.46 ±  8%      -0.0        0.43 ±  8%  perf-profile.self.cycles-pp.folio_mark_accessed
      0.14 ± 12%      -0.0        0.12 ± 17%  perf-profile.self.cycles-pp.read_tsc
      0.12 ± 23%      -0.0        0.09 ± 14%  perf-profile.self.cycles-pp.btrfs_lookup_first_ordered_range
      1.16 ±  5%      -0.0        1.14 ±  4%  perf-profile.self.cycles-pp.__filemap_add_folio
      0.19 ±  6%      -0.0        0.17 ± 12%  perf-profile.self.cycles-pp.find_free_extent
      0.30 ±  6%      -0.0        0.28 ± 10%  perf-profile.self.cycles-pp.__filemap_get_folio
      0.13 ±  9%      -0.0        0.11 ± 14%  perf-profile.self.cycles-pp.__tree_search
      0.32 ±  7%      -0.0        0.30 ±  9%  perf-profile.self.cycles-pp.btrfs_is_subpage
      0.24 ±  7%      -0.0        0.22 ± 11%  perf-profile.self.cycles-pp.__wake_up
      0.03 ±100%      -0.0        0.01 ±200%  perf-profile.self.cycles-pp.__mem_cgroup_uncharge_list
      0.11 ± 12%      -0.0        0.09 ± 18%  perf-profile.self.cycles-pp.btrfs_put_block_group
      0.55 ±  7%      -0.0        0.53 ± 12%  perf-profile.self.cycles-pp.entry_SYSCALL_64_after_hwframe
      0.81 ±  3%      -0.0        0.79 ±  5%  perf-profile.self.cycles-pp._raw_spin_lock_irqsave
      0.02 ±100%      -0.0        0.01 ±300%  perf-profile.self.cycles-pp.__update_load_avg_cfs_rq
      0.14 ± 12%      -0.0        0.12 ±  9%  perf-profile.self.cycles-pp.page_counter_uncharge
      0.22 ±  7%      -0.0        0.20 ±  8%  perf-profile.self.cycles-pp.truncate_cleanup_folio
      0.60 ±  9%      -0.0        0.58 ±  7%  perf-profile.self.cycles-pp.__mod_node_page_state
      0.25 ±  7%      -0.0        0.23 ±  9%  perf-profile.self.cycles-pp.get_pfnblock_flags_mask
      0.70 ± 13%      -0.0        0.69 ±  8%  perf-profile.self.cycles-pp.xas_descend
      0.30 ±  6%      -0.0        0.28 ± 14%  perf-profile.self.cycles-pp.alloc_pages
      0.02 ±129%      -0.0        0.01 ±299%  perf-profile.self.cycles-pp.entry_SYSCALL_64_safe_stack
      0.06 ± 14%      -0.0        0.04 ± 68%  perf-profile.self.cycles-pp.btrfs_get_64
      0.13 ±  9%      -0.0        0.11 ± 16%  perf-profile.self.cycles-pp._raw_spin_unlock_irqrestore
      0.12 ± 14%      -0.0        0.10 ± 16%  perf-profile.self.cycles-pp.btrfs_get_block_group
      0.40 ±  8%      -0.0        0.39 ±  8%  perf-profile.self.cycles-pp.btrfs_get_extent
      0.18 ±  9%      -0.0        0.16 ± 11%  perf-profile.self.cycles-pp.free_unref_page_prepare
      0.02 ±129%      -0.0        0.01 ±299%  perf-profile.self.cycles-pp.set_extent_buffer_dirty
      0.20 ± 10%      -0.0        0.18 ±  5%  perf-profile.self.cycles-pp.add_delayed_ref_head
      0.05 ± 38%      -0.0        0.04 ± 65%  perf-profile.self.cycles-pp.mutex_unlock
      0.09 ±  7%      -0.0        0.08 ± 23%  perf-profile.self.cycles-pp._find_first_bit
      0.02 ±129%      -0.0        0.01 ±300%  perf-profile.self.cycles-pp.tick_sched_do_timer
      0.09 ±  7%      -0.0        0.08 ± 26%  perf-profile.self.cycles-pp.btrfs_set_token_32
      0.22 ±  8%      -0.0        0.21 ± 12%  perf-profile.self.cycles-pp.wait_subpage_spinlock
      0.15 ± 18%      -0.0        0.14 ± 15%  perf-profile.self.cycles-pp.test_range_bit
      0.25 ± 10%      -0.0        0.24 ± 12%  perf-profile.self.cycles-pp.cpuidle_enter_state
      0.18 ± 14%      -0.0        0.17 ± 15%  perf-profile.self.cycles-pp.xa_get_order
      0.05 ± 40%      -0.0        0.04 ± 83%  perf-profile.self.cycles-pp.intel_idle_irq
      0.03 ±100%      -0.0        0.02 ±153%  perf-profile.self.cycles-pp.update_load_avg
      0.78 ±  6%      -0.0        0.76 ±  7%  perf-profile.self.cycles-pp.find_lock_entries
      0.03 ± 77%      -0.0        0.02 ±122%  perf-profile.self.cycles-pp.read_cache_folio
      0.12 ± 12%      -0.0        0.11 ± 11%  perf-profile.self.cycles-pp.__btrfs_qgroup_release_data
      0.06 ± 13%      -0.0        0.05 ± 53%  perf-profile.self.cycles-pp.__write_extent_buffer
      0.03 ±100%      -0.0        0.02 ±155%  perf-profile.self.cycles-pp.hrtimer_interrupt
      0.10 ± 19%      -0.0        0.08 ± 18%  perf-profile.self.cycles-pp.filemap_alloc_folio
      0.02 ±129%      -0.0        0.01 ±200%  perf-profile.self.cycles-pp._copy_from_user
      0.26 ±  6%      -0.0        0.25 ±  8%  perf-profile.self.cycles-pp.lock_extent
      0.19 ± 10%      -0.0        0.18 ± 13%  perf-profile.self.cycles-pp.btrfs_set_delalloc_extent
      0.03 ±100%      -0.0        0.02 ±123%  perf-profile.self.cycles-pp.__schedule
      0.02 ±129%      -0.0        0.01 ±201%  perf-profile.self.cycles-pp.__kmem_cache_alloc_node
      0.03 ±100%      -0.0        0.02 ±154%  perf-profile.self.cycles-pp.__filemap_fdatawait_range
      0.27 ±  3%      -0.0        0.26 ± 14%  perf-profile.self.cycles-pp.folio_add_lru
      0.03 ±102%      -0.0        0.02 ±123%  perf-profile.self.cycles-pp.start_transaction
      0.06 ± 15%      -0.0        0.06 ± 16%  perf-profile.self.cycles-pp.entry_SYSRETQ_unsafe_stack
      0.01 ±173%      -0.0        0.01 ±299%  perf-profile.self.cycles-pp.__switch_to_asm
      0.01 ±264%      -0.0        0.00        perf-profile.self.cycles-pp.mutex_lock
      0.10 ± 15%      -0.0        0.09 ± 13%  perf-profile.self.cycles-pp.rb_next
      0.04 ± 80%      -0.0        0.03 ±103%  perf-profile.self.cycles-pp.__filemap_fdatawrite_range
      0.09 ± 12%      -0.0        0.08 ± 19%  perf-profile.self.cycles-pp.perf_pmu_nop_void
      0.07 ± 16%      -0.0        0.07 ± 39%  perf-profile.self.cycles-pp.page_counter_try_charge
      0.01 ±264%      -0.0        0.00        perf-profile.self.cycles-pp.btrfs_wait_ordered_range
      0.01 ±264%      -0.0        0.00        perf-profile.self.cycles-pp.xas_nomem
      0.01 ±264%      -0.0        0.00        perf-profile.self.cycles-pp.btrfs_insert_replace_extent
      0.01 ±264%      -0.0        0.00        perf-profile.self.cycles-pp.btrfs_get_8
      0.01 ±264%      -0.0        0.00        perf-profile.self.cycles-pp.assert_eb_page_uptodate
      0.01 ±264%      -0.0        0.00        perf-profile.self.cycles-pp.__slab_free
      0.01 ±264%      -0.0        0.00        perf-profile.self.cycles-pp.uncharge_batch
      0.26 ±  9%      -0.0        0.26 ±  9%  perf-profile.self.cycles-pp.mem_cgroup_update_lru_size
      0.05 ± 62%      -0.0        0.05 ± 69%  perf-profile.self.cycles-pp.btrfs_drop_extents
      0.07 ± 17%      -0.0        0.07 ± 15%  perf-profile.self.cycles-pp.setup_items_for_insert
      0.17 ±  7%      -0.0        0.16 ± 13%  perf-profile.self.cycles-pp.__intel_pmu_enable_all
      0.09 ± 14%      -0.0        0.08 ± 14%  perf-profile.self.cycles-pp.__memmove
      0.13 ± 11%      -0.0        0.12 ± 12%  perf-profile.self.cycles-pp.btrfs_get_token_32
      0.57 ± 15%      -0.0        0.57 ± 14%  perf-profile.self.cycles-pp.btrfs_do_readpage
      0.04 ± 58%      -0.0        0.04 ± 82%  perf-profile.self.cycles-pp.xas_find_marked
      0.03 ±101%      -0.0        0.03 ±100%  perf-profile.self.cycles-pp.btrfs_fallocate
      0.08 ± 17%      -0.0        0.08 ± 10%  perf-profile.self.cycles-pp.btrfs_root_node
      0.20 ± 13%      -0.0        0.19 ±  9%  perf-profile.self.cycles-pp.filemap_unaccount_folio
      0.10 ± 12%      -0.0        0.10 ± 13%  perf-profile.self.cycles-pp.btrfs_get_32
      0.04 ± 58%      -0.0        0.04 ± 67%  perf-profile.self.cycles-pp.btrfs_clone
      0.35 ± 10%      -0.0        0.35 ±  6%  perf-profile.self.cycles-pp.xas_start
      0.07 ± 15%      -0.0        0.07 ±  6%  perf-profile.self.cycles-pp.ioctl
      0.07 ± 22%      -0.0        0.07 ± 17%  perf-profile.self.cycles-pp.apparmor_file_permission
      0.07 ± 45%      -0.0        0.07 ± 14%  perf-profile.self.cycles-pp.memcg_check_events
      0.46 ±  5%      -0.0        0.46 ±  5%  perf-profile.self.cycles-pp.workingset_update_node
      0.10 ± 15%      -0.0        0.10 ±  7%  perf-profile.self.cycles-pp.__wake_up_common
      0.08 ± 22%      -0.0        0.08 ± 39%  perf-profile.self.cycles-pp.native_sched_clock
      0.03 ±100%      -0.0        0.03 ±100%  perf-profile.self.cycles-pp.read_extent_buffer
      0.06 ± 13%      +0.0        0.06 ± 15%  perf-profile.self.cycles-pp.should_fail_alloc_page
      0.01 ±264%      +0.0        0.01 ±300%  perf-profile.self.cycles-pp.timekeeping_max_deferment
      0.05 ± 41%      +0.0        0.05 ± 40%  perf-profile.self.cycles-pp.rwsem_optimistic_spin
      0.35 ±  7%      +0.0        0.35 ±  6%  perf-profile.self.cycles-pp.rmqueue
      0.16 ± 17%      +0.0        0.16 ± 32%  perf-profile.self.cycles-pp.btrfs_clear_delalloc_extent
      0.15 ± 21%      +0.0        0.15 ± 10%  perf-profile.self.cycles-pp.set_state_bits
      0.25 ±  5%      +0.0        0.25 ±  8%  perf-profile.self.cycles-pp.uncharge_folio
      0.09 ± 21%      +0.0        0.09 ± 15%  perf-profile.self.cycles-pp.native_irq_return_iret
      0.86 ±  4%      +0.0        0.86 ±  6%  perf-profile.self.cycles-pp.filemap_get_entry
      0.10 ± 11%      +0.0        0.10 ± 14%  perf-profile.self.cycles-pp.rb_erase
      0.21 ±  9%      +0.0        0.21 ±  8%  perf-profile.self.cycles-pp.should_failslab
      0.02 ±129%      +0.0        0.02 ±124%  perf-profile.self.cycles-pp.btrfs_remap_file_range
      0.09 ± 19%      +0.0        0.09 ± 24%  perf-profile.self.cycles-pp.btrfs_bin_search
      0.01 ±173%      +0.0        0.02 ±153%  perf-profile.self.cycles-pp.update_irq_load_avg
      0.01 ±173%      +0.0        0.02 ±153%  perf-profile.self.cycles-pp.folio_lruvec_lock_irqsave
      0.10 ± 17%      +0.0        0.10 ± 15%  perf-profile.self.cycles-pp.policy_node
      0.01 ±264%      +0.0        0.01 ±201%  perf-profile.self.cycles-pp.__generic_remap_file_range_prep
      0.00            +0.0        0.01 ±299%  perf-profile.self.cycles-pp.vfs_fallocate
      0.00            +0.0        0.01 ±299%  perf-profile.self.cycles-pp.syscall_enter_from_user_mode
      0.00            +0.0        0.01 ±299%  perf-profile.self.cycles-pp.policy_nodemask
      0.00            +0.0        0.01 ±300%  perf-profile.self.cycles-pp.try_to_wake_up
      0.00            +0.0        0.01 ±300%  perf-profile.self.cycles-pp.__btrfs_end_transaction
      0.00            +0.0        0.01 ±300%  perf-profile.self.cycles-pp.exit_to_user_mode_prepare
      0.01 ±264%      +0.0        0.01 ±200%  perf-profile.self.cycles-pp.update_sg_lb_stats
      0.09 ± 24%      +0.0        0.09 ± 15%  perf-profile.self.cycles-pp.btrfs_search_slot
      0.09 ± 27%      +0.0        0.10 ± 17%  perf-profile.self.cycles-pp.__fget_light
      0.05 ± 39%      +0.0        0.06 ± 22%  perf-profile.self.cycles-pp.do_syscall_64
      0.04 ± 78%      +0.0        0.04 ± 51%  perf-profile.self.cycles-pp.btrfs_add_delayed_data_ref
      0.07 ± 42%      +0.0        0.08 ± 22%  perf-profile.self.cycles-pp.truncate_inode_pages_range
      0.12 ± 13%      +0.0        0.13 ± 18%  perf-profile.self.cycles-pp.menu_select
      0.03 ±130%      +0.0        0.04 ± 85%  perf-profile.self.cycles-pp.__hrtimer_run_queues
      0.10 ± 20%      +0.0        0.12 ± 23%  perf-profile.self.cycles-pp.__count_memcg_events
      0.64 ±  6%      +0.0        0.65 ±  5%  perf-profile.self.cycles-pp.__alloc_pages
      0.22 ±  5%      +0.0        0.24 ±  8%  perf-profile.self.cycles-pp.filemap_add_folio
      0.00            +0.0        0.01 ±201%  perf-profile.self.cycles-pp.__btrfs_release_delayed_node
      0.01 ±174%      +0.0        0.02 ±100%  perf-profile.self.cycles-pp.btrfs_replace_file_extents
      0.17 ±  9%      +0.0        0.18 ±  8%  perf-profile.self.cycles-pp.__mem_cgroup_charge
      0.11 ± 22%      +0.0        0.12 ± 15%  perf-profile.self.cycles-pp.try_release_extent_mapping
      0.22 ±  9%      +0.0        0.24 ± 13%  perf-profile.self.cycles-pp.free_extent_map
      0.04 ± 60%      +0.0        0.06 ± 35%  perf-profile.self.cycles-pp.rb_insert_color
      0.06 ± 45%      +0.0        0.08 ± 42%  perf-profile.self.cycles-pp.arch_scale_freq_tick
      0.02 ±129%      +0.0        0.03 ±104%  perf-profile.self.cycles-pp.ktime_get_update_offsets_now
      0.10 ± 12%      +0.0        0.12 ±  9%  perf-profile.self.cycles-pp.xas_create
      0.01 ±174%      +0.0        0.03 ±100%  perf-profile.self.cycles-pp.down_write
      0.01 ±264%      +0.0        0.02 ±123%  perf-profile.self.cycles-pp.irqtime_account_irq
      0.01 ±174%      +0.0        0.03 ±102%  perf-profile.self.cycles-pp.poll_idle
      0.14 ± 15%      +0.0        0.15 ± 14%  perf-profile.self.cycles-pp.btrfs_read_folio
      0.13 ± 21%      +0.0        0.15 ±  7%  perf-profile.self.cycles-pp.lapic_next_deadline
      0.37 ± 14%      +0.0        0.39 ± 10%  perf-profile.self.cycles-pp.get_mem_cgroup_from_mm
      0.02 ±132%      +0.0        0.04 ± 51%  perf-profile.self.cycles-pp.up_write
      0.02 ±130%      +0.0        0.04 ± 52%  perf-profile.self.cycles-pp.btrfs_del_items
      0.16 ± 53%      +0.0        0.18 ± 58%  perf-profile.self.cycles-pp.ktime_get
      0.03 ±104%      +0.0        0.05 ± 38%  perf-profile.self.cycles-pp.__x64_sys_ioctl
      0.07 ± 18%      +0.0        0.10 ± 12%  perf-profile.self.cycles-pp.folio_alloc
      0.34 ±  7%      +0.0        0.37 ±  8%  perf-profile.self.cycles-pp.perf_adjust_freq_unthr_context
      0.04 ± 79%      +0.0        0.06 ± 15%  perf-profile.self.cycles-pp.do_vfs_ioctl
      0.20 ± 11%      +0.0        0.23 ± 10%  perf-profile.self.cycles-pp.mem_cgroup_commit_charge
      0.31 ± 12%      +0.1        0.36 ±  6%  perf-profile.self.cycles-pp.free_unref_page_commit
      0.32 ±  9%      +0.1        0.37 ±  9%  perf-profile.self.cycles-pp.get_page_from_freelist
      0.26 ± 19%      +0.1        0.36 ±  5%  perf-profile.self.cycles-pp.__mod_zone_page_state
      0.24 ±  8%      +0.1        0.38 ±  6%  perf-profile.self.cycles-pp.__list_add_valid_or_report
      0.00            +0.2        0.15 ± 10%  perf-profile.self.cycles-pp.__rmqueue_pcplist
      0.00            +0.2        0.24 ±  8%  perf-profile.self.cycles-pp.free_pcppages_bulk
      0.31 ± 11%      +0.6        0.90 ±  6%  perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath
      0.00            +0.6        0.62 ± 13%  perf-profile.self.cycles-pp.rmqueue_bulk
      0.51 ±  4%      +0.9        1.44 ±  6%  perf-profile.self.cycles-pp.__list_del_entry_valid_or_report
      0.00            +1.3        1.32 ±  6%  perf-profile.self.cycles-pp.__free_one_page

xsang@inn:~$

[Index of Archives]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux