Re: [PATCH] /dev/zero: make private mapping full anonymous mapping

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



hi, Yang Shi,

On Fri, Jan 31, 2025 at 10:38:03AM -0800, Yang Shi wrote:
> 
> 
> 
> On 1/27/25 7:14 PM, kernel test robot wrote:
> > hi, All,
> > 
> > we don't have enough knowledge to understand fully the discussion for this
> > patch, we saw "NACK" but there were more discussions later.
> > so below report is just FYI what we observed in our tests. thanks
> 
> Thanks for the report. It was nack'ed because of the change to smaps/maps
> files in proc.
> 
> > 
> > Hello,
> > 
> > kernel test robot noticed a 858.5% improvement of vm-scalability.throughput on:
> > 
> > 
> > commit: 7143ee2391f1ea15e6791e129870473543634de2 ("[PATCH] /dev/zero: make private mapping full anonymous mapping")
> > url: https://github.com/intel-lab-lkp/linux/commits/Yang-Shi/dev-zero-make-private-mapping-full-anonymous-mapping/20250114-063339
> > base: https://git.kernel.org/cgit/linux/kernel/git/gregkh/char-misc.git a68d3cbfade64392507302f3a920113b60dc811f
> > patch link: https://lore.kernel.org/all/20250113223033.4054534-1-yang@xxxxxxxxxxxxxxxxxxxxxx/
> > patch subject: [PATCH] /dev/zero: make private mapping full anonymous mapping
> > 
> > testcase: vm-scalability
> > config: x86_64-rhel-9.4
> > compiler: gcc-12
> > test machine: 224 threads 4 sockets Intel(R) Xeon(R) Platinum 8380H CPU @ 2.90GHz (Cooper Lake) with 192G memory
> > parameters:
> > 
> > 	runtime: 300s
> > 	test: small-allocs
> 
> It seems this benchmark allocates huge amount of small areas (each area is
> as big as 40K) by mmap'ing /dev/zero.
> 
> This patch makes /dev/zero mapping a full anonymous mapping, so the later
> vma_link_file() is actually skipped, which needs acquire file rmap lock then
> insert the mapping into file rmap tree. The below profiling also showed
> this.
> Quoted here so that we don't have to scroll down:
> 
> >       95.60           -95.2        0.42 ±113%  perf-profile.children.cycles-pp.__mmap
> >       94.14           -93.6        0.54 ±106%  perf-profile.children.cycles-pp.__mmap_new_vma
> >       93.79           -93.6        0.21 ±171%  perf-profile.children.cycles-pp.vma_link_file
> >       93.40           -93.4        0.00        perf-profile.children.cycles-pp.rwsem_down_write_slowpath
> >       93.33           -93.3        0.00        perf-profile.children.cycles-pp.rwsem_optimistic_spin
> >       93.44           -93.2        0.24 ±178%  perf-profile.children.cycles-pp.down_write
> >       94.55           -93.1        1.40 ± 51%  perf-profile.children.cycles-pp.ksys_mmap_pgoff
> >       94.25           -93.0        1.30 ± 59%  perf-profile.children.cycles-pp.__mmap_region
> >       92.91           -92.9        0.00        perf-profile.children.cycles-pp.osq_lock
> >       94.45           -92.7        1.72 ± 34%  perf-profile.children.cycles-pp.do_mmap
> >       94.46           -92.6        1.83 ± 31%  perf-profile.children.cycles-pp.vm_mmap_pgoff
> 
> It significantly speed up mmap for this benchmark and the rmap lock
> contention is reduced significantly for both multi-processes and
> multi-threads.
> 
> The benchmark itself may exaggerate the improvement, but it may really speed
> up some real life workloads. For example, multiple applications which may
> allocate anonymous mapping by mmap'ing /dev/zero, then they may have
> contention on /dev/zero's rmap lock.
> 
> It doesn't make too much sense to link /dev/zero anonymous vmas to the file
> rmap tree. So the below patch should be able to speed up the benchmark too.

sorry for late and thanks a lot for information!

> 
> Oliver, can you please give this patch a try?

it seems this is an alternative patch?
since we applied your "/dev/zero: make private mapping full anonymous mapping"
patch upon a68d3cbfad like below:

* 7143ee2391f1e /dev/zero: make private mapping full anonymous mapping
* a68d3cbfade64 memstick: core: fix kernel-doc notation

so I applied below patch also upon a68d3cbfad.

we saw big improvement but not that big.

=========================================================================================
compiler/cpufreq_governor/kconfig/rootfs/runtime/tbox_group/test/testcase:
  gcc-12/performance/x86_64-rhel-9.4/debian-12-x86_64-20240206.cgz/300s/lkp-cpl-4sp2/small-allocs/vm-scalability

commit: 
  a68d3cbfad ("memstick: core: fix kernel-doc notation")
  52ec85cb99  <--- your patch


a68d3cbfade64392 52ec85cb99e9b31dc304eae965a
---------------- ---------------------------
         %stddev     %change         %stddev
             \          |                \
  14364828 ±  4%    +410.6%   73349239 ±  3%  vm-scalability.throughput

full comparison as below [1] just FYI.

> 
> 
> diff --git a/mm/vma.c b/mm/vma.c
> index bb2119e5a0d0..1092222c40ae 100644
> --- a/mm/vma.c
> +++ b/mm/vma.c
> @@ -1633,6 +1633,9 @@ static void unlink_file_vma_batch_process(struct
> unlink_vma_file_batch *vb)
>  void unlink_file_vma_batch_add(struct unlink_vma_file_batch *vb,
>                                struct vm_area_struct *vma)
>  {
> +       if (vma_is_anonymous(vma))
> +               return;
> +
>         if (vma->vm_file == NULL)
>                 return;
> 
> @@ -1658,6 +1661,9 @@ void unlink_file_vma(struct vm_area_struct *vma)
>  {
>         struct file *file = vma->vm_file;
> 
> +       if (vma_is_anonymous(vma))
> +               return;
> +
>         if (file) {
>                 struct address_space *mapping = file->f_mapping;
> 
> @@ -1672,6 +1678,9 @@ void vma_link_file(struct vm_area_struct *vma)
>         struct file *file = vma->vm_file;
>         struct address_space *mapping;
> 
> +       if (vma_is_anonymous(vma))
> +               return;
> +
>         if (file) {
>                 mapping = file->f_mapping;
>                 i_mmap_lock_write(mapping);
> 
> 
> Because /dev/zero's private mapping is an anonymous mapping with valid
> vm_file, so we need to bail out early if the vma is anonymous even though it
> has vm_file. IMHO, making /dev/zero private mapping a full anonymous mapping
> looks more clean.
> 

[1]
=========================================================================================
compiler/cpufreq_governor/kconfig/rootfs/runtime/tbox_group/test/testcase:
  gcc-12/performance/x86_64-rhel-9.4/debian-12-x86_64-20240206.cgz/300s/lkp-cpl-4sp2/small-allocs/vm-scalability

commit: 
  a68d3cbfad ("memstick: core: fix kernel-doc notation")
  52ec85cb99  <--- your patch


a68d3cbfade64392 52ec85cb99e9b31dc304eae965a
---------------- ---------------------------
         %stddev     %change         %stddev
             \          |                \
 5.262e+09 ±  3%     -45.0%  2.896e+09 ±  6%  cpuidle..time
   7924008 ±  3%     -79.3%    1643339 ± 11%  cpuidle..usage
   1871164 ±  4%     -22.4%    1452554 ± 12%  numa-numastat.node3.local_node
   1952164 ±  3%     -20.1%    1560294 ± 12%  numa-numastat.node3.numa_hit
    399.52           -68.2%     126.86        uptime.boot
     14507           -15.7%      12232        uptime.idle
      6.99 ±  3%    +147.9%      17.34 ±  4%  vmstat.cpu.id
      1.71          +473.6%       9.79 ±  2%  vmstat.cpu.us
     34204 ±  5%     -72.9%       9272 ±  7%  vmstat.system.cs
    266575           -21.2%     210191        vmstat.system.in
      3408 ±  5%     -99.8%       8.38 ± 48%  perf-c2c.DRAM.local
     18076 ±  3%     -99.8%      32.25 ± 27%  perf-c2c.DRAM.remote
      8082 ±  5%     -99.8%      15.50 ± 64%  perf-c2c.HITM.local
      6544 ±  6%     -99.8%      13.62 ± 51%  perf-c2c.HITM.remote
     14627 ±  4%     -99.8%      29.12 ± 53%  perf-c2c.HITM.total
      6.49 ±  3%      +8.8       15.24 ±  5%  mpstat.cpu.all.idle%
      0.63            -0.3        0.32 ±  4%  mpstat.cpu.all.irq%
      0.03 ±  2%      +0.2        0.26 ±  2%  mpstat.cpu.all.soft%
     91.17           -17.0       74.15        mpstat.cpu.all.sys%
      1.68 ±  2%      +8.3       10.03 ±  2%  mpstat.cpu.all.usr%
    337.33           -97.4%       8.88 ± 75%  mpstat.max_utilization.seconds
    352.76           -77.3%      79.95 ±  2%  time.elapsed_time
    352.76           -77.3%      79.95 ±  2%  time.elapsed_time.max
    225965 ±  7%     -16.0%     189844 ±  6%  time.involuntary_context_switches
 9.592e+08 ±  4%     +11.9%  1.074e+09        time.minor_page_faults
     20852            -8.8%      19012        time.percent_of_cpu_this_job_got
     72302           -81.4%      13425 ±  3%  time.system_time
      1260 ±  3%     +41.0%       1777        time.user_time
   5393707 ±  5%     -98.4%      86880 ± 17%  time.voluntary_context_switches
   1609925           -50.3%     800493        meminfo.Active
   1609925           -50.3%     800493        meminfo.Active(anon)
    160837 ± 33%     -63.9%      58119 ± 13%  meminfo.AnonHugePages
   4435665           -18.5%    3614714        meminfo.Cached
   1775547           -43.8%     998415        meminfo.Committed_AS
    148539           -43.7%      83699 ±  4%  meminfo.Mapped
   4245538 ±  4%     -20.9%    3356561        meminfo.PageTables
  14166291 ±  4%      -9.6%   12806082        meminfo.SUnreclaim
    929777           -88.2%     109274 ±  3%  meminfo.Shmem
  14315492 ±  4%      -9.6%   12947821        meminfo.Slab
     64129 ±  4%    +418.9%     332751 ±  3%  vm-scalability.median
     45.40 ±  5%   +1961.8        2007 ±  8%  vm-scalability.stddev%
  14364828 ±  4%    +410.6%   73349239 ±  3%  vm-scalability.throughput
    352.76           -77.3%      79.95 ±  2%  vm-scalability.time.elapsed_time
    352.76           -77.3%      79.95 ±  2%  vm-scalability.time.elapsed_time.max
    225965 ±  7%     -16.0%     189844 ±  6%  vm-scalability.time.involuntary_context_switches
 9.592e+08 ±  4%     +11.9%  1.074e+09        vm-scalability.time.minor_page_faults
     20852            -8.8%      19012        vm-scalability.time.percent_of_cpu_this_job_got
     72302           -81.4%      13425 ±  3%  vm-scalability.time.system_time
      1260 ±  3%     +41.0%       1777        vm-scalability.time.user_time
   5393707 ±  5%     -98.4%      86880 ± 17%  vm-scalability.time.voluntary_context_switches
 4.316e+09 ±  4%     +11.9%  4.832e+09        vm-scalability.workload
    265763 ±  4%     -20.5%     211398 ±  4%  numa-vmstat.node0.nr_page_table_pages
     31364 ±106%     -85.0%       4690 ±169%  numa-vmstat.node0.nr_shmem
     12205 ± 67%     -74.1%       3161 ±199%  numa-vmstat.node1.nr_mapped
    265546 ±  4%     -21.8%     207742 ±  4%  numa-vmstat.node1.nr_page_table_pages
     44052 ± 71%     -86.0%       6163 ±161%  numa-vmstat.node1.nr_shmem
    885590 ±  4%      -9.9%     797649 ±  4%  numa-vmstat.node1.nr_slab_unreclaimable
    264589 ±  4%     -21.2%     208598 ±  4%  numa-vmstat.node2.nr_page_table_pages
    881598 ±  4%     -10.0%     793829 ±  4%  numa-vmstat.node2.nr_slab_unreclaimable
    192683 ± 30%     -61.0%      75078 ± 70%  numa-vmstat.node3.nr_active_anon
    286819 ±108%     -93.0%      19993 ± 39%  numa-vmstat.node3.nr_file_pages
     13124 ± 49%     -92.3%       1006 ± 57%  numa-vmstat.node3.nr_mapped
    264499 ±  4%     -22.1%     206135 ±  2%  numa-vmstat.node3.nr_page_table_pages
    139810 ± 14%     -90.5%      13229 ± 89%  numa-vmstat.node3.nr_shmem
    880199 ±  4%     -11.8%     776210 ±  5%  numa-vmstat.node3.nr_slab_unreclaimable
    192683 ± 30%     -61.0%      75077 ± 70%  numa-vmstat.node3.nr_zone_active_anon
   1951359 ±  3%     -20.1%    1558936 ± 12%  numa-vmstat.node3.numa_hit
   1870359 ±  4%     -22.4%    1451195 ± 12%  numa-vmstat.node3.numa_local
    402515           -50.3%     200150        proc-vmstat.nr_active_anon
    170568            +1.9%     173746        proc-vmstat.nr_anon_pages
   4257257            +0.9%    4296664        proc-vmstat.nr_dirty_background_threshold
   8524925            +0.9%    8603835        proc-vmstat.nr_dirty_threshold
   1109246           -18.5%     903959        proc-vmstat.nr_file_pages
  42815276            +0.9%   43210344        proc-vmstat.nr_free_pages
     37525           -43.6%      21164 ±  4%  proc-vmstat.nr_mapped
   1059932 ±  4%     -21.1%     836810        proc-vmstat.nr_page_table_pages
    232507           -88.2%      27341 ±  3%  proc-vmstat.nr_shmem
     37297            -5.0%      35436        proc-vmstat.nr_slab_reclaimable
   3537843 ±  4%      -9.8%    3192506        proc-vmstat.nr_slab_unreclaimable
    402515           -50.3%     200150        proc-vmstat.nr_zone_active_anon
     61931 ±  8%     -83.8%      10023 ± 45%  proc-vmstat.numa_hint_faults
     15755 ± 21%     -87.1%       2039 ± 97%  proc-vmstat.numa_hint_faults_local
   6916516 ±  3%      -7.1%    6425430        proc-vmstat.numa_hit
   6568542 ±  3%      -7.5%    6077764        proc-vmstat.numa_local
    293942 ±  3%     -69.6%      89435 ± 49%  proc-vmstat.numa_pte_updates
 9.608e+08 ±  4%     +11.8%  1.074e+09        proc-vmstat.pgfault
     55981 ±  2%     -63.1%      20641 ±  2%  proc-vmstat.pgreuse
   1063552 ±  4%     -20.3%     847673 ±  4%  numa-meminfo.node0.PageTables
   3565610 ±  4%      -8.0%    3279375 ±  3%  numa-meminfo.node0.SUnreclaim
    125455 ±106%     -85.2%      18620 ±168%  numa-meminfo.node0.Shmem
   3592377 ±  4%      -7.1%    3336072 ±  4%  numa-meminfo.node0.Slab
     48482 ± 67%     -74.3%      12475 ±199%  numa-meminfo.node1.Mapped
   1062709 ±  4%     -21.7%     831966 ±  4%  numa-meminfo.node1.PageTables
   3543793 ±  4%     -10.0%    3189589 ±  4%  numa-meminfo.node1.SUnreclaim
    176171 ± 71%     -86.0%      24677 ±161%  numa-meminfo.node1.Shmem
   3593431 ±  4%     -10.4%    3220352 ±  4%  numa-meminfo.node1.Slab
   1058901 ±  4%     -21.3%     833124 ±  4%  numa-meminfo.node2.PageTables
   3527862 ±  4%     -10.2%    3168666 ±  5%  numa-meminfo.node2.SUnreclaim
   3565750 ±  4%     -10.3%    3200248 ±  5%  numa-meminfo.node2.Slab
    770405 ± 30%     -61.0%     300435 ± 70%  numa-meminfo.node3.Active
    770405 ± 30%     -61.0%     300435 ± 70%  numa-meminfo.node3.Active(anon)
   1146977 ±108%     -93.0%      80110 ± 40%  numa-meminfo.node3.FilePages
     52663 ± 47%     -91.6%       4397 ± 56%  numa-meminfo.node3.Mapped
   6368902 ± 20%     -21.2%    5021246 ±  2%  numa-meminfo.node3.MemUsed
   1058539 ±  4%     -22.2%     823061 ±  3%  numa-meminfo.node3.PageTables
   3522496 ±  4%     -12.1%    3096728 ±  6%  numa-meminfo.node3.SUnreclaim
    558943 ± 14%     -90.5%      53054 ± 89%  numa-meminfo.node3.Shmem
   3557392 ±  4%     -12.3%    3119454 ±  6%  numa-meminfo.node3.Slab
      0.82 ±  4%     -39.7%       0.50 ± 12%  perf-stat.i.MPKI
 2.714e+10 ±  2%    +185.7%  7.755e+10 ±  6%  perf-stat.i.branch-instructions
      0.11 ±  3%      +0.1        0.20 ±  5%  perf-stat.i.branch-miss-rate%
  24932893          +156.6%   63980942 ±  5%  perf-stat.i.branch-misses
     64.93           -10.1       54.87 ±  2%  perf-stat.i.cache-miss-rate%
     34508 ±  4%     -61.4%      13315 ± 10%  perf-stat.i.context-switches
      7.67           -63.7%       2.79 ±  6%  perf-stat.i.cpi
    224605           +10.8%     248972 ±  4%  perf-stat.i.cpu-clock
    696.35 ±  2%     -57.4%     296.79 ±  3%  perf-stat.i.cpu-migrations
 1.102e+11          +128.5%  2.518e+11 ±  6%  perf-stat.i.instructions
      0.14          +198.2%       0.42 ±  5%  perf-stat.i.ipc
     24.25 ±  3%    +375.8%     115.36 ±  3%  perf-stat.i.metric.K/sec
   2722043 ±  3%    +439.7%   14690226 ±  6%  perf-stat.i.minor-faults
   2722043 ±  3%    +439.7%   14690226 ±  6%  perf-stat.i.page-faults
    224605           +10.8%     248972 ±  4%  perf-stat.i.task-clock
      0.81 ±  3%     -52.5%       0.39 ± 14%  perf-stat.overall.MPKI
      0.09            -0.0        0.08 ±  2%  perf-stat.overall.branch-miss-rate%
     64.81            -6.4       58.40        perf-stat.overall.cache-miss-rate%
      7.24           -56.3%       3.17 ±  3%  perf-stat.overall.cpi
      0.14          +129.0%       0.32 ±  3%  perf-stat.overall.ipc
      9012 ±  2%     -57.5%       3827        perf-stat.overall.path-length
 2.701e+10 ±  2%    +159.6%  7.012e+10 ±  2%  perf-stat.ps.branch-instructions
  24708939          +119.2%   54173035        perf-stat.ps.branch-misses
     34266 ±  5%     -73.9%       8949 ±  7%  perf-stat.ps.context-switches
 7.941e+11            -9.1%  7.219e+11        perf-stat.ps.cpu-cycles
    693.54 ±  2%     -68.6%     217.73 ±  5%  perf-stat.ps.cpu-migrations
 1.097e+11          +108.1%  2.282e+11 ±  2%  perf-stat.ps.instructions
   2710577 ±  3%    +388.7%   13246535 ±  2%  perf-stat.ps.minor-faults
   2710577 ±  3%    +388.7%   13246536 ±  2%  perf-stat.ps.page-faults
 3.886e+13 ±  2%     -52.4%  1.849e+13        perf-stat.total.instructions
  64052898 ±  5%     -96.2%    2460331 ±166%  sched_debug.cfs_rq:/.avg_vruntime.avg
  95701822 ±  7%     -85.1%   14268127 ±116%  sched_debug.cfs_rq:/.avg_vruntime.max
  43098762 ±  6%     -96.0%    1715136 ±173%  sched_debug.cfs_rq:/.avg_vruntime.min
   9223270 ±  9%     -84.2%    1457904 ±122%  sched_debug.cfs_rq:/.avg_vruntime.stddev
      0.78 ±  2%     -77.0%       0.18 ±130%  sched_debug.cfs_rq:/.h_nr_running.avg
  43049468 ± 22%     -89.3%    4590302 ±180%  sched_debug.cfs_rq:/.left_deadline.max
   3836405 ± 37%     -85.6%     550773 ±176%  sched_debug.cfs_rq:/.left_deadline.stddev
  43049467 ± 22%     -89.3%    4590279 ±180%  sched_debug.cfs_rq:/.left_vruntime.max
   3836405 ± 37%     -85.6%     550772 ±176%  sched_debug.cfs_rq:/.left_vruntime.stddev
  64052901 ±  5%     -96.2%    2460341 ±166%  sched_debug.cfs_rq:/.min_vruntime.avg
  95701822 ±  7%     -85.1%   14268127 ±116%  sched_debug.cfs_rq:/.min_vruntime.max
  43098762 ±  6%     -96.0%    1715136 ±173%  sched_debug.cfs_rq:/.min_vruntime.min
   9223270 ±  9%     -84.2%    1457902 ±122%  sched_debug.cfs_rq:/.min_vruntime.stddev
      0.77 ±  2%     -77.4%       0.17 ±128%  sched_debug.cfs_rq:/.nr_running.avg
      1.61 ± 24%    +396.0%       7.96 ± 62%  sched_debug.cfs_rq:/.removed.runnable_avg.avg
     86.69          +424.4%     454.62 ± 24%  sched_debug.cfs_rq:/.removed.runnable_avg.max
     11.14 ± 13%    +409.8%      56.79 ± 35%  sched_debug.cfs_rq:/.removed.runnable_avg.stddev
      1.61 ± 24%    +396.0%       7.96 ± 62%  sched_debug.cfs_rq:/.removed.util_avg.avg
     86.69          +424.4%     454.62 ± 24%  sched_debug.cfs_rq:/.removed.util_avg.max
     11.14 ± 13%    +409.8%      56.79 ± 35%  sched_debug.cfs_rq:/.removed.util_avg.stddev
  43049467 ± 22%     -89.3%    4590282 ±180%  sched_debug.cfs_rq:/.right_vruntime.max
   3836405 ± 37%     -85.6%     550772 ±176%  sched_debug.cfs_rq:/.right_vruntime.stddev
    286633 ± 43%    +262.3%    1038592 ± 36%  sched_debug.cfs_rq:/.runnable_avg.avg
  34728895 ± 30%    +349.2%   1.56e+08 ± 26%  sched_debug.cfs_rq:/.runnable_avg.max
   2845573 ± 30%    +325.9%   12119045 ± 26%  sched_debug.cfs_rq:/.runnable_avg.stddev
    769.03           -69.9%     231.86 ± 84%  sched_debug.cfs_rq:/.util_avg.avg
      1621 ±  5%     -31.5%       1111 ±  8%  sched_debug.cfs_rq:/.util_avg.max
    724.17 ±  2%     -89.6%      75.66 ±147%  sched_debug.cfs_rq:/.util_est.avg
      1360 ± 15%     -39.2%     826.88 ± 37%  sched_debug.cfs_rq:/.util_est.max
    766944 ±  3%     +18.1%     905901        sched_debug.cpu.avg_idle.avg
    321459 ±  2%     -35.6%     207172 ± 10%  sched_debug.cpu.avg_idle.stddev
    195573           -72.7%      53401 ± 24%  sched_debug.cpu.clock.avg
    195596           -72.7%      53442 ± 24%  sched_debug.cpu.clock.max
    195548           -72.7%      53352 ± 24%  sched_debug.cpu.clock.min
    194424           -72.6%      53229 ± 24%  sched_debug.cpu.clock_task.avg
    194608           -72.6%      53383 ± 24%  sched_debug.cpu.clock_task.max
    181834           -77.5%      40964 ± 31%  sched_debug.cpu.clock_task.min
      4241 ±  2%     -80.6%     821.65 ±142%  sched_debug.cpu.curr->pid.avg
      9799 ±  2%     -55.4%       4365 ± 17%  sched_debug.cpu.curr->pid.max
      1365 ± 10%     -48.0%     709.44 ±  5%  sched_debug.cpu.curr->pid.stddev
    537665 ±  4%     +31.2%     705318 ± 14%  sched_debug.cpu.max_idle_balance_cost.max
      3119 ± 56%    +579.1%      21184 ± 39%  sched_debug.cpu.max_idle_balance_cost.stddev
      0.78 ±  2%     -76.3%       0.18 ±135%  sched_debug.cpu.nr_running.avg
     25773 ±  5%     -96.1%       1007 ± 41%  sched_debug.cpu.nr_switches.avg
     48669 ± 10%     -76.5%      11448 ± 13%  sched_debug.cpu.nr_switches.max
     19006 ±  7%     -98.6%     258.81 ± 64%  sched_debug.cpu.nr_switches.min
      4142 ±  8%     -66.3%       1396 ± 17%  sched_debug.cpu.nr_switches.stddev
      0.07 ± 23%     -92.9%       0.01 ± 41%  sched_debug.cpu.nr_uninterruptible.avg
    240.19 ± 16%     -82.1%      42.94 ± 41%  sched_debug.cpu.nr_uninterruptible.max
    -77.92           -88.1%      -9.25        sched_debug.cpu.nr_uninterruptible.min
     37.87 ±  5%     -85.8%       5.36 ± 13%  sched_debug.cpu.nr_uninterruptible.stddev
    195549           -72.7%      53356 ± 24%  sched_debug.cpu_clk
    194699           -73.0%      52506 ± 25%  sched_debug.ktime
      0.00          -100.0%       0.00        sched_debug.rt_rq:.rt_nr_running.avg
      0.17          -100.0%       0.00        sched_debug.rt_rq:.rt_nr_running.max
      0.01          -100.0%       0.00        sched_debug.rt_rq:.rt_nr_running.stddev
    196368           -72.4%      54191 ± 24%  sched_debug.sched_clk
      0.17 ±142%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.__cond_resched.__do_fault.do_read_fault.do_pte_missing.__handle_mm_fault
      0.19 ± 34%     -51.3%       0.09 ± 37%  perf-sched.sch_delay.avg.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region
      0.14 ± 55%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.__cond_resched.__wait_for_common.stop_two_cpus.migrate_swap.task_numa_migrate
      0.14 ± 73%     -82.5%       0.03 ±168%  perf-sched.sch_delay.avg.ms.__cond_resched.change_pud_range.isra.0.change_protection_range
      0.11 ± 59%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.__cond_resched.down_write.vma_link_file.__mmap_new_vma.__mmap_region
      0.04 ±132%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.__cond_resched.down_write_killable.exec_mmap.begin_new_exec.load_elf_binary
      0.02 ± 31%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64
      0.00 ±223%  +51950.0%       0.26 ±212%  perf-sched.sch_delay.avg.ms.__cond_resched.exit_mmap.__mmput.exit_mm.do_exit
      0.25 ± 59%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.__cond_resched.generic_perform_write.shmem_file_write_iter.vfs_write.ksys_write
      0.12 ±145%     -99.1%       0.00 ±141%  perf-sched.sch_delay.avg.ms.__cond_resched.mutex_lock.perf_poll.do_poll.constprop
      0.25 ± 41%     -81.6%       0.05 ± 69%  perf-sched.sch_delay.avg.ms.__cond_resched.stop_one_cpu.migrate_task_to.task_numa_migrate.isra
      0.11 ± 59%     -87.1%       0.01 ±198%  perf-sched.sch_delay.avg.ms.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part
      0.40 ± 50%     -97.8%       0.01 ± 30%  perf-sched.sch_delay.avg.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
      2.25 ±138%     -99.6%       0.01 ±  7%  perf-sched.sch_delay.avg.ms.devkmsg_read.vfs_read.ksys_read.do_syscall_64
      0.32 ±104%     -97.3%       0.01 ± 38%  perf-sched.sch_delay.avg.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
      0.01 ± 12%     -34.9%       0.01 ± 18%  perf-sched.sch_delay.avg.ms.irq_thread.kthread.ret_from_fork.ret_from_fork_asm
      0.01 ± 20%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown].[unknown]
      0.19 ±185%     -95.6%       0.01 ± 44%  perf-sched.sch_delay.avg.ms.schedule_hrtimeout_range.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
      0.07 ± 20%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file
      0.26 ± 17%     -98.8%       0.00 ± 10%  perf-sched.sch_delay.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
      0.03 ± 51%     -69.7%       0.01 ± 67%  perf-sched.sch_delay.avg.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
      0.01 ± 55%    +721.9%       0.10 ± 29%  perf-sched.sch_delay.avg.ms.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
      0.01 ±128%     -83.6%       0.00 ± 20%  perf-sched.sch_delay.avg.ms.wait_for_partner.fifo_open.do_dentry_open.vfs_open
      0.06 ± 31%   +1921.5%       1.23 ±165%  perf-sched.sch_delay.avg.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
      1.00 ±151%    -100.0%       0.00        perf-sched.sch_delay.max.ms.__cond_resched.__do_fault.do_read_fault.do_pte_missing.__handle_mm_fault
     25.45 ± 94%     -98.6%       0.36 ± 61%  perf-sched.sch_delay.max.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region
      4.56 ± 67%    -100.0%       0.00        perf-sched.sch_delay.max.ms.__cond_resched.__wait_for_common.stop_two_cpus.migrate_swap.task_numa_migrate
      3.55 ± 97%     -98.9%       0.04 ±189%  perf-sched.sch_delay.max.ms.__cond_resched.change_pud_range.isra.0.change_protection_range
      3.16 ± 78%    -100.0%       0.00        perf-sched.sch_delay.max.ms.__cond_resched.down_write.vma_link_file.__mmap_new_vma.__mmap_region
      0.30 ±159%    -100.0%       0.00        perf-sched.sch_delay.max.ms.__cond_resched.down_write_killable.exec_mmap.begin_new_exec.load_elf_binary
      0.03 ± 86%    -100.0%       0.00        perf-sched.sch_delay.max.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64
      0.00 ±223%  +3.2e+06%      15.79 ±259%  perf-sched.sch_delay.max.ms.__cond_resched.exit_mmap.__mmput.exit_mm.do_exit
      3.09 ± 45%    -100.0%       0.00        perf-sched.sch_delay.max.ms.__cond_resched.generic_perform_write.shmem_file_write_iter.vfs_write.ksys_write
      3.51 ± 21%     -86.1%       0.49 ± 72%  perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc_noprof.vm_area_alloc.__mmap_new_vma.__mmap_region
      3.59 ± 11%     -92.0%       0.29 ±165%  perf-sched.sch_delay.max.ms.__cond_resched.stop_one_cpu.migrate_task_to.task_numa_migrate.isra
      1.60 ± 69%     -95.7%       0.07 ±243%  perf-sched.sch_delay.max.ms.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part
      0.81 ± 43%     -98.5%       0.01 ± 43%  perf-sched.sch_delay.max.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
      1.02 ± 88%     -98.1%       0.02 ± 47%  perf-sched.sch_delay.max.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
      9.68 ± 32%     -92.2%       0.76 ± 72%  perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
     12.26 ±109%     -92.9%       0.87 ±101%  perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown]
      0.03 ±106%    -100.0%       0.00        perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown].[unknown]
     37.84 ± 47%    -100.0%       0.00        perf-sched.sch_delay.max.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file
      4.68 ± 36%     -99.8%       0.01 ± 65%  perf-sched.sch_delay.max.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
      0.36 ±186%     -96.3%       0.01 ± 90%  perf-sched.sch_delay.max.ms.wait_for_partner.fifo_open.do_dentry_open.vfs_open
     97903 ±  4%     -38.3%      60433 ± 29%  perf-sched.total_wait_and_delay.count.ms
      3.97 ±  6%    -100.0%       0.00        perf-sched.wait_and_delay.avg.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region
    302.41 ±  5%     -27.4%     219.54 ± 14%  perf-sched.wait_and_delay.avg.ms.irq_thread.kthread.ret_from_fork.ret_from_fork_asm
      1.48 ±  6%     -90.9%       0.14 ± 79%  perf-sched.wait_and_delay.avg.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
    327.16 ±  9%     -46.6%     174.81 ± 24%  perf-sched.wait_and_delay.avg.ms.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll
    369.37 ±  2%     -75.3%      91.05 ± 35%  perf-sched.wait_and_delay.avg.ms.schedule_hrtimeout_range.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
      0.96 ±  6%    -100.0%       0.00        perf-sched.wait_and_delay.avg.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file
    187.66          +120.6%     413.97 ± 14%  perf-sched.wait_and_delay.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      1831 ±  9%    -100.0%       0.00        perf-sched.wait_and_delay.count.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region
      6.17 ± 45%     -79.7%       1.25 ±142%  perf-sched.wait_and_delay.count.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
     40.50 ±  8%    +245.7%     140.00 ± 23%  perf-sched.wait_and_delay.count.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll
     13.17 ±  2%    +624.4%      95.38 ± 19%  perf-sched.wait_and_delay.count.schedule_hrtimeout_range.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
     73021 ±  3%    -100.0%       0.00        perf-sched.wait_and_delay.count.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file
     11323 ±  3%     -75.9%       2725 ± 28%  perf-sched.wait_and_delay.count.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      1887 ± 45%     -96.1%      73.88 ± 78%  perf-sched.wait_and_delay.count.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
      1238           -34.5%     811.25 ± 13%  perf-sched.wait_and_delay.count.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
     35.19 ± 57%    -100.0%       0.00        perf-sched.wait_and_delay.max.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region
     20.79 ± 19%     -95.9%       0.84 ± 93%  perf-sched.wait_and_delay.max.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
      1240 ± 20%     -14.4%       1062 ± 10%  perf-sched.wait_and_delay.max.ms.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll
    500.34           +31.2%     656.38 ± 39%  perf-sched.wait_and_delay.max.ms.schedule_hrtimeout_range.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
     58.83 ± 39%    -100.0%       0.00        perf-sched.wait_and_delay.max.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file
      1237 ± 34%    +151.7%       3114 ± 25%  perf-sched.wait_and_delay.max.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
     49.27 ±119%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.__alloc_pages_noprof.alloc_pages_mpol_noprof.folio_alloc_mpol_noprof.shmem_alloc_folio
     58.17 ±187%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.__do_fault.do_read_fault.do_pte_missing.__handle_mm_fault
      3.78 ±  5%     -97.6%       0.09 ± 37%  perf-sched.wait_time.avg.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region
      2.99 ±  4%     +15.4%       3.45 ± 10%  perf-sched.wait_time.avg.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
      3.92 ±  5%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.__wait_for_common.stop_two_cpus.migrate_swap.task_numa_migrate
      4.71 ±  8%     -99.5%       0.02 ±170%  perf-sched.wait_time.avg.ms.__cond_resched.change_pud_range.isra.0.change_protection_range
      1.67 ± 20%     -92.7%       0.12 ± 30%  perf-sched.wait_time.avg.ms.__cond_resched.down_write.__mmap_new_vma.__mmap_region.do_mmap
      2.10 ± 27%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.down_write.vma_link_file.__mmap_new_vma.__mmap_region
      0.01 ± 44%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.down_write_killable.exec_mmap.begin_new_exec.load_elf_binary
      1.67 ± 21%     -94.3%       0.10 ± 35%  perf-sched.wait_time.avg.ms.__cond_resched.down_write_killable.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64
      0.04 ±133%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64
      2.30 ± 14%     -95.5%       0.10 ± 42%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc_noprof.vm_area_alloc.__mmap_new_vma.__mmap_region
      2.00 ± 74%   +2917.4%      60.44 ± 33%  perf-sched.wait_time.avg.ms.devkmsg_read.vfs_read.ksys_read.do_syscall_64
     29.19 ±  5%     -38.5%      17.96 ± 28%  perf-sched.wait_time.avg.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.x64_sys_call
      0.37 ± 30%   +5524.5%      20.95 ± 30%  perf-sched.wait_time.avg.ms.do_wait.kernel_wait4.do_syscall_64.entry_SYSCALL_64_after_hwframe
    302.40 ±  5%     -27.4%     219.53 ± 14%  perf-sched.wait_time.avg.ms.irq_thread.kthread.ret_from_fork.ret_from_fork_asm
      1.40 ±  6%     -92.7%       0.10 ± 18%  perf-sched.wait_time.avg.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
      0.72 ±220%    -100.0%       0.00        perf-sched.wait_time.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown].[unknown]
    326.84 ±  9%     -46.6%     174.54 ± 24%  perf-sched.wait_time.avg.ms.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll
    369.18 ±  2%     -75.3%      91.04 ± 35%  perf-sched.wait_time.avg.ms.schedule_hrtimeout_range.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
      0.89 ±  6%    -100.0%       0.00        perf-sched.wait_time.avg.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file
    187.58          +120.6%     413.77 ± 14%  perf-sched.wait_time.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      2.36 ± 29%   +1759.6%      43.80 ± 33%  perf-sched.wait_time.avg.ms.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
      0.01 ±156%     -97.9%       0.00 ±264%  perf-sched.wait_time.avg.ms.wait_for_partner.fifo_open.do_dentry_open.vfs_open
    340.69 ±135%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.__alloc_pages_noprof.alloc_pages_mpol_noprof.folio_alloc_mpol_noprof.shmem_alloc_folio
    535.09 ±128%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.__do_fault.do_read_fault.do_pte_missing.__handle_mm_fault
     22.04 ± 32%     -98.4%       0.36 ± 61%  perf-sched.wait_time.max.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region
     13.57 ± 17%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.__wait_for_common.stop_two_cpus.migrate_swap.task_numa_migrate
     13.54 ± 10%     -99.7%       0.04 ±189%  perf-sched.wait_time.max.ms.__cond_resched.change_pud_range.isra.0.change_protection_range
     10.17 ± 19%     -95.2%       0.49 ± 56%  perf-sched.wait_time.max.ms.__cond_resched.down_write.__mmap_new_vma.__mmap_region.do_mmap
     11.35 ± 25%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.down_write.vma_link_file.__mmap_new_vma.__mmap_region
      0.01 ± 32%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.down_write_killable.exec_mmap.begin_new_exec.load_elf_binary
     10.62 ±  9%     -96.5%       0.38 ± 72%  perf-sched.wait_time.max.ms.__cond_resched.down_write_killable.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64
      0.20 ±199%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64
     14.42 ± 22%     -96.6%       0.49 ± 72%  perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc_noprof.vm_area_alloc.__mmap_new_vma.__mmap_region
      4.00 ± 74%  +19182.5%     772.23 ± 40%  perf-sched.wait_time.max.ms.devkmsg_read.vfs_read.ksys_read.do_syscall_64
     10.75 ± 98%   +6512.2%     710.88 ± 56%  perf-sched.wait_time.max.ms.do_wait.kernel_wait4.do_syscall_64.entry_SYSCALL_64_after_hwframe
     15.80 ±  8%     -95.2%       0.76 ± 72%  perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
     11.64 ± 61%     -98.9%       0.13 ±132%  perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown]
      2.94 ±213%    -100.0%       0.00        perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown].[unknown]
      1240 ± 20%     -14.3%       1062 ± 10%  perf-sched.wait_time.max.ms.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll
    500.11           +31.2%     656.37 ± 39%  perf-sched.wait_time.max.ms.schedule_hrtimeout_range.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
     32.65 ± 33%    -100.0%       0.00        perf-sched.wait_time.max.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file
      1237 ± 34%    +151.6%       3113 ± 25%  perf-sched.wait_time.max.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
     95.59           -95.6        0.00        perf-profile.calltrace.cycles-pp.__mmap
     95.54           -95.5        0.00        perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.__mmap
     95.54           -95.5        0.00        perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.__mmap
     94.54           -94.5        0.00        perf-profile.calltrace.cycles-pp.ksys_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe.__mmap
     94.46           -94.0        0.41 ±138%  perf-profile.calltrace.cycles-pp.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe.__mmap
     94.14           -93.7        0.40 ±136%  perf-profile.calltrace.cycles-pp.__mmap_new_vma.__mmap_region.do_mmap.vm_mmap_pgoff.ksys_mmap_pgoff
     93.79           -93.5        0.31 ±134%  perf-profile.calltrace.cycles-pp.vma_link_file.__mmap_new_vma.__mmap_region.do_mmap.vm_mmap_pgoff
     93.40           -93.4        0.00        perf-profile.calltrace.cycles-pp.rwsem_down_write_slowpath.down_write.vma_link_file.__mmap_new_vma.__mmap_region
     93.33           -93.3        0.00        perf-profile.calltrace.cycles-pp.rwsem_optimistic_spin.rwsem_down_write_slowpath.down_write.vma_link_file.__mmap_new_vma
     93.44           -93.3        0.14 ±264%  perf-profile.calltrace.cycles-pp.down_write.vma_link_file.__mmap_new_vma.__mmap_region.do_mmap
     94.45           -93.0        1.42 ± 60%  perf-profile.calltrace.cycles-pp.do_mmap.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe
     94.25           -92.9        1.33 ± 61%  perf-profile.calltrace.cycles-pp.__mmap_region.do_mmap.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64
     92.89           -92.9        0.00        perf-profile.calltrace.cycles-pp.osq_lock.rwsem_optimistic_spin.rwsem_down_write_slowpath.down_write.vma_link_file
      0.00            +1.1        1.09 ± 33%  perf-profile.calltrace.cycles-pp.dup_mmap.dup_mm.copy_process.kernel_clone.__do_sys_clone
      0.00            +1.4        1.37 ± 49%  perf-profile.calltrace.cycles-pp.setlocale
      0.00            +1.6        1.64 ± 47%  perf-profile.calltrace.cycles-pp.__irq_exit_rcu.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.acpi_safe_halt.acpi_idle_do_entry
      0.00            +1.6        1.64 ± 47%  perf-profile.calltrace.cycles-pp.handle_softirqs.__irq_exit_rcu.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.acpi_safe_halt
      0.00            +1.6        1.65 ± 43%  perf-profile.calltrace.cycles-pp.do_sys_openat2.__x64_sys_openat.do_syscall_64.entry_SYSCALL_64_after_hwframe
      0.00            +1.8        1.76 ± 44%  perf-profile.calltrace.cycles-pp.__x64_sys_openat.do_syscall_64.entry_SYSCALL_64_after_hwframe
      0.00            +1.9        1.93 ± 26%  perf-profile.calltrace.cycles-pp.dup_mm.copy_process.kernel_clone.__do_sys_clone.do_syscall_64
      0.00            +2.2        2.16 ± 44%  perf-profile.calltrace.cycles-pp.do_pte_missing.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault
      0.00            +2.2        2.23 ± 33%  perf-profile.calltrace.cycles-pp.copy_process.kernel_clone.__do_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe
      0.00            +2.4        2.37 ± 36%  perf-profile.calltrace.cycles-pp.zap_present_ptes.zap_pte_range.zap_pmd_range.unmap_page_range.unmap_vmas
      0.00            +2.5        2.48 ± 32%  perf-profile.calltrace.cycles-pp.get_cpu_sleep_time_us.get_idle_time.uptime_proc_show.seq_read_iter.vfs_read
      0.00            +2.5        2.50 ± 45%  perf-profile.calltrace.cycles-pp.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
      0.00            +2.5        2.54 ± 47%  perf-profile.calltrace.cycles-pp.__mmput.exit_mm.do_exit.do_group_exit.__x64_sys_exit_group
      0.00            +2.5        2.54 ± 47%  perf-profile.calltrace.cycles-pp.exit_mm.do_exit.do_group_exit.__x64_sys_exit_group.x64_sys_call
      0.00            +2.6        2.62 ± 35%  perf-profile.calltrace.cycles-pp.__do_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe._Fork
      0.00            +2.6        2.62 ± 35%  perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe._Fork
      0.00            +2.6        2.62 ± 35%  perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe._Fork
      0.00            +2.6        2.62 ± 35%  perf-profile.calltrace.cycles-pp.kernel_clone.__do_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe._Fork
      0.00            +2.7        2.68 ± 35%  perf-profile.calltrace.cycles-pp.get_idle_time.uptime_proc_show.seq_read_iter.vfs_read.ksys_read
      0.00            +2.8        2.77 ± 33%  perf-profile.calltrace.cycles-pp.uptime_proc_show.seq_read_iter.vfs_read.ksys_read.do_syscall_64
      0.00            +2.8        2.82 ± 32%  perf-profile.calltrace.cycles-pp._Fork
      0.00            +2.8        2.84 ± 45%  perf-profile.calltrace.cycles-pp.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
      0.00            +2.8        2.84 ± 45%  perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault
      0.00            +2.9        2.89 ± 39%  perf-profile.calltrace.cycles-pp.event_function_call.perf_event_release_kernel.perf_release.__fput.task_work_run
      0.00            +2.9        2.89 ± 39%  perf-profile.calltrace.cycles-pp.smp_call_function_single.event_function_call.perf_event_release_kernel.perf_release.__fput
      0.00            +3.1        3.10 ± 64%  perf-profile.calltrace.cycles-pp.proc_reg_read_iter.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe
      0.00            +3.1        3.10 ± 64%  perf-profile.calltrace.cycles-pp.seq_read_iter.proc_reg_read_iter.vfs_read.ksys_read.do_syscall_64
      0.00            +3.1        3.13 ± 33%  perf-profile.calltrace.cycles-pp.asm_exc_page_fault
      0.00            +3.2        3.18 ± 37%  perf-profile.calltrace.cycles-pp.seq_read_iter.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe
      0.00            +3.2        3.20 ± 28%  perf-profile.calltrace.cycles-pp.mutex_unlock.sw_perf_event_destroy._free_event.perf_event_release_kernel.perf_release
      0.00            +3.2        3.24 ± 39%  perf-profile.calltrace.cycles-pp.bprm_execve.do_execveat_common.__x64_sys_execve.do_syscall_64.entry_SYSCALL_64_after_hwframe
      0.00            +3.2        3.24 ± 36%  perf-profile.calltrace.cycles-pp.__x64_sys_exit_group.x64_sys_call.do_syscall_64.entry_SYSCALL_64_after_hwframe
      0.00            +3.2        3.24 ± 36%  perf-profile.calltrace.cycles-pp.do_exit.do_group_exit.__x64_sys_exit_group.x64_sys_call.do_syscall_64
      0.00            +3.2        3.24 ± 36%  perf-profile.calltrace.cycles-pp.do_group_exit.__x64_sys_exit_group.x64_sys_call.do_syscall_64.entry_SYSCALL_64_after_hwframe
      0.00            +3.2        3.24 ± 36%  perf-profile.calltrace.cycles-pp.x64_sys_call.do_syscall_64.entry_SYSCALL_64_after_hwframe
      0.00            +3.8        3.85 ± 39%  perf-profile.calltrace.cycles-pp.__x64_sys_execve.do_syscall_64.entry_SYSCALL_64_after_hwframe.execve
      0.00            +3.8        3.85 ± 39%  perf-profile.calltrace.cycles-pp.do_execveat_common.__x64_sys_execve.do_syscall_64.entry_SYSCALL_64_after_hwframe.execve
      0.00            +3.8        3.85 ± 39%  perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.execve
      0.00            +3.8        3.85 ± 39%  perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.execve
      0.00            +3.8        3.85 ± 39%  perf-profile.calltrace.cycles-pp.execve
      0.00            +4.0        4.04 ± 43%  perf-profile.calltrace.cycles-pp.do_filp_open.do_sys_openat2.__x64_sys_openat.do_syscall_64.entry_SYSCALL_64_after_hwframe
      0.00            +4.0        4.04 ± 43%  perf-profile.calltrace.cycles-pp.path_openat.do_filp_open.do_sys_openat2.__x64_sys_openat.do_syscall_64
      0.00            +4.1        4.10 ± 30%  perf-profile.calltrace.cycles-pp.unmap_page_range.unmap_vmas.exit_mmap.__mmput.exit_mm
      0.00            +4.2        4.18 ± 31%  perf-profile.calltrace.cycles-pp.zap_pmd_range.unmap_page_range.unmap_vmas.exit_mmap.__mmput
      0.00            +4.2        4.18 ± 31%  perf-profile.calltrace.cycles-pp.zap_pte_range.zap_pmd_range.unmap_page_range.unmap_vmas.exit_mmap
      0.00            +4.2        4.20 ± 28%  perf-profile.calltrace.cycles-pp.unmap_vmas.exit_mmap.__mmput.exit_mm.do_exit
      0.00            +4.2        4.25 ± 65%  perf-profile.calltrace.cycles-pp.generic_perform_write.shmem_file_write_iter.vfs_write.ksys_write.do_syscall_64
      0.00            +4.3        4.27 ± 26%  perf-profile.calltrace.cycles-pp.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
      0.00            +4.3        4.30 ± 22%  perf-profile.calltrace.cycles-pp.__mmput.exit_mm.do_exit.do_group_exit.get_signal
      0.00            +4.3        4.30 ± 22%  perf-profile.calltrace.cycles-pp.exit_mm.do_exit.do_group_exit.get_signal.arch_do_signal_or_restart
      0.00            +4.5        4.46 ± 59%  perf-profile.calltrace.cycles-pp.shmem_file_write_iter.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe
      0.00            +4.6        4.57 ± 58%  perf-profile.calltrace.cycles-pp.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe.write.writen
      0.00            +4.7        4.68 ± 55%  perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.write.writen.record__pushfn
      0.00            +4.7        4.68 ± 55%  perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.write.writen.record__pushfn.perf_mmap__push
      0.00            +4.7        4.68 ± 55%  perf-profile.calltrace.cycles-pp.record__pushfn.perf_mmap__push.record__mmap_read_evlist.__cmd_record.cmd_record
      0.00            +4.7        4.68 ± 55%  perf-profile.calltrace.cycles-pp.write.writen.record__pushfn.perf_mmap__push.record__mmap_read_evlist
      0.00            +4.7        4.68 ± 55%  perf-profile.calltrace.cycles-pp.writen.record__pushfn.perf_mmap__push.record__mmap_read_evlist.__cmd_record
      0.00            +4.9        4.90 ± 57%  perf-profile.calltrace.cycles-pp.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe.write
      0.00            +4.9        4.92 ± 26%  perf-profile.calltrace.cycles-pp.sw_perf_event_destroy._free_event.perf_event_release_kernel.perf_release.__fput
      0.00            +5.0        4.99 ±100%  perf-profile.calltrace.cycles-pp.__intel_pmu_enable_all.perf_rotate_context.perf_mux_hrtimer_handler.__hrtimer_run_queues.hrtimer_interrupt
      0.00            +5.0        4.99 ±100%  perf-profile.calltrace.cycles-pp.perf_rotate_context.perf_mux_hrtimer_handler.__hrtimer_run_queues.hrtimer_interrupt.__sysvec_apic_timer_interrupt
      0.00            +5.1        5.08 ±102%  perf-profile.calltrace.cycles-pp.perf_mux_hrtimer_handler.__hrtimer_run_queues.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt
      0.00            +5.1        5.14 ± 28%  perf-profile.calltrace.cycles-pp.perf_mmap__push.record__mmap_read_evlist.__cmd_record.cmd_record.run_builtin
      0.00            +5.1        5.14 ± 28%  perf-profile.calltrace.cycles-pp.record__mmap_read_evlist.__cmd_record.cmd_record.run_builtin.handle_internal_command
      0.00            +5.4        5.43 ± 25%  perf-profile.calltrace.cycles-pp._free_event.perf_event_release_kernel.perf_release.__fput.task_work_run
      0.00            +5.8        5.82 ± 94%  perf-profile.calltrace.cycles-pp.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.acpi_safe_halt.acpi_idle_do_entry
      0.00            +5.8        5.82 ± 94%  perf-profile.calltrace.cycles-pp.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.acpi_safe_halt
      0.00            +6.1        6.07 ± 90%  perf-profile.calltrace.cycles-pp.__hrtimer_run_queues.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt
      0.00            +6.6        6.62 ± 24%  perf-profile.calltrace.cycles-pp.__cmd_record.cmd_record.run_builtin.handle_internal_command.main
      0.00            +6.6        6.62 ± 24%  perf-profile.calltrace.cycles-pp.cmd_record.run_builtin.handle_internal_command.main
      0.00            +6.8        6.76 ± 18%  perf-profile.calltrace.cycles-pp.exit_mmap.__mmput.exit_mm.do_exit.do_group_exit
      0.00            +7.6        7.56 ± 76%  perf-profile.calltrace.cycles-pp.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.acpi_safe_halt.acpi_idle_do_entry.acpi_idle_enter
      0.00            +8.0        8.03 ± 27%  perf-profile.calltrace.cycles-pp.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe.read
      0.00            +8.0        8.03 ± 27%  perf-profile.calltrace.cycles-pp.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe.read
      0.00            +8.0        8.05 ± 68%  perf-profile.calltrace.cycles-pp.acpi_safe_halt.acpi_idle_do_entry.acpi_idle_enter.cpuidle_enter_state.cpuidle_enter
      0.00            +8.1        8.13 ± 28%  perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.read
      0.00            +8.1        8.13 ± 28%  perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.read
      0.00            +8.1        8.13 ± 28%  perf-profile.calltrace.cycles-pp.read
      0.00            +9.1        9.05 ± 35%  perf-profile.calltrace.cycles-pp.handle_internal_command.main
      0.00            +9.1        9.05 ± 35%  perf-profile.calltrace.cycles-pp.main
      0.00            +9.1        9.05 ± 35%  perf-profile.calltrace.cycles-pp.run_builtin.handle_internal_command.main
      0.00            +9.3        9.26 ± 30%  perf-profile.calltrace.cycles-pp.perf_event_release_kernel.perf_release.__fput.task_work_run.do_exit
      0.00            +9.3        9.26 ± 30%  perf-profile.calltrace.cycles-pp.perf_release.__fput.task_work_run.do_exit.do_group_exit
      0.00           +10.1       10.14 ± 28%  perf-profile.calltrace.cycles-pp.__fput.task_work_run.do_exit.do_group_exit.get_signal
      0.00           +10.2       10.23 ± 27%  perf-profile.calltrace.cycles-pp.task_work_run.do_exit.do_group_exit.get_signal.arch_do_signal_or_restart
      0.00           +11.0       10.98 ± 55%  perf-profile.calltrace.cycles-pp.asm_sysvec_reschedule_ipi.acpi_safe_halt.acpi_idle_do_entry.acpi_idle_enter.cpuidle_enter_state
      0.00           +20.6       20.64 ± 30%  perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe
      0.00           +20.6       20.64 ± 30%  perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe
      1.21 ±  3%     +36.6       37.80 ± 12%  perf-profile.calltrace.cycles-pp.cpuidle_enter.cpuidle_idle_call.do_idle.cpu_startup_entry.start_secondary
      1.21 ±  3%     +36.6       37.80 ± 12%  perf-profile.calltrace.cycles-pp.cpuidle_idle_call.do_idle.cpu_startup_entry.start_secondary.common_startup_64
      1.22 ±  3%     +36.8       38.00 ± 13%  perf-profile.calltrace.cycles-pp.do_idle.cpu_startup_entry.start_secondary.common_startup_64
      1.22 ±  3%     +36.9       38.10 ± 13%  perf-profile.calltrace.cycles-pp.cpu_startup_entry.start_secondary.common_startup_64
      1.22 ±  3%     +36.9       38.10 ± 13%  perf-profile.calltrace.cycles-pp.start_secondary.common_startup_64
      1.21 ±  3%     +37.2       38.43 ± 11%  perf-profile.calltrace.cycles-pp.acpi_idle_do_entry.acpi_idle_enter.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call
      1.21 ±  3%     +37.2       38.43 ± 11%  perf-profile.calltrace.cycles-pp.acpi_idle_enter.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle
      1.21 ±  3%     +37.3       38.54 ± 12%  perf-profile.calltrace.cycles-pp.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle.cpu_startup_entry
      1.22 ±  3%     +37.6       38.84 ± 12%  perf-profile.calltrace.cycles-pp.common_startup_64
      2.19 ±  3%     +53.9       56.10 ± 19%  perf-profile.calltrace.cycles-pp.asm_sysvec_apic_timer_interrupt.acpi_safe_halt.acpi_idle_do_entry.acpi_idle_enter.cpuidle_enter_state
     95.60           -95.2        0.41 ±138%  perf-profile.children.cycles-pp.__mmap
     94.14           -93.7        0.49 ±130%  perf-profile.children.cycles-pp.__mmap_new_vma
     93.79           -93.5        0.31 ±134%  perf-profile.children.cycles-pp.vma_link_file
     93.40           -93.4        0.00        perf-profile.children.cycles-pp.rwsem_down_write_slowpath
     93.33           -93.3        0.00        perf-profile.children.cycles-pp.rwsem_optimistic_spin
     94.55           -93.1        1.42 ± 60%  perf-profile.children.cycles-pp.ksys_mmap_pgoff
     92.91           -92.9        0.00        perf-profile.children.cycles-pp.osq_lock
     93.44           -92.7        0.75 ±109%  perf-profile.children.cycles-pp.down_write
     94.46           -92.6        1.84 ± 34%  perf-profile.children.cycles-pp.vm_mmap_pgoff
     94.45           -92.6        1.84 ± 34%  perf-profile.children.cycles-pp.do_mmap
     94.25           -92.6        1.66 ± 37%  perf-profile.children.cycles-pp.__mmap_region
     95.58           -44.8       50.78 ± 11%  perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe
     95.58           -44.8       50.78 ± 11%  perf-profile.children.cycles-pp.do_syscall_64
      0.00            +1.1        1.09 ± 33%  perf-profile.children.cycles-pp.dup_mmap
      0.00            +1.4        1.37 ± 49%  perf-profile.children.cycles-pp.setlocale
      0.00            +1.9        1.93 ± 26%  perf-profile.children.cycles-pp.dup_mm
      0.03 ± 70%      +2.0        1.99 ± 36%  perf-profile.children.cycles-pp.handle_softirqs
      0.00            +2.0        1.99 ± 36%  perf-profile.children.cycles-pp.__irq_exit_rcu
      0.00            +2.0        2.02 ± 38%  perf-profile.children.cycles-pp.folios_put_refs
      0.00            +2.1        2.06 ± 52%  perf-profile.children.cycles-pp._raw_spin_lock
      0.00            +2.2        2.16 ± 44%  perf-profile.children.cycles-pp.do_pte_missing
      0.00            +2.2        2.21 ± 68%  perf-profile.children.cycles-pp.link_path_walk
      0.00            +2.2        2.23 ± 33%  perf-profile.children.cycles-pp.copy_process
      0.00            +2.3        2.30 ± 40%  perf-profile.children.cycles-pp.__tlb_batch_free_encoded_pages
      0.00            +2.3        2.30 ± 40%  perf-profile.children.cycles-pp.free_pages_and_swap_cache
      0.00            +2.3        2.34 ± 46%  perf-profile.children.cycles-pp.walk_component
      0.00            +2.4        2.37 ± 36%  perf-profile.children.cycles-pp.zap_present_ptes
      0.00            +2.5        2.48 ± 32%  perf-profile.children.cycles-pp.get_cpu_sleep_time_us
      0.00            +2.6        2.62 ± 35%  perf-profile.children.cycles-pp.__do_sys_clone
      0.00            +2.6        2.62 ± 35%  perf-profile.children.cycles-pp.kernel_clone
      0.00            +2.7        2.68 ± 35%  perf-profile.children.cycles-pp.get_idle_time
      0.00            +2.8        2.77 ± 33%  perf-profile.children.cycles-pp.uptime_proc_show
      0.00            +2.9        2.91 ± 32%  perf-profile.children.cycles-pp._Fork
      0.00            +3.1        3.10 ± 64%  perf-profile.children.cycles-pp.proc_reg_read_iter
      0.00            +3.2        3.24 ± 39%  perf-profile.children.cycles-pp.bprm_execve
      0.00            +3.2        3.24 ± 36%  perf-profile.children.cycles-pp.__x64_sys_exit_group
      0.00            +3.2        3.24 ± 36%  perf-profile.children.cycles-pp.x64_sys_call
      0.00            +3.8        3.85 ± 39%  perf-profile.children.cycles-pp.__x64_sys_execve
      0.00            +3.8        3.85 ± 39%  perf-profile.children.cycles-pp.do_execveat_common
      0.00            +3.8        3.85 ± 39%  perf-profile.children.cycles-pp.execve
      0.00            +4.0        3.99 ± 38%  perf-profile.children.cycles-pp.mutex_unlock
      0.00            +4.2        4.19 ± 31%  perf-profile.children.cycles-pp.zap_pte_range
      0.00            +4.2        4.25 ± 65%  perf-profile.children.cycles-pp.generic_perform_write
      0.00            +4.3        4.29 ± 29%  perf-profile.children.cycles-pp.unmap_page_range
      0.00            +4.3        4.29 ± 29%  perf-profile.children.cycles-pp.zap_pmd_range
      0.00            +4.3        4.31 ± 51%  perf-profile.children.cycles-pp.do_filp_open
      0.00            +4.3        4.31 ± 51%  perf-profile.children.cycles-pp.path_openat
      0.19 ± 23%      +4.4        4.60 ± 26%  perf-profile.children.cycles-pp.__handle_mm_fault
      0.00            +4.5        4.46 ± 59%  perf-profile.children.cycles-pp.shmem_file_write_iter
      0.00            +4.5        4.55 ± 24%  perf-profile.children.cycles-pp.event_function_call
      0.00            +4.5        4.55 ± 24%  perf-profile.children.cycles-pp.smp_call_function_single
      0.00            +4.6        4.58 ± 30%  perf-profile.children.cycles-pp.unmap_vmas
      0.51 ±  6%      +4.6        5.14 ± 24%  perf-profile.children.cycles-pp.handle_mm_fault
      0.00            +4.7        4.68 ± 55%  perf-profile.children.cycles-pp.record__pushfn
      0.00            +4.7        4.68 ± 55%  perf-profile.children.cycles-pp.writen
      0.00            +4.8        4.80 ± 48%  perf-profile.children.cycles-pp.do_sys_openat2
      0.77 ±  3%      +4.8        5.59 ± 21%  perf-profile.children.cycles-pp.exc_page_fault
      0.76 ±  3%      +4.8        5.59 ± 21%  perf-profile.children.cycles-pp.do_user_addr_fault
      0.00            +4.9        4.90 ± 57%  perf-profile.children.cycles-pp.ksys_write
      0.00            +4.9        4.90 ± 57%  perf-profile.children.cycles-pp.vfs_write
      0.00            +4.9        4.90 ± 48%  perf-profile.children.cycles-pp.__x64_sys_openat
      0.00            +4.9        4.92 ± 26%  perf-profile.children.cycles-pp.sw_perf_event_destroy
      0.00            +5.0        4.99 ±100%  perf-profile.children.cycles-pp.perf_rotate_context
      0.00            +5.0        5.01 ± 54%  perf-profile.children.cycles-pp.write
      0.00            +5.1        5.09 ±102%  perf-profile.children.cycles-pp.perf_mux_hrtimer_handler
      0.00            +5.4        5.43 ± 25%  perf-profile.children.cycles-pp._free_event
      1.18            +5.6        6.78 ± 20%  perf-profile.children.cycles-pp.asm_exc_page_fault
      0.46            +5.6        6.07 ± 90%  perf-profile.children.cycles-pp.__hrtimer_run_queues
      0.00            +5.7        5.75 ± 39%  perf-profile.children.cycles-pp.perf_mmap__push
      0.00            +5.7        5.75 ± 39%  perf-profile.children.cycles-pp.record__mmap_read_evlist
      0.53            +5.8        6.28 ± 89%  perf-profile.children.cycles-pp.__sysvec_apic_timer_interrupt
      0.53            +5.8        6.28 ± 89%  perf-profile.children.cycles-pp.hrtimer_interrupt
      0.00            +6.6        6.65 ± 77%  perf-profile.children.cycles-pp.__intel_pmu_enable_all
      0.00            +6.8        6.85 ± 20%  perf-profile.children.cycles-pp.exit_mm
      0.58 ±  2%      +7.6        8.14 ± 75%  perf-profile.children.cycles-pp.sysvec_apic_timer_interrupt
      0.00            +7.7        7.67 ± 23%  perf-profile.children.cycles-pp.exit_mmap
      0.00            +7.7        7.67 ± 30%  perf-profile.children.cycles-pp.seq_read_iter
      0.00            +7.7        7.72 ± 80%  perf-profile.children.cycles-pp.asm_sysvec_reschedule_ipi
      0.00            +7.8        7.75 ± 23%  perf-profile.children.cycles-pp.__mmput
      0.00            +8.0        8.03 ± 27%  perf-profile.children.cycles-pp.ksys_read
      0.00            +8.0        8.03 ± 27%  perf-profile.children.cycles-pp.vfs_read
      0.00            +8.1        8.13 ± 28%  perf-profile.children.cycles-pp.read
      0.02 ±141%      +9.0        9.05 ± 35%  perf-profile.children.cycles-pp.__cmd_record
      0.02 ±141%      +9.0        9.05 ± 35%  perf-profile.children.cycles-pp.cmd_record
      0.02 ±141%      +9.0        9.05 ± 35%  perf-profile.children.cycles-pp.handle_internal_command
      0.02 ±141%      +9.0        9.05 ± 35%  perf-profile.children.cycles-pp.main
      0.02 ±141%      +9.0        9.05 ± 35%  perf-profile.children.cycles-pp.run_builtin
      0.00            +9.3        9.26 ± 30%  perf-profile.children.cycles-pp.perf_event_release_kernel
      0.00            +9.3        9.26 ± 30%  perf-profile.children.cycles-pp.perf_release
      1.02 ±  4%      +9.3       10.33 ± 27%  perf-profile.children.cycles-pp.task_work_run
      0.00           +11.0       11.05 ± 28%  perf-profile.children.cycles-pp.__fput
      0.00           +15.8       15.85 ± 25%  perf-profile.children.cycles-pp.arch_do_signal_or_restart
      0.00           +15.8       15.85 ± 25%  perf-profile.children.cycles-pp.get_signal
      0.00           +19.1       19.09 ± 19%  perf-profile.children.cycles-pp.do_exit
      0.00           +19.1       19.09 ± 19%  perf-profile.children.cycles-pp.do_group_exit
      1.70 ±  2%     +30.7       32.41 ± 21%  perf-profile.children.cycles-pp.asm_sysvec_apic_timer_interrupt
      1.22 ±  3%     +36.9       38.10 ± 13%  perf-profile.children.cycles-pp.start_secondary
      1.21 ±  3%     +37.2       38.43 ± 11%  perf-profile.children.cycles-pp.acpi_idle_do_entry
      1.21 ±  3%     +37.2       38.43 ± 11%  perf-profile.children.cycles-pp.acpi_idle_enter
      1.21 ±  3%     +37.2       38.43 ± 11%  perf-profile.children.cycles-pp.acpi_safe_halt
      1.22 ±  3%     +37.3       38.54 ± 12%  perf-profile.children.cycles-pp.cpuidle_idle_call
      1.21 ±  3%     +37.3       38.54 ± 12%  perf-profile.children.cycles-pp.cpuidle_enter
      1.21 ±  3%     +37.3       38.54 ± 12%  perf-profile.children.cycles-pp.cpuidle_enter_state
      1.22 ±  3%     +37.6       38.84 ± 12%  perf-profile.children.cycles-pp.common_startup_64
      1.22 ±  3%     +37.6       38.84 ± 12%  perf-profile.children.cycles-pp.cpu_startup_entry
      1.22 ±  3%     +37.6       38.84 ± 12%  perf-profile.children.cycles-pp.do_idle
     92.37           -92.4        0.00        perf-profile.self.cycles-pp.osq_lock
      0.00            +2.1        2.06 ± 52%  perf-profile.self.cycles-pp._raw_spin_lock
      0.00            +2.6        2.61 ± 36%  perf-profile.self.cycles-pp.smp_call_function_single
      0.00            +3.7        3.68 ± 37%  perf-profile.self.cycles-pp.mutex_unlock
      0.00            +6.6        6.65 ± 77%  perf-profile.self.cycles-pp.__intel_pmu_enable_all
      1.19 ±  3%     +29.2       30.38 ± 15%  perf-profile.self.cycles-pp.acpi_safe_halt






[Index of Archives]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux