Hello, kernel test robot noticed a 160.4% improvement of will-it-scale.per_process_ops on: commit: e38e699a42b4db5daf7dac453759fdc8ba0dab31 ("evm: Move metadata in the inode security blob to a pointer") https://github.com/robertosassu/linux evm-iint-ptr-v1-devel-v3 testcase: will-it-scale config: x86_64-rhel-9.4 compiler: gcc-12 test machine: 224 threads 2 sockets Intel(R) Xeon(R) Platinum 8480CTDX (Sapphire Rapids) with 512G memory parameters: nr_task: 100% mode: process test: open3 cpufreq_governor: performance Details are as below: --------------------------------------------------------------------------------------------------> The kernel config and materials to reproduce are available at: https://download.01.org/0day-ci/archive/20250217/202502171412.ec2e5b88-lkp@xxxxxxxxx ========================================================================================= compiler/cpufreq_governor/kconfig/mode/nr_task/rootfs/tbox_group/test/testcase: gcc-12/performance/x86_64-rhel-9.4/process/100%/debian-12-x86_64-20240206.cgz/lkp-spr-2sp4/open3/will-it-scale commit: d7a782797d ("ima: Reset IMA_NONACTION_RULE_FLAGS after post_setattr") e38e699a42 ("evm: Move metadata in the inode security blob to a pointer") d7a782797df7c64e e38e699a42b4db5daf7dac45375 ---------------- --------------------------- %stddev %change %stddev \ | \ 194348 ± 2% +42.5% 276904 ± 9% cpuidle..usage 2032 +69.8% 3451 ± 8% vmstat.system.cs 219050 ± 22% -53.4% 101970 ± 46% meminfo.AnonHugePages 224488 ± 6% -14.1% 192753 ± 6% meminfo.Mapped 1151621 -10.9% 1026208 meminfo.Shmem 0.62 ± 3% +0.1 0.70 ± 2% mpstat.cpu.all.idle% 0.16 ± 4% +0.0 0.20 ± 2% mpstat.cpu.all.usr% 3.00 +1650.0% 52.50 mpstat.max_utilization.seconds 1221925 ± 7% -16.7% 1018311 ± 12% numa-meminfo.node1.FilePages 123310 ± 9% -25.6% 91721 ± 12% numa-meminfo.node1.Mapped 1085554 -18.7% 882893 ± 12% numa-meminfo.node1.Shmem 305498 ± 7% -16.7% 254488 ± 12% numa-vmstat.node1.nr_file_pages 31041 ± 9% -25.1% 23250 ± 12% numa-vmstat.node1.nr_mapped 271405 -18.7% 220634 ± 12% numa-vmstat.node1.nr_shmem 620152 ± 2% +160.4% 1614732 ± 2% will-it-scale.224.processes 0.13 +46.2% 0.19 ± 4% will-it-scale.224.processes_idle 2768 ± 2% +160.4% 7208 ± 2% will-it-scale.per_process_ops 620152 ± 2% +160.4% 1614732 ± 2% will-it-scale.workload 32.33 ± 13% +1535.6% 528.83 ± 12% perf-c2c.DRAM.local 3917 ± 6% +54.6% 6054 ± 5% perf-c2c.DRAM.remote 5456 ± 7% +25.3% 6835 ± 7% perf-c2c.HITM.local 3562 ± 7% +22.8% 4376 ± 5% perf-c2c.HITM.remote 9019 ± 7% +24.3% 11212 ± 6% perf-c2c.HITM.total 487248 -6.8% 454278 proc-vmstat.nr_active_anon 107.00 ± 22% -53.4% 49.84 ± 46% proc-vmstat.nr_anon_transparent_hugepages 1164233 -2.7% 1133175 proc-vmstat.nr_file_pages 56483 ± 5% -14.1% 48516 ± 6% proc-vmstat.nr_mapped 287834 -10.8% 256778 proc-vmstat.nr_shmem 487248 -6.8% 454278 proc-vmstat.nr_zone_active_anon 25013 ± 34% +124.4% 56137 ± 5% proc-vmstat.numa_hint_faults 11595 ± 47% +102.3% 23462 ± 38% proc-vmstat.numa_hint_faults_local 1728407 -1.8% 1697091 proc-vmstat.numa_hit 1493274 -1.7% 1467674 proc-vmstat.numa_local 179326 ± 15% +70.1% 305113 ± 2% proc-vmstat.numa_pte_updates 1881688 -4.2% 1802948 proc-vmstat.pgalloc_normal 1350178 +3.0% 1390538 proc-vmstat.pgfault 64622 ± 17% +29.0% 83367 ± 6% proc-vmstat.pgreuse 46897964 ± 2% +23.2% 57779828 ± 5% sched_debug.cfs_rq:/.avg_vruntime.max 30929943 +10.8% 34256347 sched_debug.cfs_rq:/.avg_vruntime.min 1297707 ± 7% +112.5% 2758168 ± 11% sched_debug.cfs_rq:/.avg_vruntime.stddev 0.75 ± 11% -29.6% 0.53 ± 11% sched_debug.cfs_rq:/.h_nr_running.min 3452 ± 10% -29.5% 2433 ± 11% sched_debug.cfs_rq:/.load.min 3.00 ± 11% -29.6% 2.11 ± 10% sched_debug.cfs_rq:/.load_avg.min 46897964 ± 2% +23.2% 57779828 ± 5% sched_debug.cfs_rq:/.min_vruntime.max 30929943 +10.8% 34256347 sched_debug.cfs_rq:/.min_vruntime.min 1297707 ± 7% +112.5% 2758169 ± 11% sched_debug.cfs_rq:/.min_vruntime.stddev 0.75 ± 11% -29.6% 0.53 ± 11% sched_debug.cfs_rq:/.nr_running.min 761.36 ± 10% -27.3% 553.17 ± 7% sched_debug.cfs_rq:/.runnable_avg.min 673.31 ± 12% -24.7% 507.00 ± 9% sched_debug.cfs_rq:/.util_avg.min 407.66 +107.2% 844.62 sched_debug.cfs_rq:/.util_est.avg 755.11 ± 6% +98.5% 1498 ± 8% sched_debug.cfs_rq:/.util_est.max 7.03 ±147% +6359.7% 453.97 ± 21% sched_debug.cfs_rq:/.util_est.min 972664 -69.0% 301696 ± 18% sched_debug.cpu.avg_idle.avg 1514330 ± 4% -51.5% 734068 ± 11% sched_debug.cpu.avg_idle.max 161784 ± 3% -97.2% 4568 ± 49% sched_debug.cpu.avg_idle.min 26.03 ± 4% -13.2% 22.60 ± 8% sched_debug.cpu.clock.stddev 5158 ± 7% -25.0% 3867 ± 18% sched_debug.cpu.curr->pid.min 2243 +41.7% 3179 ± 4% sched_debug.cpu.nr_switches.avg 783.47 ± 4% +45.8% 1142 ± 2% sched_debug.cpu.nr_switches.min 0.18 -14.7% 0.16 ± 2% perf-stat.i.MPKI 1.286e+10 +143.1% 3.126e+10 perf-stat.i.branch-instructions 0.11 -0.0 0.06 perf-stat.i.branch-miss-rate% 12831155 +25.5% 16098495 perf-stat.i.branch-misses 30.76 ± 2% +3.9 34.69 ± 2% perf-stat.i.cache-miss-rate% 9505547 +150.7% 23827336 ± 2% perf-stat.i.cache-misses 32785522 ± 2% +112.8% 69779552 perf-stat.i.cache-references 1988 +71.3% 3406 ± 8% perf-stat.i.context-switches 12.29 -66.5% 4.12 perf-stat.i.cpi 266.62 +7.4% 286.36 perf-stat.i.cpu-migrations 69838 -60.9% 27322 ± 2% perf-stat.i.cycles-between-cache-misses 5.257e+10 +198.1% 1.567e+11 perf-stat.i.instructions 0.09 +188.2% 0.25 perf-stat.i.ipc 0.18 -15.8% 0.15 ± 2% perf-stat.overall.MPKI 0.10 -0.0 0.05 perf-stat.overall.branch-miss-rate% 28.59 +5.5 34.11 ± 2% perf-stat.overall.cache-miss-rate% 12.31 -66.5% 4.13 perf-stat.overall.cpi 68150 -60.2% 27152 ± 2% perf-stat.overall.cycles-between-cache-misses 0.08 +198.3% 0.24 perf-stat.overall.ipc 25547531 ± 2% +14.4% 29237729 perf-stat.overall.path-length 1.282e+10 +143.1% 3.116e+10 perf-stat.ps.branch-instructions 12599626 +27.1% 16009955 perf-stat.ps.branch-misses 9468988 +151.0% 23763879 ± 2% perf-stat.ps.cache-misses 33116170 +110.4% 69681456 perf-stat.ps.cache-references 1970 +72.2% 3393 ± 8% perf-stat.ps.context-switches 262.57 +8.5% 284.83 perf-stat.ps.cpu-migrations 5.239e+10 +198.1% 1.562e+11 perf-stat.ps.instructions 4015 +2.3% 4106 perf-stat.ps.minor-faults 4015 +2.3% 4106 perf-stat.ps.page-faults 1.583e+13 +198.0% 4.719e+13 perf-stat.total.instructions 3.48 ± 7% -88.4% 0.40 ±223% perf-sched.sch_delay.avg.ms.__cond_resched.__alloc_pages_noprof.alloc_pages_mpol_noprof.folio_alloc_mpol_noprof.shmem_alloc_folio 3.63 ± 11% -94.0% 0.22 ±223% perf-sched.sch_delay.avg.ms.__cond_resched.__do_fault.do_read_fault.do_pte_missing.__handle_mm_fault 0.13 ± 3% -74.7% 0.03 ± 7% perf-sched.sch_delay.avg.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity 0.02 ±108% +1600.0% 0.28 ± 35% perf-sched.sch_delay.avg.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64 0.77 ± 4% -45.9% 0.42 ± 18% perf-sched.sch_delay.avg.ms.__cond_resched.dput.terminate_walk.path_openat.do_filp_open 3.65 ± 4% -93.0% 0.25 ±222% perf-sched.sch_delay.avg.ms.__cond_resched.generic_perform_write.shmem_file_write_iter.vfs_write.ksys_write 0.09 ± 47% -97.4% 0.00 ±143% perf-sched.sch_delay.avg.ms.__cond_resched.kvfree_rcu_drain_ready.kfree_rcu_monitor.process_one_work.worker_thread 3.50 ± 16% -92.6% 0.26 ±223% perf-sched.sch_delay.avg.ms.__cond_resched.shmem_inode_acct_blocks.shmem_alloc_and_add_folio.shmem_get_folio_gfp.shmem_write_begin 0.05 ± 10% -74.1% 0.01 ± 85% perf-sched.sch_delay.avg.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm 0.87 ± 11% -26.2% 0.64 ± 16% perf-sched.sch_delay.avg.ms.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part 0.61 ±103% -99.3% 0.00 ± 14% perf-sched.sch_delay.avg.ms.ipmi_thread.kthread.ret_from_fork.ret_from_fork_asm 0.07 ± 16% -89.8% 0.01 ± 11% perf-sched.sch_delay.avg.ms.irq_thread.kthread.ret_from_fork.ret_from_fork_asm 3.83 -27.8% 2.76 ± 9% perf-sched.sch_delay.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown] 3.40 ± 19% -90.7% 0.32 ± 38% perf-sched.sch_delay.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi.[unknown] 0.12 ± 35% -67.6% 0.04 ± 42% perf-sched.sch_delay.avg.ms.pipe_read.vfs_read.ksys_read.do_syscall_64 0.53 ± 16% -77.4% 0.12 ± 64% perf-sched.sch_delay.avg.ms.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll 0.14 ± 37% +97.9% 0.28 ± 10% perf-sched.sch_delay.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone 0.06 ± 35% -88.2% 0.01 ± 70% perf-sched.sch_delay.avg.ms.schedule_timeout.msleep.ast_astdp_connector_helper_detect_ctx.drm_helper_probe_detect_ctx 0.07 ± 4% -49.5% 0.04 ± 50% perf-sched.sch_delay.avg.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread 0.02 ± 4% -65.4% 0.01 ± 23% perf-sched.sch_delay.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm 0.74 ± 5% -65.9% 0.25 ± 67% perf-sched.sch_delay.avg.ms.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown] 0.23 ± 45% -83.4% 0.04 ± 96% perf-sched.sch_delay.avg.ms.usleep_range_state.ipmi_thread.kthread.ret_from_fork 4.13 -83.9% 0.67 ±223% perf-sched.sch_delay.max.ms.__cond_resched.__alloc_pages_noprof.alloc_pages_mpol_noprof.folio_alloc_mpol_noprof.shmem_alloc_folio 4.14 -84.0% 0.66 ±223% perf-sched.sch_delay.max.ms.__cond_resched.__do_fault.do_read_fault.do_pte_missing.__handle_mm_fault 0.03 ±156% +21902.3% 6.42 ± 77% perf-sched.sch_delay.max.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64 18.15 ± 42% -74.4% 4.64 ± 30% perf-sched.sch_delay.max.ms.__cond_resched.dput.terminate_walk.path_openat.do_filp_open 4.15 -83.9% 0.67 ±223% perf-sched.sch_delay.max.ms.__cond_resched.generic_perform_write.shmem_file_write_iter.vfs_write.ksys_write 0.10 ± 39% -97.6% 0.00 ±143% perf-sched.sch_delay.max.ms.__cond_resched.kvfree_rcu_drain_ready.kfree_rcu_monitor.process_one_work.worker_thread 4.09 -83.7% 0.67 ±223% perf-sched.sch_delay.max.ms.__cond_resched.shmem_inode_acct_blocks.shmem_alloc_and_add_folio.shmem_get_folio_gfp.shmem_write_begin 3.28 +13.5% 3.72 ± 6% perf-sched.sch_delay.max.ms.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part 1.05 ±119% -99.6% 0.00 ± 8% perf-sched.sch_delay.max.ms.ipmi_thread.kthread.ret_from_fork.ret_from_fork_asm 0.16 ± 9% -92.7% 0.01 ± 21% perf-sched.sch_delay.max.ms.irq_thread.kthread.ret_from_fork.ret_from_fork_asm 13.59 ± 16% -64.6% 4.81 ± 41% perf-sched.sch_delay.max.ms.pipe_read.vfs_read.ksys_read.do_syscall_64 3.98 ± 4% -80.3% 0.78 ± 42% perf-sched.sch_delay.max.ms.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll 0.06 ± 81% -92.7% 0.00 ± 83% perf-sched.sch_delay.max.ms.schedule_timeout.__wait_for_common.__flush_work.__lru_add_drain_all 0.09 ± 50% -88.3% 0.01 ±101% perf-sched.sch_delay.max.ms.schedule_timeout.msleep.ast_astdp_connector_helper_detect_ctx.drm_helper_probe_detect_ctx 0.20 ± 9% +2615.7% 5.51 ± 60% perf-sched.sch_delay.max.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread 2.40 ± 36% +53.4% 3.68 ± 8% perf-sched.sch_delay.max.ms.wait_for_partner.fifo_open.do_dentry_open.vfs_open 144.41 ± 3% -31.7% 98.62 ± 7% perf-sched.total_wait_and_delay.average.ms 10694 ± 2% +33.0% 14221 ± 5% perf-sched.total_wait_and_delay.count.ms 143.83 ± 3% -31.8% 98.07 ± 7% perf-sched.total_wait_time.average.ms 3.50 ± 5% +85.0% 6.47 ± 12% perf-sched.wait_and_delay.avg.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity 411.43 ± 20% -58.3% 171.38 ± 70% perf-sched.wait_and_delay.avg.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm 0.87 ± 11% -24.7% 0.66 ± 19% perf-sched.wait_and_delay.avg.ms.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part 683.93 ± 24% -22.0% 533.76 ± 35% perf-sched.wait_and_delay.avg.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep 1.90 ± 6% -17.9% 1.56 ± 13% perf-sched.wait_and_delay.avg.ms.do_wait.kernel_wait4.do_syscall_64.entry_SYSCALL_64_after_hwframe 7.66 -27.8% 5.53 ± 9% perf-sched.wait_and_delay.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown] 188.08 ± 32% -57.5% 79.88 ± 23% perf-sched.wait_and_delay.avg.ms.pipe_read.vfs_read.ksys_read.do_syscall_64 1.39 ± 16% +20.1% 1.67 ± 7% perf-sched.wait_and_delay.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone 2265 -96.8% 72.33 ± 46% perf-sched.wait_and_delay.count.__cond_resched.dput.terminate_walk.path_openat.do_filp_open 123.00 -24.0% 93.50 ± 9% perf-sched.wait_and_delay.count.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part 102.00 +9.5% 111.67 ± 2% perf-sched.wait_and_delay.count.do_wait.kernel_wait4.do_syscall_64.entry_SYSCALL_64_after_hwframe 409.00 ± 25% +135.5% 963.33 ± 27% perf-sched.wait_and_delay.count.pipe_read.vfs_read.ksys_read.do_syscall_64 2450 ± 5% -81.0% 466.00 ± 76% perf-sched.wait_and_delay.count.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown] 1003 +348.9% 4506 ± 23% perf-sched.wait_and_delay.max.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity 3.28 +31.6% 4.31 ± 33% perf-sched.wait_and_delay.max.ms.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part 143.48 ±139% -99.7% 0.40 ±223% perf-sched.wait_time.avg.ms.__cond_resched.__alloc_pages_noprof.alloc_pages_mpol_noprof.folio_alloc_mpol_noprof.shmem_alloc_folio 116.07 ±216% -99.8% 0.22 ±223% perf-sched.wait_time.avg.ms.__cond_resched.__do_fault.do_read_fault.do_pte_missing.__handle_mm_fault 3.37 ± 5% +91.1% 6.44 ± 12% perf-sched.wait_time.avg.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity 0.02 ±108% +3587.0% 0.61 ± 34% perf-sched.wait_time.avg.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64 361.12 ±160% -99.9% 0.26 ±223% perf-sched.wait_time.avg.ms.__cond_resched.shmem_inode_acct_blocks.shmem_alloc_and_add_folio.shmem_get_folio_gfp.shmem_write_begin 411.38 ± 20% -54.1% 188.90 ± 51% perf-sched.wait_time.avg.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm 683.63 ± 24% -22.0% 533.55 ± 35% perf-sched.wait_time.avg.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep 0.79 ± 76% -74.2% 0.20 perf-sched.wait_time.avg.ms.ipmi_thread.kthread.ret_from_fork.ret_from_fork_asm 3.83 -27.8% 2.76 ± 9% perf-sched.wait_time.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown] 3.40 ± 19% -90.2% 0.33 ± 43% perf-sched.wait_time.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi.[unknown] 187.96 ± 32% -57.5% 79.84 ± 23% perf-sched.wait_time.avg.ms.pipe_read.vfs_read.ksys_read.do_syscall_64 0.02 ± 48% -91.7% 0.00 ±223% perf-sched.wait_time.avg.ms.schedule_timeout.__wait_for_common.__flush_work.__lru_add_drain_all 0.35 ± 30% -37.1% 0.22 ± 18% perf-sched.wait_time.avg.ms.usleep_range_state.ipmi_thread.kthread.ret_from_fork 1386 ±141% -100.0% 0.67 ±223% perf-sched.wait_time.max.ms.__cond_resched.__alloc_pages_noprof.alloc_pages_mpol_noprof.folio_alloc_mpol_noprof.shmem_alloc_folio 790.56 ±222% -99.9% 0.66 ±223% perf-sched.wait_time.max.ms.__cond_resched.__do_fault.do_read_fault.do_pte_missing.__handle_mm_fault 1002 +349.5% 4505 ± 23% perf-sched.wait_time.max.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity 0.03 ±156% +48113.7% 14.06 ± 79% perf-sched.wait_time.max.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64 1464 ±141% -100.0% 0.67 ±223% perf-sched.wait_time.max.ms.__cond_resched.shmem_inode_acct_blocks.shmem_alloc_and_add_folio.shmem_get_folio_gfp.shmem_write_begin 1.21 ± 98% -83.2% 0.20 perf-sched.wait_time.max.ms.ipmi_thread.kthread.ret_from_fork.ret_from_fork_asm 0.05 ± 47% -96.2% 0.00 ±223% perf-sched.wait_time.max.ms.schedule_timeout.__wait_for_common.__flush_work.__lru_add_drain_all 2.40 ± 36% +53.4% 3.68 ± 8% perf-sched.wait_time.max.ms.wait_for_partner.fifo_open.do_dentry_open.vfs_open 75.05 -72.6 2.42 ± 10% perf-profile.calltrace.cycles-pp.path_openat.do_filp_open.do_sys_openat2.__x64_sys_openat.do_syscall_64 75.05 -72.6 2.43 ± 10% perf-profile.calltrace.cycles-pp.do_filp_open.do_sys_openat2.__x64_sys_openat.do_syscall_64.entry_SYSCALL_64_after_hwframe 75.14 -72.6 2.57 ± 9% perf-profile.calltrace.cycles-pp.do_sys_openat2.__x64_sys_openat.do_syscall_64.entry_SYSCALL_64_after_hwframe.open64 75.14 -72.6 2.58 ± 9% perf-profile.calltrace.cycles-pp.__x64_sys_openat.do_syscall_64.entry_SYSCALL_64_after_hwframe.open64 75.16 -72.6 2.61 ± 9% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.open64 75.16 -72.5 2.62 ± 9% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.open64 75.18 -72.5 2.66 ± 9% perf-profile.calltrace.cycles-pp.open64 50.23 -48.6 1.66 ± 12% perf-profile.calltrace.cycles-pp.do_open.path_openat.do_filp_open.do_sys_openat2.__x64_sys_openat 25.20 -25.2 0.00 perf-profile.calltrace.cycles-pp.lockref_get.do_dentry_open.vfs_open.do_open.path_openat 25.16 -25.2 0.00 perf-profile.calltrace.cycles-pp._raw_spin_lock.lockref_get.do_dentry_open.vfs_open.do_open 25.16 -25.2 0.00 perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock.lockref_get.do_dentry_open.vfs_open 24.88 -24.9 0.00 perf-profile.calltrace.cycles-pp._raw_spin_lock.lockref_get_not_dead.__legitimize_path.try_to_unlazy.complete_walk 24.87 -24.9 0.00 perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock.lockref_get_not_dead.__legitimize_path.try_to_unlazy 24.73 -24.7 0.00 perf-profile.calltrace.cycles-pp._raw_spin_lock.dput.terminate_walk.path_openat.do_filp_open 24.72 -24.7 0.00 perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock.dput.terminate_walk.path_openat 24.92 -24.5 0.46 ± 45% perf-profile.calltrace.cycles-pp.lockref_get_not_dead.__legitimize_path.try_to_unlazy.complete_walk.do_open 24.46 -24.5 0.00 perf-profile.calltrace.cycles-pp._raw_spin_lock.dput.__fput.__x64_sys_close.do_syscall_64 24.92 -24.5 0.46 ± 45% perf-profile.calltrace.cycles-pp.__legitimize_path.try_to_unlazy.complete_walk.do_open.path_openat 24.93 -24.5 0.47 ± 45% perf-profile.calltrace.cycles-pp.try_to_unlazy.complete_walk.do_open.path_openat.do_filp_open 24.93 -24.5 0.47 ± 45% perf-profile.calltrace.cycles-pp.complete_walk.do_open.path_openat.do_filp_open.do_sys_openat2 24.45 -24.5 0.00 perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock.dput.__fput.__x64_sys_close 25.28 -24.2 1.05 ± 17% perf-profile.calltrace.cycles-pp.do_dentry_open.vfs_open.do_open.path_openat.do_filp_open 25.28 -24.2 1.07 ± 17% perf-profile.calltrace.cycles-pp.vfs_open.do_open.path_openat.do_filp_open.do_sys_openat2 24.76 -24.2 0.60 ± 6% perf-profile.calltrace.cycles-pp.dput.terminate_walk.path_openat.do_filp_open.do_sys_openat2 24.76 -24.1 0.61 ± 6% perf-profile.calltrace.cycles-pp.terminate_walk.path_openat.do_filp_open.do_sys_openat2.__x64_sys_openat 24.50 -23.9 0.57 ± 7% perf-profile.calltrace.cycles-pp.dput.__fput.__x64_sys_close.do_syscall_64.entry_SYSCALL_64_after_hwframe 0.00 +0.5 0.54 ± 6% perf-profile.calltrace.cycles-pp.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.osq_lock.__mutex_lock 0.00 +0.6 0.55 ± 5% perf-profile.calltrace.cycles-pp.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.osq_lock.__mutex_lock.evm_file_release 0.00 +0.6 0.56 ± 7% perf-profile.calltrace.cycles-pp.lockref_put_return.dput.__fput.__x64_sys_close.do_syscall_64 0.00 +0.6 0.58 ± 5% perf-profile.calltrace.cycles-pp.asm_sysvec_apic_timer_interrupt.osq_lock.__mutex_lock.evm_file_release.security_file_release 0.00 +0.6 0.60 ± 7% perf-profile.calltrace.cycles-pp.lockref_put_return.dput.terminate_walk.path_openat.do_filp_open 24.57 +72.2 96.75 perf-profile.calltrace.cycles-pp.__fput.__x64_sys_close.do_syscall_64.entry_SYSCALL_64_after_hwframe.__close 24.63 +72.4 97.02 perf-profile.calltrace.cycles-pp.__x64_sys_close.do_syscall_64.entry_SYSCALL_64_after_hwframe.__close 24.64 +72.4 97.09 perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.__close 24.65 +72.5 97.10 perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.__close 24.66 +72.5 97.15 perf-profile.calltrace.cycles-pp.__close 0.00 +95.1 95.15 perf-profile.calltrace.cycles-pp.osq_lock.__mutex_lock.evm_file_release.security_file_release.__fput 0.00 +95.5 95.52 perf-profile.calltrace.cycles-pp.__mutex_lock.evm_file_release.security_file_release.__fput.__x64_sys_close 0.00 +95.7 95.71 perf-profile.calltrace.cycles-pp.evm_file_release.security_file_release.__fput.__x64_sys_close.do_syscall_64 0.00 +95.9 95.87 perf-profile.calltrace.cycles-pp.security_file_release.__fput.__x64_sys_close.do_syscall_64.entry_SYSCALL_64_after_hwframe 99.26 -99.3 0.00 perf-profile.children.cycles-pp._raw_spin_lock 99.23 -99.2 0.00 perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath 75.05 -72.6 2.42 ± 10% perf-profile.children.cycles-pp.path_openat 75.05 -72.6 2.43 ± 10% perf-profile.children.cycles-pp.do_filp_open 75.14 -72.6 2.58 ± 9% perf-profile.children.cycles-pp.do_sys_openat2 75.14 -72.6 2.58 ± 9% perf-profile.children.cycles-pp.__x64_sys_openat 75.18 -72.5 2.68 ± 9% perf-profile.children.cycles-pp.open64 50.23 -48.6 1.66 ± 12% perf-profile.children.cycles-pp.do_open 49.26 -48.1 1.17 ± 6% perf-profile.children.cycles-pp.dput 25.20 -24.8 0.44 ± 6% perf-profile.children.cycles-pp.lockref_get 24.92 -24.4 0.54 ± 7% perf-profile.children.cycles-pp.lockref_get_not_dead 24.93 -24.4 0.54 ± 7% perf-profile.children.cycles-pp.__legitimize_path 24.93 -24.4 0.56 ± 7% perf-profile.children.cycles-pp.complete_walk 24.93 -24.4 0.55 ± 7% perf-profile.children.cycles-pp.try_to_unlazy 25.28 -24.2 1.06 ± 17% perf-profile.children.cycles-pp.do_dentry_open 25.29 -24.2 1.07 ± 17% perf-profile.children.cycles-pp.vfs_open 24.76 -24.1 0.61 ± 6% perf-profile.children.cycles-pp.terminate_walk 99.88 -0.1 99.80 perf-profile.children.cycles-pp.do_syscall_64 99.88 -0.1 99.81 perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe 0.15 ± 4% +0.0 0.17 ± 2% perf-profile.children.cycles-pp.task_tick_fair 0.24 ± 4% +0.0 0.26 perf-profile.children.cycles-pp.sched_tick 0.38 ± 5% +0.0 0.42 ± 4% perf-profile.children.cycles-pp.__hrtimer_run_queues 0.38 ± 5% +0.0 0.42 ± 4% perf-profile.children.cycles-pp.tick_nohz_handler 0.35 ± 4% +0.0 0.39 ± 3% perf-profile.children.cycles-pp.update_process_times 0.00 +0.1 0.05 perf-profile.children.cycles-pp.entry_SYSCALL_64 0.00 +0.1 0.05 perf-profile.children.cycles-pp.kmem_cache_alloc_noprof 0.00 +0.1 0.05 ± 8% perf-profile.children.cycles-pp.alloc_empty_file 0.00 +0.1 0.05 ± 13% perf-profile.children.cycles-pp.mutex_spin_on_owner 0.51 ± 6% +0.1 0.56 ± 5% perf-profile.children.cycles-pp.__sysvec_apic_timer_interrupt 0.00 +0.1 0.06 ± 16% perf-profile.children.cycles-pp.evm_iint_find 0.00 +0.1 0.06 perf-profile.children.cycles-pp.link_path_walk 0.00 +0.1 0.07 ± 5% perf-profile.children.cycles-pp.syscall_exit_to_user_mode 0.00 +0.1 0.09 perf-profile.children.cycles-pp.getname_flags 0.00 +0.1 0.11 ± 8% perf-profile.children.cycles-pp.osq_unlock 0.00 +0.1 0.12 ± 4% perf-profile.children.cycles-pp.mutex_unlock 0.00 +0.1 0.13 ± 14% perf-profile.children.cycles-pp.mutex_lock 0.05 +1.1 1.16 ± 6% perf-profile.children.cycles-pp.lockref_put_return 24.57 +72.2 96.76 perf-profile.children.cycles-pp.__fput 24.63 +72.4 97.02 perf-profile.children.cycles-pp.__x64_sys_close 24.67 +72.5 97.17 perf-profile.children.cycles-pp.__close 0.00 +95.2 95.17 perf-profile.children.cycles-pp.osq_lock 0.00 +95.5 95.53 perf-profile.children.cycles-pp.__mutex_lock 0.00 +95.7 95.71 perf-profile.children.cycles-pp.evm_file_release 0.00 +95.9 95.87 perf-profile.children.cycles-pp.security_file_release 98.69 -98.7 0.00 perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath 0.00 +0.1 0.05 ± 7% perf-profile.self.cycles-pp.mutex_spin_on_owner 0.00 +0.1 0.05 ± 13% perf-profile.self.cycles-pp.evm_iint_find 0.00 +0.1 0.11 ± 8% perf-profile.self.cycles-pp.osq_unlock 0.00 +0.1 0.11 ± 4% perf-profile.self.cycles-pp.mutex_unlock 0.00 +0.1 0.12 ± 13% perf-profile.self.cycles-pp.mutex_lock 0.06 +0.2 0.22 ± 23% perf-profile.self.cycles-pp.__fput 0.00 +0.2 0.19 ± 7% perf-profile.self.cycles-pp.__mutex_lock 0.00 +0.4 0.43 ± 6% perf-profile.self.cycles-pp.lockref_get 0.07 ± 5% +0.5 0.59 ± 29% perf-profile.self.cycles-pp.do_dentry_open 0.00 +0.5 0.54 ± 7% perf-profile.self.cycles-pp.lockref_get_not_dead 0.05 +1.1 1.16 ± 7% perf-profile.self.cycles-pp.lockref_put_return 0.00 +94.6 94.60 perf-profile.self.cycles-pp.osq_lock Disclaimer: Results have been estimated based on internal Intel analysis and are provided for informational purposes only. Any difference in system hardware or software design or configuration may affect actual performance. -- 0-DAY CI Kernel Test Service https://github.com/intel/lkp-tests/wiki