[cifs:for-next-next] [smb3] 29ed7da631: filebench.sum_operations/s 18.8% improvement

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 




Hello,

kernel test robot noticed a 18.8% improvement of filebench.sum_operations/s on:


commit: 29ed7da6313fc6048b9a60c75aae44fc8f6ddbab ("smb3: allow files to be created with backslash in name")
git://git.samba.org/sfrench/cifs-2.6.git for-next-next

testcase: filebench
test machine: 128 threads 2 sockets Intel(R) Xeon(R) Platinum 8358 CPU @ 2.60GHz (Ice Lake) with 128G memory
parameters:

	disk: 1HDD
	fs: ext4
	fs2: cifs
	test: fileserver.f
	cpufreq_governor: performance


In addition to that, the commit also has significant impact on the following tests:

+------------------+------------------------------------------------------------------------------------------------+
| testcase: change | filebench: filebench.sum_operations/s 45.9% improvement                                        |
| test machine     | 128 threads 2 sockets Intel(R) Xeon(R) Platinum 8358 CPU @ 2.60GHz (Ice Lake) with 128G memory |
| test parameters  | cpufreq_governor=performance                                                                   |
|                  | disk=1HDD                                                                                      |
|                  | fs2=cifs                                                                                       |
|                  | fs=xfs                                                                                         |
|                  | test=webserver.f                                                                               |
+------------------+------------------------------------------------------------------------------------------------+




Details are as below:
-------------------------------------------------------------------------------------------------->


The kernel config and materials to reproduce are available at:
https://download.01.org/0day-ci/archive/20231227/202312271513.cd04d453-oliver.sang@xxxxxxxxx

=========================================================================================
compiler/cpufreq_governor/disk/fs2/fs/kconfig/rootfs/tbox_group/test/testcase:
  gcc-12/performance/1HDD/cifs/ext4/x86_64-rhel-8.3/debian-11.1-x86_64-20220510.cgz/lkp-icl-2sp6/fileserver.f/filebench

commit: 
  54f95f4e6f ("smb: client: Fix minor whitespace errors and warnings")
  29ed7da631 ("smb3: allow files to be created with backslash in name")

54f95f4e6ff76fa3 29ed7da6313fc6048b9a60c75aa 
---------------- --------------------------- 
         %stddev     %change         %stddev
             \          |                \  
    421.00 ±  4%     +10.0%     463.00 ±  6%  perf-c2c.HITM.local
     32.70 ±  4%     -11.5%      28.93 ±  4%  vmstat.procs.r
     50682            +7.4%      54452        vmstat.system.cs
     72.78            +2.9%      74.91        iostat.cpu.idle
      1.17            +4.0%       1.22 ±  3%  iostat.cpu.iowait
     25.69            -8.6%      23.48        iostat.cpu.system
      0.10            +0.0        0.11        mpstat.cpu.all.soft%
     24.28            -2.5       21.80        mpstat.cpu.all.sys%
      0.36            +0.0        0.40        mpstat.cpu.all.usr%
      3136            -1.2%       3099        turbostat.Bzy_MHz
      6.19            +0.5        6.66        turbostat.C1E%
    298.22            -1.9%     292.51        turbostat.PkgWatt
    108.62           +19.2%     129.50        filebench.sum_bytes_mb/s
    275222           +18.8%     326951        filebench.sum_operations
      4586           +18.8%       5448        filebench.sum_operations/s
    417.00           +18.8%     495.33        filebench.sum_reads/s
     10.89           -15.8%       9.17        filebench.sum_time_ms/op
    834.17           +18.8%     991.00        filebench.sum_writes/s
   9095557           +15.1%   10466277        filebench.time.file_system_outputs
      9077            -8.2%       8336        filebench.time.involuntary_context_switches
      2140            -6.9%       1993        filebench.time.percent_of_cpu_this_job_got
      1453            -6.7%       1355        filebench.time.system_time
    545023           +11.9%     609623        filebench.time.voluntary_context_switches
     27722            +4.6%      29007        proc-vmstat.nr_active_anon
     51938            -6.2%      48695        proc-vmstat.nr_active_file
   2259798           +14.9%    2596743        proc-vmstat.nr_dirtied
    169458            +6.1%     179785        proc-vmstat.nr_dirty
   1133821            +1.8%    1154179        proc-vmstat.nr_file_pages
    317423            +7.0%     339622        proc-vmstat.nr_inactive_file
     44201            +3.3%      45644        proc-vmstat.nr_shmem
   1530949           +12.0%    1715152        proc-vmstat.nr_written
     27722            +4.6%      29007        proc-vmstat.nr_zone_active_anon
     51938            -6.2%      48695        proc-vmstat.nr_zone_active_file
    317423            +7.0%     339622        proc-vmstat.nr_zone_inactive_file
    168178            +6.2%     178558        proc-vmstat.nr_zone_write_pending
   4020254            +9.6%    4404885        proc-vmstat.numa_hit
   3884880            +9.9%    4270577        proc-vmstat.numa_local
   6063917            +9.1%    6613855        proc-vmstat.pgalloc_normal
   5868547            +9.2%    6405878        proc-vmstat.pgfree
      3.21 ±  4%     -20.8%       2.54 ± 21%  perf-sched.sch_delay.avg.ms.__cond_resched.dput.cifsFileInfo_put_final._cifsFileInfo_put.process_one_work
      0.00 ±152%   +1776.2%       0.07 ±163%  perf-sched.sch_delay.avg.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      0.04 ± 26%     +72.2%       0.07 ± 17%  perf-sched.sch_delay.avg.ms.__cond_resched.stop_one_cpu.migrate_task_to.task_numa_migrate.isra
      0.30 ±105%     -78.0%       0.07 ± 20%  perf-sched.sch_delay.avg.ms.__lock_sock.lock_sock_nested.tcp_sock_set_cork.__smb_send_rqst
      0.02 ±  3%     -13.4%       0.02 ±  2%  perf-sched.sch_delay.avg.ms.schedule_timeout.wait_woken.sk_wait_data.tcp_recvmsg_locked
      0.35 ±  9%     -42.7%       0.20 ± 12%  perf-sched.sch_delay.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      0.01 ±148%   +2277.1%       0.14 ±194%  perf-sched.sch_delay.max.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
    219.43 ±123%     -93.7%      13.89 ±147%  perf-sched.sch_delay.max.ms.__lock_sock.lock_sock_nested.tcp_sock_set_cork.__smb_send_rqst
     12.64 ± 15%     -27.7%       9.14 ± 13%  perf-sched.sch_delay.max.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
     21.79 ± 21%     -57.8%       9.20 ±100%  perf-sched.wait_and_delay.avg.ms.__cond_resched.kmem_cache_alloc.cifs_do_create.isra.0
      0.15 ±  3%     -11.6%       0.13 ±  7%  perf-sched.wait_and_delay.avg.ms.__lock_sock.lock_sock_nested.tcp_recvmsg.inet6_recvmsg
      0.10 ±  8%     -12.7%       0.08 ±  4%  perf-sched.wait_and_delay.avg.ms.__lock_sock.sk_wait_data.tcp_recvmsg_locked.tcp_recvmsg
      0.16           -12.9%       0.14        perf-sched.wait_and_delay.avg.ms.schedule_timeout.wait_woken.sk_wait_data.tcp_recvmsg_locked
      6.77 ±  2%     -11.3%       6.00        perf-sched.wait_and_delay.avg.ms.wait_for_response.compound_send_recv.cifs_send_recv.SMB2_open
      0.70 ±  4%     +18.0%       0.83 ±  5%  perf-sched.wait_and_delay.avg.ms.wait_for_response.compound_send_recv.cifs_send_recv.SMB2_read
      6.12 ±  4%     -18.7%       4.97 ± 14%  perf-sched.wait_and_delay.avg.ms.wait_for_response.compound_send_recv.smb2_compound_op.smb2_unlink
    881.17 ±  6%     -37.5%     551.17 ±  8%  perf-sched.wait_and_delay.count.__cond_resched.__flush_work.isra.0.__cancel_work_timer
      2525           +29.8%       3277 ±  2%  perf-sched.wait_and_delay.count.__lock_sock.lock_sock_nested.tcp_recvmsg.inet6_recvmsg
    933.50 ±  4%     +27.6%       1190 ±  5%  perf-sched.wait_and_delay.count.__lock_sock.lock_sock_nested.tcp_sendmsg.sock_sendmsg
      1383 ±  2%     +18.9%       1645 ±  3%  perf-sched.wait_and_delay.count.__lock_sock.lock_sock_nested.tcp_sock_set_cork.__smb_send_rqst
      3776 ±  2%     +31.6%       4971        perf-sched.wait_and_delay.count.__lock_sock.sk_wait_data.tcp_recvmsg_locked.tcp_recvmsg
      5344           +19.1%       6367 ±  2%  perf-sched.wait_and_delay.count.io_schedule.folio_wait_bit_common.folio_wait_writeback.__filemap_fdatawait_range
      4272 ±  5%     +15.3%       4926 ±  3%  perf-sched.wait_and_delay.count.schedule_preempt_disabled.__mutex_lock.constprop.0.compound_send_recv
      1659           +16.7%       1937        perf-sched.wait_and_delay.count.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.do_unlinkat
      1649           +16.3%       1918        perf-sched.wait_and_delay.count.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.open_last_lookups
     29390           +11.1%      32658        perf-sched.wait_and_delay.count.schedule_timeout.wait_woken.sk_wait_data.tcp_recvmsg_locked
      2032           +16.8%       2374        perf-sched.wait_and_delay.count.wait_for_response.compound_send_recv.cifs_send_recv.query_info
     11622           +14.3%      13279        perf-sched.wait_and_delay.count.wait_for_response.compound_send_recv.smb2_compound_op.smb2_query_path_info
      2034           +16.6%       2372        perf-sched.wait_and_delay.count.wait_for_response.compound_send_recv.smb2_compound_op.smb2_unlink
    157.95 ±  9%     -16.6%     131.81 ±  8%  perf-sched.wait_and_delay.max.ms.wait_for_response.compound_send_recv.smb2_compound_op.smb2_query_path_info
     21.18 ± 21%     -22.2%      16.48 ± 12%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.cifs_do_create.isra.0
      0.11 ±  4%     -14.5%       0.09 ±  4%  perf-sched.wait_time.avg.ms.__lock_sock.lock_sock_nested.tcp_recvmsg.inet6_recvmsg
      0.08 ±  6%     -10.8%       0.07 ±  3%  perf-sched.wait_time.avg.ms.__lock_sock.sk_wait_data.tcp_recvmsg_locked.tcp_recvmsg
      0.13           -13.0%       0.12        perf-sched.wait_time.avg.ms.schedule_timeout.wait_woken.sk_wait_data.tcp_recvmsg_locked
      6.66           -11.4%       5.90        perf-sched.wait_time.avg.ms.wait_for_response.compound_send_recv.cifs_send_recv.SMB2_open
      0.64 ±  3%     +19.8%       0.76 ±  5%  perf-sched.wait_time.avg.ms.wait_for_response.compound_send_recv.cifs_send_recv.SMB2_read
      6.02 ±  4%     -18.8%       4.89 ± 15%  perf-sched.wait_time.avg.ms.wait_for_response.compound_send_recv.smb2_compound_op.smb2_unlink
      5.96 ± 11%     -25.4%       4.44 ± 11%  perf-sched.wait_time.max.ms.__lock_sock.lock_sock_nested.tcp_recvmsg.inet6_recvmsg
    157.85 ±  9%     -16.6%     131.67 ±  8%  perf-sched.wait_time.max.ms.wait_for_response.compound_send_recv.smb2_compound_op.smb2_query_path_info
      0.63           +11.6%       0.71 ±  4%  perf-stat.i.MPKI
 5.193e+09            -6.7%  4.846e+09        perf-stat.i.branch-instructions
      0.74            +0.1        0.82 ±  2%  perf-stat.i.branch-miss-rate%
  29430000            +5.6%   31068143        perf-stat.i.branch-misses
      6.26            -0.2        6.10        perf-stat.i.cache-miss-rate%
  12345040            +7.8%   13312743        perf-stat.i.cache-misses
  1.96e+08           +10.3%  2.163e+08        perf-stat.i.cache-references
     52474            +7.7%      56494        perf-stat.i.context-switches
      4.88            -4.7%       4.65        perf-stat.i.cpi
 1.096e+11            -8.8%      1e+11        perf-stat.i.cpu-cycles
      1654            +6.0%       1752        perf-stat.i.cpu-migrations
      8628           -15.4%       7303 ±  2%  perf-stat.i.cycles-between-cache-misses
      0.13 ±  3%      +0.0        0.14        perf-stat.i.dTLB-load-miss-rate%
   5020142 ±  2%     +13.8%    5711312        perf-stat.i.dTLB-load-misses
    462535            +8.1%     500225        perf-stat.i.dTLB-store-misses
  1.24e+09           +13.6%  1.409e+09        perf-stat.i.dTLB-stores
  2.13e+10            -3.7%   2.05e+10        perf-stat.i.instructions
      0.86            -8.8%       0.78        perf-stat.i.metric.GHz
     91.83            -1.7%      90.27        perf-stat.i.metric.M/sec
   2029913 ±  2%      +8.6%    2204401 ±  5%  perf-stat.i.node-load-misses
   2736286            +9.3%    2991129 ±  2%  perf-stat.i.node-stores
      0.58           +12.1%       0.65        perf-stat.overall.MPKI
      0.56            +0.1        0.64        perf-stat.overall.branch-miss-rate%
      6.32            -0.2        6.17        perf-stat.overall.cache-miss-rate%
      5.16            -5.3%       4.88        perf-stat.overall.cpi
      8920           -15.5%       7538        perf-stat.overall.cycles-between-cache-misses
      0.10 ±  2%      +0.0        0.11        perf-stat.overall.dTLB-load-miss-rate%
      0.04            -0.0        0.04        perf-stat.overall.dTLB-store-miss-rate%
      0.19            +5.6%       0.20        perf-stat.overall.ipc
 5.143e+09            -6.8%  4.792e+09        perf-stat.ps.branch-instructions
  28981289            +5.6%   30608098        perf-stat.ps.branch-misses
  12187330            +7.8%   13136296        perf-stat.ps.cache-misses
  1.93e+08           +10.4%  2.131e+08        perf-stat.ps.cache-references
     51715            +7.7%      55685        perf-stat.ps.context-switches
 1.087e+11            -8.9%  9.901e+10        perf-stat.ps.cpu-cycles
      1637            +5.8%       1733        perf-stat.ps.cpu-migrations
   4939197 ±  2%     +13.9%    5624080        perf-stat.ps.dTLB-load-misses
    454820            +8.2%     492327        perf-stat.ps.dTLB-store-misses
 1.224e+09           +13.6%  1.391e+09        perf-stat.ps.dTLB-stores
 2.108e+10            -3.9%  2.027e+10        perf-stat.ps.instructions
   2009213 ±  2%      +8.4%    2178707 ±  5%  perf-stat.ps.node-load-misses
   2688455            +9.5%    2943088 ±  2%  perf-stat.ps.node-stores
 1.444e+12            -3.8%  1.388e+12        perf-stat.total.instructions
      5.12 ± 21%      -3.0        2.10 ± 39%  perf-profile.calltrace.cycles-pp.do_fault.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault
      3.74 ± 20%      -2.7        1.04 ± 52%  perf-profile.calltrace.cycles-pp.filemap_map_pages.do_read_fault.do_fault.__handle_mm_fault.handle_mm_fault
      5.65 ± 22%      -2.7        3.00 ± 55%  perf-profile.calltrace.cycles-pp.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
      3.97 ± 18%      -2.6        1.40 ± 57%  perf-profile.calltrace.cycles-pp.do_read_fault.do_fault.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
      5.52 ± 20%      -2.5        3.06 ± 16%  perf-profile.calltrace.cycles-pp.__libc_fork
      5.65 ± 22%      -2.4        3.21 ± 48%  perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault
      4.78 ± 22%      -1.9        2.84 ± 32%  perf-profile.calltrace.cycles-pp.exit_mm.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
      2.46 ± 45%      -1.9        0.52 ± 79%  perf-profile.calltrace.cycles-pp.next_uptodate_folio.filemap_map_pages.do_read_fault.do_fault.__handle_mm_fault
      4.29 ± 22%      -1.9        2.37 ± 30%  perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.__libc_fork
      4.29 ± 22%      -1.9        2.37 ± 30%  perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.__libc_fork
      4.29 ± 22%      -1.9        2.37 ± 30%  perf-profile.calltrace.cycles-pp.__do_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe.__libc_fork
      4.29 ± 22%      -1.9        2.37 ± 30%  perf-profile.calltrace.cycles-pp.kernel_clone.__do_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe.__libc_fork
      3.42 ± 29%      -1.9        1.56 ± 59%  perf-profile.calltrace.cycles-pp.exit_mmap.__mmput.exit_mm.do_exit.do_group_exit
      3.75 ± 22%      -1.8        1.92 ± 44%  perf-profile.calltrace.cycles-pp.copy_process.kernel_clone.__do_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe
      2.84 ± 33%      -1.8        1.04 ± 69%  perf-profile.calltrace.cycles-pp.dup_mm.copy_process.kernel_clone.__do_sys_clone.do_syscall_64
      3.31 ± 29%      -1.7        1.56 ± 59%  perf-profile.calltrace.cycles-pp.__mmput.exit_mm.do_exit.do_group_exit.__x64_sys_exit_group
      3.18 ±  5%      -0.4        2.75 ± 11%  perf-profile.calltrace.cycles-pp.record__mmap_read_evlist.__cmd_record.cmd_record.run_builtin.main
      3.18 ±  5%      -0.4        2.75 ± 11%  perf-profile.calltrace.cycles-pp.perf_mmap__push.record__mmap_read_evlist.__cmd_record.cmd_record.run_builtin
      0.23 ±141%      +0.7        0.95 ± 50%  perf-profile.calltrace.cycles-pp.proc_task_name.proc_pid_status.proc_single_show.seq_read_iter.seq_read
      9.40 ± 18%      -4.2        5.21 ± 34%  perf-profile.children.cycles-pp.do_user_addr_fault
      9.40 ± 18%      -4.0        5.42 ± 30%  perf-profile.children.cycles-pp.exc_page_fault
      5.12 ± 21%      -3.0        2.11 ± 39%  perf-profile.children.cycles-pp.do_fault
      5.82 ± 19%      -2.9        2.92 ± 54%  perf-profile.children.cycles-pp.__mmput
      3.85 ± 22%      -2.8        1.04 ± 52%  perf-profile.children.cycles-pp.filemap_map_pages
      5.58 ± 20%      -2.8        2.83 ± 59%  perf-profile.children.cycles-pp.exit_mmap
      3.97 ± 18%      -2.6        1.40 ± 57%  perf-profile.children.cycles-pp.do_read_fault
      5.55 ± 18%      -2.5        3.06 ± 16%  perf-profile.children.cycles-pp.__libc_fork
      5.12 ± 22%      -2.3        2.84 ± 32%  perf-profile.children.cycles-pp.exit_mm
      5.68 ± 33%      -2.1        3.62 ± 27%  perf-profile.children.cycles-pp.__schedule
      4.29 ± 22%      -1.9        2.37 ± 30%  perf-profile.children.cycles-pp.__do_sys_clone
      4.29 ± 22%      -1.9        2.37 ± 30%  perf-profile.children.cycles-pp.kernel_clone
      3.75 ± 22%      -1.8        1.92 ± 44%  perf-profile.children.cycles-pp.copy_process
      2.34 ± 41%      -1.8        0.52 ± 79%  perf-profile.children.cycles-pp.next_uptodate_folio
      2.84 ± 33%      -1.8        1.04 ± 69%  perf-profile.children.cycles-pp.dup_mm
      1.01 ± 48%      -0.7        0.30 ±100%  perf-profile.children.cycles-pp.lock_vma_under_rcu
      0.90 ± 34%      -0.6        0.29 ±100%  perf-profile.children.cycles-pp.mtree_range_walk
      3.18 ±  5%      -0.4        2.75 ± 11%  perf-profile.children.cycles-pp.record__mmap_read_evlist
      3.18 ±  5%      -0.4        2.75 ± 11%  perf-profile.children.cycles-pp.perf_mmap__push
      0.23 ±141%      +0.7        0.95 ± 50%  perf-profile.children.cycles-pp.proc_task_name
      5.69 ± 31%      +2.4        8.08 ± 11%  perf-profile.children.cycles-pp.seq_read_iter
     63.77 ±  9%      +8.5       72.28 ±  4%  perf-profile.children.cycles-pp.do_syscall_64
     63.77 ±  9%      +8.6       72.41 ±  4%  perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe
      1.88 ± 53%      -1.4        0.52 ± 79%  perf-profile.self.cycles-pp.next_uptodate_folio
      0.90 ± 34%      -0.6        0.29 ±100%  perf-profile.self.cycles-pp.mtree_range_walk


***************************************************************************************************
lkp-icl-2sp6: 128 threads 2 sockets Intel(R) Xeon(R) Platinum 8358 CPU @ 2.60GHz (Ice Lake) with 128G memory
=========================================================================================
compiler/cpufreq_governor/disk/fs2/fs/kconfig/rootfs/tbox_group/test/testcase:
  gcc-12/performance/1HDD/cifs/xfs/x86_64-rhel-8.3/debian-11.1-x86_64-20220510.cgz/lkp-icl-2sp6/webserver.f/filebench

commit: 
  54f95f4e6f ("smb: client: Fix minor whitespace errors and warnings")
  29ed7da631 ("smb3: allow files to be created with backslash in name")

54f95f4e6ff76fa3 29ed7da6313fc6048b9a60c75aa 
---------------- --------------------------- 
         %stddev     %change         %stddev
             \          |                \  
    324.60 ±  6%     -10.2%     291.63 ±  2%  sched_debug.cfs_rq:/.runnable_avg.stddev
    322.59 ±  6%      -9.9%     290.68 ±  2%  sched_debug.cfs_rq:/.util_avg.stddev
      4227 ± 10%     +22.7%       5187 ±  7%  vmstat.io.bo
     24086           +20.7%      29072 ±  3%  vmstat.system.cs
    107057 ±  9%     +43.7%     153891 ±  6%  meminfo.Dirty
    102405 ±  4%     +46.5%     149992        meminfo.FileHugePages
    326524 ±  3%     +42.8%     466341        meminfo.Inactive(file)
     47288 ± 25%     +45.5%      68805 ± 12%  numa-meminfo.node0.Dirty
    137726 ± 30%     +55.3%     213928 ± 20%  numa-meminfo.node0.Inactive(file)
     59597 ± 22%     +43.2%      85366 ± 10%  numa-meminfo.node1.Dirty
    560238           +17.2%     656377        turbostat.C1E
      1.17            +0.1        1.24        turbostat.C1E%
     12032 ± 11%     +41.4%      17014 ± 12%  turbostat.POLL
     11823 ± 25%     +45.6%      17216 ± 12%  numa-vmstat.node0.nr_dirty
     34433 ± 30%     +55.4%      53497 ± 20%  numa-vmstat.node0.nr_inactive_file
     34433 ± 30%     +55.4%      53497 ± 20%  numa-vmstat.node0.nr_zone_inactive_file
     11995 ± 26%     +45.3%      17432 ± 12%  numa-vmstat.node0.nr_zone_write_pending
     14904 ± 22%     +43.3%      21350 ± 10%  numa-vmstat.node1.nr_dirty
     15299 ± 22%     +42.3%      21763 ± 10%  numa-vmstat.node1.nr_zone_write_pending
    137.92           +45.8%     201.12        filebench.sum_bytes_mb/s
   1650593           +45.9%    2407494        filebench.sum_operations
     27507           +45.9%      40122        filebench.sum_operations/s
      8873           +45.9%      12942        filebench.sum_reads/s
      3.61           -31.4%       2.48        filebench.sum_time_ms/op
    888.17           +45.8%       1295        filebench.sum_writes/s
    845794           +44.1%    1218541        filebench.time.file_system_outputs
     86085 ±  2%      +8.4%      93282 ±  4%  filebench.time.minor_page_faults
    482544           +26.9%     612318 ±  4%  filebench.time.voluntary_context_switches
     29.97 ±  9%      -5.7       24.28 ± 15%  perf-profile.calltrace.cycles-pp.perf_release.__fput.task_work_run.do_exit.do_group_exit
     29.97 ±  9%      -5.7       24.28 ± 15%  perf-profile.calltrace.cycles-pp.perf_event_release_kernel.perf_release.__fput.task_work_run.do_exit
      1.45 ± 27%      -1.1        0.37 ±102%  perf-profile.calltrace.cycles-pp.__d_lookup_rcu.lookup_fast.open_last_lookups.path_openat.do_filp_open
      0.26 ±141%      +0.9        1.15 ± 52%  perf-profile.calltrace.cycles-pp.proc_task_name.proc_pid_status.proc_single_show.seq_read_iter.seq_read
     29.97 ±  9%      -5.7       24.28 ± 15%  perf-profile.children.cycles-pp.perf_release
     29.97 ±  9%      -5.7       24.28 ± 15%  perf-profile.children.cycles-pp.perf_event_release_kernel
      1.84 ± 33%      -1.3        0.50 ±119%  perf-profile.children.cycles-pp.__d_lookup_rcu
      2.25 ± 28%      -1.3        0.94 ± 89%  perf-profile.children.cycles-pp.intel_idle_irq
      1.73 ± 39%      -1.2        0.48 ±116%  perf-profile.children.cycles-pp.__split_vma
      0.26 ±141%      +0.9        1.15 ± 52%  perf-profile.children.cycles-pp.proc_task_name
      1.84 ± 33%      -1.3        0.50 ±119%  perf-profile.self.cycles-pp.__d_lookup_rcu
      1.44 ± 40%      -0.9        0.51 ± 73%  perf-profile.self.cycles-pp.intel_idle_irq
      1.47 ± 54%      +1.1        2.57 ± 20%  perf-profile.self.cycles-pp._raw_spin_lock
     23688            +4.3%      24697        proc-vmstat.nr_active_anon
    211445           +44.1%     304619        proc-vmstat.nr_dirtied
     26739 ±  9%     +43.9%      38482 ±  6%  proc-vmstat.nr_dirty
    848497            +4.3%     884801        proc-vmstat.nr_file_pages
     81638 ±  3%     +42.8%     116595        proc-vmstat.nr_inactive_file
     42477            +3.1%      43810        proc-vmstat.nr_shmem
    175170 ±  3%     +35.8%     237820 ±  3%  proc-vmstat.nr_written
     23688            +4.3%      24697        proc-vmstat.nr_zone_active_anon
     81638 ±  3%     +42.8%     116595        proc-vmstat.nr_zone_inactive_file
     27275 ±  9%     +43.4%      39120 ±  6%  proc-vmstat.nr_zone_write_pending
    847032            +6.9%     905154        proc-vmstat.numa_hit
    711333            +8.4%     771356        proc-vmstat.numa_local
     61861            +2.4%      63333        proc-vmstat.pgactivate
   1439695           +11.1%    1599027        proc-vmstat.pgalloc_normal
   1249954 ±  2%     +12.4%    1405256        proc-vmstat.pgfree
    278966 ± 10%     +23.1%     343274 ±  7%  proc-vmstat.pgpgout
      0.34 ±  6%     -40.5%       0.20 ± 24%  perf-sched.sch_delay.avg.ms.futex_wait_queue.__futex_wait.futex_wait.do_futex
      0.04 ± 28%     -57.8%       0.02 ± 45%  perf-sched.sch_delay.avg.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
      0.02 ± 25%     -44.4%       0.01 ± 37%  perf-sched.sch_delay.avg.ms.schedule_timeout.wait_woken.sk_wait_data.tcp_recvmsg_locked
      1.13 ± 52%     -48.4%       0.58 ± 11%  perf-sched.sch_delay.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
    345.04 ±  6%     -20.3%     275.00 ± 14%  perf-sched.sch_delay.max.ms.futex_wait_queue.__futex_wait.futex_wait.do_futex
    211.37 ± 43%     -67.7%      68.37 ± 77%  perf-sched.sch_delay.max.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
    312.80 ±  9%     -26.3%     230.55 ± 17%  perf-sched.sch_delay.max.ms.wait_for_response.compound_send_recv.cifs_send_recv.query_info
     20.18 ±  2%     -13.0%      17.56 ±  7%  perf-sched.total_wait_and_delay.average.ms
     64102           +17.2%      75124 ±  5%  perf-sched.total_wait_and_delay.count.ms
     19.80 ±  2%     -12.7%      17.28 ±  7%  perf-sched.total_wait_time.average.ms
      0.17 ± 56%     -55.9%       0.07 ± 14%  perf-sched.wait_and_delay.avg.ms.__lock_sock.lock_sock_nested.tcp_recvmsg.inet6_recvmsg
      0.39 ±  6%     -38.4%       0.24 ± 20%  perf-sched.wait_and_delay.avg.ms.futex_wait_queue.__futex_wait.futex_wait.do_futex
      0.52 ±  6%     -14.8%       0.44 ±  2%  perf-sched.wait_and_delay.avg.ms.schedule_timeout.wait_woken.sk_wait_data.tcp_recvmsg_locked
    531.50 ± 11%     +94.8%       1035 ± 16%  perf-sched.wait_and_delay.count.__lock_sock.sk_wait_data.tcp_recvmsg_locked.tcp_recvmsg
    495.50 ± 11%     -55.8%     219.17 ±100%  perf-sched.wait_and_delay.count.cifs_wait_bit_killable.__wait_on_bit_lock.out_of_line_wait_on_bit_lock.cifs_revalidate_mapping
      9.67 ± 48%    +110.3%      20.33 ± 17%  perf-sched.wait_and_delay.count.io_schedule.rq_qos_wait.wbt_wait.__rq_qos_throttle
     34.83 ±  6%     +37.3%      47.83 ± 18%  perf-sched.wait_and_delay.count.schedule_hrtimeout_range_clock.do_poll.constprop.0.do_sys_poll
      9312 ±  5%     +14.9%      10702 ±  3%  perf-sched.wait_and_delay.count.schedule_timeout.wait_woken.sk_wait_data.tcp_recvmsg_locked
      4316           +46.7%       6333        perf-sched.wait_and_delay.count.wait_for_response.compound_send_recv.cifs_send_recv.query_info
    363.37 ± 10%     -23.8%     276.88 ± 13%  perf-sched.wait_and_delay.max.ms.futex_wait_queue.__futex_wait.futex_wait.do_futex
      1504 ± 14%     +33.1%       2003        perf-sched.wait_and_delay.max.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
      0.14 ± 36%     -50.8%       0.07 ± 15%  perf-sched.wait_time.avg.ms.__lock_sock.lock_sock_nested.tcp_recvmsg.inet6_recvmsg
      0.04 ±  8%     -21.4%       0.03 ±  5%  perf-sched.wait_time.avg.ms.futex_wait_queue.__futex_wait.futex_wait.do_futex
      0.02 ± 12%     +43.5%       0.02 ± 19%  perf-sched.wait_time.avg.ms.schedule_preempt_disabled.__mutex_lock.constprop.0.compound_send_recv
      0.50 ±  5%     -13.4%       0.43 ±  2%  perf-sched.wait_time.avg.ms.schedule_timeout.wait_woken.sk_wait_data.tcp_recvmsg_locked
      0.44 ± 38%     +94.3%       0.86 ± 30%  perf-sched.wait_time.max.ms.__lock_sock.lock_sock_nested.tcp_sock_set_cork.__smb_send_rqst
      0.28 ± 30%    +397.5%       1.41 ± 69%  perf-sched.wait_time.max.ms.schedule_preempt_disabled.__mutex_lock.constprop.0.compound_send_recv
      0.19 ±  2%     +24.1%       0.24 ± 11%  perf-stat.i.MPKI
 8.285e+09            -1.9%   8.13e+09        perf-stat.i.branch-instructions
   5693586 ±  2%     +21.0%    6890289        perf-stat.i.cache-misses
  43522169           +16.9%   50865574 ±  2%  perf-stat.i.cache-references
     24767           +21.1%      29984 ±  3%  perf-stat.i.context-switches
    252.79           +10.0%     278.09 ±  3%  perf-stat.i.cpu-migrations
     59997 ±  2%     -18.8%      48703 ±  2%  perf-stat.i.cycles-between-cache-misses
 8.121e+09            +2.1%  8.292e+09        perf-stat.i.dTLB-loads
  9.31e+08           +19.2%   1.11e+09        perf-stat.i.dTLB-stores
    300.38           +16.6%     350.14 ±  2%  perf-stat.i.metric.K/sec
   1118594 ±  2%     +20.8%    1351232        perf-stat.i.node-load-misses
    227414 ±  5%     +18.4%     269333 ±  3%  perf-stat.i.node-loads
    458563 ±  2%     +24.9%     572925        perf-stat.i.node-store-misses
   1007667 ±  4%     +21.4%    1222862 ±  2%  perf-stat.i.node-stores
      0.17 ±  2%     +21.9%       0.20        perf-stat.overall.MPKI
     56414 ±  2%     -18.4%      46044        perf-stat.overall.cycles-between-cache-misses
      0.02            -0.0        0.01 ±  2%  perf-stat.overall.dTLB-store-miss-rate%
 8.205e+09            -2.0%  8.042e+09        perf-stat.ps.branch-instructions
   5523801 ±  2%     +21.8%    6730244        perf-stat.ps.cache-misses
  41259525           +18.6%   48943300 ±  2%  perf-stat.ps.cache-references
     24329           +21.3%      29518 ±  3%  perf-stat.ps.context-switches
    240.48           +11.3%     267.76 ±  3%  perf-stat.ps.cpu-migrations
 8.038e+09            +2.0%  8.198e+09        perf-stat.ps.dTLB-loads
  9.15e+08           +19.5%  1.093e+09        perf-stat.ps.dTLB-stores
   1099551 ±  2%     +21.1%    1331752        perf-stat.ps.node-load-misses
    221785 ±  5%     +19.3%     264529 ±  3%  perf-stat.ps.node-loads
    448003 ±  2%     +25.6%     562873        perf-stat.ps.node-store-misses
    936984 ±  4%     +23.7%    1159370 ±  2%  perf-stat.ps.node-stores





Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.


-- 
0-DAY CI Kernel Test Service
https://github.com/intel/lkp-tests/wiki





[Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]

  Powered by Linux