Re: [lkp-robot] [MD] 0ffbb1adf8: aim7.jobs-per-min -10.6% regression

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi all

It's the first time I received such report. So I took some time reading
the manual of lkp and aim7. And I reserved one server with fedora and
did a test with the steps in this email. It failed like this:

2018-04-25 02:43:19 echo "/fs/md0" > config
2018-04-25 02:43:19 
	(
		echo storageqe-07.lab.bos.redhat.com
		echo sync_disk_rw

		echo 1
		echo 600
		echo 2
		echo 600
		echo 1
	) | ./multitask -t

AIM Multiuser Benchmark - Suite VII v1.1, January 22, 1996
Copyright (c) 1996 - 2001 Caldera International, Inc.
All Rights Reserved.

Machine's name                                              : Machine's configuration                                     : Number of iterations to run [1 to 10]                       : 
Information for iteration #1
Starting number of operation loads [1 to 10000]             : 1) Run to crossover
2) Run to specific operation load           Enter [1 or 2]: Maximum number of operation loads to simulate [600 to 10000]: Operation load increment [1 to 100]                         : 
Using disk directory </fs/md0>
HZ is <100>
AIM Multiuser Benchmark - Suite VII Run Beginning

Tasks    jobs/min  jti  jobs/min/task      real       cpu
  600/root/lkp-tests/bin/run-local:142:in `system': Interrupt
	from /root/lkp-tests/bin/run-local:142:in `<main>'

So now I can't understand these information from this report. 
What does "-10.6% regression of aim7.jobs-per-min" mean? And
what's the usage of aim7.jobs-per-min? Could anyone help to
give some suggestions? What should I do to resolve such problem?

Best Regards
Xiao

----- Original Message -----
> From: "kernel test robot" <xiaolong.ye@xxxxxxxxx>
> To: "Xiao Ni" <xni@xxxxxxxxxx>
> Cc: linux-raid@xxxxxxxxxxxxxxx, shli@xxxxxxxxxx, "ming lei" <ming.lei@xxxxxxxxxx>, ncroxon@xxxxxxxxxx,
> neilb@xxxxxxxx, lkp@xxxxxx
> Sent: Monday, April 23, 2018 8:41:43 AM
> Subject: [lkp-robot] [MD]  0ffbb1adf8:  aim7.jobs-per-min -10.6% regression
> 
> 
> Greeting,
> 
> FYI, we noticed a -10.6% regression of aim7.jobs-per-min due to commit:
> 
> 
> commit: 0ffbb1adf8b448568b44fe44c5fcdcf485040365 ("MD: fix lock contention
> for flush bios")
> url:
> https://github.com/0day-ci/linux/commits/Xiao-Ni/MD-fix-lock-contention-for-flush-bios/20180411-040300
> 
> 
> in testcase: aim7
> on test machine: 40 threads Intel(R) Xeon(R) CPU E5-2690 v2 @ 3.00GHz with
> 384G memory
> with following parameters:
> 
> 	disk: 4BRD_12G
> 	md: RAID1
> 	fs: xfs
> 	test: sync_disk_rw
> 	load: 600
> 	cpufreq_governor: performance
> 
> test-description: AIM7 is a traditional UNIX system level benchmark suite
> which is used to test and measure the performance of multiuser system.
> test-url: https://sourceforge.net/projects/aimbench/files/aim-suite7/
> 
> 
> 
> Details are as below:
> -------------------------------------------------------------------------------------------------->
> 
> 
> To reproduce:
> 
>         git clone https://github.com/intel/lkp-tests.git
>         cd lkp-tests
>         bin/lkp install job.yaml  # job file is attached in this email
>         bin/lkp run     job.yaml
> 
> =========================================================================================
> compiler/cpufreq_governor/disk/fs/kconfig/load/md/rootfs/tbox_group/test/testcase:
>   gcc-7/performance/4BRD_12G/xfs/x86_64-rhel-7.2/600/RAID1/debian-x86_64-2016-08-31.cgz/lkp-ivb-ep01/sync_disk_rw/aim7
> 
> commit:
>   v4.16
>   0ffbb1adf8 ("MD: fix lock contention for flush bios")
> 
>            v4.16 0ffbb1adf8b448568b44fe44c5
> ---------------- --------------------------
>          %stddev     %change         %stddev
>              \          |                \
>       1632 ±  2%     -10.6%       1458        aim7.jobs-per-min
>       2207 ±  2%     +11.8%       2468        aim7.time.elapsed_time
>       2207 ±  2%     +11.8%       2468        aim7.time.elapsed_time.max
>   51186515           -51.5%   24800655
>   aim7.time.involuntary_context_switches
>     146259 ±  8%     +31.7%     192669 ±  2%  aim7.time.minor_page_faults
>      80457 ±  2%     +15.9%      93267        aim7.time.system_time
>      50.25 ±  2%     +11.8%      56.17        aim7.time.user_time
>  7.257e+08 ±  2%      +7.3%  7.787e+08
>  aim7.time.voluntary_context_switches
>     520491 ± 61%     +53.8%     800775
>     interrupts.CAL:Function_call_interrupts
>       2463 ± 18%     +31.7%       3246 ± 15%  numa-vmstat.node0.nr_mapped
>       4.06 ±  2%      -0.5        3.51        mpstat.cpu.idle%
>       0.24 ±  6%      -0.1        0.15        mpstat.cpu.iowait%
>    8829533           +16.2%   10256984        softirqs.SCHED
>   33149109 ±  2%     +12.3%   37229216        softirqs.TIMER
>    4724795 ± 33%     +38.5%    6544104        cpuidle.C1E.usage
>  7.151e+08 ± 40%     -37.8%  4.449e+08        cpuidle.C6.time
>    3881055 ±122%     -85.7%     553608 ±  2%  cpuidle.C6.usage
>      61107 ±  2%     -10.7%      54566        vmstat.io.bo
>       2.60 ± 18%     -51.9%       1.25 ± 34%  vmstat.procs.b
>     305.10           -16.1%     256.00        vmstat.procs.r
>     404271           -11.3%     358644        vmstat.system.cs
>     167773           -16.5%     140121        vmstat.system.in
>     115358 ±  9%     +39.9%     161430 ±  3%  proc-vmstat.numa_hint_faults
>      62520 ± 10%     +47.6%      92267 ±  4%
>      proc-vmstat.numa_hint_faults_local
>      20893 ± 10%     +29.0%      26948 ±  2%  proc-vmstat.numa_pages_migrated
>     116983 ±  9%     +39.5%     163161 ±  3%  proc-vmstat.numa_pte_updates
>    5504935 ±  3%     +12.3%    6179561        proc-vmstat.pgfault
>      20893 ± 10%     +29.0%      26948 ±  2%  proc-vmstat.pgmigrate_success
>       2.68 ±  3%      -0.2        2.44        turbostat.C1%
>    4724733 ± 33%     +38.5%    6544028        turbostat.C1E
>    3879529 ±122%     -85.8%     552056 ±  2%  turbostat.C6
>       0.82 ± 43%      -0.4        0.45        turbostat.C6%
>       3.62 ±  2%     -15.1%       3.08        turbostat.CPU%c1
>     176728 ±  2%     +13.3%     200310        turbostat.SMI
>  9.893e+12 ± 65%     +57.6%  1.559e+13        perf-stat.branch-instructions
>  3.022e+10 ± 65%     +46.6%   4.43e+10        perf-stat.branch-misses
>      11.31 ± 65%      +5.5       16.78        perf-stat.cache-miss-rate%
>  2.821e+10 ± 65%     +50.4%  4.243e+10        perf-stat.cache-misses
>  1.796e+14 ± 65%     +58.4%  2.845e+14        perf-stat.cpu-cycles
>   26084177 ± 65%    +176.3%   72073346        perf-stat.cpu-migrations
>  1.015e+13 ± 65%     +57.3%  1.597e+13        perf-stat.dTLB-loads
>  1.125e+12 ± 65%     +45.6%  1.638e+12        perf-stat.dTLB-stores
>  4.048e+13 ± 65%     +57.3%  6.367e+13        perf-stat.instructions
>       4910 ± 65%     +57.5%       7734
>       perf-stat.instructions-per-iTLB-miss
>    3847673 ± 65%     +57.4%    6057388        perf-stat.minor-faults
>  1.403e+10 ± 65%     +51.8%   2.13e+10        perf-stat.node-load-misses
>  1.557e+10 ± 65%     +51.0%  2.351e+10        perf-stat.node-loads
>      27.41 ± 65%     +12.1       39.52        perf-stat.node-store-miss-rate%
>  7.828e+09 ± 65%     +53.2%  1.199e+10        perf-stat.node-store-misses
>  1.216e+10 ± 65%     +50.9%  1.835e+10        perf-stat.node-stores
>    3847675 ± 65%     +57.4%    6057392        perf-stat.page-faults
>    1041337 ±  2%     +12.0%    1166774
>    sched_debug.cfs_rq:/.exec_clock.avg
>    1045329 ±  2%     +11.8%    1168250
>    sched_debug.cfs_rq:/.exec_clock.max
>    1037380 ±  2%     +12.3%    1165349
>    sched_debug.cfs_rq:/.exec_clock.min
>       3670 ± 16%     -70.1%       1098 ± 50%
>       sched_debug.cfs_rq:/.exec_clock.stddev
>     234.08 ±  3%     -22.4%     181.73
>     sched_debug.cfs_rq:/.load_avg.avg
>      33.65 ±  2%     -59.3%      13.70 ±  3%
>      sched_debug.cfs_rq:/.load_avg.min
>   36305683 ±  2%     +12.4%   40814603
>   sched_debug.cfs_rq:/.min_vruntime.avg
>   37771587 ±  2%     +11.1%   41960294
>   sched_debug.cfs_rq:/.min_vruntime.max
>   34884765 ±  2%     +13.6%   39635549
>   sched_debug.cfs_rq:/.min_vruntime.min
>    1277146 ±  9%     -21.8%     998594 ±  4%
>    sched_debug.cfs_rq:/.min_vruntime.stddev
>       1.99 ±  5%     -16.8%       1.65 ±  3%
>       sched_debug.cfs_rq:/.nr_running.max
>      19.18 ±  5%     +20.3%      23.08
>      sched_debug.cfs_rq:/.nr_spread_over.avg
>       8.47 ± 18%     +23.8%      10.49 ± 12%
>       sched_debug.cfs_rq:/.nr_spread_over.min
>      22.06 ±  7%     -18.2%      18.05 ±  6%
>      sched_debug.cfs_rq:/.runnable_load_avg.avg
>       0.08 ± 38%     -50.5%       0.04 ± 20%  sched_debug.cfs_rq:/.spread.avg
>    1277131 ±  9%     -21.8%     998584 ±  4%
>    sched_debug.cfs_rq:/.spread0.stddev
>     177058 ±  7%     -23.5%     135429        sched_debug.cpu.avg_idle.max
>      33859 ±  6%     -24.6%      25519        sched_debug.cpu.avg_idle.stddev
>    1119705 ±  2%     +11.0%    1242378        sched_debug.cpu.clock.avg
>    1119726 ±  2%     +11.0%    1242399        sched_debug.cpu.clock.max
>    1119680 ±  2%     +11.0%    1242353        sched_debug.cpu.clock.min
>    1119705 ±  2%     +11.0%    1242378        sched_debug.cpu.clock_task.avg
>    1119726 ±  2%     +11.0%    1242399        sched_debug.cpu.clock_task.max
>    1119680 ±  2%     +11.0%    1242353        sched_debug.cpu.clock_task.min
>       3.47 ± 11%     +28.4%       4.45 ±  2%  sched_debug.cpu.cpu_load[1].min
>      28.92 ±  4%      -7.8%      26.66 ±  3%  sched_debug.cpu.cpu_load[2].avg
>       5.65 ±  7%     +37.9%       7.80 ±  8%  sched_debug.cpu.cpu_load[2].min
>      29.97 ±  3%      -7.9%      27.60 ±  3%  sched_debug.cpu.cpu_load[3].avg
>       8.22 ±  4%     +31.9%      10.83 ±  7%  sched_debug.cpu.cpu_load[3].min
>      10.50 ±  5%     +24.0%      13.02 ±  6%  sched_debug.cpu.cpu_load[4].min
>       2596 ±  4%      -9.2%       2358 ±  2%  sched_debug.cpu.curr->pid.avg
>       4463 ±  4%     -13.5%       3862 ±  3%
>       sched_debug.cpu.curr->pid.stddev
>    1139237 ±  2%     +11.5%    1269690
>    sched_debug.cpu.nr_load_updates.avg
>    1146166 ±  2%     +11.2%    1274264
>    sched_debug.cpu.nr_load_updates.max
>    1133061 ±  2%     +11.4%    1262298
>    sched_debug.cpu.nr_load_updates.min
>       3943 ±  9%     -32.4%       2666 ± 23%
>       sched_debug.cpu.nr_load_updates.stddev
>       8814 ±  7%     +17.8%      10386 ±  2%
>       sched_debug.cpu.nr_uninterruptible.max
>      -3613           +32.3%      -4782
>      sched_debug.cpu.nr_uninterruptible.min
>       2999 ±  4%     +13.1%       3391
>       sched_debug.cpu.nr_uninterruptible.stddev
>      42794 ± 15%     -76.8%       9921 ± 90%
>      sched_debug.cpu.sched_goidle.stddev
>     652177           +14.4%     745857        sched_debug.cpu.ttwu_local.avg
>     684397           +17.7%     805440 ±  2%  sched_debug.cpu.ttwu_local.max
>     622628           +12.0%     697353 ±  2%  sched_debug.cpu.ttwu_local.min
>      16189 ± 33%    +128.0%      36916 ± 47%
>      sched_debug.cpu.ttwu_local.stddev
>    1119677 ±  2%     +11.0%    1242351        sched_debug.cpu_clk
>    1119677 ±  2%     +11.0%    1242351        sched_debug.ktime
>    1120113 ±  2%     +11.0%    1242771        sched_debug.sched_clk
>                                                                                 
>                                                                                                                                                                 
>                                  aim7.jobs-per-min
>                                                                                 
>   1750 +-+------------------------------------------------------------------+
>        |                                                                    |   
>   1700 +-+                         +             .+.+                .+.+   |
>   1650 +-+                        + +          .+    :             .+    :  |
>        |  .+.+.+.    +      .+.+.+   +.       +      :    +.      +      :  |
>   1600 +-+       +  : + .+.+           +..+. +        +. +  +.+. +        +.|
>        |          + :  +                    +           +       +           |
>   1550 +-+         +                                                        |
>        |                                                                    |   
>   1500 +-+                                                                  |
>   1450 +-O     O     O               O O      O O                           |
>        O   O O     O   O     O O O O      O O     O O                       |
>   1400 +-+       O       O O                                                |
>        |                                                                    |   
>   1350 +-+------------------------------------------------------------------+
>                                                                                 
>                                                                                 
> [*] bisect-good sample
> [O] bisect-bad  sample
> 
> 
> 
> Disclaimer:
> Results have been estimated based on internal Intel analysis and are provided
> for informational purposes only. Any difference in system hardware or
> software
> design or configuration may affect actual performance.
> 
> 
> Thanks,
> Xiaolong
> 
--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html



[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux