Re: [PATCH v4] md: improve io stats accounting

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On 7/16/20 7:29 PM, Song Liu wrote:
> I just noticed another issue with this work on raid456, as iostat
> shows something
> like:
> 
> Device:         rrqm/s   wrqm/s     r/s     w/s    rMB/s    wMB/s
> avgrq-sz avgqu-sz   await r_await w_await  svctm  %util
> nvme0n1        6306.50 18248.00  636.00 1280.00    45.11    76.19
> 129.65     3.03    1.23    0.67    1.51   0.76 145.50
> nvme1n1       11441.50 13234.00 1069.50  961.00    71.87    55.39
> 128.35     3.32    1.30    0.90    1.75   0.72 146.50
> nvme2n1        8280.50 16352.50  971.50 1231.00    65.53    68.65
> 124.77     3.20    1.17    0.69    1.54   0.64 142.00
> nvme3n1        6158.50 18199.50  567.00 1453.50    39.81    76.74
> 118.13     3.50    1.40    0.88    1.60   0.73 146.50
> md0               0.00     0.00 1436.00 1411.00    89.75    88.19
> 128.00    22.98    8.07    0.16   16.12   0.52 147.00
> 
> md0 here is a RAID-6 array with 4 devices. %util of > 100% is clearly
> wrong here.
> This only doesn't happen to RAID-0 or RAID-1 in my tests.
> 
> Artur, could you please take a look at this?

Hi Song,

I think it's not caused by this patch, because %util of the member
drives is affected as well. I reverted the patch and it's still
happening:

Device            r/s     rMB/s   rrqm/s  %rrqm r_await rareq-sz     w/s     wMB/s   wrqm/s  %wrqm w_await wareq-sz     d/s     dMB/s   drqm/s  %drqm d_await dareq-sz     f/s f_await  aqu-sz  %util
md0             20.00      2.50     0.00   0.00    0.00   128.00   21.00      2.62     0.00   0.00    0.00   128.00    0.00      0.00     0.00   0.00    0.00     0.00    0.00    0.00    0.00   0.00
nvme0n1         13.00      1.62   279.00  95.55    0.77   128.00    4.00      0.50   372.00  98.94 1289.00   128.00    0.00      0.00     0.00   0.00    0.00     0.00    0.00    0.00    5.17 146.70
nvme1n1         15.00      1.88   310.00  95.38    0.53   128.00   21.00      2.62   341.00  94.20 1180.29   128.00    0.00      0.00     0.00   0.00    0.00     0.00    0.00    0.00   24.80 146.90
nvme2n1         16.00      2.00   310.00  95.09    0.69   128.00   19.00      2.38   341.00  94.72  832.89   128.00    0.00      0.00     0.00   0.00    0.00     0.00    0.00    0.00   15.84 146.80
nvme3n1         18.00      2.25   403.00  95.72    0.72   128.00   16.00      2.00   248.00  93.94  765.69   128.00    0.00      0.00     0.00   0.00    0.00     0.00    0.00    0.00   12.26 114.30

I was only able to reproduce it on a VM, it doesn't occur on real
hardware (for me). What was your test configuration?

Thanks,
Artur



[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux