Re: [PATCH v4] md: improve io stats accounting

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Fri, Jul 17, 2020 at 3:44 AM Artur Paszkiewicz
<artur.paszkiewicz@xxxxxxxxx> wrote:
>
> On 7/16/20 7:29 PM, Song Liu wrote:
> > I just noticed another issue with this work on raid456, as iostat
> > shows something
> > like:
> >
> > Device:         rrqm/s   wrqm/s     r/s     w/s    rMB/s    wMB/s
> > avgrq-sz avgqu-sz   await r_await w_await  svctm  %util
> > nvme0n1        6306.50 18248.00  636.00 1280.00    45.11    76.19
> > 129.65     3.03    1.23    0.67    1.51   0.76 145.50
> > nvme1n1       11441.50 13234.00 1069.50  961.00    71.87    55.39
> > 128.35     3.32    1.30    0.90    1.75   0.72 146.50
> > nvme2n1        8280.50 16352.50  971.50 1231.00    65.53    68.65
> > 124.77     3.20    1.17    0.69    1.54   0.64 142.00
> > nvme3n1        6158.50 18199.50  567.00 1453.50    39.81    76.74
> > 118.13     3.50    1.40    0.88    1.60   0.73 146.50
> > md0               0.00     0.00 1436.00 1411.00    89.75    88.19
> > 128.00    22.98    8.07    0.16   16.12   0.52 147.00
> >
> > md0 here is a RAID-6 array with 4 devices. %util of > 100% is clearly
> > wrong here.
> > This only doesn't happen to RAID-0 or RAID-1 in my tests.
> >
> > Artur, could you please take a look at this?
>
> Hi Song,
>
> I think it's not caused by this patch, because %util of the member
> drives is affected as well. I reverted the patch and it's still
> happening:
>
> Device            r/s     rMB/s   rrqm/s  %rrqm r_await rareq-sz     w/s     wMB/s   wrqm/s  %wrqm w_await wareq-sz     d/s     dMB/s   drqm/s  %drqm d_await dareq-sz     f/s f_await  aqu-sz  %util
> md0             20.00      2.50     0.00   0.00    0.00   128.00   21.00      2.62     0.00   0.00    0.00   128.00    0.00      0.00     0.00   0.00    0.00     0.00    0.00    0.00    0.00   0.00
> nvme0n1         13.00      1.62   279.00  95.55    0.77   128.00    4.00      0.50   372.00  98.94 1289.00   128.00    0.00      0.00     0.00   0.00    0.00     0.00    0.00    0.00    5.17 146.70
> nvme1n1         15.00      1.88   310.00  95.38    0.53   128.00   21.00      2.62   341.00  94.20 1180.29   128.00    0.00      0.00     0.00   0.00    0.00     0.00    0.00    0.00   24.80 146.90
> nvme2n1         16.00      2.00   310.00  95.09    0.69   128.00   19.00      2.38   341.00  94.72  832.89   128.00    0.00      0.00     0.00   0.00    0.00     0.00    0.00    0.00   15.84 146.80
> nvme3n1         18.00      2.25   403.00  95.72    0.72   128.00   16.00      2.00   248.00  93.94  765.69   128.00    0.00      0.00     0.00   0.00    0.00     0.00    0.00    0.00   12.26 114.30
>
> I was only able to reproduce it on a VM, it doesn't occur on real
> hardware (for me). What was your test configuration?

I was testing on VM. But I didn't see this issue after reverting the
patch. Let me
test more.

Thanks,
Song


> Thanks,
> Artur



[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux