RE: IOPS Log Incorrect?

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Yes!  This seems to have fixed the issue.  Thanks very much.  Test results on 2.99 pasted below.

nvmf@nvmfjt01:~/jrg9/chandler$ cat 4k-rr-01-01-128.ini |grep -v '#'
[global]
bs=4k
rw=randread
numjobs=1
iodepth=128
thread=1
ioengine=libaio
direct=1
ramp_time=10
time_based=0
norandommap=1
write_iops_log=joe1x128map
log_avg_msec=25000
[job00]
filename=/dev/nvme0n1


nvmf@nvmfjt01:~/jrg9/chandler$ sudo fio 4k-rr-01-01-128.ini
job00: (g=0): rw=randread, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128
fio-2.99
Starting 1 thread
job00: No I/O performed by libaio, perhaps try --debug=io option for details?1s]
job00: (groupid=0, jobs=1): err= 0: pid=12127: Mon Aug 21 15:27:02 2017
   read: IOPS=391k, BW=1526MiB/s (1600MB/s)(357GiB/239570msec)
    slat (nsec): min=905, max=186037, avg=1767.28, stdev=1210.88
    clat (usec): min=64, max=3207, avg=325.34, stdev=64.39
     lat (usec): min=65, max=3208, avg=327.22, stdev=64.46
    clat percentiles (usec):
     |  1.00th=[  231],  5.00th=[  251], 10.00th=[  262], 20.00th=[  277],
     | 30.00th=[  289], 40.00th=[  297], 50.00th=[  310], 60.00th=[  326],
     | 70.00th=[  343], 80.00th=[  367], 90.00th=[  408], 95.00th=[  449],
     | 99.00th=[  545], 99.50th=[  586], 99.90th=[  701], 99.95th=[  742],
     | 99.99th=[  840]
   bw (  MiB/s): min= 1086, max= 1641, per=100.00%, avg=1537.14, stdev=88.71, samples=479
   iops        : min=353807, max=411848, avg=389345.11, stdev=17570.39, samples=9
    lat (usec) : 100=0.01%, 250=4.85%, 500=93.10%, 750=2.01%, 1000=0.04%
    lat (msec) : 2=0.01%, 4=0.01%
  cpu          : usr=31.05%, sys=68.96%, ctx=716, majf=0, minf=2121
  IO depths    : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.1%, 32=0.1%, >=64=104.4%
     submit    : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0%
     complete  : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1%
     issued rwt: total=93588018,0,0, short=0,0,0, dropped=0,0,0
     latency   : target=0, window=0, percentile=100.00%, depth=128
Run status group 0 (all jobs):
   READ: bw=1526MiB/s (1600MB/s), 1526MiB/s-1526MiB/s (1600MB/s-1600MB/s), io=357GiB (383GB), run=239570-239570msec
Disk stats (read/write):
  nvme0n1: ios=97644549/0, merge=0/0, ticks=15111756/0, in_queue=15145884, util=100.00%


nvmf@nvmfjt01:~/jrg9/chandler$ cat joe1x128map_iops.1.log
24999, 386402, 0, 0
49997, 389145, 0, 0
74996, 382756, 0, 0
99995, 387607, 0, 0
124994, 381998, 0, 0
149992, 353807, 0, 0
174991, 398791, 0, 0
200000, 411848, 0, 0
225000, 411752, 0, 0



> -----Original Message-----
> From: Sitsofe Wheeler [mailto:sitsofe@xxxxxxxxx]
> Sent: Monday, August 21, 2017 3:16 PM
> To: Mikhail Terekhov <Mikhail.Terekhov@xxxxxxxx>
> Cc: Jeff Furlong <jeff.furlong@xxxxxxx>; Gruher, Joseph R
> <joseph.r.gruher@xxxxxxxxx>; neto from Brazil
> <Antonio.Jose.Rodrigues.Neto@xxxxxxxxxx>; fio@xxxxxxxxxxxxxxx
> Subject: Re: IOPS Log Incorrect?
> 
> Joseph,
> 
> Does the suggested fix solve the problem you were seeing too?
> 
> On 21 August 2017 at 23:10, Mikhail Terekhov <Mikhail.Terekhov@xxxxxxxx>
> wrote:
> > It seems fixed the problem. Here the results for fio-2.99 and
> > fio-2.99+patch for rotating media:
> >
> >  :/tmp>cat test299_iops.1.log
> > 5000, 1448, 0, 0
> > 10641, 180, 0, 0
> > 15576, 160, 0, 0
> > 20511, 204, 0, 0
> > 25446, 222, 0, 0
> > 30381, 126, 0, 0
> > :/tmp>cat test299p1_iops.1.log
> > 5000, 1746, 0, 0
> > 10000, 1435, 0, 0
> > 15000, 1478, 0, 0
> > 20000, 1161, 0, 0
> > 25000, 1291, 0, 0
> > 30000, 1265, 0, 0
> >
> > On 08/21/17 17:12, Sitsofe Wheeler wrote:
> >>
> >> Hmm. How do things look with the following?
> >>
> >> diff --git a/time.c b/time.c
> >> index edfe779..5e38aad 100644
> >> --- a/time.c
> >> +++ b/time.c
> >> @@ -8,7 +8,7 @@ static unsigned long ns_granularity;
> >>
> >>   void timespec_add_msec(struct timespec *ts, unsigned int msec)
> >>   {
> >> -       unsigned long adj_nsec = 1000000 * msec;
> >> +       unsigned long adj_nsec = 1000000UL * msec;
> >>
> >>          ts->tv_nsec += adj_nsec;
> >>          if (adj_nsec >= 1000000000) {
> >>
> >> Perhaps because msec is an unsigned int and 1000000 fits in an
> >> unsigned int the multiplication will wrap unexpectedly even though
> >> the real answer fits in a (64 bit) unsigned long? Perhaps  adj_nsec
> >> needs to be a uint64_t ?
> >>
> >> On 21 August 2017 at 20:33, Jeff Furlong <jeff.furlong@xxxxxxx> wrote:
> >>>
> >>> Looks like when we set log_avg_msec=4000 or less it works as expected.
> >>> Also, the logged timestamps are very close to 4000 multiples.  When
> >>> we set
> >>> log_avg_msec=5000 or more the average is wrong, and also the logged
> >>> timestamp has significant jitter from 5000 multiples.  Are we
> >>> overflowing the log_avg_msec variable (32bits)?
> >>>
> >>> Regards,
> >>> Jeff
> >>>
> >>> -----Original Message-----
> >>> From: fio-owner@xxxxxxxxxxxxxxx [mailto:fio-owner@xxxxxxxxxxxxxxx]
> >>> On Behalf Of Gruher, Joseph R
> >>> Sent: Monday, August 21, 2017 11:11 AM
> >>> To: neto from Brazil <Antonio.Jose.Rodrigues.Neto@xxxxxxxxxx>;
> >>> fio@xxxxxxxxxxxxxxx
> >>> Subject: RE: IOPS Log Incorrect?
> >>>
> >>> Feature seems to work in FIO 2.15 through 2.20.  I couldn't test
> >>> 2.21 (keeps segfaulting when I try to run after compile) and then
> >>> the feature is broken in the next release which is 2.99 and still broken in
> 3.0.
> >>>
> >>>> -----Original Message-----
> >>>> From: Gruher, Joseph R
> >>>> Sent: Thursday, August 17, 2017 5:16 PM
> >>>> To: neto from Brazil <Antonio.Jose.Rodrigues.Neto@xxxxxxxxxx>;
> >>>> fio@xxxxxxxxxxxxxxx
> >>>> Subject: RE: IOPS Log Incorrect?
> >>>>
> >>>> This feature seems to work in FIO 2.15:
> >>>>
> >>>> root@nvmfjt01:/home/nvmf/jrg9/chandler# fio 4k-rr-01-01-128.ini
> >>>> job00: (g=0): rw=randread, bs=4K-4K/4K-4K/4K-4K, ioengine=libaio,
> >>>> iodepth=128
> >>>> fio-2.15
> >>>> Starting 1 thread
> >>>> Jobs: 1 (f=1): [r(1)] [78.8% done] [1804MB/0KB/0KB /s] [462K/0/0
> >>>> iops] [eta 00m:58s]
> >>>> job00: (groupid=0, jobs=1): err= 0: pid=5341: Thu Aug 17 16:57:40 2017
> >>>>    read : io=277967MB, bw=1789.6MB/s, iops=457984, runt=155375msec
> >>>>      slat (usec): min=0, max=95, avg= 1.44, stdev= 0.91
> >>>>      clat (usec): min=61, max=5168, avg=277.60, stdev=118.64
> >>>>       lat (usec): min=62, max=5170, avg=279.12, stdev=118.63
> >>>>      clat percentiles (usec):
> >>>>       |  1.00th=[  129],  5.00th=[  155], 10.00th=[  173],
> >>>> 20.00th=[ 197],
> >>>>       | 30.00th=[  215], 40.00th=[  235], 50.00th=[  255],
> >>>> 60.00th=[ 278],
> >>>>       | 70.00th=[  306], 80.00th=[  342], 90.00th=[  406],
> >>>> 95.00th=[ 470],
> >>>>       | 99.00th=[  636], 99.50th=[  716], 99.90th=[ 1176],
> >>>> 99.95th=[ 2096],
> >>>>       | 99.99th=[ 2896]
> >>>>      lat (usec) : 100=0.12%, 250=47.54%, 500=48.64%, 750=3.31%,
> >>>> 1000=0.26%
> >>>>      lat (msec) : 2=0.08%, 4=0.05%, 10=0.01%
> >>>>    cpu          : usr=28.43%, sys=68.46%, ctx=736888, majf=0, minf=816
> >>>>    IO depths    : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.1%, 32=0.1%,
> >>>>>
> >>>>> =64=137.3%
> >>>>
> >>>>       submit    : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%,
> >>>>>
> >>>>> =64=0.0%
> >>>>
> >>>>       complete  : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%,
> >>>> 64=0.0%,
> >>>>>
> >>>>> =64=0.1%
> >>>>
> >>>>       issued    : total=r=71159403/w=0/d=0, short=r=0/w=0/d=0,
> >>>> drop=r=0/w=0/d=0
> >>>>       latency   : target=0, window=0, percentile=100.00%, depth=128
> >>>> Run status group 0 (all jobs):
> >>>>     READ: io=277967MB, aggrb=1789.6MB/s, minb=1789.6MB/s,
> >>>> maxb=1789.6MB/s, mint=155375msec, maxt=155375msec Disk stats
> >>>> (read/write):
> >>>>    nvme0n1: ios=97645100/0, merge=0/0, ticks=24504996/0,
> >>>> in_queue=24519280, util=99.95%
> >>>>
> >>>> root@nvmfjt01:/home/nvmf/jrg9/chandler# cat joe1x128_iops.1.log
> >>>> 20001, 438712, 0, 0 39999, 455544, 0, 0 59999, 457418, 0, 0 79999,
> >>>> 462500, 0,
> >>>> 0 99999, 462728, 0, 0 119999, 462893, 0, 0 139999, 462508, 0, 0
> >>>>
> >>>> root@nvmfjt01:/home/nvmf/jrg9/chandler# cat 4k-rr-01-01-128.ini
> >>>> [global] bs=4k rw=randread
> >>>> numjobs=1
> >>>> iodepth=128
> >>>> thread=1
> >>>> ioengine=libaio
> >>>> direct=1
> >>>> ramp_time=60
> >>>> norandommap=1
> >>>> group_reporting=1
> >>>> write_iops_log=joe1x128
> >>>> log_avg_msec=20000
> >>>> [job00]
> >>>> filename=/dev/nvme0n1
> >>>>
> >>>> root@nvmfjt01:/home/nvmf/jrg9/chandler# uname -a Linux nvmfjt01
> >>>> 4.12.7- 041207-generic #201708160856 SMP Wed Aug 16 12:59:25 UTC
> >>>> 2017
> >>>> x86_64
> >>>> x86_64 x86_64 GNU/Linux
> >>>>
> >>>> root@nvmfjt01:/home/nvmf/jrg9/chandler# fio --version
> >>>> fio-2.15
> >>>>
> >>>>> -----Original Message-----
> >>>>> From: Gruher, Joseph R
> >>>>> Sent: Thursday, August 17, 2017 11:52 AM
> >>>>> To: neto from Brazil <Antonio.Jose.Rodrigues.Neto@xxxxxxxxxx>;
> >>>>> fio@xxxxxxxxxxxxxxx
> >>>>> Subject: RE: IOPS Log Incorrect?
> >>>>>
> >>>>> Sure.  That produces a very large log file where every entry is
> >>>>> <timestamp>, 1, 0, 4096.  See example below.
> >>>>>
> >>>>>
> >>>>> root@nvmfjt01:/home/nvmf/jrg9/chandler# cat 4k-rr-01.ini [global]
> >>>>> bs=4k rw=randread
> >>>>> numjobs=1
> >>>>> iodepth=128
> >>>>> ioengine=libaio
> >>>>> direct=1
> >>>>> runtime=3
> >>>>> write_iops_log=joe2
> >>>>>   [job00]
> >>>>> filename=/dev/nvme0n1
> >>>>>
> >>>>>
> >>>>> root@nvmfjt01:/home/nvmf/jrg9/chandler# fio 4k-rr-01.ini
> >>>>> job00: (g=0): rw=randread, bs=(R) 4096B-4096B, (W) 4096B-4096B,
> >>>>> (T)
> >>>>> 4096B- 4096B, ioengine=libaio, iodepth=128
> >>>>> fio-3.0
> >>>>> Starting 1 process
> >>>>> Jobs: 1 (f=1): [r(1)][100.0%][r=1330MiB/s,w=0KiB/s][r=340k,w=0
> >>>>> IOPS][eta 00m:00s]
> >>>>> job00: (groupid=0, jobs=1): err= 0: pid=3124: Thu Aug 17 11:49:47 2017
> >>>>>     read: IOPS=317k, BW=1238MiB/s (1298MB/s)(3715MiB/3001msec)
> >>>>>      slat (nsec): min=930, max=114792, avg=2418.97, stdev=1970.30
> >>>>>      clat (usec): min=81, max=1210, avg=400.22, stdev=74.26
> >>>>>       lat (usec): min=84, max=1220, avg=402.75, stdev=74.53
> >>>>>      clat percentiles (usec):
> >>>>>       |  1.00th=[  310],  5.00th=[  326], 10.00th=[  334],
> >>>>> 20.00th=[ 343],
> >>>>>       | 30.00th=[  355], 40.00th=[  363], 50.00th=[  375],
> >>>>> 60.00th=[ 392],
> >>>>>       | 70.00th=[  420], 80.00th=[  461], 90.00th=[  494],
> >>>>> 95.00th=[ 529],
> >>>>>       | 99.00th=[  660], 99.50th=[  734], 99.90th=[  857],
> >>>>> 99.95th=[ 889],
> >>>>>       | 99.99th=[ 1004]
> >>>>>     bw (  MiB/s): min=  960, max= 1332, per=98.62%, avg=1220.89,
> >>>>> stdev=160.84,
> >>>>> samples=5
> >>>>>     iops        : min=    1, max=    1, avg= 1.00, stdev= 0.00,
> >>>>> samples=951085
> >>>>>    lat (usec)   : 100=0.01%, 250=0.05%, 500=91.24%, 750=8.28%,
> >>>>> 1000=0.42%
> >>>>>    lat (msec)   : 2=0.01%
> >>>>>    cpu          : usr=22.13%, sys=77.87%, ctx=61, majf=0, minf=3234
> >>>>>    IO depths    : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.1%, 32=0.1%,
> >>>>> =64=99.9%
> >>>>>       submit    : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%,
> >>>>>>
> >>>>>> =64=0.0%
> >>>>>
> >>>>>       complete  : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%,
> >>>>> 64=0.0%,
> >>>>>>
> >>>>>> =64=0.1%
> >>>>>
> >>>>>       issued rwt: total=951085,0,0, short=0,0,0, dropped=0,0,0
> >>>>>       latency   : target=0, window=0, percentile=100.00%, depth=128
> >>>>> Run status group 0 (all jobs):
> >>>>>     READ: bw=1238MiB/s (1298MB/s), 1238MiB/s-1238MiB/s (1298MB/s-
> >>>>> 1298MB/s), io=3715MiB (3896MB), run=3001-3001msec Disk stats
> >>>>> (read/write):
> >>>>>    nvme0n1: ios=951085/0, merge=0/0, ticks=121124/0,
> >>>>> in_queue=120984, util=92.50%
> >>>>>
> >>>>>
> >>>>> root@nvmfjt01:/home/nvmf/jrg9/chandler# head -n 10 joe2_iops.1.log
> >>>>> 2, 1, 0,
> >>>>> 4096 2, 1, 0, 4096 2, 1, 0, 4096 2, 1, 0, 4096 2, 1, 0, 4096 2, 1,
> >>>>> 0,
> >>>>> 4096 2, 1, 0,
> >>>>> 4096 2, 1, 0, 4096 2, 1, 0, 4096 2, 1, 0, 4096
> >>>>>
> >>>>>
> >>>>> root@nvmfjt01:/home/nvmf/jrg9/chandler# tail -n 10 joe2_iops.1.log
> >>>>> 3000, 1, 0, 4096 3000, 1, 0, 4096 3000, 1, 0, 4096 3000, 1, 0,
> >>>>> 4096 3000, 1, 0, 4096 3000, 1, 0, 4096 3000, 1, 0, 4096 3000, 1,
> >>>>> 0, 4096 3000, 1, 0, 4096 3000, 1, 0,
> >>>>> 4096
> >>>>>
> >>>>>
> >>>>> Thanks,
> >>>>> Joe
> >>>>>
> >>>>>> -----Original Message-----
> >>>>>> From: neto from Brazil
> >>>>>> [mailto:Antonio.Jose.Rodrigues.Neto@xxxxxxxxxx]
> >>>>>> Sent: Thursday, August 17, 2017 11:47 AM
> >>>>>> To: Gruher, Joseph R <joseph.r.gruher@xxxxxxxxx>;
> >>>>>> fio@xxxxxxxxxxxxxxx
> >>>>>> Subject: Re: IOPS Log Incorrect?
> >>>>>>
> >>>>>> Hi Joseph,
> >>>>>>
> >>>>>> This is neto from Brazil
> >>>>>>
> >>>>>> How are you?
> >>>>>>
> >>>>>> Could you try to re-run without log_avg_msec=5000?
> >>>>>>
> >>>>>> Thanks,
> >>>>>>
> >>>>>> neto
> >>>>>>
> >>>>>> On 8/17/17, 2:38 PM, "fio-owner@xxxxxxxxxxxxxxx on behalf of
> >>>>>> Gruher, Joseph R" <fio-owner@xxxxxxxxxxxxxxx on behalf of
> >>>>>
> >>>>> joseph.r.gruher@xxxxxxxxx> wrote:
> >>>>>>
> >>>>>>      Hi, I'm using FIO 3.0 on Ubuntu 16.10 with 4.12.7 kernel
> >>>>>> (also reproduced problem with default 4.8.0-59 kernel). I'm
> >>>>>> running a 4KB random IO pattern against an Intel P3700 400GB NVMe
> >>>>>> SSD. I'm trying to view the IOPS/BW over time using the
> >>>>>> write_bw_log and write_iops_log parameters. FIO is writing the
> >>>>>> logs but the values in the logs don't align with the summary for
> >>>>>> the run.  Here's a simple example.  Notice FIO reports 398K IOPS
> >>>>>> for the run overall, but the IOPS log file shows 383K for the
> >>>>>> first entry and then values around 56K for the rest of the run,
> >>>>>> so
> >>>>>
> >>>>> one or the other must be incorrect (presumably the log file).
> >>>>>>
> >>>>>> The bandwidth log has a similar issue.
> >>>>>>
> >>>>>>
> >>>>>>      root@nvmfjt01:/home/nvmf/jrg9/chandler# cat 4k-rr-01.ini
> >>>>>>      [global]
> >>>>>>      bs=4k
> >>>>>>      rw=randread
> >>>>>>      numjobs=1
> >>>>>>      iodepth=128
> >>>>>>      ioengine=libaio
> >>>>>>      direct=1
> >>>>>>      runtime=30
> >>>>>>      write_iops_log=joe2
> >>>>>>      log_avg_msec=5000
> >>>>>>      [job00]
> >>>>>>      filename=/dev/nvme0n1
> >>>>>>
> >>>>>>
> >>>>>>      root@nvmfjt01:/home/nvmf/jrg9/chandler# fio 4k-rr-01.ini
> >>>>>>      job00: (g=0): rw=randread, bs=(R) 4096B-4096B, (W)
> >>>>>> 4096B-4096B,
> >>>>>> (T) 4096B-4096B, ioengine=libaio, iodepth=128
> >>>>>>      fio-3.0
> >>>>>>      Starting 1 process
> >>>>>>      Jobs: 1 (f=1):
> >>>>>> [r(1)][100.0%][r=1595MiB/s,w=0KiB/s][r=408k,w=0
> >>>>>> IOPS][eta 00m:00s]
> >>>>>>      job00: (groupid=0, jobs=1): err= 0: pid=2767: Thu Aug 17
> >>>>>> 10:34:09
> >>>>>> 2017
> >>>>>>      read: IOPS=398k, BW=1555MiB/s (1630MB/s)(45.6GiB/30001msec)
> >>>>>>      slat (nsec): min=824, max=117534, avg=1682.36, stdev=1290.45
> >>>>>>      clat (usec): min=96, max=2899, avg=319.29, stdev=69.36
> >>>>>>      lat (usec): min=97, max=2900, avg=321.08, stdev=69.39
> >>>>>>      clat percentiles (usec):
> >>>>>>      | 1.00th=[ 221], 5.00th=[ 243], 10.00th=[ 255], 20.00th=[ 269],
> >>>>>>      | 30.00th=[ 281], 40.00th=[ 289], 50.00th=[ 302], 60.00th=[ 318],
> >>>>>>      | 70.00th=[ 334], 80.00th=[ 363], 90.00th=[ 412], 95.00th=[ 453],
> >>>>>>      | 99.00th=[ 562], 99.50th=[ 611], 99.90th=[ 725], 99.95th=[ 775],
> >>>>>>      | 99.99th=[ 906]
> >>>>>>      bw ( MiB/s): min= 1081, max= 1594, per=99.96%, avg=1554.11,
> >>>>>> stdev=68.33,
> >>>>>> samples=59
> >>>>>>      iops : min=56293, max=383046, avg=65603.75, stdev=54419.77,
> >>>>>
> >>>>> samples=36
> >>>>>>
> >>>>>>      lat (usec) : 100=0.01%, 250=7.73%, 500=89.82%, 750=2.38%,
> >>>>>> 1000=0.07%
> >>>>>>      lat (msec) : 2=0.01%, 4=0.01%
> >>>>>>      cpu : usr=32.51%, sys=67.50%, ctx=105, majf=0, minf=153
> >>>>>>      IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.1%,
> >>>>>> 32=0.1%,
> >>>>>>>
> >>>>>>> =64=100.0%
> >>>>>>
> >>>>>>      submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%,
> >>>>>> 64=0.0%, =64=0.0%
> >>>>>>      complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%,
> >>>>>> 64=0.0%,
> >>>>>>>
> >>>>>>> =64=0.1%
> >>>>>>
> >>>>>>      issued rwt: total=11941208,0,0, short=0,0,0, dropped=0,0,0
> >>>>>>      latency : target=0, window=0, percentile=100.00%, depth=128
> >>>>>>      Run status group 0 (all jobs):
> >>>>>>      READ: bw=1555MiB/s (1630MB/s), 1555MiB/s-1555MiB/s
> >>>>>> (1630MB/s- 1630MB/s), io=45.6GiB (48.9GB), run=30001-30001msec
> >>>>>>      Disk stats (read/write):
> >>>>>>      nvme0n1: ios=11890654/0, merge=0/0, ticks=1923076/0,
> >>>>>> in_queue=1925204, util=99.69%
> >>>>>>
> >>>>>>
> >>>>>>      root@nvmfjt01:/home/nvmf/jrg9/chandler# cat joe2_iops.1.log
> >>>>>>      5000, 383046, 0, 0
> >>>>>>      10640, 56412, 0, 0
> >>>>>>      15576, 56410, 0, 0
> >>>>>>      20511, 56425, 0, 0
> >>>>>>      25446, 56394, 0, 0
> >>>>>>      30040, 57131, 0, 0
> >>>>>>
> >>>>>>
> >>>>>>      root@nvmfjt01:/home/nvmf/jrg9/chandler# fio --version
> >>>>>>      fio-3.0
> >>>>>>
> >>>>>>
> >>>>>>      root@nvmfjt01:/home/nvmf/jrg9/chandler# uname -a
> >>>>>>      Linux nvmfjt01 4.12.7-041207-generic #201708160856 SMP Wed
> >>>>>> Aug
> >>>>>> 16
> >>>>>> 12:59:25 UTC 2017 x86_64 x86_64 x86_64 GNU/Linux
> 
> --
> Sitsofe | http://sucs.org/~sits/
��.n��������+%������w��{.n�������^n�r������&��z�ޗ�zf���h���~����������_��+v���)ߣ�

[Index of Archives]     [Linux Kernel]     [Linux SCSI]     [Linux IDE]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux SCSI]

  Powered by Linux