On 10/11/14, 12:30 PM, "Jens Axboe" <axboe@xxxxxxxxx> wrote: >On 2014-10-11 10:28, Jens Axboe wrote: >> On 2014-10-10 07:32, Neto, Antonio Jose Rodrigues wrote: >>> >>> >>> On 10/8/14, 10:52 AM, "Jens Axboe" <axboe@xxxxxxxxx> wrote: >>> >>>> On 10/08/2014 08:47 AM, Neto, Antonio Jose Rodrigues wrote: >>>>> >>>>> >>>>> On 10/8/14, 10:33 AM, "Jens Axboe" <axboe@xxxxxxxxx> wrote: >>>>> >>>>>> On 10/08/2014 08:13 AM, Neto, Antonio Jose Rodrigues wrote: >>>>>>> >>>>>>> >>>>>>> On 10/8/14, 12:03 AM, "Jens Axboe" <axboe@xxxxxxxxx> wrote: >>>>>>> >>>>>>>> On 2014-10-07 21:24, Neto, Antonio Jose Rodrigues wrote: >>>>>>>>> Nossa Senhora:fio neto$ ./fio --client 10.61.109.151 >>>>>>>>> --remote-config >>>>>>>>> /root/fio.patch/fio/model >>>>>>>>> hostname=s1, be=0, 64-bit, os=Linux, arch=x86-64, >>>>>>>>> fio=fio-2.1.13-42-g3232, >>>>>>>>> flags=1 >>>>>>>>> <s1> fio: unable to open '/root/fio.patch/fio/model:70?' job file >>>>>>>>> client: host=10.61.109.151 disconnected >>>>>>>>> >>>>>>>>> Any ideas? >>>>>>>> >>>>>>>> Looks like I just forgot to zero terminate that string. It was >>>>>>>>never >>>>>>>> absolute or relative path, just luck and what was in memory. Try >>>>>>>>and >>>>>>>> pull again, I committed a fix for that. >>>>>>>> >>>>>>>> -- >>>>>>>> Jens Axboe >>>>>>>> >>>>>>>> -- >>>>>>> >>>>>>> >>>>>>> Hi Jens, >>>>>>> >>>>>>> This is neto from Brazil >>>>>>> >>>>>>> How are you? >>>>>>> >>>>>>> Seems to me it's working with absolute path now with the latest >>>>>>> commit >>>>>>> to >>>>>>> remote-config branch. >>>>>> >>>>>> Great, I verified this morning that it was an issue, we'd be >>>>>> looking at >>>>>> unitialized/allocated memory without it. >>>>>> >>>>>>> But, running the workload from my mac (connected to 2 Linux >>>>>>> clients) I >>>>>>> do >>>>>>> not see the progress. >>>>>>> >>>>>>> Nossa Senhora:fiop neto$ ./fio --client 10.61.109.151 >>>>>>>--remote-config >>>>>>> /root/fiop/model --client 10.61.109.152 --remote-config >>>>>>> /root/fiop/model >>>>>>> hostname=s2, be=0, 64-bit, os=Linux, arch=x86-64, fio=fio-2.1.13, >>>>>>> flags=1 >>>>>>> hostname=s1, be=0, 64-bit, os=Linux, arch=x86-64, >>>>>>> fio=fio-2.1.13-31-g15e3, >>>>>>> flags=1 >>>>>>> <s2> workload: (g=0): rw=read, <s1> workload: (g=0): rw=read, >>>>>>> bs=64K-64K/64K-64K/64K-64K, bs=64K-64K/64K-64K/64K-64K, >>>>>>> ioengine=libaio, >>>>>>> iodepth=1 >>>>>>> ioengine=libaio, iodepth=1 >>>>>>> <s2> ... >>>>>>> <s1> ... >>>>>>> <s1> Starting <s2> Starting 128 threads >>>>>>> 128 threads >>>>>>> Jobs: 0 (f=0) >>>>>>> >>>>>>> Any idea why? >>>>>> >>>>>> Works for me, just tried it from an OSX client. I notice that you >>>>>> don't >>>>>> seem to have updated the 's2' fio version, however. So I'd suggest >>>>>>you >>>>>> ensure you are running the same thing on all of them. >>>>>> >>>>>> -- >>>>>> Jens Axboe >>>>>> >>>>> >>>>> >>>>> Hi Jens, >>>>> >>>>> This is neto from Brazil >>>>> >>>>> How are you? >>>>> >>>>> With one client and one server it works >>>>> >>>>> Nossa Senhora:fiop neto$ ./fio --client 10.61.109.151 --remote-config >>>>> /root/fiop/model >>>>> hostname=s1, be=0, 64-bit, os=Linux, arch=x86-64, >>>>> fio=fio-2.1.13-31-g15e3, >>>>> flags=1 >>>>> <s1> workload: (g=0): rw=read, bs=64K-64K/64K-64K/64K-64K, >>>>> ioengine=libaio, iodepth=1 >>>>> <s1> ... >>>>> <s1> Starting 128 threads >>>>> Jobs: 128 (f=2048): [R(128)] [4.4% done] [1770M/0K/0K /s] [27.7K/0/0 >>>>> iops] >>>>> [eta 09m:45s] >>>>> >>>>> >>>>> >>>>> >>>>> >>>>> But with one client and 2 servers it does not work (the progress) >>>>> >>>>> >>>>> Nossa Senhora:fiop neto$ ./fio --client 10.61.109.151 --remote-config >>>>> /root/fiop/model --client 10.61.109.152 --remote-config >>>>> /root/fiop/model >>>>> hostname=s2, be=0, 64-bit, os=Linux, arch=x86-64, >>>>> fio=fio-2.1.13-31-g15e3, >>>>> flags=1 >>>>> hostname=s1, be=0, 64-bit, os=Linux, arch=x86-64, >>>>> fio=fio-2.1.13-31-g15e3, >>>>> flags=1 >>>>> <s2> workload: (g=0): rw=read, <s1> workload: (g=0): rw=read, >>>>> bs=64K-64K/64K-64K/64K-64K, bs=64K-64K/64K-64K/64K-64K, >>>>> ioengine=libaio, >>>>> iodepth=1 >>>>> ioengine=libaio, iodepth=1 >>>>> <s2> ... >>>>> <s1> ... >>>>> <s2> Starting <s1> Starting 128 threads128 threads >>>>> >>>>> Jobs: 0 (f=0) >>>>> Jobs: 0 (f=0) >>>> >>>> Weird, tested two here, running different jobs, and it summed them up >>>> fine and reported the ETA line. I will take a look, when time permits. >>>> >>>> -- >>>> Jens Axboe >>> >>> >>> Hi Jens, >>> >>> This is neto from Brazil >>> >>> How are you? >>> >>> Just a quick update to help you with the troubleshooting. >>> >>> From latest commit ... >>> >>> When I start the fio (from my mac to use 2 Linux servers) >>> >>> When the job starts, I do not see anything on the screen only this: >>> >>> Nossa Senhora:fio neto$ ./fio --client 10.61.109.151 --remote-config >>> /root/fio/write --client 10.61.109.152 --remote-config /root/fio/write >>> hostname=s2, be=0, 64-bit, os=Linux, arch=x86-64, >>> fio=fio-2.1.13-58-g3441, >>> flags=1 >>> hostname=s1, be=0, 64-bit, os=Linux, arch=x86-64, >>> fio=fio-2.1.13-58-g3441, >>> flags=1 >>> <s2> workload: (g=0): rw=write, <s1> workload: (g=0): rw=write, >>> bs=32K-32K/32K-32K/32K-32K, bs=32K-32K/32K-32K/32K-32K, >>>ioengine=libaio, >>> iodepth=4 >>> ioengine=libaio, iodepth=4 >>> <s2> ... >>> <s1> ... >>> <s2> Starting <s1> Starting 64 threads >>> 64 threads >>> Jobs: 0 (f=0) >>> >>> >>> >>> >>> After 60 seconds.... (on my config file) >>> >>> I have this: >>> >>> workload: (groupid=0, jobs=64): err= 0: pid=3644: Fri Oct 10 09:31:50 >>> 2014 >>> mixed: io=36714MB, bw=1223.2MB/s, iops=39141, runt= 30015msec >>> slat (usec): min=10, max=308, avg=20.57, stdev= 5.24 >>> clat (usec): min=265, max=295734, avg=6496.79, stdev=12165.17 >>> lat (usec): min=280, max=295748, avg=6517.62, stdev=12165.19 >>> clat percentiles (usec): >>> | 1th=[ 868], 5th=[ 1336], 10th=[ 1720], 20th=[ 2384], 30th=[ >>> 3024], >>> | 40th=[ 3696], 50th=[ 4448], 60th=[ 5344], 70th=[ 6496], 80th=[ >>> 8096], >>> | 90th=[11968], 95th=[15808], 99th=[22912], 100th=[69120], >>> 100th=[197632], >>> | 100th=[211968], 100th=[254976] >>> bw (KB /s): min= 3648, max=58112, per=1.57%, avg=19618.75, >>> stdev=4463.44 >>> lat (usec) : 500=0.06%, 750=0.45%, 1000=1.38% >>> lat (msec) : 2=12.27%, 4=29.89%, 10=41.98%, 20=12.22%, 50=1.23% >>> lat (msec) : 100=0.06%, 250=0.45%, 500=0.01% >>> cpu : usr=0.72%, sys=1.15%, ctx=1146612, majf=0, minf=223 >>> IO depths : 1=0.1%, 2=0.1%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, >>>> =64=0.0% >>> submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >>>> =64=0.0% >>> complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >>>> =64=0.0% >>> issued : total=r=1174844/w=0/d=0, short=r=0/w=0/d=0, >>> drop=r=0/w=0/d=0 >>> latency : target=0, window=0, percentile=100.00%, depth=4 >>> >>> Run status group 0 (all jobs): >>> MIXED: io=36714MB, aggrb=1223.2MB/s, minb=1223.2MB/s, >>>maxb=1223.2MB/s, >>> mint=30015msec, maxt=30015msec >>> Jobs: 0 (f=0) >>> >>> >>> >>> After 60 seconds.... ( I have this)... >>> >>> >>> <s1> 0 (f=0) >>> workload: (groupid=0, jobs=64): err= 0: pid=3607: Fri Oct 10 09:32:04 >>> 2014 >>> mixed: io=60243MB, bw=1338.6MB/s, iops=42833, runt= 45006msec >>> slat (usec): min=10, max=1097, avg=21.47, stdev= 7.11 >>> clat (usec): min=256, max=302936, avg=5944.33, stdev=9841.20 >>> lat (usec): min=275, max=302957, avg=5966.03, stdev=9841.14 >>> clat percentiles (usec): >>> | 1th=[ 820], 5th=[ 1240], 10th=[ 1656], 20th=[ 2416], 30th=[ >>> 3120], >>> | 40th=[ 4048], 50th=[ 4704], 60th=[ 5152], 70th=[ 6048], 80th=[ >>> 7328], >>> | 90th=[10304], 95th=[14912], 99th=[21888], 100th=[25728], >>> 100th=[181248], >>> | 100th=[207872], 100th=[244736] >>> bw (KB /s): min= 8256, max=53824, per=1.57%, avg=21476.50, >>> stdev=4885.09 >>> lat (usec) : 500=0.07%, 750=0.59%, 1000=1.81% >>> lat (msec) : 2=11.88%, 4=25.20%, 10=49.90%, 20=8.94%, 50=1.29% >>> lat (msec) : 100=0.04%, 250=0.26%, 500=0.01% >>> cpu : usr=0.78%, sys=1.29%, ctx=1952463, majf=0, minf=219 >>> IO depths : 1=0.1%, 2=0.1%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, >>>> =64=0.0% >>> submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >>>> =64=0.0% >>> complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >>>> =64=0.0% >>> issued : total=r=1927764/w=0/d=0, short=r=0/w=0/d=0, >>> drop=r=0/w=0/d=0 >>> latency : target=0, window=0, percentile=100.00%, depth=4 >>> >>> Run status group 0 (all jobs): >>> MIXED: io=60243MB, aggrb=1338.6MB/s, minb=1338.6MB/s, >>>maxb=1338.6MB/s, >>> mint=45006msec, maxt=45006msec >> >> It's weird, like there's some clock source issue. What happens if you >> add --eta=always as an option to fio? > >Or perhaps try this attached patch. > > >-- >Jens Axboe Hi Jens, This is neto from Brazil How are you? Just applied the patch and it's perfect. Please see below: Nossa Senhora:fiop neto$ ./fio --client 10.61.109.151 --remote-config /root/fiop/iotest --client 10.61.109.152 --remote-config /root/fio/iotest hostname=s2, be=0, 64-bit, os=Linux, arch=x86-64, fio=fio-2.1.13-59-gaa7bc, flags=1 hostname=s1, be=0, 64-bit, os=Linux, arch=x86-64, fio=fio-2.1.13-59-gaa7bc, flags=1 <s2> fio: unable to open '/root/fio/iotest' job file <s1> workload: (g=0): rw=write, bs=32K-32K/32K-32K/32K-32K, ioengine=libaio, iodepth=4 <s1> ... <s1> Starting 64 threads Jobs: 64 (f=1024): [W(64)] [43.3% done] [882.5M/0K/0K /s] [27.6K/0/0 iops] [eta 00m:34s] Thank you very much, neto -- To unsubscribe from this list: send the line "unsubscribe fio" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html