Re: [EXTERNAL] Ceph performance is too good (impossible..)...

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,

you need to flush all caches before starting read tests. With fio you can probably do this if you keep the files that it creates.

as root on all clients and all osd nodes run:

echo 3 > /proc/sys/vm/drop_caches

But fio is a little problematic for ceph because of the caches in the clients and the osd nodes. If you really need to know the read rates, go for large files, write them with dd, flush all caches, and the read the files with dd. Single threaded read dd shows less throughput compared to multiple threads dd read. readahead size does also matter.

Good luck testing.

- mike


On 12/13/16 2:37 AM, V Plus wrote:
The same..
see:
A: (g=0): rw=read, bs=5M-5M/5M-5M/5M-5M, ioengine=*libaio*, iodepth=1
...
fio-2.2.10
Starting 16 processes

A: (groupid=0, jobs=16): err= 0: pid=27579: Mon Dec 12 20:36:10 2016
  mixed: io=122515MB, bw=6120.3MB/s, iops=1224, runt= 20018msec

I think at the end, the only one way to solve this issue is to write the
image before read test....as suggested....

I have no clue why rbd engine does not work...

On Mon, Dec 12, 2016 at 4:23 PM, Will.Boege <Will.Boege@xxxxxxxxxx
<mailto:Will.Boege@xxxxxxxxxx>> wrote:

    Try adding --ioengine=libaio____

    __ __

    *From: *V Plus <v.plussharp@xxxxxxxxx <mailto:v.plussharp@xxxxxxxxx>>
    *Date: *Monday, December 12, 2016 at 2:40 PM
    *To: *"Will.Boege" <Will.Boege@xxxxxxxxxx
    <mailto:Will.Boege@xxxxxxxxxx>>
    *Subject: *Re: [EXTERNAL]  Ceph performance is too good
    (impossible..)...____

    __ __

    Hi Will, ____

    thanks very much..____

    However, I tried with your suggestions.____

    Both are *not *working...____

    1. with FIO rbd engine:
    *[RBD_TEST]
    ioengine=rbd
    clientname=admin
    pool=rbd
    rbdname=fio_test
    invalidate=1
    direct=1
    group_reporting=1
    unified_rw_reporting=1
    time_based=1
    rw=read
    bs=4MB
    numjobs=16
    ramp_time=10
    runtime=20*____

    then I run "sudo fio rbd.job" and  got:____


    RBD_TEST: (g=0): rw=read, bs=4M-4M/4M-4M/4M-4M, ioengine=rbd, iodepth=1
    ...
    fio-2.2.10
    Starting 16 processes
    rbd engine: RBD version: 0.1.10
    rbd engine: RBD version: 0.1.10
    rbd engine: RBD version: 0.1.10
    rbd engine: RBD version: 0.1.10
    rbd engine: RBD version: 0.1.10
    rbd engine: RBD version: 0.1.10
    rbd engine: RBD version: 0.1.10
    rbd engine: RBD version: 0.1.10
    rbd engine: RBD version: 0.1.10
    rbd engine: RBD version: 0.1.10
    rbd engine: RBD version: 0.1.10
    rbd engine: RBD version: 0.1.10
    rbd engine: RBD version: 0.1.10
    rbd engine: RBD version: 0.1.10
    rbd engine: RBD version: 0.1.10
    rbd engine: RBD version: 0.1.10
    Jobs: 12 (f=7): [R(5),_(1),R(4),_(1),R(2),_(2),R(1)] [100.0% done]
    [11253MB/0KB/0KB /s] [2813/0/0 iops] [eta 00m:00s]
    RBD_TEST: (groupid=0, jobs=16): err= 0: pid=17504: Mon Dec 12
    15:32:52 2016
      mixed: io=212312MB, *bw=10613MB/s*, iops=2653, runt= 20005msec____

    2. with blockalign
    [A]
    direct=1
    group_reporting=1
    unified_rw_reporting=1
    size=100%
    time_based=1
    filename=/dev/rbd0
    rw=read
    bs=5MB
    numjobs=16
    ramp_time=5
    runtime=20
    blockalign=512b

    [B]
    direct=1
    group_reporting=1
    unified_rw_reporting=1
    size=100%
    time_based=1
    filename=/dev/rbd1
    rw=read
    bs=5MB
    numjobs=16
    ramp_time=5
    runtime=20
    blockalign=512b

    sudo fio fioA.job -output a.txt & sudo  fio fioB.job -output b.txt
    & wait____

    Then I got:
    A: (groupid=0, jobs=16): err= 0: pid=19320: Mon Dec 12 15:35:32 2016
      mixed: io=88590MB, bw=4424.7MB/s, iops=884, runt= 20022msec____

    B: (groupid=0, jobs=16): err= 0: pid=19324: Mon Dec 12 15:35:32 2016
      mixed: io=88020MB, *bw=4395.6MB/s*, iops=879, runt= 20025msec____

    ..............____

    __ __

    On Mon, Dec 12, 2016 at 10:45 AM, Will.Boege <Will.Boege@xxxxxxxxxx
    <mailto:Will.Boege@xxxxxxxxxx>> wrote:____

        My understanding is that when using direct=1 on a raw block
        device FIO (aka-you) will have to handle all the sector
        alignment or the request will get buffered to perform the
        alignment.  ____

         ____

        Try adding the –blockalign=512b option to your jobs, or better
        yet just use the native FIO RBD engine.____

         ____

        Something like this (untested) - ____

         ____

        [A]____

        ioengine=rbd____

        clientname=admin____

        pool=rbd____

        rbdname=fio_test____

        direct=1____

        group_reporting=1____

        unified_rw_reporting=1____

        time_based=1____

        rw=read____

        bs=4MB____

        numjobs=16____

        ramp_time=10____

        runtime=20____

         ____

        *From: *ceph-users <ceph-users-bounces@xxxxxxxxxxxxxx
        <mailto:ceph-users-bounces@xxxxxxxxxxxxxx>> on behalf of V Plus
        <v.plussharp@xxxxxxxxx <mailto:v.plussharp@xxxxxxxxx>>
        *Date: *Sunday, December 11, 2016 at 7:44 PM
        *To: *"ceph-users@xxxxxxxxxxxxxx
        <mailto:ceph-users@xxxxxxxxxxxxxx>" <ceph-users@xxxxxxxxxxxxxx
        <mailto:ceph-users@xxxxxxxxxxxxxx>>
        *Subject: *[EXTERNAL]  Ceph performance is too good
        (impossible..)...____

         ____

        Hi Guys,____

        we have a ceph cluster with 6 machines (6 OSD per host). ____

        1. I created 2 images in Ceph, and map them to another host A
        (*/outside /*the Ceph cluster). On host A, I
        got *//dev/rbd0/* and*/ /dev/rbd1/*.____

        2. I start two fio job to perform READ test on rbd0 and rbd1.
        (fio job descriptions can be found below)____

        */"sudo fio fioA.job -output a.txt & sudo  fio fioB.job -output
        b.txt  & wait"/*____

        3. After the test, in a.txt, we got */bw=1162.7MB/s/*, in b.txt,
        we get */bw=3579.6MB/s/*.____

        The results do NOT make sense because there is only one NIC on
        host A, and its limit is 10 Gbps (1.25GB/s).____

         ____

        I suspect it is because of the cache setting.____

        But I am sure that in file *//etc/ceph/ceph.conf/* on host A,I
        already added:____

        */[client]/*____

        */rbd cache = false/*____

         ____

        Could anyone give me a hint what is missing? why....____

        Thank you very much.____

         ____

        *fioA.job:*____

        /[A]/____

        /direct=1/____

        /group_reporting=1/____

        /unified_rw_reporting=1/____

        /size=100%/____

        /time_based=1/____

        /filename=/dev/rbd0/____

        /rw=read/____

        /bs=4MB/____

        /numjobs=16/____

        /ramp_time=10/____

        /runtime=20/____

         ____

        *fioB.job:*____

        /[B]/____

        /direct=1/____

        /group_reporting=1/____

        /unified_rw_reporting=1/____

        /size=100%/____

        /time_based=1/____

        /filename=/dev/rbd1/____

        /rw=read/____

        /bs=4MB/____

        /numjobs=16/____

        /ramp_time=10/____

        /runtime=20/____

         ____

        /Thanks.../____

    __ __




_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux