Re: Performance Discrepancy Between rbd bench and fio on Ceph RBD

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Thanks a lot!
That was the part i did wrong :)

Cheers,
Mario

On Mon, Dec 9, 2024 at 5:38 PM <darren@xxxxxxxxxxxx> wrote:
>
> Hi Mario,
>
> You are not testing like for like.
>
> Either increase the numjobs on the FIO command to 16 or reduce the io-threads on the rbd bench command to 1 or add on iodepth=16 to the FIO command.
>
> Currently with the RBD bench you have 16 concurrent IO’s running at the same time but you are restricting FIO to 1.
>
> Darren
>
> > On 9 Dec 2024, at 15:31, Ml Ml <mliebherr99@xxxxxxxxxxxxxx> wrote:
> >
> > Hi all,
> >
> > I have a question regarding the performance differences between rbd
> > bench and a fio test on a Ceph RBD.
> > During testing of my Ceph environment, I noticed that rbd bench
> > reports significantly higher write speeds compared to fio,
> > which is run on an XFS filesystem mounted on the same RBD.
> >
> > Here are the test details:
> >
> > fio test:
> > ------------
> > fio --name=ceph_bench --rw=write --bs=4M --direct=1 --ioengine=libaio
> > --size=4G --numjobs=1 --filename=/mnt/ceph10/backup-cluster1/testfile
> >
> > Results:
> >    Write speed: ~70.3 MiB/s (73.7 MB/s)
> >    IOPS: ~17
> >    Latency: ~56.4 ms (average)
> >
> > rbd bench:
> > --------------
> > rbd bench --io-type write --io-size 4M --io-threads 16 --io-total 12G
> > backup-proxmox/cluster1-new --keyring
> > /etc/ceph/ceph10/ceph.client.admin.keyring --conf
> > /etc/ceph/ceph10/ceph.conf
> >
> > Results:
> >    Write speed: ~150–180 MiB/s (sustained after initial peak)
> >    IOPS: ~43–46
> >    Threads: 16
> >
> > ceph tell osd.X bench:
> > ---------------------------
> > osd.0: IOPS: 20.03, Throughput: 80.11 MB/s
> > osd.1: IOPS: 23.08, Throughput: 92.33 MB/s
> > osd.2: IOPS: 22.01, Throughput: 88.02 MB/s
> > osd.3: IOPS: 17.65, Throughput: 70.61 MB/s
> > osd.4: IOPS: 20.15, Throughput: 80.59 MB/s
> > osd.5: IOPS: 19.71, Throughput: 78.82 MB/s
> > osd.6: IOPS: 18.96, Throughput: 75.84 MB/s
> > osd.7: IOPS: 20.13, Throughput: 80.52 MB/s
> > osd.8: IOPS: 16.45, Throughput: 65.79 MB/s (hm)
> > osd.9: IOPS: 31.16, Throughput: 124.65 MB/s
> > osd.10: IOPS: 23.19, Throughput: 92.76 MB/s
> > osd.12: IOPS: 19.10, Throughput: 76.38 MB/s
> > osd.13: IOPS: 28.02, Throughput: 112.09 MB/s
> > osd.14: IOPS: 19.07, Throughput: 76.27 MB/s
> > osd.15: IOPS: 21.40, Throughput: 85.59 MB/s
> > osd.16: IOPS: 20.24, Throughput: 80.95 MB/s
> >
> >
> > I understand that rbd bench is designed to test the raw performance of
> > the RBD layer, while fio runs at the filesystem level. However, the
> > large discrepancy between the two results has left me curious. What
> > could be causing this difference? Are there any specific factors
> > related to caching, journaling, or XFS itself that could explain the
> > significantly lower performance in the fio test?
> >
> > Any insights or recommendations for further troubleshooting would be
> > greatly appreciated!
> >
> >
> >
> > Thanks in advance!
> > Best regards,
> > Mario
> > _______________________________________________
> > ceph-users mailing list -- ceph-users@xxxxxxx
> > To unsubscribe send an email to ceph-users-leave@xxxxxxx
>
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux