Re: Performance Discrepancy Between rbd bench and fio on Ceph RBD

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Another frequent cause is when someone runs from an RBD VM that is IOPS-throttled at the libvirt layer, and they do like 1KB writes.  Increasing the write size does wonders :D

> On Dec 10, 2024, at 5:47 AM, Ml Ml <mliebherr99@xxxxxxxxxxxxxx> wrote:
> 
> Thanks a lot!
> That was the part i did wrong :)
> 
> Cheers,
> Mario
> 
> On Mon, Dec 9, 2024 at 5:38 PM <darren@xxxxxxxxxxxx> wrote:
>> 
>> Hi Mario,
>> 
>> You are not testing like for like.
>> 
>> Either increase the numjobs on the FIO command to 16 or reduce the io-threads on the rbd bench command to 1 or add on iodepth=16 to the FIO command.
>> 
>> Currently with the RBD bench you have 16 concurrent IO’s running at the same time but you are restricting FIO to 1.
>> 
>> Darren
>> 
>>> On 9 Dec 2024, at 15:31, Ml Ml <mliebherr99@xxxxxxxxxxxxxx> wrote:
>>> 
>>> Hi all,
>>> 
>>> I have a question regarding the performance differences between rbd
>>> bench and a fio test on a Ceph RBD.
>>> During testing of my Ceph environment, I noticed that rbd bench
>>> reports significantly higher write speeds compared to fio,
>>> which is run on an XFS filesystem mounted on the same RBD.
>>> 
>>> Here are the test details:
>>> 
>>> fio test:
>>> ------------
>>> fio --name=ceph_bench --rw=write --bs=4M --direct=1 --ioengine=libaio
>>> --size=4G --numjobs=1 --filename=/mnt/ceph10/backup-cluster1/testfile
>>> 
>>> Results:
>>>   Write speed: ~70.3 MiB/s (73.7 MB/s)
>>>   IOPS: ~17
>>>   Latency: ~56.4 ms (average)
>>> 
>>> rbd bench:
>>> --------------
>>> rbd bench --io-type write --io-size 4M --io-threads 16 --io-total 12G
>>> backup-proxmox/cluster1-new --keyring
>>> /etc/ceph/ceph10/ceph.client.admin.keyring --conf
>>> /etc/ceph/ceph10/ceph.conf
>>> 
>>> Results:
>>>   Write speed: ~150–180 MiB/s (sustained after initial peak)
>>>   IOPS: ~43–46
>>>   Threads: 16
>>> 
>>> ceph tell osd.X bench:
>>> ---------------------------
>>> osd.0: IOPS: 20.03, Throughput: 80.11 MB/s
>>> osd.1: IOPS: 23.08, Throughput: 92.33 MB/s
>>> osd.2: IOPS: 22.01, Throughput: 88.02 MB/s
>>> osd.3: IOPS: 17.65, Throughput: 70.61 MB/s
>>> osd.4: IOPS: 20.15, Throughput: 80.59 MB/s
>>> osd.5: IOPS: 19.71, Throughput: 78.82 MB/s
>>> osd.6: IOPS: 18.96, Throughput: 75.84 MB/s
>>> osd.7: IOPS: 20.13, Throughput: 80.52 MB/s
>>> osd.8: IOPS: 16.45, Throughput: 65.79 MB/s (hm)
>>> osd.9: IOPS: 31.16, Throughput: 124.65 MB/s
>>> osd.10: IOPS: 23.19, Throughput: 92.76 MB/s
>>> osd.12: IOPS: 19.10, Throughput: 76.38 MB/s
>>> osd.13: IOPS: 28.02, Throughput: 112.09 MB/s
>>> osd.14: IOPS: 19.07, Throughput: 76.27 MB/s
>>> osd.15: IOPS: 21.40, Throughput: 85.59 MB/s
>>> osd.16: IOPS: 20.24, Throughput: 80.95 MB/s
>>> 
>>> 
>>> I understand that rbd bench is designed to test the raw performance of
>>> the RBD layer, while fio runs at the filesystem level. However, the
>>> large discrepancy between the two results has left me curious. What
>>> could be causing this difference? Are there any specific factors
>>> related to caching, journaling, or XFS itself that could explain the
>>> significantly lower performance in the fio test?
>>> 
>>> Any insights or recommendations for further troubleshooting would be
>>> greatly appreciated!
>>> 
>>> 
>>> 
>>> Thanks in advance!
>>> Best regards,
>>> Mario
>>> _______________________________________________
>>> ceph-users mailing list -- ceph-users@xxxxxxx
>>> To unsubscribe send an email to ceph-users-leave@xxxxxxx
>> 
> _______________________________________________
> ceph-users mailing list -- ceph-users@xxxxxxx
> To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux