| tis 4 sep. 18:07 (21 timmar sedan) | |||
Hi,
I have a really small homelab 3-node ceph cluster on consumer hw - thanks to Proxmox for making it easy to deploy it.
The problem I am having is very very bad transfer rates, ie 20mb/sec for both read and write on 17 OSDs with cache layer.
However during recovery the speed hover between 250 to 700mb/sec which proves that the cluster IS capable of reaching way above those 20mb/sec in KVM.
Reading the documentation, I see that during recovery "nearly all OSDs participate in resilvering a new drive" - kind of a torrent of data incoming from multiple sources at once, causing a huge deluge.
However I believe this does not happen during the normal transfers, so my question is simply - is there any hidden tunables I can enable for this with the implied cost of network and heavy usage of disks ? Will there be in the future if not ?
I have a really small homelab 3-node ceph cluster on consumer hw - thanks to Proxmox for making it easy to deploy it.
The problem I am having is very very bad transfer rates, ie 20mb/sec for both read and write on 17 OSDs with cache layer.
However during recovery the speed hover between 250 to 700mb/sec which proves that the cluster IS capable of reaching way above those 20mb/sec in KVM.
Reading the documentation, I see that during recovery "nearly all OSDs participate in resilvering a new drive" - kind of a torrent of data incoming from multiple sources at once, causing a huge deluge.
However I believe this does not happen during the normal transfers, so my question is simply - is there any hidden tunables I can enable for this with the implied cost of network and heavy usage of disks ? Will there be in the future if not ?
I have tried disabling authx, upgrading the network to 10gbit, have bigger journals, more bluestore cache and disabled the debugging logs as it has been advised on the list. The only thing that did help a bit was cache tiering, but this only helps somewhat as the ops do not get promoted unless I am very adamant about keeping programs in KVM open for very long times so that the writes/reads are promoted.
To add some to the injury, once the cache gets full - the whole 3-node cluster grinds to a full halt until I start forcefully evict data from the cache... manually!
So I am therefore guessing a really bad misconfiguration from my side.
Next step would be removing the cache layer and using those SSDs as bcache instead as it seems to yeld 5x the results, even though it does add yet another layer of complexity and RAM requirements.
Full config details:
https://pastebin.com/xUM7VF9k
To add some to the injury, once the cache gets full - the whole 3-node cluster grinds to a full halt until I start forcefully evict data from the cache... manually!
So I am therefore guessing a really bad misconfiguration from my side.
Next step would be removing the cache layer and using those SSDs as bcache instead as it seems to yeld 5x the results, even though it does add yet another layer of complexity and RAM requirements.
Full config details:
https://pastebin.com/xUM7VF9k
rados bench -p ceph_pool 30 write
Total time run: 30.983343
Total writes made: 762
Write size: 4194304
Object size: 4194304
Bandwidth (MB/sec): 98.3754
Stddev Bandwidth: 20.9586
Max bandwidth (MB/sec): 132
Min bandwidth (MB/sec): 16
Average IOPS: 24
Stddev IOPS: 5
Max IOPS: 33
Min IOPS: 4
Average Latency(s): 0.645017
Stddev Latency(s): 0.326411
Max latency(s): 2.08067
Min latency(s): 0.0355789
Cleaning up (deleting benchmark objects)
Removed 762 objects
Clean up completed and total clean up time :3.925631
Thanks,
Alex
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com