Hi Christian,
Thanks for the update.
I have 5 SSD on each node i.e. a total of 15 SSD using which I have created this RAID 0 Disk, which in Ceph becomes three OSD. Each OSD with around 4.4 TB of disk. and in total it is coming around 13.3 TB.
Do you feel local RAID is an issue here? Keeping independent disks can help recovery fast or increase the performance? please advice.
Thanks once again.
-Lokendra
On Thu, Oct 28, 2021 at 1:30 PM Christian Wuerdig <christian.wuerdig@xxxxxxxxx> wrote:
Hm, so you have 13TiB across 3 nodes with 3 OSDs total - are you running 5TB HDDs or are you running some sort of local RAID? If each OSD is a single drive then 154MiB/sec recovery seems quite decent since it's essentially backfilling that single drive. If you run a local RAID then get rid of it and expose the actual drives directly (JBOD) - you will cause yourself much fewer problems in the long runOn Thu, 28 Oct 2021 at 20:09, Lokendra Rathour <lokendrarathour@xxxxxxxxx> wrote:Hi,_______________________________________________we have been trying to test a scenario on ceph with the following configuration:cluster:
id: cc0ba1e4-68b9-4237-bc81-40b38455f713
health: HEALTH_OK
services:
mon: 3 daemons, quorum storagenode1,storagenode2,storagenode3 (age 4h)
mgr: storagenode2(active, since 22h), standbys: storagenode1, storagenode3
mds: cephfs:1 {0=storagenode1=up:active} 2 up:standby
osd: 3 osds: 3 up (since 4m), 3 in (since 4h)
rgw: 3 daemons active (storagenode1.rgw0, storagenode2.rgw0, storagenode3.rgw0)
task status:
scrub status:
mds.storagenode1: idle
data:
pools: 7 pools, 169 pgs
objects: 1.06M objects, 1.3 TiB
usage: 3.9 TiB used, 9.2 TiB / 13 TiB avail
pgs: 169 active+clean
io:
client: 43 KiB/s wr, 0 op/s rd, 3 op/s wr
recovery: 154 MiB/s, 98 objects/s
We have network links of 10GiG for all the networks used in Ceph. MTU is configured as 9000. But the Transfer rate as can be seen above is max 154 MiB/s which I feel is way low than possible.Test Case:We removed one node and added it back to the Ceph Cluster after reinstalling the OS. During this time of activity, Ceph has around 1.3 TB to rebalance in the newly added node. The time taken in such a case is approximate: 4 hours.Considering this as the production-grade setup with all production-grade infra, this time is too much.Query:
- Is there a way to optimize the recovery/rebalancing and i/o rate of Ceph?
- we found a few suggestions on the internet that we can modify the below parameters to achieve a good rate, but is this advisable
- osd max backfills, osd recovery max active, osd recovery max single start
- we have dedicated 10gig n/w infra so can we have any ideal value to reach max rate of recovery.
Any input would be helpful, we are really blocked here.--~ Lokendraskype: lokendrarathour
Dev mailing list -- dev@xxxxxxx
To unsubscribe send an email to dev-leave@xxxxxxx
_______________________________________________ Dev mailing list -- dev@xxxxxxx To unsubscribe send an email to dev-leave@xxxxxxx