Re: Rebuilding data resiliency after adding new OSD's stuck for so long at 5%

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,

As long as you see changes and "recovery" it will make progress so i guess you have just to wait...

What kind of disks did you add?

Hth
Mehmet

Am 12. September 2023 20:37:56 MESZ schrieb sharathvuthpala@xxxxxxxxx:
>We have a user-provisioned instance( Bare Metal Installation) of OpenShift cluster running on version 4.12 and we are using OpenShift Data Foundation as the Storage System. Earlier we had 3 disks attached to the storage system and 3 OSDs were available in the cluster. Today, while adding additional disks to the storage cluster, we increased the number of disks from 3 to 9, that is 3 per node. The addition of storage capacity was successful, resulting in 6 new OSDs in the cluster. 
>
>But,  after this operation, we noticed that Rebuilding Data Resiliency is stuck at 5% and not moving forward. At the same time, ceph status shows 65% of objects are misplaced and PGs are not in active+clean state. 
>
>Here is more information about the ceph cluster:
>
>
>sh-4.4$ ceph status
>  cluster:
>    id:     18bf836d-4937-4925-b964-7a026c1d548d
>    health: HEALTH_OK
>
>  services:
>    mon: 3 daemons, quorum b,u,v (age 2w)
>    mgr: a(active, since 7w)
>    mds: 1/1 daemons up, 1 hot standby
>    osd: 9 osds: 9 up (since 5h), 9 in (since 5h); 191 remapped pgs
>    rgw: 1 daemon active (1 hosts, 1 zones)
>
>  data:
>    volumes: 1/1 healthy
>    pools:   12 pools, 305 pgs
>    objects: 2.69M objects, 2.9 TiB
>    usage:   8.8 TiB used, 27 TiB / 36 TiB avail
>    pgs:     4723077/8079717 objects misplaced (58.456%)
>             188 active+remapped+backfill_wait
>             114 active+clean
>             3   active+remapped+backfilling
>
>  io:
>    client:   679 KiB/s rd, 11 MiB/s wr, 13 op/s rd, 622 op/s wr
>    recovery: 20 MiB/s, 89 keys/s, 22 objects/s
>
>
>sh-4.4$ ceph balancer status
>{
>    "active": true,
>    "last_optimize_duration": "0:00:00.000276",
>    "last_optimize_started": "Tue Sep 12 17:36:03 2023",
>    "mode": "upmap",
>    "optimize_result": "Too many objects (0.581933 > 0.050000) are misplaced; try again later",
>    "plans": []
>}
>
>One more thing we observed is that the number of misplaced objects is decreasing and also there is a drop in the percentage. What might be the reason behind Rebuilding Data Resiliency is not moving forward?
>
>Any inputs would be appreciated.
>
>Thanks
>_______________________________________________
>ceph-users mailing list -- ceph-users@xxxxxxx
>To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux