Re: Watcher Issue

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



We really don't know what the cluster state was. I assume by removing the fifth host you removed the fullest OSD(s). But again, we only get a few bits of information, so it's still basically guessing. What exactly do you mean by "so I thought to run pool repair"? What exactly did you do?

Zitat von Devender Singh <devender@xxxxxxxxxx>:

Hello Eugen

Thanks for your reply.

	ceph osd set nodeep-scrub is not stopping if repairs are running.
Reapir started another set for deepscrub+repair which is not controlled using this command.

When I started my cluster utilization as 74% and when it finished now my cluster is showing 43%(surprising figure, does a repair can shuffle that much? Even my osd’s where between 60-83%) utilized. Before starting repair I increased pool pg’s manually from 1024 to 2048 but my pool only came down to 84%, so I thought to run pool repair.. During the repair and high utilization my cluster was crying and giving SLOW OSD Communication from osd.x to osd.y and many number of times.

Not sure what was it, how utilization came that much down, If repair can change thing this much then it should have command to pause/unpause too.


Regards
Dev



On Jan 25, 2025, at 1:15 AM, Eugen Block <eblock@xxxxxx> wrote:

But they would only reveal inconsistent PGs during deep-scrub.


_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux