Good afternoon We tried to leave only one mds, stopped others, even deleted one, and turned off the requirement for stand-by mds. Nothing helped, mds remained in the status of replays. Current situation: we now have two active mds in the status of replays, and one in stand-by. At the same time, in the logs we see a message mds.0.660178 waiting for osdmap 123138 (which blocklists prior instance) At the same time, there is no activity on both mds. The launch of the cephfs-journal-tool journal inspect utility does not produce any results - the utility worked for 12 hours and did not produce anything, we stopped it. Maybe the problem is this blocking? How to remove it? Best regards! Alexey Tsivinsky e-mail: a.tsivinsky@xxxxxxxxxxxxxxxxxxxxx ________________________________________ От: Marc <Marc@xxxxxxxxxxxxxxxxx> Отправлено: 25 ноября 2024 г. 1:47 Кому: Цивинский Алексей Александрович; ceph-users@xxxxxxx Тема: RE: CephFS 16.2.10 problem > > The following problem occurred. > There is a cluster ceph 16.2.10 > The cluster was operating normally on Friday. Shut down cluster: > -Excluded all clients > Executed commands: > ceph osd set noout > ceph osd set nobackfill > ceph osd set norecover > ceph osd set norebalance > ceph osd set nodown > ceph osd set pause > Turned off the cluster, checked server maintenance. > Enabled cluster. He gathered himself, found all the nodes, and here the > problem began. After all OSD went up and all pg became available, cephfs > refused to start. > Now mds are in the replay status, and do not go to the ready status. > Previously, one of them was in the replay (laggy) status, but we > executed command: ceph config set mds mds_wipe_sessions true > After that, mds switched to the status of replays, the third in standby > status started, and mds crashes with an error stopped. > But cephfs is still unavailable. > What else can we do? > The cluster is very large, almost 200 million files. > I assume you tried to start just one mds and wait until it would come up as active (before starting the others)? _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx