Re: CephFS 16.2.10 problem

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Good afternoon

We tried to leave only one mds, stopped others, even deleted one, and turned off the requirement for stand-by mds. Nothing helped, mds remained in the status of replays.
Current situation: we now have two active mds in the status of replays, and one in stand-by.
At the same time, in the logs we see a message
mds.0.660178  waiting for osdmap 123138 (which blocklists prior instance)
At the same time, there is no activity on both mds.
The launch of the cephfs-journal-tool journal inspect utility does not produce any results - the utility worked for 12 hours and did not produce anything, we stopped it.

Maybe the problem is this blocking? How to remove it?

Best regards!

Alexey Tsivinsky
e-mail: a.tsivinsky@xxxxxxxxxxxxxxxxxxxxx
________________________________________
От: Marc <Marc@xxxxxxxxxxxxxxxxx>
Отправлено: 25 ноября 2024 г. 1:47
Кому: Цивинский Алексей Александрович; ceph-users@xxxxxxx
Тема: RE: CephFS 16.2.10 problem

>
> The following problem occurred.
> There is a cluster ceph 16.2.10
> The cluster was operating normally on Friday. Shut down cluster:
> -Excluded all clients
> Executed commands:
> ceph osd set noout
> ceph osd set nobackfill
> ceph osd set norecover
> ceph osd set norebalance
> ceph osd set nodown
> ceph osd set pause
> Turned off the cluster, checked server maintenance.
> Enabled cluster. He gathered himself, found all the nodes, and here the
> problem began. After all OSD went up and all pg became available, cephfs
> refused to start.
> Now mds are in the replay status, and do not go to the ready status.
> Previously, one of them was in the replay (laggy) status, but we
> executed command:  ceph config set mds mds_wipe_sessions true
> After that, mds switched to the status of replays, the third in standby
> status started, and mds crashes with an error stopped.
> But cephfs is still unavailable.
> What else can we do?
> The cluster is very large, almost 200 million files.
>

I assume you tried to start just one mds and wait until it would come up as active (before starting the others)?




_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux