Re: nautilus cluster down by loss of 2 mons

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,
Maybe this is https://tracker.ceph.com/issues/50004 ?
14.2.22 and 15.2.13 have a more aggressive mon trimming to prevent
this sort of issue, when osds are spamming the mons with logs.
-- dan

On Tue, Aug 31, 2021 at 9:39 PM Marcel Kuiper <ceph@xxxxxxxx> wrote:
>
> During normal operation the size is under 1G. After the network ordeal
> it was 65G. I gave the last mon all diskspace I could find under
> /var/lib/ceph and started the mon again. it is now reaching 90G and
> still growing
>
> Does anyone have an idea howmuch disk free would be needed to get the
> job done?
>
> Any other strategies to get the cluster going again??
>
> Marc schreef op 2021-08-31 20:02:
> > Could someone also explain the logics behind the decision to dump so
> > much data to the disk. Especially in container environments with
> > resource limits this is not really nice.
> >
> >> -----Original Message-----
> >> Sent: Tuesday, 31 August 2021 19:16
> >> To: ceph-users@xxxxxxx
> >> Subject:  nautilus cluster down by loss of 2 mons
> >>
> >> Hi
> >>
> >> We have a nautilus cluster that was plagued by a network failure. One
> >> of
> >> the monitors fell out of quorum
> >> Once the network settled down and all osds were back online again we
> >> got
> >> that mon synchronizing
> >>
> >> However the filesystem suddenly exploded in a minute or so from 63G
> >> usage to 93G  resulting in 100% usage
> >> At that point we decided to remove that mon from the cluster and hope
> >> to
> >> compact the database on the remaining mons so that
> >> we could add a new mon while there was less synchronizing to do
> >> because
> >> of the smaller database size
> >>
> >> Unfortunately the tell osd compact command made the database on mon nr
> >> 2
> >> grow very fast resulting in another full filesystem
> >> hence a dead cluster
> >>
> >> Can anyone advice towards the fastest recovery in this situation?
> > _______________________________________________
> > ceph-users mailing list -- ceph-users@xxxxxxx
> > To unsubscribe send an email to ceph-users-leave@xxxxxxx
> _______________________________________________
> ceph-users mailing list -- ceph-users@xxxxxxx
> To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux