Hello *, thx for taking care. I read "works as designed, be sure to have disk space for the mon available". It sounds a little odd that the growth from 50MB to ~15GB + compaction space happens within a couple of seconds, when two OSD rejoin the cluster. Does it matter if I have cephfs in use? Usually I would expect to have MDS load, but does it also cause load on the mon with many files? My OSD map seems to have low absolute numbers: ceph report | grep osdmap | grep committed report 777999536 "osdmap_first_committed": 1276, "osdmap_last_committed": 1781, If a get new disks (partitions) for the mons, is there a size recommendation? Is there a rule of thumb? BTW: Do I still need a filesystem for the partition of the mon DB? Beste Regards, derjohn On 02.10.20 16:25, Dan van der Ster wrote: > The important metric is the difference between these two values: > > # ceph report | grep osdmap | grep committed > report 3324953770 > "osdmap_first_committed": 3441952, > "osdmap_last_committed": 3442452, > > The mon stores osdmaps on disk, and trims the older versions whenever > the PGs are clean. Trimming brings the osdmap_first_committed to be > closer to osdmap_last_committed. > In a cluster with no PGs backfilling or recovering, the mon should > trim that difference to be within 500-750 epochs. > > If there are any PGs backfilling or recovering, then the mon will not > trim beyond the osdmap epoch when the pools were clean. > > So if you are accumulating gigabytes of data in the mon dir, it > suggests that you have unclean PGs/Pools. > > Cheers, dan > > > > > On Fri, Oct 2, 2020 at 4:14 PM Marc Roos <M.Roos@xxxxxxxxxxxxxxxxx> wrote: >> >> Does this also count if your cluster is not healthy because of errors >> like '2 pool(s) have no replicas configured' >> I sometimes use these pools for testing, they are empty. >> >> >> >> >> -----Original Message----- >> Cc: ceph-users >> Subject: Re: Massive Mon DB Size with noout on 14.2.11 >> >> As long as the cluster is no healthy, the OSD will require much more >> space, depending on the cluster size and other factors. Yes this is >> somewhat normal. >> >> _______________________________________________ >> ceph-users mailing list -- ceph-users@xxxxxxx >> To unsubscribe send an email to ceph-users-leave@xxxxxxx > _______________________________________________ > ceph-users mailing list -- ceph-users@xxxxxxx > To unsubscribe send an email to ceph-users-leave@xxxxxxx > -- Andreas John net-lab GmbH | Frankfurter Str. 99 | 63067 Offenbach Geschaeftsfuehrer: Andreas John | AG Offenbach, HRB40832 Tel: +49 69 8570033-1 | Fax: -2 | http://www.net-lab.net Facebook: https://www.facebook.com/netlabdotnet Twitter: https://twitter.com/netlabdotnet _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx