That seems like a sane way to do it, thanks for the clarification Wido. As a follow-up, do you have any feeling as to whether the trimming a particularly intensive task? We just had a fun afternoon where the monitors became unresponsive (no ceph status etc) for several hours, seemingly due to the leaders monitor process consuming all available ram+swap (64GB+32GB) on that monitor. This was then followed by the actual trimming of the stores (26GB->11GB), which took a few minutes and happened simultaneously across the monitors. If this is something to be expected, it'll be a good reason to plan our long backfills much more carefully in the future! > -----Original Message----- > From: ceph-users <ceph-users-bounces@xxxxxxxxxxxxxx> On Behalf Of Wido > den Hollander > Sent: 17 May 2018 15:40 > To: ceph-users@xxxxxxxxxxxxxx > Subject: Re: A question about HEALTH_WARN and monitors > holding onto cluster maps > > > > On 05/17/2018 04:37 PM, Thomas Byrne - UKRI STFC wrote: > > Hi all, > > > > > > > > As far as I understand, the monitor stores will grow while not > > HEALTH_OK as they hold onto all cluster maps. Is this true for all > > HEALTH_WARN reasons? Our cluster recently went into HEALTH_WARN > due to > > a few weeks of backfilling onto new hardware pushing the monitors data > > stores over the default 15GB threshold. Are they now prevented from > > shrinking till I increase the threshold above their current size? > > > > No, monitors will trim their data store with all PGs are active+clean, not when > they are HEALTH_OK. > > So a 'noout' flag triggers a WARN, but that doesn't prevent the MONs from > trimming for example. > > Wido > > > > > > > Cheers > > > > Tom > > > > > > > > > > > > > > > > _______________________________________________ > > ceph-users mailing list > > ceph-users@xxxxxxxxxxxxxx > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > > _______________________________________________ > ceph-users mailing list > ceph-users@xxxxxxxxxxxxxx > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com