As i can see, all pg's are active+clean: ~# ceph -s cluster: id: d168189f-6105-4223-b244-f59842404076 health: HEALTH_WARN noout,nodeep-scrub flag(s) set mons 1,2,3,4,5 are using a lot of disk space services: mon: 5 daemons, quorum 1,2,3,4,5 mgr: api1(active), standbys: api2 osd: 832 osds: 791 up, 790 in flags noout,nodeep-scrub data: pools: 10 pools, 52336 pgs objects: 47.78M objects, 238TiB usage: 854TiB used, 1.28PiB / 2.12PiB avail pgs: 52336 active+clean io: client: 929MiB/s rd, 1.16GiB/s wr, 31.85kop/s rd, 36.19kop/s wr 08.10.2018, 22:11, "Wido den Hollander" <wido@xxxxxxxx>: > On 10/08/2018 05:04 PM, Aleksei Zakharov wrote: >> Hi all, >> >> We've upgraded our cluster from jewel to luminous and re-created monitors using rocksdb. >> Now we see, that mon's are using a lot of disk space and used space only grows. It is about 17GB for now. It was ~13GB when we used leveldb and jewel release. >> >> When we added new osd's we saw that it downloads from monitors a lot of data. It was ~15GiB few days ago and it is ~18GiB today. >> One of the osd's we created uses filestore and it looks like old osd maps are not removed: >> >> ~# find /var/lib/ceph/osd/ceph-224/current/meta/ | wc -l >> 73590 >> >> I've tried to run manual compaction (ceph tell mon.NUM compact) but it doesn't help. >> >> So, how to stop this growth of data on monitors? > > What is the status of Ceph? Can you post the output of: > > $ ceph -s > > MONs do not trim their database if one or more PGs aren't active+clean. > > Wido > >> -- >> Regards, >> Aleksei Zakharov >> >> _______________________________________________ >> ceph-users mailing list >> ceph-users@xxxxxxxxxxxxxx >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com -- Regards, Aleksei Zakharov _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com