We got the same problem today while we were adding memory to OSD nodes, and it decreased monitor’s performance a lot. I noticed that the db kept increasing after an OSD is shutdown, so I guess that it is caused by the warning reports collected by mgr insights module. When I disabled the mgr insights module, the db size reduced to 1xx MB from 3x GB. > 2020年3月12日 下午2:44,Hartwig Hauschild <ml-ceph@xxxxxxxxxxxx> 写道: > > Am 10.03.2020 schrieb Wido den Hollander: >> >> >> On 3/10/20 10:48 AM, Hartwig Hauschild wrote: >>> Hi, >>> >>> I've done a bit more testing ... >>> >>> Am 05.03.2020 schrieb Hartwig Hauschild: >>>> Hi, >>>> > [ snipped ] >>> I've read somewhere in the docs that I should provide ample space (tens of >>> GB) for the store.db, found on the ML and Bugtracker that ~100GB might not >>> be a bad idea and that large clusters may require space on order of >>> magnitude greater. >>> Is there some sort of formula I can use to approximate the space required? >> >> I don't know about a formula, but make sure you have enough space. MONs >> are dedicated nodes in most production environments, so I usually >> install a 400 ~ 1000GB SSD just to make sure they don't run out of space. >> > That seems fair. >>> >>> Also: is the db supposed to grow this fast in Nautilus when it did not do >>> that in Luminous? Is that behaviour configurable somewhere? >>> >> >> The MONs need to cache the OSDMaps when not all PGs are active+clean >> thus their database grows. >> >> You can compact RocksDB in the meantime, but it won't last for ever. >> >> Just make sure the MONs have enough space. >> > Do you happen to know if that behaved differently in previous releases? I'm > just asking because I have not found anything about this yet and may need to > explain that it's different now. > > -- > Cheers, > Hardy > _______________________________________________ > ceph-users mailing list -- ceph-users@xxxxxxx > To unsubscribe send an email to ceph-users-leave@xxxxxxx _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx