I did the heap release. It did not free the memory right away, but after I restarted the OSD, it started with "just" 10GB RAM. After ~12h the container reports again 30+ GB RAM. To mention that I lowered pg_log values too: osd_min_pg_log_entries = 100 osd_max_pg_log_entries = 500 osd_target_pg_log_entries_per_osd = 30000 On Tue, Nov 16, 2021 at 10:49 PM Mark Nelson <mnelson@xxxxxxxxxx> wrote: > Yeah, if it's not memory reported by the mempools, that means it's > something we aren't tracking. Perhaps temporary allocations in some > dark corner of the code, or possibly rocksdb (though 38GB of ram is > obviously excessive). heap stats are a good idea. it's possible if > neither the heap stats nor the mempool stats are helpful (and if debug > bluestore = 5 and debug prioritycache =5 doesn't indicate any obvious > problems with autotuning code), it may require valgrind or some other > method to figure out where the memory is going. If the memory is > growing rapidly it's possible that wallclock profiling may help if you > can catch where the allocations are being made. > > > Mark > > > On 11/16/21 2:42 PM, Josh Baergen wrote: > > Hi Marius, > > > >> However, docker stats reports 38GB for that container. > >> There is a huge gap between what RAM is being used by the container > what ceph daemon osd.xxx dump_mempools reports. > > Take a look at "ceph daemon osd.XX heap stats" and see what it says. > > You might try "ceph daemon osd.XX heap release"; I didn't think that > > was supposed to be necessary with Bluestore, though. This is reaching > > the end of the sort of problems I know how to track down, though, so > > maybe others have some ideas. > > > >> How can I check if trim happens? > > I'm not sure how to dig into this, but if your "up" count = your "in" > > count in "ceph -s", it should be trimming. > > > > Josh > > _______________________________________________ > > ceph-users mailing list -- ceph-users@xxxxxxx > > To unsubscribe send an email to ceph-users-leave@xxxxxxx > > > > _______________________________________________ > ceph-users mailing list -- ceph-users@xxxxxxx > To unsubscribe send an email to ceph-users-leave@xxxxxxx > _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx