Hi Joachim Thank you for sharing the information. I appreciate it. I have successfully activated the OSD after trimming, and it's a really useful tool." Regards Mosharaf Hossain Manager, Product Development Bangladesh Online (BOL) Level 8, SAM Tower, Plot 4, Road 22, Gulshan 1, Dhaka 1212, Bangladesh Tel: +880 9609 000 999, +880 2 58815559, Ext: 14191, Fax: +880 2 2222 95757 Cell: +880 1787 680828, Web: www.bol-online.com On Wed, Oct 30, 2024 at 1:58 PM Joachim Kraftmayer < joachim.kraftmayer@xxxxxxxxx> wrote: > Hi Mosharaf, > > read this article to identify if you are facing this issue: > https://docs.clyso.com/blog/osds-with-unlimited-ram-growth/ > > Regards, Joachim > > > www.clyso.com > > Hohenzollernstr. 27, 80801 Munich > > Utting | HR: Augsburg | HRB: 25866 | USt. ID-Nr.: DE275430677 > > > > Am Mi., 30. Okt. 2024 um 08:27 Uhr schrieb Md Mosharaf Hossain < > mosharaf.hossain@xxxxxxxxxxxxxx>: > >> Dear Ceph Community, >> >> I hope this message finds you well. >> >> I am encountering an out-of-memory (OOM) issue with one of my Ceph OSDs, >> which is repeatedly getting killed by the OOM killer on my system. Below >> are the relevant details from the log: >> >> *OOM Log*: >> [Wed Oct 30 13:14:48 2024] >> >> oom-kill:constraint=CONSTRAINT_NONE,nodemask=(null),cpuset=/,mems_allowed=0,global_oom,task_memcg=/system.slice/system-ceph\x2dosd.slice,task=ceph-osd,pid=6213,uid=64045 >> [Wed Oct 30 13:14:48 2024] Out of memory: Killed process 6213 (ceph-osd) >> total-vm:216486528kB, anon-rss:211821164kB, file-rss:0kB, shmem-rss:0kB, >> UID:64045 pgtables:418836kB oom_score_adj:0 >> [Wed Oct 30 13:14:58 2024] oom_reaper: reaped process 6213 (ceph-osd), now >> anon-rss:0kB, file-rss:0kB, shmem-rss:0kB >> >> *Ceph OSD Log*: >> >> 2024-10-30T13:15:30.207+0600 7f906c74dd80 0 _get_class not permitted to >> load lua >> 2024-10-30T13:15:30.211+0600 7f906c74dd80 0 <cls> >> /build/ceph-15.2.17/src/cls/hello/cls_hello.cc:312: loading cls_hello >> 2024-10-30T13:15:30.215+0600 7f906c74dd80 0 _get_class not permitted to >> load kvs >> 2024-10-30T13:15:30.219+0600 7f906c74dd80 0 _get_class not permitted to >> load queue >> 2024-10-30T13:15:30.223+0600 7f906c74dd80 0 <cls> >> /build/ceph-15.2.17/src/cls/cephfs/cls_cephfs.cc:198: loading cephfs >> 2024-10-30T13:15:30.223+0600 7f906c74dd80 0 osd.13 299547 crush map has >> features 432629239337189376, adjusting msgr requires for clients >> 2024-10-30T13:15:30.223+0600 7f906c74dd80 0 osd.13 299547 crush map has >> features 432629239337189376 was 8705, adjusting msgr requires for mons >> 2024-10-30T13:15:30.223+0600 7f906c74dd80 0 osd.13 299547 crush map has >> features 3314933000854323200, adjusting msgr requires for osds >> 2024-10-30T13:15:30.223+0600 7f906c74dd80 1 osd.13 299547 >> check_osdmap_features require_osd_release unknown -> octopus >> 2024-10-30T13:15:31.023+0600 7f906c74dd80 0 osd.13 299547 load_pgs >> *Environment Details*: >> >> - Ceph Version: 15.2.17 (Octopus) >> - OSD: osd.13 >> - Kernel: Linux kernel version >> >> It seems that the OSD process is consuming a substantial amount of >> memory (total-vm: >> 216486528kB, anon-rss: 211821164kB), leading to OOM kills on the node. The >> OSD service restarts but continues to showing consumption excessive memory >> and OSD get down. >> >> Could you please provide guidance or suggestions on how to mitigate this >> issue? Are there any known memory management settings, configuration >> adjustments, or OSD-specific tuning parameters that could help prevent >> this >> from recurring? >> >> Any help would be greatly appreciated. >> >> Thank you for your time and assistance! >> >> >> >> Regards >> Mosharaf Hossain >> Manager, Product Development >> Bangladesh Online (BOL) >> >> Level 8, SAM Tower, Plot 4, Road 22, Gulshan 1, Dhaka 1212, Bangladesh >> Tel: +880 9609 000 999, +880 2 58815559, Ext: 14191, Fax: +880 2 2222 >> 95757 >> Cell: +880 1787 680828, Web: www.bol-online.com >> _______________________________________________ >> ceph-users mailing list -- ceph-users@xxxxxxx >> To unsubscribe send an email to ceph-users-leave@xxxxxxx >> > _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx