@Paul Emmerich Thank you very much!. That did the trick! On Sat, Jun 8, 2019 at 4:59 AM huang jun <hjwsm1989@xxxxxxxxx> wrote: > > Did you osd oom killed when cluster doing recover/backfill, or just > the client io? > The configure items you mentioned is for bluestore and the osd memory > include many other > things, like pglog, you it's important to known do you cluster is dong recover? > > Sergei Genchev <sgenchev@xxxxxxxxx> 于2019年6月8日周六 上午5:35写道: > > > > Hi, > > My OSD processes are constantly getting killed by OOM killer. My > > cluster has 5 servers, each with 18 spinning disks, running 18 OSD > > daemons in 48GB of memory. > > I was trying to limit OSD cache, according to > > http://docs.ceph.com/docs/mimic/rados/configuration/bluestore-config-ref/ > > > > [osd] > > bluestore_cache_size_ssd = 1G > > bluestore_cache_size_hdd = 768M > > > > Yet, my OSDs are using way more memory than that. I have seen as high as 3.2G > > > > KiB Mem : 47877604 total, 310172 free, 45532752 used, 2034680 buff/cache > > KiB Swap: 2097148 total, 0 free, 2097148 used. 950224 avail Mem > > > > PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ > > COMMAND > > 352516 ceph 20 0 3962504 2.8g 4164 S 2.3 6.1 4:22.98 > > ceph-osd > > 350771 ceph 20 0 3668248 2.7g 4724 S 3.0 6.0 3:56.76 > > ceph-osd > > 352777 ceph 20 0 3659204 2.7g 4672 S 1.7 5.9 4:10.52 > > ceph-osd > > 353578 ceph 20 0 3589484 2.6g 4808 S 4.6 5.8 3:37.54 > > ceph-osd > > 352280 ceph 20 0 3577104 2.6g 4704 S 5.9 5.7 3:44.58 > > ceph-osd > > 350933 ceph 20 0 3421168 2.5g 4140 S 2.6 5.4 3:38.13 > > ceph-osd > > 353678 ceph 20 0 3368664 2.4g 4804 S 4.0 5.3 12:47.12 > > ceph-osd > > 350665 ceph 20 0 3364780 2.4g 4716 S 2.6 5.3 4:23.44 > > ceph-osd > > 353101 ceph 20 0 3304288 2.4g 4676 S 4.3 5.2 3:16.53 > > ceph-osd > > ....... > > > > > > Is there any way for me to limit how much memory does OSD use? > > Thank you! > > > > ceph version 13.2.4 (b10be4d44915a4d78a8e06aa31919e74927b142e) mimic (stable) > > _______________________________________________ > > ceph-users mailing list > > ceph-users@xxxxxxxxxxxxxx > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > > > -- > Thank you! > HuangJun _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com