Unexplainable high memory usage OSD with BlueStore

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,

Recently I've seen a Ceph cluster experience a few outages due to memory
issues.

The machines:

- Intel Xeon E3 CPU
- 32GB Memory
- 8x 1.92TB SSD
- Ubuntu 16.04
- Ceph 12.2.8

Looking at one of the machines:

root@ceph22:~# free -h
              total        used        free      shared  buff/cache
available
Mem:            31G         22G        8.2G        9.8M        809M
  8.5G
Swap:          1.9G          0B        1.9G
root@ceph22:~#

As you can see, it's using 22GB of the 32GB in the system.

[osd]
bluestore_cache_size_ssd = 1G

The BlueStore Cache size for SSD has been set to 1GB, so the OSDs
shouldn't use more then that.

When dumping the mem pools each OSD claims to be using between 1.8GB and
2.2GB of memory.

$ ceph daemon osd.X dump_mempools|jq '.total.bytes'

Summing up all the values I get to a total of 15.8GB and the system is
using 22GB.

Looking at 'ps aux --sort rss' I see OSDs using almost 10% of the
memory, which would be ~3GB for a single daemon.

Sometimes these machines go OOM without a really good reason. After all
the OSDs have been restarted it runs fine, but about 10 days later the
same problem arrises again.

I know that BlueStore uses more memory, but 32GB with 8 OSDs should be a
workable setup.

Any ideas?

Wido
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux