Re: OSD memory usage

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Please excuse my brain-fart. We're using 24 disks on the servers in question. Only after discussing this further with a colleague did we realize this.

This brings us right to the minimum-spec which generally isn't a good idea.

Sincerely

-Dave


On 11/09/17 11:38 AM, bulk.schulz@xxxxxxxxxxx wrote:
[This sender failed our fraud detection checks and may not be who they appear to be. Learn about spoofing at http://aka.ms/LearnAboutSpoofing]

Hi Everyone,

I wonder if someone out there has a similar problem to this?

I keep having issues with memory usage.  I have 2 OSD servers wiith 48G
memory and 12 2TB OSDs.  I seem to have significantly more memory than
the minimum spec, but these two machines with 2TB drives seem to OOM
kill and crash periodically -- basically any time the cluster goes into
recovery for even 1 OSD this happens.

12 Drives * 2TB = 24 TB.  By using the 1GB RAM per 1TB Disk rule: I
should need only 24TB or so.

I am testing and benchmarking at this time so most changes are fine.  I
am abusing this filesystem considerably by running 14 clients with
something that is more or less dd each to a different file but that's
the point :)

When it's working, the performance is really good.  3GB/s with 3x
replicated data pool up to around 10GB/s with 1X replication (just for
kicks and giggles) My bottleneck is likely the SAS channels to those disks.

I'm using the 12.2.0 release running on Centos 7

Testing cephfs with one MDS and 3 montors.  The MON/MDS are not on the
servers in question.

Total of around 350 OSDs (all spinning disk) most of which are 1TB
drives on 15 servers that are a bit older with Xeon E5620's.

Dual QDR Infiniband (20GBit) fabrics (1 cluster and 1 client).

Any thoughts?  Am I missing some tuning parameter in /proc or something?

Thanks
-Dave
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux