full osd ssd cluster advise : replication 2x or 3x ?

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



>>Hi,

>>I've got 16 nodes cluster ssd only. Each node is 6x600GB, 10Gbit uplink.

>>We're using Intel 320 series. Cluster is running now half year as production and no problems with ssds. 

>>Replication x3 (main DC) and x2 in backup DC (10 nodes cluster there = less space).

>>From what I've notice it's just easier to remove entire node from cluster and rebuild it comparing to work on it or try to fix it. With 10gbit uplink 3.6TB is rebalanced after 40-60min. During rebalance we've >>been able to saturated 10Gbit network.

Thanks for your reply !



----- Mail original ----- 

De: "?ukasz Jagie??o" <jagiello.lukasz at gmail.com> 
?: "Alexandre DERUMIER" <aderumier at odiso.com> 
Cc: ceph-users at lists.ceph.com 
Envoy?: Mercredi 28 Mai 2014 01:25:40 
Objet: Re: full osd ssd cluster advise : replication 2x or 3x ? 


Hi, 


I've got 16 nodes cluster ssd only. Each node is 6x600GB, 10Gbit uplink. 


We're using Intel 320 series. Cluster is running now half year as production and no problems with ssds. 


Replication x3 (main DC) and x2 in backup DC (10 nodes cluster there = less space). 


>From what I've notice it's just easier to remove entire node from cluster and rebuild it comparing to work on it or try to fix it. With 10gbit uplink 3.6TB is rebalanced after 40-60min. During rebalance we've been able to saturated 10Gbit network. 



On Thu, May 22, 2014 at 9:00 AM, Alexandre DERUMIER < aderumier at odiso.com > wrote: 


Hi, 

I'm looking to build a full osd ssd cluster, with this config: 

6 nodes, 

each node 10 osd/ ssd drives (dual 10gbit network). (1journal + datas on each osd) 

ssd drive will be entreprise grade, 

maybe intel sc3500 800GB (well known ssd) 

or new Samsung SSD PM853T 960GB (don't have too much info about it for the moment, but price seem a little bit lower than intel) 


I would like to have some advise on replication level, 


Maybe somebody have experience with intel sc3500 failure rate ? 
How many chance to have 2 failing disks on 2 differents nodes at the same time (murphy's law ;). 


I think in case of disk failure, pgs should replicate fast with 10gbits links. 


So the question is: 

2x or 3x ? 


Regards, 

Alexandre 
_______________________________________________ 
ceph-users mailing list 
ceph-users at lists.ceph.com 
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com 







-- 
?ukasz Jagie??o 
lukasz<at>jagiello<dot>org 


[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux