Hi, I've got 16 nodes cluster ssd only. Each node is 6x600GB, 10Gbit uplink. We're using Intel 320 series. Cluster is running now half year as production and no problems with ssds. Replication x3 (main DC) and x2 in backup DC (10 nodes cluster there = less space). >From what I've notice it's just easier to remove entire node from cluster and rebuild it comparing to work on it or try to fix it. With 10gbit uplink 3.6TB is rebalanced after 40-60min. During rebalance we've been able to saturated 10Gbit network. On Thu, May 22, 2014 at 9:00 AM, Alexandre DERUMIER <aderumier at odiso.com>wrote: > Hi, > > I'm looking to build a full osd ssd cluster, with this config: > > 6 nodes, > > each node 10 osd/ ssd drives (dual 10gbit network). (1journal + datas on > each osd) > > ssd drive will be entreprise grade, > > maybe intel sc3500 800GB (well known ssd) > > or new Samsung SSD PM853T 960GB (don't have too much info about it for the > moment, but price seem a little bit lower than intel) > > > I would like to have some advise on replication level, > > > Maybe somebody have experience with intel sc3500 failure rate ? > How many chance to have 2 failing disks on 2 differents nodes at the same > time (murphy's law ;). > > > I think in case of disk failure, pgs should replicate fast with 10gbits > links. > > > So the question is: > > 2x or 3x ? > > > Regards, > > Alexandre > _______________________________________________ > ceph-users mailing list > ceph-users at lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > -- ?ukasz Jagie??o lukasz<at>jagiello<dot>org -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://lists.ceph.com/pipermail/ceph-users-ceph.com/attachments/20140527/5737bbe0/attachment.htm>