Hi, ceph speeds up with more nodes and more OSDs - so go for 6 nodes with mixed SSD+SATA. Udo On 23.03.2017 18:55, Alejandro Comisario wrote: > Hi everyone! > I have to install a ceph cluster (6 nodes) with two "flavors" of > disks, 3 servers with SSD and 3 servers with SATA. > > Y will purchase 24 disks servers (the ones with sata with NVE SSD for > the SATA journal) > Processors will be 2 x E5-2620v4 with HT, and ram will be 20GB for the > OS, and 1.3GB of ram per storage TB. > > The servers will have 2 x 10Gb bonding for public network and 2 x 10Gb > for cluster network. > My doubts resides, ar want to ask the community about experiences and > pains and gains of choosing between. > > Option 1 > 3 x servers just for SSD > 3 x servers jsut for SATA > > Option 2 > 6 x servers with 12 SSD and 12 SATA each > > Regarding crushmap configuration and rules everything is clear to make > sure that two pools (poolSSD and poolSATA) uses the right disks. > > But, what about performance, maintenance, architecture scalability, etc ? > > thank you very much ! > _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com