One thing that jumps out at me is using the S3700 for OS but the S3500 for journals. I would use the S3700 for journals and S3500 for the OS. Looks pretty good other than that!
From: "German Anders" <ganders@xxxxxxxxxxxx>
To: "ceph-users" <ceph-users@xxxxxxxxxxxxxx>
Sent: Monday, June 29, 2015 12:24:41 PM
Subject: infiniband implementation
To: "ceph-users" <ceph-users@xxxxxxxxxxxxxx>
Sent: Monday, June 29, 2015 12:24:41 PM
Subject: infiniband implementation
hi cephers,
Want to know if there's any 'best' practice or procedure to implement Ceph with Infiniband FDR 56gb/s for front and back end connectivity. Any crush tunning parameters, etc.
The Ceph cluster has:
- 8 OSD servers
- 2x Intel Xeon E5 8C with HT
- 2x Intel Xeon E5 8C with HT
- 128G RAM
- 2x 200G Intel DC S3700 (RAID-1) OS
- 3x 200G Intel DC S3500 - Journals
- 4x 800G Intel DC S3500 - OSD SSD & Journal on same disks
- 4x 3TB - OSD SATA
- 1x IB FDR ADPT DP
- 3 MON servers
- 2x Intel Xeon E5 6C with HT
- 128G RAM
- 2x 200G Intel SSD (RAID-1) OS
- 1x IB FDR ADP DP
All with Ubuntu 14.04.1LTS with Kern 4.0.6
Thanks in advance,
German
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com