Re: Disk/Pool Layout

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Some comments inline.
A lot of it depends on your workload, but I'd say you almost certainly need higher-grade SSDs. You can save money on memory.

What will be the role of this cluster? VM disks? Object storage? Streaming?...

Jan

On 27 Aug 2015, at 17:56, German Anders <ganders@xxxxxxxxxxxx> wrote:

Hi all,

   I'm planning to deploy a new Ceph cluster with IB FDR 56Gb/s and I've the following HW:

3x MON Servers:
   2x Intel Xeon E5-2600@v3 8C
   256GB RAM

I don't think you need that much memory, 64GB should be plenty (if that's the only role for the servers).

   1xIB FRD ADPT-DP (two ports for PUB network)
   1xGB ADPT-DP
  
   Disk Layout:
  
   SOFT-RAID:
   SCSI1 (0,0,0) (sda) - 120.0 GB ATA INTEL SSDSC2BB12 (OS-RAID1)
   SCSI2 (0,0,0) (sdb) - 120.0 GB ATA INTEL SSDSC2BB12 (OS-RAID1)

I 100% recommend going with SSDs for the /var/lib/ceph/mon storage, fast ones (but they can be fairly small). Should be the same grade as journal drives IMO.
NOT S3500!
I can recommend S3610 (just got some :)), Samsung 845 DC PRO. At least 1 DWPD rating, better go with 3 DWPD.


8x OSD Servers:
   2x Intel Xeon E5-2600@v3 10C

Go for the fastest you can afford if you need the latency - even at the expense of cores.
Go for cores if you want bigger throughput.

   256GB RAM

Again - I think too much if that's the only role for those nodes, 64GB should be plenty.


   1xIB FRD ADPT-DP (one port for PUB and one for CLUS network)
   1xGB ADPT-DP

   Disk Layout:

   SOFT-RAID:
   SCSI1 (0,0,0) (sda) - 120.0 GB ATA INTEL SSDSC2BB12 (OS-RAID1)
   SCSI2 (0,0,0) (sdb) - 120.0 GB ATA INTEL SSDSC2BB12 (OS-RAID1)

   JBOD:
   SCSI9 (0,0,0) (sdd) - 120.0 GB ATA INTEL SC3500 SSDSC2BB12 (Journal)
   SCSI9 (0,1,0) (sde) - 120.0 GB ATA INTEL SC3500 SSDSC2BB12 (Journal)
   SCSI9 (0,2,0) (sdf) - 120.0 GB ATA INTEL SC3500 SSDSC2BB12 (Journal)

No no no. Those SSDs will die a horrible death, too little endurance.
Better go with 2x 3700 in RAID1 and partition them for journals. Or just don't use journaling drives and buy better SSDs for storage.

   SCSI9 (0,3,0) (sdg) - 800.2 GB ATA INTEL SC3510 SSDSC2BB80 (Pool-SSD)
   SCSI9 (0,4,0) (sdh) - 800.2 GB ATA INTEL SC3510 SSDSC2BB80 (Pool-SSD)
   SCSI9 (0,5,0) (sdi) - 800.2 GB ATA INTEL SC3510 SSDSC2BB80 (Pool-SSD)
   SCSI9 (0,6,0) (sdj) - 800.2 GB ATA INTEL SC3510 SSDSC2BB80 (Pool-SSD)

Too little endurance.


   SCSI9 (0,7,0) (sdk) - 3.0 TB SEAGATE ST3000NM0023 (Pool-SATA)
   SCSI9 (0,8,0) (sdl) - 3.0 TB SEAGATE ST3000NM0023 (Pool-SATA)
   SCSI9 (0,9,0) (sdm) - 3.0 TB SEAGATE ST3000NM0023 (Pool-SATA)
   SCSI9 (0,10,0) (sdn) - 3.0 TB SEAGATE ST3000NM0023 (Pool-SATA)
   SCSI9 (0,11,0) (sdo) - 3.0 TB SEAGATE ST3000NM0023 (Pool-SATA)


I would like to have an expert opinion on what would be the best deploy/config disk pools and crush map? any other advice?

Thanks in advance,

Best regards,

German
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux