CRUSH map advice

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Our experimental Ceph cluster is performing terribly (with the operator 
to blame!), and while it's down to address some issues, I'm curious to 
hear advice about the following ideas.

The cluster:
- two disk nodes (6 * CPU, 16GB RAM each)
- 8 OSDs (4 each)
- 3 monitors
- 10Gb front + back networks
- 2TB Enterprise SATA drives
- HP RAID controller w/battery-backed cache
- one SSD journal drive for each two OSDs

First, I'd like to play with taking one machine down, but with the other 
node continuing to serve the cluster.  To maintain redundancy in this 
scenario, I'm thinking of setting the pool size to 4 and the min_size to 
2, with the idea that a proper CRUSH map should always keep two copies 
on each disk node.  Again, *this is for experimentation* and probably 
raises red flags for production, but I'm just asking if it's *possible*: 
  Could one node go down and the other node continue to serve r/w data? 
  Any anecdotes of performance differences between size=4 and size=3 in 
other clusters?

Second, does it make any sense to divide the CRUSH map into an extra 
level for the SSD disks, which each hold journals for two OSDs?  This 
might increase redundancy in case of a journal disk failure, but ISTR 
something about too few OSDs in a bucket causing problems with the CRUSH 
algorithm.

Thanks-

	John


[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux