Better way to use osd's of different size

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi!


We have a small production ceph cluster, based on firefly release.


It was built using hardware we already have in our site so it is not "new & shiny",

but works quite good. It was started in 2014.09 as a "proof of concept" from 4 hosts

with 3 x 1tb osd's each: 1U dual socket Intel 54XX & 55XX platforms on 1 gbit network.


Now it contains 4x12 osd nodes on shared 10Gbit network. We use it as a backstore

for running VMs under qemu+rbd.


During migration we temporarily use 1U nodes with 2tb osds and already face some

problems with uneven distribution. I know, that the best practice is to use osds of same

capacity, but it is impossible sometimes.


Now we have 24-28 spare 2tb drives and want to increase capacity on the same boxes.

What is the more right way to do it:

- replace 12x1tb drives with 12x2tb drives, so we will have 2 nodes full of 2tb drives and

other nodes remains in 12x1tb confifg

- or replace 1tb to 2tb drives in more unify way, so every node will have 6x1tb + 6x2tb drives?


I feel that the second way will give more smooth distribution among the nodes, and

outage of one node may give lesser impact on cluster. Am I right and what you can

advice me in such a situation?




Megov Igor
yuterra.ru, CIO
megov@xxxxxxxxxx
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux