Re: pros/cons of multiple OSD's per host

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 





On Wed, Aug 23, 2017 at 2:28 PM, Christian Balzer <chibi@xxxxxxx> wrote:
On Wed, 23 Aug 2017 13:38:25 +0800 Nick Tan wrote:

> Thanks for the advice Christian.  I think I'm leaning more towards the
> 'traditional' storage server with 12 disks - as you say they give a lot
> more flexibility with the performance tuning/network options etc.
>
> The cache pool is an interesting idea but as you say it can get quite
> expensive for the capacities we're looking at.  I'm interested in how
> bluestore performs without a flash/SSD WAL/DB.  In my small scale testing
> it seems much better than filestore so I was planning on building something
> without any flash/SSD.  There's always the option of adding it later if
> required.
>
Given the lack (for large writes) of double writes with Bluestore that's
to be expected.

Since you're looking mostly at largish, sequential writes and reads, a
pure HDDs cluster may be feasible.



I have a final question.  Assume I'll have 96 OSD's of 10TB each and I use erasure coding with k=5 and m=3 with the failure domain of host.

It's my understanding that with erasure coding and host failure domains, the minimum number of hosts required is equal to k+m.  Is this correct?

So, if I use 8 hosts with 12 OSDs each this means I can lose any 3 OSD's but can I lose a host, assuming there's enough free space in the pool to accomodate the missing 12 OSD's?  Or will the cluster block at this point because there's only 7 hosts which is less than k+m?

If I use 48 hosts with 2 OSDs each then I can lose up to any 3 OSD's and the cluster can recover.  And I could lose 1 host at a time assuming the recovery completes before the next host is lost (until there's 7 hosts left or there's not enough free space in the pool).  Is this a correct assessment?

Thanks,
Nick


_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux