cache pools on hypervisor servers

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Thu, 14 Aug 2014, Andrei Mikhailovsky wrote:
> Hi guys,
> 
> Could someone from the ceph team please comment on running osd cache pool on
> the hypervisors? Is this a good idea, or will it create a lot of performance
> issues?

It doesn't sound like an especially good idea.  In general you want the 
cache pool to be significantly faster than the base pool (think PCI 
attached flash).  And there won't be any particular affinity to the host 
where the VM consuming the sotrage happens to be, so I don't think there 
is a reason to put the flash in the hypervisor nodes unless there simply 
isn't anywhere else to put them.

Probably what you're after is a client-side write-thru cache?  There is 
some ongoing work to build this into qemu and possibly librbd, but nothing 
is ready yet that I know of.

sage


> 
> Anyone in the ceph community that has done this? Any results to share?
> 
> Many thanks
> 
> Andrei
> 
> ____________________________________________________________________________
>       From: "Robert van Leeuwen" <Robert.vanLeeuwen at spilgames.com>
>       To: "Andrei Mikhailovsky" <andrei at arhont.com>
>       Cc: ceph-users at lists.ceph.com
>       Sent: Thursday, 14 August, 2014 9:31:24 AM
>       Subject: RE: cache pools on hypervisor servers
> 
> > Personally I am not worried too much about the hypervisor -
> hypervisor traffic as I am using a dedicated infiniband network for
> storage.
> > It is not used for the guest to guest or the internet traffic or
> anything else. I would like to decrease or at least smooth out the
> traffic peaks between the hypervisors and the SAS/SATA osd storage
> servers.
> > I guess the ssd cache pool would enable me to do that as the
> eviction rate should be more structured compared to the random io
> writes that guest vms generate. Sounds reasonable
> 
> >>I'm very interested in the effect of caching pools in combination
> with running VMs on them so I'd be happy to hear what you find ;)
> > I will give it a try and share back the results when we get the ssd
> kit.
> Excellent, looking forward to it.
> 
> 
> >> As a side note: Running OSDs on hypervisors would not be my
> preferred choice since hypervisor load might impact Ceph performance.
> > Do you think it is not a good idea even if you have a lot of cores
> on the hypervisors?
> > Like 24 or 32 per host server?
> > According to my monitoring, our osd servers are not that stressed
> and generally have over 50% of free cpu power.
> 
> The number of cores do not really matter if they are all busy ;)
> I honestly do not know how Ceph behaves when it is CPU starved but I
> guess it might not be pretty.
> Since your whole environment will be crumbling down if your storage
> becomes unavailable it is not a risk I would take lightly.
> 
> Cheers,
> Robert van Leeuwen
> 
> 
> 
> 
> 
> 
> 


[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux