Re: bcache vs flashcache vs cache tiering

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



> -----Original Message-----
> From: Christian Balzer [mailto:chibi@xxxxxxx]
> Sent: 15 February 2017 01:42
> To: 'Ceph Users' <ceph-users@xxxxxxxxxxxxxx>
> Cc: Nick Fisk <nick@xxxxxxxxxx>; 'Gregory Farnum' <gfarnum@xxxxxxxxxx>
> Subject: Re:  bcache vs flashcache vs cache tiering
> 
> On Tue, 14 Feb 2017 22:42:21 -0000 Nick Fisk wrote:
> 
> > > -----Original Message-----
> > > From: Gregory Farnum [mailto:gfarnum@xxxxxxxxxx]
> > > Sent: 14 February 2017 21:05
> > > To: Wido den Hollander <wido@xxxxxxxx>
> > > Cc: Dongsheng Yang <dongsheng.yang@xxxxxxxxxxxx>; Nick Fisk
> > > <nick@xxxxxxxxxx>; Ceph Users <ceph-users@xxxxxxxxxxxxxx>
> > > Subject: Re:  bcache vs flashcache vs cache tiering
> > >
> > > On Tue, Feb 14, 2017 at 8:25 AM, Wido den Hollander <wido@xxxxxxxx>
> > > wrote:
> > > >
> > > >> Op 14 februari 2017 om 11:14 schreef Nick Fisk <nick@xxxxxxxxxx>:
> > > >>
> > > >>
> > > >> > -----Original Message-----
> > > >> > From: ceph-users [mailto:ceph-users-bounces@xxxxxxxxxxxxxx] On
> > > >> > Behalf Of Dongsheng Yang
> > > >> > Sent: 14 February 2017 09:01
> > > >> > To: Sage Weil <notifications@xxxxxxxxxx>
> > > >> > Cc: ceph-devel@xxxxxxxxxxxxxxx; ceph-users@xxxxxxxxxxxxxx
> > > >> > Subject:  bcache vs flashcache vs cache tiering
> > > >> >
> > > >> > Hi Sage and all,
> > > >> >      We are going to use SSDs for cache in ceph. But I am not
> > > >> > sure which one is the best solution, bcache? flashcache? or
> > > >> > cache
> > > >> tier?
> > > >>
> > > >> I would vote for cache tier. Being able to manage it from within
> > > >> Ceph, instead of having to manage X number of bcache/flashcache
> > > >> instances, appeals to me more. Also last time I looked Flashcache
> > > >> seems unmaintained and bcache might be going that way with talk
> > > >> of this new bcachefs. Another point to consider is that Ceph has
> > > >> had a lot of
> > > work done on it to ensure data consistency; I don't ever want to be
> > > in a position where I'm trying to diagnose problems that might be
> > > being caused by another layer sitting in-between Ceph and the Disk.
> > > >>
> > > >> However, I know several people on here are using bcache and
> > > >> potentially getting better performance than with cache tiering,
> > > >> so
> > > hopefully someone will give their views.
> > > >
> > > > I am using Bcache on various systems and it performs really well.
> > > > The
> > > caching layer in Ceph is slow. Promoting Objects is slow and it also
> > > involves additional RADOS lookups.
> > >
> > > Yeah. Cache tiers have gotten a lot more usable in Ceph, but the use
> > > cases where they're effective are still pretty limited and I think
> > > in-node caching has a brighter future. We just don't like to
> > > maintain the global state that makes separate caching locations
> > > viable and unless you're doing something analogous to the
> > > supercomputing "burst buffers" (which some people are!), it's going
> > > to be hard to beat something that doesn't have to pay the cost of extra network hops/bandwidth.
> > > Cache tiers are also not a feature that all the vendors support in
> > > their downstream products, so it will probably see less ongoing
> > > investment than you'd expect from such a system.
> >
> > Should that be taken as an unofficial sign that the tiering support is likely to fade away?
> >
> Nick, you also posted back in October in the "cache tiering deprecated in RHCS 2.0" thread and should remember the deafening
> silence when I asked that question.
> 
> I'm actually surprised that Greg said as much as he did now, unfortunately that doesn't really cover all the questions I had back then, in
> particular long term support and bug fixes, not necessarily more features.
> 
> We're literally about to order our next cluster and cache-tiering works like a charm for us, even in Hammer.
> With the (still undocumented) knobs in Jewel and read-forward it will be even more effective.
> 
> So given the lack of any statements that next cluster will still use the same design as the previous one, since BlueStore isn't ready,
> bcache and others haven't been tested here to my satisfaction and we know very well what works and what not.
> 
> So 3 regular (HDD OSD, journal SSD) nodes and 3 cache-tier ones. Dedicated cache-tier nodes allow for deployment of high end CPUs
> only in those nodes.
> 
> Another point in favor of cache-tiering is that it can be added at a later stage, while in-node caching requires an initial design with large
> local SSDs/NVMes or at least the space for them.
> Because the journal SSDs most people will deploy initially don't tend to be large enough to be effective when used with bcache or
> similar.

And I think that is the main advantage of tiering, you have a lot more flexibility both during implementation and further down the road. We can look at our hot tier hit ratio and just add more SSD's, either in the spare slots I leave in each chassis or like you have suggested, dedicated SSD nodes.

> 
> > I think both approaches have different strengths and probably the difference between a tiering system and a caching one is what
> causes some of the problems.
> >
> > If something like bcache is going to be the preferred approach, then I think more work needs to be done around certifying it for use
> with Ceph and allowing its behavior to be more controlled by Ceph as well. I assume there are issues around backfilling and scrubbing
> polluting the cache? Maybe you would want to be able to pass hints down from Ceph, which could also allow per pool cache
> behavior??
> >
> According to the RHCS release notes back then their idea to achieve rainbows and pink ponies was using dm-cache.
> 
> Christian
> --
> Christian Balzer        Network/Systems Engineer
> chibi@xxxxxxx   	Global OnLine Japan/Rakuten Communications
> http://www.gol.com/

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux