Re: Set cache tier pool forward state automatically!

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Wed, 3 Feb 2016 16:57:09 -0700 Robert LeBlanc wrote:

> -----BEGIN PGP SIGNED MESSAGE-----
> Hash: SHA256
> 
> My experience with Hammer is showing that setting the pool to forward
> mode is not evicting objects, nor do I think it is flushing objects.
>
Same here (with Firefly).

> We have had our pool in forward mode for weeks now and we still have
> almost the same amount of I/O to it. There has been a slight shift
> between SSD and HDD, but I think that is because some objects have
> cooled off and others have been newly accessed. You may have better
> luck adjusting the ratios, but we see that there is a big hit to our
> cluster to do that. We usually do 1% every minute or two to help
> reduce the impact of evicting the data (we usually drop the cache full
> ratio 10% or so to evict some objects and we then toggle the cache
> mode between writeback and forward periodically to warm up the cache.
> Setting it to writeback will promote so many objects at once that it
> severely impact our cluster. There is also a limit that we reach at
> about 10K IOPs when in writeback where with forward I've seen spikes
> to 64K IOPs. So we turn on writeback for 30-60 seconds (or until the
> blocked I/O is too much for us to handle), then set it to forward for
> 60-120 second, rinse and repeat until the impact of writeback isn't so
> bad, then set it back to forward for a couple more weeks).
>
That's an interesting strategy, I suppose you haven't run into the issue I
wrote about 2 days ago when switching to forward while running rdb bench?

In my case I venture that the number of really hot objects is small enough
to not overwhelm things and that 5K IOPS would be all that cluster ever
needs to provide.

Regards,

Christian
 
> Needless to say, cache tiering could use some more love. If I get some
> time, I'd like to try and help that section of code, but I have a
> couple other more pressing issues I'd like to track down first.
> -----BEGIN PGP SIGNATURE-----
> Version: Mailvelope v1.3.4
> Comment: https://www.mailvelope.com
> 
> wsFcBAEBCAAQBQJWspPSCRDmVDuy+mK58QAAsQgP/15YrzV+BRt+CGnzZL/Q
> w6PwnSdw4HBJT4OEqdg+kStCP+SqUSVCiJcdeHo5Sm40smEWVYRim3jsHBSg
> Z4Woa31XsjYbEw3HCxIoI93OPhaKszOhvktKZxu1iSnyMDDJIYMARlYIjbfc
> ToCOC/IVe2MMAEtVq+J2fm/NQy6VDGbaUuYcNtkIF41j7vKoNoE3h5qi+L0K
> cVwUhVTcuSNDuiuJOoduM/vSH6nJzmCnypH1BDTcEOYpvmbXWJ0iTdej2Oa1
> gVvV7SOcu4PkjzL9MmJB2Cjiiy/zWjUTfN01nBvIatwOjF7AE8vq2pLD9FIs
> TxmzE4UZgjwJNbkDVQsgHPCeUlEll+t3QKbokpEkQDQgvIOs6NCbj0KYpuhC
> DWtQCbgYsniT+Md1vWFMgqs0a45ulGxEKUWiUOEXgTJLHH+dbrW32MZEl1Gd
> yTKyzFarbae6tbAmaMPC8l9vaj15t7bAB0KOokMqZied7EcM1ZoFVqKRahrm
> 73mIeHiDUwZ8gi+BHKX7OwqKt3VZJYf/+rNJx+g4kp5WN0FEkUMoqF75qO4p
> 62+PuQIwh6jUpB4cDsbEJd78UGbCptJBojmsNVogU+xiSXTKQmEduP0HqQfG
> JhTLg3Un2C4/MSGbhRI26csFCzEi66iRXQWdfCITP4Um70KO6dE2C1MAveYg
> hJ7b
> =CaRF
> -----END PGP SIGNATURE-----
> ----------------
> Robert LeBlanc
> PGP Fingerprint 79A2 9CA4 6CC4 45DD A904  C70E E654 3BB2 FA62 B9F1
> 
> 
> On Wed, Feb 3, 2016 at 10:01 AM, Nick Fisk <nick@xxxxxxxxxx> wrote:
> > I think this would be better to be done outside of Ceph. It should be
> > quite simple for whatever monitoring software you are using to pick up
> > the disk failure to set the target_dirty_ratio to a very low value or
> > change the actual caching mode.
> >
> > Doing it in Ceph would be complicated as you are then asking Ceph to
> > decide when you are in an at risk scenario, ie would you want it to
> > flush your cache after a quick service reload or node reboot?
> >
> >> -----Original Message-----
> >> From: ceph-users [mailto:ceph-users-bounces@xxxxxxxxxxxxxx] On Behalf
> >> Of Mihai Gheorghe
> >> Sent: 03 February 2016 16:57
> >> To: ceph-users <ceph-users@xxxxxxxxxxxxxx>; ceph-users <ceph-
> >> users@xxxxxxxx>
> >> Subject:  Set cache tier pool forward state automatically!
> >>
> >> Hi,
> >>
> >> Is there a built in setting in ceph that would set the cache pool from
> >> writeback to forward state automatically in case of an OSD fail from
> >> the pool?
> >>
> >> Let;s say the size of the cache pool is 2. If an OSD fails ceph
> >> blocks write to the pool, making the VM that use this pool to be
> >> unaccesable. But an earlier copy of the data is present on the cold
> >> storage pool prior to the last cache flush.
> >>
> >> In this case, is it possible that when an OSD fails, the data on the
> >> cache pool to be flushed onto the cold storage pool and set the
> >> forward flag automatically on the cache pool? So that the VM can
> >> resume write to the block device as soon as the cache is flushed from
> >> the pool and read/write directly from the cold storage pool untill
> >> manual intervention on the cache pool is done to fix it and set it
> >> back to writeback?
> >>
> >> This way we can get away with a pool size of 2 without worrying for
> >> too much downtime!
> >>
> >> Hope i was explicit enough!
> >
> > _______________________________________________
> > ceph-users mailing list
> > ceph-users@xxxxxxxxxxxxxx
> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> _______________________________________________
> ceph-users mailing list
> ceph-users@xxxxxxxxxxxxxx
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> 


-- 
Christian Balzer        Network/Systems Engineer                
chibi@xxxxxxx   	Global OnLine Japan/Rakuten Communications
http://www.gol.com/
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux