v0.80 Firefly released

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



>>Do we need a journal when using this back-end? 

no,they are no journal with key value store

----- Mail original ----- 

De: "Kenneth Waegeman" <Kenneth.Waegeman at UGent.be> 
?: "Sage Weil" <sage at inktank.com> 
Cc: ceph-users at ceph.com 
Envoy?: Mercredi 7 Mai 2014 15:06:50 
Objet: Re: v0.80 Firefly released 


----- Message from Sage Weil <sage at inktank.com> --------- 
Date: Tue, 6 May 2014 18:05:19 -0700 (PDT) 
From: Sage Weil <sage@xxxxxxxxxxx> 
Subject: v0.80 Firefly released 
To: ceph-devel at vger.kernel.org, ceph-users at ceph.com 


> We did it! Firefly v0.80 is built and pushed out to the ceph.com 
> repositories. 
> 
> This release will form the basis for our long-term supported release 
> Firefly, v0.80.x. The big new features are support for erasure coding 
> and cache tiering, although a broad range of other features, fixes, 
> and improvements have been made across the code base. Highlights include: 
> 
> * *Erasure coding*: support for a broad range of erasure codes for lower 
> storage overhead and better data durability. 
> * *Cache tiering*: support for creating 'cache pools' that store hot, 
> recently accessed objects with automatic demotion of colder data to 
> a base tier. Typically the cache pool is backed by faster storage 
> devices like SSDs. 
> * *Primary affinity*: Ceph now has the ability to skew selection of 
> OSDs as the "primary" copy, which allows the read workload to be 
> cheaply skewed away from parts of the cluster without migrating any 
> data. 
> * *Key/value OSD backend* (experimental): An alternative storage backend 
> for Ceph OSD processes that puts all data in a key/value database like 
> leveldb. This provides better performance for workloads dominated by 
> key/value operations (like radosgw bucket indices). 

Nice! 
Is there already some documentation about this Key/value OSD back-end 
topic, like how to use, restrictions, ..? 
A question (referring to 
http://www.sebastien-han.fr/blog/2013/12/02/ceph-performance-interesting-things-going-on/): Do we need a journal when using this 
back-end? 
And is this compatible for use with CephFS? 

Thanks! 


> * *Standalone radosgw* (experimental): The radosgw process can now run 
> in a standalone mode without an apache (or similar) web server or 
> fastcgi. This simplifies deployment and can improve performance. 
> 
> We expect to maintain a series of stable releases based on v0.80 
> Firefly for as much as a year. In the meantime, development of Ceph 
> continues with the next release, Giant, which will feature work on the 
> CephFS distributed file system, more alternative storage backends 
> (like RocksDB and f2fs), RDMA support, support for pyramid erasure 
> codes, and additional functionality in the block device (RBD) like 
> copy-on-read and multisite mirroring. 
> 
> This release is the culmination of a huge collective effort by about 100 
> different contributors. Thank you everyone who has helped to make this 
> possible! 
> 
> Upgrade Sequencing 
> ------------------ 
> 
> * If your existing cluster is running a version older than v0.67 
> Dumpling, please first upgrade to the latest Dumpling release before 
> upgrading to v0.80 Firefly. Please refer to the :ref:`Dumpling upgrade` 
> documentation. 
> 
> * Upgrade daemons in the following order: 
> 
> 1. Monitors 
> 2. OSDs 
> 3. MDSs and/or radosgw 
> 
> If the ceph-mds daemon is restarted first, it will wait until all 
> OSDs have been upgraded before finishing its startup sequence. If 
> the ceph-mon daemons are not restarted prior to the ceph-osd 
> daemons, they will not correctly register their new capabilities 
> with the cluster and new features may not be usable until they are 
> restarted a second time. 
> 
> * Upgrade radosgw daemons together. There is a subtle change in behavior 
> for multipart uploads that prevents a multipart request that was initiated 
> with a new radosgw from being completed by an old radosgw. 
> 
> Notable changes since v0.79 
> --------------------------- 
> 
> * ceph-fuse, libcephfs: fix several caching bugs (Yan, Zheng) 
> * ceph-fuse: trim inodes in response to mds memory pressure (Yan, Zheng) 
> * librados: fix inconsistencies in API error values (David Zafman) 
> * librados: fix watch operations with cache pools (Sage Weil) 
> * librados: new snap rollback operation (David Zafman) 
> * mds: fix respawn (John Spray) 
> * mds: misc bugs (Yan, Zheng) 
> * mds: misc multi-mds fixes (Yan, Zheng) 
> * mds: use shared_ptr for requests (Greg Farnum) 
> * mon: fix peer feature checks (Sage Weil) 
> * mon: require 'x' mon caps for auth operations (Joao Luis) 
> * mon: shutdown when removed from mon cluster (Joao Luis) 
> * msgr: fix locking bug in authentication (Josh Durgin) 
> * osd: fix bug in journal replay/restart (Sage Weil) 
> * osd: many many many bug fixes with cache tiering (Samuel Just) 
> * osd: track omap and hit_set objects in pg stats (Samuel Just) 
> * osd: warn if agent cannot enable due to invalid (post-split) stats 
> (Sage Weil) 
> * rados bench: track metadata for multiple runs separately (Guang Yang) 
> * rgw: fixed subuser modify (Yehuda Sadeh) 
> * rpm: fix redhat-lsb dependency (Sage Weil, Alfredo Deza) 
> 
> For the complete release notes, please see: 
> 
> http://ceph.com/docs/master/release-notes/#v0-80-firefly 
> 
> 
> Getting Ceph 
> ------------ 
> 
> * Git at git://github.com/ceph/ceph.git 
> * Tarball at http://ceph.com/download/ceph-0.80.tar.gz 
> * For packages, see http://ceph.com/docs/master/install/get-packages 
> * For ceph-deploy, see 
> http://ceph.com/docs/master/install/install-ceph-deploy 
> 
> _______________________________________________ 
> ceph-users mailing list 
> ceph-users at lists.ceph.com 
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com 


----- End message from Sage Weil <sage at inktank.com> ----- 

-- 

Met vriendelijke groeten, 
Kenneth Waegeman 

_______________________________________________ 
ceph-users mailing list 
ceph-users at lists.ceph.com 
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com 


[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux