v0.80 Firefly released

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



----- Message from Alexandre DERUMIER <aderumier at odiso.com> ---------
    Date: Wed, 07 May 2014 15:21:55 +0200 (CEST)
    From: Alexandre DERUMIER <aderumier at odiso.com>
Subject: Re: v0.80 Firefly released
      To: Kenneth Waegeman <Kenneth.Waegeman at UGent.be>
      Cc: ceph-users at ceph.com, Sage Weil <sage at inktank.com>


>>> Do we need a journal when using this back-end?
>
> no,they are no journal with key value store

Thanks, And how can I activate this?
>
> ----- Mail original -----
>
> De: "Kenneth Waegeman" <Kenneth.Waegeman at UGent.be>
> ?: "Sage Weil" <sage at inktank.com>
> Cc: ceph-users at ceph.com
> Envoy?: Mercredi 7 Mai 2014 15:06:50
> Objet: Re: v0.80 Firefly released
>
>
> ----- Message from Sage Weil <sage at inktank.com> ---------
> Date: Tue, 6 May 2014 18:05:19 -0700 (PDT)
> From: Sage Weil <sage at inktank.com>
> Subject: v0.80 Firefly released
> To: ceph-devel at vger.kernel.org, ceph-users at ceph.com
>
>
>> We did it! Firefly v0.80 is built and pushed out to the ceph.com
>> repositories.
>>
>> This release will form the basis for our long-term supported release
>> Firefly, v0.80.x. The big new features are support for erasure coding
>> and cache tiering, although a broad range of other features, fixes,
>> and improvements have been made across the code base. Highlights include:
>>
>> * *Erasure coding*: support for a broad range of erasure codes for lower
>> storage overhead and better data durability.
>> * *Cache tiering*: support for creating 'cache pools' that store hot,
>> recently accessed objects with automatic demotion of colder data to
>> a base tier. Typically the cache pool is backed by faster storage
>> devices like SSDs.
>> * *Primary affinity*: Ceph now has the ability to skew selection of
>> OSDs as the "primary" copy, which allows the read workload to be
>> cheaply skewed away from parts of the cluster without migrating any
>> data.
>> * *Key/value OSD backend* (experimental): An alternative storage backend
>> for Ceph OSD processes that puts all data in a key/value database like
>> leveldb. This provides better performance for workloads dominated by
>> key/value operations (like radosgw bucket indices).
>
> Nice!
> Is there already some documentation about this Key/value OSD back-end
> topic, like how to use, restrictions, ..?
> A question (referring to
> http://www.sebastien-han.fr/blog/2013/12/02/ceph-performance-interesting-things-going-on/): Do we need a journal when using  
> this
> back-end?
> And is this compatible for use with CephFS?
>
> Thanks!
>
>
>> * *Standalone radosgw* (experimental): The radosgw process can now run
>> in a standalone mode without an apache (or similar) web server or
>> fastcgi. This simplifies deployment and can improve performance.
>>
>> We expect to maintain a series of stable releases based on v0.80
>> Firefly for as much as a year. In the meantime, development of Ceph
>> continues with the next release, Giant, which will feature work on the
>> CephFS distributed file system, more alternative storage backends
>> (like RocksDB and f2fs), RDMA support, support for pyramid erasure
>> codes, and additional functionality in the block device (RBD) like
>> copy-on-read and multisite mirroring.
>>
>> This release is the culmination of a huge collective effort by about 100
>> different contributors. Thank you everyone who has helped to make this
>> possible!
>>
>> Upgrade Sequencing
>> ------------------
>>
>> * If your existing cluster is running a version older than v0.67
>> Dumpling, please first upgrade to the latest Dumpling release before
>> upgrading to v0.80 Firefly. Please refer to the :ref:`Dumpling upgrade`
>> documentation.
>>
>> * Upgrade daemons in the following order:
>>
>> 1. Monitors
>> 2. OSDs
>> 3. MDSs and/or radosgw
>>
>> If the ceph-mds daemon is restarted first, it will wait until all
>> OSDs have been upgraded before finishing its startup sequence. If
>> the ceph-mon daemons are not restarted prior to the ceph-osd
>> daemons, they will not correctly register their new capabilities
>> with the cluster and new features may not be usable until they are
>> restarted a second time.
>>
>> * Upgrade radosgw daemons together. There is a subtle change in behavior
>> for multipart uploads that prevents a multipart request that was initiated
>> with a new radosgw from being completed by an old radosgw.
>>
>> Notable changes since v0.79
>> ---------------------------
>>
>> * ceph-fuse, libcephfs: fix several caching bugs (Yan, Zheng)
>> * ceph-fuse: trim inodes in response to mds memory pressure (Yan, Zheng)
>> * librados: fix inconsistencies in API error values (David Zafman)
>> * librados: fix watch operations with cache pools (Sage Weil)
>> * librados: new snap rollback operation (David Zafman)
>> * mds: fix respawn (John Spray)
>> * mds: misc bugs (Yan, Zheng)
>> * mds: misc multi-mds fixes (Yan, Zheng)
>> * mds: use shared_ptr for requests (Greg Farnum)
>> * mon: fix peer feature checks (Sage Weil)
>> * mon: require 'x' mon caps for auth operations (Joao Luis)
>> * mon: shutdown when removed from mon cluster (Joao Luis)
>> * msgr: fix locking bug in authentication (Josh Durgin)
>> * osd: fix bug in journal replay/restart (Sage Weil)
>> * osd: many many many bug fixes with cache tiering (Samuel Just)
>> * osd: track omap and hit_set objects in pg stats (Samuel Just)
>> * osd: warn if agent cannot enable due to invalid (post-split) stats
>> (Sage Weil)
>> * rados bench: track metadata for multiple runs separately (Guang Yang)
>> * rgw: fixed subuser modify (Yehuda Sadeh)
>> * rpm: fix redhat-lsb dependency (Sage Weil, Alfredo Deza)
>>
>> For the complete release notes, please see:
>>
>> http://ceph.com/docs/master/release-notes/#v0-80-firefly
>>
>>
>> Getting Ceph
>> ------------
>>
>> * Git at git://github.com/ceph/ceph.git
>> * Tarball at http://ceph.com/download/ceph-0.80.tar.gz
>> * For packages, see http://ceph.com/docs/master/install/get-packages
>> * For ceph-deploy, see
>> http://ceph.com/docs/master/install/install-ceph-deploy
>>
>> _______________________________________________
>> ceph-users mailing list
>> ceph-users at lists.ceph.com
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>
> ----- End message from Sage Weil <sage at inktank.com> -----
>
> --
>
> Met vriendelijke groeten,
> Kenneth Waegeman
>
> _______________________________________________
> ceph-users mailing list
> ceph-users at lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


----- End message from Alexandre DERUMIER <aderumier at odiso.com> -----

-- 

Met vriendelijke groeten,
Kenneth Waegeman




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux