Explicit F2FS support (was: v0.80 Firefly released)

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Wed, 7 May 2014, Andrey Korolyov wrote:
> Hello,
> 
> first of all, congratulations to Inktank and thank you for your awesome work!
> 
> Although exploiting native f2fs abilities, as with btrfs, sounds
> awesome for a matter of performance, I wonder when kv db is able to
> practically give users with 'legacy' file systems ability to conduct
> CoW operations as fast as on the log-based fs, with small or no
> performance impact, what`s the primary idea behind introducing
> interface bounded to the specific filesystem in same time? Of course I
> believe that f2fs will outperform almost every competitor at its field
> - non-rotating media operations, but I would be grateful if someone
> can shed light on this development choice.

There are multiple directions to pursue, and they may be more or less 
suitable for different types of workloads.  f2fs is obviously targetted at 
SSD-backed nodes and more general purpose workloads, the main win being 
that we should be able to eliminate the ceph journal entirely.  The 
RocksDB support is a drop-in replacement for leveldb, which we already use 
for lots of OSD metadata.  There is also work in replacing the file-based 
strategy for storing objects entirely in a kv store (whether it is leveldb 
or rocksdb or something else like kinetic or NVMKV).  Some of this will be 
generally useful, and some will work better for specific types of 
workloads.

sage

> 
> On Wed, May 7, 2014 at 5:05 AM, Sage Weil <sage at inktank.com> wrote:
> > We did it!  Firefly v0.80 is built and pushed out to the ceph.com
> > repositories.
> >
> > This release will form the basis for our long-term supported release
> > Firefly, v0.80.x.  The big new features are support for erasure coding
> > and cache tiering, although a broad range of other features, fixes,
> > and improvements have been made across the code base.  Highlights include:
> >
> > * *Erasure coding*: support for a broad range of erasure codes for lower
> >   storage overhead and better data durability.
> > * *Cache tiering*: support for creating 'cache pools' that store hot,
> >   recently accessed objects with automatic demotion of colder data to
> >   a base tier.  Typically the cache pool is backed by faster storage
> >   devices like SSDs.
> > * *Primary affinity*: Ceph now has the ability to skew selection of
> >   OSDs as the "primary" copy, which allows the read workload to be
> >   cheaply skewed away from parts of the cluster without migrating any
> >   data.
> > * *Key/value OSD backend* (experimental): An alternative storage backend
> >   for Ceph OSD processes that puts all data in a key/value database like
> >   leveldb.  This provides better performance for workloads dominated by
> >   key/value operations (like radosgw bucket indices).
> > * *Standalone radosgw* (experimental): The radosgw process can now run
> >   in a standalone mode without an apache (or similar) web server or
> >   fastcgi.  This simplifies deployment and can improve performance.
> >
> > We expect to maintain a series of stable releases based on v0.80
> > Firefly for as much as a year.  In the meantime, development of Ceph
> > continues with the next release, Giant, which will feature work on the
> > CephFS distributed file system, more alternative storage backends
> > (like RocksDB and f2fs), RDMA support, support for pyramid erasure
> > codes, and additional functionality in the block device (RBD) like
> > copy-on-read and multisite mirroring.
> >
> > This release is the culmination of a huge collective effort by about 100
> > different contributors.  Thank you everyone who has helped to make this
> > possible!
> >
> > Upgrade Sequencing
> > ------------------
> >
> > * If your existing cluster is running a version older than v0.67
> >   Dumpling, please first upgrade to the latest Dumpling release before
> >   upgrading to v0.80 Firefly.  Please refer to the :ref:`Dumpling upgrade`
> >   documentation.
> >
> > * Upgrade daemons in the following order:
> >
> >     1. Monitors
> >     2. OSDs
> >     3. MDSs and/or radosgw
> >
> >   If the ceph-mds daemon is restarted first, it will wait until all
> >   OSDs have been upgraded before finishing its startup sequence.  If
> >   the ceph-mon daemons are not restarted prior to the ceph-osd
> >   daemons, they will not correctly register their new capabilities
> >   with the cluster and new features may not be usable until they are
> >   restarted a second time.
> >
> > * Upgrade radosgw daemons together.  There is a subtle change in behavior
> >   for multipart uploads that prevents a multipart request that was initiated
> >   with a new radosgw from being completed by an old radosgw.
> >
> > Notable changes since v0.79
> > ---------------------------
> >
> > * ceph-fuse, libcephfs: fix several caching bugs (Yan, Zheng)
> > * ceph-fuse: trim inodes in response to mds memory pressure (Yan, Zheng)
> > * librados: fix inconsistencies in API error values (David Zafman)
> > * librados: fix watch operations with cache pools (Sage Weil)
> > * librados: new snap rollback operation (David Zafman)
> > * mds: fix respawn (John Spray)
> > * mds: misc bugs (Yan, Zheng)
> > * mds: misc multi-mds fixes (Yan, Zheng)
> > * mds: use shared_ptr for requests (Greg Farnum)
> > * mon: fix peer feature checks (Sage Weil)
> > * mon: require 'x' mon caps for auth operations (Joao Luis)
> > * mon: shutdown when removed from mon cluster (Joao Luis)
> > * msgr: fix locking bug in authentication (Josh Durgin)
> > * osd: fix bug in journal replay/restart (Sage Weil)
> > * osd: many many many bug fixes with cache tiering (Samuel Just)
> > * osd: track omap and hit_set objects in pg stats (Samuel Just)
> > * osd: warn if agent cannot enable due to invalid (post-split) stats (Sage Weil)
> > * rados bench: track metadata for multiple runs separately (Guang Yang)
> > * rgw: fixed subuser modify (Yehuda Sadeh)
> > * rpm: fix redhat-lsb dependency (Sage Weil, Alfredo Deza)
> >
> > For the complete release notes, please see:
> >
> >    http://ceph.com/docs/master/release-notes/#v0-80-firefly
> >
> >
> > Getting Ceph
> > ------------
> >
> > * Git at git://github.com/ceph/ceph.git
> > * Tarball at http://ceph.com/download/ceph-0.80.tar.gz
> > * For packages, see http://ceph.com/docs/master/install/get-packages
> > * For ceph-deploy, see http://ceph.com/docs/master/install/install-ceph-deploy
> >
> 
> 


[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux