Re: How are you using Ceph?

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Matt,

On Mon, 17 Sep 2012, Matt W. Benjamin wrote:
> Hi
> 
> Just FYI, on the NFS integration front.  A pnfs files (RFC5661)-capable 
> NFSv4 re-exporter for Ceph has been committed to the Ganesha NFSv4 
> server development branch.  We're continuing to enhance and elaborate 
> this.  We have had on our (full) plates for a while to return Ceph 
> client library changes.  We've finished pullup and rebasing of these, 
> are doing some final testing of a couple things in preparation to push a 
> branch for review.

This is great news!  I'm interested to hear how the Ganesha bits map pNFS 
server instances to OSDs.. is it just matching IP addresses or something?

sage

> 
> Regards,
> 
> Matt
> 
> ----- "Sage Weil" <sage@xxxxxxxxxxx> wrote:
> 
> > On Mon, 17 Sep 2012, Tren Blackburn wrote:
> > > On Mon, Sep 17, 2012 at 5:05 PM, Smart Weblications GmbH - Florian
> > > Wiessner <f.wiessner@xxxxxxxxxxxxxxxxxxxxx> wrote:
> > > >
> > > > Hi,
> > > >
> > > > i use ceph to provide storage via rbd for our virtualization
> > cluster delivering
> > > > KVM based high availability Virtual Machines to my customers. I
> > also use it
> > > > as rbd device with ocfs2 on top of it for a 4 node webserver
> > cluster as shared
> > > > storage - i do this, because unfortunatelly cephfs is not ready
> > yet ;)
> > > >
> > > Hi Florian;
> > > 
> > > When you say "cephfs is not ready yet", what parts about it are not
> > > ready? There are vague rumblings about that in general, but I'd
> > love
> > > to see specific issues. I understand multiple *active* mds's are
> > not
> > > supported, but what other issues are you aware of?
> > 
> > Inktank is not yet supporting it because we do not have the QA in
> > place 
> > and general hardening that will make us feel comfortable recommending
> > it 
> > for customers.  That said, it works pretty well for most workloads. 
> > In 
> > particular, if you stay away from the snapshots and multi-mds, you
> > should 
> > be quite stable.
> > 
> > The engineering team here is about to do a bit of a pivot and refocus
> > on 
> > the file system now that the object store and RBD are in pretty good 
> > shape.  That will mean both core fs/mds stability and features as well
> > as 
> > integration efforts (NFS/CIFS/Hadoop).
> > 
> > 'Ready' is in the eye of the beholder.  There are a few people using
> > the 
> > fs successfully in production, but not too many.
> > 
> > sage
> > 
> > 
> >  > 
> > > And if there's a page documenting this already, I apologize...and
> > > would appreciate a link :)
> > > 
> > > t.
> > > --
> > > To unsubscribe from this list: send the line "unsubscribe
> > ceph-devel" in
> > > the body of a message to majordomo@xxxxxxxxxxxxxxx
> > > More majordomo info at  http://vger.kernel.org/majordomo-info.html
> > > 
> > > 
> > --
> > To unsubscribe from this list: send the line "unsubscribe ceph-devel"
> > in
> > the body of a message to majordomo@xxxxxxxxxxxxxxx
> > More majordomo info at  http://vger.kernel.org/majordomo-info.html
> 
> -- 
> Matt Benjamin
> The Linux Box
> 206 South Fifth Ave. Suite 150
> Ann Arbor, MI  48104
> 
> http://linuxbox.com
> 
> tel. 734-761-4689
> fax. 734-769-8938
> cel. 734-216-5309
> --
> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> the body of a message to majordomo@xxxxxxxxxxxxxxx
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> 
> 
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux