Re: How's cephfs going?

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Mon, 2017-07-17 at 02:59 +0000, 许雪寒 wrote:
> Hi, everyone.
> 
> We intend to use cephfs of Jewel version, however, we don’t know its status.
> Is it production ready in Jewel? Does it still have lots of bugs? Is it a
> major effort of the current ceph development? And who are using cephfs now?

Hello,

I've been using CephFS in production in a small-scale, conservative deployment
for the past year — to provide a small backing store, spanning two datacenters,
for a highly-available web service that supplies host configuration data and
APT package repositories for a fleet of ~1000 Linux workstations.

The motivation for using CephFS was not performance, but availability and
correctness.  It replaced a fairly complicated stack involving corosync,
pacemaker, XFS, DRBD, and mdadm.

Previous production systems using DRBD had proven unreliable in practice;
network glitches between datacenters would cause DRBD to enter a split-brain
state — which, for this application, was tolerable.

Bad, however, was when DRBD failed to keep the two halves of the mirrored
filesystem in sync — at one point, they had diverged to include over a gigabyte
of differences, which caused XFS to snap to read-only mode when it detected
internal inconsistencies.

CephFS, by contrast, has been remarkably solid, even in the face of networking
interruptions.  As well as remaining available, data integrity has been
problem-free.

(In addition to Ceph's own scrubbing capabilities, I've had automated
monitoring checks verifying the file-sizes and checksums of files in each APT
repository against those recorded in their indexes, with zero discrepancies in
a year of operation.)

Write performance wasn't terrific, but my deployment has been on constrained
hardware: running all of the Ceph daemons, plus the local kernel mount, plus
the nginx web server, all on the same hosts — with quadruple replication and
only a single gigabit link per machine — is not recommended for write
throughput.

The only issue I came across was a bug in the ceph.ko driver, which would
occasionally trip a null pointer exception in the kernel.  It was possible to
avoid this bug by mounting with the noasyncreaddir flag; the bug has long since
been fixed.

Kind regards,
David
-- 
David McBride <dwm37@xxxxxxxxx>
Unix Specialist, University Information Services

Attachment: signature.asc
Description: This is a digitally signed message part

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux