Fwd: Ceph Filesystem - Production?

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Fri, Aug 29, 2014 at 8:36 AM, James Devine <fxmulder at gmail.com> wrote:
>
> On Thu, Aug 28, 2014 at 1:30 PM, Gregory Farnum <greg at inktank.com> wrote:
>>
>> On Thu, Aug 28, 2014 at 10:36 AM, Brian C. Huffman
>> <bhuffman at etinternational.com> wrote:
>> > Is Ceph Filesystem ready for production servers?
>> >
>> > The documentation says it's not, but I don't see that mentioned anywhere
>> > else.
>> > http://ceph.com/docs/master/cephfs/
>>
>> Everybody has their own standards, but Red Hat isn't supporting it for
>> general production use at this time. If you're brave you could test it
>> under your workload for a while and see how it comes out; the known
>> issues are very much workload-dependent (or just general concerns over
>> polish).
>> -Greg
>> Software Engineer #42 @ http://inktank.com | http://ceph.com
>> _______________________________________________
>> ceph-users mailing list
>> ceph-users at lists.ceph.com
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>
>
> I've been testing it with our webstats since it gets live hits but isn't
> customer affecting.  Seems the MDS server has problems every few days
> requiring me to umount and remount the ceph disk to resolve.  Not sure if
> the issue is resolved in development versions but as of 0.80.5 we seem to be
> hitting it.  I set the log verbosity to 20 so there's tons of logs but ends
> with

The cephfs client is supposed to be able to handle MDS takeover.
what's symptom makes you umount and remount the cephfs ?

>
> 2014-08-24 07:10:19.682015 7f2b575e7700 10 mds.0.14  laggy, deferring
> client_request(client.92141:6795587 getattr pAsLsXsFs #10000026dc1)
> 2014-08-24 07:10:19.682021 7f2b575e7700  5 mds.0.14 is_laggy 19.324963 > 15
> since last acked beacon
> 2014-08-24 07:10:20.358011 7f2b554e2700 10 mds.0.14 beacon_send up:active
> seq 127220 (currently up:active)
> 2014-08-24 07:10:21.515899 7f2b575e7700  5 mds.0.14 is_laggy 21.158841 > 15
> since last acked beacon
> 2014-08-24 07:10:21.515912 7f2b575e7700 10 mds.0.14  laggy, deferring
> client_session(request_renewcaps seq 26766)
> 2014-08-24 07:10:21.515915 7f2b575e7700  5 mds.0.14 is_laggy 21.158857 > 15
> since last acked beacon
> 2014-08-24 07:10:21.981148 7f2b575e7700 10 mds.0.snap check_osd_map
> need_to_purge={}
> 2014-08-24 07:10:21.981176 7f2b575e7700  5 mds.0.14 is_laggy 21.624117 > 15
> since last acked beacon
> 2014-08-24 07:10:23.170528 7f2b575e7700  5 mds.0.14 handle_mds_map epoch 93
> from mon.0
> 2014-08-24 07:10:23.175367 7f2b532d5700  0 -- 10.251.188.124:6800/985 >>
> 10.251.188.118:0/2461578479 pipe(0x5588a80 sd=23 :6800 s=2 pgs=91 cs=1 l=0
> c=0x2cbfb20).fault with nothing to send, going to standby
> 2014-08-24 07:10:23.175376 7f2b533d6700  0 -- 10.251.188.124:6800/985 >>
> 10.251.188.55:0/306923677 pipe(0x5588d00 sd=22 :6800 s=2 pgs=7 cs=1 l=0
> c=0x2cbf700).fault with nothing to send, going to standby
> 2014-08-24 07:10:23.175380 7f2b531d4700  0 -- 10.251.188.124:6800/985 >>
> 10.251.188.31:0/2854230502 pipe(0x5589480 sd=24 :6800 s=2 pgs=881 cs=1 l=0
> c=0x2cbfde0).fault with nothing to send, going to standby
> 2014-08-24 07:10:23.175438 7f2b534d7700  0 -- 10.251.188.124:6800/985 >>
> 10.251.188.68:0/2928927296 pipe(0x5588800 sd=21 :6800 s=2 pgs=7 cs=1 l=0
> c=0x2cbf5a0).fault with nothing to send, going to standby
> 2014-08-24 07:10:23.184201 7f2b575e7700 10 mds.0.14      my compat
> compat={},rocompat={},incompat={1=base v0.20,2=client writeable
> ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds
> uses versioned encoding,6=dirfrag is stored in omap,7=mds uses inline data}
> 2014-08-24 07:10:23.184255 7f2b575e7700 10 mds.0.14  mdsmap compat
> compat={},rocompat={},incompat={1=base v0.20,2=client writeable
> ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds
> uses versioned encoding,6=dirfrag is stored in omap}
> 2014-08-24 07:10:23.184264 7f2b575e7700 10 mds.-1.-1 map says i am
> 10.251.188.124:6800/985 mds.-1.-1 state down:dne
> 2014-08-24 07:10:23.184275 7f2b575e7700 10 mds.-1.-1  peer mds gid 94665
> removed from map
> 2014-08-24 07:10:23.184282 7f2b575e7700  1 mds.-1.-1 handle_mds_map i
> (10.251.188.124:6800/985) dne in the mdsmap, respawning myself
> 2014-08-24 07:10:23.184284 7f2b575e7700  1 mds.-1.-1 respawn
> 2014-08-24 07:10:23.184286 7f2b575e7700  1 mds.-1.-1  e: '/usr/bin/ceph-mds'
> 2014-08-24 07:10:23.184288 7f2b575e7700  1 mds.-1.-1  0: '/usr/bin/ceph-mds'
> 2014-08-24 07:10:23.184289 7f2b575e7700  1 mds.-1.-1  1: '-i'
> 2014-08-24 07:10:23.184290 7f2b575e7700  1 mds.-1.-1  2:
> 'ceph-cluster1-mds2'
> 2014-08-24 07:10:23.184291 7f2b575e7700  1 mds.-1.-1  3: '--pid-file'
> 2014-08-24 07:10:23.184292 7f2b575e7700  1 mds.-1.-1  4:
> '/var/run/ceph/mds.ceph-cluster1-mds2.pid'
> 2014-08-24 07:10:23.184293 7f2b575e7700  1 mds.-1.-1  5: '-c'
> 2014-08-24 07:10:23.184294 7f2b575e7700  1 mds.-1.-1  6:
> '/etc/ceph/ceph.conf'
> 2014-08-24 07:10:23.184295 7f2b575e7700  1 mds.-1.-1  7: '--cluster'
> 2014-08-24 07:10:23.184296 7f2b575e7700  1 mds.-1.-1  8: 'ceph'
> 2014-08-24 07:10:23.274640 7f2b575e7700  1 mds.-1.-1  exe_path
> /usr/bin/ceph-mds
> 2014-08-24 07:10:23.606875 7f4c55abb800  0 ceph version 0.80.5
> (38b73c67d375a2552d8ed67843c8a65c2c0feba6), process ceph-mds, pid 987
> 2014-08-24 07:10:49.024862 7f4c506ad700  1 mds.-1.0 handle_mds_map standby
> 2014-08-24 07:10:49.199676 7f4c506ad700  0 mds.-1.0 handle_mds_beacon no
> longer laggy
> 2014-08-24 07:10:50.215240 7f4c506ad700  1 mds.-1.0 handle_mds_map standby
> 2014-08-24 07:10:51.290407 7f4c506ad700  1 mds.-1.0 handle_mds_map standby
>
>

Did you use active/standby MDS setup? Did the  MDS use lots of memory
before it crashed?

Regards
Yan, Zheng

>
> _______________________________________________
> ceph-users mailing list
> ceph-users at lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>


[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux