Crash and strange things on MDS

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hey everyone,

It's my first post here to expose a potential issue I found today using
Ceph 0.56.1.

The cluster configuration is, briefly: 27 osd of ~900GB and 3 MON/MDS.
All nodes are running Exherbo (source-based distribution) with Ceph
0.56.1 and Linux 3.7.0. We are only using CephFS on this cluster which
is mounted on ~60 clients (increasing each day). Objects are replicated
three times and the cluster handles only 7GB of data atm for 350k
objects.

In certain conditions (I don't know them atm), some clients hang,
generate CPU overloads (kworker) and are unable to make any IO on
Ceph. The active MDS have ~20Mbps in/out during the issue (less than
2Mbps in normal activity). I don't know if it's directly linked but we
also observe a lot of missing files at the same time.

The problem is similar to this one [1].

A restart of the client or the MDS was enough before today, but we found
a new behavior: the active MDS consumes a lot of CPU during 3 to 5 hours
with ~25% clients hanging.

In logs I found a segfault with this backtrace [2] and 100,000 dumped
events during the first hang. We observed another hang which produces
lot of these events (in debug mode):
   - "mds.0.server FAIL on ESTALE but attempting recovery"
   - "mds.0.server reply_request -116 (Stale NFS file handle)
      client_request(client.10991:1031 getattr As #1000004bab0
      RETRY=132)"

We have no profiling tools available on these nodes, and I don't know
what I should search in the 35 GB log file.

Note: the segmentation fault occured only once but the problem was
observed four times on this cluster.

Any help may be appreciated.

References:
[1] http://www.spinics.net/lists/ceph-devel/msg04903.html
[2] ceph version 0.56.1 (e4a541624df62ef353e754391cbbb707f54b16f7)
    1: /usr/bin/ceph-mds() [0x817e82]
    2: (()+0xf140) [0x7f9091d30140]
    3: (MDCache::request_drop_foreign_locks(MDRequest*)+0x21) [0x5b9dc1]
    4: (MDCache::request_drop_locks(MDRequest*)+0x19) [0x5baae9]
    5: (MDCache::request_cleanup(MDRequest*)+0x60) [0x5bab70]
    6: (MDCache::request_kill(MDRequest*)+0x80) [0x5bae90]
    7: (Server::journal_close_session(Session*, int)+0x372) [0x549aa2]
    8: (Server::kill_session(Session*)+0x137) [0x549c67]
    9: (Server::find_idle_sessions()+0x12a6) [0x54b0d6]
    10: (MDS::tick()+0x338) [0x4da928]
    11: (SafeTimer::timer_thread()+0x1af) [0x78151f]
    12: (SafeTimerThread::entry()+0xd) [0x782bad]
    13: (()+0x7ddf) [0x7f9091d28ddf]
    14: (clone()+0x6d) [0x7f90909cc24d]

Cheers,
-- 
Kevin Decherf - @Kdecherf
GPG C610 FE73 E706 F968 612B E4B2 108A BD75 A81E 6E2F
http://kdecherf.com
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux