"mds daemon damaged" after restarting MDS - Filesystem DOWN

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Not sure if the previous message went through since I was not subscribed. If yes sorry for the spam.

Dear all

Running nautilus 14.2.7. The data in the FS are important and cannot be lost.

Today I increased the PGS of the volume pool from 8k to 16k. The active mds started reporting slow ops. (the filesystem is not in the volume pool). After few hours the FS was very slow, I reduced the backfill to 1 and since the situation was not improving, I restarted the MDS (no other standby MDSs. it was a single mds).

After that the crash. The mds does not goes back up with this error:

020-02-07 07:03:32.477 7fbf69647700 -1 NetHandler create_socket couldn't create socket (97) Address family not supported by protocol 2020-02-07 07:03:32.541 7fbf65e6a700  1 mds.ceph-mon-01 Updating MDS map to version 48461 from mon.2 2020-02-07 07:03:37.613 7fbf65e6a700  1 mds.ceph-mon-01 Updating MDS map to version 48462 from mon.2 2020-02-07 07:03:37.613 7fbf65e6a700  1 mds.ceph-mon-01 Map has assigned me to become a standby 2020-02-07 07:14:11.789 7fbf66e42700 -1 received  signal: Terminated from /sbin/init  (PID: 1) UID: 0 2020-02-07 07:14:11.789 7fbf66e42700 -1 mds.ceph-mon-01 *** got signal Terminated *** 2020-02-07 07:14:11.789 7fbf66e42700  1 mds.ceph-mon-01 suicide! Wanted state up:standby 2020-02-07 07:14:12.565 7fbf65e6a700  0 ms_deliver_dispatch: unhandled message 0x563fcb438d00 mdsmap(e 48465) v1 from mon.2 v1:10.3.78.32:6789/0 2020-02-07 07:25:16.782 7f26c39de2c0  0 set uid:gid to 64045:64045 (ceph:ceph) 2020-02-07 07:25:16.782 7f26c39de2c0  0 ceph version 14.2.7 (3d58626ebeec02d8385a4cefb92c6cbc3a45bfe8) nautilus (stable), process ceph-mds, pid 3724 2020-02-07 07:25:16.782 7f26c39de2c0  0 pidfile_write: ignore empty --pid-file 2020-02-07 07:25:16.786 7f26b5326700 -1 NetHandler create_socket couldn't create socket (97) Address family not supported by protocol 2020-02-07 07:25:16.790 7f26b1b49700  1 mds.ceph-mon-01 Updating MDS map to version 48472 from mon.0 2020-02-07 07:25:17.691 7f26b1b49700  1 mds.ceph-mon-01 Updating MDS map to version 48473 from mon.0 2020-02-07 07:25:17.691 7f26b1b49700  1 mds.ceph-mon-01 Map has assigned me to become a standby 2020-02-07 07:29:50.306 7f26b2b21700 -1 received  signal: Terminated from /sbin/init  (PID: 1) UID: 0 2020-02-07 07:29:50.306 7f26b2b21700 -1 mds.ceph-mon-01 *** got signal Terminated *** 2020-02-07 07:29:50.306 7f26b2b21700  1 mds.ceph-mon-01 suicide! Wanted state up:standby 2020-02-07 07:29:50.526 7f26b5b27700  1 mds.beacon.ceph-mon-01 discarding unexpected beacon reply down:dne seq 70 dne 2020-02-07 07:29:52.802 7f26b1b49700  0 ms_deliver_dispatch: unhandled message 0x55ef110ab200 mdsmap(e 48474) v1 from mon.0 v1:10.3.78.22:6789/0

Rebooting did not help

I asked #CEPH OFTC and they suggested to bring up another "fresh" mds. I did that, and they do not start, going to standby. LOGS:

2020-02-07 07:12:46.696 7fe4b388b2c0  0 set uid:gid to 64045:64045 (ceph:ceph) 2020-02-07 07:12:46.696 7fe4b388b2c0  0 ceph version 14.2.7 (3d58626ebeec02d8385a4cefb92c6cbc3a45bfe8) nautilus (stable), process ceph-mds, pid 74742 2020-02-07 07:12:46.696 7fe4b388b2c0  0 pidfile_write: ignore empty --pid-file 2020-02-07 07:12:46.704 7fe4a19f6700  1 mds.ceph-mon-02 Updating MDS map to version 48462 from mon.0 2020-02-07 07:12:47.456 7fe4a19f6700  1 mds.ceph-mon-02 Updating MDS map to version 48463 from mon.0 2020-02-07 07:12:47.456 7fe4a19f6700  1 mds.ceph-mon-02 Map has assigned me to become a standby 2020-02-07 07:14:16.615 7fe4a29ce700 -1 received  signal: Terminated from /sbin/init  (PID: 1) UID: 0 2020-02-07 07:14:16.615 7fe4a29ce700 -1 mds.ceph-mon-02 *** got signal Terminated *** 2020-02-07 07:14:16.615 7fe4a29ce700  1 mds.ceph-mon-02 suicide! Wanted state up:standby 2020-02-07 07:14:16.947 7fe4a51d3700  1 mds.beacon.ceph-mon-02 discarding unexpected beacon reply down:dne seq 24 dne 2020-02-07 07:14:18.715 7fe4a19f6700  0 ms_deliver_dispatch: unhandled message 0x5602fbc6df80 mdsmap(e 48466) v1 from mon.0 v2:10.3.78.22:3300/0 2020-02-07 07:25:02.093 7f3c2f92a2c0  0 set uid:gid to 64045:64045 (ceph:ceph) 2020-02-07 07:25:02.093 7f3c2f92a2c0  0 ceph version 14.2.7 (3d58626ebeec02d8385a4cefb92c6cbc3a45bfe8) nautilus (stable), process ceph-mds, pid 75471 2020-02-07 07:25:02.093 7f3c2f92a2c0  0 pidfile_write: ignore empty --pid-file 2020-02-07 07:25:02.097 7f3c1da95700  1 mds.ceph-mon-02 Updating MDS map to version 48471 from mon.2 2020-02-07 07:25:06.413 7f3c1da95700  1 mds.ceph-mon-02 Updating MDS map to version 48472 from mon.2 2020-02-07 07:25:06.413 7f3c1da95700  1 mds.ceph-mon-02 Map has assigned me to become a standby 2020-02-07 07:29:56.869 7f3c1ea6d700 -1 received  signal: Terminated from /sbin/init  (PID: 1) UID: 0 2020-02-07 07:29:56.869 7f3c1ea6d700 -1 mds.ceph-mon-02 *** got signal Terminated *** 2020-02-07 07:29:56.869 7f3c1ea6d700  1 mds.ceph-mon-02 suicide! Wanted state up:standby 2020-02-07 07:29:58.113 7f3c1da95700  0 ms_deliver_dispatch: unhandled message 0x563c5df33f80 mdsmap(e 48475) v1 from mon.2 v2:10.3.78.32:3300/0

Here ceph status

  cluster:
    id:     a8dde71d-ca7b-4cf5-bd38-8989c6a27011
    health: HEALTH_ERR
            1 filesystem is degraded
            1 filesystem is offline
            1 mds daemon damaged
            2 daemons have recently crashed

  services:
    mon: 3 daemons, quorum ceph-mon-01,ceph-mon-02,ceph-mon-03 (age 41m)
    mgr: ceph-mon-02(active, since 41m), standbys: ceph-mon-03, ceph-mon-01
    mds: pawsey-sync-fs:0/1, 1 damaged
    osd: 925 osds: 715 up (since 2h), 715 in (since 23h)
    rgw: 3 daemons active (radosgw-01, radosgw-02, radosgw-03)

  data:
    pools:   24 pools, 26569 pgs
    objects: 52.64M objects, 199 TiB
    usage:   685 TiB used, 6.7 PiB / 7.3 PiB avail
    pgs:     26513 active+clean
             54    active+clean+scrubbing+deep
             2     active+clean+scrubbing

Ceph osd ls detail: https://pastebin.com/raw/bxi4HSa5

the metadata pool is on NVMe

Can anyone give me some help?

Any command run like journal repairs do not work as they expect the MDs to be up.

Thanks

Cheers


--
Luca Cervigni
Infrastructure Architect

Tel. +61864368802
Pawsey Supercomputing Centre
1 Bryce Ave, Kensington WA 6151
Australia
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux