Re: mds stuck in standby, not one active

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Ceph fs dump:

e62
enable_multiple, ever_enabled_multiple: 1,1
default compat: compat={},rocompat={},incompat={1=base v0.20,2=client writeable ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no anchor table,9=file layout v2,10=snaprealm v2}legacy client fscid: 1

Filesystem 'ceph_fs' (1)
fs_name ceph_fs
epoch   62
flags   12
created 2022-11-28T12:05:17.203346+0000
modified        2022-12-15T12:09:14.091724+0000
tableserver     0
root    0
session_timeout 60
session_autoclose       300
max_file_size   1099511627776
required_client_features        {}
last_failure    0
last_failure_osd_epoch  196035
compat  compat={},rocompat={},incompat={1=base v0.20,2=client writeable ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no anchor table,9=file layout v2,10=snaprealm v2}
max_mds 1
in      0
up      {}
failed  0
damaged
stopped
data_pools      [4]
metadata_pool   5
inline_data     disabled
balancer
standby_count_wanted    1

f
Standby daemons:

[mds.ceph_fs.store5.gnlqqm{-1:152180029} state up:standby seq 1 join_fscid=1 addr [v2:192.168.50.135:6800/3548272808,v1:192.168.50.135:6801/3548272808] compat {c=[1],r=[1],i=[1]}] [mds.ceph_fs.store6.fxgvoj{-1:152416137} state up:standby seq 1 join_fscid=1 addr [v2:192.168.50.136:7024/1339959968,v1:192.168.50.136:7025/1339959968] compat {c=[1],r=[1],i=[1]}] [mds.ceph_fs.store4.mhvpot{-1:152477853} state up:standby seq 1 join_fscid=1 addr [v2:192.168.50.134:6800/3098669884,v1:192.168.50.134:6801/3098669884] compat {c=[1],r=[1],i=[1]}] [mds.ceph_fs.store3.vcnwzh{-1:152481783} state up:standby seq 1 join_fscid=1 addr [v2:192.168.50.133:6800/77378788,v1:192.168.50.133:6801/77378788] compat {c=[1],r=[1],i=[1]}]
dumped fsmap epoch 62

Ceph Status:

  cluster:
    id:     8c774934-1535-11ec-973e-525400130e4f
    health: HEALTH_ERR
            1 filesystem is degraded
            1 filesystem has a failed mds daemon
            1 filesystem is offline
            26 daemons have recently crashed

  services:
    mon: 2 daemons, quorum cephadm-vm,store2 (age 2d)
    mgr: store1.uevcpd(active, since 2d), standbys: cephadm-vm.zwagng
    mds: 0/1 daemons up (1 failed), 4 standby
    osd: 312 osds: 312 up (since 8h), 312 in (since 17h)

  data:
    volumes: 0/1 healthy, 1 failed
    pools:   7 pools, 289 pgs
    objects: 2.62M objects, 9.8 TiB
    usage:   29 TiB used, 1.9 PiB / 1.9 PiB avail
    pgs:     286 active+clean
             3   active+clean+scrubbing+deep

  io:
    client:   945 KiB/s rd, 3.3 MiB/s wr, 516 op/s rd, 562 op/s wr

Ceph Health detail:

HEALTH_ERR 1 filesystem is degraded; 1 filesystem has a failed mds daemon; 1 filesystem is offline; 26 daemons have recently crashed
[WRN] FS_DEGRADED: 1 filesystem is degraded
    fs ceph_fs is degraded
[WRN] FS_WITH_FAILED_MDS: 1 filesystem has a failed mds daemon
    fs ceph_fs has 1 failed mds
[ERR] MDS_ALL_DOWN: 1 filesystem is offline
    fs ceph_fs is offline because no MDS is active for it.
[WRN] RECENT_CRASH: 26 daemons have recently crashed
    osd.323 crashed on host store7 at 2022-12-12T14:03:23.857874Z
    osd.323 crashed on host store7 at 2022-12-12T14:03:43.945625Z
    osd.323 crashed on host store7 at 2022-12-12T14:04:03.282797Z
    osd.323 crashed on host store7 at 2022-12-12T14:04:22.612037Z
    osd.323 crashed on host store7 at 2022-12-12T14:04:41.630473Z
    osd.323 crashed on host store7 at 2022-12-12T14:34:49.237008Z
    osd.323 crashed on host store7 at 2022-12-12T14:35:09.903922Z
    osd.323 crashed on host store7 at 2022-12-12T14:35:28.621955Z
    osd.323 crashed on host store7 at 2022-12-12T14:35:46.985517Z
    osd.323 crashed on host store7 at 2022-12-12T14:36:05.375758Z
    osd.323 crashed on host store7 at 2022-12-12T15:01:57.235785Z
    osd.323 crashed on host store7 at 2022-12-12T15:02:16.581335Z
    osd.323 crashed on host store7 at 2022-12-12T15:02:33.212653Z
    osd.323 crashed on host store7 at 2022-12-12T15:02:49.775560Z
    osd.323 crashed on host store7 at 2022-12-12T15:03:06.303861Z
    mgr.cephadm-vm.zwagng crashed on host cephadm-vm at 2022-12-13T13:21:41.149773Z     mgr.cephadm-vm.zwagng crashed on host cephadm-vm at 2022-12-13T13:22:15.413105Z     mgr.cephadm-vm.zwagng crashed on host cephadm-vm at 2022-12-13T13:23:39.888401Z     mgr.cephadm-vm.zwagng crashed on host cephadm-vm at 2022-12-13T13:27:56.458529Z     mgr.cephadm-vm.zwagng crashed on host cephadm-vm at 2022-12-13T13:31:03.791532Z     mgr.cephadm-vm.zwagng crashed on host cephadm-vm at 2022-12-13T13:34:24.023106Z
    osd.98 crashed on host store3 at 2022-12-13T16:11:38.064735Z
    mgr.store1.uevcpd crashed on host store1 at 2022-12-13T18:39:33.091261Z
    osd.322 crashed on host store6 at 2022-12-14T06:06:14.193437Z
    osd.234 crashed on host store8 at 2022-12-15T02:32:13.009795Z
    osd.311 crashed on host store8 at 2022-12-15T02:32:18.407978Z

As suggested I was going to upgrade the ceph cluster to 16.2.7 to fix the mds issue, but it seems none of the running standby daemons is responding.

Am 15.12.2022 um 19:08 schrieb Patrick Donnelly:
On Thu, Dec 15, 2022 at 7:24 AM Mevludin Blazevic
<mblazevic@xxxxxxxxxxxxxx> wrote:
Hi,

while upgrading to ceph pacific 6.2.7, the upgrade process stuck exactly
at the mds daemons. Before, I have tried to increase/shrink the
placement size of them, but nothing happens. Currently I have 4/3
running daemons. One daemon should be stopped and removed.

Do you suggest to force remove these daemons or what could be the
preferred workaround?
Hard to say without more information. Please share:

ceph fs dump
ceph status
ceph health detail

--
Mevludin Blazevic, M.Sc.

University of Koblenz-Landau
Computing Centre (GHRKO)
Universitaetsstrasse 1
D-56070 Koblenz, Germany
Room A023
Tel: +49 261/287-1326

_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux