=?eucgb2312_cn?q?=BB=D8=B8=B4=3A_Re=3A_Cephfs_-_MDS_all_up=3Astandby=2C_not_becoming_up=3Aactive?=

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



We are experiencing the same when upgrading to 16.2.6 with cephadm.

I tried

ceph fs set cephfs max_mds 1
ceph fs set cephfs allow_standby_replay false

, but still all MDS goes to standby. It seems all ranks are marked failed. Do we have a way to clear this flag?

Please help. Our cluster is down. Thanks.

# ceph fs status
cephfs - 0 clients
======
RANK  STATE   MDS  ACTIVITY  DNS  INOS  DIRS  CAPS
0    failed
1    failed
          POOL              TYPE     USED  AVAIL
   cephfs.cephfs.meta     metadata   114G   404G
   cephfs.cephfs.data       data    84.9T  17.6T
cephfs.cephfs.data_ssd     data       0    606G
cephfs.cephfs.data_mixed    data    9879G   404G
    STANDBY MDS
cephfs.gpu023.aetiph
cephfs.gpu018.ovxvoz
cephfs.gpu006.ddpekw
cephfs.gpu024.rpfbnh
MDS version: ceph version 16.2.6 (ee28fb57e47e9f88813e24bbf4c14496ca299d31) pacific (stable)

# ceph fs dump
e41422
enable_multiple, ever_enabled_multiple: 0,1
default compat: compat={},rocompat={},incompat={1=base v0.20,2=client writeable ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no anchor table,9=file layout v2,10=snaprealm v2}
legacy client fscid: 2

Filesystem 'cephfs' (2)
fs_name cephfs
epoch   41422
flags   12
created 2020-09-15T04:10:53.585782+0000
modified        2021-09-17T15:05:26.239956+0000
tableserver     0
root    0
session_timeout 60
session_autoclose       300
max_file_size   1099511627776
required_client_features        {}
last_failure    0
last_failure_osd_epoch  43315
compat  compat={},rocompat={},incompat={1=base v0.20,2=client writeable ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no anchor table,9=file layout v2,10=snaprealm v2}
max_mds 1
in      0,1
up      {}
failed  0,1
damaged
stopped
data_pools      [5,13,16]
metadata_pool   4
inline_data     disabled
balancer
standby_count_wanted    1


Standby daemons:

[mds.cephfs.gpu023.aetiph{-1:7908668} state up:standby seq 1 join_fscid=2 addr [v2:202.38.247.186:6800/3495351337,v1:202.38.247.186:6801/3495351337] compat {c=[1],r=[1],i=[7ff]}]
[mds.cephfs.gpu018.ovxvoz{ffffffff:78cdb8} state up:standby seq 1 join_fscid=2 addr [v2:202.38.247.181:1a90/94680caa,v1:202.38.247.181:1a91/94680caa] compat {c=[1],r=[1],i=[7ff]}]
[mds.cephfs.gpu006.ddpekw{ffffffff:78f84f} state up:standby seq 1 join_fscid=2 addr [v2:202.38.247.175:1a90/fdd0fd1a,v1:202.38.247.175:1a91/fdd0fd1a] compat {c=[1],r=[1],i=[7ff]}]
[mds.cephfs.gpu024.rpfbnh{ffffffff:78fc4e} state up:standby seq 1 join_fscid=2 addr [v2:202.38.247.187:1a90/4e2e69dc,v1:202.38.247.187:1a91/4e2e69dc] compat {c=[1],r=[1],i=[7ff]}]
dumped fsmap epoch 41422

发件人: Patrick Donnelly<mailto:pdonnell@xxxxxxxxxx>
发送时间: 2021年9月17日 22:42
收件人: Eric Dold<mailto:dold.eric@xxxxxxxxx>
抄送: ceph-users<mailto:ceph-users@xxxxxxx>
主题:  Re: Cephfs - MDS all up:standby, not becoming up:active

On Fri, Sep 17, 2021 at 8:54 AM Eric Dold <dold.eric@xxxxxxxxx> wrote:
>
> Hi,
>
> I get the same after upgrading to 16.2.6. All mds daemons are standby.
>
> After setting
> ceph fs set cephfs max_mds 1
> ceph fs set cephfs allow_standby_replay false
> the mds still wants to be standby.
>
> 2021-09-17T14:40:59.371+0200 7f810a58f600  0 ceph version 16.2.6
> (ee28fb57e47e9f88813e24bbf4c14496ca299d31) pacific (stable), process
> ceph-mds, pid 7113
> 2021-09-17T14:40:59.371+0200 7f810a58f600  1 main not setting numa affinity
> 2021-09-17T14:40:59.371+0200 7f810a58f600  0 pidfile_write: ignore empty
> --pid-file
> 2021-09-17T14:40:59.375+0200 7f8105cf1700  1 mds.ceph3 Updating MDS map to
> version 226251 from mon.0
> 2021-09-17T14:41:00.455+0200 7f8105cf1700  1 mds.ceph3 Updating MDS map to
> version 226252 from mon.0
> 2021-09-17T14:41:00.455+0200 7f8105cf1700  1 mds.ceph3 Monitors have
> assigned me to become a standby.
>
> setting add_incompat 1 does also not work:
> # ceph fs compat cephfs add_incompat 1
> Error EINVAL: adding a feature requires a feature string
>
> Any ideas?

Please share `ceph fs dump`.


--
Patrick Donnelly, Ph.D.
He / Him / His
Principal Software Engineer
Red Hat Sunnyvale, CA
GPG: 19F28A586F808C2402351B93C3301A3E258DD79D

_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx

_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux