On Tue, Dec 13, 2022 at 2:21 PM Mevludin Blazevic <mblazevic@xxxxxxxxxxxxxx> wrote: > > Hi, > > thanks for the quick response! > > CEPH STATUS: > > cluster: > id: 8c774934-1535-11ec-973e-525400130e4f > health: HEALTH_ERR > 7 failed cephadm daemon(s) > There are daemons running an older version of ceph > 1 filesystem is degraded > 1 filesystem has a failed mds daemon > 1 filesystem is offline > 1 filesystem is online with fewer MDS than max_mds > 23 daemons have recently crashed > > services: > mon: 2 daemons, quorum cephadm-vm,store2 (age 12d) > mgr: store1.uevcpd(active, since 34m), standbys: cephadm-vm.zwagng > mds: 0/1 daemons up (1 failed), 4 standby > osd: 324 osds: 318 up (since 3h), 318 in (since 2h) > > data: > volumes: 0/1 healthy, 1 failed > pools: 6 pools, 257 pgs > objects: 2.61M objects, 9.8 TiB > usage: 29 TiB used, 2.0 PiB / 2.0 PiB avail > pgs: 257 active+clean > > io: > client: 0 B/s rd, 2.8 MiB/s wr, 435 op/s rd, 496 op/s wr > > FS DUMP: > > e60 > enable_multiple, ever_enabled_multiple: 1,1 > default compat: compat={},rocompat={},incompat={1=base v0.20,2=client > writeable ranges,3=default file layouts on dirs,4=dir inode in separate > object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no > anchor table,9=file layout v2,10=snaprealm v2} > legacy client fscid: 1 > > Filesystem 'ceph_fs' (1) > fs_name ceph_fs > epoch 58 > flags 32 > created 2022-11-28T12:05:17.203346+0000 > modified 2022-12-13T19:03:46.707236+0000 > tableserver 0 > root 0 > session_timeout 60 > session_autoclose 300 > max_file_size 1099511627776 > required_client_features {} > last_failure 0 > last_failure_osd_epoch 196035 > compat compat={},rocompat={},incompat={1=base v0.20,2=client writeable > ranges,3=default file layouts on dirs,4=dir inode in separate > object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no > anchor table,9=file layout v2,10=snaprealm v2} > max_mds 2 > in 0 > up {} > failed 0 > damaged > stopped > data_pools [4] > metadata_pool 5 > inline_data disabled > balancer > standby_count_wanted 1 > > > Standby daemons: > > [mds.ceph_fs.store5.gnlqqm{-1:152180029} state up:standby seq 1 > join_fscid=1 addr > [v2:192.168.50.135:6800/3548272808,v1:192.168.50.135:6801/3548272808] > compat {c=[1],r=[1],i=[1]}] > [mds.ceph_fs.store6.fxgvoj{ffffffff:915af89} state up:standby seq 1 > join_fscid=1 addr > [v2:192.168.50.136:1b70/4fde2aa0,v1:192.168.50.136:1b71/4fde2aa0] compat > {c=[1],r=[1],i=[1]}] > [mds.ceph_fs.store4.mhvpot{ffffffff:916a09d} state up:standby seq 1 > join_fscid=1 addr > [v2:192.168.50.134:1a90/b8b1f33c,v1:192.168.50.134:1a91/b8b1f33c] compat > {c=[1],r=[1],i=[1]}] > [mds.ceph_fs.store3.vcnwzh{ffffffff:916aff7} state up:standby seq 1 > join_fscid=1 addr > [v2:192.168.50.133:1a90/49cb4e4,v1:192.168.50.133:1a91/49cb4e4] compat > {c=[1],r=[1],i=[1]}] > dumped fsmap epoch 60 You're encountering a bug fixed in v16.2.7. Please upgrade to the latest version. -- Patrick Donnelly, Ph.D. He / Him / His Principal Software Engineer Red Hat, Inc. GPG: 19F28A586F808C2402351B93C3301A3E258DD79D _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx