Re: Cluster inaccessible

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Thanks for your reply.

I tried something similar (but wrong) based on your messages in the
referenced threads, but I was missing the "gdb commands...end"
sequence so I just kept hitting my breakpoint over and over. However,
I was able to get the monitor running with your guidance.

Indeed, ceph rmfailed was run leading up to the crash.

Here is the output of ceph fs dump:

# ceph fs dump
dumped fsmap epoch 68022
e68022
enable_multiple, ever_enabled_multiple: 1,1
default compat: compat={},rocompat={},incompat={1=base v0.20,2=client
writeable ranges,3=default file layouts on dirs,4=dir inode in
separate object,5=mds uses versioned encoding,6=dirfrag is stored in
omap,8=no anchor table,9=file layout v2,10=snaprealm v2}
legacy client fscid: 1

Filesystem 'cephfs' (1)
fs_name cephfs
epoch 68022
flags 12
created 2021-08-29T00:22:55.564386+0000
modified 2021-10-10T13:45:03.446746+0000
tableserver 0
root 0
session_timeout 60
session_autoclose 300
max_file_size 1099511627776
required_client_features {}
last_failure 0
last_failure_osd_epoch 10491
compat compat={},rocompat={},incompat={1=base v0.20,2=client writeable
ranges,3=default file layouts on dirs,4=dir inode in separate
object,5=mds uses versioned encoding,6=dirfrag is stored in omap,7=mds
uses inline data,8=no anchor table,9=file layout v2,10=snaprealm v2}
max_mds 1
in 0
up {0=414534}
failed
damaged
stopped
data_pools [2]
metadata_pool 3
inline_data disabled
balancer
standby_count_wanted 1
[mds.ceph-metadata-04{0:414534} state up:active seq 51 laggy since
2021-10-10T13:45:03.427312+0000 addr
[v2:10.100.5.54:6800/1373631131,v1:10.100.5.54:6801/1373631131] compat
{c=[1],r=[1],i=[7ff]}]


Filesystem 'prod' (2)
fs_name prod
epoch 109b5
flags 12
created 2021-08-29T00:30:46.944134+0000
modified 2021-10-10T03:16:41.859392+0000
tableserver 0
root 0
session_timeout 60
session_autoclose 300
max_file_size 1099511627776
required_client_features {}
last_failure 0
last_failure_osd_epoch 9953
compat compat={},rocompat={},incompat={1=base v0.20,2=client writeable
ranges,3=default file layouts on dirs,4=dir inode in separate
object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no
anchor table,9=file layout v2,10=snaprealm v2}
max_mds 1
in 0,1,2
up {}
failed
damaged
stopped 3
data_pools [5]
metadata_pool 4
inline_data disabled
balancer
standby_count_wanted 1


Filesystem 'beta' (3)
fs_name beta
epoch 68022
flags 12
created 2021-08-29T00:31:49.339070+0000
modified 2021-10-10T13:45:03.446747+0000
tableserver 0
root 0
session_timeout 60
session_autoclose 300
max_file_size 1099511627776
required_client_features {}
last_failure 0
last_failure_osd_epoch 10492
compat compat={},rocompat={},incompat={1=base v0.20,2=client writeable
ranges,3=default file layouts on dirs,4=dir inode in separate
object,5=mds uses versioned encoding,6=dirfrag is stored in omap,7=mds
uses inline data,8=no anchor table,9=file layout v2,10=snaprealm v2}
max_mds 1
in 0
up {0=414465}
failed
damaged
stopped
data_pools [8]
metadata_pool 7
inline_data disabled
balancer
standby_count_wanted 1
[mds.ceph-metadata-01{0:414465} state up:active seq 68 laggy since
2021-10-10T13:45:03.427278+0000 addr
[v2:10.100.5.51:6800/391530905,v1:10.100.5.51:6801/391530905] compat
{c=[1],r=[1],i=[7ff]}]


Standby daemons:

[mds.ceph-metadata-05{ffffffff:6534c} state up:standby seq 1 laggy
since 07e5-0a-0aT0d:2d:03.068547+0000 addr
[v2:10.100.5.55:1a90/ea781ada,v1:10.100.5.55:1a91/ea781ada] compat
{c=[1],r=[1],i=[7ff]}]
[mds.ceph-metadata-03{ffffffff:653e4} state up:standby seq 1 laggy
since 07e5-0a-0aT0d:2d:03.06855c+0000 addr
[v2:10.100.5.53:1a90/64326853,v1:10.100.5.53:1a91/64326853] compat
{c=[1],r=[1],i=[7ff]}]
[mds.ceph-metadata-02{ffffffff:653e7} state up:standby seq 1 laggy
since 07e5-0a-0aT0d:2d:03.068570+0000 addr
[v2:10.100.5.52:1a90/63094e0,v1:10.100.5.52:1a91/63094e0] compat
{c=[1],r=[1],i=[7ff]}]
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux