Hello!
3 days have passed and I can't get cephFS work again.
I read a lot of the available documentation, posts [0] that mention
"magic" words, threads[1], blogs, etc...
And tried the suggested commands:
* ceph fs set cephfs max_mds 1
* ceph fs set cephfs allow_standby_replay false
* ceph fs compat cephfs add_incompat 7 "mds uses inline data"
* ceph fs set cephfs down false
* ceph fs set cephfs joinable true...
* Please, dear mds, become online, please....
But I still have:
# ceph mds stat
cephfs:1 1 up:standby
## ceph fs dump
e1953
enable_multiple, ever_enabled_multiple: 0,1
default compat: compat={},rocompat={},incompat={1=base v0.20,2=client
writeable ranges,3=default file layouts on dirs,4=dir inode in separate
object,5=mds uses versioned encoding,6=dirfrag is stored in omap,7=mds
uses inline data,8=no anchor table,9=file layout v2,10=snaprealm v2}
legacy client fscid: 1
Filesystem 'cephfs' (1)
fs_name cephfs
epoch 1953
flags 12
created 2021-12-29T14:01:39.824756+0000
modified 2021-12-30T13:37:30.470750+0000
tableserver 0
root 0
session_timeout 60
session_autoclose 300
max_file_size 1099511627776
required_client_features {}
last_failure 0
last_failure_osd_epoch 219687
compat compat={},rocompat={},incompat={1=base v0.20,2=client
writeable ranges,3=default file layouts on dirs,4=dir inode in separate
object,5=mds uses versioned encoding,6=dirfrag is stored in omap,7=mds
uses inline data,8=no anchor table,9=file layout v2,10=snaprealm v2}
max_mds 1
in 0
up {}
failed
damaged
stopped 1
data_pools [14]
metadata_pool 13
inline_data disabled
balancer
standby_count_wanted 0
Standby daemons:
[mds.cephfs.ceph-mds3.mfmaeh{-1:21505137} *state up:standby *seq 1
join_fscid=1 addr
[v2:192.168.15.207:6800/4170490328,v1:192.168.15.207:6801/4170490328]
compat {c=[1],r=[1],i=[77f]}]
dumped fsmap epoch 1953
Please,
When you have returned from celebrating the new year, do you have any
ideas that might help me?
Happy new year!
Javier.-
[0]
https://forum.proxmox.com/threads/ceph-16-2-6-cephfs-failed-after-upgrade-from-16-2-5.97742/
[1]
https://lists.ceph.io/hyperkitty/list/ceph-users@xxxxxxx/thread/KQ5A5OWRIUEOJBC7VILBGDIKPQGJQIWN/
El 28/12/21 a las 15:02, Tecnologia Charne.Net escribió:
Today, I upgraded from Pacific 16.2.6 to 16.2.7.
Since some items in dashboard weren't enabled
(Cluster->Hosts->Versions, for example) because I haven't cephadm
enabled, I activaded it and adopting every mon, mgr, osd on cluster,
following instructions in
https://docs.ceph.com/en/pacific/cephadm/adoption/
Everything was fine until point 10: Redeploy MDS daemons....
I have now:
# ceph health detail
HEALTH_ERR 1 filesystem is degraded; 1 filesystem has a failed mds
daemon; 1 filesystem is offline
[WRN] FS_DEGRADED: 1 filesystem is degraded
fs cephfs is degraded
[WRN] FS_WITH_FAILED_MDS: 1 filesystem has a failed mds daemon
fs cephfs has 2 failed mdss
[ERR] MDS_ALL_DOWN: 1 filesystem is offline
fs cephfs is offline because no MDS is active for it.
# ceph fs status
cephfs - 0 clients
======
RANK STATE MDS ACTIVITY DNS INOS DIRS CAPS
0 failed
1 failed
POOL TYPE USED AVAIL
cephfs_metadata metadata 1344M 20.8T
cephfs_data data 530G 8523G
STANDBY MDS
cephfs.mon1.qhueuv
cephfs.mon2.zrswzj
cephfs.mon3.cusflb
MDS version: ceph version 16.2.5-387-g7282d81d
(7282d81d2c500b5b0e929c07971b72444c6ac424) pacific (stable)
# ceph fs dump
e1777
enable_multiple, ever_enabled_multiple: 1,1
default compat: compat={},rocompat={},incompat={1=base v0.20,2=client
writeable ranges,3=default file layouts on dirs,4=dir inode in
separate object,5=mds uses versioned encoding,6=dirfrag is stored in
omap,7=mds uses inline data,8=no anchor table,9=file layout
v2,10=snaprealm v2}
legacy client fscid: 1
Filesystem 'cephfs' (1)
fs_name cephfs
epoch 1776
flags 12
created 2019-07-03T14:11:34.215467+0000
modified 2021-12-28T17:42:18.197012+0000
tableserver 0
root 0
session_timeout 60
session_autoclose 300
max_file_size 1099511627776
required_client_features {}
last_failure 0
last_failure_osd_epoch 218775
compat compat={},rocompat={},incompat={1=base v0.20,2=client
writeable ranges,3=default file layouts on dirs,4=dir inode in
separate object,5=mds uses versioned encoding,6=dirfrag is stored in
omap,7=mds uses inline data,8=no anchor table,9=file layout
v2,10=snaprealm v2}
max_mds 1
in 0,1
up {}
failed 0,1
damaged
stopped
data_pools [14]
metadata_pool 13
inline_data disabled
balancer
standby_count_wanted 1
Standby daemons:
[mds.cephfs.mon1.qhueuv{-1:21378633} state up:standby seq 1
join_fscid=1 addr
[v2:192.168.15.200:6800/3327091876,v1:192.168.15.200:6801/3327091876]
compat {c=[1],r=[1],i=[77f]}]
[mds.cephfs.mon2.zrswzj{-1:21384283} state up:standby seq 1
join_fscid=1 addr
[v2:192.168.15.203:6800/838079265,v1:192.168.15.203:6801/838079265]
compat {c=[1],r=[1],i=[77f]}]
[mds.cephfs.mon3.cusflb{-1:21393659} state up:standby seq 1
join_fscid=1 addr
[v2:192.168.15.205:6800/1887883707,v1:192.168.15.205:6801/1887883707]
compat {c=[1],r=[1],i=[77f]}]
dumped fsmap epoch 1777
Any clue will be most welcomed!
Thanks in advance.
Javier.-
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx