Hi Eugen,
Thanks for the tip! I just ran:
ceph orch daemon restart mgr.pr-md-01.jemmdf
(my specific mgr instance)
And it restarted my primary mgr daemon, and in the process failed over
to my standby mgr daemon on another server. That went smoothly.
Unfortunately, I still cannot get 'ceph fs status' to work (on any node)...
# ceph fs status
Error EINVAL: Traceback (most recent call last):
File "/usr/share/ceph/mgr/mgr_module.py", line 1811, in _handle_command
return CLICommand.COMMANDS[cmd['prefix']].call(self, cmd, inbuf)
File "/usr/share/ceph/mgr/mgr_module.py", line 474, in call
return self.func(mgr, **kwargs)
File "/usr/share/ceph/mgr/status/module.py", line 109, in
handle_fs_status
assert metadata
AssertionError
-erich
On 5/2/24 11:07 AM, Eugen Block wrote:
Yep, seen this a couple of times during upgrades. I’ll have to check my
notes if I wrote anything down for that. But try a mgr failover first,
that could help.
Zitat von Erich Weiler <weiler@xxxxxxxxxxxx>:
Hi All,
For a while now I've been using 'ceph fs status' to show current MDS
active servers, filesystem status, etc. I recently took down my MDS
servers and added RAM to them (one by one, so the filesystem stayed
online). After doing that with my four MDS servers (I had two active
and two standby), all looks OK, 'ceph -s' reports HEALTH_OK. But when
I do 'ceph fs status' now, I get this:
# ceph fs status
Error EINVAL: Traceback (most recent call last):
File "/usr/share/ceph/mgr/mgr_module.py", line 1811, in _handle_command
return CLICommand.COMMANDS[cmd['prefix']].call(self, cmd, inbuf)
File "/usr/share/ceph/mgr/mgr_module.py", line 474, in call
return self.func(mgr, **kwargs)
File "/usr/share/ceph/mgr/status/module.py", line 109, in
handle_fs_status
assert metadata
AssertionError
This is on ceph 18.2.1 reef. This is very odd - can anyone think of a
reason why 'ceph fs status' would stop working after taking each of
the servers down for maintenance?
The filesystem is online and working just fine however. This ceph
instance is deployed via the cephadm method on RHEL 9.3, so the
everything is containerized in podman.
Thanks again,
erich
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx