Re: 'ceph fs status' no longer works?

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Excellent!  Restarting all the MDS daemons fixed it.  Thank you.

This kinda feels like a bug.

-erich

On 5/2/24 12:44 PM, Bandelow, Gunnar wrote:
Hi Erich,

im not sure about this specific error message, but "ceph fs status" sometimes did fail me end of last year/in the beginning of the year.

Restarting ALL mon, mgr AND mds fixed it at the time.

Best regards,
Gunnar


=======================================================

Gunnar Bandelow (dipl. phys.)

Universitätsrechenzentrum (URZ)
Universität Greifswald
Felix-Hausdorff-Straße 18
17489 Greifswald
Germany


--- Original Nachricht ---
*Betreff: * Re: 'ceph fs status' no longer works?
*Von: *"Erich Weiler" <weiler@xxxxxxxxxxxx <mailto:weiler@xxxxxxxxxxxx>>
*An: *"Eugen Block" <eblock@xxxxxx <mailto:eblock@xxxxxx>>, ceph-users@xxxxxxx <mailto:ceph-users@xxxxxxx>
*Datum: *02-05-2024 21:05



    Hi Eugen,

    Thanks for the tip!  I just ran:

    ceph orch daemon restart mgr.pr-md-01.jemmdf

    (my specific mgr instance)

    And it restarted my primary mgr daemon, and in the process failed over
    to my standby mgr daemon on another server.  That went smoothly.

    Unfortunately, I still cannot get 'ceph fs status' to work (on any
    node)...

    # ceph fs status
    Error EINVAL: Traceback (most recent call last):
        File "/usr/share/ceph/mgr/mgr_module.py", line 1811, in
    _handle_command
          return CLICommand.COMMANDS[cmd['prefix']].call(self, cmd, inbuf)
        File "/usr/share/ceph/mgr/mgr_module.py", line 474, in call
          return self.func(mgr, **kwargs)
        File "/usr/share/ceph/mgr/status/module.py", line 109, in
    handle_fs_status
          assert metadata
    AssertionError

    -erich

    On 5/2/24 11:07 AM, Eugen Block wrote:
     > Yep, seen this a couple of times during upgrades. I’ll have to
    check my
     > notes if I wrote anything down for that. But try a mgr failover
    first,
     > that could help.
     >
     > Zitat von Erich Weiler <weiler@xxxxxxxxxxxx
    <mailto:weiler@xxxxxxxxxxxx>>:
     >
     >> Hi All,
     >>
     >> For a while now I've been using 'ceph fs status' to show current
    MDS
     >> active servers, filesystem status, etc.  I recently took down my
    MDS
     >> servers and added RAM to them (one by one, so the filesystem stayed
     >> online).  After doing that with my four MDS servers (I had two
    active
>> and two standby), all looks OK, 'ceph -s' reports HEALTH_OK. But when
     >> I do 'ceph fs status' now, I get this:
     >>
     >> # ceph fs status
     >> Error EINVAL: Traceback (most recent call last):
     >>   File "/usr/share/ceph/mgr/mgr_module.py", line 1811, in
    _handle_command
     >>     return CLICommand.COMMANDS[cmd['prefix']].call(self, cmd, inbuf)
     >>   File "/usr/share/ceph/mgr/mgr_module.py", line 474, in call
     >>     return self.func(mgr, **kwargs)
     >>   File "/usr/share/ceph/mgr/status/module.py", line 109, in
     >> handle_fs_status
     >>     assert metadata
     >> AssertionError
     >>
     >> This is on ceph 18.2.1 reef.  This is very odd - can anyone
    think of a
     >> reason why 'ceph fs status' would stop working after taking each of
     >> the servers down for maintenance?
     >>
     >> The filesystem is online and working just fine however.  This ceph
     >> instance is deployed via the cephadm method on RHEL 9.3, so the
     >> everything is containerized in podman.
     >>
     >> Thanks again,
     >> erich
     >> _______________________________________________
     >> ceph-users mailing list -- ceph-users@xxxxxxx
    <mailto:ceph-users@xxxxxxx>
     >> To unsubscribe send an email to ceph-users-leave@xxxxxxx
    <mailto:ceph-users-leave@xxxxxxx>
     >
     >
     > _______________________________________________
     > ceph-users mailing list -- ceph-users@xxxxxxx
    <mailto:ceph-users@xxxxxxx>
     > To unsubscribe send an email to ceph-users-leave@xxxxxxx
    <mailto:ceph-users-leave@xxxxxxx>
    _______________________________________________
    ceph-users mailing list -- ceph-users@xxxxxxx
    <mailto:ceph-users@xxxxxxx>
    To unsubscribe send an email to ceph-users-leave@xxxxxxx
    <mailto:ceph-users-leave@xxxxxxx>

_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux