Re: luminous 12.2.2 traceback (ceph fs status)

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Yes, it include all the available pools on the cluster:

# ceph df
GLOBAL:
    SIZE       AVAIL      RAW USED     %RAW USED
    53650G     42928G       10722G         19.99
POOLS:
    NAME                ID     USED      %USED     MAX AVAIL     OBJECTS
    volumes             13     2979G     33.73         5854G      767183
    db                  18      856G      4.65        17563G     1657174
    cephfs_data         22       880         0         5854G           6
    cephfs_metadata     23      977k         0         5854G          65

# rados lspools
volumes
db
cephfs_data
cephfs_metadata

The goods news is that after restarting the ceph-mgr, it started to work :) but like you said, it would be nice to know how the system got into this.

Thanks a lot John :)

Best,


German

2017-12-11 12:17 GMT-03:00 John Spray <jspray@xxxxxxxxxx>:
On Mon, Dec 11, 2017 at 3:13 PM, German Anders <ganders@xxxxxxxxxxxx> wrote:
> Hi John,
>
> how are you? no problem :) . Unfortunately the error on the 'ceph fs status'
> command is still happening:

OK, can you check:
 - does the "ceph df" output include all the pools?
 - does restarting ceph-mgr clear the issue?

We probably need to modify this code to handle stats-less pools
anyway, but I'm curious about how the system got into this state.

John


> # ceph fs status
> Error EINVAL: Traceback (most recent call last):
>   File "/usr/lib/ceph/mgr/status/module.py", line 301, in handle_command
>     return self.handle_fs_status(cmd)
>   File "/usr/lib/ceph/mgr/status/module.py", line 219, in handle_fs_status
>     stats = pool_stats[pool_id]
> KeyError: (15L,)
>
>
>
> German
> 2017-12-11 12:08 GMT-03:00 John Spray <jspray@xxxxxxxxxx>:
>>
>> On Mon, Dec 4, 2017 at 6:37 PM, German Anders <ganders@xxxxxxxxxxxx>
>> wrote:
>> > Hi,
>> >
>> > I just upgrade a ceph cluster from version 12.2.0 (rc) to 12.2.2
>> > (stable),
>> > and i'm getting a traceback while trying to run:
>> >
>> > # ceph fs status
>> >
>> > Error EINVAL: Traceback (most recent call last):
>> >   File "/usr/lib/ceph/mgr/status/module.py", line 301, in handle_command
>> >     return self.handle_fs_status(cmd)
>> >   File "/usr/lib/ceph/mgr/status/module.py", line 219, in
>> > handle_fs_status
>> >     stats = pool_stats[pool_id]
>> > KeyError: (15L,)
>> >
>> >
>> > # ceph fs ls
>> > name: cephfs, metadata pool: cephfs_metadata, data pools: [cephfs_data ]
>> >
>> >
>> > Any ideas?
>>
>> (I'm a bit late but...)
>>
>> Is this still happening or did it self-correct?  It could have been
>> happening when the pool had just been created but the mgr hadn't heard
>> about any stats from the OSDs about that pool yet (which we should
>> fix, anyway)
>>
>> John
>>
>>
>> >
>> > Thanks in advance,
>> >
>> > Germ
>> > an
>> >
>> > _______________________________________________
>> > ceph-users mailing list
>> > ceph-users@xxxxxxxxxxxxxx
>> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>> >
>
>

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux