Yes, it include all the available pools on the cluster:
# ceph df
GLOBAL:
SIZE AVAIL RAW USED %RAW USED
53650G 42928G 10722G 19.99
POOLS:
NAME ID USED %USED MAX AVAIL OBJECTS
volumes 13 2979G 33.73 5854G 767183
db 18 856G 4.65 17563G 1657174
cephfs_data 22 880 0 5854G 6
cephfs_metadata 23 977k 0 5854G 65
# rados lspools
volumes
db
cephfs_data
cephfs_metadata
The goods news is that after restarting the ceph-mgr, it started to work :) but like you said, it would be nice to know how the system got into this.
Thanks a lot John :)
Best,
German
2017-12-11 12:17 GMT-03:00 John Spray <jspray@xxxxxxxxxx>:
On Mon, Dec 11, 2017 at 3:13 PM, German Anders <ganders@xxxxxxxxxxxx> wrote:
> Hi John,
>
> how are you? no problem :) . Unfortunately the error on the 'ceph fs status'
> command is still happening:
OK, can you check:
- does the "ceph df" output include all the pools?
- does restarting ceph-mgr clear the issue?
We probably need to modify this code to handle stats-less pools
anyway, but I'm curious about how the system got into this state.
John
> # ceph fs status
> Error EINVAL: Traceback (most recent call last):
> File "/usr/lib/ceph/mgr/status/module.py", line 301, in handle_command
> return self.handle_fs_status(cmd)
> File "/usr/lib/ceph/mgr/status/module.py", line 219, in handle_fs_status
> stats = pool_stats[pool_id]
> KeyError: (15L,)
>
>
>
> German
> 2017-12-11 12:08 GMT-03:00 John Spray <jspray@xxxxxxxxxx>:
>>
>> On Mon, Dec 4, 2017 at 6:37 PM, German Anders <ganders@xxxxxxxxxxxx>
>> wrote:
>> > Hi,
>> >
>> > I just upgrade a ceph cluster from version 12.2.0 (rc) to 12.2.2
>> > (stable),
>> > and i'm getting a traceback while trying to run:
>> >
>> > # ceph fs status
>> >
>> > Error EINVAL: Traceback (most recent call last):
>> > File "/usr/lib/ceph/mgr/status/module.py", line 301, in handle_command
>> > return self.handle_fs_status(cmd)
>> > File "/usr/lib/ceph/mgr/status/module.py", line 219, in
>> > handle_fs_status
>> > stats = pool_stats[pool_id]
>> > KeyError: (15L,)
>> >
>> >
>> > # ceph fs ls
>> > name: cephfs, metadata pool: cephfs_metadata, data pools: [cephfs_data ]
>> >
>> >
>> > Any ideas?
>>
>> (I'm a bit late but...)
>>
>> Is this still happening or did it self-correct? It could have been
>> happening when the pool had just been created but the mgr hadn't heard
>> about any stats from the OSDs about that pool yet (which we should
>> fix, anyway)
>>
>> John
>>
>>
>> >
>> > Thanks in advance,
>> >
>> > Germ
>> > an
>> >
>> > _______________________________________________
>> > ceph-users mailing list
>> > ceph-users@xxxxxxxxxxxxxx
>> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph. com
>> >
>
>
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com