ceph rbd iscsi gwcli Non-existent images

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,
I would appreciate any help/hints to solve this issue
   iscis (gwcli) cannot see the images anymore

This configuration worked fine for many months
What changed was that ceph is "nearly full"

I am in the process of cleaning it up ( by deleting objects from one of the
pools)
and I do see reads and writes on the cluster as well as images info so not
sure what gwcli does not like

( targetcli ls not working either - just froze )

Below some info

ceph version
ceph version 13.2.2 (02899bfda814146b021136e9d8e80eba494e1126) mimic
(stable)

 gwcli --version
gwcli - 2.7

 ceph osd dump | grep ratio

full_ratio 0.96
backfillfull_ratio 0.92
nearfull_ratio 0.9

[root@osd02 ~]# rbd -p rbd info rep01
rbd image 'rep01':
        size 7 TiB in 1835008 objects
        order 22 (4 MiB objects)
        id: 15b366b8b4567
        block_name_prefix: rbd_data.15b366b8b4567
        format: 2
        features: layering, exclusive-lock, object-map, fast-diff,
deep-flatten
        op_features:
        flags:
        create_timestamp: Thu Nov  1 15:57:52 2018
[root@osd02 ~]# rbd -p rbd info vmware01
rbd image 'vmware01':
        size 6 TiB in 1572864 objects
        order 22 (4 MiB objects)
        id: 16d3f6b8b4567
        block_name_prefix: rbd_data.16d3f6b8b4567
        format: 2
        features: layering, exclusive-lock, object-map, fast-diff,
deep-flatten
        op_features:
        flags:
        create_timestamp: Thu Nov 29 13:56:28 2018
[root@osd02 ~]# ceph df
GLOBAL:
    SIZE       AVAIL       RAW USED     %RAW USED
    33 TiB     7.5 TiB       25 TiB         77.16
POOLS:
    NAME                ID     USED        %USED     MAX AVAIL     OBJECTS
    cephfs_metadata     22     173 MiB      0.01       1.4 TiB         469
    cephfs_data         23     1.7 TiB     69.78       775 GiB      486232
    rbd                 24      11 TiB     93.74       775 GiB     2974077
[root@osd02 ~]# ceph health detail
HEALTH_ERR 2 nearfull osd(s); 2 pool(s) nearfull; Module 'prometheus' has
failed: IOError("Port 9283 not free on '10.10.35.20'",)
OSD_NEARFULL 2 nearfull osd(s)
    osd.12 is near full
    osd.17 is near full
POOL_NEARFULL 2 pool(s) nearfull
    pool 'cephfs_data' is nearfull
    pool 'rbd' is nearfull



gwcli
/iscsi-target...nner-21faa413> info
Client Iqn .. iqn.1998-01.com.vmware:banner-21faa413
Ip Address ..
Alias      ..
Logged In  ..
Auth
- chap .. cephuser/PASSWORD
Group Name ..
Luns
- rbd.rep01    .. lun_id=0
- rbd.vmware01 .. lun_id=1



osd02 journal: client update failed on
iqn.1998-01.com.vmware:banner-21faa413 : Non-existent images
['rbd.vmware01'] requested for iqn.1998-01.com.vmware:banner-21faa413
Aug  7 14:15:39 osd02 journal: 127.0.0.1 - - [07/Aug/2020 14:15:39] "PUT
/api/_clientlun/iqn.1998-01.com.vmware:banner-21faa413 HTTP/1.1" 500 -
Aug  7 14:15:39 osd02 journal: _clientlun change on 127.0.0.1 failed with
500
Aug  7 14:15:39 osd02 journal: 127.0.0.1 - - [07/Aug/2020 14:15:39] "DELETE
/api/clientlun/iqn.1998-01.com.vmware:banner-21faa413 HTTP/1.1" 500 -
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux