Re: iSCSI over RBD

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



I too experienced this with that kernel as well as the elrepo kernel. 

On Jan 19, 2018 2:13 PM, "Steven Vacaroaia" <stef97@xxxxxxxxx> wrote:
Hi Joshua,

I was under the impression that kernel  3.10.0-693 will work with iscsi
 
Unfortunately  I still cannot create a disk because qfull_time_out is not supported 

What am I missing / do it wrong ?

2018-01-19 15:06:45,216     INFO [lun.py:601:add_dev_to_lio()] - (LUN.add_dev_to_lio) Adding image 'rbd.disk2' to LIO
2018-01-19 15:06:45,295    ERROR [lun.py:634:add_dev_to_lio()] - Could not set LIO device attribute cmd_time_out/qfull_time_out for device: rbd.disk2. Kernel not supported. - error(Cannot find attribute: qfull_time_out)
2018-01-19 15:06:45,300    ERROR [rbd-target-api:731:_disk()] - LUN alloc problem - Could not set LIO device attribute cmd_time_out/qfull_time_out for device: rbd.disk2. Kernel not supported. - error(Cannot find attribute: qfull_time_out)


Many thanks

Steven

On 4 January 2018 at 22:40, Joshua Chen <cschen@xxxxxxxxxxxxxxxxxxx> wrote:
Hello Steven,
  I am using CentOS 7.4.1708 with kernel 3.10.0-693.el7.x86_64
  and the following packages:

ceph-iscsi-cli-2.5-9.el7.centos.noarch.rpm
ceph-iscsi-config-2.3-12.el7.centos.noarch.rpm
libtcmu-1.3.0-0.4.el7.centos.x86_64.rpm
libtcmu-devel-1.3.0-0.4.el7.centos.x86_64.rpm
python-rtslib-2.1.fb64-2.el7.centos.noarch.rpm
python-rtslib-doc-2.1.fb64-2.el7.centos.noarch.rpm
targetcli-2.1.fb47-0.1.20170815.git5bf3517.el7.centos.noarch.rpm
tcmu-runner-1.3.0-0.4.el7.centos.x86_64.rpm
tcmu-runner-debuginfo-1.3.0-0.4.el7.centos.x86_64.rpm


Cheers
Joshua


On Fri, Jan 5, 2018 at 2:14 AM, Steven Vacaroaia <stef97@xxxxxxxxx> wrote:
Hi Joshua,

How did you manage to use iSCSI gateway ?
I would like to do that but still waiting for a patched kernel 

What kernel/OS did you use and/or how did you patch it ?

Tahnsk
Steven

On 4 January 2018 at 04:50, Joshua Chen <cschen@xxxxxxxxxxxxxxxxxxx> wrote:
Dear all,
  Although I managed to run gwcli and created some iqns, or luns,
but I do need some working config example so that my initiator could connect and get the lun.

  I am familiar with targetcli and I used to do the following ACL style connection rather than password, 
the targetcli setting tree is here:

(or see this page)

#targetcli ls
o- / ......................................................................................................................... [...]
  o- backstores .............................................................................................................. [...]
  | o- block .................................................................................................. [Storage Objects: 1]
  | | o- vmware_5t .......................................................... [/dev/rbd/rbd/vmware_5t (5.0TiB) write-thru activated]
  | |   o- alua ................................................................................................... [ALUA Groups: 1]
  | |     o- default_tg_pt_gp ....................................................................... [ALUA state: Active/optimized]
  | o- fileio ................................................................................................. [Storage Objects: 0]
  | o- pscsi .................................................................................................. [Storage Objects: 0]
  | o- ramdisk ................................................................................................ [Storage Objects: 0]
  | o- user:rbd ............................................................................................... [Storage Objects: 0]
  o- iscsi ............................................................................................................ [Targets: 1]
  | o- iqn.2017-12.asiaa.cephosd1:vmware5t ............................................................................... [TPGs: 1]
  |   o- tpg1 .................................................................................................. [gen-acls, no-auth]
  |     o- acls ......................................................................................................... [ACLs: 12]
  |     | o- iqn.1994-05.com.redhat:15dbed23be9e .................................................................. [Mapped LUNs: 1]
  |     | | o- mapped_lun0 ............................................................................. [lun0 block/vmware_5t (rw)]
  |     | o- iqn.1994-05.com.redhat:15dbed23be9e-ovirt1 ........................................................... [Mapped LUNs: 1]
  |     | | o- mapped_lun0 ............................................................................. [lun0 block/vmware_5t (rw)]
  |     | o- iqn.1994-05.com.redhat:2af344ba6ae5-ceph-admin-test .................................................. [Mapped LUNs: 1]
  |     | | o- mapped_lun0 ............................................................................. [lun0 block/vmware_5t (rw)]
  |     | o- iqn.1994-05.com.redhat:67669afedddf .................................................................. [Mapped LUNs: 1]
  |     | | o- mapped_lun0 ............................................................................. [lun0 block/vmware_5t (rw)]
  |     | o- iqn.1994-05.com.redhat:67669afedddf-ovirt3 ........................................................... [Mapped LUNs: 1]
  |     | | o- mapped_lun0 ............................................................................. [lun0 block/vmware_5t (rw)]
  |     | o- iqn.1994-05.com.redhat:a7c1ec3c43f7 .................................................................. [Mapped LUNs: 1]
  |     | | o- mapped_lun0 ............................................................................. [lun0 block/vmware_5t (rw)]
  |     | o- iqn.1994-05.com.redhat:a7c1ec3c43f7-ovirt2 ........................................................... [Mapped LUNs: 1]
  |     | | o- mapped_lun0 ............................................................................. [lun0 block/vmware_5t (rw)]
  |     | o- iqn.1994-05.com.redhat:b01662ec2129-ceph-node2 ....................................................... [Mapped LUNs: 1]
  |     | | o- mapped_lun0 ............................................................................. [lun0 block/vmware_5t (rw)]
  |     | o- iqn.1994-05.com.redhat:d46b42a1915b-ceph-node3 ....................................................... [Mapped LUNs: 1]
  |     | | o- mapped_lun0 ............................................................................. [lun0 block/vmware_5t (rw)]
  |     | o- iqn.1994-05.com.redhat:e7692a10f661-ceph-node1 ....................................................... [Mapped LUNs: 1]
  |     | | o- mapped_lun0 ............................................................................. [lun0 block/vmware_5t (rw)]
  |     | o- iqn.1998-01.com.vmware:localhost-0f904dfd ............................................................ [Mapped LUNs: 1]
  |     | | o- mapped_lun0 ............................................................................. [lun0 block/vmware_5t (rw)]
  |     | o- iqn.1998-01.com.vmware:localhost-6af62e4c ............................................................ [Mapped LUNs: 1]
  |     |   o- mapped_lun0 ............................................................................. [lun0 block/vmware_5t (rw)]
  |     o- luns .......................................................................................................... [LUNs: 1]
  |     | o- lun0 .................................................... [block/vmware_5t (/dev/rbd/rbd/vmware_5t) (default_tg_pt_gp)]
  |     o- portals .................................................................................................... [Portals: 1]
  |       o- 172.20.0.12:3260 ................................................................................................. [OK]
  o- loopback ......................................................................................................... [Targets: 0]
  o- xen_pvscsi ....................................................................................................... [Targets: 0]






My targetcli setup procedure is like this, could someone translate it to gwcli equivalent procedure?
sorry for asking for this due to lack of documentation and examples.
thanks in adavance 

Cheers
Joshua




targetcli /backstores/block create name=vmware_5t dev=/dev/rbd/rbd/vmware_5t
targetcli /iscsi/ create iqn.2017-12.asiaa.cephosd1:vmware5t
targetcli /iscsi/iqn.2017-12.asiaa.cephosd1:vmware5t/tpg1/portals delete ip_address=0.0.0.0 ip_port=3260

targetcli
  cd /iscsi/iqn.2017-12.asiaa.cephosd1:vmware5t/tpg1
        portals/  create 172.20.0.12
        acls/
                create iqn.1994-05.com.redhat:e7692a10f661-ceph-node1
                create iqn.1994-05.com.redhat:b01662ec2129-ceph-node2
                create iqn.1994-05.com.redhat:d46b42a1915b-ceph-node3
                create iqn.1994-05.com.redhat:15dbed23be9e
                create iqn.1994-05.com.redhat:a7c1ec3c43f7
                create iqn.1994-05.com.redhat:67669afedddf
                create iqn.1994-05.com.redhat:15dbed23be9e-ovirt1
                create iqn.1994-05.com.redhat:a7c1ec3c43f7-ovirt2
                create iqn.1994-05.com.redhat:67669afedddf-ovirt3
                create iqn.1994-05.com.redhat:2af344ba6ae5-ceph-admin-test
                create iqn.1998-01.com.vmware:localhost-6af62e4c
                create iqn.1998-01.com.vmware:localhost-0f904dfd
                cd ..
                set attribute generate_node_acls=1
                cd luns
                create /backstores/block/vmware_5t




On Thu, Jan 4, 2018 at 10:55 AM, Joshua Chen <cschen@xxxxxxxxxxxxxxxxxxx> wrote:
I had the same problem before, mine is CentOS, and when I created
/iscsi/create iqn_bla-bla
it goes
ocal LIO instance already has LIO configured with a target - unable to  continue



then finally the solution happened to be, turn off target service

systemctl stop target
systemctl disable target


somehow they are doing the same thing, you need to disable 'target' service (targetcli) in order to allow gwcli (rbd-target-api) do it's job.

Cheers
Joshua

On Thu, Jan 4, 2018 at 2:39 AM, Mike Christie <mchristi@xxxxxxxxxx> wrote:
On 12/25/2017 03:13 PM, Joshua Chen wrote:
> Hello folks,
>   I am trying to share my ceph rbd images through iscsi protocol.
>
> I am trying iscsi-gateway
> http://docs.ceph.com/docs/master/rbd/iscsi-overview/
>
>
> now
>
> systemctl start rbd-target-api
> is working and I could run gwcli
> (at a CentOS 7.4 osd node)
>
> gwcli
> /> ls
> o- /
> .........................................................................................................................
> [...]
>   o- clusters
> ........................................................................................................
> [Clusters: 1]
>   | o- ceph
> ............................................................................................................
> [HEALTH_OK]
>   |   o- pools
> ..........................................................................................................
> [Pools: 1]
>   |   | o- rbd
> ...........................................................................
> [(x3), Commit: 0b/25.9T (0%), Used: 395M]
>   |   o- topology
> ................................................................................................
> [OSDs: 9,MONs: 3]
>   o- disks
> ..........................................................................................................
> [0b, Disks: 0]
>   o- iscsi-target
> .....................................................................................................
> [Targets: 0]
>
>
> but when I created iscsi-target, I got
>
> Local LIO instance already has LIO configured with a target - unable to
> continue
>
>
> /> /iscsi-target create
> iqn.2003-01.org.linux-iscsi.ceph-node1.x8664:sn.571e1ab51af2
> Local LIO instance already has LIO configured with a target - unable to
> continue
> />
>


Could you send the output of

targetcli ls

?

What distro are you using?

You might just have a target setup from a non gwcli source. Maybe from
the distro targetcli systemd tools.



_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com





_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux