You are correct I am using the cephadm to create the iscsi portals. The
cluster had been one I was learning a lot with and I wondered if it was
because of the number of creations and deletions of things, so I rebuilt
the cluster, now I am getting this response even when creating my first
iscsi target. Here is the output of the gwcli ls:
sh-4.4# gwcli ls
o- /
........................................................................................................................
[...]
o- cluster
........................................................................................................
[Clusters: 1]
| o- ceph
.........................................................................................................
[HEALTH_WARN]
| o- pools
.........................................................................................................
[Pools: 8]
| | o- .rgw.root
............................................................ [(x3),
Commit: 0.00Y/71588776M (0%), Used: 1323b]
| | o- cephfs_data
.......................................................... [(x3),
Commit: 0.00Y/71588776M (0%), Used: 1639b]
| | o- cephfs_metadata
...................................................... [(x3), Commit:
0.00Y/71588776M (0%), Used: 3434b]
| | o- default.rgw.control
.................................................. [(x3), Commit:
0.00Y/71588776M (0%), Used: 0.00Y]
| | o- default.rgw.log
...................................................... [(x3), Commit:
0.00Y/71588776M (0%), Used: 3702b]
| | o- default.rgw.meta
...................................................... [(x3), Commit:
0.00Y/71588776M (0%), Used: 382b]
| | o- device_health_metrics
................................................ [(x3), Commit:
0.00Y/71588776M (0%), Used: 0.00Y]
| | o- rhv-ceph-ssd
..................................................... [(x3), Commit:
0.00Y/7868560896K (0%), Used: 511746b]
| o- topology
..............................................................................................
[OSDs: 36,MONs: 3]
o- disks
......................................................................................................
[0.00Y, Disks: 0]
o- iscsi-targets
..............................................................................
[DiscoveryAuth: None, Targets: 1]
o- iqn.2001-07.com.ceph:1668466555428
............................................................... [Auth:
None, Gateways: 1]
o- disks
.........................................................................................................
[Disks: 0]
o- gateways
...........................................................................................
[Up: 1/1, Portals: 1]
| o- host.containers.internal
........................................................................
[192.168.105.145 (UP)]
o- host-groups
.................................................................................................
[Groups : 0]
o- hosts
......................................................................................
[Auth: ACL_ENABLED, Hosts: 0]
sh-4.4#
Randy
On 11/9/2022 6:36 PM, Xiubo Li wrote:
On 10/11/2022 02:21, Randy Morgan wrote:
I am trying to create a second iscsi target and I keep getting an
error when I create the second target:
Failed to update target 'iqn.2001-07.com.ceph:1667946365517'
disk create/update failed on host.containers.internal. LUN allocation
failure
I think you were using the cephadm to add the iscsi targets, not the
gwcli or Rest APIs directly.
Before we hit other issues were login failures, that because there
were two gateways using the same IP address. Please share your `gwcli
ls` output to see what the 'host.containers.internal' gateway's config.
Thanks!
I am running ceph Pacific: *Version*
16.2.7 (dd0603118f56ab514f133c8d2e3adfc983942503)
pacific (stable)
All of the information I can find on this problem is from 3 years ago
and doesn't seem to apply any more. Does anyone know how to correct
this problem?
Randy
--
Randy Morgan
IT Manager
Department of Chemistry/BioChemistry
Brigham Young University
randym@xxxxxxxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx