[root@OSKVM1 ~]# grep -v "^#" /etc/nova/nova.conf|grep -v ^$
[DEFAULT]
instance_usage_audit = True
instance_usage_audit_period = hour
notify_on_state_change = vm_and_task_state
notification_driver = messagingv2
rbd_user=cinder
rbd_secret_uuid=1989f7a6-4ecb-4738-abbf-2962c29b2bbb
rpc_backend = rabbit
auth_strategy = keystone
my_ip = 10.1.0.4
network_api_class = nova.network.neutronv2.api.API
security_group_api = neutron
linuxnet_interface_driver = nova.network.linux_net.NeutronLinuxBridgeInterfaceDriver
firewall_driver = nova.virt.firewall.NoopFirewallDriver
enabled_apis=osapi_compute,metadata
[api_database]
connection = mysql://nova:nova@controller/nova
[barbican]
[cells]
[cinder]
os_region_name = RegionOne
[conductor]
[cors]
[cors.subdomain]
[database]
[ephemeral_storage_encryption]
[glance]
host = controller
[guestfs]
[hyperv]
[image_file_url]
[ironic]
[keymgr]
[keystone_authtoken]
auth_uri = http://controller:5000
auth_url = http://controller:35357
auth_plugin = password
project_domain_id = default
user_domain_id = default
project_name = service
username = nova
password = nova
[libvirt]
inject_password=false
inject_key=false
inject_partition=-2
live_migration_flag=VIR_MIGRATE_UNDEFINE_SOURCE, VIR_MIGRATE_PEER2PEER, VIR_MIGRATE_LIVE, VIR_MIGRATE_PERSIST_DEST, VIR_MIGRATE_TUNNELLED
disk_cachemodes ="network=writeback"
images_type=rbd
images_rbd_pool=vms
images_rbd_ceph_conf =/etc/ceph/ceph.conf
rbd_user=cinder
rbd_secret_uuid=1989f7a6-4ecb-4738-abbf-2962c29b2bbb
hw_disk_discard=unmap
[matchmaker_redis]
[matchmaker_ring]
[metrics]
[neutron]
url = "" href="http://controller:9696">http://controller:9696
auth_url = http://controller:35357
auth_plugin = password
project_domain_id = default
user_domain_id = default
region_name = RegionOne
project_name = service
username = neutron
password = neutron
service_metadata_proxy = True
metadata_proxy_shared_secret = XXXXX
[osapi_v21]
[oslo_concurrency]
lock_path = /var/lib/nova/tmp
[oslo_messaging_amqp]
[oslo_messaging_qpid]
[oslo_messaging_rabbit]
rabbit_host = controller
rabbit_userid = openstack
rabbit_password = XXXXX
[oslo_middleware]
[rdp]
[serial_console]
[spice]
[ssl]
[trusted_computing]
[upgrade_levels]
[vmware]
[vnc]
enabled = True
vncserver_listen = 0.0.0.0
novncproxy_base_url = http://controller:6080/vnc_auto.html
vncserver_proxyclient_address = $my_ip
[workarounds]
[xenserver]
[zookeeper]
[root@OSKVM1 ceph]# ls -ltr
total 24
-rwxr-xr-x 1 root root 92 May 10 12:58 rbdmap
-rw------- 1 root root 0 Jun 28 11:05 tmpfDt6jw
-rw-r--r-- 1 root root 63 Jul 5 12:59 ceph.client.admin.keyring
-rw-r--r-- 1 glance glance 64 Jul 5 14:51 ceph.client.glance.keyring
-rw-r--r-- 1 cinder cinder 64 Jul 5 14:53 ceph.client.cinder.keyring
-rw-r--r-- 1 cinder cinder 71 Jul 5 14:54 ceph.client.cinder-backup.keyring
-rwxrwxrwx 1 root root 438 Jul 7 14:19 ceph.conf
[root@OSKVM1 ceph]# more ceph.client.cinder.keyring
[client.cinder]
key = AQCIAHxX9ga8LxAAU+S3Vybdu+Cm2bP3lplGnA==
[root@OSKVM1 ~]# rados lspools
rbd
volumes
images
backups
vms
[root@OSKVM1 ~]# rbd -p rbd ls
[root@OSKVM1 ~]# rbd -p volumes ls
volume-27717a88-3c80-420f-8887-4ca5c5b94023
volume-3bd22868-cb2a-4881-b9fb-ae91a6f79cb9
volume-b9cf7b94-cfb6-4b55-816c-10c442b23519
[root@OSKVM1 ~]# rbd -p images ls
9aee6c4e-3b60-49d5-8e17-33953e384a00
a8b45c8a-a5c8-49d8-a529-1e4088bdbf3f
[root@OSKVM1 ~]# rbd -p vms ls
[root@OSKVM1 ~]# rbd -p backup
I could create cinder and attach it to one of already built nova instance.
[root@OSKVM1 ceph]# nova volume-list
WARNING: Command volume-list is deprecated and will be removed after Nova 13.0.0 is released. Use python-cinderclient or openstackclient instead.
+--------------------------------------+-----------+------------------+------+-------------+--------------------------------------+
| ID | Status | Display Name | Size | Volume Type | Attached to |
+--------------------------------------+-----------+------------------+------+-------------+--------------------------------------+
| 14a572d0-2834-40d6-9650-cb3e18271963 | available | nova-vol_gg | 10 | - | |
| 3bd22868-cb2a-4881-b9fb-ae91a6f79cb9 | in-use | nova-vol_1 | 2 | - | d06f7c3b-5bbd-4597-99ce-fa981d2e10db |
| 27717a88-3c80-420f-8887-4ca5c5b94023 | available | cinder-ceph-vol1 | 10 | - | |
+--------------------------------------+-----------+------------------+------+-------------+--------------------------------------+
Hi Kees,I regenerated the UUID as per your suggestion.Now i have same UUID in host1 and host2.I could create volumes and attach them to existing VMs.I could create new glance images.But still finding the same error while instance launch via GUI.2016-07-08 11:23:25.067 86007 INFO nova.compute.resource_tracker [req-4b7eccc8-0bf5-4f55-a941-4c93e97ef5df - - - - -] Auditing locally available compute resources for node controller
2016-07-08 11:23:25.527 86007 INFO nova.compute.resource_tracker [req-4b7eccc8-0bf5-4f55-a941-4c93e97ef5df - - - - -] Total usable vcpus: 40, total allocated vcpus: 0
2016-07-08 11:23:25.527 86007 INFO nova.compute.resource_tracker [req-4b7eccc8-0bf5-4f55-a941-4c93e97ef5df - - - - -] Final resource view: name=controller phys_ram=193168MB used_ram=1024MB phys_disk=8168GB used_disk=1GB total_vcpus=40 used_vcpus=0 pci_stats=None
2016-07-08 11:23:25.560 86007 INFO nova.compute.resource_tracker [req-4b7eccc8-0bf5-4f55-a941-4c93e97ef5df - - - - -] Compute_service record updated for OSKVM1:controller
2016-07-08 11:24:25.065 86007 INFO nova.compute.resource_tracker [req-4b7eccc8-0bf5-4f55-a941-4c93e97ef5df - - - - -] Auditing locally available compute resources for node controller
2016-07-08 11:24:25.561 86007 INFO nova.compute.resource_tracker [req-4b7eccc8-0bf5-4f55-a941-4c93e97ef5df - - - - -] Total usable vcpus: 40, total allocated vcpus: 0
2016-07-08 11:24:25.562 86007 INFO nova.compute.resource_tracker [req-4b7eccc8-0bf5-4f55-a941-4c93e97ef5df - - - - -] Final resource view: name=controller phys_ram=193168MB used_ram=1024MB phys_disk=8168GB used_disk=1GB total_vcpus=40 used_vcpus=0 pci_stats=None
2016-07-08 11:24:25.603 86007 INFO nova.compute.resource_tracker [req-4b7eccc8-0bf5-4f55-a941-4c93e97ef5df - - - - -] Compute_service record updated for OSKVM1:controller
2016-07-08 11:25:18.138 86007 INFO nova.compute.manager [req-3173f5b7-fa02-420c-954b-e21c3ce8d183 289598890db341f4af45ce5c57c41ba3 713114f3b9e54501a35a79e84c1e6c9d - - -] [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] Starting instance...
2016-07-08 11:25:18.255 86007 INFO nova.compute.claims [req-3173f5b7-fa02-420c-954b-e21c3ce8d183 289598890db341f4af45ce5c57c41ba3 713114f3b9e54501a35a79e84c1e6c9d - - -] [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] Attempting claim: memory 512 MB, disk 1 GB
2016-07-08 11:25:18.255 86007 INFO nova.compute.claims [req-3173f5b7-fa02-420c-954b-e21c3ce8d183 289598890db341f4af45ce5c57c41ba3 713114f3b9e54501a35a79e84c1e6c9d - - -] [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] Total memory: 193168 MB, used: 1024.00 MB
2016-07-08 11:25:18.256 86007 INFO nova.compute.claims [req-3173f5b7-fa02-420c-954b-e21c3ce8d183 289598890db341f4af45ce5c57c41ba3 713114f3b9e54501a35a79e84c1e6c9d - - -] [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] memory limit: 289752.00 MB, free: 288728.00 MB
2016-07-08 11:25:18.256 86007 INFO nova.compute.claims [req-3173f5b7-fa02-420c-954b-e21c3ce8d183 289598890db341f4af45ce5c57c41ba3 713114f3b9e54501a35a79e84c1e6c9d - - -] [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] Total disk: 8168 GB, used: 1.00 GB
2016-07-08 11:25:18.257 86007 INFO nova.compute.claims [req-3173f5b7-fa02-420c-954b-e21c3ce8d183 289598890db341f4af45ce5c57c41ba3 713114f3b9e54501a35a79e84c1e6c9d - - -] [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] disk limit: 8168.00 GB, free: 8167.00 GB
2016-07-08 11:25:18.271 86007 INFO nova.compute.claims [req-3173f5b7-fa02-420c-954b-e21c3ce8d183 289598890db341f4af45ce5c57c41ba3 713114f3b9e54501a35a79e84c1e6c9d - - -] [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] Claim successful
2016-07-08 11:25:18.747 86007 INFO nova.virt.libvirt.driver [req-3173f5b7-fa02-420c-954b-e21c3ce8d183 289598890db341f4af45ce5c57c41ba3 713114f3b9e54501a35a79e84c1e6c9d - - -] [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] Creating image
2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [req-3173f5b7-fa02-420c-954b-e21c3ce8d183 289598890db341f4af45ce5c57c41ba3 713114f3b9e54501a35a79e84c1e6c9d - - -] [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] Instance failed to spawn
2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] Traceback (most recent call last):
2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] File "/usr/lib/python2.7/site-packages/nova/compute/manager.py", line 2156, in _build_resources
2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] yield resources
2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] File "/usr/lib/python2.7/site-packages/nova/compute/manager.py", line 2009, in _build_and_run_instance
2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] block_device_info=block_device_info)
2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] File "/usr/lib/python2.7/site-packages/nova/virt/libvirt/driver.py", line 2527, in spawn
2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] admin_pass=admin_password)
2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] File "/usr/lib/python2.7/site-packages/nova/virt/libvirt/driver.py", line 2953, in _create_image
2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] instance, size, fallback_from_host)
2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] File "/usr/lib/python2.7/site-packages/nova/virt/libvirt/driver.py", line 6406, in _try_fetch_image_cache
2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] size=size)
2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] File "/usr/lib/python2.7/site-packages/nova/virt/libvirt/imagebackend.py", line 240, in cache
2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] *args, **kwargs)
2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] File "/usr/lib/python2.7/site-packages/nova/virt/libvirt/imagebackend.py", line 811, in create_image
2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] prepare_template(target=base, max_size=size, *args, **kwargs)
2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] File "/usr/lib/python2.7/site-packages/oslo_concurrency/lockutils.py", line 254, in inner
2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] return f(*args, **kwargs)
2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] File "/usr/lib/python2.7/site-packages/nova/virt/libvirt/imagebackend.py", line 230, in fetch_func_sync
2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] fetch_func(target=target, *args, **kwargs)
2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] File "/usr/lib/python2.7/site-packages/nova/virt/libvirt/driver.py", line 2947, in clone_fallback_to_fetch
2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] libvirt_utils.fetch_image(*args, **kwargs)
2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] File "/usr/lib/python2.7/site-packages/nova/virt/libvirt/utils.py", line 408, in fetch_image
2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] max_size=max_size)
2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] File "/usr/lib/python2.7/site-packages/nova/virt/images.py", line 123, in fetch_to_raw
2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] max_size=max_size)
2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] File "/usr/lib/python2.7/site-packages/nova/virt/images.py", line 113, in fetch
2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] IMAGE_API.download(context, image_href, dest_path=path)
2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] File "/usr/lib/python2.7/site-packages/nova/image/api.py", line 182, in download
2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] dst_path=dest_path)
2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] File "/usr/lib/python2.7/site-packages/nova/image/glance.py", line 367, in download
2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] _reraise_translated_image_exception(image_id)
2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] File "/usr/lib/python2.7/site-packages/nova/image/glance.py", line 613, in _reraise_translated_image_exception
2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] six.reraise(new_exc, None, exc_trace)
2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] File "/usr/lib/python2.7/site-packages/nova/image/glance.py", line 365, in download
2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] image_chunks = self._client.call(context, 1, 'data', image_id)
2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] File "/usr/lib/python2.7/site-packages/nova/image/glance.py", line 231, in call
2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] result = getattr(client.images, method)(*args, **kwargs)
2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] File "/usr/lib/python2.7/site-packages/glanceclient/v1/images.py", line 148, in data
2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] % urlparse.quote(str(image_id)))
2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] File "/usr/lib/python2.7/site-packages/glanceclient/common/http.py", line 280, in get
2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] return self._request('GET', url, **kwargs)
2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] File "/usr/lib/python2.7/site-packages/glanceclient/common/http.py", line 272, in _request
2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] resp, body_iter = self._handle_response(resp)
2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] File "/usr/lib/python2.7/site-packages/glanceclient/common/http.py", line 93, in _handle_response
2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] raise exc.from_response(resp, resp.content)
2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] Invalid: 400 Bad Request: Unknown scheme 'file' found in URI (HTTP 400)
2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7]
2016-07-08 11:25:19.575 86007 INFO nova.compute.manager [req-3173f5b7-fa02-420c-954b-e21c3ce8d183 289598890db341f4af45ce5c57c41ba3 713114f3b9e54501a35a79e84c1e6c9d - - -] [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] Terminating instance
2016-07-08 11:25:19.583 86007 INFO nova.virt.libvirt.driver [-] [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] During wait destroy, instance disappeared.
2016-07-08 11:25:19.665 86007 INFO nova.virt.libvirt.driver [req-3173f5b7-fa02-420c-954b-e21c3ce8d183 289598890db341f4af45ce5c57c41ba3 713114f3b9e54501a35a79e84c1e6c9d - - -] [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] Deleting instance files /var/lib/nova/instances/bf4839c8-2af6-4959-9158-fe411e1cfae7_del
2016-07-08 11:25:19.666 86007 INFO nova.virt.libvirt.driver [req-3173f5b7-fa02-420c-954b-e21c3ce8d183 289598890db341f4af45ce5c57c41ba3 713114f3b9e54501a35a79e84c1e6c9d - - -] [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] Deletion of /var/lib/nova/instances/bf4839c8-2af6-4959-9158-fe411e1cfae7_del complete
2016-07-08 11:25:26.073 86007 INFO nova.compute.resource_tracker [req-4b7eccc8-0bf5-4f55-a941-4c93e97ef5df - - - - -] Auditing locally available compute resources for node controller
2016-07-08 11:25:26.477 86007 INFO nova.compute.resource_tracker [req-4b7eccc8-0bf5-4f55-a941-4c93e97ef5df - - - - -] Total usable vcpus: 40, total allocated vcpus: 0
2016-07-08 11:25:26.478 86007 INFO nova.compute.resource_tracker [req-4b7eccc8-0bf5-4f55-a941-4c93e97ef5df - - - - -] Final resource view: name=controller phys_ram=193168MB used_ram=1024MB phys_disk=8168GB used_disk=1GB total_vcpus=40 used_vcpus=0 pci_stats=None
RegardsGaurav GoyalOn Fri, Jul 8, 2016 at 10:17 AM, Kees Meijs <kees@xxxxxxxx> wrote:Hi,
I'd recommend generating an UUID and use it for all your compute nodes.
This way, you can keep your configuration in libvirt constant.
Regards,
Kees
On 08-07-16 16:15, Gaurav Goyal wrote:
>
> For below section, should i generate separate UUID for both compte hosts?
>
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com