Re: (no subject)

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



[root@OSKVM1 ~]# grep -v "^#" /etc/nova/nova.conf|grep -v ^$

[DEFAULT]

instance_usage_audit = True

instance_usage_audit_period = hour

notify_on_state_change = vm_and_task_state

notification_driver = messagingv2

rbd_user=cinder

rbd_secret_uuid=1989f7a6-4ecb-4738-abbf-2962c29b2bbb

rpc_backend = rabbit

auth_strategy = keystone

my_ip = 10.1.0.4

network_api_class = nova.network.neutronv2.api.API

security_group_api = neutron

linuxnet_interface_driver = nova.network.linux_net.NeutronLinuxBridgeInterfaceDriver

firewall_driver = nova.virt.firewall.NoopFirewallDriver

enabled_apis=osapi_compute,metadata

[api_database]

connection = mysql://nova:nova@controller/nova

[barbican]

[cells]

[cinder]

os_region_name = RegionOne

[conductor]

[cors]

[cors.subdomain]

[database]

[ephemeral_storage_encryption]

[glance]

host = controller

[guestfs]

[hyperv]

[image_file_url]

[ironic]

[keymgr]

[keystone_authtoken]

auth_uri = http://controller:5000

auth_url = http://controller:35357

auth_plugin = password

project_domain_id = default

user_domain_id = default

project_name = service

username = nova

password = nova

[libvirt]

inject_password=false

inject_key=false

inject_partition=-2

live_migration_flag=VIR_MIGRATE_UNDEFINE_SOURCE, VIR_MIGRATE_PEER2PEER, VIR_MIGRATE_LIVE, VIR_MIGRATE_PERSIST_DEST, VIR_MIGRATE_TUNNELLED

disk_cachemodes ="network=writeback"

images_type=rbd

images_rbd_pool=vms

images_rbd_ceph_conf =/etc/ceph/ceph.conf

rbd_user=cinder

rbd_secret_uuid=1989f7a6-4ecb-4738-abbf-2962c29b2bbb

hw_disk_discard=unmap

[matchmaker_redis]

[matchmaker_ring]

[metrics]

[neutron]

url = "" href="http://controller:9696">http://controller:9696

auth_url = http://controller:35357

auth_plugin = password

project_domain_id = default

user_domain_id = default

region_name = RegionOne

project_name = service

username = neutron

password = neutron

service_metadata_proxy = True

metadata_proxy_shared_secret = XXXXX

[osapi_v21]

[oslo_concurrency]

lock_path = /var/lib/nova/tmp

[oslo_messaging_amqp]

[oslo_messaging_qpid]

[oslo_messaging_rabbit]

rabbit_host = controller

rabbit_userid = openstack

rabbit_password = XXXXX

[oslo_middleware]

[rdp]

[serial_console]

[spice]

[ssl]

[trusted_computing]

[upgrade_levels]

[vmware]

[vnc]

enabled = True

vncserver_listen = 0.0.0.0

novncproxy_base_url = http://controller:6080/vnc_auto.html

vncserver_proxyclient_address = $my_ip

[workarounds]

[xenserver]

[zookeeper]


[root@OSKVM1 ceph]# ls -ltr

total 24

-rwxr-xr-x 1 root   root    92 May 10 12:58 rbdmap

-rw------- 1 root   root     0 Jun 28 11:05 tmpfDt6jw

-rw-r--r-- 1 root   root    63 Jul  5 12:59 ceph.client.admin.keyring

-rw-r--r-- 1 glance glance  64 Jul  5 14:51 ceph.client.glance.keyring

-rw-r--r-- 1 cinder cinder  64 Jul  5 14:53 ceph.client.cinder.keyring

-rw-r--r-- 1 cinder cinder  71 Jul  5 14:54 ceph.client.cinder-backup.keyring

-rwxrwxrwx 1 root   root   438 Jul  7 14:19 ceph.conf

[root@OSKVM1 ceph]# more ceph.client.cinder.keyring

[client.cinder]

key = AQCIAHxX9ga8LxAAU+S3Vybdu+Cm2bP3lplGnA==

[root@OSKVM1 ~]# rados lspools

rbd

volumes

images

backups

vms

[root@OSKVM1 ~]# rbd -p rbd ls

[root@OSKVM1 ~]# rbd -p volumes ls

volume-27717a88-3c80-420f-8887-4ca5c5b94023

volume-3bd22868-cb2a-4881-b9fb-ae91a6f79cb9

volume-b9cf7b94-cfb6-4b55-816c-10c442b23519

[root@OSKVM1 ~]# rbd -p images ls

9aee6c4e-3b60-49d5-8e17-33953e384a00

a8b45c8a-a5c8-49d8-a529-1e4088bdbf3f

[root@OSKVM1 ~]# rbd -p vms ls

[root@OSKVM1 ~]# rbd -p backup 


I could create cinder and  attach it to one of already built nova instance.

[root@OSKVM1 ceph]# nova volume-list

WARNING: Command volume-list is deprecated and will be removed after Nova 13.0.0 is released. Use python-cinderclient or openstackclient instead.

+--------------------------------------+-----------+------------------+------+-------------+--------------------------------------+

| ID                                   | Status    | Display Name     | Size | Volume Type | Attached to                          |

+--------------------------------------+-----------+------------------+------+-------------+--------------------------------------+

| 14a572d0-2834-40d6-9650-cb3e18271963 | available | nova-vol_gg      | 10   | -           |                                      |

| 3bd22868-cb2a-4881-b9fb-ae91a6f79cb9 | in-use    | nova-vol_1       | 2    | -           | d06f7c3b-5bbd-4597-99ce-fa981d2e10db |

| 27717a88-3c80-420f-8887-4ca5c5b94023 | available | cinder-ceph-vol1 | 10   | -           |                                      |

+--------------------------------------+-----------+------------------+------+-------------+--------------------------------------+


On Fri, Jul 8, 2016 at 11:33 AM, Gaurav Goyal <er.gauravgoyal@xxxxxxxxx> wrote:
Hi Kees,

I regenerated the UUID as per your suggestion. 
Now i have same UUID in host1 and host2.
I could create volumes and attach them to existing VMs.

I could create new glance images. 

But still finding the same error while instance launch via GUI.


2016-07-08 11:23:25.067 86007 INFO nova.compute.resource_tracker [req-4b7eccc8-0bf5-4f55-a941-4c93e97ef5df - - - - -] Auditing locally available compute resources for node controller

2016-07-08 11:23:25.527 86007 INFO nova.compute.resource_tracker [req-4b7eccc8-0bf5-4f55-a941-4c93e97ef5df - - - - -] Total usable vcpus: 40, total allocated vcpus: 0

2016-07-08 11:23:25.527 86007 INFO nova.compute.resource_tracker [req-4b7eccc8-0bf5-4f55-a941-4c93e97ef5df - - - - -] Final resource view: name=controller phys_ram=193168MB used_ram=1024MB phys_disk=8168GB used_disk=1GB total_vcpus=40 used_vcpus=0 pci_stats=None

2016-07-08 11:23:25.560 86007 INFO nova.compute.resource_tracker [req-4b7eccc8-0bf5-4f55-a941-4c93e97ef5df - - - - -] Compute_service record updated for OSKVM1:controller

2016-07-08 11:24:25.065 86007 INFO nova.compute.resource_tracker [req-4b7eccc8-0bf5-4f55-a941-4c93e97ef5df - - - - -] Auditing locally available compute resources for node controller

2016-07-08 11:24:25.561 86007 INFO nova.compute.resource_tracker [req-4b7eccc8-0bf5-4f55-a941-4c93e97ef5df - - - - -] Total usable vcpus: 40, total allocated vcpus: 0

2016-07-08 11:24:25.562 86007 INFO nova.compute.resource_tracker [req-4b7eccc8-0bf5-4f55-a941-4c93e97ef5df - - - - -] Final resource view: name=controller phys_ram=193168MB used_ram=1024MB phys_disk=8168GB used_disk=1GB total_vcpus=40 used_vcpus=0 pci_stats=None

2016-07-08 11:24:25.603 86007 INFO nova.compute.resource_tracker [req-4b7eccc8-0bf5-4f55-a941-4c93e97ef5df - - - - -] Compute_service record updated for OSKVM1:controller

2016-07-08 11:25:18.138 86007 INFO nova.compute.manager [req-3173f5b7-fa02-420c-954b-e21c3ce8d183 289598890db341f4af45ce5c57c41ba3 713114f3b9e54501a35a79e84c1e6c9d - - -] [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] Starting instance...

2016-07-08 11:25:18.255 86007 INFO nova.compute.claims [req-3173f5b7-fa02-420c-954b-e21c3ce8d183 289598890db341f4af45ce5c57c41ba3 713114f3b9e54501a35a79e84c1e6c9d - - -] [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] Attempting claim: memory 512 MB, disk 1 GB

2016-07-08 11:25:18.255 86007 INFO nova.compute.claims [req-3173f5b7-fa02-420c-954b-e21c3ce8d183 289598890db341f4af45ce5c57c41ba3 713114f3b9e54501a35a79e84c1e6c9d - - -] [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] Total memory: 193168 MB, used: 1024.00 MB

2016-07-08 11:25:18.256 86007 INFO nova.compute.claims [req-3173f5b7-fa02-420c-954b-e21c3ce8d183 289598890db341f4af45ce5c57c41ba3 713114f3b9e54501a35a79e84c1e6c9d - - -] [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] memory limit: 289752.00 MB, free: 288728.00 MB

2016-07-08 11:25:18.256 86007 INFO nova.compute.claims [req-3173f5b7-fa02-420c-954b-e21c3ce8d183 289598890db341f4af45ce5c57c41ba3 713114f3b9e54501a35a79e84c1e6c9d - - -] [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] Total disk: 8168 GB, used: 1.00 GB

2016-07-08 11:25:18.257 86007 INFO nova.compute.claims [req-3173f5b7-fa02-420c-954b-e21c3ce8d183 289598890db341f4af45ce5c57c41ba3 713114f3b9e54501a35a79e84c1e6c9d - - -] [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] disk limit: 8168.00 GB, free: 8167.00 GB

2016-07-08 11:25:18.271 86007 INFO nova.compute.claims [req-3173f5b7-fa02-420c-954b-e21c3ce8d183 289598890db341f4af45ce5c57c41ba3 713114f3b9e54501a35a79e84c1e6c9d - - -] [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] Claim successful

2016-07-08 11:25:18.747 86007 INFO nova.virt.libvirt.driver [req-3173f5b7-fa02-420c-954b-e21c3ce8d183 289598890db341f4af45ce5c57c41ba3 713114f3b9e54501a35a79e84c1e6c9d - - -] [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] Creating image

2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [req-3173f5b7-fa02-420c-954b-e21c3ce8d183 289598890db341f4af45ce5c57c41ba3 713114f3b9e54501a35a79e84c1e6c9d - - -] [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] Instance failed to spawn

2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] Traceback (most recent call last):

2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7]   File "/usr/lib/python2.7/site-packages/nova/compute/manager.py", line 2156, in _build_resources

2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7]     yield resources

2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7]   File "/usr/lib/python2.7/site-packages/nova/compute/manager.py", line 2009, in _build_and_run_instance

2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7]     block_device_info=block_device_info)

2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7]   File "/usr/lib/python2.7/site-packages/nova/virt/libvirt/driver.py", line 2527, in spawn

2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7]     admin_pass=admin_password)

2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7]   File "/usr/lib/python2.7/site-packages/nova/virt/libvirt/driver.py", line 2953, in _create_image

2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7]     instance, size, fallback_from_host)

2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7]   File "/usr/lib/python2.7/site-packages/nova/virt/libvirt/driver.py", line 6406, in _try_fetch_image_cache

2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7]     size=size)

2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7]   File "/usr/lib/python2.7/site-packages/nova/virt/libvirt/imagebackend.py", line 240, in cache

2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7]     *args, **kwargs)

2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7]   File "/usr/lib/python2.7/site-packages/nova/virt/libvirt/imagebackend.py", line 811, in create_image

2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7]     prepare_template(target=base, max_size=size, *args, **kwargs)

2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7]   File "/usr/lib/python2.7/site-packages/oslo_concurrency/lockutils.py", line 254, in inner

2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7]     return f(*args, **kwargs)

2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7]   File "/usr/lib/python2.7/site-packages/nova/virt/libvirt/imagebackend.py", line 230, in fetch_func_sync

2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7]     fetch_func(target=target, *args, **kwargs)

2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7]   File "/usr/lib/python2.7/site-packages/nova/virt/libvirt/driver.py", line 2947, in clone_fallback_to_fetch

2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7]     libvirt_utils.fetch_image(*args, **kwargs)

2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7]   File "/usr/lib/python2.7/site-packages/nova/virt/libvirt/utils.py", line 408, in fetch_image

2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7]     max_size=max_size)

2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7]   File "/usr/lib/python2.7/site-packages/nova/virt/images.py", line 123, in fetch_to_raw

2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7]     max_size=max_size)

2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7]   File "/usr/lib/python2.7/site-packages/nova/virt/images.py", line 113, in fetch

2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7]     IMAGE_API.download(context, image_href, dest_path=path)

2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7]   File "/usr/lib/python2.7/site-packages/nova/image/api.py", line 182, in download

2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7]     dst_path=dest_path)

2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7]   File "/usr/lib/python2.7/site-packages/nova/image/glance.py", line 367, in download

2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7]     _reraise_translated_image_exception(image_id)

2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7]   File "/usr/lib/python2.7/site-packages/nova/image/glance.py", line 613, in _reraise_translated_image_exception

2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7]     six.reraise(new_exc, None, exc_trace)

2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7]   File "/usr/lib/python2.7/site-packages/nova/image/glance.py", line 365, in download

2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7]     image_chunks = self._client.call(context, 1, 'data', image_id)

2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7]   File "/usr/lib/python2.7/site-packages/nova/image/glance.py", line 231, in call

2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7]     result = getattr(client.images, method)(*args, **kwargs)

2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7]   File "/usr/lib/python2.7/site-packages/glanceclient/v1/images.py", line 148, in data

2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7]     % urlparse.quote(str(image_id)))

2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7]   File "/usr/lib/python2.7/site-packages/glanceclient/common/http.py", line 280, in get

2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7]     return self._request('GET', url, **kwargs)

2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7]   File "/usr/lib/python2.7/site-packages/glanceclient/common/http.py", line 272, in _request

2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7]     resp, body_iter = self._handle_response(resp)

2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7]   File "/usr/lib/python2.7/site-packages/glanceclient/common/http.py", line 93, in _handle_response

2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7]     raise exc.from_response(resp, resp.content)

2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] Invalid: 400 Bad Request: Unknown scheme 'file' found in URI (HTTP 400)

2016-07-08 11:25:19.126 86007 ERROR nova.compute.manager [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] 

2016-07-08 11:25:19.575 86007 INFO nova.compute.manager [req-3173f5b7-fa02-420c-954b-e21c3ce8d183 289598890db341f4af45ce5c57c41ba3 713114f3b9e54501a35a79e84c1e6c9d - - -] [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] Terminating instance

2016-07-08 11:25:19.583 86007 INFO nova.virt.libvirt.driver [-] [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] During wait destroy, instance disappeared.

2016-07-08 11:25:19.665 86007 INFO nova.virt.libvirt.driver [req-3173f5b7-fa02-420c-954b-e21c3ce8d183 289598890db341f4af45ce5c57c41ba3 713114f3b9e54501a35a79e84c1e6c9d - - -] [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] Deleting instance files /var/lib/nova/instances/bf4839c8-2af6-4959-9158-fe411e1cfae7_del

2016-07-08 11:25:19.666 86007 INFO nova.virt.libvirt.driver [req-3173f5b7-fa02-420c-954b-e21c3ce8d183 289598890db341f4af45ce5c57c41ba3 713114f3b9e54501a35a79e84c1e6c9d - - -] [instance: bf4839c8-2af6-4959-9158-fe411e1cfae7] Deletion of /var/lib/nova/instances/bf4839c8-2af6-4959-9158-fe411e1cfae7_del complete

2016-07-08 11:25:26.073 86007 INFO nova.compute.resource_tracker [req-4b7eccc8-0bf5-4f55-a941-4c93e97ef5df - - - - -] Auditing locally available compute resources for node controller

2016-07-08 11:25:26.477 86007 INFO nova.compute.resource_tracker [req-4b7eccc8-0bf5-4f55-a941-4c93e97ef5df - - - - -] Total usable vcpus: 40, total allocated vcpus: 0

2016-07-08 11:25:26.478 86007 INFO nova.compute.resource_tracker [req-4b7eccc8-0bf5-4f55-a941-4c93e97ef5df - - - - -] Final resource view: name=controller phys_ram=193168MB used_ram=1024MB phys_disk=8168GB used_disk=1GB total_vcpus=40 used_vcpus=0 pci_stats=None



Regards
Gaurav Goyal

On Fri, Jul 8, 2016 at 10:17 AM, Kees Meijs <kees@xxxxxxxx> wrote:
Hi,

I'd recommend generating an UUID and use it for all your compute nodes.
This way, you can keep your configuration in libvirt constant.

Regards,
Kees

On 08-07-16 16:15, Gaurav Goyal wrote:
>
> For below section, should i generate separate UUID for both compte hosts?
>

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux