Re: Ceph Volume Issue

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,

please share some details about your cluster (especially the hardware)

- how many OSDs are there? How many disks per an OSD machine?
- Do you use dedicated (SSD) OSD journals?
- RAM size, CPUs model, network card bandwidth/model
- Do you have a dedicated cluster network?
- How many VMs (in the whole cluster) are running?
- The total number (in the whole cluster) of attached rbd images
- The total number (in the whole cluster) of concurrent writers

Also,
- does the same problem occur when multiple threads/processes are writing into a single rbd image?
- are there anything "interesting" in qemu, kernel (both on hypervisors and OSDs), OSDs logs?

Best regards,
      Alexey


On Wed, Nov 16, 2016 at 9:10 AM, <Mehul1.Jani@xxxxxxx> wrote:

Hi All,

 

We have a Ceph Storage Cluster and it’s been integrated with our Openstack private cloud.

We have created a Pool for Volume which allows our Openstack Private Cloud user to create a volume from image and boot from volume.

Additionally our images(both Ubuntu1404 and CentOS 7) are in a raw format.

 

One of our use cases is to attach multiple volumes other than “boot volume”.

We have observed that when we attach multiple volumes, and try to simultaneous writes to these attached volumes for example via the “dd command” , all these processes go into “D state (uninterruptible sleep)”.

Also we can see in vmstat output that “bo” values trickling down to zero.

We have checked the network utilization on the compute node which does not show any issues.

 

Finally after a while system becomes unresponsive and only way to resolve is to reboot the VM.

 

Some of our version details are as follows.

 

Ceph version : 0.80.7

Libvirt version : 1.2.2

Openstack Version : Juno (Mirantis 6.0)

 

Please do let me know if anyone has faced a similar issue or have any pointers.

 

Any direction will be helpful.

 

Thanks,

Mehul

 

 


"Confidentiality Warning: This message and any attachments are intended only for the use of the intended recipient(s), are confidential and may be privileged. If you are not the intended recipient, you are hereby notified that any review, re-transmission, conversion to hard copy, copying, circulation or other use of this message and any attachments is strictly prohibited. If you are not the intended recipient, please notify the sender immediately by return email and delete this message and any attachments from your system.

Virus Warning: Although the company has taken reasonable precautions to ensure no viruses are present in this email. The company cannot accept responsibility for any loss or damage arising from the use of this email or attachment."


_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux