When you say client, we're actually doing everything through Openstack
vms and cinder block devices.
librbd and librados are:
/usr/lib/librbd.so.1.0.0
/usr/lib/librados.so.2
But I think this problem may have been related to a disk going back. We
got Disk I/O errors over the weekend and are replacing a disk, and I
think the blocked requests may have all been associated with PGs that
included the bad OSD/disk.
Would this make sense?
Jon
On 9/15/2016 3:49 AM, Wido den Hollander wrote:
Op 13 september 2016 om 18:54 schreef "WRIGHT, JON R (JON R)" <jonrodwright@xxxxxxxxx>:
VM Client OS: ubuntu 14.04
Openstack: kilo
libvirt: 1.2.12
nova-compute-kvm: 1:2015.1.4-0ubuntu2
What librados/librbd version are you running on the client?
Wido
Jon
On 9/13/2016 11:17 AM, Wido den Hollander wrote:
Op 13 september 2016 om 15:58 schreef "WRIGHT, JON R (JON R)" <jonrodwright@xxxxxxxxx>:
Yes, I do have old clients running. The clients are all vms. Is it
typical that vm clients have to be rebuilt after a ceph upgrade?
No, not always, but it is just that I saw this happening recently after a Jewel upgrade.
What version are the client(s) still running?
Wido
Thanks,
Jon
On 9/12/2016 4:05 PM, Wido den Hollander wrote:
Op 12 september 2016 om 18:47 schreef "WRIGHT, JON R (JON R)" <jonrodwright@xxxxxxxxx>:
Since upgrading to Jewel from Hammer, we're started to see HEALTH_WARN
because of 'blocked requests > 32 sec'. Seems to be related to writes.
Has anyone else seen this? Or can anyone suggest what the problem might be?
Do you by any chance have old clients connecting? I saw this after a Jewel upgrade as well and it was because of very old clients still connecting to the cluster.
Wido
Thanks!
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com