Re: KVM problems when rebalance occurs

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



-----BEGIN PGP SIGNED MESSAGE-----
Hash: SHA256

There has been a lot of "discussion" about osd_backfill_scan[min,max]
lately. My experience with hammer has been opposite that of what
people have said before. Increasing those values for us has reduced
the load of recovery and has prevented a lot of the disruption seen in
our cluster caused by backfilling. It does increase the amount of time
to do the recovery (a new node added to the cluster took about 3-4
hours before, now takes about 24 hours).

We are currently using these values and seem to work well for us.
osd_max_backfills = 1
osd_backfill_scan_min = 16
osd_recovery_max_active = 1
osd_backfill_scan_max = 32

I would be interested in your results if you try these values.
-----BEGIN PGP SIGNATURE-----
Version: Mailvelope v1.3.2
Comment: https://www.mailvelope.com

wsFcBAEBCAAQBQJWjUu/CRDmVDuy+mK58QAArdMQAI+0Er/sdN7TF7knGey2
5wJ6Ie81KJlrt/X9fIMpFdwkU2g5ET+sdU9R2hK4XcBpkonfGvwS8Ctha5Aq
XOJPrN4bMMeDK9Z4angK86ioLJevTH7tzp3FZL0U4Kbt1s9ZpwF6t+wlvkKl
mt6Tkj4VKr0917TuXqk58AYiZTYcEjGAb0QUe/gC24yFwZYrPO0vUVb4gmTQ
klNKAdTinGSn4Ynj+lBsEstWGVlTJiL3FA6xRBTz1BSjb4vtb2SoIFwHlAp+
GO+bKSh19YIasXCZfRqC/J2XcNauOIVfb4l4viV23JN2fYavEnLCnJSglYjF
Rjxr0wK+6NhRl7naJ1yGNtdMkw+h+nu/xsbYhNqT0EVq1d0nhgzh6ZjAhW1w
oRiHYA4KNn2uWiUgigpISFi4hJSP4CEPToO8jbhXhARs0H6v33oWrR8RYKxO
dFz+Lxx969rpDkk+1nRks9hTeIF+oFnW7eezSiR6TILYxvCZQ0ThHXQsL4ph
bvUr0FQmdV3ukC+Xwa/cePIlVY6JsIQfOlqmrtG7caTZWLvLUDwrwcleb272
243GXlbWCxoI7+StJDHPnY2k7NHLvbN2yG3f5PZvZaBgqqyAP8Fnq6CDtTIE
vZ/p+ZcuRw8lqoDgjjdiFyMmhQnFcCtDo3vtIy/UXDw23AVsI5edUyyP/sHt
ruPt
=X7SH
-----END PGP SIGNATURE-----
----------------
Robert LeBlanc
PGP Fingerprint 79A2 9CA4 6CC4 45DD A904  C70E E654 3BB2 FA62 B9F1


On Wed, Jan 6, 2016 at 7:13 AM, nick <nick@xxxxxxx> wrote:
> Heya,
> we are using a ceph cluster (6 Nodes with each having 10x4TB HDD + 2x SSD (for
> journal)) in combination with KVM virtualization. All our virtual machine hard
> disks are stored on the ceph cluster. The ceph cluster was updated to the
> 'infernalis' release recently.
>
> We are experiencing problems during cluster maintenance. A normal workflow for
> us looks like this:
>
> - set the noout flag for the cluster
> - stop all OSDs on one node
> - update the node
> - reboot the node
> - start all OSDs
> - wait for the backfilling to finish
> - unset the noout flag
>
> After we start all OSDs on the node again the cluster backfills and tries to
> get all the OSDs in sync. During the beginning of this process we experience
> 'stalls' in our running virtual machines. On some the load raises to a very
> high value. On others a running webserver responses only with 5xx HTTP codes.
> It takes around 5-6 minutes until all is ok again. After those 5-6 minutes the
> cluster is still backfilling, but the virtual machines behave normal again.
>
> I already set the following parameters in ceph.conf on the nodes to have a
> better rebalance traffic/user traffic ratio:
>
> """
> [osd]
> osd max backfills = 1
> osd backfill scan max = 8
> osd backfill scan min = 4
> osd recovery max active = 1
> osd recovery op priority = 1
> osd op threads = 8
> """
>
> It helped a bit, but we are still experiencing the above written problems. It
> feels like that for a short time some virtual hard disks are locked. Our ceph
> nodes are using bonded 10G network interfaces for the 'OSD network', so I do
> not think that network is a bottleneck.
>
> After reading this blog post:
> http://dachary.org/?p=2182
> I wonder if there is really a 'read lock' during the object push.
>
> Does anyone know more about this or do others have the same problems and were
> able to fix it?
>
> Best Regards
> Nick
>
> --
> Sebastian Nickel
> Nine Internet Solutions AG, Albisriederstr. 243a, CH-8047 Zuerich
> Tel +41 44 637 40 00 | Support +41 44 637 40 40 | www.nine.ch
> _______________________________________________
> ceph-users mailing list
> ceph-users@xxxxxxxxxxxxxx
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux