On 06/10/2014 02:00 PM,
gluster-users-request@xxxxxxxxxxx wrote:
You could try this:From: Laurent Chouinard <laurent.chouinard@xxxxxxxxxxx> To: Pranith Kumar Karampuri <pkarampu@xxxxxxxxxx> Cc: "gluster-users@xxxxxxxxxxx" <gluster-users@xxxxxxxxxxx> Subject: Re: Unavailability during self-heal for large volumes Message-ID: <95ea1865fac2484980d020c6a3b7f0cd@xxxxxxxxxxxxxxxxxxxxxxxxxxx> Content-Type: text/plain; charset="utf-8"Laurent, This has been improved significantly in afr-v2 (enhanced version ofreplication http://www.gluster.org/author/andrew-lau/ Controlling glusterfsd CPU outbreaks with cgroupsSome of you may that same feeling when adding a new brick to your
gluster replicated volume which already has an excess of 1TB data
already on there and suddenly your gluster server has shot up to
500% CPU usage. What's worse is when my hosts run along side oVirt
so while gluster hogged all the CPU, my VMs started to crawl, even
running simple commands like My first attempt I limited the NIC's bandwidth to 200Mbps rather than the 2x1Gbps aggregated link and this calmed glusterfsd down to a healthy 50%. A temporary fix which however meant clients accessing gluster storage would be bottlenecked by that shared limit. So off to the mailing list - a great suggestion from James/purpleidea (https://ttboj.wordpress.com/code/puppet-gluster/) on using cgroups. The concept is simple, we limit the total CPU glusterfsd sees so
when it comes to doing the checksums for self heals, replication
etc. They won't have the high priority which other services such
as running VMs would have. This effectively slows down
replication rate in return for lower CPU usage. Kind regards, Netbulae B.V. http://www.netbulae.eu |
_______________________________________________ Gluster-users mailing list Gluster-users@xxxxxxxxxxx http://supercolony.gluster.org/mailman/listinfo/gluster-users