Memory leak in rebalance

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



I'm curious if this was ever resolved/commented on? Were you able to
finish the rebalance?

On Tue, Jan 31, 2012 at 4:37 AM, Emir Imamagic <eimamagi at srce.hr> wrote:
> Hello,
>
> we are using glusterfs 3.2.5 and have a distributed volume with over 10M
> directories. We recently added new node and initiated rebalance. After
> several days glusterfsd consumed all the memory and was killed by the
> kernel. At that stage it was still doing the layout rebalance and got over
> 9M directories.
>
> Rebalance reports failed status:
> # gluster volume rebalance vol status
> rebalance failed
>
> In the glusterd log however I found:
> ?I [glusterd-rebalance.c:473:glusterd_defrag_start] 0-rebalance: rebalance
> on /etc/glusterd/mount/vol complete
>
> I stumbled upon the patch on gluster-devel which mentions memory leak
> related rebalance:
> http://dev.gluster.com/pipermail/glusterfs/2011-June/003369.html. But can't
> figure out if this was included in the 3.2.5 release.
>
> Another question is - is it safe to initiate rebalance of the data?
>
> Thanks in advance
> --
> Emir Imamagic
> Sektor za racunalne sustave
> Sveuciliste u Zagrebu, Sveucilisni racunski centar (Srce), www.srce.unizg.hr
> Emir.Imamagic at srce.hr, tel: +385 1 616 5809, fax: +385 1 616 5559
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users


[Index of Archives]     [Gluster Development]     [Linux Filesytems Development]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux