Memory leak in rebalance

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hello,

we are using glusterfs 3.2.5 and have a distributed volume with over 10M 
directories. We recently added new node and initiated rebalance. After 
several days glusterfsd consumed all the memory and was killed by the 
kernel. At that stage it was still doing the layout rebalance and got 
over 9M directories.

Rebalance reports failed status:
# gluster volume rebalance vol status
rebalance failed

In the glusterd log however I found:
  I [glusterd-rebalance.c:473:glusterd_defrag_start] 0-rebalance: 
rebalance on /etc/glusterd/mount/vol complete

I stumbled upon the patch on gluster-devel which mentions memory leak 
related rebalance: 
http://dev.gluster.com/pipermail/glusterfs/2011-June/003369.html. But 
can't figure out if this was included in the 3.2.5 release.

Another question is - is it safe to initiate rebalance of the data?

Thanks in advance
-- 
Emir Imamagic
Sektor za racunalne sustave
Sveuciliste u Zagrebu, Sveucilisni racunski centar (Srce), www.srce.unizg.hr
Emir.Imamagic at srce.hr, tel: +385 1 616 5809, fax: +385 1 616 5559


[Index of Archives]     [Gluster Development]     [Linux Filesytems Development]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux