Self-Heal Daemon not Running

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On 09/25/2013 01:06 PM, Andrew Lau wrote:
> On Wed, Sep 25, 2013 at 2:28 PM, Ravishankar N <ravishankar at redhat.com 
> <mailto:ravishankar at redhat.com>>wrote:
>
>     On 09/25/2013 06:16 AM, Andrew Lau wrote:
>>     That's where I found the 200+ entries
>>
>>     [ root at hv01 ]gluster volume heal STORAGE info split-brain
>>     Gathering Heal info on volume STORAGE has been successful
>>
>>     Brick hv01:/data1
>>     Number of entries: 271
>>     at            path on brick
>>
>>     2013-09-25 00:04:29 /6682d31f-39ce-4896-99ef-14e1c9682585/dom_md/ids
>>     2013-09-25 00:04:29
>>     /6682d31f-39ce-4896-99ef-14e1c9682585/images/5599c7c7-0c25-459a-9d7d-80190a7c739b/0593d351-2ab1-49cd-a9b6-c94c897ebcc7
>>     2013-09-24 23:54:29 <gfid:9c83f7e4-6982-4477-816b-172e4e640566>
>>     2013-09-24 23:54:29 <gfid:91e98909-c217-417b-a3c1-4cf0f2356e14>
>>     <snip>
>>
>>     Brick hv02:/data1
>>     Number of entries: 0
>>
>>     When I run the same command on hv02, it will show the reverse
>>     (the other node having 0 entries).
>>
>>     I remember last time having to delete these files individually on
>>     another split-brain case, but I was hoping there was a better
>>     solution than going through 200+ entries.
>>
>     While I haven't tried it out myself, Jeff Darcy has written a
>     script
>     (https://github.com/jdarcy/glusterfs/tree/heal-script/extras/heal_script)
>     which helps in automating the process. He has detailed it's usage
>     in his blog post
>     http://hekafs.org/index.php/2012/06/healing-split-brain/
>
>     Hope this helps.
>     -Ravi
>
>
> That didn't end up working, ImportError: No module named volfilter
>
Oh, you need to download all 4 python scripts in the heal_script folder.
> But I didn't end up spending much time with it as the number of 
> entries magically reduced to 10, I removed the files and the 
> split-brain info reports 0 entries. Still wondering why there's 
> different file sizes on the two bricks.
>
>
>>     Cheers.
>>
>>
>>     On Wed, Sep 25, 2013 at 10:39 AM, Mohit Anchlia
>>     <mohitanchlia at gmail.com <mailto:mohitanchlia at gmail.com>> wrote:
>>
>>         What's the output of
>>         |gluster volume heal $VOLUME info ||split||-brain|
>>
>>
>>         On Tue, Sep 24, 2013 at 5:33 PM, Andrew Lau
>>         <andrew at andrewklau.com <mailto:andrew at andrewklau.com>> wrote:
>>
>>             Found the BZ
>>             https://bugzilla.redhat.com/show_bug.cgi?id=960190 - so I
>>             restarted one of the volumes and it seems to have
>>             restarted the all daemons again.
>>
>>             Self heal started again, but I seem to have split-brain
>>             issues everywhere. There's over 100 different entries on
>>             each node, what's the best way to restore this now? Short
>>             of having to manually go through and delete 200+ files.
>>             It looks like a full split brain as the file sizes on the
>>             different nodes are out of balance by about 100GB or so.
>>
>>             Any suggestions would be much appreciated!
>>
>>             Cheers.
>>
>>             On Tue, Sep 24, 2013 at 10:32 PM, Andrew Lau
>>             <andrew at andrewklau.com <mailto:andrew at andrewklau.com>> wrote:
>>
>>                 Hi,
>>
>>                 Right now, I have a 2x1 replica. Ever since I had to
>>                 reinstall one of the gluster servers, there's been
>>                 issues with split-brain. The self-heal daemon doesn't
>>                 seem to be running on either of the nodes.
>>
>>                 To reinstall the gluster server (the original brick
>>                 data was intact but the OS had to be reinstalled)
>>                 - Reinstalled gluster
>>                 - Copied over the old uuid from backup
>>                 - gluster peer probe
>>                 - gluster volume sync $othernode all
>>                 - mount -t glusterfs localhost:STORAGE /mnt
>>                 - find /mnt -noleaf -print0 | xargs --null stat
>>                 >/dev/null 2>/var/log/glusterfs/mnt-selfheal.log
>>
>>                 I let it resync and it was working fine, atleast so I
>>                 thought. I just came back a few days later to see
>>                 there's a miss match in the brick volumes. One is
>>                 50GB ahead of the other.
>>
>>                 # gluster volume heal STORAGE info
>>                 Status: self-heal-daemon is not running on
>>                 966456a1-b8a6-4ca8-9da7-d0eb96997cbe
>>
>>                 /var/log/gluster/glustershd.log doesn't seem to have
>>                 any recent logs, only those from when the two
>>                 original gluster servers were running.
>>
>>                 # gluster volume status
>>
>>                 Self-heal Daemon on localhostN/ANN/A
>>
>>                 Any suggestions would be much appreciated!
>>
>>                 Cheers
>>                 Andrew.
>>
>

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://supercolony.gluster.org/pipermail/gluster-users/attachments/20130925/b2f741cd/attachment.html>


[Index of Archives]     [Gluster Development]     [Linux Filesytems Development]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux