Replacing a failed brick

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Ravi,

Thanks for the tips. When I run a volume status:
gluster> volume status test-a
Status of volume: test-a
Gluster process                                         Port    Online  Pid
------------------------------------------------------------------------------
Brick 10.250.4.63:/localmnt/g1lv2                       49152   Y       8072
Brick 10.250.4.65:/localmnt/g2lv2                       49152   Y       3403
Brick 10.250.4.63:/localmnt/g1lv3                       49153   Y       8081
Brick 10.250.4.65:/localmnt/g2lv3                       49153   Y       3410
Brick 10.250.4.63:/localmnt/g1lv4                       49154   Y       8090
Brick 10.250.4.65:/localmnt/g2lv4                       49154   Y       3417
Brick 10.250.4.63:/localmnt/g1lv5                       49155   Y       8099
Brick 10.250.4.65:/localmnt/g2lv5                       N/A     N       N/A
Brick 10.250.4.63:/localmnt/g1lv1                       49156   Y       8576
Brick 10.250.4.65:/localmnt/g2lv1                       49156   Y       3431
NFS Server on localhost                                 2049    Y       3440
Self-heal Daemon on localhost                           N/A     Y       3445
NFS Server on 10.250.4.63                               2049    Y       8586
Self-heal Daemon on 10.250.4.63                         N/A     Y       8593

There are no active volume tasks
--

Attempting to start the volume results in:
gluster> volume start test-a force
volume start: test-a: failed: Failed to get extended attribute
trusted.glusterfs.volume-id for brick dir /localmnt/g2lv5. Reason : No data
available
--

It doesn't like when I try to fire off a heal either:
gluster> volume heal test-a
Launching Heal operation on volume test-a has been unsuccessful
--

Although that did lead me to this:
gluster> volume heal test-a info
Gathering Heal info on volume test-a has been successful

Brick 10.250.4.63:/localmnt/g1lv2
Number of entries: 0

Brick 10.250.4.65:/localmnt/g2lv2
Number of entries: 0

Brick 10.250.4.63:/localmnt/g1lv3
Number of entries: 0

Brick 10.250.4.65:/localmnt/g2lv3
Number of entries: 0

Brick 10.250.4.63:/localmnt/g1lv4
Number of entries: 0

Brick 10.250.4.65:/localmnt/g2lv4
Number of entries: 0

Brick 10.250.4.63:/localmnt/g1lv5
Number of entries: 0

Brick 10.250.4.65:/localmnt/g2lv5
Status: Brick is Not connected
Number of entries: 0

Brick 10.250.4.63:/localmnt/g1lv1
Number of entries: 0

Brick 10.250.4.65:/localmnt/g2lv1
Number of entries: 0
--

So perhaps I need to re-connect the brick?

Cheers,
Dave



On Fri, Aug 16, 2013 at 12:43 AM, Ravishankar N <ravishankar at redhat.com>wrote:

>  On 08/15/2013 10:05 PM, David Gibbons wrote:
>
> Hi There,
>
>  I'm currently testing Gluster for possible production use. I haven't
> been able to find the answer to this question in the forum arch or in the
> public docs. It's possible that I don't know which keywords to search for.
>
>  Here's the question (more details below): let's say that one of my
> bricks "fails" -- *not* a whole node failure but a single brick failure
> within the node. How do I replace a single brick on a node and force a sync
> from one of the replicas?
>
>  I have two nodes with 5 bricks each:
>  gluster> volume info test-a
>
>  Volume Name: test-a
> Type: Distributed-Replicate
> Volume ID: e8957773-dd36-44ae-b80a-01e22c78a8b4
> Status: Started
> Number of Bricks: 5 x 2 = 10
> Transport-type: tcp
> Bricks:
> Brick1: 10.250.4.63:/localmnt/g1lv2
> Brick2: 10.250.4.65:/localmnt/g2lv2
> Brick3: 10.250.4.63:/localmnt/g1lv3
> Brick4: 10.250.4.65:/localmnt/g2lv3
> Brick5: 10.250.4.63:/localmnt/g1lv4
> Brick6: 10.250.4.65:/localmnt/g2lv4
> Brick7: 10.250.4.63:/localmnt/g1lv5
> Brick8: 10.250.4.65:/localmnt/g2lv5
> Brick9: 10.250.4.63:/localmnt/g1lv1
> Brick10: 10.250.4.65:/localmnt/g2lv1
>
>  I formatted 10.250.4.65:/localmnt/g2lv5 (to simulate a "failure"). What
> is the next step? I have tried various combinations of removing and
> re-adding the brick, replacing the brick, etc. I read in a previous message
> to this list that replace-brick was for planned changes which makes sense,
> so that's probably not my next step.
>
> You must first check if the 'formatted' brick 10.250.4.65:/localmnt/g2lv5
> is online using the `gluster volume status` command. If not start the
> volume using `gluster volume start <VOLNAME>force`. You can then use the
> gluster volume heal command which would copy the data from the other
> replica brick into your formatted brick.
> Hope this helps.
> -Ravi
>
>
>  Cheers,
> Dave
>
>
> _______________________________________________
> Gluster-users mailing listGluster-users at gluster.orghttp://supercolony.gluster.org/mailman/listinfo/gluster-users
>
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://supercolony.gluster.org/pipermail/gluster-users/attachments/20130816/00c31469/attachment.html>


[Index of Archives]     [Gluster Development]     [Linux Filesytems Development]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux