Fwd: Unexpected behaviour during replication heal

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



----- Forwarded Message -----
From: "Darren Austin" <darren-lists at widgit.com>
To: "Mohit Anchlia" <mohitanchlia at gmail.com>
Sent: Wednesday, 29 June, 2011 11:13:30 AM
Subject: Re: Unexpected behaviour during replication heal

----- Original Message -----
> Did you recently upgrade?

I was able to reproduce this problem on both 3.2.0 and 3.2.1.
It wasn't an upgrade situation - I deleted the volumes and re-created
them for each test.

> Can you also post gluster volume info and your gluster vol files?

'gluster volume info':
Volume Name: data-volume
Type: Replicate
Status: Started
Number of Bricks: 2
Transport-type: tcp
Bricks:
Brick1: 10.234.158.226:/data
Brick2: 10.49.14.115:/data

glusterd.vol (same on both servers):
volume management
type mgmt/glusterd
option working-directory /etc/glusterd
option transport-type socket,rdma
option transport.socket.keepalive-time 10
option transport.socket.keepalive-interval 2
end-volume

Other vol files and brick info were posted with my first description of
the issue :)

HTH,
Darren.

-- Darren Austin - Systems Administrator, Widgit Software.
Tel: +44 (0)1926 333680. Web: http://www.widgit.com/
26 Queen Street, Cubbington, Warwickshire, CV32 7NA.

-- 
Darren Austin - Systems Administrator, Widgit Software.
Tel: +44 (0)1926 333680.    Web: http://www.widgit.com/
26 Queen Street, Cubbington, Warwickshire, CV32 7NA.



[Index of Archives]     [Gluster Development]     [Linux Filesytems Development]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux