even /var/log/glusterfs directory logs are also present in 00300/glusterfs and 002500/glusterfs.
Regards,On Mon, Mar 14, 2016 at 5:09 PM, ABHISHEK PALIWAL <abhishpaliwal@xxxxxxxxx> wrote:
2./var/lib/glusterd1. /var/log/glusterfsI am not getting you which glusterd directory you are asking. if you are asking about the /var/lib/glusterd directory then which I shared earlier is the same.I have two directories related to gluster--On Mon, Mar 14, 2016 at 4:12 PM, Atin Mukherjee <amukherj@xxxxxxxxxx> wrote:
On 03/14/2016 03:59 PM, ABHISHEK PALIWAL wrote:
> I have only these glusterd files available on the nodes
Look for etc-*-glusterd*.log in /var/log/glusterfs, that represents the
glusterd log file.
>
> Regards,
> Abhishek
>
> On Mon, Mar 14, 2016 at 3:43 PM, Atin Mukherjee <amukherj@xxxxxxxxxx
> <mailto:amukherj@xxxxxxxxxx>> wrote:
>
>
>
> On 03/14/2016 02:18 PM, ABHISHEK PALIWAL wrote:
> >
> >
> > On Mon, Mar 14, 2016 at 12:12 PM, Atin Mukherjee <amukherj@xxxxxxxxxx <mailto:amukherj@xxxxxxxxxx>
> <mailto:Gluster-devel@xxxxxxxxxxx <mailto:Gluster-devel@xxxxxxxxxxx>>> > <mailto:amukherj@xxxxxxxxxx <mailto:amukherj@xxxxxxxxxx>>> wrote:
> >
> >
> >
> > On 03/14/2016 10:52 AM, ABHISHEK PALIWAL wrote:
> > > Hi Team,
> > >
> > > I am facing some issue with peer status and because of that
> remove-brick
> > > on replica volume is getting failed.
> > >
> > > Here. is the scenario what I am doing with gluster:
> > >
> > > 1. I have two boards A & B and gluster is running on both of
> the boards.
> > > 2. On board I have created a replicated volume with one
> brick on each
> > > board.
> > > 3. Created one glusterfs mount point where both of brick are
> mounted.
> > > 4. start the volume with nfs.disable=true.
> > > 5. Till now everything is in sync between both of bricks.
> > >
> > > Now when I manually plug-out the board B from the slot and
> plug-in it again.
> > >
> > > 1. After bootup the board B I have started the glusted on
> the board B.
> > >
> > > Following are the some gluster command output on Board B
> after the step 1.
> > >
> > > # gluster peer status
> > > Number of Peers: 2
> > >
> > > Hostname: 10.32.0.48
> > > Uuid: f4ebe3c5-b6a4-4795-98e0-732337f76faf
> > > State: Accepted peer request (Connected)
> > >
> > > Hostname: 10.32.0.48
> > > Uuid: 4bf982c0-b21b-415c-b870-e72f36c7f2e7
> > > State: Peer is connected and Accepted (Connected)
> > >
> > > Why this peer status is showing two peer with different UUID?
> > GlusterD doesn't generate a new UUID on init if it has already
> generated
> > an UUID earlier. This clearly indicates that on reboot of board B
> > content of /var/lib/glusterd were wiped off. I've asked this
> question to
> > you multiple times that is it the case?
> >
> >
> > Yes I am following the same which is mentioned in the link:
> >
> >
> http://www.gluster.org/community/documentation/index.php/Resolving_Peer_Rejected
> >
> > but why it is showing two peer enteries?
> >
> > >
> > > # gluster volume info
> > >
> > > Volume Name: c_glusterfs
> > > Type: Replicate
> > > Volume ID: c11f1f13-64a0-4aca-98b5-91d609a4a18d
> > > Status: Started
> > > Number of Bricks: 1 x 2 = 2
> > > Transport-type: tcp
> > > Bricks:
> > > Brick1: 10.32.0.48:/opt/lvmdir/c2/brick
> > > Brick2: 10.32.1.144:/opt/lvmdir/c2/brick
> > > Options Reconfigured:
> > > performance.readdir-ahead: on
> > > network.ping-timeout: 4
> > > nfs.disable: on
> > > # gluster volume heal c_glusterfs info
> > > c_glusterfs: Not able to fetch volfile from glusterd
> > > Volume heal failed.
> > > # gluster volume status c_glusterfs
> > > Status of volume: c_glusterfs
> > > Gluster process TCP Port RDMA Port
> > Online
> > > Pid
> > >
> >
> ------------------------------------------------------------------------------
> > >
> > > Brick 10.32.1.144:/opt/lvmdir/c2/brick N/A N/A
> N
> > > N/A
> > > Self-heal Daemon on localhost N/A N/A
> Y
> > > 3922
> > >
> > > Task Status of Volume c_glusterfs
> > >
> >
> ------------------------------------------------------------------------------
> > >
> > > There are no active volume tasks
> > > --
> > >
> > > At the same time Board A have the following gluster commands
> outcome:
> > >
> > > # gluster peer status
> > > Number of Peers: 1
> > >
> > > Hostname: 10.32.1.144
> > > Uuid: c6b64e36-76da-4e98-a616-48e0e52c7006
> > > State: Peer in Cluster (Connected)
> > >
> > > Why it is showing the older UUID of host 10.32.1.144 when this
> > UUID has
> > > been changed and new UUID is
> 267a92c3-fd28-4811-903c-c1d54854bda9
> > >
> > >
> > > # gluster volume heal c_glusterfs info
> > > c_glusterfs: Not able to fetch volfile from glusterd
> > > Volume heal failed.
> > > # gluster volume status c_glusterfs
> > > Status of volume: c_glusterfs
> > > Gluster process TCP Port RDMA Port
> > Online
> > > Pid
> > >
> >
> ------------------------------------------------------------------------------
> > >
> > > Brick 10.32.0.48:/opt/lvmdir/c2/brick 49169 0
> Y
> > > 2427
> > > Brick 10.32.1.144:/opt/lvmdir/c2/brick N/A N/A
> N
> > > N/A
> > > Self-heal Daemon on localhost N/A N/A
> Y
> > > 3388
> > > Self-heal Daemon on 10.32.1.144 N/A N/A
> Y
> > > 3922
> > >
> > > Task Status of Volume c_glusterfs
> > >
> >
> ------------------------------------------------------------------------------
> > >
> > > There are no active volume tasks
> > >
> > > As you see in the "gluster volume status" showing that Brick
> > > "10.32.1.144:/opt/lvmdir/c2/brick " is offline so We have
> tried to
> > > remove it but getting "volume remove-brick c_glusterfs replica 1
> > > 10.32.1.144:/opt/lvmdir/c2/brick force : FAILED : Incorrect
> brick
> > > 10.32.1.144:/opt/lvmdir/c2/brick for volume c_glusterfs"
> error on the
> > > Board A.
> > >
> > > Please reply on this post because I am always getting this error
> > in this
> > > scenario.
> > >
> > > For more detail I am also adding the logs of both of the
> board which
> > > having some manual created file in which you can find the
> output of
> > > glulster command from both of the boards
> > >
> > > in logs
> > > 00030 is board A
> > > 00250 is board B.
> > This attachment doesn't help much. Could you attach full
> glusterd log
> > files from both the nodes?
> > >
> >
> > inside this attachment you will found full glusterd log file
> > 00300/glusterd/ and 002500/glusterd/
> No, that contains the configuration files.
> >
> > > Thanks in advance waiting for the reply.
> > >
> > > Regards,
> > > Abhishek
> > >
> > >
> > > Regards
> > > Abhishek Paliwal
> > >
> > >
> > > _______________________________________________
> > > Gluster-devel mailing list
> > > Gluster-devel@xxxxxxxxxxx <mailto:Gluster-devel@xxxxxxxxxxx>
> > > http://www.gluster.org/mailman/listinfo/gluster-devel
> > >
> >
> >
> >
> >
> > --
> >
> >
> >
> >
> > Regards
> > Abhishek Paliwal
>
>
>
>
> --
>
>
>
>
> Regards
> Abhishek Paliwal
Regards
Abhishek Paliwal
--
Regards
Abhishek Paliwal
_______________________________________________ Gluster-devel mailing list Gluster-devel@xxxxxxxxxxx http://www.gluster.org/mailman/listinfo/gluster-devel