Re: Messup with peer status!!

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



[1970-01-01 00:02:05.860202] D [MSGID: 0]
[store.c:501:gf_store_iter_new] 0-: Returning with 0
[1970-01-01 00:02:05.860518] D [MSGID: 0]
[store.c:613:gf_store_iter_get_next] 0-: Returning with 0
[1970-01-01 00:02:05.860545] D [MSGID: 0]
[glusterd-store.c:2567:glusterd_store_update_volinfo] 0-management: key
= type value = 2
[1970-01-01 00:02:05.860583] D [MSGID: 0]
[store.c:613:gf_store_iter_get_next] 0-: Returning with 0
[1970-01-01 00:02:05.860609] D [MSGID: 0]
[glusterd-store.c:2567:glusterd_store_update_volinfo] 0-management: key
= count value = 2
[1970-01-01 00:02:05.860650] D [MSGID: 0]
[store.c:613:gf_store_iter_get_next] 0-: Returning with 0
[1970-01-01 00:02:05.860676] D [MSGID: 0]
[glusterd-store.c:2567:glusterd_store_update_volinfo] 0-management: key
= status value = 1
[1970-01-01 00:02:05.860717] D [MSGID: 0]
[store.c:613:gf_store_iter_get_next] 0-: Returning with 0
[1970-01-01 00:02:05.860743] D [MSGID: 0]
[glusterd-store.c:2567:glusterd_store_update_volinfo] 0-management: key
= sub_count value = 2
[1970-01-01 00:02:05.860780] D [MSGID: 0]
[store.c:613:gf_store_iter_get_next] 0-: Returning with 0
[1970-01-01 00:02:05.860806] D [MSGID: 0]
[glusterd-store.c:2567:glusterd_store_update_volinfo] 0-management: key
= stripe_count value = 1
[1970-01-01 00:02:05.860842] D [MSGID: 0]
[store.c:613:gf_store_iter_get_next] 0-: Returning with 0
[1970-01-01 00:02:05.860868] D [MSGID: 0]
[glusterd-store.c:2567:glusterd_store_update_volinfo] 0-management: key
= replica_count value = 2
[1970-01-01 00:02:05.860905] D [MSGID: 0]
[store.c:613:gf_store_iter_get_next] 0-: Returning with 0
[1970-01-01 00:02:05.860931] D [MSGID: 0]
[glusterd-store.c:2567:glusterd_store_update_volinfo] 0-management: key
= disperse_count value = 0
[1970-01-01 00:02:05.860967] D [MSGID: 0]
[store.c:613:gf_store_iter_get_next] 0-: Returning with 0
[1970-01-01 00:02:05.860994] D [MSGID: 0]
[glusterd-store.c:2567:glusterd_store_update_volinfo] 0-management: key
= redundancy_count value = 0
[1970-01-01 00:02:05.861030] D [MSGID: 0]
[store.c:613:gf_store_iter_get_next] 0-: Returning with 0
[1970-01-01 00:02:05.861056] D [MSGID: 0]
[glusterd-store.c:2567:glusterd_store_update_volinfo] 0-management: key
= version value = 42
[1970-01-01 00:02:05.861093] D [MSGID: 0]
[store.c:613:gf_store_iter_get_next] 0-: Returning with 0
[1970-01-01 00:02:05.861118] D [MSGID: 0]
[glusterd-store.c:2567:glusterd_store_update_volinfo] 0-management: key
= transport-type value = 0
[1970-01-01 00:02:05.861155] D [MSGID: 0]
[store.c:613:gf_store_iter_get_next] 0-: Returning with 0
[1970-01-01 00:02:05.861182] D [MSGID: 0]
[glusterd-store.c:2567:glusterd_store_update_volinfo] 0-management: key
= volume-id value = d86e215c-1710-4b33-8076-fbf8e075d3e7
[1970-01-01 00:02:05.861290] D [MSGID: 0]
[store.c:613:gf_store_iter_get_next] 0-: Returning with 0
[1970-01-01 00:02:05.861317] D [MSGID: 0]
[glusterd-store.c:2567:glusterd_store_update_volinfo] 0-management: key
= username value = db1d21cb-3feb-41da-88d0-2fc7a34cdb3a
[1970-01-01 00:02:05.861361] D [MSGID: 0]
[store.c:613:gf_store_iter_get_next] 0-: Returning with 0
[1970-01-01 00:02:05.861387] D [MSGID: 0]
[glusterd-store.c:2567:glusterd_store_update_volinfo] 0-management: key
= password value = df5bf0b7-34dd-4f0d-a01b-62d2b67aa8b0
[1970-01-01 00:02:05.861426] D [MSGID: 0]
[store.c:613:gf_store_iter_get_next] 0-: Returning with 0
[1970-01-01 00:02:05.861455] D [MSGID: 0]
[glusterd-store.c:2567:glusterd_store_update_volinfo] 0-management: key
= op-version value = 3
[1970-01-01 00:02:05.861503] D [MSGID: 0]
[store.c:613:gf_store_iter_get_next] 0-: Returning with 0
[1970-01-01 00:02:05.861530] D [MSGID: 0]
[glusterd-store.c:2567:glusterd_store_update_volinfo] 0-management: key
= client-op-version value = 3
[1970-01-01 00:02:05.861568] D [MSGID: 0]
[store.c:613:gf_store_iter_get_next] 0-: Returning with 0
[1970-01-01 00:02:05.861594] D [MSGID: 0]
[glusterd-store.c:2567:glusterd_store_update_volinfo] 0-management: key
= quota-version value = 0
[1970-01-01 00:02:05.861632] D [MSGID: 0]
[store.c:613:gf_store_iter_get_next] 0-: Returning with 0
[1970-01-01 00:02:05.861658] D [MSGID: 0]
[glusterd-store.c:2567:glusterd_store_update_volinfo] 0-management: key
= parent_volname value = N/A
[1970-01-01 00:02:05.861696] D [MSGID: 0]
[store.c:613:gf_store_iter_get_next] 0-: Returning with 0
[1970-01-01 00:02:05.861722] D [MSGID: 0]
[glusterd-store.c:2567:glusterd_store_update_volinfo] 0-management: key
= restored_from_snap value = 00000000-0000-0000-0000-000000000000
[1970-01-01 00:02:05.861762] D [MSGID: 0]
[store.c:613:gf_store_iter_get_next] 0-: Returning with 0
[1970-01-01 00:02:05.861788] D [MSGID: 0]
[glusterd-store.c:2567:glusterd_store_update_volinfo] 0-management: key
= snap-max-hard-limit value = 256
[1970-01-01 00:02:05.861825] D [MSGID: 0]
[store.c:613:gf_store_iter_get_next] 0-: Returning with 0
[1970-01-01 00:02:05.861851] D [MSGID: 0]
[glusterd-store.c:2567:glusterd_store_update_volinfo] 0-management: key
= nfs.disable value = on
[1970-01-01 00:02:05.861940] D [MSGID: 0]
[glusterd-store.c:2725:glusterd_store_update_volinfo] 0-management:
Parsed as Volume-set:key=nfs.disable,value:on
[1970-01-01 00:02:05.861978] D [MSGID: 0]
[store.c:613:gf_store_iter_get_next] 0-: Returning with 0
[1970-01-01 00:02:05.862004] D [MSGID: 0]
[glusterd-store.c:2567:glusterd_store_update_volinfo] 0-management: key
= network.ping-timeout value = 4
[1970-01-01 00:02:05.862039] D [MSGID: 0]
[glusterd-store.c:2725:glusterd_store_update_volinfo] 0-management:
Parsed as Volume-set:key=network.ping-timeout,value:4
[1970-01-01 00:02:05.862077] D [MSGID: 0]
[store.c:613:gf_store_iter_get_next] 0-: Returning with 0
[1970-01-01 00:02:05.862104] D [MSGID: 0]
[glusterd-store.c:2567:glusterd_store_update_volinfo] 0-management: key
= performance.readdir-ahead value = on
[1970-01-01 00:02:05.862140] D [MSGID: 0]
[glusterd-store.c:2725:glusterd_store_update_volinfo] 0-management:
Parsed as Volume-set:key=performance.readdir-ahead,value:on
[1970-01-01 00:02:05.862178] D [MSGID: 0]
[store.c:613:gf_store_iter_get_next] 0-: Returning with 0
[1970-01-01 00:02:05.862217] D [MSGID: 0]
[glusterd-store.c:2567:glusterd_store_update_volinfo] 0-management: key
= brick-0 value = 10.32.0.48:-opt-lvmdir-c2-brick
[1970-01-01 00:02:05.862257] D [MSGID: 0]
[store.c:613:gf_store_iter_get_next] 0-: Returning with 0
[1970-01-01 00:02:05.862283] D [MSGID: 0]
[glusterd-store.c:2567:glusterd_store_update_volinfo] 0-management: key
= brick-1 value = 10.32.1.144:-opt-lvmdir-c2-brick

On 03/16/2016 11:04 AM, ABHISHEK PALIWAL wrote:
> Hi Atin,
> 
> Please tell me the line number where you areseeing that glusterd has
> restored value from the disk files in Board B file.
> 
> Regards,
> Abhishek
> 
> On Tue, Mar 15, 2016 at 11:31 AM, ABHISHEK PALIWAL
> <abhishpaliwal@xxxxxxxxx <mailto:abhishpaliwal@xxxxxxxxx>> wrote:
> 
> 
> 
>     On Tue, Mar 15, 2016 at 11:10 AM, Atin Mukherjee
>     <amukherj@xxxxxxxxxx <mailto:amukherj@xxxxxxxxxx>> wrote:
> 
> 
> 
>         On 03/15/2016 10:54 AM, ABHISHEK PALIWAL wrote:
>         > Hi Atin,
>         >
>         > Is these files are ok? or you need some other files.
>         I just started going through the log files you shared. I've few
>         questions for you looking at the log:
>         1. Are you sure the log what you have provided from board B is
>         post a
>         reboot? If you claim that a reboot wipes of /var/lib/glusterd/
>         then why
>         am I seeing that glusterd has restored value from the disk files?
> 
> 
>     Yes these logs from Board B after reboot. Could you please explain
>     me the line number where you are seeing that glusterd has restored
>     value from the disk files.
> 
> 
>         2. From the content of glusterd configurations which you shared
>         earlier
>         the peer UUIDs are 4bf982c0-b21b-415c-b870-e72f36c7f2e7,
>         4bf982c0-b21b-415c-b870-e72f36c7f2e7 002500/glusterd/peers &
>         c6b64e36-76da-4e98-a616-48e0e52c7006 from 000300/glusterd/peers.
>         They
>         don't even exist in glusterd.log.
> 
>         Somehow I have a feeling that the sequence of log and configurations
>         files you shared don't match!
> 
> 
>     There is two UUID file present in 002500/glusterd/peers
>     1. 4bf982c0-b21b-415c-b870-e72f36c7f2e7
>     Content of this file is:
>     uuid=4bf982c0-b21b-415c-b870-e72f36c7f2e7
>     state=10
>     hostname1=10.32.0.48
>     I have a question from where this UUID is coming?
> 
>     2. 98a28041-f853-48ac-bee0-34c592eeb827
>     Content of this file is:
>     uuid=f4ebe3c5-b6a4-4795-98e0-732337f76faf //This uuid is belogs to
>     000300(10.32.0.48) board you can check this in both of the glusterd
>     log file
>     state=4 //what this state field display in this file?
>     hostname1=10.32.0.48
> 
> 
>     There is only one UUID file is present on 00030/glusterd/peers
> 
>     c6b64e36-76da-4e98-a616-48e0e52c7006 //This is the old UUID of the
>     002500 board before reboot
> 
>     content of this file is:
> 
>     uuid=267a92c3-fd28-4811-903c-c1d54854bda9 //This is new UUID
>     generated by the 002500 board after reboot you can check this as
>     well in glusterd file of 00030 board.
>     state=3
>     hostname1=10.32.1.144
> 
> 
>         ~Atin
> 
>         >
>         > Regards,
>         > Abhishek
>         >
>         > On Mon, Mar 14, 2016 at 6:12 PM, ABHISHEK PALIWAL
>         > <abhishpaliwal@xxxxxxxxx <mailto:abhishpaliwal@xxxxxxxxx>
>         <mailto:abhishpaliwal@xxxxxxxxx
>         <mailto:abhishpaliwal@xxxxxxxxx>>> wrote:
>         >
>         >     You mean etc*-glusterd-*.log file from both of the boards?
>         >
>         >     if yes please find the attachment for the same.
>         >
>         >     On Mon, Mar 14, 2016 at 5:27 PM, Atin Mukherjee <amukherj@xxxxxxxxxx <mailto:amukherj@xxxxxxxxxx>
>         >     <mailto:amukherj@xxxxxxxxxx <mailto:amukherj@xxxxxxxxxx>>> wrote:
>         >
>         >
>         >
>         >         On 03/14/2016 05:09 PM, ABHISHEK PALIWAL wrote:
>         >         > I am not getting you which glusterd directory you are asking. if you are
>         >         > asking about the /var/lib/glusterd directory then which I shared earlier
>         >         > is the same.
>         >         1. Go to /var/log/glusterfs directory
>         >         2. Look for glusterd log file
>         >         3. attach the log
>         >         Do it for both the boards.
>         >         >
>         >         > I have two directories related to gluster
>         >         >
>         >         > 1. /var/log/glusterfs
>         >         > 2./var/lib/glusterd
>         >         >
>         >         > On Mon, Mar 14, 2016 at 4:12 PM, Atin Mukherjee <amukherj@xxxxxxxxxx <mailto:amukherj@xxxxxxxxxx>
>         <mailto:amukherj@xxxxxxxxxx <mailto:amukherj@xxxxxxxxxx>>
>         >         > <mailto:amukherj@xxxxxxxxxx <mailto:amukherj@xxxxxxxxxx>
>         <mailto:amukherj@xxxxxxxxxx <mailto:amukherj@xxxxxxxxxx>>>> wrote:
>         >         >
>         >         >
>         >         >
>         >         >     On 03/14/2016 03:59 PM, ABHISHEK PALIWAL wrote:
>         >         >     > I have only these glusterd files available on the nodes
>         >         >     Look for etc-*-glusterd*.log in /var/log/glusterfs, that represents the
>         >         >     glusterd log file.
>         >         >     >
>         >         >     > Regards,
>         >         >     > Abhishek
>         >         >     >
>         >         >     > On Mon, Mar 14, 2016 at 3:43 PM, Atin Mukherjee <amukherj@xxxxxxxxxx <mailto:amukherj@xxxxxxxxxx>
>         <mailto:amukherj@xxxxxxxxxx <mailto:amukherj@xxxxxxxxxx>>
>         >         <mailto:amukherj@xxxxxxxxxx <mailto:amukherj@xxxxxxxxxx>
>         <mailto:amukherj@xxxxxxxxxx <mailto:amukherj@xxxxxxxxxx>>>
>         >         >     > <mailto:amukherj@xxxxxxxxxx
>         <mailto:amukherj@xxxxxxxxxx> <mailto:amukherj@xxxxxxxxxx
>         <mailto:amukherj@xxxxxxxxxx>>
>         >         <mailto:amukherj@xxxxxxxxxx
>         <mailto:amukherj@xxxxxxxxxx> <mailto:amukherj@xxxxxxxxxx
>         <mailto:amukherj@xxxxxxxxxx>>>>> wrote:
>         >         >     >
>         >         >     >
>         >         >     >
>         >         >     >     On 03/14/2016 02:18 PM, ABHISHEK PALIWAL
>         wrote:
>         >         >     >     >
>         >         >     >     >
>         >         >     >     > On Mon, Mar 14, 2016 at 12:12 PM, Atin
>         Mukherjee
>         >         <amukherj@xxxxxxxxxx <mailto:amukherj@xxxxxxxxxx>
>         <mailto:amukherj@xxxxxxxxxx <mailto:amukherj@xxxxxxxxxx>>
>         >         <mailto:amukherj@xxxxxxxxxx
>         <mailto:amukherj@xxxxxxxxxx> <mailto:amukherj@xxxxxxxxxx
>         <mailto:amukherj@xxxxxxxxxx>>>
>         >         >     <mailto:amukherj@xxxxxxxxxx
>         <mailto:amukherj@xxxxxxxxxx> <mailto:amukherj@xxxxxxxxxx
>         <mailto:amukherj@xxxxxxxxxx>>
>         >         <mailto:amukherj@xxxxxxxxxx
>         <mailto:amukherj@xxxxxxxxxx> <mailto:amukherj@xxxxxxxxxx
>         <mailto:amukherj@xxxxxxxxxx>>>>
>         >         >     >     > <mailto:amukherj@xxxxxxxxxx
>         <mailto:amukherj@xxxxxxxxxx>
>         >         <mailto:amukherj@xxxxxxxxxx
>         <mailto:amukherj@xxxxxxxxxx>> <mailto:amukherj@xxxxxxxxxx
>         <mailto:amukherj@xxxxxxxxxx>
>         >         <mailto:amukherj@xxxxxxxxxx <mailto:amukherj@xxxxxxxxxx>>>
>         >         >     <mailto:amukherj@xxxxxxxxxx
>         <mailto:amukherj@xxxxxxxxxx> <mailto:amukherj@xxxxxxxxxx
>         <mailto:amukherj@xxxxxxxxxx>>
>         >         <mailto:amukherj@xxxxxxxxxx
>         <mailto:amukherj@xxxxxxxxxx> <mailto:amukherj@xxxxxxxxxx
>         <mailto:amukherj@xxxxxxxxxx>>>>>> wrote:
>         >         >     >     >
>         >         >     >     >
>         >         >     >     >
>         >         >     >     >     On 03/14/2016 10:52 AM, ABHISHEK
>         PALIWAL wrote:
>         >         >     >     >     > Hi Team,
>         >         >     >     >     >
>         >         >     >     >     > I am facing some issue with peer
>         status and
>         >         because of
>         >         >     that
>         >         >     >     remove-brick
>         >         >     >     >     > on replica volume is getting failed.
>         >         >     >     >     >
>         >         >     >     >     > Here. is the scenario what I am
>         doing with
>         >         gluster:
>         >         >     >     >     >
>         >         >     >     >     > 1. I have two boards A & B and
>         gluster is
>         >         running on
>         >         >     both of
>         >         >     >     the boards.
>         >         >     >     >     > 2. On  board I have created a
>         replicated
>         >         volume with one
>         >         >     >     brick on each
>         >         >     >     >     > board.
>         >         >     >     >     > 3. Created one glusterfs mount
>         point where
>         >         both of
>         >         >     brick are
>         >         >     >     mounted.
>         >         >     >     >     > 4. start the volume with
>         nfs.disable=true.
>         >         >     >     >     > 5. Till now everything is in sync
>         between
>         >         both of bricks.
>         >         >     >     >     >
>         >         >     >     >     > Now when I manually plug-out the
>         board B
>         >         from the slot and
>         >         >     >     plug-in it again.
>         >         >     >     >     >
>         >         >     >     >     > 1. After bootup the board B I have
>         started
>         >         the glusted on
>         >         >     >     the board B.
>         >         >     >     >     >
>         >         >     >     >     > Following are the some gluster command
>         >         output on Board B
>         >         >     >     after the step 1.
>         >         >     >     >     >
>         >         >     >     >     > # gluster peer status
>         >         >     >     >     > Number of Peers: 2
>         >         >     >     >     >
>         >         >     >     >     > Hostname: 10.32.0.48
>         >         >     >     >     > Uuid:
>         f4ebe3c5-b6a4-4795-98e0-732337f76faf
>         >         >     >     >     > State: Accepted peer request
>         (Connected)
>         >         >     >     >     >
>         >         >     >     >     > Hostname: 10.32.0.48
>         >         >     >     >     > Uuid:
>         4bf982c0-b21b-415c-b870-e72f36c7f2e7
>         >         >     >     >     > State: Peer is connected and Accepted
>         >         (Connected)
>         >         >     >     >     >
>         >         >     >     >     > Why this peer status is showing
>         two peer with
>         >         >     different UUID?
>         >         >     >     >     GlusterD doesn't generate a new UUID
>         on init
>         >         if it has
>         >         >     already
>         >         >     >     generated
>         >         >     >     >     an UUID earlier. This clearly
>         indicates that
>         >         on reboot
>         >         >     of board B
>         >         >     >     >     content of /var/lib/glusterd were
>         wiped off.
>         >         I've asked this
>         >         >     >     question to
>         >         >     >     >     you multiple times that is it the case?
>         >         >     >     >
>         >         >     >     >
>         >         >     >     > Yes I am following the same which is
>         mentioned in
>         >         the link:
>         >         >     >     >
>         >         >     >     >
>         >         >     >
>         >         >
>         >       
>          http://www.gluster.org/community/documentation/index.php/Resolving_Peer_Rejected
>         >         >     >     >
>         >         >     >     > but why it is showing two peer enteries?
>         >         >     >     >
>         >         >     >     >     >
>         >         >     >     >     > # gluster volume info
>         >         >     >     >     >
>         >         >     >     >     > Volume Name: c_glusterfs
>         >         >     >     >     > Type: Replicate
>         >         >     >     >     > Volume ID:
>         c11f1f13-64a0-4aca-98b5-91d609a4a18d
>         >         >     >     >     > Status: Started
>         >         >     >     >     > Number of Bricks: 1 x 2 = 2
>         >         >     >     >     > Transport-type: tcp
>         >         >     >     >     > Bricks:
>         >         >     >     >     > Brick1:
>         10.32.0.48:/opt/lvmdir/c2/brick
>         >         >     >     >     > Brick2:
>         10.32.1.144:/opt/lvmdir/c2/brick
>         >         >     >     >     > Options Reconfigured:
>         >         >     >     >     > performance.readdir-ahead: on
>         >         >     >     >     > network.ping-timeout: 4
>         >         >     >     >     > nfs.disable: on
>         >         >     >     >     > # gluster volume heal c_glusterfs info
>         >         >     >     >     > c_glusterfs: Not able to fetch
>         volfile from
>         >         glusterd
>         >         >     >     >     > Volume heal failed.
>         >         >     >     >     > # gluster volume status c_glusterfs
>         >         >     >     >     > Status of volume: c_glusterfs
>         >         >     >     >     > Gluster process
>         >          TCP Port
>         >         >     RDMA Port
>         >         >     >     >     Online
>         >         >     >     >     > Pid
>         >         >     >     >     >
>         >         >     >     >
>         >         >     >
>         >         >
>         >         
>         ------------------------------------------------------------------------------
>         >         >     >     >     >
>         >         >     >     >     > Brick 10.32.1.144:/opt/lvmdir/c2/brick
>         >         N/A       N/A
>         >         >     >         N
>         >         >     >     >     > N/A
>         >         >     >     >     > Self-heal Daemon on localhost
>         >          N/A       N/A
>         >         >     >         Y
>         >         >     >     >     > 3922
>         >         >     >     >     >
>         >         >     >     >     > Task Status of Volume c_glusterfs
>         >         >     >     >     >
>         >         >     >     >
>         >         >     >
>         >         >
>         >         
>         ------------------------------------------------------------------------------
>         >         >     >     >     >
>         >         >     >     >     > There are no active volume tasks
>         >         >     >     >     > --
>         >         >     >     >     >
>         >         >     >     >     > At the same time Board A have the
>         following
>         >         gluster
>         >         >     commands
>         >         >     >     outcome:
>         >         >     >     >     >
>         >         >     >     >     > # gluster peer status
>         >         >     >     >     > Number of Peers: 1
>         >         >     >     >     >
>         >         >     >     >     > Hostname: 10.32.1.144
>         >         >     >     >     > Uuid:
>         c6b64e36-76da-4e98-a616-48e0e52c7006
>         >         >     >     >     > State: Peer in Cluster (Connected)
>         >         >     >     >     >
>         >         >     >     >     > Why it is showing the older UUID
>         of host
>         >         10.32.1.144
>         >         >     when this
>         >         >     >     >     UUID has
>         >         >     >     >     > been changed and new UUID is
>         >         >     >     267a92c3-fd28-4811-903c-c1d54854bda9
>         >         >     >     >     >
>         >         >     >     >     >
>         >         >     >     >     > # gluster volume heal c_glusterfs info
>         >         >     >     >     > c_glusterfs: Not able to fetch
>         volfile from
>         >         glusterd
>         >         >     >     >     > Volume heal failed.
>         >         >     >     >     > # gluster volume status c_glusterfs
>         >         >     >     >     > Status of volume: c_glusterfs
>         >         >     >     >     > Gluster process
>         >          TCP Port
>         >         >     RDMA Port
>         >         >     >     >     Online
>         >         >     >     >     > Pid
>         >         >     >     >     >
>         >         >     >     >
>         >         >     >
>         >         >
>         >         
>         ------------------------------------------------------------------------------
>         >         >     >     >     >
>         >         >     >     >     > Brick 10.32.0.48:/opt/lvmdir/c2/brick
>         >          49169     0
>         >         >     >         Y
>         >         >     >     >     > 2427
>         >         >     >     >     > Brick 10.32.1.144:/opt/lvmdir/c2/brick
>         >         N/A       N/A
>         >         >     >         N
>         >         >     >     >     > N/A
>         >         >     >     >     > Self-heal Daemon on localhost
>         >          N/A       N/A
>         >         >     >         Y
>         >         >     >     >     > 3388
>         >         >     >     >     > Self-heal Daemon on 10.32.1.144
>         >          N/A       N/A
>         >         >     >         Y
>         >         >     >     >     > 3922
>         >         >     >     >     >
>         >         >     >     >     > Task Status of Volume c_glusterfs
>         >         >     >     >     >
>         >         >     >     >
>         >         >     >
>         >         >
>         >         
>         ------------------------------------------------------------------------------
>         >         >     >     >     >
>         >         >     >     >     > There are no active volume tasks
>         >         >     >     >     >
>         >         >     >     >     > As you see in the "gluster volume
>         status"
>         >         showing that
>         >         >     Brick
>         >         >     >     >     > "10.32.1.144:/opt/lvmdir/c2/brick " is
>         >         offline so We have
>         >         >     >     tried to
>         >         >     >     >     > remove it but getting "volume
>         remove-brick
>         >         c_glusterfs
>         >         >     replica 1
>         >         >     >     >     > 10.32.1.144:/opt/lvmdir/c2/brick
>         force :
>         >         FAILED :
>         >         >     Incorrect
>         >         >     >     brick
>         >         >     >     >     > 10.32.1.144:/opt/lvmdir/c2/brick
>         for volume
>         >         c_glusterfs"
>         >         >     >     error on the
>         >         >     >     >     > Board A.
>         >         >     >     >     >
>         >         >     >     >     > Please reply on this post because I am
>         >         always getting
>         >         >     this error
>         >         >     >     >     in this
>         >         >     >     >     > scenario.
>         >         >     >     >     >
>         >         >     >     >     > For more detail I am also adding
>         the logs of
>         >         both of the
>         >         >     >     board which
>         >         >     >     >     > having some manual created file in
>         which you
>         >         can find the
>         >         >     >     output of
>         >         >     >     >     > glulster command from both of the
>         boards
>         >         >     >     >     >
>         >         >     >     >     > in logs
>         >         >     >     >     > 00030 is board A
>         >         >     >     >     > 00250 is board B.
>         >         >     >     >     This attachment doesn't help much.
>         Could you
>         >         attach full
>         >         >     >     glusterd log
>         >         >     >     >     files from both the nodes?
>         >         >     >     >     >
>         >         >     >     >
>         >         >     >     > inside this attachment you will found full
>         >         glusterd log file
>         >         >     >     > 00300/glusterd/ and 002500/glusterd/
>         >         >     >     No, that contains the configuration files.
>         >         >     >     >
>         >         >     >     >     > Thanks in advance waiting for the
>         reply.
>         >         >     >     >     >
>         >         >     >     >     > Regards,
>         >         >     >     >     > Abhishek
>         >         >     >     >     >
>         >         >     >     >     >
>         >         >     >     >     > Regards
>         >         >     >     >     > Abhishek Paliwal
>         >         >     >     >     >
>         >         >     >     >     >
>         >         >     >     >     >
>         _______________________________________________
>         >         >     >     >     > Gluster-devel mailing list
>         >         >     >     >     > Gluster-devel@xxxxxxxxxxx
>         <mailto:Gluster-devel@xxxxxxxxxxx>
>         >         <mailto:Gluster-devel@xxxxxxxxxxx
>         <mailto:Gluster-devel@xxxxxxxxxxx>>
>         >         >     <mailto:Gluster-devel@xxxxxxxxxxx
>         <mailto:Gluster-devel@xxxxxxxxxxx>
>         >         <mailto:Gluster-devel@xxxxxxxxxxx
>         <mailto:Gluster-devel@xxxxxxxxxxx>>>
>         >         <mailto:Gluster-devel@xxxxxxxxxxx
>         <mailto:Gluster-devel@xxxxxxxxxxx>
>         <mailto:Gluster-devel@xxxxxxxxxxx
>         <mailto:Gluster-devel@xxxxxxxxxxx>>
>         >         >     <mailto:Gluster-devel@xxxxxxxxxxx
>         <mailto:Gluster-devel@xxxxxxxxxxx>
>         >         <mailto:Gluster-devel@xxxxxxxxxxx
>         <mailto:Gluster-devel@xxxxxxxxxxx>>>>
>         >         >     >     <mailto:Gluster-devel@xxxxxxxxxxx
>         <mailto:Gluster-devel@xxxxxxxxxxx>
>         >         <mailto:Gluster-devel@xxxxxxxxxxx
>         <mailto:Gluster-devel@xxxxxxxxxxx>>
>         >         >     <mailto:Gluster-devel@xxxxxxxxxxx
>         <mailto:Gluster-devel@xxxxxxxxxxx>
>         >         <mailto:Gluster-devel@xxxxxxxxxxx
>         <mailto:Gluster-devel@xxxxxxxxxxx>>>
>         >         <mailto:Gluster-devel@xxxxxxxxxxx
>         <mailto:Gluster-devel@xxxxxxxxxxx>
>         <mailto:Gluster-devel@xxxxxxxxxxx
>         <mailto:Gluster-devel@xxxxxxxxxxx>>
>         >         >     <mailto:Gluster-devel@xxxxxxxxxxx
>         <mailto:Gluster-devel@xxxxxxxxxxx>
>         >         <mailto:Gluster-devel@xxxxxxxxxxx
>         <mailto:Gluster-devel@xxxxxxxxxxx>>>>>
>         >         >     >     >     >
>         >         http://www.gluster.org/mailman/listinfo/gluster-devel
>         >         >     >     >     >
>         >         >     >     >
>         >         >     >     >
>         >         >     >     >
>         >         >     >     >
>         >         >     >     > --
>         >         >     >     >
>         >         >     >     >
>         >         >     >     >
>         >         >     >     >
>         >         >     >     > Regards
>         >         >     >     > Abhishek Paliwal
>         >         >     >
>         >         >     >
>         >         >     >
>         >         >     >
>         >         >     > --
>         >         >     >
>         >         >     >
>         >         >     >
>         >         >     >
>         >         >     > Regards
>         >         >     > Abhishek Paliwal
>         >         >
>         >         >
>         >         >
>         >         >
>         >         > --
>         >         >
>         >         >
>         >         >
>         >         >
>         >         > Regards
>         >         > Abhishek Paliwal
>         >
>         >
>         >
>         >
>         >     --
>         >
>         >
>         >
>         >
>         >     Regards
>         >     Abhishek Paliwal
>         >
>         >
>         >
>         >
>         > --
>         >
>         >
>         >
>         >
>         > Regards
>         > Abhishek Paliwal
> 
> 
> 
> 
>     -- 
> 
> 
> 
> 
>     Regards
>     Abhishek Paliwal
> 
> 
> 
> 
> -- 
> 
> 
> 
> 
> Regards
> Abhishek Paliwal
_______________________________________________
Gluster-devel mailing list
Gluster-devel@xxxxxxxxxxx
http://www.gluster.org/mailman/listinfo/gluster-devel



[Index of Archives]     [Gluster Users]     [Ceph Users]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Security]     [Bugtraq]     [Linux]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux