Re: Previously replaced brick not coming up after reboot

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



glusterfs 3.12.12

2018-08-16 9:26 GMT+02:00 Serkan Çoban <cobanserkan@xxxxxxxxx>:
> What is your gluster version? There was a bug in 3.10, when you reboot
> a node some bricks may not come online but it fixed in later versions.
>
> On 8/16/18, Hu Bert <revirii@xxxxxxxxxxxxxx> wrote:
>> Hi there,
>>
>> 2 times i had to replace a brick on 2 different servers; replace went
>> fine, heal took very long but finally finished. From time to time you
>> have to reboot the server (kernel upgrades), and i've noticed that the
>> replaced brick doesn't come up after the reboot. Status after reboot:
>>
>> gluster volume status
>> Status of volume: shared
>> Gluster process                             TCP Port  RDMA Port  Online
>> Pid
>> ------------------------------------------------------------------------------
>> Brick gluster11:/gluster/bricksda1/shared   49164     0          Y
>> 6425
>> Brick gluster12:/gluster/bricksda1/shared   49152     0          Y
>> 2078
>> Brick gluster13:/gluster/bricksda1/shared   49152     0          Y
>> 2478
>> Brick gluster11:/gluster/bricksdb1/shared   49165     0          Y
>> 6452
>> Brick gluster12:/gluster/bricksdb1/shared   49153     0          Y
>> 2084
>> Brick gluster13:/gluster/bricksdb1/shared   49153     0          Y
>> 2497
>> Brick gluster11:/gluster/bricksdc1/shared   49166     0          Y
>> 6479
>> Brick gluster12:/gluster/bricksdc1/shared   49154     0          Y
>> 2090
>> Brick gluster13:/gluster/bricksdc1/shared   49154     0          Y
>> 2485
>> Brick gluster11:/gluster/bricksdd1/shared   49168     0          Y
>> 7897
>> Brick gluster12:/gluster/bricksdd1_new/shared  49157     0          Y
>> 7632
>> Brick gluster13:/gluster/bricksdd1_new/shared  N/A       N/A        N
>>      N/A
>> Self-heal Daemon on localhost               N/A       N/A        Y
>> 25483
>> Self-heal Daemon on gluster13               N/A       N/A        Y
>> 2463
>> Self-heal Daemon on gluster12               N/A       N/A        Y
>> 17619
>>
>> Task Status of Volume shared
>> ------------------------------------------------------------------------------
>> There are no active volume tasks
>>
>> Here gluster13:/gluster/bricksdd1_new/shared is not up. Related log
>> message after reboot in glusterd.log:
>>
>> [2018-08-16 05:22:52.986757] W [socket.c:593:__socket_rwv]
>> 0-management: readv on
>> /var/run/gluster/02d086b75bfc97f2cce96fe47e26dcf3.socket failed (No
>> data available)
>> [2018-08-16 05:22:52.987648] I [MSGID: 106005]
>> [glusterd-handler.c:6071:__glusterd_brick_rpc_notify] 0-management:
>> Brick gluster13:/gluster/bricksdd1_new/shared has disconnected from
>> glusterd.
>> [2018-08-16 05:22:52.987908] E [rpc-clnt.c:350:saved_frames_unwind]
>> (-->
>> /usr/lib/x86_64-linux-gnu/libglusterfs.so.0(_gf_log_callingfn+0x13e)[0x7fdbaa398b8e]
>> (--> /usr/lib/x86_64-
>> linux-gnu/libgfrpc.so.0(saved_frames_unwind+0x1d1)[0x7fdbaa15f111]
>> (-->
>> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(saved_frames_destroy+0xe)[0x7fdbaa15f23e]
>> (--> /usr/lib/x86_64-linu
>> x-gnu/libgfrpc.so.0(rpc_clnt_connection_cleanup+0x91)[0x7fdbaa1608d1]
>> (-->
>> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(rpc_clnt_notify+0x288)[0x7fdbaa1613f8]
>> ))))) 0-management: force
>> d unwinding frame type(brick operations) op(--(4)) called at
>> 2018-08-16 05:22:52.941332 (xid=0x2)
>> [2018-08-16 05:22:52.988058] W [dict.c:426:dict_set]
>> (-->/usr/lib/x86_64-linux-gnu/glusterfs/3.12.12/xlator/mgmt/glusterd.so(+0xd1e59)
>> [0x7fdba4f9ce59]
>> -->/usr/lib/x86_64-linux-gnu/libglusterfs.so.0(dict_set_int32+0x2b)
>> [0x7fdbaa39122b]
>> -->/usr/lib/x86_64-linux-gnu/libglusterfs.so.0(dict_set+0xd3)
>> [0x7fdbaa38fa13] ) 0-dict: !this || !value for key=index [I
>> nvalid argument]
>> [2018-08-16 05:22:52.988092] E [MSGID: 106060]
>> [glusterd-syncop.c:1014:gd_syncop_mgmt_brick_op] 0-management: Error
>> setting index on brick status rsp dict
>>
>> This problem could be related to my previous mail. After executing
>> "gluster volume start shared force" the brick comes up, resulting in
>> healing the brick (and in high load, too). Is there any possibility to
>> track down why this happens and how to ensure that the brick comes up
>> at boot?
>>
>>
>> Best regards
>> Hubert
>> _______________________________________________
>> Gluster-users mailing list
>> Gluster-users@xxxxxxxxxxx
>> https://lists.gluster.org/mailman/listinfo/gluster-users
>>
_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
https://lists.gluster.org/mailman/listinfo/gluster-users




[Index of Archives]     [Gluster Development]     [Linux Filesytems Development]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux