On Wed, May 31, 2017 at 04:08:06PM +0530, ABHISHEK PALIWAL wrote: > We are using 3.7.6 and on link https://review.gluster.org/#/c/16279 status > is "can't merge" Note that 3.7.x will not get any updates anymore. We currently maintain version 3.8.x, 3.10.x and 3.11.x. See the release schedele for more details: https://www.gluster.org/community/release-schedule/ Niels > > On Wed, May 31, 2017 at 4:05 PM, Amar Tumballi <atumball@xxxxxxxxxx> wrote: > > > This is already part of 3.11.0 release? > > > > On Wed, May 31, 2017 at 3:47 PM, ABHISHEK PALIWAL <abhishpaliwal@xxxxxxxxx > > > wrote: > > > >> Hi Atin, > >> > >> Could you please let us know any time plan for deliver of this patch. > >> > >> Regards, > >> Abhishek > >> > >> On Tue, May 9, 2017 at 6:37 PM, ABHISHEK PALIWAL <abhishpaliwal@xxxxxxxxx > >> > wrote: > >> > >>> Actually it is very risky if it will reproduce in production thats is > >>> why I said it is on high priority as want to resolve it before production. > >>> > >>> On Tue, May 9, 2017 at 6:20 PM, Atin Mukherjee <amukherj@xxxxxxxxxx> > >>> wrote: > >>> > >>>> > >>>> > >>>> On Tue, May 9, 2017 at 6:10 PM, ABHISHEK PALIWAL < > >>>> abhishpaliwal@xxxxxxxxx> wrote: > >>>> > >>>>> Hi Atin, > >>>>> > >>>>> Thanks for your reply. > >>>>> > >>>>> > >>>>> Its urgent because this error is very rarely reproducible we have seen > >>>>> this 2 3 times in our system till now. > >>>>> > >>>>> We have delivery in near future so that we want it asap. Please try to > >>>>> review it internally. > >>>>> > >>>> > >>>> I don't think your statements justified the reason of urgency as (a) > >>>> you have mentioned it to be *rarely* reproducible and (b) I am still > >>>> waiting for a real use case where glusterd will go through multiple > >>>> restarts in a loop? > >>>> > >>>> > >>>>> Regards, > >>>>> Abhishek > >>>>> > >>>>> On Tue, May 9, 2017 at 5:58 PM, Atin Mukherjee <amukherj@xxxxxxxxxx> > >>>>> wrote: > >>>>> > >>>>>> > >>>>>> > >>>>>> On Tue, May 9, 2017 at 3:37 PM, ABHISHEK PALIWAL < > >>>>>> abhishpaliwal@xxxxxxxxx> wrote: > >>>>>> > >>>>>>> + Muthu-vingeshwaran > >>>>>>> > >>>>>>> On Tue, May 9, 2017 at 11:30 AM, ABHISHEK PALIWAL < > >>>>>>> abhishpaliwal@xxxxxxxxx> wrote: > >>>>>>> > >>>>>>>> Hi Atin/Team, > >>>>>>>> > >>>>>>>> We are using gluster-3.7.6 with setup of two brick and while > >>>>>>>> restart of system I have seen that the glusterd daemon is getting failed > >>>>>>>> from start. > >>>>>>>> > >>>>>>>> > >>>>>>>> At the time of analyzing the logs from etc-glusterfs.......log file > >>>>>>>> I have received the below logs > >>>>>>>> > >>>>>>>> > >>>>>>>> [2017-05-06 03:33:39.798087] I [MSGID: 100030] > >>>>>>>> [glusterfsd.c:2348:main] 0-/usr/sbin/glusterd: Started running > >>>>>>>> /usr/sbin/glusterd version 3.7.6 (args: /usr/sbin/glusterd -p > >>>>>>>> /var/run/glusterd.pid --log-level INFO) > >>>>>>>> [2017-05-06 03:33:39.807859] I [MSGID: 106478] > >>>>>>>> [glusterd.c:1350:init] 0-management: Maximum allowed open file descriptors > >>>>>>>> set to 65536 > >>>>>>>> [2017-05-06 03:33:39.807974] I [MSGID: 106479] > >>>>>>>> [glusterd.c:1399:init] 0-management: Using /system/glusterd as working > >>>>>>>> directory > >>>>>>>> [2017-05-06 03:33:39.826833] I [MSGID: 106513] > >>>>>>>> [glusterd-store.c:2047:glusterd_restore_op_version] 0-glusterd: > >>>>>>>> retrieved op-version: 30706 > >>>>>>>> [2017-05-06 03:33:39.827515] E [MSGID: 106206] > >>>>>>>> [glusterd-store.c:2562:glusterd_store_update_volinfo] > >>>>>>>> 0-management: Failed to get next store iter > >>>>>>>> [2017-05-06 03:33:39.827563] E [MSGID: 106207] > >>>>>>>> [glusterd-store.c:2844:glusterd_store_retrieve_volume] > >>>>>>>> 0-management: Failed to update volinfo for c_glusterfs volume > >>>>>>>> [2017-05-06 03:33:39.827625] E [MSGID: 106201] > >>>>>>>> [glusterd-store.c:3042:glusterd_store_retrieve_volumes] > >>>>>>>> 0-management: Unable to restore volume: c_glusterfs > >>>>>>>> [2017-05-06 03:33:39.827722] E [MSGID: 101019] > >>>>>>>> [xlator.c:428:xlator_init] 0-management: Initialization of volume > >>>>>>>> 'management' failed, review your volfile again > >>>>>>>> [2017-05-06 03:33:39.827762] E [graph.c:322:glusterfs_graph_init] > >>>>>>>> 0-management: initializing translator failed > >>>>>>>> [2017-05-06 03:33:39.827784] E [graph.c:661:glusterfs_graph_activate] > >>>>>>>> 0-graph: init failed > >>>>>>>> [2017-05-06 03:33:39.828396] W [glusterfsd.c:1238:cleanup_and_exit] > >>>>>>>> (-->/usr/sbin/glusterd(glusterfs_volumes_init-0x1b0b8) > >>>>>>>> [0x1000a648] -->/usr/sbin/glusterd(glusterfs_process_volfp-0x1b210) > >>>>>>>> [0x1000a4d8] -->/usr/sbin/glusterd(cleanup_and_exit-0x1beac) > >>>>>>>> [0x100097ac] ) 0-: received signum (0), shutting down > >>>>>>>> > >>>>>>> > >>>>>> Abhishek, > >>>>>> > >>>>>> This patch needs to be thoroughly reviewed to ensure that it doesn't > >>>>>> cause any regression given this touches on the core store management > >>>>>> functionality of glusterd. AFAICT, we get into an empty info file only when > >>>>>> volume set operation is executed and in parallel one of the glusterd > >>>>>> instance in other nodes have been brought down and whole sequence of > >>>>>> operation happens in a loop. The test case through which you can get into > >>>>>> this situation is not something you'd hit in production. Please help me to > >>>>>> understand the urgency here. > >>>>>> > >>>>>> Also in one of the earlier thread, I did mention the workaround of > >>>>>> this issue back to Xin through http://lists.gluster.org/piper > >>>>>> mail/gluster-users/2017-January/029600.html > >>>>>> > >>>>>> "If you end up in having a 0 byte info file you'd need to copy the same info file from other node and put it there and restart glusterd" > >>>>>> > >>>>>> > >>>>>>>> > >>>>>>>> I have found one of the existing case is there and also solution > >>>>>>>> patch is available but the status of that patch in "cannot merge". Also the > >>>>>>>> "info" file is empty and "info.tmp" file present in "lib/glusterd/vol" > >>>>>>>> directory. > >>>>>>>> > >>>>>>>> Below is the link of the existing case. > >>>>>>>> > >>>>>>>> https://review.gluster.org/#/c/16279/5 > >>>>>>>> > >>>>>>>> please let me know what is the plan of community to provide the > >>>>>>>> solution of this problem and in which version. > >>>>>>>> > >>>>>>>> Regards > >>>>>>>> Abhishek Paliwal > >>>>>>>> > >>>>>>> > >>>>>>> > >>>>>>> > >>>>>>> -- > >>>>>>> > >>>>>>> > >>>>>>> > >>>>>>> > >>>>>>> Regards > >>>>>>> Abhishek Paliwal > >>>>>>> > >>>>>> > >>>>>> > >>>>> > >>>>> > >>>>> -- > >>>>> > >>>>> > >>>>> > >>>>> > >>>>> Regards > >>>>> Abhishek Paliwal > >>>>> > >>>> > >>>> > >>> > >>> > >>> -- > >>> > >>> > >>> > >>> > >>> Regards > >>> Abhishek Paliwal > >>> > >> > >> > >> > >> -- > >> > >> > >> > >> > >> Regards > >> Abhishek Paliwal > >> > >> _______________________________________________ > >> Gluster-devel mailing list > >> Gluster-devel@xxxxxxxxxxx > >> http://lists.gluster.org/mailman/listinfo/gluster-devel > >> > > > > > > > > -- > > Amar Tumballi (amarts) > > > > > > -- > > > > > Regards > Abhishek Paliwal > _______________________________________________ > Gluster-devel mailing list > Gluster-devel@xxxxxxxxxxx > http://lists.gluster.org/mailman/listinfo/gluster-devel
Attachment:
signature.asc
Description: PGP signature
_______________________________________________ Gluster-devel mailing list Gluster-devel@xxxxxxxxxxx http://lists.gluster.org/mailman/listinfo/gluster-devel