Re: NFS not start on localhost

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Sat, Oct 18, 2014 at 01:24:12PM +0200, Demeter Tibor wrote:
> Hi, 
> 
> [root@node0 ~]# tail -n 20 /var/log/glusterfs/nfs.log 
> [2014-10-18 07:41:06.136035] E [graph.c:307:glusterfs_graph_init] 0-nfs-server: initializing translator failed 
> [2014-10-18 07:41:06.136040] E [graph.c:502:glusterfs_graph_activate] 0-graph: init failed 
> pending frames: 
> frame : type(0) op(0) 
> 
> patchset: git://git.gluster.com/glusterfs.git 
> signal received: 11 
> time of crash: 2014-10-18 07:41:06configuration details: 
> argp 1 
> backtrace 1 
> dlfcn 1 
> fdatasync 1 
> libpthread 1 
> llistxattr 1 
> setfsid 1 
> spinlock 1 
> epoll.h 1 
> xattr.h 1 
> st_atim.tv_nsec 1 
> package-string: glusterfs 3.5.2 

This definitely is a gluster/nfs issue. For whatever reasone, the
gluster/nfs server crashes :-/ The log does not show enough details,
some more lines before this are needed.

There might be an issue where the NFS RPC-services can not register. I
think I have seen similar crashes before, but never found the cause. You
should check with the 'rpcinfo' command to see if there are any NFS
RPC-services registered (nfs, lockd, mount, lockmgr). If there are any,
verify that there are no other nfs processes running, this includes
NFS-mounts in /etc/fstab and similar.

Could you file a bug, attach the full (gzipped) nfs.log? Try to explain
as much details of the setup as you can, and add a link to the archives
of this thread. Please post the url to the bug in a response to this
thread. A crashing process is never good, even when is could be caused
by external processes.

Link to file a bug:
- https://bugzilla.redhat.com/enter_bug.cgi?product=GlusterFS&component=nfs&version=3.5.2

Thanks,
Niels


> 
> Udv: 
> 
> Demeter Tibor 
> 
> Email: tdemeter @itsmart.hu 
> Skype: candyman_78 
> Phone: +36 30 462 0500 
> Web : www.it smart.hu 
> 
> IT SMART KFT. 
> 2120 Dunakeszi Wass Albert utca 2. I. em 9. 
> Telefon: +36 30 462-0500 Fax: +36 27 637-486 
> 
> [EN] This message and any attachments are confidential and privileged and intended for the use of the addressee only. If you have received this communication in error, please notify the sender by replay e-mail and delete this message from your system. Please note that Internet e-mail guarantees neither the confidentiality nor the proper receipt of the message sent. The data deriving from our correspondence with you are included in a file of ITSMART Ltd which exclusive purpose is to manage the communications of the company; under the understanding that, in maintaining said correspondence, you authorize the treatment of such data for the mentioned purpose. You are entitled to exercise your rights of access, rectification, cancellation and opposition by addressing such written application to address above. 
> [HUN] Ez az üzenet és annak bármely csatolt anyaga bizalmas, a nyilvános közléstol védett, kizárólag a címzett használhatja fel. Ha Ön nem az üzenet címzettje, úgy kérjük válaszüzenetben értesítse errol a feladót és törölje az üzenetet a rendszerbol. Kérjük vegye figyelembe, hogy az email-en történo információtovábbítás kockázattal járhat, nem garantálja sem a csatorna bizalmasságát, sem a kézbesítést. A levél az ITSMART Informatikai Kft. kommunikációjának eszköze, az adatokat kizárólag erre a célra használjuk. Jogosult tájékoztatást kérni személyes adatai kezelésérol, kérheti azok helyesbítését, illetve törlését írásos kérelemben a fenti e-mail címen. 
> 
> ----- Eredeti üzenet -----
> 
> > Maybe share the last 15-20 lines of you /var/log/glusterfs/nfs.log for the
> > consideration of everyone on the list? Thanks.
> 
> > From: Demeter Tibor <tdemeter@xxxxxxxxxx>;
> > To: Anirban Ghoshal <chalcogen_eg_oxygen@xxxxxxxxx>;
> > Cc: gluster-users <gluster-users@xxxxxxxxxxx>;
> > Subject: Re:  NFS not start on localhost
> > Sent: Sat, Oct 18, 2014 10:36:36 AM
> 
> > 
> > Hi,
> 
> > I've try out these things:
> 
> > - nfs.disable on-of
> > - iptables disable
> > - volume stop-start
> 
> > but same.
> > So, when I make a new volume everything is fine.
> > After reboot the NFS won't listen on local host (only on server has brick0)
> 
> > Centos7 with last ovirt
> 
> > Regards,
> 
> > Tibor
> 
> > ----- Eredeti üzenet -----
> 
> > > It happens with me sometimes. Try `tail -n 20 /var/log/glusterfs/nfs.log`.
> > > You will probably find something out that will help your cause. In general,
> > > if you just wish to start the thing up without going into the why of it,
> > > try
> > > `gluster volume set engine nfs.disable on` followed by ` gluster volume set
> > > engine nfs.disable off`. It does the trick quite often for me because it is
> > > a polite way to askmgmt/glusterd to try and respawn the nfs server process
> > > if need be. But, keep in mind that this will call a (albeit small) service
> > > interruption to all clients accessing volume engine over nfs.
> > 
> 
> > > Thanks,
> > 
> > > Anirban
> > 
> 
> > > On Saturday, 18 October 2014 1:03 AM, Demeter Tibor <tdemeter@xxxxxxxxxx>
> > > wrote:
> > 
> 
> > > Hi,
> > 
> 
> > > I have make a glusterfs with nfs support.
> > 
> 
> > > I don't know why, but after a reboot the nfs does not listen on localhost,
> > > only on gs01.
> > 
> 
> > > [root@node0 ~]# gluster volume info engine
> > 
> 
> > > Volume Name: engine
> > 
> > > Type: Replicate
> > 
> > > Volume ID: 2ea009bf-c740-492e-956d-e1bca76a0bd3
> > 
> > > Status: Started
> > 
> > > Number of Bricks: 1 x 2 = 2
> > 
> > > Transport-type: tcp
> > 
> > > Bricks:
> > 
> > > Brick1: gs00.itsmart.cloud:/gluster/engine0
> > 
> > > Brick2: gs01.itsmart.cloud:/gluster/engine1
> > 
> > > Options Reconfigured:
> > 
> > > storage.owner-uid: 36
> > 
> > > storage.owner-gid: 36
> > 
> > > performance.quick-read: off
> > 
> > > performance.read-ahead: off
> > 
> > > performance.io-cache: off
> > 
> > > performance.stat-prefetch: off
> > 
> > > cluster.eager-lock: enable
> > 
> > > network.remote-dio: enable
> > 
> > > cluster.quorum-type: auto
> > 
> > > cluster.server-quorum-type: server
> > 
> > > auth.allow: *
> > 
> > > nfs.disable: off
> > 
> 
> > > [root@node0 ~]# gluster volume status engine
> > 
> > > Status of volume: engine
> > 
> > > Gluster process Port Online Pid
> > 
> > > ------------------------------------------------------------------------------
> > 
> > > Brick gs00.itsmart.cloud:/gluster/engine0 50158 Y 3250
> > 
> > > Brick gs01.itsmart.cloud:/gluster/engine1 50158 Y 5518
> > 
> > > NFS Server on localhost N/A N N/A
> > 
> > > Self-heal Daemon on localhost N/A Y 3261
> > 
> > > NFS Server on gs01.itsmart.cloud 2049 Y 5216
> > 
> > > Self-heal Daemon on gs01.itsmart.cloud N/A Y 5223
> > 
> 
> > > Does anybody help me?
> > 
> 
> > > Thanks in advance.
> > 
> 
> > > Tibor
> > 
> 
> > > _______________________________________________
> > 
> > > Gluster-users mailing list
> > 
> > > Gluster-users@xxxxxxxxxxx
> > 
> > > http://supercolony.gluster.org/mailman/listinfo/gluster-users
> > 

> _______________________________________________
> Gluster-users mailing list
> Gluster-users@xxxxxxxxxxx
> http://supercolony.gluster.org/mailman/listinfo/gluster-users

_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
http://supercolony.gluster.org/mailman/listinfo/gluster-users





[Index of Archives]     [Gluster Development]     [Linux Filesytems Development]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux