Re: [Nfs-ganesha-devel] Questions on ganesha HA and shared storage size

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



OK, I think we are now closer to the end of the story.
Recompiling with your instructions, and slightly changing the release name to match the convention in epel, the new RPMS produce something working!
So it means essentially that:

1) the RPMS in epel are broken, they should really be fixed;
2) the RPMS, produced by exporting the tarball from git, even by selecting the correct branch, and the spec file from epel are broken as well;
3) following your procedure produce working packages, but with revision “0.3” instead of the required “3” (not a real problem, easy to fix).

I have created a repo in my institute, that can be used from outside too, in case someone is interested, it may be accessed by yum in this way:

[nfs-ganesha-infn]
name=NFS-Ganesha-INFN
baseurl=http://classis01.roma1.infn.it/RPMS/nfs-ganesha/el$releasever/$basearch/
enabled=1
skip_if_unavailable=1
gpgcheck=0


I have also updated the puppet module to use this repo as failover, for the moment. I will publish the module soon in puppetforge, but you can already use it from here:

https://github.com/desalvo/puppet-ganesha

The module is still working only for gluster, but it’s very easy to add more.
Thanks for all the help,

	Alessandro

> Il giorno 15/giu/2015, alle ore 21:40, Alessandro De Salvo <Alessandro.DeSalvo@xxxxxxxxxxxxx> ha scritto:
> 
> Hi Malahal,
> I have downloaded and used the tarball created by git for the stable 2.2.0 branch, so it should have been consistent. Also, I have used the spec file from Epel to build the rpms. I’m going to try your procedure as well now, to see if anything changes.
> Thanks,
> 
> 	Alessandro
> 
>> Il giorno 15/giu/2015, alle ore 20:10, Malahal Naineni <malahal@xxxxxxxxxx> ha scritto:
>> 
>> I am not familiar with tirpc code, but how are you building ganesha
>> rpms? Did you do "git submodule update" to get the latest tirpc code
>> when you built those rpms? Can somebody familiar with tirpc chime in?
>> 
>> The one I use is below:
>> # git submodule status
>> b1a82463c4029315fac085a9d0d6bef766847ed7 src/libntirpc (v1.2.0-2-gb1a8246)
>> 
>> The way I build ganesha 2.2 rpms is:
>> #1. git clone <repo> && git checkout V2.2-stable
>> #2. git submodule update --init
>> #3. cmake ./src -DDEBUG_SYMS=ON -DUSE_DBUS=ON -DUSE_ADMIN_TOOLS=ON -DUSE_GUI_ADMIN_TOOLS=OFF
>> #4. make dist
>> #5. rpmbuild --with utils -ta nfs-ganesha-2.2*
>> 
>> Regards, Malahal.
>> PS: there were some efforts to make tirpc as an rpm by itself. Not sure where
>> that is.
>> 
>> Alessandro De Salvo [Alessandro.DeSalvo@xxxxxxxxxxxxx] wrote:
>>> OK, thanks, so, any hint on what I could check now?
>>> I have tried even without any VFS, so with just the nfs-ganesha rpm installed and with an empty ganesha.conf, but still the same problem. The same configuration with ganesha 2.1.0 was working, on the same server.
>>> Any idea? I have sent you the logs but please tell me if you need more.
>>> Thanks,
>>> 
>>> 	Alessandro
>>> 
>>>> Il giorno 15/giu/2015, alle ore 18:47, Malahal Naineni <malahal@xxxxxxxxxx> ha scritto:
>>>> 
>>>> We do run ganesha on RHEL7.0 (same as CentOS7.0), and I don't think 7.1
>>>> would be much different. We do run GPFS FSAL only (no VFS_FSAL).
>>>> 
>>>> Regards, Malahal.
>>>> 
>>>> Alessandro De Salvo [Alessandro.DeSalvo@xxxxxxxxxxxxx] wrote:
>>>>> Hi,
>>>>> any news on this? Did you have the chance to look into that?
>>>>> I'd also be curious to know if anyone tried nfs ganesha on CentOS 7.1
>>>>> and if it was really working, as I also tried on a standalone, clean
>>>>> machine, and I see the very same behavior, even without gluster.
>>>>> Thanks,
>>>>> 
>>>>> 	Alessandro
>>>>> 
>>>>> On Fri, 2015-06-12 at 14:34 +0200, Alessandro De Salvo wrote:
>>>>>> Hi,
>>>>>> looking at the code and having recompiled adding some more debug, I
>>>>>> might be wrong, but it seems that in nfs_rpc_dispatcher_thread.c,
>>>>>> fuction nfs_rpc_dequeue_req, the threads enter the while (!(wqe->flags &
>>>>>> Wqe_LFlag_SyncDone)) and never exit from there.
>>>>>> I do not know if it's normal or not as I should read better the code.
>>>>>> Cheers,
>>>>>> 
>>>>>> 	Alessandro
>>>>>> 
>>>>>> On Fri, 2015-06-12 at 09:35 +0200, Alessandro De Salvo wrote:
>>>>>>> Hi Malahal,
>>>>>>> 
>>>>>>> 
>>>>>>>> Il giorno 12/giu/2015, alle ore 01:23, Malahal Naineni <malahal@xxxxxxxxxx> ha scritto:
>>>>>>>> 
>>>>>>>> The logs indicate that ganesha was started successfully without any
>>>>>>>> exports.  gstack output seemed normal as well -- threads were waiting to
>>>>>>>> serve requests.
>>>>>>> 
>>>>>>> Yes, no exports as it was the default config before enabling Ganesha on any gluster volume.
>>>>>>> 
>>>>>>>> 
>>>>>>>> Assuming that you are running "showmount -e" on the same system, there
>>>>>>>> shouldn't be any firewall coming into the picture.
>>>>>>> 
>>>>>>> Yes it was the case in my last attempt, from the same machine. I also tried from another machine, but the result was the same. The firewall (firewalld, as it's a CentOS 7.1) is disabled anyways.
>>>>>>> 
>>>>>>>> If you are running
>>>>>>>> "showmount" from some other system, make sure there is no firewall
>>>>>>>> dropping the packets.
>>>>>>>> 
>>>>>>>> I think you need tcpdump trace to figure out the problem. My wireshark
>>>>>>>> trace showed two requests from the client to complete the "showmount -e"
>>>>>>>> command:
>>>>>>>> 
>>>>>>>> 1. Client sent "GETPORT" call to port 111 (rpcbind) to get the port number
>>>>>>>> of MOUNT.
>>>>>>>> 2. Then it sent "EXPORT" call to mountd port (port it got in response to #1).
>>>>>>> 
>>>>>>> Yes, I did it already, and indeed it showed the two requests, so the portmapper works fine, but it hangs on the second request.
>>>>>>> Also "rpcinfo -t localhost portmapper" returns successfully, while "rpcinfo -t localhost nfs" hangs.
>>>>>>> The output of rpcinfo -p is the following:
>>>>>>> 
>>>>>>>  program vers proto   port  service
>>>>>>>  100000    4   tcp    111  portmapper
>>>>>>>  100000    3   tcp    111  portmapper
>>>>>>>  100000    2   tcp    111  portmapper
>>>>>>>  100000    4   udp    111  portmapper
>>>>>>>  100000    3   udp    111  portmapper
>>>>>>>  100000    2   udp    111  portmapper
>>>>>>>  100024    1   udp  56082  status
>>>>>>>  100024    1   tcp  41858  status
>>>>>>>  100003    3   udp   2049  nfs
>>>>>>>  100003    3   tcp   2049  nfs
>>>>>>>  100003    4   udp   2049  nfs
>>>>>>>  100003    4   tcp   2049  nfs
>>>>>>>  100005    1   udp  45611  mountd
>>>>>>>  100005    1   tcp  55915  mountd
>>>>>>>  100005    3   udp  45611  mountd
>>>>>>>  100005    3   tcp  55915  mountd
>>>>>>>  100021    4   udp  48775  nlockmgr
>>>>>>>  100021    4   tcp  51621  nlockmgr
>>>>>>>  100011    1   udp   4501  rquotad
>>>>>>>  100011    1   tcp   4501  rquotad
>>>>>>>  100011    2   udp   4501  rquotad
>>>>>>>  100011    2   tcp   4501  rquotad
>>>>>>> 
>>>>>>>> 
>>>>>>>> What does "rpcinfo -p <server-ip>" show?
>>>>>>>> 
>>>>>>>> Do you have selinux enabled? I am not sure if that is playing any role
>>>>>>>> here...
>>>>>>> 
>>>>>>> Nope, it's disabled:
>>>>>>> 
>>>>>>> # uname -a
>>>>>>> Linux node2 3.10.0-229.4.2.el7.x86_64 #1 SMP Wed May 13 10:06:09 UTC 2015 x86_64 x86_64 x86_64 GNU/Linux
>>>>>>> 
>>>>>>> 
>>>>>>> Thanks for the help,
>>>>>>> 
>>>>>>>  Alessandro
>>>>>>> 
>>>>>>>> 
>>>>>>>> Regards, Malahal.
>>>>>>>> 
>>>>>>>> Alessandro De Salvo [Alessandro.DeSalvo@xxxxxxxxxxxxx] wrote:
>>>>>>>>> Hi,
>>>>>>>>> this was an extract from the old logs, before Soumya's suggestion of
>>>>>>>>> changing the rquota port in the conf file. The new logs are attached
>>>>>>>>> (ganesha-20150611.log.gz) as well as the gstack of the ganesha process
>>>>>>>>> while I was executing the hanging showmount
>>>>>>>>> (ganesha-20150611.gstack.gz).
>>>>>>>>> Thanks,
>>>>>>>>> 
>>>>>>>>> Alessandro
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>>> On Thu, 2015-06-11 at 11:37 -0500, Malahal Naineni wrote:
>>>>>>>>>> Soumya Koduri [skoduri@xxxxxxxxxx] wrote:
>>>>>>>>>>> CCin ganesha-devel to get more inputs.
>>>>>>>>>>> 
>>>>>>>>>>> In case of ipv6 enabled, only v6 interfaces are used by NFS-Ganesha.
>>>>>>>>>> 
>>>>>>>>>> I am not a network expert but I have seen IPv4 traffic over IPv6
>>>>>>>>>> interface while fixing few things before. This may be normal.
>>>>>>>>>> 
>>>>>>>>>>> commit - git show 'd7e8f255' , which got added in v2.2 has more details.
>>>>>>>>>>> 
>>>>>>>>>>>> # netstat -ltaupn | grep 2049
>>>>>>>>>>>> tcp6       4      0 :::2049                 :::*
>>>>>>>>>>>> LISTEN      32080/ganesha.nfsd
>>>>>>>>>>>> tcp6       1      0 x.x.x.2:2049      x.x.x.2:33285     CLOSE_WAIT
>>>>>>>>>>>> -
>>>>>>>>>>>> tcp6       1      0 127.0.0.1:2049          127.0.0.1:39555
>>>>>>>>>>>> CLOSE_WAIT  -
>>>>>>>>>>>> udp6       0      0 :::2049                 :::*
>>>>>>>>>>>> 32080/ganesha.nfsd
>>>>>>>>>>> 
>>>>>>>>>>>>>> I have enabled the full debug already, but I see nothing special. Before exporting any volume the log shows no error, even when I do a showmount (the log is attached, ganesha.log.gz). If I do the same after exporting a volume nfs-ganesha does not even start, complaining for not being able to bind the IPv6 ruota socket, but in fact there is nothing listening on IPv6, so it should not happen:
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> tcp6       0      0 :::111                  :::*                    LISTEN      7433/rpcbind
>>>>>>>>>>>>>> tcp6       0      0 :::2224                 :::*                    LISTEN      9054/ruby
>>>>>>>>>>>>>> tcp6       0      0 :::22                   :::*                    LISTEN      1248/sshd
>>>>>>>>>>>>>> udp6       0      0 :::111                  :::*                                7433/rpcbind
>>>>>>>>>>>>>> udp6       0      0 fe80::8c2:27ff:fef2:123 :::*                                31238/ntpd
>>>>>>>>>>>>>> udp6       0      0 fe80::230:48ff:fed2:123 :::*                                31238/ntpd
>>>>>>>>>>>>>> udp6       0      0 fe80::230:48ff:fed2:123 :::*                                31238/ntpd
>>>>>>>>>>>>>> udp6       0      0 fe80::230:48ff:fed2:123 :::*                                31238/ntpd
>>>>>>>>>>>>>> udp6       0      0 ::1:123                 :::*                                31238/ntpd
>>>>>>>>>>>>>> udp6       0      0 fe80::5484:7aff:fef:123 :::*                                31238/ntpd
>>>>>>>>>>>>>> udp6       0      0 :::123                  :::*                                31238/ntpd
>>>>>>>>>>>>>> udp6       0      0 :::824                  :::*                                7433/rpcbind
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> The error, as shown in the attached ganesha-after-export.log.gz logfile, is the following:
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> 10/06/2015 02:07:47 : epoch 55777fb5 : node2 : ganesha.nfsd-26195[main] Bind_sockets_V6 :DISP :WARN :Cannot bind RQUOTA tcp6 socket, error 98 (Address already in use)
>>>>>>>>>>>>>> 10/06/2015 02:07:47 : epoch 55777fb5 : node2 : ganesha.nfsd-26195[main] Bind_sockets :DISP :FATAL :Error binding to V6 interface. Cannot continue.
>>>>>>>>>>>>>> 10/06/2015 02:07:48 : epoch 55777fb5 : node2 : ganesha.nfsd-26195[main] glusterfs_unload :FSAL :DEBUG :FSAL Gluster unloaded
>>>>>>>>>> 
>>>>>>>>>> The above messages indicate that someone tried to restart ganesha. But
>>>>>>>>>> ganesha failed to come up because RQUOTA port (default is 875) is
>>>>>>>>>> already in use by an old ganesha instance or some other program holding
>>>>>>>>>> it. The new instance of ganesha will die, but if you are using systemd,
>>>>>>>>>> it will try to restart automatically. We have disabled systemd auto
>>>>>>>>>> restart in our environment as it was causing issues for debugging.
>>>>>>>>>> 
>>>>>>>>>> What version of ganesha is this?
>>>>>>>>>> 
>>>>>>>>>> Regards, Malahal.
>>>>>>>> 
>>>>>>>> 
>>>>>>>> 
>>>>>>> _______________________________________________
>>>>>>> Gluster-users mailing list
>>>>>>> Gluster-users@xxxxxxxxxxx
>>>>>>> http://www.gluster.org/mailman/listinfo/gluster-users
>>>>>> 
>>>>>> 
>>>>>> _______________________________________________
>>>>>> Gluster-users mailing list
>>>>>> Gluster-users@xxxxxxxxxxx
>>>>>> http://www.gluster.org/mailman/listinfo/gluster-users
>>>>> 
>>>>> 
>>>> 
>>> 
>> 
>> 
> 
> _______________________________________________
> Gluster-users mailing list
> Gluster-users@xxxxxxxxxxx
> http://www.gluster.org/mailman/listinfo/gluster-users

Attachment: smime.p7s
Description: S/MIME cryptographic signature

_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
http://www.gluster.org/mailman/listinfo/gluster-users

[Index of Archives]     [Gluster Development]     [Linux Filesytems Development]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux