Hi Soumya, again I apologize for delay in response. I'll try to file a bug. Meantime I'm sending AVCs and version number. AVC are collected between two reboots, in both cases I manually started nfs-ganesha.service and nfs-ganesha-lock.service failed to start. uname -r 3.10.0-514.21.1.el7.x86_64 sestatus -v SELinux status: enabled SELinuxfs mount: /sys/fs/selinux SELinux root directory: /etc/selinux Loaded policy name: targeted Current mode: enforcing Mode from config file: enforcing Policy MLS status: enabled Policy deny_unknown status: allowed Max kernel policy version: 28 Process contexts: Current context: unconfined_u:unconfined_r:unconfined_t:s0-s0:c0.c1023 Init context: system_u:system_r:init_t:s0 File contexts: Controlling terminal: unconfined_u:object_r:user_tty_device_t:s0 /etc/passwd system_u:object_r:passwd_file_t:s0 /etc/shadow system_u:object_r:shadow_t:s0 /bin/bash system_u:object_r:shell_exec_t:s0 /bin/login system_u:object_r:login_exec_t:s0 /bin/sh system_u:object_r:bin_t:s0 -> system_u:object_r:shell_exec_t:s0 /sbin/agetty system_u:object_r:getty_exec_t:s0 /sbin/init system_u:object_r:bin_t:s0 -> system_u:object_r:init_exec_t:s0 /usr/sbin/sshd system_u:object_r:sshd_exec_t:s0 sudo systemctl start nfs-ganesha.service systemctl status -l nfs-ganesha-lock.service ● nfs-ganesha-lock.service - NFS status monitor for NFSv2/3 locking. Loaded: loaded (/usr/lib/systemd/system/nfs-ganesha-lock.service; static; vendor preset: disabled) Active: failed (Result: exit-code) since Sun 2017-05-28 14:12:48 UTC; 9s ago Process: 1991 ExecStart=/usr/sbin/rpc.statd --no-notify $STATDARGS (code=exited, status=1/FAILURE) mynode0.localdomain systemd[1]: Starting NFS status monitor for NFSv2/3 locking.... mynode0.localdomain rpc.statd[1992]: Version 1.3.0 starting mynode0.localdomain rpc.statd[1992]: Flags: TI-RPC mynode0.localdomain rpc.statd[1992]: Failed to open directory sm: Permission denied mynode0.localdomain systemd[1]: nfs-ganesha-lock.service: control process exited, code=exited status=1 mynode0.localdomain systemd[1]: Failed to start NFS status monitor for NFSv2/3 locking.. mynode0.localdomain systemd[1]: Unit nfs-ganesha-lock.service entered failed state. mynode0.localdomain systemd[1]: nfs-ganesha-lock.service failed. sudo ausearch -m AVC,USER_AVC,SELINUX_ERR,USER_SELINUX_ERR -i ---- type=SYSCALL msg=audit(05/28/2017 14:04:32.160:25) : arch=x86_64 syscall=bind success=yes exit=0 a0=0xf a1=0x7ffc757feb60 a2=0x10 a3=0x22 items=0 ppid=1149 pid=1157 auid=unset uid=root gid=root euid=root suid=root fsuid=root egid=root sgid=root fsgid=root tty=(none) ses=unset comm=glusterd exe=/usr/sbin/glusterfsd subj=system_u:system_r:glusterd_t:s0 key=(null) type=AVC msg=audit(05/28/2017 14:04:32.160:25) : avc: denied { name_bind } for pid=1157 comm=glusterd src=61000 scontext=system_u:system_r:glusterd_t:s0 tcontext=system_u:object_r:ephemeral_port_t:s0 tclass=tcp_socket ---- type=SYSCALL msg=audit(05/28/2017 14:11:16.141:26) : arch=x86_64 syscall=bind success=no exit=EACCES(Permission denied) a0=0xf a1=0x7ffffbf92620 a2=0x10 a3=0x22 items=0 ppid=1139 pid=1146 auid=unset uid=root gid=root euid=root suid=root fsuid=root egid=root sgid=root fsgid=root tty=(none) ses=unset comm=glusterd exe=/usr/sbin/glusterfsd subj=system_u:system_r:glusterd_t:s0 key=(null) type=AVC msg=audit(05/28/2017 14:11:16.141:26) : avc: denied { name_bind } for pid=1146 comm=glusterd src=61000 scontext=system_u:system_r:glusterd_t:s0 tcontext=system_u:object_r:ephemeral_port_t:s0 tclass=tcp_socket ---- type=SYSCALL msg=audit(05/28/2017 14:12:48.068:75) : arch=x86_64 syscall=openat success=no exit=EACCES(Permission denied) a0=0xffffffffffffff9c a1=0x7efdc1ec3e10 a2=O_RDONLY|O_NONBLOCK|O_DIRECTORY|O_CLOEXEC a3=0x0 items=0 ppid=1991 pid=1992 auid=unset uid=root gid=root euid=root suid=root fsuid=root egid=root sgid=root fsgid=root tty=(none) ses=unset comm=rpc.statd exe=/usr/sbin/rpc.statd subj=system_u:system_r:rpcd_t:s0 key=(null) type=AVC msg=audit(05/28/2017 14:12:48.068:75) : avc: denied { read } for pid=1992 comm=rpc.statd name=sm dev="fuse" ino=12866274077597183313 scontext=system_u:system_r:rpcd_t:s0 tcontext=system_u:object_r:fusefs_t:s0 tclass=dir ---- type=SYSCALL msg=audit(05/28/2017 14:12:48.080:76) : arch=x86_64 syscall=open success=no exit=EACCES(Permission denied) a0=0x7efdc1ec3dd0 a1=O_RDONLY a2=0x7efdc1ec3de8 a3=0x5 items=0 ppid=1991 pid=1992 auid=unset uid=root gid=root euid=root suid=root fsuid=root egid=root sgid=root fsgid=root tty=(none) ses=unset comm=rpc.statd exe=/usr/sbin/rpc.statd subj=system_u:system_r:rpcd_t:s0 key=(null) type=AVC msg=audit(05/28/2017 14:12:48.080:76) : avc: denied { read } for pid=1992 comm=rpc.statd name=state dev="fuse" ino=12362789396445498341 scontext=system_u:system_r:rpcd_t:s0 tcontext=system_u:object_r:fusefs_t:s0 tclass=file ---- type=SYSCALL msg=audit(05/28/2017 14:17:37.177:26) : arch=x86_64 syscall=bind success=no exit=EACCES(Permission denied) a0=0xf a1=0x7ffdfa768c70 a2=0x10 a3=0x22 items=0 ppid=1155 pid=1162 auid=unset uid=root gid=root euid=root suid=root fsuid=root egid=root sgid=root fsgid=root tty=(none) ses=unset comm=glusterd exe=/usr/sbin/glusterfsd subj=system_u:system_r:glusterd_t:s0 key=(null) type=AVC msg=audit(05/28/2017 14:17:37.177:26) : avc: denied { name_bind } for pid=1162 comm=glusterd src=61000 scontext=system_u:system_r:glusterd_t:s0 tcontext=system_u:object_r:ephemeral_port_t:s0 tclass=tcp_socket ---- type=SYSCALL msg=audit(05/28/2017 14:17:46.401:56) : arch=x86_64 syscall=kill success=no exit=EACCES(Permission denied) a0=0x560 a1=SIGKILL a2=0x7fd684000078 a3=0x0 items=0 ppid=1 pid=1167 auid=unset uid=root gid=root euid=root suid=root fsuid=root egid=root sgid=root fsgid=root tty=(none) ses=unset comm=glusterd exe=/usr/sbin/glusterfsd subj=system_u:system_r:glusterd_t:s0 key=(null) type=AVC msg=audit(05/28/2017 14:17:46.401:56) : avc: denied { sigkill } for pid=1167 comm=glusterd scontext=system_u:system_r:glusterd_t:s0 tcontext=system_u:system_r:cluster_t:s0 tclass=process ---- type=SYSCALL msg=audit(05/28/2017 14:17:45.400:55) : arch=x86_64 syscall=kill success=no exit=EACCES(Permission denied) a0=0x560 a1=SIGTERM a2=0x7fd684000038 a3=0x99 items=0 ppid=1 pid=1167 auid=unset uid=root gid=root euid=root suid=root fsuid=root egid=root sgid=root fsgid=root tty=(none) ses=unset comm=glusterd exe=/usr/sbin/glusterfsd subj=system_u:system_r:glusterd_t:s0 key=(null) type=AVC msg=audit(05/28/2017 14:17:45.400:55) : avc: denied { signal } for pid=1167 comm=glusterd scontext=system_u:system_r:glusterd_t:s0 tcontext=system_u:system_r:cluster_t:s0 tclass=process ---- type=SYSCALL msg=audit(05/28/2017 14:18:56.024:67) : arch=x86_64 syscall=openat success=no exit=EACCES(Permission denied) a0=0xffffffffffffff9c a1=0x7ff662e9be10 a2=O_RDONLY|O_NONBLOCK|O_DIRECTORY|O_CLOEXEC a3=0x0 items=0 ppid=1949 pid=1950 auid=unset uid=root gid=root euid=root suid=root fsuid=root egid=root sgid=root fsgid=root tty=(none) ses=unset comm=rpc.statd exe=/usr/sbin/rpc.statd subj=system_u:system_r:rpcd_t:s0 key=(null) type=AVC msg=audit(05/28/2017 14:18:56.024:67) : avc: denied { read } for pid=1950 comm=rpc.statd name=sm dev="fuse" ino=12866274077597183313 scontext=system_u:system_r:rpcd_t:s0 tcontext=system_u:object_r:fusefs_t:s0 tclass=dir ---- type=SYSCALL msg=audit(05/28/2017 14:18:56.034:68) : arch=x86_64 syscall=open success=no exit=EACCES(Permission denied) a0=0x7ff662e9bdd0 a1=O_RDONLY a2=0x7ff662e9bde8 a3=0x5 items=0 ppid=1949 pid=1950 auid=unset uid=root gid=root euid=root suid=root fsuid=root egid=root sgid=root fsgid=root tty=(none) ses=unset comm=rpc.statd exe=/usr/sbin/rpc.statd subj=system_u:system_r:rpcd_t:s0 key=(null) type=AVC msg=audit(05/28/2017 14:18:56.034:68) : avc: denied { read } for pid=1950 comm=rpc.statd name=state dev="fuse" ino=12362789396445498341 scontext=system_u:system_r:rpcd_t:s0 tcontext=system_u:object_r:fusefs_t:s0 tclass=file On Mon, May 15, 2017 at 11:56 AM, Soumya Koduri <skoduri@xxxxxxxxxx> wrote: > > > On 05/12/2017 06:27 PM, Adam Ru wrote: >> >> Hi Soumya, >> >> Thank you very much for last response – very useful. >> >> I apologize for delay, I had to find time for another testing. >> >> I updated instructions that I provided in previous e-mail. *** means >> that the step was added. >> >> Instructions: >> - Clean installation of CentOS 7.3 with all updates, 3x node, >> resolvable IPs and VIPs >> - Stopped firewalld (just for testing) >> - *** SELinux in permissive mode (I had to, will explain bellow) >> - Install "centos-release-gluster" to get "centos-gluster310" repo >> and install following (nothing else): >> --- glusterfs-server >> --- glusterfs-ganesha >> - Passwordless SSH between all nodes >> (/var/lib/glusterd/nfs/secret.pem and secret.pem.pub on all nodes) >> - systemctl enable and start glusterd >> - gluster peer probe <other nodes> >> - gluster volume set all cluster.enable-shared-storage enable >> - systemctl enable and start pcsd.service >> - systemctl enable pacemaker.service (cannot be started at this moment) >> - Set password for hacluster user on all nodes >> - pcs cluster auth <node 1> <node 2> <node 3> -u hacluster -p blabla >> - mkdir /var/run/gluster/shared_storage/nfs-ganesha/ >> - touch /var/run/gluster/shared_storage/nfs-ganesha/ganesha.conf (not >> sure if needed) >> - vi /var/run/gluster/shared_storage/nfs-ganesha/ganesha-ha.conf and >> insert configuration >> - Try list files on other nodes: ls >> /var/run/gluster/shared_storage/nfs-ganesha/ >> - gluster nfs-ganesha enable >> - *** systemctl enable pacemaker.service (again, since pacemaker was >> disabled at this point) >> - *** Check owner of "state", "statd", "sm" and "sm.bak" in >> /var/lib/nfs/ (I had to: chown rpcuser:rpcuser >> /var/lib/nfs/statd/state) >> - Check on other nodes that nfs-ganesha.service is running and "pcs >> status" shows started resources >> - gluster volume create mynewshare replica 3 transport tcp >> node1:/<dir> node2:/<dir> node3:/<dir> >> - gluster volume start mynewshare >> - gluster vol set mynewshare ganesha.enable on >> >> At this moment, this is status of important (I think) services: >> >> -- corosync.service disabled >> -- corosync-notifyd.service disabled >> -- glusterd.service enabled >> -- glusterfsd.service disabled >> -- pacemaker.service enabled >> -- pcsd.service enabled >> -- nfs-ganesha.service disabled >> -- nfs-ganesha-config.service static >> -- nfs-ganesha-lock.service static >> >> -- corosync.service active (running) >> -- corosync-notifyd.service inactive (dead) >> -- glusterd.service active (running) >> -- glusterfsd.service inactive (dead) >> -- pacemaker.service active (running) >> -- pcsd.service active (running) >> -- nfs-ganesha.service active (running) >> -- nfs-ganesha-config.service inactive (dead) >> -- nfs-ganesha-lock.service active (running) >> >> May I ask you a few questions please? >> >> 1. Could you please confirm that services above has correct status/state? > > > Looks good to the best of my knowledge. > >> >> 2. When I restart a node then nfs-ganesha is not running. Of course I >> cannot enable it since it needs to be enabled after shared storage is >> mounted. What is best practice to start it automatically so I don’t >> have to worry about restarting node? Should I create a script that >> will check whether shared storage was mounted and then start >> nfs-ganesha? How do you do this in production? > > > That's right.. We have plans to address this in near future (probably by > having a new .service which mounts shared_storage before starting > nfs-ganesha). But until then ..yes having a custom defined script to do so > is the only way to automate it. > > >> >> 3. SELinux is an issue, is that a known bug? >> >> When I restart a node and start nfs-ganesha.service with SELinux in >> permissive mode: >> >> sudo grep 'statd' /var/log/messages >> May 12 12:05:46 mynode1 rpc.statd[2415]: Version 1.3.0 starting >> May 12 12:05:46 mynode1 rpc.statd[2415]: Flags: TI-RPC >> May 12 12:05:46 mynode1 rpc.statd[2415]: Failed to read >> /var/lib/nfs/statd/state: Success >> May 12 12:05:46 mynode1 rpc.statd[2415]: Initializing NSM state >> May 12 12:05:52 mynode1 rpc.statd[2415]: Received SM_UNMON_ALL request >> from mynode1.localdomain while not monitoring any hosts >> >> systemctl status nfs-ganesha-lock.service --full >> ● nfs-ganesha-lock.service - NFS status monitor for NFSv2/3 locking. >> Loaded: loaded (/usr/lib/systemd/system/nfs-ganesha-lock.service; >> static; vendor preset: disabled) >> Active: active (running) since Fri 2017-05-12 12:05:46 UTC; 1min 43s >> ago >> Process: 2414 ExecStart=/usr/sbin/rpc.statd --no-notify $STATDARGS >> (code=exited, status=0/SUCCESS) >> Main PID: 2415 (rpc.statd) >> CGroup: /system.slice/nfs-ganesha-lock.service >> └─2415 /usr/sbin/rpc.statd --no-notify >> >> May 12 12:05:46 mynode1.localdomain systemd[1]: Starting NFS status >> monitor for NFSv2/3 locking.... >> May 12 12:05:46 mynode1.localdomain rpc.statd[2415]: Version 1.3.0 >> starting >> May 12 12:05:46 mynode1.localdomain rpc.statd[2415]: Flags: TI-RPC >> May 12 12:05:46 mynode1.localdomain rpc.statd[2415]: Failed to read >> /var/lib/nfs/statd/state: Success >> May 12 12:05:46 mynode1.localdomain rpc.statd[2415]: Initializing NSM >> state >> May 12 12:05:46 mynode1.localdomain systemd[1]: Started NFS status >> monitor for NFSv2/3 locking.. >> May 12 12:05:52 mynode1.localdomain rpc.statd[2415]: Received >> SM_UNMON_ALL request from mynode1.localdomain while not monitoring any >> hosts >> >> >> When I restart a node and start nfs-ganesha.service with SELinux in >> enforcing mode: >> >> >> sudo grep 'statd' /var/log/messages >> May 12 12:14:01 mynode1 rpc.statd[1743]: Version 1.3.0 starting >> May 12 12:14:01 mynode1 rpc.statd[1743]: Flags: TI-RPC >> May 12 12:14:01 mynode1 rpc.statd[1743]: Failed to open directory sm: >> Permission denied >> May 12 12:14:01 mynode1 rpc.statd[1743]: Failed to open >> /var/lib/nfs/statd/state: Permission denied >> >> systemctl status nfs-ganesha-lock.service --full >> ● nfs-ganesha-lock.service - NFS status monitor for NFSv2/3 locking. >> Loaded: loaded (/usr/lib/systemd/system/nfs-ganesha-lock.service; >> static; vendor preset: disabled) >> Active: failed (Result: exit-code) since Fri 2017-05-12 12:14:01 >> UTC; 1min 21s ago >> Process: 1742 ExecStart=/usr/sbin/rpc.statd --no-notify $STATDARGS >> (code=exited, status=1/FAILURE) >> >> May 12 12:14:01 mynode1.localdomain systemd[1]: Starting NFS status >> monitor for NFSv2/3 locking.... >> May 12 12:14:01 mynode1.localdomain rpc.statd[1743]: Version 1.3.0 >> starting >> May 12 12:14:01 mynode1.localdomain rpc.statd[1743]: Flags: TI-RPC >> May 12 12:14:01 mynode1.localdomain rpc.statd[1743]: Failed to open >> directory sm: Permission denied >> May 12 12:14:01 mynode1.localdomain systemd[1]: >> nfs-ganesha-lock.service: control process exited, code=exited status=1 >> May 12 12:14:01 mynode1.localdomain systemd[1]: Failed to start NFS >> status monitor for NFSv2/3 locking.. >> May 12 12:14:01 mynode1.localdomain systemd[1]: Unit >> nfs-ganesha-lock.service entered failed state. >> May 12 12:14:01 mynode1.localdomain systemd[1]: nfs-ganesha-lock.service >> failed. > > > Cant remember right now. Could you please paste the AVCs you get, and > se-linux packages version. Or preferably please file a bug. We can get the > details verified from selinux members. > > Thanks, > Soumya > > >> >> On Fri, May 5, 2017 at 8:10 PM, Soumya Koduri <skoduri@xxxxxxxxxx> wrote: >>> >>> >>> >>> On 05/05/2017 08:04 PM, Adam Ru wrote: >>>> >>>> >>>> Hi Soumya, >>>> >>>> Thank you for the answer. >>>> >>>> Enabling Pacemaker? Yes, you’re completely right, I didn’t do it. Thank >>>> you. >>>> >>>> I spent some time by testing and I have some results. This is what I >>>> did: >>>> >>>> - Clean installation of CentOS 7.3 with all updates, 3x node, >>>> resolvable IPs and VIPs >>>> - Stopped firewalld (just for testing) >>>> - Install "centos-release-gluster" to get "centos-gluster310" repo and >>>> install following (nothing else): >>>> --- glusterfs-server >>>> --- glusterfs-ganesha >>>> - Passwordless SSH between all nodes (/var/lib/glusterd/nfs/secret.pem >>>> and secret.pem.pub on all nodes) >>>> - systemctl enable and start glusterd >>>> - gluster peer probe <other nodes> >>>> - gluster volume set all cluster.enable-shared-storage enable >>>> - systemctl enable and start pcsd.service >>>> - systemctl enable pacemaker.service (cannot be started at this moment) >>>> - Set password for hacluster user on all nodes >>>> - pcs cluster auth <node 1> <node 2> <node 3> -u hacluster -p blabla >>>> - mkdir /var/run/gluster/shared_storage/nfs-ganesha/ >>>> - touch /var/run/gluster/shared_storage/nfs-ganesha/ganesha.conf (not >>>> sure if needed) >>>> - vi /var/run/gluster/shared_storage/nfs-ganesha/ganesha-ha.conf and >>>> insert configuration >>>> - Try list files on other nodes: ls >>>> /var/run/gluster/shared_storage/nfs-ganesha/ >>>> - gluster nfs-ganesha enable >>>> - Check on other nodes that nfs-ganesha.service is running and "pcs >>>> status" shows started resources >>>> - gluster volume create mynewshare replica 3 transport tcp node1:/<dir> >>>> node2:/<dir> node3:/<dir> >>>> - gluster volume start mynewshare >>>> - gluster vol set mynewshare ganesha.enable on >>>> >>>> After these steps, all VIPs are pingable and I can mount >>>> node1:/mynewshare >>>> >>>> Funny thing is that pacemaker.service is disabled again (something >>>> disabled it). This is status of important (I think) services: >>> >>> >>> >>> yeah. We too had observed this recently. We guess probably pcs cluster >>> setup >>> command first destroys existing cluster (if any) which may be disabling >>> pacemaker too. >>> >>>> >>>> systemctl list-units --all >>>> # corosync.service loaded active running >>>> # glusterd.service loaded active running >>>> # nfs-config.service loaded inactive dead >>>> # nfs-ganesha-config.service loaded inactive dead >>>> # nfs-ganesha-lock.service loaded active running >>>> # nfs-ganesha.service loaded active running >>>> # nfs-idmapd.service loaded inactive dead >>>> # nfs-mountd.service loaded inactive dead >>>> # nfs-server.service loaded inactive dead >>>> # nfs-utils.service loaded inactive dead >>>> # pacemaker.service loaded active running >>>> # pcsd.service loaded active running >>>> >>>> systemctl list-unit-files --all >>>> # corosync-notifyd.service disabled >>>> # corosync.service disabled >>>> # glusterd.service enabled >>>> # glusterfsd.service disabled >>>> # nfs-blkmap.service disabled >>>> # nfs-config.service static >>>> # nfs-ganesha-config.service static >>>> # nfs-ganesha-lock.service static >>>> # nfs-ganesha.service disabled >>>> # nfs-idmap.service static >>>> # nfs-idmapd.service static >>>> # nfs-lock.service static >>>> # nfs-mountd.service static >>>> # nfs-rquotad.service disabled >>>> # nfs-secure-server.service static >>>> # nfs-secure.service static >>>> # nfs-server.service disabled >>>> # nfs-utils.service static >>>> # nfs.service disabled >>>> # nfslock.service static >>>> # pacemaker.service disabled >>>> # pcsd.service enabled >>>> >>>> I enabled pacemaker again on all nodes and restart all nodes one by one. >>>> >>>> After reboot all VIPs are gone and I can see that nfs-ganesha.service >>>> isn’t running. When I start it on at least two nodes then VIPs are >>>> pingable again and I can mount NFS again. But there is still some issue >>>> in the setup because when I check nfs-ganesha-lock.service I get: >>>> >>>> systemctl -l status nfs-ganesha-lock.service >>>> ● nfs-ganesha-lock.service - NFS status monitor for NFSv2/3 locking. >>>> Loaded: loaded (/usr/lib/systemd/system/nfs-ganesha-lock.service; >>>> static; vendor preset: disabled) >>>> Active: failed (Result: exit-code) since Fri 2017-05-05 13:43:37 UTC; >>>> 31min ago >>>> Process: 6203 ExecStart=/usr/sbin/rpc.statd --no-notify $STATDARGS >>>> (code=exited, status=1/FAILURE) >>>> >>>> May 05 13:43:37 node0.localdomain systemd[1]: Starting NFS status >>>> monitor for NFSv2/3 locking.... >>>> May 05 13:43:37 node0.localdomain rpc.statd[6205]: Version 1.3.0 >>>> starting >>>> May 05 13:43:37 node0.localdomain rpc.statd[6205]: Flags: TI-RPC >>>> May 05 13:43:37 node0.localdomain rpc.statd[6205]: Failed to open >>>> directory sm: Permission denied >>> >>> >>> >>> Okay this issue was fixed and the fix should be present in 3.10 too - >>> https://review.gluster.org/#/c/16433/ >>> >>> Please check '/var/log/messages' for statd related errors and cross-check >>> permissions of that directory. You could manually chown owner:group of >>> /var/lib/nfs/statd/sm directory for now and then restart nfs-ganesha* >>> services. >>> >>> Thanks, >>> Soumya >>> >>>> May 05 13:43:37 node0.localdomain rpc.statd[6205]: Failed to open >>>> /var/lib/nfs/statd/state: Permission denied >>>> May 05 13:43:37 node0.localdomain systemd[1]: nfs-ganesha-lock.service: >>>> control process exited, code=exited status=1 >>>> May 05 13:43:37 node0.localdomain systemd[1]: Failed to start NFS status >>>> monitor for NFSv2/3 locking.. >>>> May 05 13:43:37 node0.localdomain systemd[1]: Unit >>>> nfs-ganesha-lock.service entered failed state. >>>> May 05 13:43:37 node0.localdomain systemd[1]: nfs-ganesha-lock.service >>>> failed. >>>> >>>> Thank you, >>>> >>>> Kind regards, >>>> >>>> Adam >>>> >>>> On Wed, May 3, 2017 at 10:32 AM, Mahdi Adnan <mahdi.adnan@xxxxxxxxxxx >>>> <mailto:mahdi.adnan@xxxxxxxxxxx>> wrote: >>>> >>>> Hi, >>>> >>>> >>>> Same here, when i reboot the node i have to manually execute "pcs >>>> cluster start gluster01" and pcsd already enabled and started. >>>> >>>> Gluster 3.8.11 >>>> >>>> Centos 7.3 latest >>>> >>>> Installed using CentOS Storage SIG repository >>>> >>>> >>>> >>>> -- >>>> >>>> Respectfully* >>>> **Mahdi A. Mahdi* >>>> >>>> >>>> ------------------------------------------------------------------------ >>>> *From:* gluster-users-bounces@xxxxxxxxxxx >>>> <mailto:gluster-users-bounces@xxxxxxxxxxx> >>>> <gluster-users-bounces@xxxxxxxxxxx >>>> <mailto:gluster-users-bounces@xxxxxxxxxxx>> on behalf of Adam Ru >>>> <ad.ruckel@xxxxxxxxx <mailto:ad.ruckel@xxxxxxxxx>> >>>> *Sent:* Wednesday, May 3, 2017 12:09:58 PM >>>> *To:* Soumya Koduri >>>> *Cc:* gluster-users@xxxxxxxxxxx <mailto:gluster-users@xxxxxxxxxxx> >>>> *Subject:* Re: Gluster and NFS-Ganesha - cluster is >>>> >>>> down after reboot >>>> >>>> Hi Soumya, >>>> >>>> thank you very much for your reply. >>>> >>>> I enabled pcsd during setup and after reboot during troubleshooting >>>> I manually started it and checked resources (pcs status). They were >>>> not running. I didn’t find what was wrong but I’m going to try it >>>> again. >>>> >>>> I’ve thoroughly checked >>>> >>>> >>>> http://gluster.readthedocs.io/en/latest/Administrator%20Guide/NFS-Ganesha%20GlusterFS%20Integration/ >>>> >>>> >>>> <http://gluster.readthedocs.io/en/latest/Administrator%20Guide/NFS-Ganesha%20GlusterFS%20Integration/> >>>> and I can confirm that I followed all steps with one exception. I >>>> installed following RPMs: >>>> glusterfs-server >>>> glusterfs-fuse >>>> glusterfs-cli >>>> glusterfs-ganesha >>>> nfs-ganesha-xfs >>>> >>>> and the guide referenced above specifies: >>>> glusterfs-server >>>> glusterfs-api >>>> glusterfs-ganesha >>>> >>>> glusterfs-api is a dependency of one of RPMs that I installed so >>>> this is not a problem. But I cannot find any mention to install >>>> nfs-ganesha-xfs. >>>> >>>> I’ll try to setup the whole environment again without installing >>>> nfs-ganesha-xfs (I assume glusterfs-ganesha has all required >>>> binaries). >>>> >>>> Again, thank you for you time to answer my previous message. >>>> >>>> Kind regards, >>>> Adam >>>> >>>> On Tue, May 2, 2017 at 8:49 AM, Soumya Koduri <skoduri@xxxxxxxxxx >>>> <mailto:skoduri@xxxxxxxxxx>> wrote: >>>> >>>> Hi, >>>> >>>> On 05/02/2017 01:34 AM, Rudolf wrote: >>>> >>>> Hi Gluster users, >>>> >>>> First, I'd like to thank you all for this amazing >>>> open-source! Thank you! >>>> >>>> I'm working on home project – three servers with Gluster and >>>> NFS-Ganesha. My goal is to create HA NFS share with three >>>> copies of each >>>> file on each server. >>>> >>>> My systems are CentOS 7.3 Minimal install with the latest >>>> updates and >>>> the most current RPMs from "centos-gluster310" repository. >>>> >>>> I followed this tutorial: >>>> >>>> >>>> http://blog.gluster.org/2015/10/linux-scale-out-nfsv4-using-nfs-ganesha-and-glusterfs-one-step-at-a-time/ >>>> >>>> >>>> <http://blog.gluster.org/2015/10/linux-scale-out-nfsv4-using-nfs-ganesha-and-glusterfs-one-step-at-a-time/> >>>> (second half that describes multi-node HA setup) >>>> >>>> with a few exceptions: >>>> >>>> 1. All RPMs are from "centos-gluster310" repo that is >>>> installed by "yum >>>> -y install centos-release-gluster" >>>> 2. I have three nodes (not four) with "replica 3" volume. >>>> 3. I created empty ganesha.conf and not empty >>>> ganesha-ha.conf >>>> in >>>> "/var/run/gluster/shared_storage/nfs-ganesha/" (referenced >>>> blog post is >>>> outdated, this is now requirement) >>>> 4. ganesha-ha.conf doesn't have "HA_VOL_SERVER" since this >>>> isn't needed >>>> anymore. >>>> >>>> >>>> Please refer to >>>> >>>> >>>> http://gluster.readthedocs.io/en/latest/Administrator%20Guide/NFS-Ganesha%20GlusterFS%20Integration/ >>>> >>>> >>>> <http://gluster.readthedocs.io/en/latest/Administrator%20Guide/NFS-Ganesha%20GlusterFS%20Integration/> >>>> >>>> It is being updated with latest changes happened wrt setup. >>>> >>>> When I finish configuration, all is good. >>>> nfs-ganesha.service is active >>>> and running and from client I can ping all three VIPs and I >>>> can mount >>>> NFS. Copied files are replicated to all nodes. >>>> >>>> But when I restart nodes (one by one, with 5 min. delay >>>> between) then I >>>> cannot ping or mount (I assume that all VIPs are down). So >>>> my setup >>>> definitely isn't HA. >>>> >>>> I found that: >>>> # pcs status >>>> Error: cluster is not currently running on this node >>>> >>>> >>>> This means pcsd service is not up. Did you enable (systemctl >>>> enable pcsd) pcsd service so that is comes up post reboot >>>> automatically. If not please start it manually. >>>> >>>> >>>> and nfs-ganesha.service is in inactive state. Btw. I didn't >>>> enable >>>> "systemctl enable nfs-ganesha" since I assume that this is >>>> something >>>> that Gluster does. >>>> >>>> >>>> Please check /var/log/ganesha.log for any errors/warnings. >>>> >>>> We recommend not to enable nfs-ganesha.service (by default), as >>>> the shared storage (where the ganesha.conf file resides now) >>>> should be up and running before nfs-ganesha gets started. >>>> So if enabled by default it could happen that shared_storage >>>> mount point is not yet up and it resulted in nfs-ganesha service >>>> failure. If you would like to address this, you could have a >>>> cron job which keeps checking the mount point health and then >>>> start nfs-ganesha service. >>>> >>>> Thanks, >>>> Soumya >>>> >>>> >>>> I assume that my issue is that I followed instructions in >>>> blog post from >>>> 2015/10 that are outdated. Unfortunately I cannot find >>>> anything better – >>>> I spent whole day by googling. >>>> >>>> Would you be so kind and check the instructions in blog post >>>> and let me >>>> know what steps are wrong / outdated? Or please do you have >>>> more current >>>> instructions for Gluster+Ganesha setup? >>>> >>>> Thank you. >>>> >>>> Kind regards, >>>> Adam >>>> >>>> >>>> >>>> _______________________________________________ >>>> Gluster-users mailing list >>>> Gluster-users@xxxxxxxxxxx <mailto:Gluster-users@xxxxxxxxxxx> >>>> http://lists.gluster.org/mailman/listinfo/gluster-users >>>> <http://lists.gluster.org/mailman/listinfo/gluster-users> >>>> >>>> >>>> >>>> >>>> -- >>>> Adam >>>> >>>> >>>> >>>> >>>> -- >>>> Adam >> >> >> >> > -- Adam _______________________________________________ Gluster-users mailing list Gluster-users@xxxxxxxxxxx http://lists.gluster.org/mailman/listinfo/gluster-users