Hi hvjunk >From the logs it seems like the ganesha cluster is not yet created/running and performing refresh config is failing.As the cluster is not running and volume "ganesha" is not yet exported due to this performing refresh config on volume will fail. For creating ganesha cluster,there are few pre-requisite of adding services like nlm,nfs,rpc-bind,high-availability,mountd,rquota to firewalld. After executing create-cluster block ,It will validate the HA Status of ganesha via gdeploy.If the ganesha cluster is up and running it will reflect HA status as "HEALTHY". After then performing refresh-config on nfs exported volume will succeed. Attaching the sample config file to create ganesha cluster and exporting ganesha volume to client - [hosts] dhcp37-102.lab.eng.blr.redhat.com dhcp37-92.lab.eng.blr.redhat.com dhcp37-119.lab.eng.blr.redhat.com dhcp37-122.lab.eng.blr.redhat.com [backend-setup] devices=/dev/sdb,/dev/sdc,/dev/sdd vgs=vg1,vg2,vg3 pools=pool1,pool2,pool3 lvs=lv1,lv2,lv3 mountpoints=/gluster/brick1,/gluster/brick2,/gluster/brick3 brick_dirs=/gluster/brick1/1,/gluster/brick2/1,/gluster/brick3/1 [yum] action=install repolist= gpgcheck=no update=no packages=glusterfs-ganesha [firewalld] action=add ports=111/tcp,2049/tcp,54321/tcp,5900/tcp,5900-6923/tcp,5666/tcp,16514/tcp services=glusterfs,nlm,nfs,rpc-bind,high-availability,mountd,rquota [volume] action=create volname=ganesha transport=tcp replica_count=2 force=yes [nfs-ganesha] action=create-cluster ha-name=ganesha-ha-360 cluster-nodes=dhcp37-102.lab.eng.blr.redhat.com,dhcp37-92.lab.eng.blr.redhat.com,dhcp37-119.lab.eng.blr.redhat.com,dhcp37-122.lab.eng.blr.redhat.com vip=10.70.36.217,10.70.36.218,10.70.36.219,10.70.36.220 volname=ganesha #This will export volume via nfs-ganesha [clients] action=mount volname=10.70.37.153:/ganesha #VIP from which the volume needs to be mounted to client hosts=10.70.46.30 #Client IP client_mount_points=/mnt/ganesha fstype=nfs options=rw,relatime,vers=4.0,rsize=1048576,wsize=1048576,namlen=255,hard,proto=tcp,port=0,timeo=600,retrans=2,sec=sys,clientaddr=10.70.44.153,local_lock=none,addr=10.70.46.30 Regards, Manisha Saini ----- Original Message ----- From: "Soumya Koduri" <skoduri@xxxxxxxxxx> To: "hvjunk" <hvjunk@xxxxxxxxx>, gluster-users@xxxxxxxxxxx, "Sachidananda URS" <surs@xxxxxxxxxx>, "Manisha Saini" <msaini@xxxxxxxxxx>, "Arthy Loganathan" <aloganat@xxxxxxxxxx> Sent: Wednesday, May 10, 2017 10:45:57 AM Subject: Re: gdeploy not starting all the daemons for NFS-ganesha :( CCin Sac, Manisha, Arthy who could help with troubleshooting. Thanks, Soumya On 05/09/2017 08:31 PM, hvjunk wrote: > Hi there, > > Given the following config file, what am I doing wrong causing the error at the bottom & no mounted /gluster_shared_storage? > > Hendrik > > [root@linked-clone-of-centos-linux ~]# cat t.conf > [hosts] > 10.10.10.11 > 10.10.10.12 > 10.10.10.13 > > [backend-setup] > devices=/dev/sdb > mountpoints=/gluster/brick1 > brick_dirs=/gluster/brick1/one > pools=pool1 > > #Installing nfs-ganesha > [yum] > action=install > repolist= > gpgcheck=no > update=no > packages=glusterfs-ganesha > > [service] > action=start > service=glusterd > > [service] > action=enable > service=glusterd > > #This will create a volume. Skip this section if your volume already exists > [volume] > action=create > volname=ganesha > transport=tcp > replica_count=3 > arbiter_count=1 > force=yes > > [clients] > action=mount > volname=ganesha > hosts=10.10.10.11,10.10.10.12,10.10.10.13 > fstype=glusterfs > client_mount_points=/mnt/ganesha_mnt > > > #Creating a high availability cluster and exporting the volume > [nfs-ganesha] > action=create-cluster > ha-name=ganesha-ha-360 > cluster-nodes=10.10.10.11,10.10.10.12 > vip=10.10.10.31,10.10.10.41 > volname=ganesha > > > [nfs-ganesha] > action=export-volume > volname=ganesha > > [nfs-ganesha] > action=refresh-config > volname=ganesha > [root@linked-clone-of-centos-linux ~]# > > > [root@linked-clone-of-centos-linux ~]# gdeploy -c t.conf -k > > PLAY [gluster_servers] ************************************************************************************* > > TASK [Clean up filesystem signature] *********************************************************************** > skipping: [10.10.10.11] => (item=/dev/sdb) > skipping: [10.10.10.12] => (item=/dev/sdb) > skipping: [10.10.10.13] => (item=/dev/sdb) > > TASK [Create Physical Volume] ****************************************************************************** > changed: [10.10.10.13] => (item=/dev/sdb) > changed: [10.10.10.11] => (item=/dev/sdb) > changed: [10.10.10.12] => (item=/dev/sdb) > > PLAY RECAP ************************************************************************************************* > 10.10.10.11 : ok=1 changed=1 unreachable=0 failed=0 > 10.10.10.12 : ok=1 changed=1 unreachable=0 failed=0 > 10.10.10.13 : ok=1 changed=1 unreachable=0 failed=0 > > > PLAY [gluster_servers] ************************************************************************************* > > TASK [Create volume group on the disks] ******************************************************************** > changed: [10.10.10.12] => (item={u'brick': u'/dev/sdb', u'vg': u'GLUSTER_vg1'}) > changed: [10.10.10.13] => (item={u'brick': u'/dev/sdb', u'vg': u'GLUSTER_vg1'}) > changed: [10.10.10.11] => (item={u'brick': u'/dev/sdb', u'vg': u'GLUSTER_vg1'}) > > PLAY RECAP ************************************************************************************************* > 10.10.10.11 : ok=1 changed=1 unreachable=0 failed=0 > 10.10.10.12 : ok=1 changed=1 unreachable=0 failed=0 > 10.10.10.13 : ok=1 changed=1 unreachable=0 failed=0 > > > PLAY [gluster_servers] ************************************************************************************* > > TASK [Create logical volume named metadata] **************************************************************** > changed: [10.10.10.12] => (item=GLUSTER_vg1) > changed: [10.10.10.13] => (item=GLUSTER_vg1) > changed: [10.10.10.11] => (item=GLUSTER_vg1) > > TASK [create data LV that has a size which is a multiple of stripe width] ********************************** > changed: [10.10.10.13] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', u'vg': u'GLUSTER_vg1'}) > changed: [10.10.10.11] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', u'vg': u'GLUSTER_vg1'}) > changed: [10.10.10.12] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', u'vg': u'GLUSTER_vg1'}) > > TASK [Convert the logical volume] ************************************************************************** > changed: [10.10.10.11] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', u'vg': u'GLUSTER_vg1'}) > changed: [10.10.10.13] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', u'vg': u'GLUSTER_vg1'}) > changed: [10.10.10.12] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', u'vg': u'GLUSTER_vg1'}) > > TASK [create stripe-aligned thin volume] ******************************************************************* > changed: [10.10.10.13] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', u'vg': u'GLUSTER_vg1'}) > changed: [10.10.10.11] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', u'vg': u'GLUSTER_vg1'}) > changed: [10.10.10.12] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', u'vg': u'GLUSTER_vg1'}) > > TASK [Change the attributes of the logical volume] ********************************************************* > changed: [10.10.10.11] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', u'vg': u'GLUSTER_vg1'}) > changed: [10.10.10.13] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', u'vg': u'GLUSTER_vg1'}) > changed: [10.10.10.12] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', u'vg': u'GLUSTER_vg1'}) > > PLAY RECAP ************************************************************************************************* > 10.10.10.11 : ok=5 changed=5 unreachable=0 failed=0 > 10.10.10.12 : ok=5 changed=5 unreachable=0 failed=0 > 10.10.10.13 : ok=5 changed=5 unreachable=0 failed=0 > > > PLAY [gluster_servers] ************************************************************************************* > > TASK [Create an xfs filesystem] **************************************************************************** > changed: [10.10.10.13] => (item=/dev/GLUSTER_vg1/GLUSTER_lv1) > changed: [10.10.10.12] => (item=/dev/GLUSTER_vg1/GLUSTER_lv1) > changed: [10.10.10.11] => (item=/dev/GLUSTER_vg1/GLUSTER_lv1) > > PLAY RECAP ************************************************************************************************* > 10.10.10.11 : ok=1 changed=1 unreachable=0 failed=0 > 10.10.10.12 : ok=1 changed=1 unreachable=0 failed=0 > 10.10.10.13 : ok=1 changed=1 unreachable=0 failed=0 > > > PLAY [gluster_servers] ************************************************************************************* > > TASK [Create the backend disks, skips if present] ********************************************************** > changed: [10.10.10.13] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', u'path': u'/gluster/brick1'}) > changed: [10.10.10.12] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', u'path': u'/gluster/brick1'}) > changed: [10.10.10.11] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', u'path': u'/gluster/brick1'}) > > TASK [Mount the volumes] *********************************************************************************** > changed: [10.10.10.11] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', u'path': u'/gluster/brick1'}) > changed: [10.10.10.12] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', u'path': u'/gluster/brick1'}) > changed: [10.10.10.13] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', u'path': u'/gluster/brick1'}) > > PLAY RECAP ************************************************************************************************* > 10.10.10.11 : ok=2 changed=2 unreachable=0 failed=0 > 10.10.10.12 : ok=2 changed=2 unreachable=0 failed=0 > 10.10.10.13 : ok=2 changed=2 unreachable=0 failed=0 > > > PLAY [gluster_servers] ************************************************************************************* > > TASK [Set SELinux labels on the bricks] ******************************************************************** > changed: [10.10.10.11] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', u'path': u'/gluster/brick1'}) > changed: [10.10.10.12] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', u'path': u'/gluster/brick1'}) > changed: [10.10.10.13] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', u'path': u'/gluster/brick1'}) > > TASK [Restore the SELinux context] ************************************************************************* > changed: [10.10.10.11] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', u'path': u'/gluster/brick1'}) > changed: [10.10.10.12] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', u'path': u'/gluster/brick1'}) > changed: [10.10.10.13] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', u'path': u'/gluster/brick1'}) > > PLAY RECAP ************************************************************************************************* > 10.10.10.11 : ok=2 changed=2 unreachable=0 failed=0 > 10.10.10.12 : ok=2 changed=2 unreachable=0 failed=0 > 10.10.10.13 : ok=2 changed=2 unreachable=0 failed=0 > > > PLAY [gluster_servers] ************************************************************************************* > > TASK [Create/Enabling yum repos] *************************************************************************** > skipping: [10.10.10.12] => (item=) > skipping: [10.10.10.11] => (item=) > skipping: [10.10.10.13] => (item=) > > TASK [Clean up the metadata] ******************************************************************************* > skipping: [10.10.10.12] > skipping: [10.10.10.11] > skipping: [10.10.10.13] > > TASK [Yum update] ****************************************************************************************** > skipping: [10.10.10.12] > skipping: [10.10.10.11] > skipping: [10.10.10.13] > > TASK [Installs/Removes a package using yum] **************************************************************** > changed: [10.10.10.12] > changed: [10.10.10.13] > changed: [10.10.10.11] > > PLAY RECAP ************************************************************************************************* > 10.10.10.11 : ok=1 changed=1 unreachable=0 failed=0 > 10.10.10.12 : ok=1 changed=1 unreachable=0 failed=0 > 10.10.10.13 : ok=1 changed=1 unreachable=0 failed=0 > > > PLAY [gluster_servers] ************************************************************************************* > > TASK [Enable or disable services] ************************************************************************** > changed: [10.10.10.12] => (item=glusterd) > changed: [10.10.10.11] => (item=glusterd) > changed: [10.10.10.13] => (item=glusterd) > > PLAY RECAP ************************************************************************************************* > 10.10.10.11 : ok=1 changed=1 unreachable=0 failed=0 > 10.10.10.12 : ok=1 changed=1 unreachable=0 failed=0 > 10.10.10.13 : ok=1 changed=1 unreachable=0 failed=0 > > > PLAY [gluster_servers] ************************************************************************************* > > TASK [start/stop/restart/reload services] ****************************************************************** > changed: [10.10.10.11] => (item=glusterd) > changed: [10.10.10.12] => (item=glusterd) > changed: [10.10.10.13] => (item=glusterd) > > PLAY RECAP ************************************************************************************************* > 10.10.10.11 : ok=1 changed=1 unreachable=0 failed=0 > 10.10.10.12 : ok=1 changed=1 unreachable=0 failed=0 > 10.10.10.13 : ok=1 changed=1 unreachable=0 failed=0 > > > PLAY [gluster_servers] ************************************************************************************* > > TASK [Create the brick dirs, skips if present] ************************************************************* > changed: [10.10.10.11] => (item=/gluster/brick1/one) > changed: [10.10.10.12] => (item=/gluster/brick1/one) > changed: [10.10.10.13] => (item=/gluster/brick1/one) > > PLAY RECAP ************************************************************************************************* > 10.10.10.11 : ok=1 changed=1 unreachable=0 failed=0 > 10.10.10.12 : ok=1 changed=1 unreachable=0 failed=0 > 10.10.10.13 : ok=1 changed=1 unreachable=0 failed=0 > > > PLAY [master] ********************************************************************************************** > > TASK [Creates a Trusted Storage Pool] ********************************************************************** > changed: [10.10.10.11] > > TASK [Pause for a few seconds] ***************************************************************************** > Pausing for 5 seconds > (ctrl+C then 'C' = continue early, ctrl+C then 'A' = abort) > ok: [10.10.10.11] > > PLAY RECAP ************************************************************************************************* > 10.10.10.11 : ok=2 changed=1 unreachable=0 failed=0 > > > PLAY [master] ********************************************************************************************** > > TASK [Creates a volume] ************************************************************************************ > changed: [10.10.10.11] > > PLAY RECAP ************************************************************************************************* > 10.10.10.11 : ok=1 changed=1 unreachable=0 failed=0 > > > PLAY [master] ********************************************************************************************** > > TASK [Starts a volume] ************************************************************************************* > changed: [10.10.10.11] > > PLAY RECAP ************************************************************************************************* > 10.10.10.11 : ok=1 changed=1 unreachable=0 failed=0 > > > PLAY [clients] ********************************************************************************************* > > TASK [Create the dir to mount the volume, skips if present] ************************************************ > changed: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'}) > changed: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'}) > changed: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'}) > > PLAY RECAP ************************************************************************************************* > 10.10.10.11 : ok=1 changed=1 unreachable=0 failed=0 > 10.10.10.12 : ok=1 changed=1 unreachable=0 failed=0 > 10.10.10.13 : ok=1 changed=1 unreachable=0 failed=0 > > > PLAY [clients] ********************************************************************************************* > > TASK [Mount the volumes, if fstype is glusterfs] *********************************************************** > changed: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'}) > changed: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'}) > changed: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'}) > > PLAY RECAP ************************************************************************************************* > 10.10.10.11 : ok=1 changed=1 unreachable=0 failed=0 > 10.10.10.12 : ok=1 changed=1 unreachable=0 failed=0 > 10.10.10.13 : ok=1 changed=1 unreachable=0 failed=0 > > > PLAY [clients] ********************************************************************************************* > > TASK [Gathering Facts] ************************************************************************************* > ok: [10.10.10.11] > ok: [10.10.10.12] > ok: [10.10.10.13] > > TASK [Uncomment STATD_PORT for rpc.statd to listen on] ***************************************************** > skipping: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'}) > skipping: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'}) > skipping: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'}) > > TASK [Uncomment LOCKD_TCPPORT for rpc.lockd to listen on] ************************************************** > skipping: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'}) > skipping: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'}) > skipping: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'}) > > TASK [Uncomment LOCKD_UDPPORT for rpc.lockd to listen on] ************************************************** > skipping: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'}) > skipping: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'}) > skipping: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'}) > > TASK [Uncomment MOUNTD_PORT for rpc.mountd to listen on] *************************************************** > skipping: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'}) > skipping: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'}) > skipping: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'}) > > TASK [Restart nfs service (RHEL 6 only)] ******************************************************************* > skipping: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'}) > skipping: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'}) > skipping: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'}) > > TASK [Restart rpc-statd service] *************************************************************************** > skipping: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'}) > skipping: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'}) > skipping: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'}) > > TASK [Restart nfs-config service] ************************************************************************** > skipping: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'}) > skipping: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'}) > skipping: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'}) > > TASK [Restart nfs-mountd service] ************************************************************************** > skipping: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'}) > skipping: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'}) > skipping: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'}) > > TASK [Restart nfslock service (RHEL 6 & 7)] **************************************************************** > skipping: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'}) > skipping: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'}) > skipping: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'}) > > TASK [Mount the volumes if fstype is NFS] ****************************************************************** > skipping: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'}) > skipping: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'}) > skipping: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'}) > > PLAY RECAP ************************************************************************************************* > 10.10.10.11 : ok=1 changed=0 unreachable=0 failed=0 > 10.10.10.12 : ok=1 changed=0 unreachable=0 failed=0 > 10.10.10.13 : ok=1 changed=0 unreachable=0 failed=0 > > > PLAY [clients] ********************************************************************************************* > > TASK [Mount the volumes, if fstype is CIFS] **************************************************************** > skipping: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'}) > skipping: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'}) > skipping: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'}) > > PLAY RECAP ************************************************************************************************* > 10.10.10.11 : ok=0 changed=0 unreachable=0 failed=0 > 10.10.10.12 : ok=0 changed=0 unreachable=0 failed=0 > 10.10.10.13 : ok=0 changed=0 unreachable=0 failed=0 > > > PLAY [master] ********************************************************************************************** > > TASK [Delete lines from Ganesha export file] *************************************************************** > > TASK [Add lines to Ganesha export file] ******************************************************************** > > TASK [Update lines to Ganesha export file] ***************************************************************** > > TASK [Add a block to Ganesha export file] ****************************************************************** > > TASK [Refresh NFS-Ganesha config] ************************************************************************** > fatal: [10.10.10.11]: FAILED! => {"changed": true, "cmd": "/usr/libexec/ganesha/ganesha-ha.sh --refresh-config \"/var/run/gluster/shared_storage/nfs-ganesha\" \"ganesha\"", "delta": "0:00:00.227074", "end": "2017-05-09 16:54:45.442219", "failed": true, "rc": 1, "start": "2017-05-09 16:54:45.215145", "stderr": "grep: /var/run/gluster/shared_storage/nfs-ganesha/ganesha-ha.conf: No such file or directory\ngrep: /var/run/gluster/shared_storage/nfs-ganesha/ganesha-ha.conf: No such file or directory\nError: cluster is not currently running on this node\ngrep: /var/run/gluster/shared_storage/nfs-ganesha/exports/export.ganesha.conf: No such file or directory", "stderr_lines": ["grep: /var/run/gluster/shared_storage/nfs-ganesha/ganesha-ha.conf: No such file or directory", "grep: /var/run/gluster/shared_storage/nfs-ganesha/ganesha-ha.conf: No such file or directory", "Error: cluster is not currently running on this node", "grep: /var/run/gluster/shared_storage/nfs-ganesha/exports/exp or > t.ganesha.conf: No such file or directory"], "stdout": "Error: refresh-config failed on localhost.", "stdout_lines": ["Error: refresh-config failed on localhost."]} > to retry, use: --limit @/tmp/tmpacx8yi/ganesha-refresh-config.retry > > PLAY RECAP ************************************************************************************************* > 10.10.10.11 : ok=0 changed=0 unreachable=0 failed=1 > > Ignoring errors... > You can view the generated configuration files inside /tmp/tmpacx8yi > [root@linked-clone-of-centos-linux ~]# > _______________________________________________ > Gluster-users mailing list > Gluster-users@xxxxxxxxxxx > http://lists.gluster.org/mailman/listinfo/gluster-users > _______________________________________________ Gluster-users mailing list Gluster-users@xxxxxxxxxxx http://lists.gluster.org/mailman/listinfo/gluster-users