gluster nfs-ganesha enable fails and is driving me crazy

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,

I hope someone could help me with this issue is driving me crazy: I'm trying to setup an high available NFS server using gluster and ganesha,it seems that all the required steps work right, except that when I issue on glstr02.carcano.local the command

gluster nfs-ganesha enable

I got the following output with error

Enabling NFS-Ganesha requires Gluster-NFS to be disabled across the trusted pool. Do you still want to continue?
 (y/n) y
This will take a few minutes to complete. Please wait ..
nfs-ganesha: failed: Commit failed on glstr01.carcano.local. Error: Failed to set up HA config for NFS-Ganesha. Please check the log file for details

and in log files

==> etc-glusterfs-glusterd.vol.log <==
[2015-12-07 20:42:43.888793] E [MSGID: 106062] [glusterd-op-sm.c:3647:glusterd_op_ac_lock] 0-management: Unable to acquire volname [2015-12-07 20:42:44.244133] W [common-utils.c:1685:gf_string2boolean] (-->/usr/lib64/glusterfs/3.7.6/xlator/mgmt/glusterd.so(glusterd_op_stage_validate+0x143) [0x7fa3da411223] -->/usr/lib64/glusterfs/3.7.6/xlator/mgmt/glusterd.so(glusterd_op_stage_set_ganesha+0xa8) [0x7fa3da45f8d8] -->/usr/lib64/libglusterfs.so.0(gf_string2boolean+0x157) [0x7fa3e59efde7] ) 0-management: argument invalid [Invalid argument] [2015-12-07 20:42:44.428305] I [MSGID: 106132] [glusterd-proc-mgmt.c:83:glusterd_proc_stop] 0-management: nfs already stopped [2015-12-07 20:42:44.432079] I [MSGID: 106474] [glusterd-ganesha.c:403:check_host_list] 0-management: ganesha host found Hostname is glstr01.carcano.local [2015-12-07 20:42:44.705564] I [MSGID: 106474] [glusterd-ganesha.c:349:is_ganesha_host] 0-management: ganesha host found Hostname is glstr01.carcano.local [2015-12-07 20:42:48.525223] E [MSGID: 106470] [glusterd-ganesha.c:264:glusterd_op_set_ganesha] 0-management: Initial NFS-Ganesha set up failed [2015-12-07 20:42:48.525320] E [MSGID: 106123] [glusterd-op-sm.c:5311:glusterd_op_ac_commit_op] 0-management: Commit of operation 'Volume (null)' failed: -1 [2015-12-07 20:42:48.541707] E [MSGID: 106062] [glusterd-op-sm.c:3710:glusterd_op_ac_unlock] 0-management: Unable to acquire volname [2015-12-07 20:42:49.674994] I [MSGID: 106164] [glusterd-handshake.c:1248:__server_get_volume_info] 0-glusterd: Received get volume info req [2015-12-07 20:42:44.705289] I [MSGID: 106474] [glusterd-ganesha.c:403:check_host_list] 0-management: ganesha host found Hostname is glstr01.carcano.local

I suppose the actual error is "E [MSGID: 106062] [glusterd-op-sm.c:3647:glusterd_op_ac_lock] 0-management: Unable to acquire volname"

This is really driving me crazy: if I start nfs-ganesha service alone, configured to use the gluster volume it works - if I issue on another client

showmount -e glstr01.carcano.local

it shows my "vol1" . that is the exported volume

I really don't know what am I failing

I'm using gluster 3.7.6-1 and nfs-ganesha-2.3.0-1 on CentOS 6.7. IPv6 is enabled and Network Manager as well as iptables and ip6tables are disabled. I tried with both selinux permissive and enforcing.

I installed the packages issuing:

yum install -y glusterfs-ganesha pacemaker pacemaker-cli pacemaker-libs corosync

here are the configuration files:

/etc/hosts
127.0.0.1 localhost.localdomain localhost.localdomain localhost4 localhost4.localdomain4 localhost ::1 localhost.localdomain localhost.localdomain localhost6 localhost6.localdomain6 localhost
192.168.66.250 glstr01.carcano.local glstr01
192.168.66.251 glstr02.carcano.local glstr02
192.168.65.250 glstr01v.carcano.local
192.168.65.251 glstr02v.carcano.local

/etc/ganesha/ganesha-ha.conf

HA_NAME="ganesha-ha-360"
HA_VOL_SERVER="glstr01.carcano.local"
HA_CLUSTER_NODES="glstr01.carcano.local,glstr02.carcano.local"
VIP_server1="192.168.65.250"
VIP_server2="192.168.65.251"

/etc/ganesha/ganesha.conf

EXPORT
{
    Export_Id = 77;
    Path = /vol1;
    Pseudo = /vol1;
    Access_Type = RW;
    FSAL {
        Name = GLUSTER;
        Hostname = localhost;
        Volume = vol1;
    }
}

gluster volume info

Volume Name: gluster_shared_storage
Type: Replicate
Volume ID: a899b299-a43b-4119-9411-4c68fd72a550
Status: Started
Number of Bricks: 1 x 2 = 2
Transport-type: tcp
Bricks:
Brick1: glstr02:/var/lib/glusterd/ss_brick
Brick2: glstr01.carcano.local:/var/lib/glusterd/ss_brick
Options Reconfigured:
nfs.disable: on
performance.readdir-ahead: on
cluster.enable-shared-storage: enable

Volume Name: vol1
Type: Replicate
Volume ID: d5632acb-972c-4840-9dd2-527723312419
Status: Started
Number of Bricks: 1 x 2 = 2
Transport-type: tcp
Bricks:
Brick1: glstr01:/gluster/disk1/contents
Brick2: glstr02:/gluster/disk1/contents
Options Reconfigured:
nfs.disable: on
performance.readdir-ahead: on
cluster.enable-shared-storage: enable

luster volume status
Status of volume: gluster_shared_storage
Gluster process                             TCP Port  RDMA Port Online  Pid
------------------------------------------------------------------------------
Brick glstr02:/var/lib/glusterd/ss_brick    49152     0 Y       1898
Brick glstr01.carcano.local:/var/lib/gluste
rd/ss_brick                                 49152     0 Y       1934
Self-heal Daemon on localhost               N/A       N/A Y 2125
Self-heal Daemon on glstr02                 N/A       N/A Y 2038

Task Status of Volume gluster_shared_storage
------------------------------------------------------------------------------
There are no active volume tasks

Status of volume: vol1
Gluster process                             TCP Port  RDMA Port Online  Pid
------------------------------------------------------------------------------
Brick glstr01:/gluster/disk1/contents       49153     0 Y       2094
Brick glstr02:/gluster/disk1/contents       49153     0 Y       1945
Self-heal Daemon on localhost               N/A       N/A Y 2125
Self-heal Daemon on glstr02                 N/A       N/A Y 2038

Task Status of Volume vol1
------------------------------------------------------------------------------
There are no active volume tasks

/usr/libexec/ganesha/ganesha-ha.sh --status
Error: cluster is not currently running on this node
Error: cluster is not currently running on this node
Error: cluster is not currently running on this node

Kind regards

Marco





_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
http://www.gluster.org/mailman/listinfo/gluster-users



[Index of Archives]     [Gluster Development]     [Linux Filesytems Development]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux