Marco Antonio Carcano
2015-Dec-08  08:46 UTC
[Gluster-users] gluster nfs-ganesha enable fails and is driving me crazy
Hi,
I hope someone could help me with this issue is driving me crazy: I'm 
trying to setup an high available NFS server using gluster and 
ganesha,it seems that all the required steps work right, except that 
when I issue on glstr02.carcano.local the command
gluster nfs-ganesha enable
I got the following output with error
Enabling NFS-Ganesha requires Gluster-NFS to be disabled across the 
trusted pool. Do you still want to continue?
  (y/n) y
This will take a few minutes to complete. Please wait ..
nfs-ganesha: failed: Commit failed on glstr01.carcano.local. Error: 
Failed to set up HA config for NFS-Ganesha. Please check the log file 
for details
and in log files
==> etc-glusterfs-glusterd.vol.log <=[2015-12-07 20:42:43.888793] E
[MSGID: 106062]
[glusterd-op-sm.c:3647:glusterd_op_ac_lock] 0-management: Unable to 
acquire volname
[2015-12-07 20:42:44.244133] W [common-utils.c:1685:gf_string2boolean] 
(-->/usr/lib64/glusterfs/3.7.6/xlator/mgmt/glusterd.so(glusterd_op_stage_validate+0x143)
[0x7fa3da411223] 
-->/usr/lib64/glusterfs/3.7.6/xlator/mgmt/glusterd.so(glusterd_op_stage_set_ganesha+0xa8)
[0x7fa3da45f8d8] 
-->/usr/lib64/libglusterfs.so.0(gf_string2boolean+0x157) 
[0x7fa3e59efde7] ) 0-management: argument invalid [Invalid argument]
[2015-12-07 20:42:44.428305] I [MSGID: 106132] 
[glusterd-proc-mgmt.c:83:glusterd_proc_stop] 0-management: nfs already 
stopped
[2015-12-07 20:42:44.432079] I [MSGID: 106474] 
[glusterd-ganesha.c:403:check_host_list] 0-management: ganesha host 
found Hostname is glstr01.carcano.local
[2015-12-07 20:42:44.705564] I [MSGID: 106474] 
[glusterd-ganesha.c:349:is_ganesha_host] 0-management: ganesha host 
found Hostname is glstr01.carcano.local
[2015-12-07 20:42:48.525223] E [MSGID: 106470] 
[glusterd-ganesha.c:264:glusterd_op_set_ganesha] 0-management: Initial 
NFS-Ganesha set up failed
[2015-12-07 20:42:48.525320] E [MSGID: 106123] 
[glusterd-op-sm.c:5311:glusterd_op_ac_commit_op] 0-management: Commit of 
operation 'Volume (null)' failed: -1
[2015-12-07 20:42:48.541707] E [MSGID: 106062] 
[glusterd-op-sm.c:3710:glusterd_op_ac_unlock] 0-management: Unable to 
acquire volname
[2015-12-07 20:42:49.674994] I [MSGID: 106164] 
[glusterd-handshake.c:1248:__server_get_volume_info] 0-glusterd: 
Received get volume info req
[2015-12-07 20:42:44.705289] I [MSGID: 106474] 
[glusterd-ganesha.c:403:check_host_list] 0-management: ganesha host 
found Hostname is glstr01.carcano.local
I suppose the actual error is "E [MSGID: 106062] 
[glusterd-op-sm.c:3647:glusterd_op_ac_lock] 0-management: Unable to 
acquire volname"
This is really driving me crazy: if I start nfs-ganesha service alone, 
configured to use the gluster volume it works - if I issue on another 
client
showmount -e glstr01.carcano.local
it shows my "vol1" . that is the exported volume
I really don't know what am I failing
I'm using gluster 3.7.6-1 and nfs-ganesha-2.3.0-1 on CentOS 6.7. IPv6 is 
enabled and Network Manager as well as iptables and ip6tables are 
disabled. I tried with both selinux permissive and enforcing.
I installed the packages issuing:
yum install -y glusterfs-ganesha pacemaker pacemaker-cli pacemaker-libs 
corosync
here are the configuration files:
/etc/hosts
127.0.0.1    localhost.localdomain    localhost.localdomain 
localhost4    localhost4.localdomain4    localhost
::1    localhost.localdomain    localhost.localdomain localhost6 
localhost6.localdomain6    localhost
192.168.66.250 glstr01.carcano.local glstr01
192.168.66.251 glstr02.carcano.local glstr02
192.168.65.250 glstr01v.carcano.local
192.168.65.251 glstr02v.carcano.local
/etc/ganesha/ganesha-ha.conf
HA_NAME="ganesha-ha-360"
HA_VOL_SERVER="glstr01.carcano.local"
HA_CLUSTER_NODES="glstr01.carcano.local,glstr02.carcano.local"
VIP_server1="192.168.65.250"
VIP_server2="192.168.65.251"
/etc/ganesha/ganesha.conf
EXPORT
{
     Export_Id = 77;
     Path = /vol1;
     Pseudo = /vol1;
     Access_Type = RW;
     FSAL {
         Name = GLUSTER;
         Hostname = localhost;
         Volume = vol1;
     }
}
gluster volume info
Volume Name: gluster_shared_storage
Type: Replicate
Volume ID: a899b299-a43b-4119-9411-4c68fd72a550
Status: Started
Number of Bricks: 1 x 2 = 2
Transport-type: tcp
Bricks:
Brick1: glstr02:/var/lib/glusterd/ss_brick
Brick2: glstr01.carcano.local:/var/lib/glusterd/ss_brick
Options Reconfigured:
nfs.disable: on
performance.readdir-ahead: on
cluster.enable-shared-storage: enable
Volume Name: vol1
Type: Replicate
Volume ID: d5632acb-972c-4840-9dd2-527723312419
Status: Started
Number of Bricks: 1 x 2 = 2
Transport-type: tcp
Bricks:
Brick1: glstr01:/gluster/disk1/contents
Brick2: glstr02:/gluster/disk1/contents
Options Reconfigured:
nfs.disable: on
performance.readdir-ahead: on
cluster.enable-shared-storage: enable
luster volume status
Status of volume: gluster_shared_storage
Gluster process                             TCP Port  RDMA Port Online  Pid
------------------------------------------------------------------------------ 
Brick glstr02:/var/lib/glusterd/ss_brick    49152     0 Y       1898
Brick glstr01.carcano.local:/var/lib/gluste
rd/ss_brick                                 49152     0 Y       1934
Self-heal Daemon on localhost               N/A       N/A Y 2125
Self-heal Daemon on glstr02                 N/A       N/A Y 2038
Task Status of Volume gluster_shared_storage
------------------------------------------------------------------------------ 
There are no active volume tasks
Status of volume: vol1
Gluster process                             TCP Port  RDMA Port Online  Pid
------------------------------------------------------------------------------ 
Brick glstr01:/gluster/disk1/contents       49153     0 Y       2094
Brick glstr02:/gluster/disk1/contents       49153     0 Y       1945
Self-heal Daemon on localhost               N/A       N/A Y 2125
Self-heal Daemon on glstr02                 N/A       N/A Y 2038
Task Status of Volume vol1
------------------------------------------------------------------------------ 
There are no active volume tasks
/usr/libexec/ganesha/ganesha-ha.sh --status
Error: cluster is not currently running on this node
Error: cluster is not currently running on this node
Error: cluster is not currently running on this node
Kind regards
Marco
Kaleb KEITHLEY
2015-Dec-08  12:30 UTC
[Gluster-users] gluster nfs-ganesha enable fails and is driving me crazy
On 12/08/2015 03:46 AM, Marco Antonio Carcano wrote:> Hi, >> > /etc/ganesha/ganesha-ha.conf > > HA_NAME="ganesha-ha-360" > HA_VOL_SERVER="glstr01.carcano.local" > HA_CLUSTER_NODES="glstr01.carcano.local,glstr02.carcano.local" > VIP_server1="192.168.65.250" > VIP_server2="192.168.65.251" >change your /etc/ganesha/ganesha-ha.conf file: HA_NAME="ganesha-ha-360" HA_VOL_SERVER="glstr01.carcano.local" HA_CLUSTER_NODES="glstr01.carcano.local,glstr02.carcano.local" VIP_glstr01.carcano.local="192.168.65.250" VIP_glstr02.carcano.local="192.168.65.251" I'd change the HA_NAME to something else too, but as long as you don't set up another cluster on the same network you should be fine. -- Kaleb