1.The gluster server has set volume option nfs.disable to: off Volume Name: gv0 Type: Disperse Volume ID: 429100e4-f56d-4e28-96d0-ee837386aa84 Status: Started Snapshot Count: 0 Number of Bricks: 1 x (2 + 1) = 3 Transport-type: tcp Bricks: Brick1: gfs1:/brick1/gv0 Brick2: gfs2:/brick1/gv0 Brick3: gfs3:/brick1/gv0 Options Reconfigured: transport.address-family: inet storage.fips-mode-rchecksum: on nfs.disable: off 2. The process has start. [root at gfs1 ~]# ps -ef | grep glustershd root 1117 1 0 10:12 ? 00:00:00 /usr/sbin/glusterfs -s localhost --volfile-id shd/gv0 -p /var/run/gluster/shd/gv0/gv0-shd.pid -l /var/log/glusterfs/glustershd.log -S /var/run/gluster/ca97b99a29c04606.socket --xlator-option *replicate*.node-uuid=323075ea-2b38-427c-a9aa-70ce18e94208 --process-name glustershd --client-pid=-6 3.But the status of gv0 is not correct,for it's status of NFS Server is not online. [root at gfs1 ~]# gluster volume status gv0 Status of volume: gv0 Gluster process TCP Port RDMA Port Online Pid ------------------------------------------------------------------------------ Brick gfs1:/brick1/gv0 49154 0 Y 4180 Brick gfs2:/brick1/gv0 49154 0 Y 1222 Brick gfs3:/brick1/gv0 49154 0 Y 1216 Self-heal Daemon on localhost N/A N/A Y 1117 NFS Server on localhost N/A N/A N N/A Self-heal Daemon on gfs2 N/A N/A Y 1138 NFS Server on gfs2 N/A N/A N N/A Self-heal Daemon on gfs3 N/A N/A Y 1131 NFS Server on gfs3 N/A N/A N N/A Task Status of Volume gv0 ------------------------------------------------------------------------------ There are no active volume tasks 4.So, I cann't mount the gv0 on my client. [root at kvms1 ~]# mount -t nfs gfs1:/gv0 /mnt/test mount.nfs: Connection refused Please Help! Thanks! sz_cuitao at 163.com -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20200401/74c5138b/attachment.html>
"sz_cuitao at 163.com" <sz_cuitao at 163.com> writes:> [1:multipart/alternative Hide] > > > [1/1:text/plain Show] > > > [1/2:text/html Hide Save:noname (10kB)] > > 1.The gluster server has set volume option nfs.disable to: off > > Volume Name: gv0 > Type: Disperse > Volume ID: 429100e4-f56d-4e28-96d0-ee837386aa84 > Status: Started > Snapshot Count: 0 > Number of Bricks: 1 x (2 + 1) = 3 > Transport-type: tcp > Bricks: > Brick1: gfs1:/brick1/gv0 > Brick2: gfs2:/brick1/gv0 > Brick3: gfs3:/brick1/gv0 > Options Reconfigured: > transport.address-family: inet > storage.fips-mode-rchecksum: on > nfs.disable: offI can't remember how to start the NFS process integrated with Gluster, but the documentation now suggest you better use NFS-Ganesha. Best regards, Olivier> > 2. The process has start. > > [root at gfs1 ~]# ps -ef | grep glustershd > root 1117 1 0 10:12 ? 00:00:00 /usr/sbin/glusterfs -s localhost --volfile-id shd/gv0 -p > /var/run/gluster/shd/gv0/gv0-shd.pid -l /var/log/glusterfs/glustershd.log -S /var/run/gluster/ca97b99a29c04606.socket > --xlator-option *replicate*.node-uuid=323075ea-2b38-427c-a9aa-70ce18e94208 --process-name glustershd --client-pid=-6 > > 3.But the status of gv0 is not correct,for it's status of NFS Server is not online. > > [root at gfs1 ~]# gluster volume status gv0 > Status of volume: gv0 > Gluster process TCP Port RDMA Port Online Pid > ------------------------------------------------------------------------------ > Brick gfs1:/brick1/gv0 49154 0 Y 4180 > Brick gfs2:/brick1/gv0 49154 0 Y 1222 > Brick gfs3:/brick1/gv0 49154 0 Y 1216 > Self-heal Daemon on localhost N/A N/A Y 1117 > NFS Server on localhost N/A N/A N N/A > Self-heal Daemon on gfs2 N/A N/A Y 1138 > NFS Server on gfs2 N/A N/A N N/A > Self-heal Daemon on gfs3 N/A N/A Y 1131 > NFS Server on gfs3 N/A N/A N N/A > > Task Status of Volume gv0 > ------------------------------------------------------------------------------ > There are no active volume tasks > > 4.So, I cann't mount the gv0 on my client. > > [root at kvms1 ~]# mount -t nfs gfs1:/gv0 /mnt/test > mount.nfs: Connection refused > > Please Help! > Thanks! > > -------------------------------------------------------------------------------------------------------------------------- > sz_cuitao at 163.com > > [2:text/plain Hide] > > ________ > > > > Community Meeting Calendar: > > Schedule - > Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC > Bridge: https://bluejeans.com/441850968 > > Gluster-users mailing list > Gluster-users at gluster.org > https://lists.gluster.org/mailman/listinfo/gluster-users--
If you are like me and cannot yet switch to Ganesha (it doesn't work in
our workload yet; I need to get back to working with the community on
that...)
What I would have expected in the process list was a glusterfs process with
"nfs" in the name.
here it is from one of my systems:
root     57927     1  0 Mar31 ?        00:00:00 /usr/sbin/glusterfs -s localhost
--volfile-id gluster/nfs -p /var/run/gluster/nfs/nfs.pid -l
/var/log/glusterfs/nfs.log -S /var/run/gluster/933ab0ad241fab5f.socket
My guess - but you'd have to confirm this with the logs - is your gluster
build does not have gnfs built in. Since they wish us to move to
Ganesha, it is often off by default. For my own builds, I enable it in
the spec file.
So you should have this installed:
/usr/lib64/glusterfs/7.2/xlator/nfs/server.so
If that isn't there, you likely need to adjust your spec file and
rebuild.
As others mentioned, the suggestion is to use Ganesha if possible,
which is a separate project.
I hope this helps!
PS here is a sniip from the spec file I use, with an erikj comment for
what I adjusted:
# gnfs
# if you wish to compile an rpm with the legacy gNFS server xlator
# rpmbuild -ta @PACKAGE_NAME at -@PACKAGE_VERSION at .tar.gz --with gnfs
%{?_without_gnfs:%global _with_gnfs --disable-gnfs}
# erikj force enable
%global _with_gnfs --enable-gnfs
# end erikj
On Wed, Apr 01, 2020 at 11:57:16AM +0800, sz_cuitao at 163.com
wrote:> 1.The gluster server has set volume option nfs.disable to: off
> 
> Volume Name: gv0
> Type: Disperse
> Volume ID: 429100e4-f56d-4e28-96d0-ee837386aa84
> Status: Started
> Snapshot Count: 0
> Number of Bricks: 1 x (2 + 1) = 3
> Transport-type: tcp
> Bricks:
> Brick1: gfs1:/brick1/gv0
> Brick2: gfs2:/brick1/gv0
> Brick3: gfs3:/brick1/gv0
> Options Reconfigured:
> transport.address-family: inet
> storage.fips-mode-rchecksum: on
> nfs.disable: off
> 
> 2. The process has start.
> 
> [root at gfs1 ~]# ps -ef | grep glustershd
> root       1117      1  0 10:12 ?        00:00:00 /usr/sbin/glusterfs -s
> localhost --volfile-id shd/gv0 -p /var/run/gluster/shd/gv0/gv0-shd.pid -l
/var/
> log/glusterfs/glustershd.log -S /var/run/gluster/ca97b99a29c04606.socket
> --xlator-option *replicate*.node-uuid=323075ea-2b38-427c-a9aa-70ce18e94208
> --process-name glustershd --client-pid=-6
> 
> 
> 3.But the status of gv0 is not correct,for it's status of NFS Server is
not
> online.
> 
> [root at gfs1 ~]# gluster volume status gv0
> Status of volume: gv0
> Gluster process                             TCP Port  RDMA Port  Online 
Pid
>
------------------------------------------------------------------------------
> Brick gfs1:/brick1/gv0                      49154     0          Y      
4180
> Brick gfs2:/brick1/gv0                      49154     0          Y      
1222
> Brick gfs3:/brick1/gv0                      49154     0          Y      
1216
> Self-heal Daemon on localhost               N/A       N/A        Y      
1117
> NFS Server on localhost                     N/A       N/A        N      
N/A
> Self-heal Daemon on gfs2                    N/A       N/A        Y      
1138
> NFS Server on gfs2                          N/A       N/A        N      
N/A
> Self-heal Daemon on gfs3                    N/A       N/A        Y      
1131
> NFS Server on gfs3                          N/A       N/A        N      
N/A
> 
> Task Status of Volume gv0
>
------------------------------------------------------------------------------
> There are no active volume tasks
> 
> 4.So, I cann't mount the gv0 on my client.
> 
> [root at kvms1 ~]# mount -t nfs  gfs1:/gv0 /mnt/test
> mount.nfs: Connection refused
> 
> 
> Please Help!
> Thanks!
> 
> 
> 
> 
> 
>
???????????????????????????????????????????????????????????????????????????????
> sz_cuitao at 163.com
> ________
> 
> 
> 
> Community Meeting Calendar:
> 
> Schedule -
> Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC
> Bridge: https://bluejeans.com/441850968 
> 
> Gluster-users mailing list
> Gluster-users at gluster.org
> https://lists.gluster.org/mailman/listinfo/gluster-users 
Erik Jacobson
Software Engineer
erik.jacobson at hpe.com
+1 612 851 0550 Office
Eagan, MN
hpe.com