1.The gluster server has set volume option nfs.disable to: off Volume Name: gv0 Type: Disperse Volume ID: 429100e4-f56d-4e28-96d0-ee837386aa84 Status: Started Snapshot Count: 0 Number of Bricks: 1 x (2 + 1) = 3 Transport-type: tcp Bricks: Brick1: gfs1:/brick1/gv0 Brick2: gfs2:/brick1/gv0 Brick3: gfs3:/brick1/gv0 Options Reconfigured: transport.address-family: inet storage.fips-mode-rchecksum: on nfs.disable: off 2. The process has start. [root at gfs1 ~]# ps -ef | grep glustershd root 1117 1 0 10:12 ? 00:00:00 /usr/sbin/glusterfs -s localhost --volfile-id shd/gv0 -p /var/run/gluster/shd/gv0/gv0-shd.pid -l /var/log/glusterfs/glustershd.log -S /var/run/gluster/ca97b99a29c04606.socket --xlator-option *replicate*.node-uuid=323075ea-2b38-427c-a9aa-70ce18e94208 --process-name glustershd --client-pid=-6 3.But the status of gv0 is not correct,for it's status of NFS Server is not online. [root at gfs1 ~]# gluster volume status gv0 Status of volume: gv0 Gluster process TCP Port RDMA Port Online Pid ------------------------------------------------------------------------------ Brick gfs1:/brick1/gv0 49154 0 Y 4180 Brick gfs2:/brick1/gv0 49154 0 Y 1222 Brick gfs3:/brick1/gv0 49154 0 Y 1216 Self-heal Daemon on localhost N/A N/A Y 1117 NFS Server on localhost N/A N/A N N/A Self-heal Daemon on gfs2 N/A N/A Y 1138 NFS Server on gfs2 N/A N/A N N/A Self-heal Daemon on gfs3 N/A N/A Y 1131 NFS Server on gfs3 N/A N/A N N/A Task Status of Volume gv0 ------------------------------------------------------------------------------ There are no active volume tasks 4.So, I cann't mount the gv0 on my client. [root at kvms1 ~]# mount -t nfs gfs1:/gv0 /mnt/test mount.nfs: Connection refused Please Help! Thanks! sz_cuitao at 163.com -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20200401/74c5138b/attachment.html>
"sz_cuitao at 163.com" <sz_cuitao at 163.com> writes:> [1:multipart/alternative Hide] > > > [1/1:text/plain Show] > > > [1/2:text/html Hide Save:noname (10kB)] > > 1.The gluster server has set volume option nfs.disable to: off > > Volume Name: gv0 > Type: Disperse > Volume ID: 429100e4-f56d-4e28-96d0-ee837386aa84 > Status: Started > Snapshot Count: 0 > Number of Bricks: 1 x (2 + 1) = 3 > Transport-type: tcp > Bricks: > Brick1: gfs1:/brick1/gv0 > Brick2: gfs2:/brick1/gv0 > Brick3: gfs3:/brick1/gv0 > Options Reconfigured: > transport.address-family: inet > storage.fips-mode-rchecksum: on > nfs.disable: offI can't remember how to start the NFS process integrated with Gluster, but the documentation now suggest you better use NFS-Ganesha. Best regards, Olivier> > 2. The process has start. > > [root at gfs1 ~]# ps -ef | grep glustershd > root 1117 1 0 10:12 ? 00:00:00 /usr/sbin/glusterfs -s localhost --volfile-id shd/gv0 -p > /var/run/gluster/shd/gv0/gv0-shd.pid -l /var/log/glusterfs/glustershd.log -S /var/run/gluster/ca97b99a29c04606.socket > --xlator-option *replicate*.node-uuid=323075ea-2b38-427c-a9aa-70ce18e94208 --process-name glustershd --client-pid=-6 > > 3.But the status of gv0 is not correct,for it's status of NFS Server is not online. > > [root at gfs1 ~]# gluster volume status gv0 > Status of volume: gv0 > Gluster process TCP Port RDMA Port Online Pid > ------------------------------------------------------------------------------ > Brick gfs1:/brick1/gv0 49154 0 Y 4180 > Brick gfs2:/brick1/gv0 49154 0 Y 1222 > Brick gfs3:/brick1/gv0 49154 0 Y 1216 > Self-heal Daemon on localhost N/A N/A Y 1117 > NFS Server on localhost N/A N/A N N/A > Self-heal Daemon on gfs2 N/A N/A Y 1138 > NFS Server on gfs2 N/A N/A N N/A > Self-heal Daemon on gfs3 N/A N/A Y 1131 > NFS Server on gfs3 N/A N/A N N/A > > Task Status of Volume gv0 > ------------------------------------------------------------------------------ > There are no active volume tasks > > 4.So, I cann't mount the gv0 on my client. > > [root at kvms1 ~]# mount -t nfs gfs1:/gv0 /mnt/test > mount.nfs: Connection refused > > Please Help! > Thanks! > > -------------------------------------------------------------------------------------------------------------------------- > sz_cuitao at 163.com > > [2:text/plain Hide] > > ________ > > > > Community Meeting Calendar: > > Schedule - > Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC > Bridge: https://bluejeans.com/441850968 > > Gluster-users mailing list > Gluster-users at gluster.org > https://lists.gluster.org/mailman/listinfo/gluster-users--
If you are like me and cannot yet switch to Ganesha (it doesn't work in our workload yet; I need to get back to working with the community on that...) What I would have expected in the process list was a glusterfs process with "nfs" in the name. here it is from one of my systems: root 57927 1 0 Mar31 ? 00:00:00 /usr/sbin/glusterfs -s localhost --volfile-id gluster/nfs -p /var/run/gluster/nfs/nfs.pid -l /var/log/glusterfs/nfs.log -S /var/run/gluster/933ab0ad241fab5f.socket My guess - but you'd have to confirm this with the logs - is your gluster build does not have gnfs built in. Since they wish us to move to Ganesha, it is often off by default. For my own builds, I enable it in the spec file. So you should have this installed: /usr/lib64/glusterfs/7.2/xlator/nfs/server.so If that isn't there, you likely need to adjust your spec file and rebuild. As others mentioned, the suggestion is to use Ganesha if possible, which is a separate project. I hope this helps! PS here is a sniip from the spec file I use, with an erikj comment for what I adjusted: # gnfs # if you wish to compile an rpm with the legacy gNFS server xlator # rpmbuild -ta @PACKAGE_NAME at -@PACKAGE_VERSION at .tar.gz --with gnfs %{?_without_gnfs:%global _with_gnfs --disable-gnfs} # erikj force enable %global _with_gnfs --enable-gnfs # end erikj On Wed, Apr 01, 2020 at 11:57:16AM +0800, sz_cuitao at 163.com wrote:> 1.The gluster server has set volume option nfs.disable to: off > > Volume Name: gv0 > Type: Disperse > Volume ID: 429100e4-f56d-4e28-96d0-ee837386aa84 > Status: Started > Snapshot Count: 0 > Number of Bricks: 1 x (2 + 1) = 3 > Transport-type: tcp > Bricks: > Brick1: gfs1:/brick1/gv0 > Brick2: gfs2:/brick1/gv0 > Brick3: gfs3:/brick1/gv0 > Options Reconfigured: > transport.address-family: inet > storage.fips-mode-rchecksum: on > nfs.disable: off > > 2. The process has start. > > [root at gfs1 ~]# ps -ef | grep glustershd > root 1117 1 0 10:12 ? 00:00:00 /usr/sbin/glusterfs -s > localhost --volfile-id shd/gv0 -p /var/run/gluster/shd/gv0/gv0-shd.pid -l /var/ > log/glusterfs/glustershd.log -S /var/run/gluster/ca97b99a29c04606.socket > --xlator-option *replicate*.node-uuid=323075ea-2b38-427c-a9aa-70ce18e94208 > --process-name glustershd --client-pid=-6 > > > 3.But the status of gv0 is not correct,for it's status of NFS Server is not > online. > > [root at gfs1 ~]# gluster volume status gv0 > Status of volume: gv0 > Gluster process TCP Port RDMA Port Online Pid > ------------------------------------------------------------------------------ > Brick gfs1:/brick1/gv0 49154 0 Y 4180 > Brick gfs2:/brick1/gv0 49154 0 Y 1222 > Brick gfs3:/brick1/gv0 49154 0 Y 1216 > Self-heal Daemon on localhost N/A N/A Y 1117 > NFS Server on localhost N/A N/A N N/A > Self-heal Daemon on gfs2 N/A N/A Y 1138 > NFS Server on gfs2 N/A N/A N N/A > Self-heal Daemon on gfs3 N/A N/A Y 1131 > NFS Server on gfs3 N/A N/A N N/A > > Task Status of Volume gv0 > ------------------------------------------------------------------------------ > There are no active volume tasks > > 4.So, I cann't mount the gv0 on my client. > > [root at kvms1 ~]# mount -t nfs gfs1:/gv0 /mnt/test > mount.nfs: Connection refused > > > Please Help! > Thanks! > > > > > > ??????????????????????????????????????????????????????????????????????????????? > sz_cuitao at 163.com> ________ > > > > Community Meeting Calendar: > > Schedule - > Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC > Bridge: https://bluejeans.com/441850968 > > Gluster-users mailing list > Gluster-users at gluster.org > https://lists.gluster.org/mailman/listinfo/gluster-usersErik Jacobson Software Engineer erik.jacobson at hpe.com +1 612 851 0550 Office Eagan, MN hpe.com