Sherry Reese
2020-Jan-24 01:37 UTC
[Gluster-users] No possible to mount a gluster volume via /etc/fstab?
Hello everyone, I am using the following entry on a CentOS server. gluster01.home:/videos /data2/plex/videos glusterfs _netdev 0 0 gluster01.home:/photos /data2/plex/photos glusterfs _netdev 0 0 I am able to use sudo mount -a to mount the volumes without any problems. When I reboot my server, nothing is mounted. I can see errors in /var/log/glusterfs/data2-plex-photos.log: ... [2020-01-24 01:24:18.302191] I [glusterfsd.c:2594:daemonize] 0-glusterfs: Pid of current running process is 3679 [2020-01-24 01:24:18.310017] E [MSGID: 101075] [common-utils.c:505:gf_resolve_ip6] 0-resolver: getaddrinfo failed (family:2) (Name or service not known) [2020-01-24 01:24:18.310046] E [name.c:266:af_inet_client_get_remote_sockaddr] 0-glusterfs: DNS resolution failed on host gluster01.home [2020-01-24 01:24:18.310187] I [MSGID: 101190] [event-epoll.c:682:event_dispatch_epoll_worker] 0-epoll: Started thread with index 0 ... I am able to to nslookup on gluster01 and gluster01.home without problems, so "DNS resolution failed" is confusing to me. What happens here? Output of my volumes. sudo gluster volume status Status of volume: documents Gluster process TCP Port RDMA Port Online Pid ------------------------------------------------------------------------------ Brick gluster01.home:/data/documents 49152 0 Y 5658 Brick gluster02.home:/data/documents 49152 0 Y 5340 Brick gluster03.home:/data/documents 49152 0 Y 5305 Self-heal Daemon on localhost N/A N/A Y 5679 Self-heal Daemon on gluster03.home N/A N/A Y 5326 Self-heal Daemon on gluster02.home N/A N/A Y 5361 Task Status of Volume documents ------------------------------------------------------------------------------ There are no active volume tasks Status of volume: photos Gluster process TCP Port RDMA Port Online Pid ------------------------------------------------------------------------------ Brick gluster01.home:/data/photos 49153 0 Y 5779 Brick gluster02.home:/data/photos 49153 0 Y 5401 Brick gluster03.home:/data/photos 49153 0 Y 5366 Self-heal Daemon on localhost N/A N/A Y 5679 Self-heal Daemon on gluster03.home N/A N/A Y 5326 Self-heal Daemon on gluster02.home N/A N/A Y 5361 Task Status of Volume photos ------------------------------------------------------------------------------ There are no active volume tasks Status of volume: videos Gluster process TCP Port RDMA Port Online Pid ------------------------------------------------------------------------------ Brick gluster01.home:/data/videos 49154 0 Y 5883 Brick gluster02.home:/data/videos 49154 0 Y 5452 Brick gluster03.home:/data/videos 49154 0 Y 5416 Self-heal Daemon on localhost N/A N/A Y 5679 Self-heal Daemon on gluster03.home N/A N/A Y 5326 Self-heal Daemon on gluster02.home N/A N/A Y 5361 Task Status of Volume videos ------------------------------------------------------------------------------ There are no active volume tasks On the server (Ubuntu) following versions are installed. glusterfs-client/bionic,now 7.2-ubuntu1~bionic1 armhf [installed,automatic] glusterfs-common/bionic,now 7.2-ubuntu1~bionic1 armhf [installed,automatic] glusterfs-server/bionic,now 7.2-ubuntu1~bionic1 armhf [installed] On the client (CentOS) following versions are installed. sudo rpm -qa | grep gluster glusterfs-client-xlators-7.2-1.el7.x86_64 glusterfs-cli-7.2-1.el7.x86_64 glusterfs-libs-7.2-1.el7.x86_64 glusterfs-7.2-1.el7.x86_64 glusterfs-api-7.2-1.el7.x86_64 libvirt-daemon-driver-storage-gluster-4.5.0-23.el7_7.3.x86_64 centos-release-gluster7-1.0-1.el7.centos.noarch glusterfs-fuse-7.2-1.el7.x86_64 I tried to disable IPv6 on the client voa sysctl with following parameters. net.ipv6.conf.all.disable_ipv6 = 1 net.ipv6.conf.default.disable_ipv6 = 1 That did not help. Volumes are configured with inet. sudo gluster volume info videos Volume Name: videos Type: Replicate Volume ID: 8fddde82-66b3-447f-8860-ed3768c51876 Status: Started Snapshot Count: 0 Number of Bricks: 1 x 3 = 3 Transport-type: tcp Bricks: Brick1: gluster01.home:/data/videos Brick2: gluster02.home:/data/videos Brick3: gluster03.home:/data/videos Options Reconfigured: features.ctime: on transport.address-family: inet nfs.disable: on performance.client-io-threads: off I tried turning off ctime but that did not work either. Any ideas? How do I do this correctly? Cheers Sherry -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20200124/cc8a2773/attachment.html>
Hu Bert
2020-Jan-24 05:50 UTC
[Gluster-users] No possible to mount a gluster volume via /etc/fstab?
Hi Sherry, maybe at the time, when the mount from /etc/fstab should take place, name resolution is not yet working? In your case i'd try to place proper entries in /etc/hosts and test it with a reboot. regards Hubert Am Fr., 24. Jan. 2020 um 02:37 Uhr schrieb Sherry Reese <s.reese4u at gmail.com>:> > Hello everyone, > > I am using the following entry on a CentOS server. > > gluster01.home:/videos /data2/plex/videos glusterfs _netdev 0 0 > gluster01.home:/photos /data2/plex/photos glusterfs _netdev 0 0 > > I am able to use sudo mount -a to mount the volumes without any problems. When I reboot my server, nothing is mounted. > > I can see errors in /var/log/glusterfs/data2-plex-photos.log: > > ... > [2020-01-24 01:24:18.302191] I [glusterfsd.c:2594:daemonize] 0-glusterfs: Pid of current running process is 3679 > [2020-01-24 01:24:18.310017] E [MSGID: 101075] [common-utils.c:505:gf_resolve_ip6] 0-resolver: getaddrinfo failed (family:2) (Name or service not known) > [2020-01-24 01:24:18.310046] E [name.c:266:af_inet_client_get_remote_sockaddr] 0-glusterfs: DNS resolution failed on host gluster01.home > [2020-01-24 01:24:18.310187] I [MSGID: 101190] [event-epoll.c:682:event_dispatch_epoll_worker] 0-epoll: Started thread with index 0 > ... > > I am able to to nslookup on gluster01 and gluster01.home without problems, so "DNS resolution failed" is confusing to me. What happens here? > > Output of my volumes. > > sudo gluster volume status > Status of volume: documents > Gluster process TCP Port RDMA Port Online Pid > ------------------------------------------------------------------------------ > Brick gluster01.home:/data/documents 49152 0 Y 5658 > Brick gluster02.home:/data/documents 49152 0 Y 5340 > Brick gluster03.home:/data/documents 49152 0 Y 5305 > Self-heal Daemon on localhost N/A N/A Y 5679 > Self-heal Daemon on gluster03.home N/A N/A Y 5326 > Self-heal Daemon on gluster02.home N/A N/A Y 5361 > > Task Status of Volume documents > ------------------------------------------------------------------------------ > There are no active volume tasks > > Status of volume: photos > Gluster process TCP Port RDMA Port Online Pid > ------------------------------------------------------------------------------ > Brick gluster01.home:/data/photos 49153 0 Y 5779 > Brick gluster02.home:/data/photos 49153 0 Y 5401 > Brick gluster03.home:/data/photos 49153 0 Y 5366 > Self-heal Daemon on localhost N/A N/A Y 5679 > Self-heal Daemon on gluster03.home N/A N/A Y 5326 > Self-heal Daemon on gluster02.home N/A N/A Y 5361 > > Task Status of Volume photos > ------------------------------------------------------------------------------ > There are no active volume tasks > > Status of volume: videos > Gluster process TCP Port RDMA Port Online Pid > ------------------------------------------------------------------------------ > Brick gluster01.home:/data/videos 49154 0 Y 5883 > Brick gluster02.home:/data/videos 49154 0 Y 5452 > Brick gluster03.home:/data/videos 49154 0 Y 5416 > Self-heal Daemon on localhost N/A N/A Y 5679 > Self-heal Daemon on gluster03.home N/A N/A Y 5326 > Self-heal Daemon on gluster02.home N/A N/A Y 5361 > > Task Status of Volume videos > ------------------------------------------------------------------------------ > There are no active volume tasks > > On the server (Ubuntu) following versions are installed. > > glusterfs-client/bionic,now 7.2-ubuntu1~bionic1 armhf [installed,automatic] > glusterfs-common/bionic,now 7.2-ubuntu1~bionic1 armhf [installed,automatic] > glusterfs-server/bionic,now 7.2-ubuntu1~bionic1 armhf [installed] > > On the client (CentOS) following versions are installed. > > sudo rpm -qa | grep gluster > glusterfs-client-xlators-7.2-1.el7.x86_64 > glusterfs-cli-7.2-1.el7.x86_64 > glusterfs-libs-7.2-1.el7.x86_64 > glusterfs-7.2-1.el7.x86_64 > glusterfs-api-7.2-1.el7.x86_64 > libvirt-daemon-driver-storage-gluster-4.5.0-23.el7_7.3.x86_64 > centos-release-gluster7-1.0-1.el7.centos.noarch > glusterfs-fuse-7.2-1.el7.x86_64 > > I tried to disable IPv6 on the client voa sysctl with following parameters. > > net.ipv6.conf.all.disable_ipv6 = 1 > net.ipv6.conf.default.disable_ipv6 = 1 > > That did not help. > > Volumes are configured with inet. > > sudo gluster volume info videos > > Volume Name: videos > Type: Replicate > Volume ID: 8fddde82-66b3-447f-8860-ed3768c51876 > Status: Started > Snapshot Count: 0 > Number of Bricks: 1 x 3 = 3 > Transport-type: tcp > Bricks: > Brick1: gluster01.home:/data/videos > Brick2: gluster02.home:/data/videos > Brick3: gluster03.home:/data/videos > Options Reconfigured: > features.ctime: on > transport.address-family: inet > nfs.disable: on > performance.client-io-threads: off > > I tried turning off ctime but that did not work either. > > Any ideas? How do I do this correctly? > > Cheers > Sherry > ________ > > Community Meeting Calendar: > > APAC Schedule - > Every 2nd and 4th Tuesday at 11:30 AM IST > Bridge: https://bluejeans.com/441850968 > > NA/EMEA Schedule - > Every 1st and 3rd Tuesday at 01:00 PM EDT > Bridge: https://bluejeans.com/441850968 > > Gluster-users mailing list > Gluster-users at gluster.org > https://lists.gluster.org/mailman/listinfo/gluster-users