Amar Tumballi Suryanarayan
2019-Jan-24 13:49 UTC
[Gluster-users] Access to Servers hangs after stop one server...
Also note that, this way of mounting with a 'static' volfile is not recommended as you wouldn't get any features out of gluster's Software Defined Storage behavior. this was an approach we used to have say 8 years before. With the introduction of management daemon called glusterd, the way of dealing with volfiles have changed, and it is created with gluster CLI. About having /etc/fstab not hang when a server is down, search for 'backup-volfile-server' option with glusterfs, and that should be used. Regards, Amar On Thu, Jan 24, 2019 at 7:17 PM Diego Remolina <dijuremo at gmail.com> wrote:> Show us output of: > > gluster v status > > Have you configured firewall rules properly for all ports being used? > > Diego > > On Thu, Jan 24, 2019 at 8:44 AM Gilberto Nunes <gilberto.nunes32 at gmail.com> > wrote: > >> >I think your mount statement in /etc/fstab is only referencing ONE of >> the gluster servers. >> > >> >Please take a look at "More redundant mount" section: >> > >> >https://www.jamescoyle.net/how-to/439-mount-a-glusterfs-volume >> > >> >Then try taking down one of the gluster servers and report back results. >> >> Guys! I have followed the very same instruction that found in the James's >> website. >> One of method his mentioned in that website, is create a file into >> /etc/glusterfs directory, named datastore.vol, for instance, with this >> content: >> >> volume remote1 >> type protocol/client >> option transport-type tcp >> option remote-host server1 >> option remote-subvolume /data/storage >> end-volume >> >> volume remote2 >> type protocol/client >> option transport-type tcp >> option remote-host server2 >> option remote-subvolume /data/storage >> end-volume >> >> volume remote3 >> type protocol/client >> option transport-type tcp >> option remote-host server3 >> option remote-subvolume /data/storage >> end-volume >> >> volume replicate >> type cluster/replicate >> subvolumes remote1 remote2 remote3 >> end-volume >> >> volume writebehind >> type performance/write-behind >> option window-size 1MB >> subvolumes replicate >> end-volume >> >> volume cache >> type performance/io-cache >> option cache-size 512MB >> subvolumes writebehind >> end-volume >> >> >> and then include this line into fstab: >> >> /etc/glusterfs/datastore.vol [MOUNT] glusterfs rw,allow_other, >> default_permissions,max_read=131072 0 0 >> >> What I doing wrong??? >> >> Thanks >> >> >> >> >> >> >> --- >> Gilberto Nunes Ferreira >> >> (47) 3025-5907 >> (47) 99676-7530 - Whatsapp / Telegram >> >> Skype: gilberto.nunes36 >> >> >> >> >> >> Em qui, 24 de jan de 2019 ?s 11:27, Scott Worthington < >> scott.c.worthington at gmail.com> escreveu: >> >>> I think your mount statement in /etc/fstab is only referencing ONE of >>> the gluster servers. >>> >>> Please take a look at "More redundant mount" section: >>> >>> https://www.jamescoyle.net/how-to/439-mount-a-glusterfs-volume >>> >>> Then try taking down one of the gluster servers and report back results. >>> >>> On Thu, Jan 24, 2019 at 8:24 AM Gilberto Nunes < >>> gilberto.nunes32 at gmail.com> wrote: >>> >>>> Yep! >>>> But as I mentioned in previously e-mail, even with 3 or 4 servers this >>>> issues occurr. >>>> I don't know what's happen. >>>> >>>> --- >>>> Gilberto Nunes Ferreira >>>> >>>> (47) 3025-5907 >>>> (47) 99676-7530 - Whatsapp / Telegram >>>> >>>> Skype: gilberto.nunes36 >>>> >>>> >>>> >>>> >>>> >>>> Em qui, 24 de jan de 2019 ?s 10:43, Diego Remolina <dijuremo at gmail.com> >>>> escreveu: >>>> >>>>> Glusterfs needs quorum, so if you have two servers and one goes down, >>>>> there is no quorum, so all writes stop until the server comes back up. You >>>>> can add a third server as an arbiter which does not store data in the >>>>> bricks, but still uses some minimal space (to keep metadata for the files). >>>>> >>>>> HTH, >>>>> >>>>> DIego >>>>> >>>>> On Wed, Jan 23, 2019 at 3:06 PM Gilberto Nunes < >>>>> gilberto.nunes32 at gmail.com> wrote: >>>>> >>>>>> Hit there... >>>>>> >>>>>> I have set up two server as replica, like this: >>>>>> >>>>>> gluster vol create Vol01 server1:/data/storage server2:/data/storage >>>>>> >>>>>> Then I create a config file in client, like this: >>>>>> volume remote1 >>>>>> type protocol/client >>>>>> option transport-type tcp >>>>>> option remote-host server1 >>>>>> option remote-subvolume /data/storage >>>>>> end-volume >>>>>> >>>>>> volume remote2 >>>>>> type protocol/client >>>>>> option transport-type tcp >>>>>> option remote-host server2 >>>>>> option remote-subvolume /data/storage >>>>>> end-volume >>>>>> >>>>>> volume replicate >>>>>> type cluster/replicate >>>>>> subvolumes remote1 remote2 >>>>>> end-volume >>>>>> >>>>>> volume writebehind >>>>>> type performance/write-behind >>>>>> option window-size 1MB >>>>>> subvolumes replicate >>>>>> end-volume >>>>>> >>>>>> volume cache >>>>>> type performance/io-cache >>>>>> option cache-size 512MB >>>>>> subvolumes writebehind >>>>>> end-volume >>>>>> >>>>>> And add this line in /etc/fstab >>>>>> >>>>>> /etc/glusterfs/datastore.vol /mnt glusterfs defaults,_netdev 0 0 >>>>>> >>>>>> After mount /mnt, I can access the servers. So far so good! >>>>>> But when I make server1 crash, I was unable to access /mnt or even >>>>>> use >>>>>> gluster vol status >>>>>> on server2 >>>>>> >>>>>> Everything hangon! >>>>>> >>>>>> I have tried with replicated, distributed and replicated-distributed >>>>>> too. >>>>>> I am using Debian Stretch, with gluster package installed via apt, >>>>>> provided by Standard Debian Repo, glusterfs-server 3.8.8-1 >>>>>> >>>>>> I am sorry if this is a newbie question, but glusterfs share it's >>>>>> not suppose to keep online if one server goes down? >>>>>> >>>>>> Any adviced will be welcome >>>>>> >>>>>> Best >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> --- >>>>>> Gilberto Nunes Ferreira >>>>>> >>>>>> (47) 3025-5907 >>>>>> (47) 99676-7530 - Whatsapp / Telegram >>>>>> >>>>>> Skype: gilberto.nunes36 >>>>>> >>>>>> >>>>>> >>>>>> _______________________________________________ >>>>>> Gluster-users mailing list >>>>>> Gluster-users at gluster.org >>>>>> https://lists.gluster.org/mailman/listinfo/gluster-users >>>>> >>>>> _______________________________________________ >>>> Gluster-users mailing list >>>> Gluster-users at gluster.org >>>> https://lists.gluster.org/mailman/listinfo/gluster-users >>> >>> _______________________________________________ > Gluster-users mailing list > Gluster-users at gluster.org > https://lists.gluster.org/mailman/listinfo/gluster-users-- Amar Tumballi (amarts) -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20190124/ec100172/attachment.html>
Gilberto Nunes
2019-Jan-24 13:55 UTC
[Gluster-users] Access to Servers hangs after stop one server...
Thanks, I'll check it out. --- Gilberto Nunes Ferreira (47) 3025-5907 (47) 99676-7530 - Whatsapp / Telegram Skype: gilberto.nunes36 Em qui, 24 de jan de 2019 ?s 11:50, Amar Tumballi Suryanarayan < atumball at redhat.com> escreveu:> Also note that, this way of mounting with a 'static' volfile is not > recommended as you wouldn't get any features out of gluster's Software > Defined Storage behavior. > > this was an approach we used to have say 8 years before. With the > introduction of management daemon called glusterd, the way of dealing with > volfiles have changed, and it is created with gluster CLI. > > About having /etc/fstab not hang when a server is down, search for > 'backup-volfile-server' option with glusterfs, and that should be used. > > Regards, > Amar > > On Thu, Jan 24, 2019 at 7:17 PM Diego Remolina <dijuremo at gmail.com> wrote: > >> Show us output of: >> >> gluster v status >> >> Have you configured firewall rules properly for all ports being used? >> >> Diego >> >> On Thu, Jan 24, 2019 at 8:44 AM Gilberto Nunes < >> gilberto.nunes32 at gmail.com> wrote: >> >>> >I think your mount statement in /etc/fstab is only referencing ONE of >>> the gluster servers. >>> > >>> >Please take a look at "More redundant mount" section: >>> > >>> >https://www.jamescoyle.net/how-to/439-mount-a-glusterfs-volume >>> > >>> >Then try taking down one of the gluster servers and report back results. >>> >>> Guys! I have followed the very same instruction that found in the >>> James's website. >>> One of method his mentioned in that website, is create a file into >>> /etc/glusterfs directory, named datastore.vol, for instance, with this >>> content: >>> >>> volume remote1 >>> type protocol/client >>> option transport-type tcp >>> option remote-host server1 >>> option remote-subvolume /data/storage >>> end-volume >>> >>> volume remote2 >>> type protocol/client >>> option transport-type tcp >>> option remote-host server2 >>> option remote-subvolume /data/storage >>> end-volume >>> >>> volume remote3 >>> type protocol/client >>> option transport-type tcp >>> option remote-host server3 >>> option remote-subvolume /data/storage >>> end-volume >>> >>> volume replicate >>> type cluster/replicate >>> subvolumes remote1 remote2 remote3 >>> end-volume >>> >>> volume writebehind >>> type performance/write-behind >>> option window-size 1MB >>> subvolumes replicate >>> end-volume >>> >>> volume cache >>> type performance/io-cache >>> option cache-size 512MB >>> subvolumes writebehind >>> end-volume >>> >>> >>> and then include this line into fstab: >>> >>> /etc/glusterfs/datastore.vol [MOUNT] glusterfs rw,allow_other, >>> default_permissions,max_read=131072 0 0 >>> >>> What I doing wrong??? >>> >>> Thanks >>> >>> >>> >>> >>> >>> >>> --- >>> Gilberto Nunes Ferreira >>> >>> (47) 3025-5907 >>> (47) 99676-7530 - Whatsapp / Telegram >>> >>> Skype: gilberto.nunes36 >>> >>> >>> >>> >>> >>> Em qui, 24 de jan de 2019 ?s 11:27, Scott Worthington < >>> scott.c.worthington at gmail.com> escreveu: >>> >>>> I think your mount statement in /etc/fstab is only referencing ONE of >>>> the gluster servers. >>>> >>>> Please take a look at "More redundant mount" section: >>>> >>>> https://www.jamescoyle.net/how-to/439-mount-a-glusterfs-volume >>>> >>>> Then try taking down one of the gluster servers and report back results. >>>> >>>> On Thu, Jan 24, 2019 at 8:24 AM Gilberto Nunes < >>>> gilberto.nunes32 at gmail.com> wrote: >>>> >>>>> Yep! >>>>> But as I mentioned in previously e-mail, even with 3 or 4 servers this >>>>> issues occurr. >>>>> I don't know what's happen. >>>>> >>>>> --- >>>>> Gilberto Nunes Ferreira >>>>> >>>>> (47) 3025-5907 >>>>> (47) 99676-7530 - Whatsapp / Telegram >>>>> >>>>> Skype: gilberto.nunes36 >>>>> >>>>> >>>>> >>>>> >>>>> >>>>> Em qui, 24 de jan de 2019 ?s 10:43, Diego Remolina <dijuremo at gmail.com> >>>>> escreveu: >>>>> >>>>>> Glusterfs needs quorum, so if you have two servers and one goes down, >>>>>> there is no quorum, so all writes stop until the server comes back up. You >>>>>> can add a third server as an arbiter which does not store data in the >>>>>> bricks, but still uses some minimal space (to keep metadata for the files). >>>>>> >>>>>> HTH, >>>>>> >>>>>> DIego >>>>>> >>>>>> On Wed, Jan 23, 2019 at 3:06 PM Gilberto Nunes < >>>>>> gilberto.nunes32 at gmail.com> wrote: >>>>>> >>>>>>> Hit there... >>>>>>> >>>>>>> I have set up two server as replica, like this: >>>>>>> >>>>>>> gluster vol create Vol01 server1:/data/storage server2:/data/storage >>>>>>> >>>>>>> Then I create a config file in client, like this: >>>>>>> volume remote1 >>>>>>> type protocol/client >>>>>>> option transport-type tcp >>>>>>> option remote-host server1 >>>>>>> option remote-subvolume /data/storage >>>>>>> end-volume >>>>>>> >>>>>>> volume remote2 >>>>>>> type protocol/client >>>>>>> option transport-type tcp >>>>>>> option remote-host server2 >>>>>>> option remote-subvolume /data/storage >>>>>>> end-volume >>>>>>> >>>>>>> volume replicate >>>>>>> type cluster/replicate >>>>>>> subvolumes remote1 remote2 >>>>>>> end-volume >>>>>>> >>>>>>> volume writebehind >>>>>>> type performance/write-behind >>>>>>> option window-size 1MB >>>>>>> subvolumes replicate >>>>>>> end-volume >>>>>>> >>>>>>> volume cache >>>>>>> type performance/io-cache >>>>>>> option cache-size 512MB >>>>>>> subvolumes writebehind >>>>>>> end-volume >>>>>>> >>>>>>> And add this line in /etc/fstab >>>>>>> >>>>>>> /etc/glusterfs/datastore.vol /mnt glusterfs defaults,_netdev 0 0 >>>>>>> >>>>>>> After mount /mnt, I can access the servers. So far so good! >>>>>>> But when I make server1 crash, I was unable to access /mnt or even >>>>>>> use >>>>>>> gluster vol status >>>>>>> on server2 >>>>>>> >>>>>>> Everything hangon! >>>>>>> >>>>>>> I have tried with replicated, distributed and replicated-distributed >>>>>>> too. >>>>>>> I am using Debian Stretch, with gluster package installed via apt, >>>>>>> provided by Standard Debian Repo, glusterfs-server 3.8.8-1 >>>>>>> >>>>>>> I am sorry if this is a newbie question, but glusterfs share it's >>>>>>> not suppose to keep online if one server goes down? >>>>>>> >>>>>>> Any adviced will be welcome >>>>>>> >>>>>>> Best >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> --- >>>>>>> Gilberto Nunes Ferreira >>>>>>> >>>>>>> (47) 3025-5907 >>>>>>> (47) 99676-7530 - Whatsapp / Telegram >>>>>>> >>>>>>> Skype: gilberto.nunes36 >>>>>>> >>>>>>> >>>>>>> >>>>>>> _______________________________________________ >>>>>>> Gluster-users mailing list >>>>>>> Gluster-users at gluster.org >>>>>>> https://lists.gluster.org/mailman/listinfo/gluster-users >>>>>> >>>>>> _______________________________________________ >>>>> Gluster-users mailing list >>>>> Gluster-users at gluster.org >>>>> https://lists.gluster.org/mailman/listinfo/gluster-users >>>> >>>> _______________________________________________ >> Gluster-users mailing list >> Gluster-users at gluster.org >> https://lists.gluster.org/mailman/listinfo/gluster-users > > > > -- > Amar Tumballi (amarts) >-------------- next part -------------- An HTML attachment was scrubbed... URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20190124/912a7295/attachment.html>
Scott Worthington
2019-Jan-24 14:04 UTC
[Gluster-users] Access to Servers hangs after stop one server...
Amar, Is this documentation relevant for Diego? https://gluster.readthedocs.io/en/latest/Administrator%20Guide/Setting%20Up%20Clients/#manual-mount "If backupvolfile-server option is added while mounting fuse client, when the first volfile server fails, then the server specified in backupvolfile-server option is used as volfile server to mount the client." Or is there 'better' documentation? On Thu, Jan 24, 2019 at 8:51 AM Amar Tumballi Suryanarayan < atumball at redhat.com> wrote:> Also note that, this way of mounting with a 'static' volfile is not > recommended as you wouldn't get any features out of gluster's Software > Defined Storage behavior. > > this was an approach we used to have say 8 years before. With the > introduction of management daemon called glusterd, the way of dealing with > volfiles have changed, and it is created with gluster CLI. > > About having /etc/fstab not hang when a server is down, search for > 'backup-volfile-server' option with glusterfs, and that should be used. > > Regards, > Amar > > On Thu, Jan 24, 2019 at 7:17 PM Diego Remolina <dijuremo at gmail.com> wrote: > >> Show us output of: >> >> gluster v status >> >> Have you configured firewall rules properly for all ports being used? >> >> Diego >> >> On Thu, Jan 24, 2019 at 8:44 AM Gilberto Nunes < >> gilberto.nunes32 at gmail.com> wrote: >> >>> >I think your mount statement in /etc/fstab is only referencing ONE of >>> the gluster servers. >>> > >>> >Please take a look at "More redundant mount" section: >>> > >>> >https://www.jamescoyle.net/how-to/439-mount-a-glusterfs-volume >>> > >>> >Then try taking down one of the gluster servers and report back results. >>> >>> Guys! I have followed the very same instruction that found in the >>> James's website. >>> One of method his mentioned in that website, is create a file into >>> /etc/glusterfs directory, named datastore.vol, for instance, with this >>> content: >>> >>> volume remote1 >>> type protocol/client >>> option transport-type tcp >>> option remote-host server1 >>> option remote-subvolume /data/storage >>> end-volume >>> >>> volume remote2 >>> type protocol/client >>> option transport-type tcp >>> option remote-host server2 >>> option remote-subvolume /data/storage >>> end-volume >>> >>> volume remote3 >>> type protocol/client >>> option transport-type tcp >>> option remote-host server3 >>> option remote-subvolume /data/storage >>> end-volume >>> >>> volume replicate >>> type cluster/replicate >>> subvolumes remote1 remote2 remote3 >>> end-volume >>> >>> volume writebehind >>> type performance/write-behind >>> option window-size 1MB >>> subvolumes replicate >>> end-volume >>> >>> volume cache >>> type performance/io-cache >>> option cache-size 512MB >>> subvolumes writebehind >>> end-volume >>> >>> >>> and then include this line into fstab: >>> >>> /etc/glusterfs/datastore.vol [MOUNT] glusterfs rw,allow_other, >>> default_permissions,max_read=131072 0 0 >>> >>> What I doing wrong??? >>> >>> Thanks >>> >>> >>> >>> >>> >>> >>> --- >>> Gilberto Nunes Ferreira >>> >>> (47) 3025-5907 >>> (47) 99676-7530 - Whatsapp / Telegram >>> >>> Skype: gilberto.nunes36 >>> >>> >>> >>> >>> >>> Em qui, 24 de jan de 2019 ?s 11:27, Scott Worthington < >>> scott.c.worthington at gmail.com> escreveu: >>> >>>> I think your mount statement in /etc/fstab is only referencing ONE of >>>> the gluster servers. >>>> >>>> Please take a look at "More redundant mount" section: >>>> >>>> https://www.jamescoyle.net/how-to/439-mount-a-glusterfs-volume >>>> >>>> Then try taking down one of the gluster servers and report back results. >>>> >>>> On Thu, Jan 24, 2019 at 8:24 AM Gilberto Nunes < >>>> gilberto.nunes32 at gmail.com> wrote: >>>> >>>>> Yep! >>>>> But as I mentioned in previously e-mail, even with 3 or 4 servers this >>>>> issues occurr. >>>>> I don't know what's happen. >>>>> >>>>> --- >>>>> Gilberto Nunes Ferreira >>>>> >>>>> (47) 3025-5907 >>>>> (47) 99676-7530 - Whatsapp / Telegram >>>>> >>>>> Skype: gilberto.nunes36 >>>>> >>>>> >>>>> >>>>> >>>>> >>>>> Em qui, 24 de jan de 2019 ?s 10:43, Diego Remolina <dijuremo at gmail.com> >>>>> escreveu: >>>>> >>>>>> Glusterfs needs quorum, so if you have two servers and one goes down, >>>>>> there is no quorum, so all writes stop until the server comes back up. You >>>>>> can add a third server as an arbiter which does not store data in the >>>>>> bricks, but still uses some minimal space (to keep metadata for the files). >>>>>> >>>>>> HTH, >>>>>> >>>>>> DIego >>>>>> >>>>>> On Wed, Jan 23, 2019 at 3:06 PM Gilberto Nunes < >>>>>> gilberto.nunes32 at gmail.com> wrote: >>>>>> >>>>>>> Hit there... >>>>>>> >>>>>>> I have set up two server as replica, like this: >>>>>>> >>>>>>> gluster vol create Vol01 server1:/data/storage server2:/data/storage >>>>>>> >>>>>>> Then I create a config file in client, like this: >>>>>>> volume remote1 >>>>>>> type protocol/client >>>>>>> option transport-type tcp >>>>>>> option remote-host server1 >>>>>>> option remote-subvolume /data/storage >>>>>>> end-volume >>>>>>> >>>>>>> volume remote2 >>>>>>> type protocol/client >>>>>>> option transport-type tcp >>>>>>> option remote-host server2 >>>>>>> option remote-subvolume /data/storage >>>>>>> end-volume >>>>>>> >>>>>>> volume replicate >>>>>>> type cluster/replicate >>>>>>> subvolumes remote1 remote2 >>>>>>> end-volume >>>>>>> >>>>>>> volume writebehind >>>>>>> type performance/write-behind >>>>>>> option window-size 1MB >>>>>>> subvolumes replicate >>>>>>> end-volume >>>>>>> >>>>>>> volume cache >>>>>>> type performance/io-cache >>>>>>> option cache-size 512MB >>>>>>> subvolumes writebehind >>>>>>> end-volume >>>>>>> >>>>>>> And add this line in /etc/fstab >>>>>>> >>>>>>> /etc/glusterfs/datastore.vol /mnt glusterfs defaults,_netdev 0 0 >>>>>>> >>>>>>> After mount /mnt, I can access the servers. So far so good! >>>>>>> But when I make server1 crash, I was unable to access /mnt or even >>>>>>> use >>>>>>> gluster vol status >>>>>>> on server2 >>>>>>> >>>>>>> Everything hangon! >>>>>>> >>>>>>> I have tried with replicated, distributed and replicated-distributed >>>>>>> too. >>>>>>> I am using Debian Stretch, with gluster package installed via apt, >>>>>>> provided by Standard Debian Repo, glusterfs-server 3.8.8-1 >>>>>>> >>>>>>> I am sorry if this is a newbie question, but glusterfs share it's >>>>>>> not suppose to keep online if one server goes down? >>>>>>> >>>>>>> Any adviced will be welcome >>>>>>> >>>>>>> Best >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> --- >>>>>>> Gilberto Nunes Ferreira >>>>>>> >>>>>>> (47) 3025-5907 >>>>>>> (47) 99676-7530 - Whatsapp / Telegram >>>>>>> >>>>>>> Skype: gilberto.nunes36 >>>>>>> >>>>>>> >>>>>>> >>>>>>> _______________________________________________ >>>>>>> Gluster-users mailing list >>>>>>> Gluster-users at gluster.org >>>>>>> https://lists.gluster.org/mailman/listinfo/gluster-users >>>>>> >>>>>> _______________________________________________ >>>>> Gluster-users mailing list >>>>> Gluster-users at gluster.org >>>>> https://lists.gluster.org/mailman/listinfo/gluster-users >>>> >>>> _______________________________________________ >> Gluster-users mailing list >> Gluster-users at gluster.org >> https://lists.gluster.org/mailman/listinfo/gluster-users > > > > -- > Amar Tumballi (amarts) > _______________________________________________ > Gluster-users mailing list > Gluster-users at gluster.org > https://lists.gluster.org/mailman/listinfo/gluster-users-------------- next part -------------- An HTML attachment was scrubbed... URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20190124/9e481308/attachment.html>