Jiffin Tony Thottan
2017-May-03 05:49 UTC
[Gluster-users] "gluster volume set all cluster.enable-shared-storage enable"
On 02/05/17 15:27, hvjunk wrote:> Good day, > > I?m busy setting up/testing NFS-HA with GlusterFS storage across VMs running Debian 8. GlusterFS volume to be "replica 3 arbiter 1" > > In the NFS-ganesha information I?ve gleamed thus far, it mentions the "gluster volume set all cluster.enable-shared-storage enable?. > > My first question is this: is that shared volume that gets created/setup, suppose to be resilient across reboots? > It appears to not be the case in my test setup thus far, that that mount doesn?t get recreated/remounted after a reboot.Following is the script which creates shared storage and mount it in the node, plus an entry will be added to /etc/fstab https://github.com/gluster/glusterfs/blob/master/extras/hook-scripts/set/post/S32gluster_enable_shared_storage.sh But there is a possibility such that, if glusterd(I hope u have enabled enabled glusterd service) is not started before systemd tries mount the shared storage then it will fail. -- Jiffin> > If the mount is not resilient, ie. not recreated/mounted by glusterfs and neither added to the /etc/fstab by glusterfs, why the initial auto mount by glusterfs and not afterwards with a reboot? > > The biggest ?issue? I have found with glusterfs is the interaction with SystemD and mounts that fails and don?t get properly retried later (Will email separately on that issue) during bootstrapping of the cluster, and that is why I need to confirm the reasoning/etc. on this initial auto-mounting, but then the need to manually add it into the /etc/fstab > > Thank you > Hendrik > > _______________________________________________ > Gluster-users mailing list > Gluster-users at gluster.org > http://lists.gluster.org/mailman/listinfo/gluster-users
hvjunk
2017-May-09 13:48 UTC
[Gluster-users] bootstrapping cluster "failure" condition fix for local mounts (like: "gluster volume set all cluster.enable-shared-storage enable")
> On 03 May 2017, at 07:49 , Jiffin Tony Thottan <jthottan at redhat.com> wrote: > > > > On 02/05/17 15:27, hvjunk wrote: >> Good day, >> >> I?m busy setting up/testing NFS-HA with GlusterFS storage across VMs running Debian 8. GlusterFS volume to be "replica 3 arbiter 1" >> >> In the NFS-ganesha information I?ve gleamed thus far, it mentions the "gluster volume set all cluster.enable-shared-storage enable?. >> >> My first question is this: is that shared volume that gets created/setup, suppose to be resilient across reboots? >> It appears to not be the case in my test setup thus far, that that mount doesn?t get recreated/remounted after a reboot. > > Following is the script which creates shared storage and mount it in the node, plus an entry will be added to /etc/fstab > https://github.com/gluster/glusterfs/blob/master/extras/hook-scripts/set/post/S32gluster_enable_shared_storage.sh > > But there is a possibility such that, if glusterd(I hope u have enabled enabled glusterd service) is not started before > systemd tries mount the shared storage then it will fail.Thank Jiffin, I since found that (1) you need to wait a bit for the cluster to ?settle? with that script having executed, before you reboot the cluster (As you might see in my bitbucket ansible scripts in https://bitbucket.org/dismyne/gluster-ansibles/src <https://bitbucket.org/dismyne/gluster-ansibles/src> ) ? something to add in the manuals perhaps to warn people to wait for that script to finish before rebooting node/vm/server(s)? (2) the default configuration, can?t bootstrap the /gluster_shared_storage volume/directory reliably from a clean shutdown-reboot of the whole cluster!!! The problem: SystemD and it?s wanting to have the control over /etc/fstab and the mounting, and and and?. (and I?ll not empty my mind about L.P. based on his remarks in: https://github.com/systemd/systemd/issues/4468#issuecomment-255711912 <https://github.com/systemd/systemd/issues/4468#issuecomment-255711912> after my struggling with this issue) To have a reliably bootstrapped (from all nodes down booting up) I'm using the following SystemD service and helper script(s) to have the gluster cluster node mount their local mounts (like /gluster_shared_storage) reliably: https://bitbucket.org/dismyne/gluster-ansibles/src/24b62dcc858364ee3744d351993de0e8e35c2680/ansible/files/glusterfsmounts.service-centos?at=master <https://bitbucket.org/dismyne/gluster-ansibles/src/24b62dcc858364ee3744d351993de0e8e35c2680/ansible/files/glusterfsmounts.service-centos?at=master> https://bitbucket.org/dismyne/gluster-ansibles/src/24b62dcc858364ee3744d351993de0e8e35c2680/ansible/files/test-mounts.sh?at=master <https://bitbucket.org/dismyne/gluster-ansibles/src/24b62dcc858364ee3744d351993de0e8e35c2680/ansible/files/test-mounts.sh?at=master>> > -- > Jiffin >> >> If the mount is not resilient, ie. not recreated/mounted by glusterfs and neither added to the /etc/fstab by glusterfs, why the initial auto mount by glusterfs and not afterwards with a reboot? >> >> The biggest ?issue? I have found with glusterfs is the interaction with SystemD and mounts that fails and don?t get properly retried later (Will email separately on that issue) during bootstrapping of the cluster, and that is why I need to confirm the reasoning/etc. on this initial auto-mounting, but then the need to manually add it into the /etc/fstab >> >> Thank you >> Hendrik >> >> _______________________________________________ >> Gluster-users mailing list >> Gluster-users at gluster.org >> http://lists.gluster.org/mailman/listinfo/gluster-users >-------------- next part -------------- An HTML attachment was scrubbed... URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20170509/cfa03f3c/attachment.html> -------------- next part -------------- A non-text attachment was scrubbed... Name: glusterfsmounts.service-centos Type: application/octet-stream Size: 416 bytes Desc: not available URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20170509/cfa03f3c/attachment.obj> -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20170509/cfa03f3c/attachment-0001.html> -------------- next part -------------- A non-text attachment was scrubbed... Name: test-mounts.sh Type: application/octet-stream Size: 1546 bytes Desc: not available URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20170509/cfa03f3c/attachment-0001.obj> -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20170509/cfa03f3c/attachment-0002.html>