similar to: Kernel NFS on GlusterFS

Displaying 20 results from an estimated 20000 matches similar to: "Kernel NFS on GlusterFS"

2018 Mar 07
0
Kernel NFS on GlusterFS
Gluster does the sync part better than corosync. It's not an active/passive failover system. It more all active. Gluster handles the recovery once all nodes are back online. That requires the client tool chain to understand that a write goes to all storage devices not just the active one. 3.10 is a long term support release. Upgrading to 3.12 or 4 is not a significant issue once a replacement
2017 Oct 05
2
Access from multiple hosts where users have different uid/gid
I have a setup with multiple hosts, each of them are administered separately. So there are no unified uid/gid for the users. When mounting a GlusterFS volume, a file owned by user1 on host1 might become owned by user2 on host2. I was looking into POSIX ACL or bindfs, but that won't help me much. What did other people do with this kind of problem? -------------- next part -------------- An
2017 Dec 20
2
glusterfs, ganesh, and pcs rules
Hi, I've just created again the gluster with NFS ganesha. Glusterfs version 3.8 When I run the command gluster nfs-ganesha enable - it returns a success. However, looking at the pcs status, I see this: [root at tlxdmz-nfs1 ~]# pcs status Cluster name: ganesha-nfs Stack: corosync Current DC: tlxdmz-nfs2 (version 1.1.16-12.el7_4.5-94ff4df) - partition with quorum Last updated: Wed Dec 20
2017 Dec 24
1
glusterfs, ganesh, and pcs rules
I checked, and I have it like this: # Name of the HA cluster created. # must be unique within the subnet HA_NAME="ganesha-nfs" # # The gluster server from which to mount the shared data volume. HA_VOL_SERVER="tlxdmz-nfs1" # # N.B. you may use short names or long names; you may not use IP addrs. # Once you select one, stay with it as it will be mildly unpleasant to # clean up
2018 May 30
1
peer detach fails
All, I added a third peer for a arbiter brick host to replica 2 cluster. Then I realized I can't use it since it has no infiniband like the other two hosts (infiniband and ethernet for clients). So I removed the new arbiter bricks from all of the volumes. However, I can't detach the peer as it keeps saying there are bricks it hosts. Nothing in volume status or info shows that host to be
2017 Jun 05
2
Gluster and NFS-Ganesha - cluster is down after reboot
Hi hvjunk, could you please tell me have you had time to check my previous post? Could you please send me mentioned link to your Gluster Ansible scripts? Thank you, Adam On Sun, May 28, 2017 at 2:47 PM, Adam Ru <ad.ruckel at gmail.com> wrote: > Hi hvjunk (Hi Hendrik), > > "centos-release-gluster" installs "centos-gluster310". I assume it > picks the
2018 May 04
2
shard corruption bug
Il giorno ven 4 mag 2018 alle ore 14:06 Jim Kinney <jim.kinney at gmail.com> ha scritto: > It stopped being an outstanding issue at 3.12.7. I think it's now fixed. So, is not possible to extend and rebalance a working cluster with sharded data ? Can someone confirm this ? Maybe the ones that hit the bug in the past
2018 May 30
2
shard corruption bug
What shard corruption bug? bugzilla url? I'm running into some odd behavior in my lab with shards and RHEV/KVM data, trying to figure out if it's related. Thanks. On Fri, May 4, 2018 at 11:13 AM, Jim Kinney <jim.kinney at gmail.com> wrote: > I upgraded my ovirt stack to 3.12.9, added a brick to a volume and left it > to settle. No problems. I am now running replica 4
2017 Dec 21
0
glusterfs, ganesh, and pcs rules
Hi, In your ganesha-ha.conf do you have your virtual ip adresses set something like this?: VIP_tlxdmz-nfs1="192.168.22.33" VIP_tlxdmz-nfs2="192.168.22.34" Renaud De?: gluster-users-bounces at gluster.org [mailto:gluster-users-bounces at gluster.org] De la part de Hetz Ben Hamo Envoy??: 20 d?cembre 2017 04:35 ??: gluster-users at gluster.org Objet?: [Gluster-users]
2017 Jun 05
0
Gluster and NFS-Ganesha - cluster is down after reboot
Sorry, got sidetracked with invoicing etc. https://bitbucket.org/dismyne/gluster-ansibles/src/6df23803df43/ansible/files/?at=master <https://bitbucket.org/dismyne/gluster-ansibles/src/6df23803df43/ansible/files/?at=master> The .service files are the stuff going into SystemD, and they call the test-mounts.sh scripts. The playbook installing higher up in the directory > On 05 Jun 2017,
2017 Jun 06
1
Gluster and NFS-Ganesha - cluster is down after reboot
----- Original Message ----- From: "hvjunk" <hvjunk at gmail.com> To: "Adam Ru" <ad.ruckel at gmail.com> Cc: gluster-users at gluster.org Sent: Monday, June 5, 2017 9:29:03 PM Subject: Re: [Gluster-users] Gluster and NFS-Ganesha - cluster is down after reboot Sorry, got sidetracked with invoicing etc.
2018 Feb 19
2
NFS Ganesha HA w/ GlusterFS
On 2/19/2018 2:39 AM, TomK wrote: + gluster users as well. Just read another post on the mailing lists about a similar ask from Nov which didn't really have a clear answer. Perhaps there's a way to get NFSv4 work with GlusterFS without NFS Ganesha then? Cheers, Tom > Hey All, > > I've setup GlusterFS on two virtuals and enabled NFS Ganesha on each > node.? ATM the
2018 Feb 19
2
NFS Ganesha HA w/ GlusterFS
On 2/19/2018 10:55 AM, Kaleb S. KEITHLEY wrote: Yep, I noticed a couple of pages including this for 'storhaug configuration' off google. Adding 'mailing list' to the search didn't help alot: https://sourceforge.net/p/nfs-ganesha/mailman/message/35929089/ https://www.spinics.net/lists/gluster-users/msg33018.html Hence the ask here. storhaug feels like it's not moving
2017 Dec 07
4
GlusterFS, Pacemaker, OCF resource agents on CentOS 7
Hi guys I'm wondering if anyone here is using the GlusterFS OCF resource agents with Pacemaker on CentOS 7? yum install centos-release-gluster yum install glusterfs-server glusterfs-resource-agents The reason I ask is that there seem to be a few problems with them on 3.10, but these problems are so severe that I'm struggling to believe I'm not just doing something wrong. I created
2018 May 04
0
shard corruption bug
I upgraded my ovirt stack to 3.12.9, added a brick to a volume and left it to settle. No problems. I am now running replica 4 (preparing to remove a brick and host to replica 3). On Fri, 2018-05-04 at 14:24 +0000, Gandalf Corvotempesta wrote: > Il giorno ven 4 mag 2018 alle ore 14:06 Jim Kinney <jim.kinney at gmail. > com> > ha scritto: > > It stopped being an outstanding
2018 Feb 19
3
NFS Ganesha HA w/ GlusterFS
On 2/19/2018 12:09 PM, Kaleb S. KEITHLEY wrote: Sounds good and no problem at all. Will look out for this update in the future. In the meantime, three's a few things I'll try including your suggestion. Was looking for a sense of direction with the projects and now you've given that. Ty. Appreciated! Cheers, Tom > On 02/19/2018 11:37 AM, TomK wrote: >> On 2/19/2018
2018 Feb 19
0
NFS Ganesha HA w/ GlusterFS
On 02/19/2018 10:24 AM, TomK wrote: > On 2/19/2018 2:39 AM, TomK wrote: > + gluster users as well.? Just read another post on the mailing lists > about a similar ask from Nov which didn't really have a clear answer. That's funny because I've answered questions like this several times. Gluster+Ganesha+Pacemaker-based HA is available up to GlusterFS 3.10.x. If you need HA,
2018 May 30
0
shard corruption bug
https://docs.gluster.org/en/latest/release-notes/3.12.6/ The major issue in 3.12.6 is not present in 3.12.7. Bugzilla ID listed in link. On May 29, 2018 8:50:56 PM EDT, Dan Lavu <dan at redhat.com> wrote: >What shard corruption bug? bugzilla url? I'm running into some odd >behavior >in my lab with shards and RHEV/KVM data, trying to figure out if it's >related. >
2017 Dec 08
0
GlusterFS, Pacemaker, OCF resource agents on CentOS 7
Hi, Can u please explain for what purpose pacemaker cluster used here? Regards, Jiffin On Thursday 07 December 2017 06:59 PM, Tomalak Geret'kal wrote: > > Hi guys > > I'm wondering if anyone here is using the GlusterFS OCF resource > agents with Pacemaker on CentOS 7? > > yum install centos-release-gluster > yum install glusterfs-server
2018 Feb 19
0
NFS Ganesha HA w/ GlusterFS
On 02/19/2018 11:37 AM, TomK wrote: > On 2/19/2018 10:55 AM, Kaleb S. KEITHLEY wrote: > Yep, I noticed a couple of pages including this for 'storhaug > configuration' off google.? Adding 'mailing list' to the search didn't > help alot: > > https://sourceforge.net/p/nfs-ganesha/mailman/message/35929089/ > >