similar to: Arbiter and geo-replication

Displaying 20 results from an estimated 4000 matches similar to: "Arbiter and geo-replication"

2017 Sep 21
0
Arbiter and geo-replication
Hi all! Today I have a small gluster replication on 2 machines. My plan is to scale this, I though need some feedback that how I plan things is in the right direction. First of all I have understood the need of an arbiter. When I scale this, say that I just have 2 replica and 1 arbiter, when I add another two machines can I still use the same physical machine as the arbiter? Or when I add
2018 Mar 02
0
geo-replication
Hi Kotresh, I am expecting my hardware to show up next week. My plan is to run gluster version 3.12 on centos 7. Has the issue been fixed in version 3.12? Thanks a lot for your help! /Marcus On Fri, Mar 02, 2018 at 05:12:13PM +0530, Kotresh Hiremath Ravishankar wrote: > Hi Marcus, > > There are no issues with geo-rep and disperse volumes. It works with > disperse volume > being
2018 Mar 02
1
geo-replication
Hi again, I have been testing and reading up on other solutions and just wanted to check if my ideas are ok. I have been looking at dispersed volumes and wonder if there are any problems running replicated-distributed cluster on the master node and a dispersed-distributed cluster on the slave side of a geo-replication. Second thought, running disperesed on both sides, is that a problem (Master:
2018 Feb 07
1
geo-replication
Thank you for your help! Just to make things clear to me (and get a better understanding of gluster): So, if I make the slave cluster just distributed and node 1 goes down, data (say file.txt) that belongs to node 1 will not be synced. When node 1 comes back up does the master not realize that file.txt has not been synced and makes sure that it is synced when it has contact with node 1 again? So
2018 Feb 07
0
geo-replication
We are happy to help you out. Please find the answers inline. On Tue, Feb 6, 2018 at 4:39 PM, Marcus Peders?n <marcus.pedersen at slu.se> wrote: > Hi all, > > I am planning my new gluster system and tested things out in > a bunch of virtual machines. > I need a bit of help to understand how geo-replication behaves. > > I have a master gluster cluster replica 2 > (in
2018 Feb 21
0
Geo replication snapshot error
Hi, Thanks for reporting the issue. This seems to be a bug. Could you please raise a bug at https://bugzilla.redhat.com/ under community/glusterfs ? We will take a look at it and fix it. Thanks, Kotresh HR On Wed, Feb 21, 2018 at 2:01 PM, Marcus Peders?n <marcus.pedersen at slu.se> wrote: > Hi all, > I use gluster 3.12 on centos 7. > I am writing a snapshot program for my
2018 Feb 06
0
geo-replication
Hi again, I made some more tests and the behavior I get is that if any of the slaves are down the geo-replication stops working. It this the way distributed volumes work, if one server goes down the entire system stops to work? The servers that are online do not continue to work? Sorry, for asking stupid questions. Best regards Marcus On Tue, Feb 06, 2018 at 12:09:40PM +0100, Marcus Peders?n
2018 Jul 13
2
Upgrade to 4.1.1 geo-replication does not work
Hi Kotresh, Yes, all nodes have the same version 4.1.1 both master and slave. All glusterd are crashing on the master side. Will send logs tonight. Thanks, Marcus ################ Marcus Peders?n Systemadministrator Interbull Centre ################ Sent from my phone ################ Den 13 juli 2018 11:28 skrev Kotresh Hiremath Ravishankar <khiremat at redhat.com>: Hi Marcus, Is the
2018 Feb 06
4
geo-replication
Hi all, I am planning my new gluster system and tested things out in a bunch of virtual machines. I need a bit of help to understand how geo-replication behaves. I have a master gluster cluster replica 2 (in production I will use an arbiter and replicatied/distributed) and the geo cluster is distributed with 2 machines. (in production I will have the geo cluster distributed) Everything is up
2018 Feb 21
2
Geo replication snapshot error
Hi all, I use gluster 3.12 on centos 7. I am writing a snapshot program for my geo-replicated cluster. Now when I started to run tests with my application I have found a very strange behavior regarding geo-replication in gluster. I have setup my geo-replication according to the docs: http://docs.gluster.org/en/latest/Administrator%20Guide/Geo%20Replication/ Both master and slave clusters are
2017 Jul 29
1
Not possible to stop geo-rep after adding arbiter to replica 2
I managed to force stopping geo replication using the "force" parameter after the "stop" but there are still other issues related to the fact that my geo replication setup was created before I added the additional arbiter node to my replca. For example when I would like to stop my volume I simply can't and I get the following error: volume stop: myvolume: failed: Staging
2018 Apr 23
0
Geo-replication faulty
Hi all, I setup my gluster cluster with geo-replication a couple of weeks ago and everything worked fine! Today I descovered that one of the master nodes geo-replication status is faulty. On master side: Distributed-replicatied 2 x (2 + 1) = 6 On slave side: Replicated 1 x (2 + 1) = 3 After checking logs I see that the master node has the following error: OSError: Permission denied Looking at
2018 Jul 10
0
Geo replication manual rsync
Hi all, I have setup a gluster system with geo replication (Centos 7, gluster 3.12). I have moved about 30 TB to the cluster. It seems that it goes realy show for the data to be syncronized to geo replication. It has been active for weeks and still just 9TB has ended up on the slave side. I pause the replication once a day and make a snapshot with a script. Does this slow things up? Is it possible
2017 Jul 29
0
Not possible to stop geo-rep after adding arbiter to replica 2
Adding Rahul and Kothresh who are SME on geo replication Thanks & Regards Karan Sandha On Sat, Jul 29, 2017 at 3:37 PM, mabi <mabi at protonmail.ch> wrote: > Hello > > To my two node replica volume I have added an arbiter node for safety > purpose. On that volume I also have geo replication running and would like > to stop it is status "Faulty" and keeps
2017 Jul 29
2
Not possible to stop geo-rep after adding arbiter to replica 2
Hello To my two node replica volume I have added an arbiter node for safety purpose. On that volume I also have geo replication running and would like to stop it is status "Faulty" and keeps trying over and over to sync without success. I am using GlusterFS 3.8.11. So in order to stop geo-rep I use: gluster volume geo-replication myvolume gfs1geo.domain.tld::myvolume-geo stop but it
2023 Feb 21
2
Gluster 11.0 upgrade
Hi Xavi, Copy the same info file worked well and the gluster 11 arbiter is now up and running and all the nodes are communication the way they should. Just another note on something I discovered on my virt machines. All the three nodes has been upgarded to 11.0 and are working. If I run: gluster volume get all cluster.op-version I get: Option Value ------
2018 Jan 23
0
Understanding client logs
Marcus, Please paste the name-version-release of the primary glusterfs package on your system. If possible, also describe the typical workload that happens at the mount via the user application. On Tue, Jan 23, 2018 at 7:43 PM, Marcus Peders?n <marcus.pedersen at slu.se> wrote: > Hi all, > I have problem pin pointing an error, that users of > my system experience processes that
2018 Apr 03
1
Tune and optimize dispersed cluster
Hi all, I have setup a dispersed cluster (2+1), version 3.12. The way our users run I guessed that we would get the penalties with dispersed cluster and I was right.... A calculation that usually takes about 48 hours (on a replicaited cluster), now took about 60 hours. There is alot of "small" reads/writes going on in these programs. Is there a way to tune, optimize a dispersed cluster
2018 Jan 23
2
Understanding client logs
Hi all, I have problem pin pointing an error, that users of my system experience processes that crash. The thing that have changed since the craches started is that I added a gluster cluster. Of cause the users start to attack my gluster cluster. I started looking at logs, starting from the client side. I just need help to understand how to read it in the right way. I can see that every ten
2023 Feb 20
2
Gluster 11.0 upgrade
Hi again Xavi, I did some more testing on my virt machines with same setup: Number of Bricks: 1 x (2 + 1) = 3 If I do it the same way, I upgrade the arbiter first, I get the same behavior that the bricks do not start and the other nodes does not "see" the upgraded node. If I upgrade one of the other nodes (non arbiter) and restart glusterd on both the arbiter and the other the arbiter