search for: quorum

Displaying 20 results from an estimated 389 matches for "quorum".

2017 Dec 21
0
Gluster replicate 3 arbiter 1 in split brain. gluster cli seems unaware
...uota: on features.quota: on features.barrier: disable features.scrub: Active features.bitrot: on nfs.rpc-auth-allow: on server.allow-insecure: on user.cifs: off features.shard: off cluster.shd-wait-qlength: 10000 cluster.locking-scheme: granular cluster.data-self-heal-algorithm: full cluster.server-quorum-type: server cluster.quorum-type: auto cluster.eager-lock: enable network.remote-dio: enable performance.low-prio-threads: 32 performance.io-cache: off performance.read-ahead: off performance.quick-read: off nfs.disable: on transport.address-family: inet server.outstanding-rpc-limit: 512 > Why...
2017 Dec 22
2
Gluster replicate 3 arbiter 1 in split brain. gluster cli seems unaware
...ble > features.scrub: Active > features.bitrot: on > nfs.rpc-auth-allow: on > server.allow-insecure: on > user.cifs: off > features.shard: off > cluster.shd-wait-qlength: 10000 > cluster.locking-scheme: granular > cluster.data-self-heal-algorithm: full > cluster.server-quorum-type: server > cluster.quorum-type: auto > cluster.eager-lock: enable > network.remote-dio: enable > performance.low-prio-threads: 32 > performance.io-cache: off > performance.read-ahead: off > performance.quick-read: off > nfs.disable: on > transport.address-family: inet...
2017 Dec 21
2
Gluster replicate 3 arbiter 1 in split brain. gluster cli seems unaware
...t; > Compare the hashes of the file on both bricks to see if the data actually > differs. If the data is the same it makes choosing the proper replica > easier. > > Any idea how you got in this situation? Did you have a loss of NW > connectivity? I see you are using server side quorum, maybe check the logs > for any loss of quorum? I wonder if there was a loos of quorum and there > was some sort of race condition hit: > > http://docs.gluster.org/en/latest/Administrator%20Guide/ > arbiter-volumes-and-quorum/#server-quorum-and-some-pitfalls > > "Unlike i...
2017 Dec 22
0
Gluster replicate 3 arbiter 1 in split brain. gluster cli seems unaware
.../fedora27.qcow2 Status: Connected Number of entries: 1 Now heal info shows 0 entries, and the two data bricks have the same md5sum, so it's back in sync. I have a few questions after all of this: 1) How can a split brain happen in a replica 3 arbiter 1 setup with both server- and client quorum enabled? 2) Why was it not able to self heal, when tro bricks seemed in sync with their changelogs? 3) Why could I not see the file in heal info split-brain? 4) Why could I not fix this through the cli split-brain resolution tool? 5) Is it possible to force a sync in a volume? Or maybe test sync st...
2017 Dec 22
0
Gluster replicate 3 arbiter 1 in split brain. gluster cli seems unaware
...1 > > > > Now heal info shows 0 entries, and the two data bricks have the same > md5sum, so it's back in sync. > > > > I have a few questions after all of this: > > 1) How can a split brain happen in a replica 3 arbiter 1 setup with > both server- and client quorum enabled? > 2) Why was it not able to self heal, when tro bricks seemed in sync > with their changelogs? > 3) Why could I not see the file in heal info split-brain? > 4) Why could I not fix this through the cli split-brain resolution tool? > 5) Is it possible to force a sync in a volu...
2018 Feb 26
2
Quorum in distributed-replicate volume
On Mon, Feb 26, 2018 at 05:45:27PM +0530, Karthik Subrahmanya wrote: > > "In a replica 2 volume... If we set the client-quorum option to > > auto, then the first brick must always be up, irrespective of the > > status of the second brick. If only the second brick is up, the > > subvolume becomes read-only." > > > By default client-quorum is "none" in replica 2 volume. I'm not...
2018 Feb 26
2
Quorum in distributed-replicate volume
I've configured 6 bricks as distributed-replicated with replica 2, expecting that all active bricks would be usable so long as a quorum of at least 4 live bricks is maintained. However, I have just found http://docs.gluster.org/en/latest/Administrator%20Guide/Split%20brain%20and%20ways%20to%20deal%20with%20it/ Which states that "In a replica 2 volume... If we set the client-quorum option to auto, then the first brick must a...
2018 Feb 26
0
Quorum in distributed-replicate volume
Hi Dave, On Mon, Feb 26, 2018 at 4:45 PM, Dave Sherohman <dave at sherohman.org> wrote: > I've configured 6 bricks as distributed-replicated with replica 2, > expecting that all active bricks would be usable so long as a quorum of > at least 4 live bricks is maintained. > The client quorum is configured per replica sub volume and not for the entire volume. Since you have a distributed-replicated volume with replica 2, the data will have 2 copies, and considering your scenario of quorum to be taken on the total numbe...
2018 Feb 27
0
Quorum in distributed-replicate volume
On Mon, Feb 26, 2018 at 6:14 PM, Dave Sherohman <dave at sherohman.org> wrote: > On Mon, Feb 26, 2018 at 05:45:27PM +0530, Karthik Subrahmanya wrote: > > > "In a replica 2 volume... If we set the client-quorum option to > > > auto, then the first brick must always be up, irrespective of the > > > status of the second brick. If only the second brick is up, the > > > subvolume becomes read-only." > > > > > By default client-quorum is "none" in replic...
2005 Apr 17
2
Quorum error
Had a problem starting Oracle after expanding an EMC Metalun. We get the following errors: >WARNING: OemInit2: Opened file(/oradata/dbf/quorum.dbf 8), tid = main:1024 file = oem.c, line = 491 {Sun Apr 17 10:33:41 2005 } >ERROR: ReadOthersDskInfo(): ReadFile(/oradata/dbf/quorum.dbf) failed(5) - (0) bytes read, tid = main:1024 file = oem.c, line = 1396 {Sun Apr 17 10:33:41 2005 } >ERROR: UpdateDskNodeInfo(): WriteFile(/oradata/...
2009 May 04
2
FW: Oracle 9204 installation on linux x86-64 on ocfs
Hello All, I have installed Oracle Cluster Manager on linux x86-64 nit. I am using ocfs file system for quorum file. But I am getting following error. Please see ocfs configureation below. I would appreciate, if someone could help me to understand if I am doing something wrong. Thanks in advance. --------------------------------------------------cm.log file ---------------------------- oracm, version[ AMD64...
2006 Jan 09
0
[PATCH 01/11] ocfs2: event-driven quorum
This patch separates o2net and o2quo from knowing about one another as much as possible. This is the first in a series of patches that will allow userspace cluster interaction. Quorum is separated out first, and will ultimately only be associated with the disk heartbeat as a separate module. To do so, this patch performs the following changes: * o2hb_notify() is added to handle injection of events in a synchronous manner. All locking is preserved as expected. * disk hear...
2017 Jun 10
4
How to remove dead peer, osrry urgent again :(
Since my node died on friday I have a dead peer (vna) that needs to be removed. I had major issues this morning that I haven't resolve yet with all VM's going offline when I rebooted a node which I *hope * was due to quorum issues as I now have four peers in the cluster, one dead, three live. Confidence level is not high. -- Lindsay Mathieson
2017 Jul 04
2
I need a sanity check.
2006 Oct 13
1
Cluster Quorum Question/Problem
...etings all, I am in need of professional insight. I have a 2 node cluster running CentOS, mysql, apache, etc. I have on each system a fiber HBA connected to a fiber SAN. Each system shows the devices sdb and sdc for each of the connections on the HBA. I have sdc1 mounted on both machines as /quorum. When I right to the /quorum from one of the nodes, the file doesn't show up on the other. However, if I unmount and remount the device on the system that can't see the new file, the file shows up, but is broken/red and cannot be accessed by the node that couldn't see it in the fi...
2017 Sep 22
2
AFR: Fail lookups when quorum not met
...The proposed fix [2] was to fail look ups when we definitely know the lookup can't be trusted (by virtue of AFR xattrs indicating the replies we got from the up bricks are indeed bad). Note that this fix is *only* for replica 3 or arbiter volumes (not replica 2, where there is no notion of quorum). But we want to 'harden' the fix by? not allowing any look ups at all if quorum is not met (or) it is met but there are no good copies. Some implications of this: -If a file ends up in data/meta data split-brain in replica 3/arbiter (rare occurrence), we won't be able to delete it...
2018 Feb 27
2
Quorum in distributed-replicate volume
On Tue, Feb 27, 2018 at 12:00:29PM +0530, Karthik Subrahmanya wrote: > I will try to explain how you can end up in split-brain even with cluster > wide quorum: Yep, the explanation made sense. I hadn't considered the possibility of alternating outages. Thanks! > > > It would be great if you can consider configuring an arbiter or > > > replica 3 volume. > > > > I can. My bricks are 2x850G and 4x11T, so I can repurpos...
2007 Aug 09
0
About quorum and fencing
Hi, In the ocfs2 FAQ, it is written: "A node has quorum when: * it sees an odd number of heartbeating nodes and has network connectivity to more than half of them. OR, * it sees an even number of heartbeating nodes and has network connectivity to at least half of them *and* has connectivity to the heartbeating...
2017 Oct 09
0
[Gluster-devel] AFR: Fail lookups when quorum not met
...s when we definitely know the lookup can't be trusted (by >> virtue of AFR xattrs indicating the replies we got from the up bricks are >> indeed bad). >> >> Note that this fix is *only* for replica 3 or arbiter volumes (not replica >> 2, where there is no notion of quorum). But we want to 'harden' the fix by >> not allowing any look ups at all if quorum is not met (or) it is met but >> there are no good copies. >> >> Some implications of this: >> >> -If a file ends up in data/meta data split-brain in replica 3/arbiter (rar...
2017 Dec 21
0
Gluster replicate 3 arbiter 1 in split brain. gluster cli seems unaware
...;path to brick>/pat/to/file Compare the hashes of the file on both bricks to see if the data actually differs. If the data is the same it makes choosing the proper replica easier. Any idea how you got in this situation? Did you have a loss of NW connectivity? I see you are using server side quorum, maybe check the logs for any loss of quorum? I wonder if there was a loos of quorum and there was some sort of race condition hit: http://docs.gluster.org/en/latest/Administrator%20Guide/arbiter-volumes-and-quorum/#server-quorum-and-some-pitfalls "Unlike in client-quorum where the volume b...