Displaying 20 results from an estimated 397 matches for "quorum".
2017 Dec 21
0
Gluster replicate 3 arbiter 1 in split brain. gluster cli seems unaware
...uota: on
features.quota: on
features.barrier: disable
features.scrub: Active
features.bitrot: on
nfs.rpc-auth-allow: on
server.allow-insecure: on
user.cifs: off
features.shard: off
cluster.shd-wait-qlength: 10000
cluster.locking-scheme: granular
cluster.data-self-heal-algorithm: full
cluster.server-quorum-type: server
cluster.quorum-type: auto
cluster.eager-lock: enable
network.remote-dio: enable
performance.low-prio-threads: 32
performance.io-cache: off
performance.read-ahead: off
performance.quick-read: off
nfs.disable: on
transport.address-family: inet
server.outstanding-rpc-limit: 512
> Why...
2017 Dec 22
2
Gluster replicate 3 arbiter 1 in split brain. gluster cli seems unaware
...ble
> features.scrub: Active
> features.bitrot: on
> nfs.rpc-auth-allow: on
> server.allow-insecure: on
> user.cifs: off
> features.shard: off
> cluster.shd-wait-qlength: 10000
> cluster.locking-scheme: granular
> cluster.data-self-heal-algorithm: full
> cluster.server-quorum-type: server
> cluster.quorum-type: auto
> cluster.eager-lock: enable
> network.remote-dio: enable
> performance.low-prio-threads: 32
> performance.io-cache: off
> performance.read-ahead: off
> performance.quick-read: off
> nfs.disable: on
> transport.address-family: inet...
2017 Dec 21
2
Gluster replicate 3 arbiter 1 in split brain. gluster cli seems unaware
...t;
> Compare the hashes of the file on both bricks to see if the data actually
> differs. If the data is the same it makes choosing the proper replica
> easier.
>
> Any idea how you got in this situation? Did you have a loss of NW
> connectivity? I see you are using server side quorum, maybe check the logs
> for any loss of quorum? I wonder if there was a loos of quorum and there
> was some sort of race condition hit:
>
> http://docs.gluster.org/en/latest/Administrator%20Guide/
> arbiter-volumes-and-quorum/#server-quorum-and-some-pitfalls
>
> "Unlike i...
2017 Dec 22
0
Gluster replicate 3 arbiter 1 in split brain. gluster cli seems unaware
.../fedora27.qcow2
Status: Connected
Number of entries: 1
Now heal info shows 0 entries, and the two data bricks have the same
md5sum, so it's back in sync.
I have a few questions after all of this:
1) How can a split brain happen in a replica 3 arbiter 1 setup with
both server- and client quorum enabled?
2) Why was it not able to self heal, when tro bricks seemed in sync
with their changelogs?
3) Why could I not see the file in heal info split-brain?
4) Why could I not fix this through the cli split-brain resolution tool?
5) Is it possible to force a sync in a volume? Or maybe test sync
st...
2017 Dec 22
0
Gluster replicate 3 arbiter 1 in split brain. gluster cli seems unaware
...1
>
>
>
> Now heal info shows 0 entries, and the two data bricks have the same
> md5sum, so it's back in sync.
>
>
>
> I have a few questions after all of this:
>
> 1) How can a split brain happen in a replica 3 arbiter 1 setup with
> both server- and client quorum enabled?
> 2) Why was it not able to self heal, when tro bricks seemed in sync
> with their changelogs?
> 3) Why could I not see the file in heal info split-brain?
> 4) Why could I not fix this through the cli split-brain resolution tool?
> 5) Is it possible to force a sync in a volu...
2018 Feb 26
2
Quorum in distributed-replicate volume
On Mon, Feb 26, 2018 at 05:45:27PM +0530, Karthik Subrahmanya wrote:
> > "In a replica 2 volume... If we set the client-quorum option to
> > auto, then the first brick must always be up, irrespective of the
> > status of the second brick. If only the second brick is up, the
> > subvolume becomes read-only."
> >
> By default client-quorum is "none" in replica 2 volume.
I'm not...
2018 Feb 26
2
Quorum in distributed-replicate volume
I've configured 6 bricks as distributed-replicated with replica 2,
expecting that all active bricks would be usable so long as a quorum of
at least 4 live bricks is maintained.
However, I have just found
http://docs.gluster.org/en/latest/Administrator%20Guide/Split%20brain%20and%20ways%20to%20deal%20with%20it/
Which states that "In a replica 2 volume... If we set the client-quorum
option to auto, then the first brick must a...
2018 Feb 26
0
Quorum in distributed-replicate volume
Hi Dave,
On Mon, Feb 26, 2018 at 4:45 PM, Dave Sherohman <dave at sherohman.org> wrote:
> I've configured 6 bricks as distributed-replicated with replica 2,
> expecting that all active bricks would be usable so long as a quorum of
> at least 4 live bricks is maintained.
>
The client quorum is configured per replica sub volume and not for the
entire volume.
Since you have a distributed-replicated volume with replica 2, the data
will have 2 copies,
and considering your scenario of quorum to be taken on the total numbe...
2018 Feb 27
0
Quorum in distributed-replicate volume
On Mon, Feb 26, 2018 at 6:14 PM, Dave Sherohman <dave at sherohman.org> wrote:
> On Mon, Feb 26, 2018 at 05:45:27PM +0530, Karthik Subrahmanya wrote:
> > > "In a replica 2 volume... If we set the client-quorum option to
> > > auto, then the first brick must always be up, irrespective of the
> > > status of the second brick. If only the second brick is up, the
> > > subvolume becomes read-only."
> > >
> > By default client-quorum is "none" in replic...
2005 Apr 17
2
Quorum error
Had a problem starting Oracle after expanding an EMC Metalun. We get the
following errors:
>WARNING: OemInit2: Opened file(/oradata/dbf/quorum.dbf 8), tid =
main:1024 file = oem.c, line = 491 {Sun Apr 17 10:33:41 2005 }
>ERROR: ReadOthersDskInfo(): ReadFile(/oradata/dbf/quorum.dbf)
failed(5) - (0) bytes read, tid = main:1024 file = oem.c, line = 1396
{Sun Apr 17 10:33:41 2005 }
>ERROR: UpdateDskNodeInfo(): WriteFile(/oradata/...
2009 May 04
2
FW: Oracle 9204 installation on linux x86-64 on ocfs
Hello All,
I have installed Oracle Cluster Manager on linux x86-64 nit. I am using ocfs file system for quorum file. But I am getting following error. Please see ocfs configureation below. I would appreciate, if someone could help me to understand if I am doing something wrong. Thanks in advance.
--------------------------------------------------cm.log file ----------------------------
oracm, version[ AMD64...
2006 Jan 09
0
[PATCH 01/11] ocfs2: event-driven quorum
This patch separates o2net and o2quo from knowing about one another as much
as possible. This is the first in a series of patches that will allow
userspace cluster interaction. Quorum is separated out first, and will
ultimately only be associated with the disk heartbeat as a separate module.
To do so, this patch performs the following changes:
* o2hb_notify() is added to handle injection of events in a synchronous
manner. All locking is preserved as expected.
* disk hear...
2017 Jun 10
4
How to remove dead peer, osrry urgent again :(
Since my node died on friday I have a dead peer (vna) that needs to be
removed.
I had major issues this morning that I haven't resolve yet with all VM's
going offline when I rebooted a node which I *hope * was due to quorum
issues as I now have four peers in the cluster, one dead, three live.
Confidence level is not high.
--
Lindsay Mathieson
2017 Jul 04
2
I need a sanity check.
2006 Oct 13
1
Cluster Quorum Question/Problem
...etings all,
I am in need of professional insight. I have a 2 node cluster running
CentOS, mysql, apache, etc. I have on each system a fiber HBA connected to
a fiber SAN. Each system shows the devices sdb and sdc for each of the
connections on the HBA. I have sdc1 mounted on both machines as /quorum.
When I right to the /quorum from one of the nodes, the file doesn't show up
on the other. However, if I unmount and remount the device on the system
that can't see the new file, the file shows up, but is broken/red and cannot
be accessed by the node that couldn't see it in the fi...
2017 Sep 22
2
AFR: Fail lookups when quorum not met
...The proposed
fix [2] was to fail look ups when we definitely know the lookup can't be
trusted (by virtue of AFR xattrs indicating the replies we got from the
up bricks are indeed bad).
Note that this fix is *only* for replica 3 or arbiter volumes (not
replica 2, where there is no notion of quorum). But we want to 'harden'
the fix by? not allowing any look ups at all if quorum is not met (or)
it is met but there are no good copies.
Some implications of this:
-If a file ends up in data/meta data split-brain in replica 3/arbiter
(rare occurrence), we won't be able to delete it...
2018 Feb 27
2
Quorum in distributed-replicate volume
On Tue, Feb 27, 2018 at 12:00:29PM +0530, Karthik Subrahmanya wrote:
> I will try to explain how you can end up in split-brain even with cluster
> wide quorum:
Yep, the explanation made sense. I hadn't considered the possibility of
alternating outages. Thanks!
> > > It would be great if you can consider configuring an arbiter or
> > > replica 3 volume.
> >
> > I can. My bricks are 2x850G and 4x11T, so I can repurpos...
2007 Aug 09
0
About quorum and fencing
Hi,
In the ocfs2 FAQ, it is written:
"A node has quorum when:
* it sees an odd number of heartbeating nodes and has network
connectivity to more than half of them.
OR,
* it sees an even number of heartbeating nodes and has network
connectivity to at least half of them *and* has connectivity to
the heartbeating...
2017 Oct 09
0
[Gluster-devel] AFR: Fail lookups when quorum not met
...s when we definitely know the lookup can't be trusted (by
>> virtue of AFR xattrs indicating the replies we got from the up bricks are
>> indeed bad).
>>
>> Note that this fix is *only* for replica 3 or arbiter volumes (not replica
>> 2, where there is no notion of quorum). But we want to 'harden' the fix by
>> not allowing any look ups at all if quorum is not met (or) it is met but
>> there are no good copies.
>>
>> Some implications of this:
>>
>> -If a file ends up in data/meta data split-brain in replica 3/arbiter (rar...
2017 Dec 21
0
Gluster replicate 3 arbiter 1 in split brain. gluster cli seems unaware
...;path to brick>/pat/to/file
Compare the hashes of the file on both bricks to see if the data actually differs. If the data is the same it makes choosing the proper replica easier.
Any idea how you got in this situation? Did you have a loss of NW connectivity? I see you are using server side quorum, maybe check the logs for any loss of quorum? I wonder if there was a loos of quorum and there was some sort of race condition hit:
http://docs.gluster.org/en/latest/Administrator%20Guide/arbiter-volumes-and-quorum/#server-quorum-and-some-pitfalls
"Unlike in client-quorum where the volume b...