Well...
I do the follow:
gluster vol create VMS replica 3 arbiter 1 pve01:/DATA/brick1
pve02:/DATA/brick1.5 pve01:/DATA/arbiter1.5 pve02:/DATA/brick2 pv
e01:/DATA/brick2.5 pve02:/DATA/arbiter2.5 force
And now I have:
gluster vol info
Volume Name: VMS
Type: Distributed-Replicate
Volume ID: 1bd712f5-ccb9-4322-8275-abe363d1ffdd
Status: Started
Snapshot Count: 0
Number of Bricks: 2 x (2 + 1) = 6
Transport-type: tcp
Bricks:
Brick1: pve01:/DATA/brick1
Brick2: pve02:/DATA/brick1.5
Brick3: pve01:/DATA/arbiter1.5 (arbiter)
Brick4: pve02:/DATA/brick2
Brick5: pve01:/DATA/brick2.5
Brick6: pve02:/DATA/arbiter2.5 (arbiter)
Options Reconfigured:
cluster.quorum-count: 1
cluster.quorum-reads: false
cluster.self-heal-daemon: enable
cluster.heal-timeout: 10
storage.fips-mode-rchecksum: on
transport.address-family: inet
nfs.disable: on
performance.client-io-threads: off
This values I have put it myself, in order to see if could improve the time
to make the volume available, when pve01 goes down with ifupdown
cluster.quorum-count: 1
cluster.quorum-reads: false
cluster.self-heal-daemon: enable
cluster.heal-timeout: 10
Nevertheless, it took more than 1 minutes to the volume VMS available in
the other host (pve02).
Is there any trick to reduce this time ?
Thanks
---
Gilberto Nunes Ferreira
Em qua., 5 de ago. de 2020 ?s 08:57, Gilberto Nunes <
gilberto.nunes32 at gmail.com> escreveu:
> hum I see... like this:
> [image: image.png]
> ---
> Gilberto Nunes Ferreira
>
> (47) 3025-5907
> (47) 99676-7530 - Whatsapp / Telegram
>
> Skype: gilberto.nunes36
>
>
>
>
>
> Em qua., 5 de ago. de 2020 ?s 02:14, Computerisms Corporation <
> bob at computerisms.ca> escreveu:
>
>> check the example of the chained configuration on this page:
>>
>>
>>
https://access.redhat.com/documentation/en-us/red_hat_gluster_storage/3.3/html/administration_guide/creating_arbitrated_replicated_volumes
>>
>> and apply it to two servers...
>>
>> On 2020-08-04 8:25 p.m., Gilberto Nunes wrote:
>> > Hi Bob!
>> >
>> > Could you, please, send me more detail about this configuration?
>> > I will appreciate that!
>> >
>> > Thank you
>> > ---
>> > Gilberto Nunes Ferreira
>> >
>> > (47) 3025-5907
>> > **
>> > (47) 99676-7530 - Whatsapp / Telegram
>> >
>> > Skype: gilberto.nunes36
>> >
>> >
>> >
>> >
>> >
>> > Em ter., 4 de ago. de 2020 ?s 23:47, Computerisms Corporation
>> > <bob at computerisms.ca <mailto:bob at
computerisms.ca>> escreveu:
>> >
>> > Hi Gilberto,
>> >
>> > My understanding is there can only be one arbiter per
replicated
>> > set. I
>> > don't have a lot of practice with gluster, so this could
be bad
>> advice,
>> > but the way I dealt with it on my two servers was to use 6
bricks as
>> > distributed-replicated (this is also relatively easy to
migrate to 3
>> > servers if that happens for you in the future):
>> >
>> > Server1 Server2
>> > brick1 brick1.5
>> > arbiter1.5 brick2
>> > brick2.5 arbiter2.5
>> >
>> > On 2020-08-04 7:00 p.m., Gilberto Nunes wrote:
>> > > Hi there.
>> > > I have two physical servers deployed as replica 2 and,
obviously,
>> > I got
>> > > a split-brain.
>> > > So I am thinking in use two virtual machines,each one in
physical
>> > > servers....
>> > > Then this two VMS act as a artiber of gluster set....
>> > >
>> > > Is this doable?
>> > >
>> > > Thanks
>> > >
>> > > ________
>> > >
>> > >
>> > >
>> > > Community Meeting Calendar:
>> > >
>> > > Schedule -
>> > > Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC
>> > > Bridge: https://bluejeans.com/441850968
>> > >
>> > > Gluster-users mailing list
>> > > Gluster-users at gluster.org <mailto:Gluster-users at
gluster.org>
>> > > https://lists.gluster.org/mailman/listinfo/gluster-users
>> > >
>> > ________
>> >
>> >
>> >
>> > Community Meeting Calendar:
>> >
>> > Schedule -
>> > Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC
>> > Bridge: https://bluejeans.com/441850968
>> >
>> > Gluster-users mailing list
>> > Gluster-users at gluster.org <mailto:Gluster-users at
gluster.org>
>> > https://lists.gluster.org/mailman/listinfo/gluster-users
>> >
>>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL:
<http://lists.gluster.org/pipermail/gluster-users/attachments/20200805/edf300fe/attachment.html>
-------------- next part --------------
A non-text attachment was scrubbed...
Name: image.png
Type: image/png
Size: 54749 bytes
Desc: not available
URL:
<http://lists.gluster.org/pipermail/gluster-users/attachments/20200805/edf300fe/attachment.png>