Well... I do the follow: gluster vol create VMS replica 3 arbiter 1 pve01:/DATA/brick1 pve02:/DATA/brick1.5 pve01:/DATA/arbiter1.5 pve02:/DATA/brick2 pv e01:/DATA/brick2.5 pve02:/DATA/arbiter2.5 force And now I have: gluster vol info Volume Name: VMS Type: Distributed-Replicate Volume ID: 1bd712f5-ccb9-4322-8275-abe363d1ffdd Status: Started Snapshot Count: 0 Number of Bricks: 2 x (2 + 1) = 6 Transport-type: tcp Bricks: Brick1: pve01:/DATA/brick1 Brick2: pve02:/DATA/brick1.5 Brick3: pve01:/DATA/arbiter1.5 (arbiter) Brick4: pve02:/DATA/brick2 Brick5: pve01:/DATA/brick2.5 Brick6: pve02:/DATA/arbiter2.5 (arbiter) Options Reconfigured: cluster.quorum-count: 1 cluster.quorum-reads: false cluster.self-heal-daemon: enable cluster.heal-timeout: 10 storage.fips-mode-rchecksum: on transport.address-family: inet nfs.disable: on performance.client-io-threads: off This values I have put it myself, in order to see if could improve the time to make the volume available, when pve01 goes down with ifupdown cluster.quorum-count: 1 cluster.quorum-reads: false cluster.self-heal-daemon: enable cluster.heal-timeout: 10 Nevertheless, it took more than 1 minutes to the volume VMS available in the other host (pve02). Is there any trick to reduce this time ? Thanks --- Gilberto Nunes Ferreira Em qua., 5 de ago. de 2020 ?s 08:57, Gilberto Nunes < gilberto.nunes32 at gmail.com> escreveu:> hum I see... like this: > [image: image.png] > --- > Gilberto Nunes Ferreira > > (47) 3025-5907 > (47) 99676-7530 - Whatsapp / Telegram > > Skype: gilberto.nunes36 > > > > > > Em qua., 5 de ago. de 2020 ?s 02:14, Computerisms Corporation < > bob at computerisms.ca> escreveu: > >> check the example of the chained configuration on this page: >> >> >> https://access.redhat.com/documentation/en-us/red_hat_gluster_storage/3.3/html/administration_guide/creating_arbitrated_replicated_volumes >> >> and apply it to two servers... >> >> On 2020-08-04 8:25 p.m., Gilberto Nunes wrote: >> > Hi Bob! >> > >> > Could you, please, send me more detail about this configuration? >> > I will appreciate that! >> > >> > Thank you >> > --- >> > Gilberto Nunes Ferreira >> > >> > (47) 3025-5907 >> > ** >> > (47) 99676-7530 - Whatsapp / Telegram >> > >> > Skype: gilberto.nunes36 >> > >> > >> > >> > >> > >> > Em ter., 4 de ago. de 2020 ?s 23:47, Computerisms Corporation >> > <bob at computerisms.ca <mailto:bob at computerisms.ca>> escreveu: >> > >> > Hi Gilberto, >> > >> > My understanding is there can only be one arbiter per replicated >> > set. I >> > don't have a lot of practice with gluster, so this could be bad >> advice, >> > but the way I dealt with it on my two servers was to use 6 bricks as >> > distributed-replicated (this is also relatively easy to migrate to 3 >> > servers if that happens for you in the future): >> > >> > Server1 Server2 >> > brick1 brick1.5 >> > arbiter1.5 brick2 >> > brick2.5 arbiter2.5 >> > >> > On 2020-08-04 7:00 p.m., Gilberto Nunes wrote: >> > > Hi there. >> > > I have two physical servers deployed as replica 2 and, obviously, >> > I got >> > > a split-brain. >> > > So I am thinking in use two virtual machines,each one in physical >> > > servers.... >> > > Then this two VMS act as a artiber of gluster set.... >> > > >> > > Is this doable? >> > > >> > > Thanks >> > > >> > > ________ >> > > >> > > >> > > >> > > Community Meeting Calendar: >> > > >> > > Schedule - >> > > Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC >> > > Bridge: https://bluejeans.com/441850968 >> > > >> > > Gluster-users mailing list >> > > Gluster-users at gluster.org <mailto:Gluster-users at gluster.org> >> > > https://lists.gluster.org/mailman/listinfo/gluster-users >> > > >> > ________ >> > >> > >> > >> > Community Meeting Calendar: >> > >> > Schedule - >> > Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC >> > Bridge: https://bluejeans.com/441850968 >> > >> > Gluster-users mailing list >> > Gluster-users at gluster.org <mailto:Gluster-users at gluster.org> >> > https://lists.gluster.org/mailman/listinfo/gluster-users >> > >> >-------------- next part -------------- An HTML attachment was scrubbed... URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20200805/edf300fe/attachment.html> -------------- next part -------------- A non-text attachment was scrubbed... Name: image.png Type: image/png Size: 54749 bytes Desc: not available URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20200805/edf300fe/attachment.png>
I'm in trouble here. When I shutdown the pve01 server, the shared folder over glusterfs is EMPTY! It's supposed to be a qcow2 file inside it. The content is show right, just after I power on pve01 backup... Some advice? Thanks --- Gilberto Nunes Ferreira (47) 3025-5907 (47) 99676-7530 - Whatsapp / Telegram Skype: gilberto.nunes36 Em qua., 5 de ago. de 2020 ?s 11:07, Gilberto Nunes < gilberto.nunes32 at gmail.com> escreveu:> Well... > I do the follow: > > gluster vol create VMS replica 3 arbiter 1 pve01:/DATA/brick1 > pve02:/DATA/brick1.5 pve01:/DATA/arbiter1.5 pve02:/DATA/brick2 pv > e01:/DATA/brick2.5 pve02:/DATA/arbiter2.5 force > > And now I have: > gluster vol info > > Volume Name: VMS > Type: Distributed-Replicate > Volume ID: 1bd712f5-ccb9-4322-8275-abe363d1ffdd > Status: Started > Snapshot Count: 0 > Number of Bricks: 2 x (2 + 1) = 6 > Transport-type: tcp > Bricks: > Brick1: pve01:/DATA/brick1 > Brick2: pve02:/DATA/brick1.5 > Brick3: pve01:/DATA/arbiter1.5 (arbiter) > Brick4: pve02:/DATA/brick2 > Brick5: pve01:/DATA/brick2.5 > Brick6: pve02:/DATA/arbiter2.5 (arbiter) > Options Reconfigured: > cluster.quorum-count: 1 > cluster.quorum-reads: false > cluster.self-heal-daemon: enable > cluster.heal-timeout: 10 > storage.fips-mode-rchecksum: on > transport.address-family: inet > nfs.disable: on > performance.client-io-threads: off > > This values I have put it myself, in order to see if could improve the > time to make the volume available, when pve01 goes down with ifupdown > cluster.quorum-count: 1 > cluster.quorum-reads: false > cluster.self-heal-daemon: enable > cluster.heal-timeout: 10 > > Nevertheless, it took more than 1 minutes to the volume VMS available in > the other host (pve02). > Is there any trick to reduce this time ? > > Thanks > > --- > Gilberto Nunes Ferreira > > > > > > > Em qua., 5 de ago. de 2020 ?s 08:57, Gilberto Nunes < > gilberto.nunes32 at gmail.com> escreveu: > >> hum I see... like this: >> [image: image.png] >> --- >> Gilberto Nunes Ferreira >> >> (47) 3025-5907 >> (47) 99676-7530 - Whatsapp / Telegram >> >> Skype: gilberto.nunes36 >> >> >> >> >> >> Em qua., 5 de ago. de 2020 ?s 02:14, Computerisms Corporation < >> bob at computerisms.ca> escreveu: >> >>> check the example of the chained configuration on this page: >>> >>> >>> https://access.redhat.com/documentation/en-us/red_hat_gluster_storage/3.3/html/administration_guide/creating_arbitrated_replicated_volumes >>> >>> and apply it to two servers... >>> >>> On 2020-08-04 8:25 p.m., Gilberto Nunes wrote: >>> > Hi Bob! >>> > >>> > Could you, please, send me more detail about this configuration? >>> > I will appreciate that! >>> > >>> > Thank you >>> > --- >>> > Gilberto Nunes Ferreira >>> > >>> > (47) 3025-5907 >>> > ** >>> > (47) 99676-7530 - Whatsapp / Telegram >>> > >>> > Skype: gilberto.nunes36 >>> > >>> > >>> > >>> > >>> > >>> > Em ter., 4 de ago. de 2020 ?s 23:47, Computerisms Corporation >>> > <bob at computerisms.ca <mailto:bob at computerisms.ca>> escreveu: >>> > >>> > Hi Gilberto, >>> > >>> > My understanding is there can only be one arbiter per replicated >>> > set. I >>> > don't have a lot of practice with gluster, so this could be bad >>> advice, >>> > but the way I dealt with it on my two servers was to use 6 bricks >>> as >>> > distributed-replicated (this is also relatively easy to migrate to >>> 3 >>> > servers if that happens for you in the future): >>> > >>> > Server1 Server2 >>> > brick1 brick1.5 >>> > arbiter1.5 brick2 >>> > brick2.5 arbiter2.5 >>> > >>> > On 2020-08-04 7:00 p.m., Gilberto Nunes wrote: >>> > > Hi there. >>> > > I have two physical servers deployed as replica 2 and, >>> obviously, >>> > I got >>> > > a split-brain. >>> > > So I am thinking in use two virtual machines,each one in >>> physical >>> > > servers.... >>> > > Then this two VMS act as a artiber of gluster set.... >>> > > >>> > > Is this doable? >>> > > >>> > > Thanks >>> > > >>> > > ________ >>> > > >>> > > >>> > > >>> > > Community Meeting Calendar: >>> > > >>> > > Schedule - >>> > > Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC >>> > > Bridge: https://bluejeans.com/441850968 >>> > > >>> > > Gluster-users mailing list >>> > > Gluster-users at gluster.org <mailto:Gluster-users at gluster.org> >>> > > https://lists.gluster.org/mailman/listinfo/gluster-users >>> > > >>> > ________ >>> > >>> > >>> > >>> > Community Meeting Calendar: >>> > >>> > Schedule - >>> > Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC >>> > Bridge: https://bluejeans.com/441850968 >>> > >>> > Gluster-users mailing list >>> > Gluster-users at gluster.org <mailto:Gluster-users at gluster.org> >>> > https://lists.gluster.org/mailman/listinfo/gluster-users >>> > >>> >>-------------- next part -------------- An HTML attachment was scrubbed... URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20200805/746ac4bd/attachment.html> -------------- next part -------------- A non-text attachment was scrubbed... Name: image.png Type: image/png Size: 54749 bytes Desc: not available URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20200805/746ac4bd/attachment.png>
If I understood you correctly, you are looking for this: Option: network.ping-timeout Default Value: 42 Description: Time duration for which the client waits to check if the server is responsive. Best Regards, Strahil Nikolov ?? 5 ?????? 2020 ?. 17:07:10 GMT+03:00, Gilberto Nunes <gilberto.nunes32 at gmail.com> ??????:>Well... >I do the follow: > >gluster vol create VMS replica 3 arbiter 1 pve01:/DATA/brick1 >pve02:/DATA/brick1.5 pve01:/DATA/arbiter1.5 pve02:/DATA/brick2 pv >e01:/DATA/brick2.5 pve02:/DATA/arbiter2.5 force > >And now I have: >gluster vol info > >Volume Name: VMS >Type: Distributed-Replicate >Volume ID: 1bd712f5-ccb9-4322-8275-abe363d1ffdd >Status: Started >Snapshot Count: 0 >Number of Bricks: 2 x (2 + 1) = 6 >Transport-type: tcp >Bricks: >Brick1: pve01:/DATA/brick1 >Brick2: pve02:/DATA/brick1.5 >Brick3: pve01:/DATA/arbiter1.5 (arbiter) >Brick4: pve02:/DATA/brick2 >Brick5: pve01:/DATA/brick2.5 >Brick6: pve02:/DATA/arbiter2.5 (arbiter) >Options Reconfigured: >cluster.quorum-count: 1 >cluster.quorum-reads: false >cluster.self-heal-daemon: enable >cluster.heal-timeout: 10 >storage.fips-mode-rchecksum: on >transport.address-family: inet >nfs.disable: on >performance.client-io-threads: off > >This values I have put it myself, in order to see if could improve the >time >to make the volume available, when pve01 goes down with ifupdown >cluster.quorum-count: 1 >cluster.quorum-reads: false >cluster.self-heal-daemon: enable >cluster.heal-timeout: 10 > >Nevertheless, it took more than 1 minutes to the volume VMS available >in >the other host (pve02). >Is there any trick to reduce this time ? > >Thanks > >--- >Gilberto Nunes Ferreira > > > > > > >Em qua., 5 de ago. de 2020 ?s 08:57, Gilberto Nunes < >gilberto.nunes32 at gmail.com> escreveu: > >> hum I see... like this: >> [image: image.png] >> --- >> Gilberto Nunes Ferreira >> >> (47) 3025-5907 >> (47) 99676-7530 - Whatsapp / Telegram >> >> Skype: gilberto.nunes36 >> >> >> >> >> >> Em qua., 5 de ago. de 2020 ?s 02:14, Computerisms Corporation < >> bob at computerisms.ca> escreveu: >> >>> check the example of the chained configuration on this page: >>> >>> >>> >https://access.redhat.com/documentation/en-us/red_hat_gluster_storage/3.3/html/administration_guide/creating_arbitrated_replicated_volumes >>> >>> and apply it to two servers... >>> >>> On 2020-08-04 8:25 p.m., Gilberto Nunes wrote: >>> > Hi Bob! >>> > >>> > Could you, please, send me more detail about this configuration? >>> > I will appreciate that! >>> > >>> > Thank you >>> > --- >>> > Gilberto Nunes Ferreira >>> > >>> > (47) 3025-5907 >>> > ** >>> > (47) 99676-7530 - Whatsapp / Telegram >>> > >>> > Skype: gilberto.nunes36 >>> > >>> > >>> > >>> > >>> > >>> > Em ter., 4 de ago. de 2020 ?s 23:47, Computerisms Corporation >>> > <bob at computerisms.ca <mailto:bob at computerisms.ca>> escreveu: >>> > >>> > Hi Gilberto, >>> > >>> > My understanding is there can only be one arbiter per >replicated >>> > set. I >>> > don't have a lot of practice with gluster, so this could be >bad >>> advice, >>> > but the way I dealt with it on my two servers was to use 6 >bricks as >>> > distributed-replicated (this is also relatively easy to >migrate to 3 >>> > servers if that happens for you in the future): >>> > >>> > Server1 Server2 >>> > brick1 brick1.5 >>> > arbiter1.5 brick2 >>> > brick2.5 arbiter2.5 >>> > >>> > On 2020-08-04 7:00 p.m., Gilberto Nunes wrote: >>> > > Hi there. >>> > > I have two physical servers deployed as replica 2 and, >obviously, >>> > I got >>> > > a split-brain. >>> > > So I am thinking in use two virtual machines,each one in >physical >>> > > servers.... >>> > > Then this two VMS act as a artiber of gluster set.... >>> > > >>> > > Is this doable? >>> > > >>> > > Thanks >>> > > >>> > > ________ >>> > > >>> > > >>> > > >>> > > Community Meeting Calendar: >>> > > >>> > > Schedule - >>> > > Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC >>> > > Bridge: https://bluejeans.com/441850968 >>> > > >>> > > Gluster-users mailing list >>> > > Gluster-users at gluster.org ><mailto:Gluster-users at gluster.org> >>> > > https://lists.gluster.org/mailman/listinfo/gluster-users >>> > > >>> > ________ >>> > >>> > >>> > >>> > Community Meeting Calendar: >>> > >>> > Schedule - >>> > Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC >>> > Bridge: https://bluejeans.com/441850968 >>> > >>> > Gluster-users mailing list >>> > Gluster-users at gluster.org <mailto:Gluster-users at gluster.org> >>> > https://lists.gluster.org/mailman/listinfo/gluster-users >>> > >>> >>