As you mentioned qcow2 files, check the virt group (/var/lib/glusterfs/group
or something like that). It has optimal setttins for VMs and is used by oVirt.
WARNING: If you decide to enable the group, which will also enable sharding,
NEVER EVER DISABLE SHARDING -> ONCE ENABLED STAYS ENABLED !!!
Sharding helps reduce loocking during replica heals.
WARNING2: As virt group uses sharding (fixes the size of file into shard size),
you should consider cluster.favorite-child-policy with value ctime/mtime.
Best Regards,
Strahil Nikolov
?? 6 ?????? 2020 ?. 1:56:58 GMT+03:00, Gilberto Nunes <gilberto.nunes32 at
gmail.com> ??????:>Ok...Thanks a lot Strahil
>
>This gluster volume set VMS cluster.favorite-child-policy size do the
>trick
>to me here!
>
>Cheers
>---
>Gilberto Nunes Ferreira
>
>(47) 3025-5907
>(47) 99676-7530 - Whatsapp / Telegram
>
>Skype: gilberto.nunes36
>
>
>
>
>
>Em qua., 5 de ago. de 2020 ?s 18:15, Strahil Nikolov
><hunter86_bg at yahoo.com>
>escreveu:
>
>> This could happen if you have pending heals. Did you reboot that node
>> recently ?
>> Did you set automatic unsplit-brain ?
>>
>> Check for pending heals and files in splitbrain.
>>
>> If not, you can check
>>
>https://docs.gluster.org/en/latest/Troubleshooting/resolving-splitbrain/
>> (look at point 5).
>>
>> Best Regards,
>> Strahil Nikolov
>>
>> ?? 5 ?????? 2020 ?. 23:41:57 GMT+03:00, Gilberto Nunes <
>> gilberto.nunes32 at gmail.com> ??????:
>> >I'm in trouble here.
>> >When I shutdown the pve01 server, the shared folder over glusterfs
>is
>> >EMPTY!
>> >It's supposed to be a qcow2 file inside it.
>> >The content is show right, just after I power on pve01 backup...
>> >
>> >Some advice?
>> >
>> >
>> >Thanks
>> >
>> >---
>> >Gilberto Nunes Ferreira
>> >
>> >(47) 3025-5907
>> >(47) 99676-7530 - Whatsapp / Telegram
>> >
>> >Skype: gilberto.nunes36
>> >
>> >
>> >
>> >
>> >
>> >Em qua., 5 de ago. de 2020 ?s 11:07, Gilberto Nunes <
>> >gilberto.nunes32 at gmail.com> escreveu:
>> >
>> >> Well...
>> >> I do the follow:
>> >>
>> >> gluster vol create VMS replica 3 arbiter 1 pve01:/DATA/brick1
>> >> pve02:/DATA/brick1.5 pve01:/DATA/arbiter1.5 pve02:/DATA/brick2
pv
>> >> e01:/DATA/brick2.5 pve02:/DATA/arbiter2.5 force
>> >>
>> >> And now I have:
>> >> gluster vol info
>> >>
>> >> Volume Name: VMS
>> >> Type: Distributed-Replicate
>> >> Volume ID: 1bd712f5-ccb9-4322-8275-abe363d1ffdd
>> >> Status: Started
>> >> Snapshot Count: 0
>> >> Number of Bricks: 2 x (2 + 1) = 6
>> >> Transport-type: tcp
>> >> Bricks:
>> >> Brick1: pve01:/DATA/brick1
>> >> Brick2: pve02:/DATA/brick1.5
>> >> Brick3: pve01:/DATA/arbiter1.5 (arbiter)
>> >> Brick4: pve02:/DATA/brick2
>> >> Brick5: pve01:/DATA/brick2.5
>> >> Brick6: pve02:/DATA/arbiter2.5 (arbiter)
>> >> Options Reconfigured:
>> >> cluster.quorum-count: 1
>> >> cluster.quorum-reads: false
>> >> cluster.self-heal-daemon: enable
>> >> cluster.heal-timeout: 10
>> >> storage.fips-mode-rchecksum: on
>> >> transport.address-family: inet
>> >> nfs.disable: on
>> >> performance.client-io-threads: off
>> >>
>> >> This values I have put it myself, in order to see if could
improve
>> >the
>> >> time to make the volume available, when pve01 goes down with
>ifupdown
>> >> cluster.quorum-count: 1
>> >> cluster.quorum-reads: false
>> >> cluster.self-heal-daemon: enable
>> >> cluster.heal-timeout: 10
>> >>
>> >> Nevertheless, it took more than 1 minutes to the volume VMS
>available
>> >in
>> >> the other host (pve02).
>> >> Is there any trick to reduce this time ?
>> >>
>> >> Thanks
>> >>
>> >> ---
>> >> Gilberto Nunes Ferreira
>> >>
>> >>
>> >>
>> >>
>> >>
>> >>
>> >> Em qua., 5 de ago. de 2020 ?s 08:57, Gilberto Nunes <
>> >> gilberto.nunes32 at gmail.com> escreveu:
>> >>
>> >>> hum I see... like this:
>> >>> [image: image.png]
>> >>> ---
>> >>> Gilberto Nunes Ferreira
>> >>>
>> >>> (47) 3025-5907
>> >>> (47) 99676-7530 - Whatsapp / Telegram
>> >>>
>> >>> Skype: gilberto.nunes36
>> >>>
>> >>>
>> >>>
>> >>>
>> >>>
>> >>> Em qua., 5 de ago. de 2020 ?s 02:14, Computerisms
Corporation <
>> >>> bob at computerisms.ca> escreveu:
>> >>>
>> >>>> check the example of the chained configuration on this
page:
>> >>>>
>> >>>>
>> >>>>
>> >
>>
>https://access.redhat.com/documentation/en-us/red_hat_gluster_storage/3.3/html/administration_guide/creating_arbitrated_replicated_volumes
>> >>>>
>> >>>> and apply it to two servers...
>> >>>>
>> >>>> On 2020-08-04 8:25 p.m., Gilberto Nunes wrote:
>> >>>> > Hi Bob!
>> >>>> >
>> >>>> > Could you, please, send me more detail about this
>configuration?
>> >>>> > I will appreciate that!
>> >>>> >
>> >>>> > Thank you
>> >>>> > ---
>> >>>> > Gilberto Nunes Ferreira
>> >>>> >
>> >>>> > (47) 3025-5907
>> >>>> > **
>> >>>> > (47) 99676-7530 - Whatsapp / Telegram
>> >>>> >
>> >>>> > Skype: gilberto.nunes36
>> >>>> >
>> >>>> >
>> >>>> >
>> >>>> >
>> >>>> >
>> >>>> > Em ter., 4 de ago. de 2020 ?s 23:47, Computerisms
Corporation
>> >>>> > <bob at computerisms.ca <mailto:bob at
computerisms.ca>> escreveu:
>> >>>> >
>> >>>> > Hi Gilberto,
>> >>>> >
>> >>>> > My understanding is there can only be one
arbiter per
>> >replicated
>> >>>> > set. I
>> >>>> > don't have a lot of practice with
gluster, so this could
>be
>> >bad
>> >>>> advice,
>> >>>> > but the way I dealt with it on my two servers
was to use 6
>> >bricks
>> >>>> as
>> >>>> > distributed-replicated (this is also
relatively easy to
>> >migrate to
>> >>>> 3
>> >>>> > servers if that happens for you in the
future):
>> >>>> >
>> >>>> > Server1 Server2
>> >>>> > brick1 brick1.5
>> >>>> > arbiter1.5 brick2
>> >>>> > brick2.5 arbiter2.5
>> >>>> >
>> >>>> > On 2020-08-04 7:00 p.m., Gilberto Nunes
wrote:
>> >>>> > > Hi there.
>> >>>> > > I have two physical servers deployed as
replica 2 and,
>> >>>> obviously,
>> >>>> > I got
>> >>>> > > a split-brain.
>> >>>> > > So I am thinking in use two virtual
machines,each one
>in
>> >>>> physical
>> >>>> > > servers....
>> >>>> > > Then this two VMS act as a artiber of
gluster set....
>> >>>> > >
>> >>>> > > Is this doable?
>> >>>> > >
>> >>>> > > Thanks
>> >>>> > >
>> >>>> > > ________
>> >>>> > >
>> >>>> > >
>> >>>> > >
>> >>>> > > Community Meeting Calendar:
>> >>>> > >
>> >>>> > > Schedule -
>> >>>> > > Every 2nd and 4th Tuesday at 14:30 IST
/ 09:00 UTC
>> >>>> > > Bridge: https://bluejeans.com/441850968
>> >>>> > >
>> >>>> > > Gluster-users mailing list
>> >>>> > > Gluster-users at gluster.org
>> ><mailto:Gluster-users at gluster.org>
>> >>>> > >
>https://lists.gluster.org/mailman/listinfo/gluster-users
>> >>>> > >
>> >>>> > ________
>> >>>> >
>> >>>> >
>> >>>> >
>> >>>> > Community Meeting Calendar:
>> >>>> >
>> >>>> > Schedule -
>> >>>> > Every 2nd and 4th Tuesday at 14:30 IST /
09:00 UTC
>> >>>> > Bridge: https://bluejeans.com/441850968
>> >>>> >
>> >>>> > Gluster-users mailing list
>> >>>> > Gluster-users at gluster.org
><mailto:Gluster-users at gluster.org>
>> >>>> >
https://lists.gluster.org/mailman/listinfo/gluster-users
>> >>>> >
>> >>>>
>> >>>
>>