Oh I see... I was confused because the terms... Now I read this and
everything becomes clear...
https://staged-gluster-docs.readthedocs.io/en/release3.7.0beta1/Features/shard/
https://access.redhat.com/documentation/en-us/red_hat_gluster_storage/3.3/html/configuring_red_hat_virtualization_with_red_hat_gluster_storage/chap-hosting_virtual_machine_images_on_red_hat_storage_volumes
Should I use cluster.granular-entrey-heal-enable too, since I am working
with big files?
Thanks
---
Gilberto Nunes Ferreira
(47) 3025-5907
(47) 99676-7530 - Whatsapp / Telegram
Skype: gilberto.nunes36
Em qui., 6 de ago. de 2020 ?s 09:32, Gilberto Nunes <
gilberto.nunes32 at gmail.com> escreveu:
> What do you mean "sharding"? Do you mean sharing folders between
two
> servers to host qcow2 or raw vm images?
> Here I am using Proxmox which uses qemu but not virsh.
>
> Thanks
> ---
> Gilberto Nunes Ferreira
>
> (47) 3025-5907
> (47) 99676-7530 - Whatsapp / Telegram
>
> Skype: gilberto.nunes36
>
>
>
>
>
> Em qui., 6 de ago. de 2020 ?s 01:09, Strahil Nikolov <
> hunter86_bg at yahoo.com> escreveu:
>
>> As you mentioned qcow2 files, check the virt group
>> (/var/lib/glusterfs/group or something like that). It has optimal
setttins
>> for VMs and is used by oVirt.
>>
>> WARNING: If you decide to enable the group, which will also enable
>> sharding, NEVER EVER DISABLE SHARDING -> ONCE ENABLED STAYS ENABLED
!!!
>> Sharding helps reduce loocking during replica heals.
>>
>> WARNING2: As virt group uses sharding (fixes the size of file into
shard
>> size), you should consider cluster.favorite-child-policy with value
>> ctime/mtime.
>>
>> Best Regards,
>> Strahil Nikolov
>>
>> ?? 6 ?????? 2020 ?. 1:56:58 GMT+03:00, Gilberto Nunes <
>> gilberto.nunes32 at gmail.com> ??????:
>> >Ok...Thanks a lot Strahil
>> >
>> >This gluster volume set VMS cluster.favorite-child-policy size do
the
>> >trick
>> >to me here!
>> >
>> >Cheers
>> >---
>> >Gilberto Nunes Ferreira
>> >
>> >(47) 3025-5907
>> >(47) 99676-7530 - Whatsapp / Telegram
>> >
>> >Skype: gilberto.nunes36
>> >
>> >
>> >
>> >
>> >
>> >Em qua., 5 de ago. de 2020 ?s 18:15, Strahil Nikolov
>> ><hunter86_bg at yahoo.com>
>> >escreveu:
>> >
>> >> This could happen if you have pending heals. Did you reboot
that node
>> >> recently ?
>> >> Did you set automatic unsplit-brain ?
>> >>
>> >> Check for pending heals and files in splitbrain.
>> >>
>> >> If not, you can check
>> >>
>>
>https://docs.gluster.org/en/latest/Troubleshooting/resolving-splitbrain/
>> >> (look at point 5).
>> >>
>> >> Best Regards,
>> >> Strahil Nikolov
>> >>
>> >> ?? 5 ?????? 2020 ?. 23:41:57 GMT+03:00, Gilberto Nunes <
>> >> gilberto.nunes32 at gmail.com> ??????:
>> >> >I'm in trouble here.
>> >> >When I shutdown the pve01 server, the shared folder over
glusterfs
>> >is
>> >> >EMPTY!
>> >> >It's supposed to be a qcow2 file inside it.
>> >> >The content is show right, just after I power on pve01
backup...
>> >> >
>> >> >Some advice?
>> >> >
>> >> >
>> >> >Thanks
>> >> >
>> >> >---
>> >> >Gilberto Nunes Ferreira
>> >> >
>> >> >(47) 3025-5907
>> >> >(47) 99676-7530 - Whatsapp / Telegram
>> >> >
>> >> >Skype: gilberto.nunes36
>> >> >
>> >> >
>> >> >
>> >> >
>> >> >
>> >> >Em qua., 5 de ago. de 2020 ?s 11:07, Gilberto Nunes <
>> >> >gilberto.nunes32 at gmail.com> escreveu:
>> >> >
>> >> >> Well...
>> >> >> I do the follow:
>> >> >>
>> >> >> gluster vol create VMS replica 3 arbiter 1
pve01:/DATA/brick1
>> >> >> pve02:/DATA/brick1.5 pve01:/DATA/arbiter1.5
pve02:/DATA/brick2 pv
>> >> >> e01:/DATA/brick2.5 pve02:/DATA/arbiter2.5 force
>> >> >>
>> >> >> And now I have:
>> >> >> gluster vol info
>> >> >>
>> >> >> Volume Name: VMS
>> >> >> Type: Distributed-Replicate
>> >> >> Volume ID: 1bd712f5-ccb9-4322-8275-abe363d1ffdd
>> >> >> Status: Started
>> >> >> Snapshot Count: 0
>> >> >> Number of Bricks: 2 x (2 + 1) = 6
>> >> >> Transport-type: tcp
>> >> >> Bricks:
>> >> >> Brick1: pve01:/DATA/brick1
>> >> >> Brick2: pve02:/DATA/brick1.5
>> >> >> Brick3: pve01:/DATA/arbiter1.5 (arbiter)
>> >> >> Brick4: pve02:/DATA/brick2
>> >> >> Brick5: pve01:/DATA/brick2.5
>> >> >> Brick6: pve02:/DATA/arbiter2.5 (arbiter)
>> >> >> Options Reconfigured:
>> >> >> cluster.quorum-count: 1
>> >> >> cluster.quorum-reads: false
>> >> >> cluster.self-heal-daemon: enable
>> >> >> cluster.heal-timeout: 10
>> >> >> storage.fips-mode-rchecksum: on
>> >> >> transport.address-family: inet
>> >> >> nfs.disable: on
>> >> >> performance.client-io-threads: off
>> >> >>
>> >> >> This values I have put it myself, in order to see if
could improve
>> >> >the
>> >> >> time to make the volume available, when pve01 goes
down with
>> >ifupdown
>> >> >> cluster.quorum-count: 1
>> >> >> cluster.quorum-reads: false
>> >> >> cluster.self-heal-daemon: enable
>> >> >> cluster.heal-timeout: 10
>> >> >>
>> >> >> Nevertheless, it took more than 1 minutes to the
volume VMS
>> >available
>> >> >in
>> >> >> the other host (pve02).
>> >> >> Is there any trick to reduce this time ?
>> >> >>
>> >> >> Thanks
>> >> >>
>> >> >> ---
>> >> >> Gilberto Nunes Ferreira
>> >> >>
>> >> >>
>> >> >>
>> >> >>
>> >> >>
>> >> >>
>> >> >> Em qua., 5 de ago. de 2020 ?s 08:57, Gilberto Nunes
<
>> >> >> gilberto.nunes32 at gmail.com> escreveu:
>> >> >>
>> >> >>> hum I see... like this:
>> >> >>> [image: image.png]
>> >> >>> ---
>> >> >>> Gilberto Nunes Ferreira
>> >> >>>
>> >> >>> (47) 3025-5907
>> >> >>> (47) 99676-7530 - Whatsapp / Telegram
>> >> >>>
>> >> >>> Skype: gilberto.nunes36
>> >> >>>
>> >> >>>
>> >> >>>
>> >> >>>
>> >> >>>
>> >> >>> Em qua., 5 de ago. de 2020 ?s 02:14, Computerisms
Corporation <
>> >> >>> bob at computerisms.ca> escreveu:
>> >> >>>
>> >> >>>> check the example of the chained
configuration on this page:
>> >> >>>>
>> >> >>>>
>> >> >>>>
>> >> >
>> >>
>> >
>>
https://access.redhat.com/documentation/en-us/red_hat_gluster_storage/3.3/html/administration_guide/creating_arbitrated_replicated_volumes
>> >> >>>>
>> >> >>>> and apply it to two servers...
>> >> >>>>
>> >> >>>> On 2020-08-04 8:25 p.m., Gilberto Nunes
wrote:
>> >> >>>> > Hi Bob!
>> >> >>>> >
>> >> >>>> > Could you, please, send me more detail
about this
>> >configuration?
>> >> >>>> > I will appreciate that!
>> >> >>>> >
>> >> >>>> > Thank you
>> >> >>>> > ---
>> >> >>>> > Gilberto Nunes Ferreira
>> >> >>>> >
>> >> >>>> > (47) 3025-5907
>> >> >>>> > **
>> >> >>>> > (47) 99676-7530 - Whatsapp / Telegram
>> >> >>>> >
>> >> >>>> > Skype: gilberto.nunes36
>> >> >>>> >
>> >> >>>> >
>> >> >>>> >
>> >> >>>> >
>> >> >>>> >
>> >> >>>> > Em ter., 4 de ago. de 2020 ?s 23:47,
Computerisms Corporation
>> >> >>>> > <bob at computerisms.ca
<mailto:bob at computerisms.ca>> escreveu:
>> >> >>>> >
>> >> >>>> > Hi Gilberto,
>> >> >>>> >
>> >> >>>> > My understanding is there can only
be one arbiter per
>> >> >replicated
>> >> >>>> > set. I
>> >> >>>> > don't have a lot of practice
with gluster, so this could
>> >be
>> >> >bad
>> >> >>>> advice,
>> >> >>>> > but the way I dealt with it on my
two servers was to use 6
>> >> >bricks
>> >> >>>> as
>> >> >>>> > distributed-replicated (this is also
relatively easy to
>> >> >migrate to
>> >> >>>> 3
>> >> >>>> > servers if that happens for you in
the future):
>> >> >>>> >
>> >> >>>> > Server1 Server2
>> >> >>>> > brick1 brick1.5
>> >> >>>> > arbiter1.5 brick2
>> >> >>>> > brick2.5 arbiter2.5
>> >> >>>> >
>> >> >>>> > On 2020-08-04 7:00 p.m., Gilberto
Nunes wrote:
>> >> >>>> > > Hi there.
>> >> >>>> > > I have two physical servers
deployed as replica 2 and,
>> >> >>>> obviously,
>> >> >>>> > I got
>> >> >>>> > > a split-brain.
>> >> >>>> > > So I am thinking in use two
virtual machines,each one
>> >in
>> >> >>>> physical
>> >> >>>> > > servers....
>> >> >>>> > > Then this two VMS act as a
artiber of gluster set....
>> >> >>>> > >
>> >> >>>> > > Is this doable?
>> >> >>>> > >
>> >> >>>> > > Thanks
>> >> >>>> > >
>> >> >>>> > > ________
>> >> >>>> > >
>> >> >>>> > >
>> >> >>>> > >
>> >> >>>> > > Community Meeting Calendar:
>> >> >>>> > >
>> >> >>>> > > Schedule -
>> >> >>>> > > Every 2nd and 4th Tuesday at
14:30 IST / 09:00 UTC
>> >> >>>> > > Bridge:
https://bluejeans.com/441850968
>> >> >>>> > >
>> >> >>>> > > Gluster-users mailing list
>> >> >>>> > > Gluster-users at gluster.org
>> >> ><mailto:Gluster-users at gluster.org>
>> >> >>>> > >
>> >https://lists.gluster.org/mailman/listinfo/gluster-users
>> >> >>>> > >
>> >> >>>> > ________
>> >> >>>> >
>> >> >>>> >
>> >> >>>> >
>> >> >>>> > Community Meeting Calendar:
>> >> >>>> >
>> >> >>>> > Schedule -
>> >> >>>> > Every 2nd and 4th Tuesday at 14:30
IST / 09:00 UTC
>> >> >>>> > Bridge:
https://bluejeans.com/441850968
>> >> >>>> >
>> >> >>>> > Gluster-users mailing list
>> >> >>>> > Gluster-users at gluster.org
>> ><mailto:Gluster-users at gluster.org>
>> >> >>>> >
https://lists.gluster.org/mailman/listinfo/gluster-users
>> >> >>>> >
>> >> >>>>
>> >> >>>
>> >>
>>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL:
<http://lists.gluster.org/pipermail/gluster-users/attachments/20200806/8ae6bbe4/attachment.html>