Strahil Nikolov
2024-Oct-20 20:34 UTC
[Gluster-users] How much disk can fail after a catastrophic failure occur?
If it's replica 2, you can loose up to 1 replica per distribution group.For example, if you have a volume TEST with such setup: server1:/brick1 server2:/brick1 server1:/brick2 server2:/brick2 You can loose any brick of the replica "/brick1" and any brick in the replica "/brick2". So if you loose server1:/brick1 and server2:/brick2 -> no data loss will be experienced. As usual, consider if you can add an arbiter for your volumes. Best Regards, Strahil Nikolov ? ??????, 19 ???????? 2024 ?. ? 18:32:40 ?. ???????+3, Gilberto Ferreira <gilberto.nunes32 at gmail.com> ??????: Hi there.I have 2 servers with this number of disks in each side: pve01:~# df | grep disco /dev/sdd ? ? ? ? ?1.0T ?9.4G 1015G ? 1% /disco1TB-0 /dev/sdh ? ? ? ? ?1.0T ?9.3G 1015G ? 1% /disco1TB-3 /dev/sde ? ? ? ? ?1.0T ?9.5G 1015G ? 1% /disco1TB-1 /dev/sdf ? ? ? ? ?1.0T ?9.4G 1015G ? 1% /disco1TB-2 /dev/sdg ? ? ? ? ?2.0T ? 19G ?2.0T ? 1% /disco2TB-1 /dev/sdc ? ? ? ? ?2.0T ? 19G ?2.0T ? 1% /disco2TB-0 /dev/sdj ? ? ? ? ?1.0T ?9.2G 1015G ? 1% /disco1TB-4 I have a?Type: Distributed-Replicate glusterSo my question is: how much disk can be in fail state after losing data or something? Thanks in advance --- Gilberto Nunes Ferreira ? ________ Community Meeting Calendar: Schedule - Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC Bridge: https://meet.google.com/cpu-eiue-hvk Gluster-users mailing list Gluster-users at gluster.org https://lists.gluster.org/mailman/listinfo/gluster-users -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20241020/4a424036/attachment.html>
Gilberto Ferreira
2024-Oct-21 11:39 UTC
[Gluster-users] How much disk can fail after a catastrophic failure occur?
Ok! I got it about how many disks I can lose and so on. But regard the arbiter isse, I always set this parameters in the gluster volume, in order to avoid split-brain and I might add that work pretty well to me. I already have a Proxmox VE cluster with 2 nodes and about 50 vms, running different Linux distro - and Windows as well - with Cpanel and other stuff, in production. Anyway here the parameters I had have used: gluster vol set VMS cluster.heal-timeout 5 gluster vol heal VMS enable gluster vol set VMS cluster.quorum-reads false gluster vol set VMS cluster.quorum-count 1 gluster vol set VMS network.ping-timeout 2 gluster vol set VMS cluster.favorite-child-policy mtime gluster vol heal VMS granular-entry-heal enable gluster vol set VMS cluster.data-self-heal-algorithm full gluster vol set VMS features.shard on gluster vol set VMS performance.write-behind off gluster vol set VMS performance.flush-behind off --- Gilberto Nunes Ferreira Em dom., 20 de out. de 2024 ?s 17:34, Strahil Nikolov <hunter86_bg at yahoo.com> escreveu:> If it's replica 2, you can loose up to 1 replica per distribution group. > For example, if you have a volume TEST with such setup: > > server1:/brick1 > server2:/brick1 > server1:/brick2 > server2:/brick2 > > You can loose any brick of the replica "/brick1" and any brick in the > replica "/brick2". So if you loose server1:/brick1 and server2:/brick2 -> > no data loss will be experienced. > As usual, consider if you can add an arbiter for your volumes. > > Best Regards, > Strahil Nikolov > > ? ??????, 19 ???????? 2024 ?. ? 18:32:40 ?. ???????+3, Gilberto Ferreira < > gilberto.nunes32 at gmail.com> ??????: > > > Hi there. > I have 2 servers with this number of disks in each side: > > pve01:~# df | grep disco > /dev/sdd 1.0T 9.4G 1015G 1% /disco1TB-0 > /dev/sdh 1.0T 9.3G 1015G 1% /disco1TB-3 > /dev/sde 1.0T 9.5G 1015G 1% /disco1TB-1 > /dev/sdf 1.0T 9.4G 1015G 1% /disco1TB-2 > /dev/sdg 2.0T 19G 2.0T 1% /disco2TB-1 > /dev/sdc 2.0T 19G 2.0T 1% /disco2TB-0 > /dev/sdj 1.0T 9.2G 1015G 1% /disco1TB-4 > > I have a Type: Distributed-Replicate gluster > So my question is: how much disk can be in fail state after losing data or > something? > > Thanks in advance > > --- > > > Gilberto Nunes Ferreira > > > > > > > ________ > > > > Community Meeting Calendar: > > Schedule - > Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC > Bridge: https://meet.google.com/cpu-eiue-hvk > Gluster-users mailing list > Gluster-users at gluster.org > https://lists.gluster.org/mailman/listinfo/gluster-users >-------------- next part -------------- An HTML attachment was scrubbed... URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20241021/863e5f46/attachment.html>
Possibly Parallel Threads
- How much disk can fail after a catastrophic failure occur?
- How much disk can fail after a catastrophic failure occur?
- Add an arbiter when have multiple bricks at same server.
- Add an arbiter when have multiple bricks at same server.
- Add an arbiter when have multiple bricks at same server.