similar to: Possible to preload data on a georeplication target? First sync taking forever...

Displaying 20 results from an estimated 100 matches similar to: "Possible to preload data on a georeplication target? First sync taking forever..."

2017 Oct 24
2
active-active georeplication?
hi everybody, Have glusterfs released a feature named active-active georeplication? If yes, in which version it is released? If no, is it planned to have this feature? -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20171024/0656b41f/attachment.html>
2017 Sep 17
2
georeplication sync deamon
hi all, I want to know some more detail about glusterfs georeplication, more about syncdeamon, if 'file A' was mirorred in slave volume , a change happen to 'file A', then how the syncdeamon act? 1. transfer the whole 'file A' to slave 2. transfer the changes of file A to slave thx lot -------------- next part -------------- An HTML attachment was scrubbed... URL:
2017 Oct 24
0
active-active georeplication?
Hi, No, gluster doesn't support active-active geo-replication. It's not planned in near future. We will let you know when it's planned. Thanks, Kotresh HR On Tue, Oct 24, 2017 at 11:19 AM, atris adam <atris.adam at gmail.com> wrote: > hi everybody, > > Have glusterfs released a feature named active-active georeplication? If > yes, in which version it is released?
2017 Oct 24
0
active-active georeplication?
thx for reply, that was so much interesting to me. How can I get these news about glusterfs new features? On Tue, Oct 24, 2017 at 5:54 PM, Vijay Bellur <vbellur at redhat.com> wrote: > > Halo replication [1] could be of interest here. This functionality is > available since 3.11 and the current plan is to have it fully supported in > a 4.x release. > > Note that Halo
2017 Oct 24
2
active-active georeplication?
Halo replication [1] could be of interest here. This functionality is available since 3.11 and the current plan is to have it fully supported in a 4.x release. Note that Halo replication is built on existing synchronous replication in Gluster and differs from the current geo-replication implementation. Kotresh's response is spot on for the current geo-replication implementation. Regards,
2017 Dec 21
1
seeding my georeplication
Thanks for your response (6 months ago!) but I have only just got around to following up on this. Unfortunately, I had already copied and shipped the data to the second datacenter before copying the GFIDs so I already stumbled before the first hurdle! I have been using the scripts in the extras/geo-rep provided for an earlier version upgrade. With a bit of tinkering, these have given me a file
2017 Jun 23
2
seeding my georeplication
I have a ~600tb distributed gluster volume that I want to start using geo replication on. The current volume is on 6 100tb bricks on 2 servers My plan is: 1) copy each of the bricks to a new arrays on the servers locally 2) move the new arrays to the new servers 3) create the volume on the new servers using the arrays 4) fix the layout on the new volume 5) start georeplication (which should be
2018 Feb 08
0
georeplication over ssh.
Hi Alvin, Yes, geo-replication sync happens via SSH. Ther server port 24007 is of glusterd. glusterd will be listening in this port and all volume management communication happens via RPC. Thanks, Kotresh HR On Wed, Feb 7, 2018 at 8:29 PM, Alvin Starr <alvin at netvel.net> wrote: > I am running gluster 3.8.9 and trying to setup a geo-replicated volume > over ssh, > > It looks
2023 Nov 28
0
Is there a way to short circuit the georeplication process?
We have an application that is storing an insane number of small files. We have run some tests with enabling geo-replication and letting it run but on our smallest data set it takes 10 days and our largest data set will likely take over 100 days. Would there be any way to take a copy of the data brick and convert that into a replicated image and then enable replication from the time of the
2018 Feb 07
2
georeplication over ssh.
I am running gluster 3.8.9 and trying to setup a geo-replicated volume over ssh, It looks like the volume create command is trying to directly access the server over port 24007. The docs imply that all communications are over ssh. What am I missing? -- Alvin Starr || land: (905)513-7688 Netvel Inc. || Cell: (416)806-0133 alvin at netvel.net
2013 Jul 18
1
Gluster & PHP - stat problem?
Has anyone ever ran into a problem in which PHP's stat() call to a file on a Gluster-backed volume randomly fails, yet /usr/bin/stat never fails? Running strace against both reveals that the underlying system calls succeed. I realize this is probably a PHP problem since I cannot replicate with a non-PHP-based script; however, was hoping someone on this list might have seen this before. RHEL
2018 Feb 08
0
georeplication over ssh.
Ccing glusterd team for information On Thu, Feb 8, 2018 at 10:02 AM, Alvin Starr <alvin at netvel.net> wrote: > That makes for an interesting problem. > > I cannot open port 24007 to allow RPC access. > > On 02/07/2018 11:29 PM, Kotresh Hiremath Ravishankar wrote: > > Hi Alvin, > > Yes, geo-replication sync happens via SSH. Ther server port 24007 is of >
2018 Feb 08
2
georeplication over ssh.
That makes for an interesting problem. I cannot open port 24007 to allow RPC access. On 02/07/2018 11:29 PM, Kotresh Hiremath Ravishankar wrote: > Hi Alvin, > > Yes, geo-replication sync happens via SSH. Ther server port 24007 is > of glusterd. > glusterd will be listening in this port and all volume management > communication > happens via RPC. > > Thanks, >
2012 Jun 29
2
compile glusterfs for debian squeeze
Hello, I'm compiling glusterfs for a debian squeeze. When I do a make command, I see These parameter: GlusterFS configure summary =========================== FUSE client: yes Infiniband verbs: yes epoll IO multiplex: yes argp-standalone: no fusermount: no readline: no georeplication: yes I would like to create a package that can be used both as a client and a server. I'm not interested
2017 Jun 26
1
"Rotating" .glusterfs/changelogs
Hello all, I'm trying to find a way to rotate the metadata changelogs. I've so far learned (by ndevos in #gluster) that changelog is needed for certain services, among those, georeplication, but not entirely sure about the extent. Is there a way to rotate these logs so that it takes up less space? This is not an entirely critical issue, but it seems kinda silly when we have a 3 GB volume
2012 Nov 15
0
Why does geo-replication stop when a replica member goes down
Hi, We are testing glusterfs. We have a setup like this: Site A: 4 nodes, 2 bricks per node, 1 volume, distributed, replicated, replica count 2 Site B: 2 nodes, 2 bricks per node, 1 volume, distributed georeplication setup: master: site A, node 1. slave:site B, node 1, ssh replicasets on Site A: node 1, brick 1 + node 3, brick 1 node 2, brick 1 + node 4, brick 1 node 2, brick 2 + node 3, brick
2018 May 08
2
Compiling 3.13.2 under FreeBSD 11.1?
On Mon, May 7, 2018 at 9:19 PM, Kaleb S. KEITHLEY <kkeithle at redhat.com> wrote: > > See https://review.gluster.org/19974 Many thanks Kaleb. Your patch did the trick and I did manage to compile, however I get a Segmentation fault when trying to execute gluster. I'm using the following options to configure (taken from the glusterfs 3.11.1 port in the FreeBSD port repository):
2012 Dec 03
1
configure: error: python does not have ctypes support
Hi, I am trying to install glusterfs 3.3.1 from source code. At the time of configuration i am getting the following error * configure: error: python does not have ctypes support* On my system python version is: 2.4.3 Kindly advice on fixing the error. Thanks n Regards Neetu Sharma Bangalore -------------- next part -------------- An HTML attachment was scrubbed... URL:
2018 May 07
0
Compiling 3.13.2 under FreeBSD 11.1?
On 05/07/2018 04:29 AM, Roman Serbski wrote: > Hello, > > Has anyone managed to successfully compile the latest 3.13.2 under > FreeBSD 11.1? ./autogen.sh and ./configure seem to work but make > fails: See https://review.gluster.org/19974 3.13 reached EOL with 4.0. There will be a fix posted for 4.0 soon. In the mean time I believe your specific problem with 3.13.2 should be
2010 Feb 12
1
[PATCH] Converter: Explicitly preload sym53c8xx when running mkinitrd
This is belt and braces, because we've already added it to scsi_hostadapter. --- lib/Sys/VirtV2V/Converter/Linux.pm | 2 +- 1 files changed, 1 insertions(+), 1 deletions(-) diff --git a/lib/Sys/VirtV2V/Converter/Linux.pm b/lib/Sys/VirtV2V/Converter/Linux.pm index 375da48..d5a93a7 100644 --- a/lib/Sys/VirtV2V/Converter/Linux.pm +++ b/lib/Sys/VirtV2V/Converter/Linux.pm @@ -341,7 +341,7 @@