search for: replicator

Displaying 20 results from an estimated 12085 matches for "replicator".

2011 Feb 19
2
reading simulations
Hi to all the people (again), I'm doing some simulations with the memisc package of an own function, but I've the problem that I didn't know how to read the result of such simulations. My function is: > Torre<-function(a1,N1,a2,N2) + {Etorre<-(a1*N1)/(1+a1*N1) + Efuera<-(a2*N2)/(1+a2*N2) + if(Etorre>Efuera)Subir=TRUE + if (Etorre<Efuera)Subir=FALSE +
2012 Feb 28
2
Dovecot clustering with dsync-based replication
...or not. Aggregator ---------- When running in Dovecot director-based setup, all of the Dovecot backends (where replication plugin runs) also run "aggregator" process. Its job is very simple: It proxies the notifications from mail plugin and sends them via a single TCP connection to the replicator process running in Dovecot proxies. This is simply an optimization to avoid tons of short lived TCP connections directly from replication plugin to director server. When not running in Dovecot director setup (i.e. there is only a single Dovecot instance that handles all of the users), there is no...
2017 Oct 26
0
not healing one file
Hey Richard, Could you share the following informations please? 1. gluster volume info <volname> 2. getfattr output of that file from all the bricks getfattr -d -e hex -m . <brickpath/filepath> 3. glustershd & glfsheal logs Regards, Karthik On Thu, Oct 26, 2017 at 10:21 AM, Amar Tumballi <atumball at redhat.com> wrote: > On a side note, try recently released health
2017 Mar 21
3
replicator crashing - oom
I have the following in my log: Mar 21 14:46:59 bubba dovecot: replicator: Panic: data stack: Out of memory when allocating 1073741864 bytes Mar 21 14:46:59 bubba dovecot: replicator: Error: Raw backtrace: /usr/local/lib/dovecot/libdovecot.so.0(+0x97c90) [0x7f4638a7cc90] -> /usr/local/lib/dovecot/libdovecot.so.0(+0x97d6e) [0x7f4638a7cd6e] -> /usr/local/lib/dove...
2017 Oct 26
3
not healing one file
On a side note, try recently released health report tool, and see if it does diagnose any issues in setup. Currently you may have to run it in all the three machines. On 26-Oct-2017 6:50 AM, "Amar Tumballi" <atumball at redhat.com> wrote: > Thanks for this report. This week many of the developers are at Gluster > Summit in Prague, will be checking this and respond next
2017 Oct 26
2
not healing one file
Hi Karthik, thanks for taking a look at this. I'm not working with gluster long enough to make heads or tails out of the logs. The logs are attached to this mail and here is the other information: # gluster volume info home Volume Name: home Type: Replicate Volume ID: fe6218ae-f46b-42b3-a467-5fc6a36ad48a Status: Started Snapshot Count: 1 Number of Bricks: 1 x 3 = 3 Transport-type: tcp
2019 Mar 07
2
Removing a mailbox from a dovecot cluster
Le mar. 5 mars 2019 ? 10:08, Gerald Galster via dovecot <dovecot at dovecot.org> a ?crit : > > you could try to stop replication for the user you are deleting: > doveadm replicator remove [-a replicator_socket_path] username > > Good idea! But I have a problem. I tried to stop the replication (doveadm replicator remove <myuser>) for an user on both server. Verified with "doveadm replicator status <myuser>", it doesn't return, so I assume the re...
2017 Aug 08
2
How to delete geo-replication session?
Do you see any session listed when Geo-replication status command is run(without any volume name) gluster volume geo-replication status Volume stop force should work even if Geo-replication session exists. From the error it looks like node "arbiternode.domain.tld" in Master cluster is down or not reachable. regards Aravinda VK On 08/07/2017 10:01 PM, mabi wrote: > Hi, >
2018 May 08
2
replicator: User listing returned failure
...othing happens automatically. >> But mx2:~# doveadm -D sync -u abuse at example.com -d -N -l 30 -U >> successfully executed and getting mails from mx1. prw-rw---- 1 dovecot mail 0 May 8 06:29 replication-notify-fifo > Does it work if you issue: > > doveadm replicator replicate '*' mx1:~# doveadm replicator replicate '*' 2 users updated > doveadm replicator status '*' I already posted this output. > Do you have different hostnames for your servers? Sure. You may see in listing that short hostnames are mx1 and mx2.
2017 Mar 22
2
replicator crashing - oom
Where would I find the core file? I'm not finding anything obvious. The replicator path is /usr/local/libexec/dovecot/replicator Daniel On 3/22/2017 12:52 AM, Aki Tuomi wrote: > Can you provide us gdb bt full dump? > > gdb /usr/libexec/dovecot/replicator /path/to/core > > on some systems, it's /usr/lib/dovecot/replicator > > Aki > > On 21.03.2017...
2016 Jul 31
4
Sieve Script Replication Gliches (Report #2)
Hi, I've observed some odd behaviour with dsync replication between two hosts, specifically to do with sieve script replication. In short, I have two hosts which replicate in a master-master type setup where almost all of the reads and writes happen to just one of the two hosts. They are both running 2.2.devel (9dc6403), which is close to the latest 2.2 -git . Pigeonhole is running
2017 Aug 01
3
How to delete geo-replication session?
Hi, I would like to delete a geo-replication session on my GluterFS 3.8.11 replicat 2 volume in order to re-create it. Unfortunately the "delete" command does not work as you can see below: $ sudo gluster volume geo-replication myvolume gfs1geo.domain.tld::myvolume-geo delete Staging failed on arbiternode.domain.tld. Error: Geo-replication session between myvolume and
2017 Aug 08
0
How to delete geo-replication session?
When I run the "gluster volume geo-replication status" I see my geo replication session correctly including the volume name under the "VOL" column. I see my two nodes (node1 and node2) but not arbiternode as I have added it later after setting up geo-replication. For more details have a quick look at my previous post here:
2017 Aug 08
1
How to delete geo-replication session?
Sorry I missed your previous mail. Please perform the following steps once a new node is added - Run gsec create command again gluster system:: execute gsec_create - Run Geo-rep create command with force and run start force gluster volume geo-replication <mastervol> <slavehost>::<slavevol> create push-pem force gluster volume geo-replication <mastervol>
2012 Feb 05
2
Would difference in size (and content) of a file on replicated bricks be healed?
Hi... Started playing with gluster. And the heal functions is my "target" for testing. Short description of my test ---------------------------- * 4 replicas on single machine * glusterfs mounted locally * Create file on glusterfs-mounted directory: date >data.txt * Append to file on one of the bricks: hostname >>data.txt * Trigger a self-heal with: stat data.txt =>
2011 Mar 04
1
swapping database mid replication
The Xapian documentation talks about a situation where you don't want to do two database swaps while in the process of replicating. It says... To confuse the replication system, the following needs to happen: 1. Start with two databases, A and B. 2. Start a replication of database A. 3. While the replication is in progress, swap B in place of A (ie, by moving the files around, such that B
2017 Sep 29
1
2.2.32 'doveadm replicator replicate -f' segfault
Very minor bug; not specifying the user mask with 'doveadm replicator replicate -f' causes a segfault: server:~# doveadm replicator replicate -f Segmentation fault server:~# doveadm replicator replicate -f '*' 123 users updated server:~# gdb /usr/bin/doveadm core.2418 GNU gdb (GDB) Red Hat Enterprise Linux 7.6.1-100.el7 This GDB was configured as &quo...
2019 Dec 17
2
DSync replication error
Please tell me about Dsync replicator. The following are set to enable the "replication" plug-in /etc/dovecot/conf.d/10-mail.conf *---mail_plugins = notify replication---* And I made the following file /etc/dovecot/conf.d/30-dsync.conf *---service replicator { process_min_avail = 1} ds...
2018 Feb 21
2
Geo replication snapshot error
Hi all, I use gluster 3.12 on centos 7. I am writing a snapshot program for my geo-replicated cluster. Now when I started to run tests with my application I have found a very strange behavior regarding geo-replication in gluster. I have setup my geo-replication according to the docs: http://docs.gluster.org/en/latest/Administrator%20Guide/Geo%20Replication/ Both master and slave clusters are
2011 May 07
1
Gluster "Peer Rejected"
Hello All, I have 8 servers.? 7 of the 8 say that gbe02 is in state State: Peer Rejected (Connected). gbe08 says it is connected to the other?7? but they are all State: Peer Rejected (Connected) So it would appear that gbe02 is out of sync with the group. I triggered a manual self heal by doing a the recommended ./find on a gluster mount. I'm stuck... I cannot find ANY docs on this