similar to: Fixing split brain

Displaying 20 results from an estimated 400 matches similar to: "Fixing split brain"

2017 Sep 16
2
dsync panic in mbox_lock
Hello I am trying to setup replication with dovecot-2.2.29.1, and for some users, I get a reproductible panic: # doveadm -v -o plugin/mail_replica=remoteprefix:root at mail2.example.net sync -d -u jdoe dsync-local(jdoe): Panic: file mbox-lock.c: line 799 (mbox_lock): assertion failed: (lock_type == F_RDLCK || mbox->mbox_lock_type != F_RDLCK) Abort User is not overquota, and filesystem
2013 Sep 13
1
glusterfs-3.4.1qa2 released
RPM: http://bits.gluster.org/pub/gluster/glusterfs/3.4.1qa2/ SRC: http://bits.gluster.org/pub/gluster/glusterfs/src/glusterfs-3.4.1qa2.tar.gz This release is made off jenkins-release-42 -- Gluster Build System
2013 Nov 29
2
BUG: Authentication client sent unknown handshake command
Hi After upgrading the kernel, everything is fine, except dovecot authentication. I get this trange thing (data after REQUEST? changed just in case it contains anything sensitive): Nov 29 16:56:01 volanges dovecot: auth: Error: BUG: Authentication client sent unknown handshake command: REQUEST?6970356762?616?6?235264ef69dbd1665538af54d12fdaea?session_pid=453?req... Nov 29 16:56:01 volanges
2009 Jun 24
3
Auth worker max line size
Hello I have been playing with some exotic authentication scheme with Dovecot and PAM. That involves sending really large base64 encoded data as the IMAP password, and I have hit a line limit in Dovecot, with AUTH_WORKER_MAX_LINE_LENGTH set to 1024. This limit is especially frustrating since other parts of the software use much larger limits: MAX_INBUF_SIZE 4096 MAX_IMAP_LINE 8192
2013 Jul 26
5
[FEEDBACK] Governance of GlusterFS project
Hello everyone, We are in the process of formalizing the governance model of the GlusterFS project. Historically, the governance of the project has been loosely structured. This is an invitation to all of you to participate in this discussion and provide your feedback and suggestions on how we should evolve a formal model. Feedback from this thread will be considered to the extent possible in
2012 Jul 24
19
what best for anti-spam filter?
what anti-spam for you used ? dspam?spammassian? amavisd-new ? what is best ?
2013 May 05
5
dovecot 2.2.0 corrupts mailboxes?
Hi On april 17th, I upgraded from dovecot 2.1.13 to 2.2.0. Since that time, I had two different users that reported received three incident of messages that disapeared from their mailboxes. The mailbox format is mbox on local FFS filesystem (no NFS), and I use filesystem quotas (but both users are far from filling their quotas). When the message disapeared, it was always a whole rand of dates.
2012 Jun 23
1
pop3-throttle
Hello I am having a hard time with users using POP while leaving mailboxes of several gigabyte cumulated. This causes a lot of disk I/O and kills performancs for everyone. I try to encourage people migrating to IMAP, but that migration will take some time, and therefore I am looking for alterantive ways to workaround the problem. I found pop3-throttle-plugin.c, which seems a smart way to solve
2013 Jan 19
1
[PATCH] support for NetBSD 6.0 libquota
Hi NetBSD 6.0 introduced a new improved quota subsystem and an unified API (libquota) to handle the new and old quota implementation. dovecot is unable to use the new quota implementation right now, and will even fail to build with the old quota API on NetBSD 6.0: Here are dovecot patches to support NetBSD libquota: http://ftp.espci.fr/shadow/manu/dovecot-libquota.tgz -- Emmanuel Dreyfus
2017 Sep 17
1
dsync panic in mbox_lock
Aki Tuomi <aki.tuomi at dovecot.fi> wrote: > We have seen this before but so far the actual bug has not been > reproducible for us. Can you post your doveconf -n Here it is: https://ftp.espci.fr/shadow/manu/conf.log > and also logs with > doveadm -Dv -o plugin/mail_replica=remoteprefix:root at mail2.example.net > sync -d -u jdoe https://ftp.espci.fr/shadow/manu/sync.log
2013 Mar 07
4
[Gluster-devel] glusterfs-3.4.0alpha2 released
RPM: http://bits.gluster.org/pub/gluster/glusterfs/3.4.0alpha2/ SRC: http://bits.gluster.org/pub/gluster/glusterfs/src/glusterfs-3.4.0alpha2.tar.gz This release is made off jenkins-release-19 -- Gluster Build System _______________________________________________ Gluster-devel mailing list Gluster-devel at nongnu.org https://lists.nongnu.org/mailman/listinfo/gluster-devel
2011 Oct 05
1
Directory listings not working
Hello, I just finished installing gluster on two machines in server mode in EC2. I have mounted it via fuse on one of the boxes. Here is my volume info: # gluster volume info Volume Name: fast Type: Stripe Status: Started Number of Bricks: 2 Transport-type: tcp Bricks: Brick1: server1:/data2 Brick2: server2:/data All of this works great, and I can write files at a fairly high throughput.
2017 Sep 16
0
dsync panic in mbox_lock
> On September 16, 2017 at 5:28 AM Emmanuel Dreyfus <manu at netbsd.org> wrote: > > > Hello > > I am trying to setup replication with dovecot-2.2.29.1, and for some > users, I get a reproductible panic: > > # doveadm -v -o plugin/mail_replica=remoteprefix:root at mail2.example.net > sync -d -u jdoe > dsync-local(jdoe): Panic: file mbox-lock.c: line 799
2017 Sep 02
0
mail_log_events and dsync
Hello mail_log_events is handy to track what happened to a given message. Unfortunatly, it seems dsync activity is not captured. This causes messages to appear or vanish without a log trace. Did I miss a setting to get it? How should I track how something went wrong with dsync? -- Emmanuel Dreyfus http://hcpnet.free.fr/pubz manu at netbsd.org
2013 Apr 11
2
[PATCHES] NetBSD support, authentication buffer size
Hi Here are a few unintegrated patches, just tested against 2.2rc7: 1) NetBSD's getmntinfo uses struct statvfs while other BSD use struct statfs http://ftp.espci.fr/shadow/manu/patch-ak 2) NetBSD 5.x net_getunixcred() support. Build on NetBSD, but not tested (I am testing on NetBSD 6.0): http://ftp.espci.fr/shadow/manu/patch-src_lib_net.c 3) Increase authentication buffer size so that it
2013 Jan 07
0
access a file on one node, split brain, while it's normal on another node
Hi, everyone: We have a glusterfs clusters, version is 3.2.7. The volume info is as below: Volume Name: gfs1 Type: Distributed-Replicate Status: Started Number of Bricks: 94 x 3 = 282 Transport-type: tcp We native mount the volume in all nodes. When we access the file ?/XMTEXT/gfs1_000/000/000/095? on one nodes, the error is split brain. While we can access the same file on
2009 Jun 11
2
Issue with files on glusterfs becoming unreadable.
elbert at host1:~$ dpkg -l|grep glusterfs ii glusterfs-client 1.3.8-0pre2 GlusterFS fuse client ii glusterfs-server 1.3.8-0pre2 GlusterFS fuse server ii libglusterfs0 1.3.8-0pre2 GlusterFS libraries and translator modules I have 2 hosts set up to use AFR with
2020 Jan 10
2
Dovecot HA/Resilience
On Thu, Jan 09, 2020 at 06:51:36PM +0200, Aki Tuomi wrote: > You can do it using replication, > https://wiki.dovecot.org/Replication Last time I tried, it did not work with mbox. Did that change? The document does not tell about the format. -- Emmanuel Dreyfus manu at netbsd.org
2020 Jan 10
2
Dovecot HA/Resilience
On Fri, Jan 10, 2020 at 09:07:24AM +0200, Aki Tuomi wrote: > Replication is not supported with mbox. Most features are not. It would be nice if the document about replication could tell what setup works. -- Emmanuel Dreyfus manu at netbsd.org
2020 Jan 10
3
Dovecot HA/Resilience
Thank you all for the replies.... I have the test environment with the same configuration. But I have been asked to go with same environment for HA/Resilience in Live. Yes, I have only one Live server. It is configured in "Maildir" format. The data stores on a Network / Shared Storage (But definitely not local disk, its a mount point). I have been asked to create a HA/Resilience for