similar to: postmaster_address parameter scope

Displaying 20 results from an estimated 20000 matches similar to: "postmaster_address parameter scope"

2018 Jul 07
2
Sieve-filter and default postmaster_address -> error?
I wanted to re-filter my inbox after adding some filters for my DMARC reports. When I didn?t have postmaster_address in my 15-lda.conf, I got an error from sieve-filter that it was invalid (postmaster@). I added the setting (postmaster_address = ler at lerctr.org<mailto:ler at lerctr.org>), and the error went away. Is this expected? Dovecot 2.3.2 Pigeonhole 0.5.2 thebighonker.lerctr.org
2018 Jul 10
0
Sieve-filter and default postmaster_address -> error?
Op 07/07/2018 om 22:22 schreef Larry Rosenman: > > I wanted to re-filter my inbox after adding some filters for my DMARC > reports.? When I didn?t have postmaster_address in my 15-lda.conf, I > got an error from sieve-filter that it was invalid (postmaster@).? I > added the setting (postmaster_address = ler at lerctr.org > <mailto:ler at lerctr.org>), and the error went
2009 Dec 06
3
postmaster_address setting not given
Hi, I am trying to get dovecot working with postfix, and am coming up against this error message. I have this setting configured in the 'protocol lda' section. I am running v1.1.11 dovecot -n -c /etc/dovecot/dovecot-postfix.conf # 1.1.11: /etc/dovecot/dovecot-postfix.conf # OS: Linux 2.6.31-14-generic x86_64 Ubuntu 9.10 reiserfs log_timestamp: %Y-%m-%d %H:%M:%S protocols: imap
2011 Jan 16
1
Dovecot LDA problem: dovecot: deliver(dave): Fatal: postmaster_address setting not given
I have been trying to get the virtual domain and mailboxes working on an newly built Ubuntu 10.04 server and am stuck on this error. System is using a postgres database and PostfixAdmin. Thanks in advance for any help. Dave ==> /var/log/mail.err <== Jan 16 13:21:26 radius dovecot: deliver(dave at mydomain.com): Fatal: postmaster_address setting not given ==> /var/log/mail.log
2008 Aug 20
1
config file parsing, tab characters
Hello, I am using the latest dovecot 1.1.2 with postfix, and using doveco deliver command for LDA. I really love dovecot, and how it has progressed over the years. However I have noticed that if I put a TAB character in the dovecot.conf file, it sometimes does not parse the file correctlty, and complains about missing postmaster_address etc, even though they are there. So I reformated my
2015 Apr 18
0
Installing fts-solr on dovecot (centos 6) - [request for help]
On 2015-04-18 19:23, Kevin Laurie wrote: > Hello, Hello, > I have been trying to get full-text-search on dovecot. I followed the > tutorial on this site:- > https://extremeshok.com/6622/enabling-apache-solr-4-10-using-jetty-with-dovecot-2-2-for-fulltext-search-results-on-centos-6-iredmail-compatible/ > > The instruction on the page here seems wrong:- > sed -i -e
2008 Jan 07
3
Problem with passwords surrounded by curly braces
Hello, one of my users set his password to something surrounded by curly braces, and promptly all access to his mailbox was refused. After looking through the logs I found an entry "Unknown password scheme xxxx", where "xxxx" is the user's password, but without the curly braces. I then browsed through the documentation and there was explained that it is possible to prefix
2017 Aug 28
0
self-heal not working
Great, can you raise a bug for the issue so that it is easier to keep track (plus you'll be notified if the patch is posted) of it? The general guidelines are @ https://gluster.readthedocs.io/en/latest/Contributors-Guide/Bug-Reporting-Guidelines but you just need to provide whatever you described in this email thread in the bug: i.e. volume info, heal info, getfattr and stat output of
2017 Aug 28
0
self-heal not working
On 08/28/2017 01:29 PM, mabi wrote: > Excuse me for my naive questions but how do I reset the afr.dirty > xattr on the file to be healed? and do I need to do that through a > FUSE mount? or simply on every bricks directly? > > Directly on the bricks: `setfattr -n trusted.afr.dirty -v 0x000000000000000000000000
2017 Aug 28
2
self-heal not working
Thank you for the command. I ran it on all my nodes and now finally the the self-heal daemon does not report any files to be healed. Hopefully this scenario can get handled properly in newer versions of GlusterFS. > -------- Original Message -------- > Subject: Re: [Gluster-users] self-heal not working > Local Time: August 28, 2017 10:41 AM > UTC Time: August 28, 2017 8:41 AM >
2017 Aug 28
0
self-heal not working
On 08/28/2017 01:57 AM, Ben Turner wrote: > ----- Original Message ----- >> From: "mabi" <mabi at protonmail.ch> >> To: "Ravishankar N" <ravishankar at redhat.com> >> Cc: "Ben Turner" <bturner at redhat.com>, "Gluster Users" <gluster-users at gluster.org> >> Sent: Sunday, August 27, 2017 3:15:33 PM >>
2017 Aug 27
0
self-heal not working
Thanks Ravi for your analysis. So as far as I understand nothing to worry about but my question now would be: how do I get rid of this file from the heal info? > -------- Original Message -------- > Subject: Re: [Gluster-users] self-heal not working > Local Time: August 27, 2017 3:45 PM > UTC Time: August 27, 2017 1:45 PM > From: ravishankar at redhat.com > To: mabi <mabi at
2017 Aug 27
2
self-heal not working
----- Original Message ----- > From: "mabi" <mabi at protonmail.ch> > To: "Ravishankar N" <ravishankar at redhat.com> > Cc: "Ben Turner" <bturner at redhat.com>, "Gluster Users" <gluster-users at gluster.org> > Sent: Sunday, August 27, 2017 3:15:33 PM > Subject: Re: [Gluster-users] self-heal not working > >
2017 Aug 25
0
self-heal not working
Hi Ravi, Did you get a chance to have a look at the log files I have attached in my last mail? Best, Mabi > -------- Original Message -------- > Subject: Re: [Gluster-users] self-heal not working > Local Time: August 24, 2017 12:08 PM > UTC Time: August 24, 2017 10:08 AM > From: mabi at protonmail.ch > To: Ravishankar N <ravishankar at redhat.com> > Ben Turner
2017 Aug 28
3
self-heal not working
Excuse me for my naive questions but how do I reset the afr.dirty xattr on the file to be healed? and do I need to do that through a FUSE mount? or simply on every bricks directly? > -------- Original Message -------- > Subject: Re: [Gluster-users] self-heal not working > Local Time: August 28, 2017 5:58 AM > UTC Time: August 28, 2017 3:58 AM > From: ravishankar at redhat.com >
2017 Aug 24
2
self-heal not working
Thanks for confirming the command. I have now enabled DEBUG client-log-level, run a heal and then attached the glustershd log files of all 3 nodes in this mail. The volume concerned is called myvol-pro, the other 3 volumes have no problem so far. Also note that in the mean time it looks like the file has been deleted by the user and as such the heal info command does not show the file name
2015 May 12
1
dovecot-lda panics when postmaster_address is given on command line
Hi, When passing a postmaster_address to dovecot-lda on the command line with -o, it panics: $ ./dovecot-lda -o postmaster_address=anything lda(onno,)Panic: file lda-settings.c: line 80 (lda_settings_check): assertion failed: (set->postmaster_address[0] == SETTING_STRVAR_UNEXPANDED[0]) It looks like postmaster_address has already been expanded at that point when it's given on the
2017 Aug 23
2
self-heal not working
I just saw the following bug which was fixed in 3.8.15: https://bugzilla.redhat.com/show_bug.cgi?id=1471613 Is it possible that the problem I described in this post is related to that bug? > -------- Original Message -------- > Subject: Re: [Gluster-users] self-heal not working > Local Time: August 22, 2017 11:51 AM > UTC Time: August 22, 2017 9:51 AM > From: ravishankar at
2017 Aug 24
0
self-heal not working
Unlikely. In your case only the afr.dirty is set, not the afr.volname-client-xx xattr. `gluster volume set myvolume diagnostics.client-log-level DEBUG` is right. On 08/23/2017 10:31 PM, mabi wrote: > I just saw the following bug which was fixed in 3.8.15: > > https://bugzilla.redhat.com/show_bug.cgi?id=1471613 > > Is it possible that the problem I described in this post is
2017 Aug 27
2
self-heal not working
Yes, the shds did pick up the file for healing (I saw messages like " got entry: 1985e233-d5ee-4e3e-a51a-cf0b5f9f2aea") but no error afterwards. Anyway I reproduced it by manually setting the afr.dirty bit for a zero byte file on all 3 bricks. Since there are no afr pending xattrs indicating good/bad copies and all files are zero bytes, the data self-heal algorithm just picks the