Displaying 20 results from an estimated 10000 matches similar to: "Can`t get over 1024 processes on FreeBSD - possible bug?"
2012 Sep 17
1
Can`t get over 1024 processes
Hello,
I have configuration question. Running FreeBSD 9.0 / AMD64 / Dovecot
2.1.8 version and cannot get over 1024 processes. I`ve increased
process_limit in imap and pop3 and still have messages in syslog like:
maxproc limit exceeded by uid 89, please see tuning(7) and login.conf(5).
Did I forget to increase some other value? System values are higher -
kern.maxprocperuid: 8192 and
2003 Apr 06
1
load testing and tuning a 4GB RAM server
Hello everyone,
First of all, great job on the 4.8-R. We have been a long standing user of
FreeBSD and are very happy with everything.
Now my question. I am trying to stress test a new Dell PowerEdge server
and find the limits of its hardware and my tuning. Here are the server
stats:
* 2x Xeon 2.8 with SMP compiled, hyperthreading NOT compiled in kernel
* 4 GB of RAM, 8 GB of swap on Raid 1
2017 Feb 09
1
dovecot logout issues
hi
we are using dovecot version 2.2.7 (config file given below)
centos 6, qmail, vpopmail, mysql
server configuration
hex core processor, 16 gb ram 1 X 600 gb 15 k rpm for main drive and 2 X 2000 gb hdd for data (No raid)
busy server with around 4000 email ids --- load is around 2 to 10
the issue is that SQUIRRELMAIL webmail users suddenly lose connection while they are working on the webmail.
2020 Sep 07
2
Btrfs RAID-10 performance
Dne 07.09.2020 v 12:43 Sami Ketola napsal(a):
>> On 7. Sep 2020, at 12.38, Miloslav H?la <miloslav.hula at gmail.com> wrote:
>>
>> Hello,
>>
>> I sent this into the Linux Kernel Btrfs mailing list and I got reply: "RAID-1 would be preferable" (https://lore.kernel.org/linux-btrfs/7b364356-7041-7d18-bd77-f60e0e2e2112 at lechevalier.se/T/). May I ask you
2023 Jan 07
1
replicator: Panic: data stack: Out of memory when allocating 268435496 bytes
ok a little bit more info ....
My servers all run under FreeBSD 12.xx
which was also the base for the apple operating system origionally.
setting default_vsz_limit = 0 i seem to remember trying with no so great
results
setting to zero can cause memory over runs (espically with replication)
etc i found that when i used the config i sent eariler (vsz_limit is
defaulted (not set )) everything
2012 Aug 31
1
New log entries with 2.0.19?
I just upgraded my Ubuntu server from 11.10 to 12.04.1 and that upgraded
Dovecot to 2.0.19. For some reason I'm getting a lot more (5 times)
disconnect in IDLE and inactivity time outs. I'm using only IMAP with
two users. Both users access the server from multiple devices, phone,
tablet and desktop computer. I had previously set the
max_userip_connections to 20 and rarely got the
2012 Jun 14
1
disable_plaintext_auth = no as no effect on IMAP/POP3 logins
Hello
In my installation the disable_plaintext_auth does not appear to take
effect.
I can see that the value is correct using doveconf -a but it doesn't
change anything.
Whenever attempting to log in using IMAP I get this:
* BAD [ALERT] Plaintext authentication not allowed without SSL/TLS, but
your client did it anyway. If anyone was listening, the password was
exposed.
ls NO
2011 Mar 24
3
Dovecot 2.0.9: Thunderbird 3.1.9: Very slow while open sent folder with many (4k>) mails
hi,
one customer told us, that he has performance problem in his Thunderbird
3.1.9, while open his sent folder, which has round about 4000 mails in
it, and 20k in his Inbox. For example TB, tries to sync the sent-folder
list (over SSL) from the sent folder for now over 15min.
In the log, I can't see any strange things, so for me, Dovecot works ok.
Maybe, I should use better cache options?
2012 Jun 08
2
difference between client_limit and process_limit
Hi,
What is the real difference between client and process limit? According
to documentation (http://wiki2.dovecot.org/Services#Service_limits):
client_limit: Maximum number of simultaneous client connections. If set
to 0, default_client_limit is used instead.
process_limit: Maximum number of processes that can exist for this
service. If set to 0, default_process_limit is used instead.
But
2014 May 11
2
questions about process_limit
Hello all!
# dovecot --version
2.1.17
Sometimes I have this in the logfile:
May 11 16:55:52 master: Warning: service(imap-login): process_limit (100) reached, client connections are being dropped
May 11 17:35:03 master: Warning: service(imap-login): process_limit (100) reached, client connections are being dropped
May 11 17:36:27 master: Warning: service(imap-login): process_limit (100)
2023 Jan 06
1
replicator: Panic: data stack: Out of memory when allocating 268435496 bytes
How problematic is it to have
default_vsz_limit = 0
in dovecot.conf? macOS+MacPorts had this as a requirement even.
Gerben
> On 6 Jan 2023, at 16:49, Paul Kudla <paul at scom.ca> wrote:
>
>
> i ran into this as well
>
> here is the full config for mine with replication
>
> # cat dovecot.conf
> # 2.3.14 (cee3cbc0d): /usr/local/etc/dovecot/dovecot.conf
> #
2015 Sep 13
1
concerning dovecot settings for high volume server
----- Original Message -----
From: Tony Morehen [mailto:tmorehen at ajmconsulting.ca]
To: dovecot at dovecot.org
Sent: Sat, 12 Sep 2015 17:57:27 -0400
Subject: Re: concerning dovecot settings for high volume server
You may be running up against Linux system/user limits. Run
$ cat /proc/sys/kernel/pid_max
and
$ ulimit -a
That should give some insight into your problem.
On 12/09/2015 2:53 PM,
2018 Nov 14
2
huge increase in storage activity afther dovecot upgrade
Hi,
we upgraded our servers from version 2.2.13 to 2.2.27. After the upgrade we
notice a 10x increase in traffic with the nfs storage an errors like this
in the logfile:
Nov 12 09:48:16 mail dovecot: imap(...): Error: Corrupted index cache file
/.../dovecot.index.cache: invalid record size
Nov 12 09:48:16 mail dovecot: imap(...): Error:
unlink(/.../dovecot.index.cache) failed: No such file or
2018 Dec 18
2
High Load average on NFS Spool - v.2.1.15 & 2.2.13
I have two servers pointing to an NFS mounted mail spool with dovecot.?
Since I recently switched from using Dovecot v1.X, I have been
experiencing high CPU use with the two Dovecot servers. I am not certain
why they are not well behaved.? Here is the configuration information.
This configuration is currently running at a load average of 17.
/usr/sbin/dovecot -n
# 2.1.15:
2018 Nov 14
2
huge increase in storage activity afther dovecot upgrade
Thank you !
I was little concerned that the following settings are not in line with the
new version:
mail_nfs_index = yes
mail_nfs_storage = yes
mail_fsync = always
mailbox_list_index = yes
maildir_stat_dirs = yes
mmap_disable = yes
On Wed, Nov 14, 2018 at 10:19 AM Aki Tuomi <aki.tuomi at open-xchange.com>
wrote:
> It should eventually wind down once all the problems are fixed. Of
2016 Sep 12
5
Sieve filtering SPAM and redirect
Hi every body. I'm having difficulties to setup my filters correctly
and I really need some help.
This is my environment:
Mail that is received on the system passes through a sieve_before
script that checks the message header for SPAM tag and store it into
the "Junk" folder. If no spam tag is found, mail goes to inbox.
My horde webmail is integrated with sieve, so this allow
2017 Aug 03
1
rawlog
> On August 3, 2017 at 8:57 PM Alexander Dalloz <ad+lists at uni-x.org> wrote:
>
>
> Am 03.08.2017 um 01:04 schrieb Matt Bryant:
> > Hi,
> >
> > Trying to get rawlog working on dovecot 2.2.31 configured as per
> >
> > https://wiki2.dovecot.org/Debugging/Rawlog
> >
> > but
> >
> > a) it doesnt appear to be loggin anything
2020 Jan 06
4
Upgrade 2.2.27 to 2.3.9.2: master(imap): net_connect_unix(imap) failed: Resource temporarily unavailable
Hi,
After upgrading Dovecot from version 2.2.27 to 2.3.9.2, we are
sporadically seeing lots of these errors in the error log on many of our
servers:
imap-login: Error: master(imap): net_connect_unix(imap) failed:
Resource temporarily unavailable -
http://wiki2.dovecot.org/SocketUnavailable
The issue is causing significant delays and/or timeouts on login.
From what I can tell, this
2018 Dec 18
2
High Load average on NFS Spool - v.2.1.15 & 2.2.13
I have, but I will be happy to review it once again.
On 12/18/18 2:14 PM, admin wrote:
> Am Dienstag, den 18.12.2018, 14:06 -0500 schrieb Albert E. Whale, CEH
> CHS CISA CISSP:
>>
>> I have two servers pointing to an NFS mounted mail spool with
>> dovecot.? Since I recently switched from using Dovecot v1.X, I have
>> been experiencing high CPU use with the two
2018 Feb 17
4
2.2.33.2 - Virtual Mailbox Issue bug maybe ?
Been experimenting with virtual mailbox and have a question :-
I have defined 2 virtual mailboxes 'all' and 'combined' as follows ...
virtual/all/dovecot-virtual
virtual/combined
inthread refs x-mailbox INBOX
virtual/combined/dovecot-virtual
!INBOX
INBOX/Sent
INBOX/Drafts
all
When new mail comes I would assume that both virtual mailboxes should
actually see the message