similar to: How to read/set ulimit for non-root asterisk process ?

Displaying 20 results from an estimated 8000 matches similar to: "How to read/set ulimit for non-root asterisk process ?"

2020 Sep 16
1
dovecot 2.2.36.4 problem with ulimit
Hi, perhaps this? > with new debian9: > open files (-n) 1024 Regards Urban Am 16.09.20 um 12:57 schrieb Maciej Milaszewski: > Hi > Limits: > > Where all working fine: > > core file size????????? (blocks, -c) 0 > data seg size?????????? (kbytes, -d) unlimited > scheduling priority???????????? (-e) 0 > file size?????????????? (blocks,
2020 Sep 16
2
dovecot 2.2.36.4 problem with ulimit
Hi I update os from debian8 to debian9 # 2.2.36.4 (baf9232c1): /etc/dovecot/dovecot.conf # Pigeonhole version 0.4.24.2 (aaba65b7) # OS: Linux 4.9.0-13-amd64 x86_64 Debian 9.13 All works fine but sometimes I get: Sep 16 09:17:00 dovecot4 dovecot: master: Error: service(pop3): fork() failed: Resource temporarily unavailable (ulimit -u 257577 reached?) Sep 16 09:17:00 dovecot4 dovecot: master:
2017 Mar 29
2
cannot login to imap under load
Hi Steffen, On 29-03-17 12:38, Steffen Kaiser wrote: > On Tue, 28 Mar 2017, Gerard Ranke wrote: > >> dovecot: master: Error: service(imap): fork() failed: Resource >> temporarily unavailable >> dovecot: master: Error: service(imap): command startup failed, >> throttling for 2 secs > > check out the ulimits for the Dovecot process. > > -- Steffen Kaiser
2018 Jan 14
2
Lmtp Memory Limit
Hi, i am using dovecot 2.2.33.2 on CentOS 7.4. Since i upgraded from CentOS 7.2. to CentOS 7.4. (without upgrading dovecot), my dovecot sieve-pipe scripts crash with Out of memory: Out of memory (allocated 262144) (tried to allocate 8793 bytes) There are some memory limits in dovecot or sieve? Can i change this value? Kernel limitks: [root at xxx software]# ulimit -a core file size
2008 Aug 04
1
pam max locked memory issue after updating to 5.2 and rebooting
We were previously running 5.1 x86_64 and recently updated to 5.2 using yum. Under 5.1 we were having problems when running jobs using torque and the solution had been to add the following items to the files noted "* soft memlock unlimited" in /etc/security/limits.conf "session required pam_limits.so" in /etc/pam.d/{rsh,sshd} This changed the max
2006 Apr 04
1
Too many open files
Dear all, we have encounter problem when starting asterisk in the foreground, "asterisk -vvvvgc" with more 100 SIP calls concurrently. we have set ulimit to the highest value. still has this problem. Is this the problem keeping asterisk in the foreground or this is a bug in SVN 1.2 16771? Apr 5 08:48:36 WARNING[14887]: channel.c:562 ast_channel_alloc: Channel allocation
2018 Jan 15
1
Aw: Re: Lmtp Memory Limit
2008 Oct 21
4
Too many open files
Hi, other error in "open files": dovecot: Oct 20 10:52:36 Fatal: execv(/usr/local/sbin/post-login-imap-dovecot) failed: Too many open files dovecot: Oct 20 10:52:36 Error: child 1038 (imap) returned error 84 (exec() failed) dovecot: Oct 20 10:55:51 Fatal: execv(/usr/local/sbin/post-login-imap-dovecot) failed: Too many open files dovecot: Oct 20 10:55:51 Error: child 4853 (imap)
2016 Feb 17
2
New glibc for CentOS-6 and CentOS-7 and CVE-2015-7547
On 02/17/2016 08:10 AM, Michael H wrote: >> The easy answer is yes .. glibc requires so many things to be restarted, >> that is the best bet. Or certainly the easiest. >> >> Note: in CentOS 7, there is also a kernel update which is rated as >> Important .. so you should boot to that anyway: >>
2005 Dec 21
3
Another "Too many files open" post
I'm running Solaris 9 Sparc with Dovecot alpha 5 and using maildir format.. I noticed in the log file the following error: dovecot: Dec 20 15:54:18 Error: imap(xxxxxxxx): open(/export/home/xxxxxxxx/mail/.Sent Messages/cur/1135013649.002916.mbox:2,Sab) failed: Too many open files I called the guy and asked him what he was doing... synching his folders. It hung for a while but finally
2017 Aug 23
2
socketpair failed: Too many open files on Debian 9
I haven't done this on the old, working machine. So there must be a difference between Debian 7 and 9 how open files are handled? Regards Patrick Aki Tuomi schrieb: > You probably need to increase ulimit -n > > Aki > > > On 23.08.2017 14:10, Patrick Westenberg wrote: >> Hi @all, >> >> after re-installing one of my two frontends/proxy-servers I get
2016 Feb 17
1
New glibc for CentOS-6 and CentOS-7 and CVE-2015-7547
> The easy answer is yes .. glibc requires so many things to be restarted, > that is the best bet. Or certainly the easiest. > > Note: in CentOS 7, there is also a kernel update which is rated as > Important .. so you should boot to that anyway: > https://lists.centos.org/pipermail/centos-announce/2016-February/021705.html > > Here is a good link to figure out what to
2015 Sep 13
1
concerning dovecot settings for high volume server
----- Original Message ----- From: Tony Morehen [mailto:tmorehen at ajmconsulting.ca] To: dovecot at dovecot.org Sent: Sat, 12 Sep 2015 17:57:27 -0400 Subject: Re: concerning dovecot settings for high volume server You may be running up against Linux system/user limits. Run $ cat /proc/sys/kernel/pid_max and $ ulimit -a That should give some insight into your problem. On 12/09/2015 2:53 PM,
2017 May 25
2
Re: can't establish more than 1000 connections with virsh
在 2017年05月25日 18:37, Daniel P. Berrange 写道: > On Thu, May 25, 2017 at 06:20:51PM +0800, dw wrote: >> Hi: >> >> I'm trying to connect with libvirtd with virsh from a remote PC,but only >> can establish 1000 connections. >> >> If try more connections,prompt: >> >> "error: failed to connect to the hypervisor >>
2010 Sep 09
2
set lockedmem to higher value?
We have some ORACLE PGA problem and ORACLE document say we can serup lockmem(Max lock memory) to higher value. $ulimit -a time(cpu-seconds) unlimited file(blocks) unlimited coredump(blocks) 0 data(kbytes) unlimited stack(kbytes) 10240 lockedmem(kbytes) 3145728 memory(kbytes) unlimited nofiles(descriptors) 65536 processes 2047 Can anyone tell me how to set lockedmem to higher value on
2017 Aug 23
3
socketpair failed: Too many open files on Debian 9
Hi @all, after re-installing one of my two frontends/proxy-servers I get the following error messages after some time (sometimes after 1h, sometimes after 24h): 11:23:55 imap-login: Error: socketpair() failed: Too many open files 11:23:55 imap-login: Error: socketpair() failed: Too many open files 11:23:56 imap-login: Error: socketpair() failed: Too many open files 11:23:56 imap-login: Error:
2007 Aug 27
3
rsync out of memory at 8 MB although ulimit is 512MB
Hello again, I encountered something amazing. First I thought there is not enough memory allowed through ulimit. ulimit is now set to (almost) 512MB but rsync still gets out fo memory at 8MB. Can anyone tell me why? That's my configuration: rsync version 2.6.2 from AIX 5.3 to SuSE Linux 9 (also has rsync 2.6.2) ulimit -a (AIX) ulimit -a AIX (source): -------------------------
2012 Sep 11
3
[LLVMdev] Fwd: Build Error from Intrinsics.td
ulimit -s = 8192 set "ulimit -c unlimited" On Tue, Sep 11, 2012 at 3:03 PM, Ryan Taylor <ryta1203 at gmail.com> wrote: > John, > > Thanks for responding. No, I don't see a limit from ulimit. It's > definitely with the tblgen though, I have the same errors trying to compile > clang. > > > On Tue, Sep 11, 2012 at 2:57 PM, John Criswell
2020 Sep 16
0
dovecot 2.2.36.4 problem with ulimit
Hi Limits: Where all working fine: core file size????????? (blocks, -c) 0 data seg size?????????? (kbytes, -d) unlimited scheduling priority???????????? (-e) 0 file size?????????????? (blocks, -f) unlimited pending signals???????????????? (-i) 257970 max locked memory?????? (kbytes, -l) 64 max memory size???????? (kbytes, -m) unlimited open files????????????????????? (-n) 65536 pipe
2015 Jan 26
3
imap-login: Fatal: pipe() failed: Too many open files
Am 26.01.2015 um 02:13 schrieb Leander Sch?fer: > I just checked my ulimit again and it really seems like it has more > than enough - so I still don't understand what I've configured wrong > here ;/ > > root at WM-01 [~]$ su -m dovecot -c "ulimit -a" > socket buffer size (bytes, -b) unlimited > core file size (blocks, -c) unlimited > data