similar to: dovecot-imapd eating 100% cpu every once in a while

Displaying 20 results from an estimated 300 matches similar to: "dovecot-imapd eating 100% cpu every once in a while"

2010 Oct 10
3
pop3 TCP_CORK too late error
I was straceing a pop3 process and noticed that the TCP_CORK option isn't set soon enough: epoll_wait(8, {{EPOLLOUT, {u32=37481984, u64=37481984}}}, 38, 207) = 1 write(41, "iTxPBrNlaNFao+yQzLhuO4/+tQ5cuiKSe"..., 224) = 224 epoll_ctl(8, EPOLL_CTL_MOD, 41, {EPOLLIN|EPOLLPRI|EPOLLERR|EPOLLHUP, {u32=37481984, u64=37481984}}) = 0 pread(19,
2010 Mar 21
3
2.0.beta3 imapd running hot
Hi, since a couple of weeks I occasionally have an imapd going crazy on me, using up 100% CPU. Current revision is 10962 29865 vmail 20 0 47820 3296 1708 R 99.7 0.9 131:07.85 imap vmail 29865 86.1 0.8 47820 3296 ? R Mar20 131:20 dovecot/imap [berni 2001:a60:f001:1:219:66ff:fe8b:a6e IDLE strace -p <pid> shows a repeating pattern of epoll_wait(8, {{EPOLLOUT,
2019 Mar 08
1
Dovecot v2.3.5 released
On 7.3.2019 23.37, A. Schulze via dovecot wrote: > > Am 07.03.19 um 17:33 schrieb Aki Tuomi via dovecot: > >>> test-http-client-errors.c:2989: Assert failed: FALSE >>> connection timed out ................................................. : FAILED > Hello Aki, > >> Are you running with valgrind or on really slow system? > I'm not aware my buildsystem
2004 Aug 30
0
[PATCH] [RFC] epoll based ioloop handler (now with patch)
Sorry, I forgot to attach patch to previous mail. -- Andrey Panin | Linux and UNIX system administrator pazke at donpac.ru | PGP key: wwwkeys.pgp.net -------------- next part -------------- diff -urpNX /usr/share/dontdiff dovecot-cvs.vanilla/configure.in dovecot-cvs/configure.in --- dovecot-cvs.vanilla/configure.in 2004-08-09 23:14:02.000000000 +0400 +++ dovecot-cvs/configure.in 2004-08-25
2012 Jun 26
1
Segmentation fault with latest 1.1 revision
Hello, I am trying 1.1 branch and I experience a segmentation fault upon ALRM signal. This looks like a race condition. I have my tincd daemon instantiated manually in if-up.d/jmuchemb (without IF_TINC_NET) and when if-up.d/tinc runs, it sends a ALRM signal that makes tincd crash. It fails here: Core was generated by `tincd -D -n jmuchemb -d -o ConnectTo srv -o srv.Address 81.x.y.z -o
2009 Feb 24
4
dovecot1.2 segfault
I've tried 2 builds of dovecot, based on a amd64 rebuild of the experimental package on http://xi.rename-it.nl/debian/ I've tried to include as much useful info as possible, sorry if I've missed out anything of use. Let me know as I'm more then happy to assist in any way possible. $ telnet host 110 Escape character is '^]'. +OK Dovecot ready. USER ian at domain.com +OK
2017 Jun 17
1
client reconnect fails (was gluster heal entry reappears)
Hi Ravi, back to our client-cannot-reconnect-to-gluster-brick problem ... > Von: Ravishankar N [ravishankar at redhat.com] > Gesendet: Montag, 29. Mai 2017 06:34 > An: Markus Stockhausen; gluster-users at gluster.org > Betreff: Re: [Gluster-users] gluster heal entry reappears > > > On 05/28/2017 10:31 PM, Markus Stockhausen wrote: > > Hi, > > > > I'm
2007 Dec 11
1
1.1.beta10 pop3 process hangs with 100% CPU
Hello, we have observed pop3 process which got stuck consuming all available CPU. It seems that it happened because of some kind of abnormal POP3 connection termination. Here is strace info for this process: 13:36:05.866190 writev(1, [{"508qWWH96If+uVXeH2Zxl/hkn+plVwmI"..., 3975}, {"HP1oxt+np0o4Xtz27VQBtxx0zWfGuA3r"..., 193}], 2) = -1 EPIPE (Broken pipe) 13:36:05.866250 ---
2014 Nov 12
1
closed fd causes: lmtp(18385): Panic: epoll_ctl(del, 11) failed: Bad file descriptor
-----BEGIN PGP SIGNED MESSAGE----- Hash: SHA1 See the end of the message for a possible delete of a closed fd from epoll() problem. Every now and then I get a Panic after a successfull LMTP delivery: 2014-11-12 13:25:31 lmtp(20890): Info: Connect from local 2014-11-12 13:25:32 lmtp(20890): Info: Disconnect from local: Connection closed (in reset) 2014-11-12 13:25:34 lmtp(20890): Info: Connect
2018 Sep 09
0
[PATCH net-next v8 5/7] net: vhost: introduce bitmap for vhost_poll
On Tue, Aug 21, 2018 at 8:45 AM Jason Wang <jasowang at redhat.com> wrote: > > > > On 2018?08?19? 20:11, xiangxia.m.yue at gmail.com wrote: > > From: Tonghao Zhang <xiangxia.m.yue at gmail.com> > > > > The bitmap of vhost_dev can help us to check if the > > specified poll is scheduled. This patch will be used > > for next two patches. > >
2023 Aug 30
2
[libnbd PATCH 0/2] (Attempt to) fix Rust on BSD-based builds
I managed to get a build of the async Rust handle compiling on FreeBSD (although the cirrus CI appears to not actually run 'make check' on non-Linux machines, at least when run on my fork): https://gitlab.com/ebblake/libnbd/-/jobs/4985192286 However, I'd really like Tage's review on patch 2 to see if my Rust makes sense. Eric Blake (2): maint: Favor 4-space indent in .rs files
2019 Oct 24
3
winbind : suspend nightmare
On Wed, 2019-10-23 at 22:21 -0700, Jeremy Allison wrote: > On Wed, Oct 23, 2019 at 11:58:33PM +0000, Jon Gerdes wrote: > > winbind has a concept of offline and online but I don't know what > > that > > is, nor how nss works with it. I've tried using smbcontrol to tell > > winbind it is offline or online but that does not seem to work. > > Restarting
2008 Aug 24
2
Unusual bug in glusterfsd
Hi, I'm rather new to this project, having stumbled across it earlier this afternoon, so forgive me if I'm still trying to find my way around. I was in the need of an alternative to NFS that would let me spread the task of sharing my downloaded source code files across a couple of boxes, and GlusterFS looked like a great candidate, having had no luck with Coda or OpenAFS. I also want
2019 May 16
6
[PATCH net 0/4] Prevent vhost kthread from hogging CPU
Hi: This series try to prvernt a guest triggerable CPU hogging through vhost kthread. This is done by introducing and checking the weight after each requrest. The patch has been tested with reproducer of vsock and virtio-net. Only compile test is done for vhost-scsi. Please review. This addresses CVE-2019-3900. Jason Wang (4): vhost: introduce vhost_exceeds_weight() vhost_net: fix possible
2020 May 29
0
[PATCH 1/6] vhost: allow device that does not depend on vhost worker
vDPA device currently relays the eventfd via vhost worker. This is inefficient due the latency of wakeup and scheduling, so this patch tries to introduce a use_worker attribute for the vhost device. When use_worker is not set with vhost_dev_init(), vhost won't try to allocate a worker thread and the vhost_poll will be processed directly in the wakeup function. This help for vDPA since it
2019 May 17
0
[PATCH V2 1/4] vhost: introduce vhost_exceeds_weight()
We used to have vhost_exceeds_weight() for vhost-net to: - prevent vhost kthread from hogging the cpu - balance the time spent between TX and RX This function could be useful for vsock and scsi as well. So move it to vhost.c. Device must specify a weight which counts the number of requests, or it can also specific a byte_weight which counts the number of bytes that has been processed.
2004 Oct 25
0
[PATCH] move iolist functions into separate file
Hello, attached patch extracts iolist related functions into separate file and makes it a little more generic. This functions will be used by the new ioloop handler based on the BSD kqueue API. Please consider applying. Best regards. -- Andrey Panin | Linux and UNIX system administrator pazke at donpac.ru | PGP key: wwwkeys.pgp.net -------------- next part -------------- diff -udrpNX
2019 May 17
9
[PATCH V2 0/4] Prevent vhost kthread from hogging CPU
Hi: This series try to prevent a guest triggerable CPU hogging through vhost kthread. This is done by introducing and checking the weight after each requrest. The patch has been tested with reproducer of vsock and virtio-net. Only compile test is done for vhost-scsi. Please review. This addresses CVE-2019-3900. Changs from V1: - fix user-ater-free in vosck patch Jason Wang (4): vhost:
2019 May 17
9
[PATCH V2 0/4] Prevent vhost kthread from hogging CPU
Hi: This series try to prevent a guest triggerable CPU hogging through vhost kthread. This is done by introducing and checking the weight after each requrest. The patch has been tested with reproducer of vsock and virtio-net. Only compile test is done for vhost-scsi. Please review. This addresses CVE-2019-3900. Changs from V1: - fix user-ater-free in vosck patch Jason Wang (4): vhost:
2019 Sep 27
0
[RFC PATCH 04/13] vsock: add 'transport' member in the struct vsock_sock
As a preparation to support multiple transports, this patch adds the 'transport' member at the 'struct vsock_sock'. This new field is initialized during the creation in the __vsock_create() function. This patch also renames the global 'transport' pointer to 'transport_single', since for now we're only supporting a single transport registered at run-time.