similar to: Asterisk dead but pid file exists - gdb asterisk core.13089

Displaying 20 results from an estimated 60000 matches similar to: "Asterisk dead but pid file exists - gdb asterisk core.13089"

2014 Oct 20
1
2.2.14 Panic in sync_expunge_range()
I am getting some panics after upgrading from 2.2.13 to 2.2.14 This panic happens for one user only, he is subscribed to 86 folders, on two of them this panic happens quite often - several times a day. The mbox folders seems OK, less than 30M with 30 and 200 messages. Panic: file mail-index-sync-update.c: line 250 (sync_expunge_range): assertion failed: (count > 0) hmk GNU gdb 6.8
2018 Dec 20
2
Fatal: master: service(indexer-worker): child 493 killed with signal 11 (core dumped)
I'm really sorry, but it was a long day yesterday. I forgot to mention that I used alpine:edge (based on the alpine master branch) image to set up the build environment. So there were some version changes. dovecot --version: 2.3.3 (dcead646b) apk info musl: musl-1.1.20-r2 description: the musl c library (libc) implementation Configuration was unchanged wrt to my first mail. Maybe it's
2018 Dec 21
2
Fatal: master: service(indexer-worker): child 493 killed with signal 11 (core dumped)
On 21.12.2018 15.01, Tim Mohlmann via dovecot wrote: > Hi, I now this is kinda impolite to ask, but I would like some feedback > on this issue. I'm representing the Mailu user group here, plus the fact > that my own production server is affected and starts to crash due to > file corruption as a result of the indexer-worker segfaults. Last night > my server became unavailable
2016 Sep 03
2
doveadm quota recalc returns a Segmentation fault
Hi, with the latest dovecot-ee version (2.2.25.4 (5cb7c97)) if I run ?doveadm quota recalc -u alessio at skye.it? dovecot return a ?Segmentation fault? (but the quota is update fine). Since dovecot version 2.2.24.3 (55cdc32) works fine. This is my quota backends configuration: plugin { quota = maildir:UserQuota quota2 = dict:Quota Usage::noenforcing:proxy::sqlquota } dict { sqlquota
2019 Oct 07
3
Sieve redirect is broken in 2.3.7.2 - signal 11
Hi Stephan, Here it is: Program received signal SIGSEGV, Segmentation fault. p_strdup (pool=pool at entry=0x55555579bc20, str=0x6d65642e6c69616d <error: Cannot access memory at address 0x6d65642e6c69616d>) at strfuncs.c:51 51????? strfuncs.c: No such file or directory. (gdb) bt full #0? p_strdup (pool=pool at entry=0x55555579bc20, str=0x6d65642e6c69616d <error: Cannot access memory at
2019 Oct 07
1
Sieve redirect is broken in 2.3.7.2 - signal 11
That's a good thought. I'm building from the alpine repo and I indeed see pigeonhole is 0.5.5 https://git.alpinelinux.org/aports/tree/main/dovecot/APKBUILD?h=3.8-stable Let me build it with 0.5.7.2 Regards, ? Laszlo On 2019. 10. 07. 10:45, Stephan Bosch via dovecot wrote: > > > On 07/10/2019 10:17, Demonhost via dovecot wrote: >> Hi Stephan, >> >> Here it
2017 Sep 29
1
2.2.32 'doveadm replicator replicate -f' segfault
Very minor bug; not specifying the user mask with 'doveadm replicator replicate -f' causes a segfault: server:~# doveadm replicator replicate -f Segmentation fault server:~# doveadm replicator replicate -f '*' 123 users updated server:~# gdb /usr/bin/doveadm core.2418 GNU gdb (GDB) Red Hat Enterprise Linux 7.6.1-100.el7 This GDB was configured as
2018 Dec 19
2
Fatal: master: service(indexer-worker): child 493 killed with signal 11 (core dumped)
Hope this helps. There are no debugging symbols available for dovecot in the alpine repository. (gdb) bt full #0? a_crash () at ./arch/x86_64/atomic_arch.h:108 No locals. #1? free (p=0x55d949250660) at src/malloc/malloc.c:467 ??????? extra = 65 ??????? base = 0x55d94925060f "" ??????? len = 94391723427649 ??????? self = 0x55d949250650 ??????? next = <optimized out> ???????
2004 Aug 06
3
icecast-2.0-kh34
In message "[icecast] icecast-2.0-kh34" on 24.04.2004, Karl Heyes <karl@xiph.org> writes: KH> On Fri, 2004-04-23 at 09:58, Wild Karl-Heinz wrote: KH> ... [snip] >> ... but it coredumps every 10 secs. KH> There is a segv bug in kh34 triggered when clients move from one mount KH> to another. That should be fixed now in kh35. the result with gdb .... (gdb)
2020 Jun 19
2
SegFault when using TunnelServer=yes
Hi all, I have a network with about ~800. The network is a mix of tinc 1.0 and 1.1 nodes. It is gradually expanding for several years now. The problem is that at some point it seams the daemon can not handle the processing of the new connection and the edges. There are 3 major nodes in the system and every other node initially makes connection to one of them. Now after a lot of debugging
2007 May 19
2
Crash in blur.c (SIGSEGV)
Hello. System info: OS: FC6 x86_64 video: Nvidia GF 7900 GT, driver version 97.55 compiz version: git config backend used: ccp I can reproduce this crash every time on my system. When blur plugin is active and I try to access any right-click menu, regular application menu or drop-down list, compiz crashes. Backtrace produced by crashhandler plugin shows crash in blur.c in function
2020 Jul 27
3
SegFault when using TunnelServer=yes
Hi, thank for getting back. I'll answer the questions, but I've already gave up on tinc and switch to zerotier-one. On 2020-07-27 5:10 p.m., borg at uu3.net wrote: > Hi. I have few questions out of curiosity.. Cant help for now with > your problem... > > What version is crashing? 1.1 or 1.0 ? 1.1 is crashing > > How your network is segmented..? > I use tinc myself
2019 Sep 16
1
Dovecot 2.3.7.2 lucene segfault and coredump
Hi Stephan, thanks for your replay. This is bt full output: (gdb) bt full #0 clear (this=0x0) at /usr/src/debug/clucene-core-2.3.3.4/src/core/CLucene/util/VoidMap.h:123 No locals. #1 lucene::search::FieldSortedHitQueue::_shutdown () at /usr/src/debug/clucene-core-2.3.3.4/src/core/CLucene/search/FieldSortedHitQueue.cpp:60 No locals. #2 0x00007f7dc5eabe96 in _lucene_shutdown () at
2018 Dec 08
2
doveadm batch crash
Hi I'm having an issue with doveadm batch - the following command always crashes: doveadm batch : mailbox status all inbox Whereas the following work as expected: doveadm mailbox status all inbox doveadm batch : mailbox list Any assistance gratefully received. Debug output from the crash: Debug: Loading modules from directory: /usr/lib64/dovecot Debug: Module loaded:
2007 Sep 11
2
Asterisk 1.4.11, res_features.so, SegFault
Hi All, I have a really strange issue occuring where if I run "show dialplan" or "dialplan show" or "dialplan show parkedcalls", then asterisk dumps core. It only appears to happen with contexts that are created within res_features. I am able to display all my other dialplans, but, every time I try to just do a normal "dialplan show" asterisk core dumps
2016 Oct 06
2
[imap-login] SSL related crashes using the latest 2.2.25
I'm running Dovecot as proxy in front of some IMAP/POP3 Dovecot & Courier-IMAP servers and in the last couple of days I've been seeing a lot of imap-login crashes (signal 11) on both 2.2.18 and 2.2.25, all SSL related. The following backtraces are taken running 2.2.25, built from source on a test system similar to the live proxy servers. OS: CentOS 6.8 64bit Packages:
2019 Oct 06
2
Sieve redirect is broken in 2.3.7.2 - signal 11
I think alpine does not have that https://pkgs.alpinelinux.org/packages?name=dovecot*&branch=v3.8&repo=main&arch=x86_64 https://git.alpinelinux.org/aports/tree/main/dovecot/APKBUILD?h=3.8-stable What else would help? strace? On 2019. 10. 06. 23:05, Stephan Bosch via dovecot wrote: > > > On 06/10/2019 23:03, Demonhost wrote: >> Hi Stephan, >> >> I managed
2007 Jan 03
3
Asterisk Core Dump in app_queue - Anyone seen?
Anyone seen this? It ocurred on a 'reload app_queue.so' command. Asterisk version is 1.2.9.1. Tried again, but it was not immediately reproducable. Doug. (gdb) bt #0 reload_queues () at app_queue.c:3339 #1 0xb778a7a8 in reload () at app_queue.c:4012 #2 0x0805bb44 in ast_module_reload (name=0x8137cc7 "app_queue.so") at loader.c:257 #3 0x08092b3f in handle_reload (fd=33,
2019 Mar 26
2
v2.2.27 Panic: file rfc822-parser.h: line 23 (rfc822_parser_deinit): assertion failed: (ctx->data <= ctx->end)
After some investigation, it turns out it is non trivial to install dovecot-dbg on debian jessie. Sorry I can't investigate further. Jason Aki Tuomi wrote on 25/3/19 6:12 pm: > Can you install dovecot-dbg and try gdb again? > > Aki > > On 25.3.2019 3.20, Jason Lewis via dovecot wrote: >> Hi, >> >> I've been having an issue with the indexer giving me
2019 Sep 16
2
Dovecot 2.3.7.2 lucene segfault and coredump
Hi, I have other information. I analyzed the core dump: (gdb) core 22534 [New LWP 22534] Reading symbols from /usr/libexec/dovecot/doveadm-server...Reading symbols from /usr/lib/debug/usr/libexec/dovecot/doveadm-server.debug...done. done. [Thread debugging using libthread_db enabled] Using host libthread_db library "/lib64/libthread_db.so.1". Core was generated by