similar to: nfs_flush_fcntl

Displaying 20 results from an estimated 1000 matches similar to: "nfs_flush_fcntl"

2008 Nov 10
1
nfs_flush_fcntl failed: No locks available
Hi there, I've been seeing this error in our logs quite frequently on our nfs storage (v3): 2008-11-10T13:24:26+00:00 mail8 dovecot: IMAP(XXX at YYY.com): nfs_flush_fcntl: fcntl(/var/spool/mail/XXX/Maildir/dovecot.index.cache, F_RDLCK) failed: No locks available Which is because we don't run lockd on our servers. Why is dovecot trying to use fcntl() ? I explicitly set it to use dotlocks
2017 Sep 16
2
dsync panic in mbox_lock
Hello I am trying to setup replication with dovecot-2.2.29.1, and for some users, I get a reproductible panic: # doveadm -v -o plugin/mail_replica=remoteprefix:root at mail2.example.net sync -d -u jdoe dsync-local(jdoe): Panic: file mbox-lock.c: line 799 (mbox_lock): assertion failed: (lock_type == F_RDLCK || mbox->mbox_lock_type != F_RDLCK) Abort User is not overquota, and filesystem
2012 Jan 02
1
error bad file number with compressed mbox files
Hello, can dsync convert from compressed mbox to compressed mdbox format? When I use compressed mbox files, either with gzip or with bzip2, I can read the mails as usual, but I find the following errors in dovecots log file: imap(userxy): Error: nfs_flush_fcntl: fcntl(/home/hrz/userxy/Mail/mymbox.gz, F_RDLCK) failed: Bad file number imap(userxy): Error: nfs_flush_fcntl:
2008 Aug 07
3
GlusterFS
Hi everybody, We use a Dovecot setup with GlusterFS. Dovecot 1.1.2 and GlusterFS 1.3.9. I enabled the following options (I don't have posix-locks translators): lock_method = dotlock dotlock_use_excl = no mmap_disable = yes mail_nfs_index = yes mail_nfs_storage = yes I receive the following error message. Aug 7 09:38:51 mta2 dovecot: POP3(somebody at domain.tld): nfs_flush_fcntl:
2014 Dec 29
2
smbd's using up 100% of all cpu's and load avg slowly going up
Hi again, Sorry, I forgot to include the fd used... 13 -> /var/lib/samba/private/sam.ldb.d/DC=EXAMPLE,DC=COM.ldb 16 -> /var/lib/samba/private/sam.ldb.d/metadata.tdb the others refers to /var/run/samba/winbindd/pipe Regards, Bruno Andrade. On 12/29/2014 11:56 AM, Bruno Andrade wrote: > Hi, > > Here is some output of "strace -ff -p" to the process using ~100% CPU >
2014 Dec 29
2
smbd's using up 100% of all cpu's and load avg slowly going up
Hi, Yes, I already thought about it. I have the sernet repo installed.. Is possible to yum update and jump from 4.1.6 to 4.1.14 directly, without problems or should I follow some kind of procedure to avoid future problems? Regards, Bruno Andrade. On 12/29/2014 01:46 PM, Rowland Penny wrote: > On 29/12/14 12:02, Bruno Andrade wrote: >> Hi again, >> >> Sorry, I forgot to
2017 May 02
2
samba process use 100% cpu
Hi! I need some help. We use samba4 as AD, and now when clients connect to server, samba process stuck at 100% cpu. samba Version: 4.3.4 Release: 13.el6 top: 3777 root 20 0 131m 46m 28m R 99.7 0.3 219:20.53 /usr/local/samba4//sbin/samba -D 24541 csertam 20 0 49260 11m 9048 S 25.1 0.1 0:01.56 smbd -D 7080 squid 20 0 926m 908m 6428 S 9.9 6.2 11:43.50
2017 May 09
2
samba process use 100% cpu
Did you install it from binary packages ? compiled from source ? Post your smb.conf here. ----- Original Message ----- From: "samba" <samba at lists.samba.org> To: "samba" <samba at lists.samba.org> Sent: Tuesday, May 9, 2017 1:01:10 PM Subject: Re: [Samba] samba process use 100% cpu Anybody? 2017. 05. 02. 10:48 keltezéssel, Papp Bence via samba írta: > Hi!
2018 Mar 16
3
performance problem on bridgehead DC
Hi everybody ! I am encountering performance problem on my bridgehead DC. I have 19 DC (Debian Stretch / Samba 4.6.7 from Tranquil.it repo) and they all synchronized on a main bridgehead DC. This performance problem first appears when the bridgehead DC was on Debian Jessie and kernel was updated with meltdown/spectre patch from Debian (3.16.51-3+deb8u1) I added "nopti" option to grub
2008 Oct 02
2
some errors on 1.1.3
Hi, we recently upgraded dovecot from 1.0.5 to 1.1.3+latest patches. After the upgrade I saw some errors in the logs and I'm not sure whether they're a cause for concern. Our setup is 2 servers running openSUSE 11.0 and a NetApp Filer as storage backend. Following is the config: protocols: imap imaps pop3 pop3s login_dir: /var/run/dovecot/login login_executable(default):
2015 Oct 07
3
Hyper-V Gen 2 waiting for ldlinux.e64
> On Wed, Oct 07, 2015 at 05:06:41PM -0500, Clements, James via Syslinux > wrote: > > > On Wed, Oct 07, 2015 at 06:01:30PM +0200, Geert Stappers via Syslinux > wrote: > > > is for BIOS, not for EFI. So that suggestion can be scraped. > > > > > > Which makes > > > * text which is on the screen of the netbooting (virtual) machine > > >
2016 Apr 21
2
Setting up replication?
On 4/21/2016 12:01, Timo Sirainen wrote: > On 21 Apr 2016, at 19:51, Knute Johnson <dovecot at knutejohnson.com> wrote: >> On 4/21/2016 04:33, Aropalo Tommi wrote: >>>> I'm trying to set up replication between two servers and I've hit a >>>> snag. I have two users and am using mbox files. If anybody has a >>>> sample config file for a
2008 Aug 25
2
Dovecot 1.1.2
Hello, I upgraded dovecot to 1.1.2 this weekend and i get those errors: grep dovecot /var/log/net/maillog |grep Panic | cut -d " " -f 11-100| sort | uniq -c | sort -rn 7547 file index-sync.c: line 39: assertion failed: (seq_range_exists(&ibox->recent_flags, uid)) 89 file index-mail.c: line 1091: assertion failed: (! mail->data.destroying_stream) 18 file
2013 May 31
1
dsync assert failure in 2.2.2
Hi, with 2.2.2 and today's hg, dsync crashes with dsync(root): Panic: file ../../../../../src/lib-storage/index/mbox/mbox-lock.c: line 797 (mbox_lock): assertion failed: (lock_type == F_RDLCK || mbox->mbox_lock_type != F_RDLCK) when I run USER=root 2.2-hg/bin/dsync -c etc/dovecot.conf -f -o mail_location=mbox:/tmp/imap/fwadmin.tmp:INBOX=/tmp/imap/fwadmin.tmp/INBOX mirror
2018 Dec 12
4
vfs_fruit causes delay in listing directories for Windows clients
Listing directories with many files (10000+) from a Windows client is nociceably slower when vfs_fruit is enabled on the samba server compared to the same setup without vfs_fruit. On my setup it's roughly 2.5 times slower. To me it looks like this is caused by the getxattr call which is only present with vfs_fruit activated and introduces an additional delay of ~ 0.00033 s per listed
2020 Oct 11
2
Preparing for replication: dsync-local(testaccount): Panic: file mbox-lock.c
I am preparing a bit for setting up replication. However when I manually try to dsync an account, the first time I execute this[1] command it seems to be ok. The 2nd time I am getting this error[2]. If I add -1 (one way syncing) the error disappears. Does this mean I will have problems with setting up replication between these two servers? [1] [@ ~]# doveadm sync -n inbox -u testaccount
2017 Sep 26
3
dsync: assert panic on mbox // no INBOX sync on hybrid mbox/maildir
Hi, This (hopefully!) might be a case of user error, as I'm new to dovecot... Alternatively it might be some confusion in an index, since I've been mucking about trying different things in my configuration files and had frequent crashes. I'm trying to sync 2 (LAN) hosts, ideally to have full automatic replication happening. They both have hybrid mbox INBOX and maildir++ folders,
2005 Sep 18
1
Can't connect to auth server at default: Connection refused
We've being running with Dovecot 1.0-stable since Tuesday, and on the whole it's gone very well (meaning nobody's noticed!). My best estimate is we're using half the disk accesses and 2/3 the CPU of UW-IMAP/Qpopper. We've quite a mixture of error messages in the logs, the most worrying of which is that several times we've had a few seconds of "xxx-login: Can't
2017 Sep 16
0
dsync panic in mbox_lock
> On September 16, 2017 at 5:28 AM Emmanuel Dreyfus <manu at netbsd.org> wrote: > > > Hello > > I am trying to setup replication with dovecot-2.2.29.1, and for some > users, I get a reproductible panic: > > # doveadm -v -o plugin/mail_replica=remoteprefix:root at mail2.example.net > sync -d -u jdoe > dsync-local(jdoe): Panic: file mbox-lock.c: line 799
2004 Dec 01
2
1.0-test CVS HEAD index problems
Running current CVS HEAD I get: dovecot: Dec 01 13:27:38 Info: Dovecot v1.0-test52 starting up dovecot: Dec 01 13:33:34 Info: imap-login: Login: jtl [69.162.177.245] dovecot: Dec 01 13:41:07 Error: IMAP(jtl): UIDVALIDITY changed (1100294646 -> 110192643 5) in mbox file /var/spool/mail/j/t/jtl dovecot: Dec 01 13:41:07 Error: IMAP(jtl): UIDVALIDITY changed (1100294646 -> 110192643 5) in