similar to: RC7: BUG! and patch [Was: Re: rc7 bug? [Was: deliver LDA and INBOX location] (fwd)] (fwd)

Displaying 20 results from an estimated 5000 matches similar to: "RC7: BUG! and patch [Was: Re: rc7 bug? [Was: deliver LDA and INBOX location] (fwd)] (fwd)"

2006 Sep 05
2
rc7 bug? [Was: deliver LDA and INBOX location] (fwd)
Anyone had any thoughts on the item below? If the problem is with my config, I'd like to be guided towards how I might resolve it. If it is a bug in rc7, it would be good to fix it, and I'd be happy to beta-test. -- : David Lee I.T. Service : : Senior Systems Programmer Computer Centre : :
2006 Aug 31
1
deliver LDA and INBOX location
(OS: Fedora Core 5; dovecot: 1.0 rc7) On a typical UNIX-like OS, the INBOXes are in "/var/spool/mail/" using the user identifier: so user 'fred' has INBOX "/var/spool/mail/fred". We have a well-established different convention which subdivides this, based on the last two digits of the uid: "/var/spool/mail/12/fred" (for fred's uid as something ending
2006 Oct 03
2
dovecot, procmail and deliver
(Using dovecot 1.0 RC7 on Fedora Core 5) <scene set> Hitherto we have used UW-IMAP on a "farm" of Linux machines mounting NFS from a NetApp. (The UW-IMAP author doesn't like use of NFS, but with careful use of NFS mount arguments ('noac,actimeo=0' etc.) and trying to ensure that all activity for a given user takes place within one machine in the farm, we seem to
2006 Aug 21
4
RC7: its issues or mine?
Background: I'm new to dovecot (although with many years Washington IMAP behind me). We're considering migrating from Washington IMAP to dovecot on the main service here, and have just started trying dovecot, using RC7. Washington, IMAP has the usual(-ish) "/var/spool/mail" shared area for the INBOX (trad. UNIX "From " format); a user's folders default to being
2006 Oct 11
2
1.0rc8 status report
A quick status report on how 1.0rc8 behaved in service for a few hours with several hundred simultaneous users, at a site very new to dovecot. Oh, and a question at the end. Summary: Reasonable for a first shot but one significant problem, requiring backing off. Background: We have a long-established UW-IMAP service for a user population of about 20,000 based on a few Linux (Redhat) machines
2006 Oct 05
0
bug with rsh/ssh connections
(Using rc7 code) There seems to be a bug in rsh/ssh-style connections. For reference, on a normal imap connection (port 143, 993, etc.) things are OK. The imap session ends (from a client-end command "a logout") thus: --- a logout * BYE Logging out a OK Logout completed. Connection closed by foreign host. unix-prompt% --- But when the connection had been established using rsh or ssh
2006 Aug 17
17
1.0 RC7 released
http://dovecot.org/releases/dovecot-1.0.rc7.tar.gz http://dovecot.org/releases/dovecot-1.0.rc7.tar.gz.sig Can everyone now agree that there are no more hangs? :) * Require that Dovecot master process's version number matches the child process's, unless version_ignore=yes. Usually it's an accidental installation problem if the version numbers don't match. * Maildir: Create
2006 Aug 17
17
1.0 RC7 released
http://dovecot.org/releases/dovecot-1.0.rc7.tar.gz http://dovecot.org/releases/dovecot-1.0.rc7.tar.gz.sig Can everyone now agree that there are no more hangs? :) * Require that Dovecot master process's version number matches the child process's, unless version_ignore=yes. Usually it's an accidental installation problem if the version numbers don't match. * Maildir: Create
2007 Feb 09
5
resilience suggestion
On the whole we are pleased with our trials of dovecot to replace UW-IMAP. But (ah!) we have hit one particular problem, in which we think dovecot could probably benefit from a resilience improvement. We're running dovecot on Fedora Core 5 (FC5), with passwd map details supplied by NIS. We have found that "nscd" sometimes thinks that a username is invalid, even though it is valid.
2006 Oct 07
0
No subject
valid incoming email. It is rare, but it does occasionally happen on large, busy systems. Clearly it is fundamentally an "nscd" bug. But that bug is nevertheless out there, in the wild, on such systems, potentially affecting dovecot's delivery of valid user email. We have had a source code hack since October (in "deliver.c", simply replacing a "return ret"
2006 Oct 30
2
Upgrade from RC7 to RC10 didn't go too well...
Hi, On the weekend I tried to upgrade from RC7 to RC10. Clearly, I have to change some things before I can do this. What does it take to get to RC10? As is, my mail client got an error trying to list messages in /INBOX (via IMAP) and the mail.err log shows the following messages: Oct 29 10:27:32 siona dovecot: IMAP(archangel): open(/var/mail/archangel/inbox, O_CREAT) failed: Not a
2006 Oct 10
1
RC8 failing...
I just tried upgrading from RC7 to RC8 this morning, and I'm seeing an issue I've never seen before. On my first POP3 login, all is fine, but any subsequent logins seem to fail with the message: dovecot: Oct 10 11:04:31 Error: Maximum number of mail processes exceeded In the dovecot log file. I'm also oddly seeing the message: Oct 10 11:01:05 popbkup pop3-login: [ID 799321
2006 Oct 12
2
1.0rc8: another problem? Possibly 64-bit index?
Yesterday I gave a status report about 10.rc8 in production, which mentioned a problem about "Login process died too early..." Timo suggested a patch for logging error messages. I've applied this. Others suggested increasing "login_max_processes_count". That was already way above our likely maximum, but I've doubled it anyway. Today, I've just repeated the
2006 Oct 20
4
1.0.rc10 status report
(Background: Relatively new to dovecot; looking to do transparent replacement of long-established UW-IMAP on cluster of Linux boxes which NFS-mount a shared "/var/spool/mail".) With rc8, where I had already increased "login_max_processes_count" from default 128 to 1024, we had still hit the issue of too many logins crashing dovecot, so that trial had only lasted a couple of
2009 Mar 05
1
[PATCH] OCFS2: Pagecache usage optimization on OCFS2
Hi. I introduced "is_partially_uptodate" aops for OCFS2. A page can have multiple buffers and even if a page is not uptodate, some buffers can be uptodate on pagesize != blocksize environment. This aops checks that all buffers which correspond to a part of a file that we want to read are uptodate. If so, we do not have to issue actual read IO to HDD even if a page is not uptodate
2001 Apr 11
0
replicating lists (fwd) (PR#907)
Filed as a bug, as suggested by Brian R., Jonathan. Jonathan Rougier Science Laboratories Department of Mathematical Sciences South Road University of Durham Durham DH1 3LE tel: +44 (0)191 374 2361, fax: +44 (0)191 374 7388 http://www.maths.dur.ac.uk/stats/people/jcr/jcr.html ---------- Forwarded message ---------- Date: Wed, 11 Apr 2001 09:15:26 +0100
2002 Nov 22
4
Small change to plot.xy
Hi everyone, Is there any reason why we should not automatically coerce a factor supplied as an argument to col in a plotting function? The following modification (to R-1.6.1) seems pretty harmless > plot.xy function (xy, type, pch = 1, lty = "solid", col = par("fg"), bg = NA, cex = 1, ...) { if (is.factor(col)) col <- codes(col)
2013 Apr 19
7
Re: [BUG REPORT] Kernel panic on 3.9.0-rc7-4-gbb33db7
(cc''ing btrfs people) On Fri, Apr 19, 2013 at 11:33:20AM +0800, Wanlong Gao wrote: > RIP: 0010:[<ffffffff812484d3>] [<ffffffff812484d3>] ftrace_raw_event_block_bio_complete+0x73/0xf0 ... > [<ffffffff811b6c10>] bio_endio+0x80/0x90 > [<ffffffffa0790d26>] btrfs_end_bio+0xf6/0x190 [btrfs] > [<ffffffff811b6bcd>] bio_endio+0x3d/0x90 >
2001 Apr 11
5
replicating lists
Hi Everyone, At the moment it is not possible to replicate complex lists, but only simple ones: > rep(list(fred = 1:10), 10) # works fine > rep(list(fred = 1:10, happy = "squash"), 10) Error in rep(list(fred = 1:10, happy = "squash"), 10) : Unimplemented feature in rep There is nothing in ?rep that suggests that the latter should not work, and I think it would
2006 Oct 13
1
Segfault in in rc7 when index does not exists
This is the trace (gdb) run Starting program: /root/tmp/dovecot-1.0.rc7/src/imap/imap x select inbox Program received signal SIGSEGV, Segmentation fault. mail_index_write_base_header (index=0x80e2a28, hdr=0xaff66028) at mail-index.c:1313 1313 memcpy(index->map->mmap_base, hdr, hdr_size); (gdb) bt #0 mail_index_write_base_header (index=0x80e2a28, hdr=0xaff66028) at