Displaying 20 results from an estimated 1000 matches similar to: "Panic when runing "doveadm mailbox status ... vsize""
2009 Feb 04
4
pop3_lock_session question
Hello, I'm preparing to convert from qpopper + UW-IMAP to dovecot.
So far testing has gone very well. One problem we haven't figured
out is that long-running POP sessions keep the mailbox locked, so that
the MDA times out while trying to deliver. We're using "maildrop" as
our MDA if that matters.
We don't see this issue for dovecot IMAP sessions. During long-running
2009 Apr 24
3
1 Dovecot proxy to 2 real IMAP servers
Hello all.
I have 2 Dovecot IMAP servers with different mailboxes. What serves
different email domains.
I want to add one Dovecot Proxy server and make him understand based on
user domain - were it need to transfer a connection.
I use LDAP auth based on "mail" attribute.
It is possible?
P.S. Proxy documentation at dovecot.org not help at all and seems to not
complited. :-(
--
2007 Nov 24
7
[Bug 1393] New: patch modifies gnome-ssh-askpass to optionally use one-time password
https://bugzilla.mindrot.org/show_bug.cgi?id=1393
Summary: patch modifies gnome-ssh-askpass to optionally use one-
time password
Classification: Unclassified
Product: Portable OpenSSH
Version: 4.7p1
Platform: All
URL: http://www.swcp.com/~pgsery
OS/Version: Linux
Status: NEW
Keywords:
2008 Jun 13
3
[Bug 983] Required authentication
https://bugzilla.mindrot.org/show_bug.cgi?id=983
Damien Miller <djm at mindrot.org> changed:
What |Removed |Added
----------------------------------------------------------------------------
CC| |pgsery at swcp.com
--- Comment #18 from Damien Miller <djm at mindrot.org> 2008-06-13 13:56:12 ---
*** Bug
2009 Feb 04
3
auth_debug_passwords
Hi. I'm new to Dovecot and about to start using it in production. In
the config file, I set the option, auth_debug_passwords, to yes. I do
not see any failed passwords logged, however. It did cause more verbose
authentication logging, but failed passwords are still hidden.
I have also set these options to yes, because I thought they might be
required for auth_debug_passwords to work:
2010 Jan 07
4
[Bug 983] Required authentication
https://bugzilla.mindrot.org/show_bug.cgi?id=983
--- Comment #26 from Paul Sery <pgsery at swcp.com> 2010-01-08 08:09:14 EST ---
The configuration below is incorrect. When using protocol 2, it should
read:
...
RequiredAuthentications2 password
Also, there's no need to specify publickey in conjunction with other
authentication methods because it will always be tried first (as
specified
2010 Nov 12
1
[Bug 983] Required authentication
https://bugzilla.mindrot.org/show_bug.cgi?id=983
Paul Sery <pgsery at swcp.com> changed:
What |Removed |Added
----------------------------------------------------------------------------
Attachment #1768|0 |1
is obsolete| |
--- Comment #30 from Paul Sery <pgsery at swcp.com> 2010-11-12
2024 Apr 08
2
Exceptional slowness with read.csv
? Sun, 7 Apr 2024 23:47:52 -0600
Dave Dixon <ddixon at swcp.com> ?????:
> > second_records <- read.csv(file_name, skip = 2459465, nrows = 5)
It may or may not be important that read.csv defaults to header =
TRUE. Having skipped 2459465 lines, it may attempt to parse the next
one as a header, so the second call read.csv() should probably include
header = FALSE.
Bert's advice
2006 Nov 29
2
--max-vsize option
The R memory docs say that the --*-vsize option takes an integer
argument and then 'G', 'M', 'K', or 'k'. When I start R using
R --max-vsize=10G
I receive the warning:
WARNING: --max-vsize=10G=10'M': too large and ignored
The system that I'm working on is a 64-bit Sun server with 40G of
memory. What is the correct syntax for this command?
Daniel
2011 Jul 21
1
--max-vsize
Hi,
In both R 2.13 and the SVN trunk, I observe odd behaviour with the
--max-vsize command-line argument:
1. passing a largeish value (about 260M or greater) makes mem.limits()
report NA for the vsize limit; gc() continues to report a value...
2. ...but that value (and the actual limit) is wrong by a factor of 8.
I attach a patch for issue 2, lightly tested. I believe that fixing
issue 1
2024 Apr 08
2
Exceptional slowness with read.csv
I solved the mystery, but not the problem. The problem is that there's
an unclosed quote somewhere in those 5 additional records I'm trying to
access. So read.csv is reading million-character fields. It's slow at
that. That mystery solved.
However, the the problem persists: how to fix what is obvious to the
naked eye - a quote not adjacent to a comma - but that read.csv can't
2024 Apr 08
2
Exceptional slowness with read.csv
Hi Dave,
That's rather frustrating. I've found vroom (from the package vroom) to be
helpful with large files like this.
Does the following give you any better luck?
vroom(file_name, delim = ",", skip = 2459465, n_max = 5)
Of course, when you know you've got errors & the files are big like that it
can take a bit of work resolving things. The command line tools awk
1999 May 15
2
vsize and nsize
I am running R version ??? under Redhat 5.2. It seems as though the
--nsize object has no effct on the size of the allocated Ncells as
determined using gc(). Yes, I have that much data....
That is if I envoke R with
R --vsize 100 --nsize 5000000
then type
gc()
I get
free total
Ncells 92202 200000
Vcells 12928414 13107200
Thanks
Tony Long
Ecology and Evolutionary Biology
Steinhaus
2000 Mar 13
1
check does not accept --vsize option (PR#481)
Full_Name: Markus Neteler
Version: 1.0.0
OS: Linux 2.2.10/i686
Submission from: (NULL) (130.75.72.37)
Hi,
I wanted to "check" the R.GRASS GIS interface from Roger Bivand:
http://www.geog.uni-hannover.de/grass/statsgrasslist.html
using
R CMD check --vsize=10M GRASS
but:
[error message shortened]
> G <- gmeta()
Error: heap memory (6144 Kb) exhausted [needed 1024 Kb more]
1999 Apr 12
3
--nsize and --vsize
Martin M has suggested I widen this discussion to R-devel, and
> I agree that we should increase them,
> but I'm not sure at all about the amount.
>
> The default could even depend on the architecture (via "./configure")..
Views, please.
------------- Begin Forwarded Message -------------
Is is not time we increased the defaults a bit? As the base gets bigger
I hit
2010 Oct 25
1
Maildir filename extensions S=<size> and W=<vsize>
I'd like to add S=<size> and W=<vsize> to the filenames in my Maildir
storage for newly delivered mails. But I cannot find where I can make
Dovecot's LDA add those. Anybody?
http://wiki2.dovecot.org/MailboxFormat/Maildir says:
A maildir filename with those fields would look something like: 1035478339.27041_118.foo.org,S=1000,W=1030:2,S
which is exactly what I want, but it
2024 Apr 08
1
Exceptional slowness with read.csv
No idea, but have you tried using ?scan to read those next 5 rows? It might
give you a better idea of the pathologies that are causing problems. For
example, an unmatched quote might result in some huge number of characters
trying to be read into a single element of a character variable. As your
previous respondent said, resolving such problems can be a challenge.
Cheers,
Bert
On Mon, Apr 8,
2008 May 13
2
how add size (, S=size) and virtual size (, W=vsize) fields to a maildir filename ?
Hi to all,
I'm a sys admin in a college, and we're using Dovecot as IMAP/POP3
server and delivery; we're also patch dovecot to add managesieve
capability. I've a question:
I've read on dovecot's wiki, that is possible improve the performance on
maildir files by adding ,S=<size>,W=<vsize> fields on filename. How is
possible to do that ? I've search on all
2019 Aug 11
2
doveadm / vsize
Hi,
i want to keep track of user mailbox size,
I'm using
doveadm mailbox status -u <user> vsize INBOX
for a given mailbox this is under reporting ( by a lot)
is this recursive? what shoud i be using to get an accurate disk usage metric?
thanks
Paul.
--
-------------------------
Paul Macdonald
IFDNRG Ltd
Web and video hosting
-------------------------
t: 0131 5548070
m:
2004 Jul 20
1
--max-vsize and --max-nsize linux?
Hi,
somtimes i have trivial recodings like this:
> dim(tt)
[1] 252382 98
system.time(for(i in 2:length(tt)){
tt[,i][is.na(tt[,i])] <- 0
})
...and a win2000(XP2000+,1GB) machine makes it in several minutes, but
my linux notebook (XP2.6GHZ,512MB) don't get success after some hours.
I recognize that the cpu load is most time relative small, but the hardisk