search for: seling

Displaying 20 results from an estimated 38 matches for "seling".

Did you mean: seeing
2019 Jan 14
2
mdbox + zlib performing less than just mdbox
I have test environment to determine what would be best settings. I have been told that enabling zlib compression would be good to save iops on storage. But doing the test now, I get worse results. [@test2 ~]# pr -m -t mail04-mdbox-vdb-append-64kb-6.log mail04-mdbox-vdb-append-64kb-8.log |less Logi Sele Appe Logi Sele Appe 100% 100% 100% 100% 100%
2019 Jan 14
0
mdbox + zlib performing less than just mdbox
mdbox format is a cross between mbox and sdbox. The idea is that it keeps up to mdbox_rotate_size sized mbox files which contain mails. Using zlib here will not help, because zlib is not applied to the full mdbox file, but individual mails within. Also not sure why you think that adding compression would make things faster? =) I am not fully sure how cephfs works but you might get better
2007 Jan 25
4
Benchmarking
I wasted too much time today doing this, but they do show a few interesting things: 1) mmap_disable=yes is faster than mmap_disable=no. I didn't expect this. I'm not yet sure why this is, but possibly because then it updates the mail index by reading the changes from dovecot.index.log, instead of reopening and re-mmaping dovecot.index. At least that's the biggest difference between
2007 Nov 30
3
Zimbra benchmarking
Now that I have a working kvm setup, I thought I'd finally try how Zimbra works. This is mainly some microbenchmarking, so it may not have much to do with actual performance in real life. Setup: - 1GB memory given to kvm (from host's 2GB) - Intel Core 2 6600 (kvm uses only one CPU) - CentOS 5 - 15GB qcow2 image on XFS filesystem - Zimbra 5.0 RC2 RHEL5 x86_64 - Dovecot latest hg,
2019 Jan 14
0
What does the mdbox_rotate_size influence?
I wondered why mdbox_rotate_size is 2MB by default. I thought if I increased it to 16MB, maybe there would be less disk io, but I dont see any difference. Furthermore I read in some thread that increasing it from 2MB, could give problems when deleting messages, can someone explain this? Does anyone have this on a cephfs mount, does it make sense to set this to 4MB? [@test2 ~]# pr -m -t
2007 May 12
3
dbmail benchmarking
I thought I'd try benchmarking with dbmail (v2.2.4) to see how much slower a SQL backend could actually be. Skip to bottom for the conclusions. Originally I ran the tests with the databases being in XFS filesystem. MySQL's performance was horrible. It went 3-7x faster with ext3. MySQL 5.0.30 backend (innodb): ./imaptest clients=1 - append=100 seed=1 secs=30 msgs=1000000 logout=0
2009 Mar 30
2
dbox benchmarks
http://hg.dovecot.org/dovecot-dbox-redesign/ Looks like multi-dbox scales pretty nicely. Even after 100k messages the peak saved msgs/sec is the same as the initial saved msgs/sec, even if the average slows down somewhat. I tested this by first deleting mailbox, then running "imaptest" for a second to get saving to start writing several fields to dovecot.index.cache file. Then ran
2007 Mar 29
1
uiddir mailbox format with benchmarks
I cleaned up the lib-storage code a bit so that it's easier to write support for new mailbox formats. It could still use a bit of cleaning up. Anyway, I thought I'd try how easy it would be to implement Cyrus-like mail store which would consist of only Dovecot's index files and "<uid>." message files. Flags and everything else is kept only in Dovecot's index files
2011 May 05
5
Dovecot imaptest on RHEL4/GFS1, RHEL6/GFS2, NFS and local storage results
We have done some benchmarking tests using dovecot 2.0.12 to find the best shared filesystem for hosting many users, here I share with you the results, notice the bad perfomance of all the shared filesystems against the local storage. Is there any specific optimization/tunning on dovecot for use GFS2 on rhel6??, we have configured the director to make the user mailbox persistent in a node, we will
2007 Mar 14
2
Benchmarking CVS HEAD vs. v1.0
Some new features in CVS HEAD are: - v1.0 saves most of the data to dovecot.index.cache only when the client is FETCHing the messages. With mboxes the message headers are saved while parsing the mbox, which is almost the same thing. With CVS HEAD the dovecot.index.cache file is updated already when the new messages are being saved (with deliver or IMAP APPEND). This avoids reading the message
2007 Jan 09
3
min() return factor class values
Um texto embutido e sem conjunto de caracteres especificado associado... Nome: n?o dispon?vel Url: https://stat.ethz.ch/pipermail/r-help/attachments/20070109/d9ceecae/attachment.pl
2017 Nov 01
0
IMAP Benchmarking, active/active cluster with GFS2
Hi All, We're on migration stage to a new 2-node active/active cluster with shared iSCSI storage,? everything is ready, but we are in the process of I/O testing, before we going to in production. Cluster info: 2 nodes with CentOS 7.4 balancer ( keepalived + ldirectord ), 2 dovecot directors, 2 dovecot IMAP backends, 2 Postfix SMTP, gfs2 ( corosync + dlm ), LMTP between Postfix and Dovecot
2007 Jul 12
3
v1.1 status and benchmarks
v1.1 plans have changed a bit. I'll release v1.1.alpha1 soon and hope to have a stable v1.1 in a month or two. The rest of the features that didn't make it into v1.1 will go to v1.2. I'll write more about this when v1.1 alpha is released. I also did a bit of benchmarking. v1.1's performance improvements are looking pretty great, it seems to be twice as fast as v1.0. v1.0:
2007 Jul 20
1
BOA (Bayesian Output Analysis)
Um texto embutido e sem conjunto de caracteres especificado associado... Nome: n?o dispon?vel Url: https://stat.ethz.ch/pipermail/r-help/attachments/20070719/d3991989/attachment.pl
2006 Nov 29
2
Dummies multiplied with other variable
Hi, I would like to estimate something like y = a + b*d2*y + c*d3*y where the dummies are created from some vector d with three (actually many more) levels using factor(). But either there is included the variable y or d1*y. How could I get rid of these? Example: x = c(1,2,3,4,5,6,7,8) y = c(3,6,2,8,7,6,2,4) d = c(1,1,1,2,3,2,3,3) fd = factor(d) lm(x ~ fd*y) gives: Coefficients: (Intercept)
2015 Jan 14
1
Questions regarding imaptest
Hi, The measurements were created under the following conditions : ? operating system : Red Hat Enterprise Linux Server release 6.6 (Santiago) kernel in version 2.6.32-504.el6.x86_64 ? virtual server (VMware) with an Intel(R) Xeon(R) 4vCPU E5649 x 2,526 GHz (2 cores per virtual socket) and 4 GB RAM ? 7,200 RPM SATA 1TB (FC SAN IBM System Storage N3400) ? all file systems had been formated in
2007 Mar 28
2
imaptest10 and stalled messages
Greetings - I've now got as far as playing with the imaptest10 test utility to see if I can stress-test our development server. imaptest10 is built against dovecot-1.0rc28 It may just be that I'm excessively heavy handed, but when I let imaptest10 rip with the command... ./imaptest10 user=test%03d host=testserver.imap.york.ac.uk clients=50 mbox=./dovecot.mbox msgs=1000 secs=30
2018 Jul 04
2
Benchmarks with imaptest
Hey guys, Has anybody ever used imaptest to benchmark *ONLY* reads? I'm trying to run a battery of tests in my new cluster and apparently no combination of parameters works, because fetch instruction doesn't run. Below is the command I'm using and its partial output imaptest/src/imaptest host=$SERVER_IP port=143 user=teste%d at example%d.com pass=$TEST_PASS users=10 domains=10
2009 Mar 05
3
Maildir dirty syncs
http://dovecot.org/patches/1.1/maildir-dirty-syncs.diff This patch adds a new maildir_very_dirty_syncs setting. If set to "yes", Dovecot assumes it's the only one changing the cur/ directory (so other MDAs can add mails to new/ without problems). This makes it possible to avoid rescanning the cur/ directory all the time looking for filenames. It also looks like (in stress testing)
2008 Feb 19
3
compiling error imaptest.c
dovecot-1.1.beta16]# gcc imaptest.c -o imaptest -g -Wall -W -I. -Isrc/lib-mail -Isrc/lib -Isrc/lib-imap -Isrc/lib-storage/index/mbox -DHAVE_CONFIG_H src/lib-storage/index/mbox/mbox-from.o src/lib-imap/libimap.a src/lib-mail/libmail.a src/lib/liblib.a imaptest.c: In client_append function: imaptest.c:1492: error: too many arguments for i_stream_create_limit function