similar to: why rsync update large directory so slow?

Displaying 20 results from an estimated 300 matches similar to: "why rsync update large directory so slow?"

2015 Nov 25
0
Slow backup
>>> I have a large directory with about 0 Milillion files in it to backup by network. Server and client both windows. >>> >>>In order to make backup faster, I created many rsync intances with "-avvu" options. It seems worked, and only take about 5 hours to complete backup job. >>>When I restarted the client bat script, and expect very little time to
2006 Feb 24
1
How to use multiple link-dest directories?
Hi guys, I hope this is not considered a stupid question, but it really seems to be something that needs to be covered with a decent example in the man page. I have looked in every piece of documentation I can think of and have found no answer. So I want to use the multiple link-dest directories feature as of 2.6.4. (I am using 2.6.6) My understanding of how this should work is that I can
2004 Jan 12
1
Performance problem
Hi, I am running rsync version 2.5.7 (stock distro on redhat linux, ES and 9) to rsync a directory with ~300k files in it from a machine on a 10Mbit internet tap to a machine with a 100Mbit internet tap. The problem is that I am only getting about 500Kbps during the transfer. I have tested link speed by scping a file between the two systems, performance was approximately 7.5Mbit/sec. Ping time
2005 Aug 21
1
Rsync 2.6.3 does not show to be deleted files. Bug?
Hi, I am running rsync 2.6.3 on Mac OS X 10.4. I have set up rsync to act as rsync server on one machine with a module entry like this: syslog facility = local5 [src-rw] path = /usr/local/src comment = whole /usr/local/src tree hosts allow = [ip addresses deleted] hosts deny = * uid = gerben gid = staff read only = no I have an
2002 Oct 16
2
HELP !!! Problem with file timestamps updating "weird" during rsync data pull
Running rsync 2.3.1, 2.4.1, and 2.5.5 on Solaris 8 on various system. I'm using rsync to pull data over for collecting performance data for graphing in Orca. What I'm seeing is the timestamps from time to time of the data files is being setup "weird" on the system the data is being pulled to. The timestamp should match that of the system the data is pulled from right ?
2010 Oct 22
0
NUT and AsciiDoc (was: Next AsciiDoc release)
Hi Stuart, sorry for the lag in answering, I know that you're awaiting feedback for the final release... 2010/10/19 Stuart Rackham > Hi Arnaud > > I successfully compiled all the docs in svn:// > svn.debian.org/nut/branches/AsciiDoc using the current AsciiDoc trunk, and > yes, the chunked manual only has a single toc. But there was the added > problem of the article
2005 Feb 22
1
Multiple OCFS intances in the same LAN
Hi, is it possible and safe to use more than one RAC-Cluster using OCFS in the same LAN? I'm considering to following setup: a 2-node production and a 2-node test system, the production data is stored in an EMC Symmetrix and for the test-system the production data is copied once to a BCV disk internally. This will give me an new independant copy of the production disk with the same contents,
2006 Jun 16
2
How normal is this memory usage?
The following is from just two invocations of "mongrel_rails". Is this normal? 57M x 3 seems a little high to me. This is the latest pre-release, Debian sarge, on my VPS. Two Rails apps running against SQLite3, with Pound in front. top - 10:50:12 up 47 days, 12:09, 1 user, load average: 0.06, 0.03, 0.01 Tasks: 52 total, 2 running, 50 sleeping, 0 stopped, 0 zombie Cpu(s):
2007 Aug 02
3
Dovecot strong or not for a big Webmail architecture
Hello, Do you think Dovecot could handle millions of active users in a big architecture ? This cluster could be for example (each server is a bi quad Xeon 2.66 Ghz) : - 40 Dovecot servers - 4 LVS - 20 Apache+PHP - 2 Openldap - 20 Postfix + ClamAV + SpamAssassin - 1 NFS Netapps For millions of user it could be multiple different clusters of 40 Dovecot servers with 1 Netapp for each cluster. I
2017 Feb 21
0
[PATCH 3/3] dib: rename "aux" to "in_target.aux"
Sadly, there are elements (hello "gentoo"!) that, during the cleanup.d phase, wipe out almost everything in the /tmp of the guest, including the /tmp/aux where virt-dib mounts the auxiliary data. Since that removal excludes things starting with "in_target" (mostly to avoid wiping the "in_target.d" that disk-image-create itself sets up, then rename our "aux"
2008 Sep 30
1
[LLVMdev] Integer handling
WRT: Message-based concurrence, you might what to check-out the papers here: http://www.malhar.net/sriram/
2007 Jan 29
1
Error in merging
Hi R, The error I get if I try to merge the zoo objects, intra1, intra2, and intra3 are as follows: > z=merge(intra1,intra2,intra3) Error in dimnames(x) <- dn : length of 'dimnames' [2] not equal to array extent Does this error mean that when merged, two columns have the same column name and so we get this error and hence not able to merge? Need a concurrence...
2007 Apr 11
1
bind or samba configuration preventing browsing network
I have a networking problem where I am not certain if the problem is samba or bind. I am still pretty much a nb at linux. The machine in question is running openSuSE 10.2 and is named rd1. I had samba working fine before I started to make it a WINS server and DNS host. I have a small LAN with no real administration functionality. The network is used for simple file sharing and dial-up
2002 Dec 18
2
patch for openssh3.5p1 - adds logging option
this patch adds a LogFile option to sshd_config. it just logs messages directly to a file instead of stderr or syslog. the largest change is an additional argument to log_init() in log.c for the log file name (and then changes to the rest of the tools to add a NULL arg). galt -------------- next part -------------- diff -urN openssh-3.5p1-orig/log.c openssh-3.5p1/log.c ---
2009 Sep 26
1
R as a web service
Dear R-helpers, I have been inquired about the possibility of developing a web distributed scoring system: a model is created in a central location, users fill a form in their browsers, and the central server calls this model and returns a YES/NO answer to them. I am tempted into using R for this assignment. I have used Rapache for similar tasks, but I am afraid that it is too of a novelty for
2015 Nov 06
2
[PATCH supermin] build: use a custom test driver
Use a custom test driver for running the tests: based on the test-driver provided by automake, it adds the running time of the test in each .trs file. --- configure.ac | 1 + guestfs-test-driver | 151 ++++++++++++++++++++++++++++++++++++++++++++++++++++ tests/Makefile.am | 2 + 3 files changed, 154 insertions(+) create mode 100755 guestfs-test-driver diff --git a/configure.ac
2009 Dec 15
1
[PATCH] The autotest timeout is now a command line configurable option.
By default it's 120 ms, but can be changed through command line arguments. Signed-off-by: Darryl L. Pierce <dpierce at redhat.com> --- autotest.sh | 16 ++++++++++------ 1 files changed, 10 insertions(+), 6 deletions(-) diff --git a/autotest.sh b/autotest.sh index c67931a..bcd9bd5 100755 --- a/autotest.sh +++ b/autotest.sh @@ -62,6 +62,7 @@ Usage: $ME [-n test_name] [LOGFILE] -i:
2015 Jun 10
3
Failed to init inotify - Too many open files
Hello, I've a problem on my system with inotify. In the smbd logfile are shown a lot messages like this: [2015/06/10 11:15:21.644453, 0, pid=57030, effective(12700, 100), real(0, 0)] smbd/notify_inotify.c:297(inotify_setup) Failed to init inotify - Too many open files [2015/06/10 11:15:23.968497, 0, pid=57030, effective(12700, 100), real(0, 0)] smbd/notify_inotify.c:297(inotify_setup)
2016 Aug 03
0
[PATCH] dib: rework run of extra-data.d hooks (RHBZ#1362354)
Instead of running them before lanching the appliance with the disks and then uploading the result after root.d hooks run, mount the root in the local temporary directory using FUSE and then run the hooks on the guest. Other than being closer to what diskimage-builder does, it also avoids issues with the extra-data.d scripts assuming to find things already, and we don't error out while trying
2010 Jul 16
2
Creating functions of "many" arguments
Dear users, My apologies for the simple question. I'd like to create a function where the number of arguments is as big as the size of my data set. Supose I have n observations in my data, how can I write a function like fun <- function (x1,x2,....,xn) {x1*x2*.......*xn} Thanks in advance for your help! Axel. [[alternative HTML version deleted]]