Displaying 20 results from an estimated 2000 matches similar to: "How match.c hash_search works with multiple checksums that have identical tags"
2004 Jan 27
1
Init array to -1 with memset()?
The match.c code has a loop that initializes an array to -1. I'm
considering changing this to a memset() of 0xFF over all the array's
bytes, but that depends on a system's representation of a -1 being "all
bit on". Should I be anal about this and add a configure check to make
sure that we're not running on some weird system where this is not true?
Or should I just let
2003 Oct 05
2
Possible security hole
Maybe security related mails should be sent elsewhere? I didn't notice
any so here it goes:
sender.c:receive_sums()
s->count = read_int(f);
..
s->sums = (struct sum_buf *)malloc(sizeof(s->sums[0])*s->count);
if (!s->sums) out_of_memory("receive_sums");
for (i=0; i < (int) s->count;i++) {
s->sums[i].sum1 = read_int(f);
2012 Feb 01
0
timeout during hash_search
Hi,
I'm experiencing timeouts during synchronization of large files.
The scenario is as follows: Locally I have a large file (e.g. 20GB),
which is already present at the server. When I try to resync it with:
rsync --inplace --ignore-times -vvvv --compress bigfile rsync://xxx at ...
The '--ignore-times' is only to force a comparison. The real world scenario
would be the sync of a
2002 Mar 23
1
why variable last_i is needed in match.c rsync source ?
Hi all
I see the rsync source and rsync makes hashing table and search hashing
table tag_table to find the index of array struct sum_buf , which is a
element of struct sum_struct.
According to the source code, variable last_i is used to encourage
adjacent matches allowing the RLL coding of the output to work more
efficiently.
Why last_i makes more efficiency?
I can't understanding what
2002 Aug 05
5
[patch] read-devices
Greetings,
I'd like to propose a new option to rsync, which causes it to read
device files as if they were regular files. This includes pipes,
character devices and block devices (I'm not sure about sockets). The
main motivation is cases where you need to synchronize a large amount of
data that is not available as regular files, as in the following scenarios:
* Keep a copy of a block
2004 Apr 21
1
rsync-2.6.1pre-1 hang
Hi,
I am running mentioned rsync version in daemon mode on a x86 machine
which is a Gentoo Linux running kernel 2.4.26 and glibc 2.3.3_pre20040207.
During the sync from another machine, the rsync daemon hangs; client receives
no data and server waits on select() call. when this happens, rsync server process
can only be killed by SIGKILL; no timeout occurs on server side.
ssh is not being used for
2004 Jun 16
3
[Bug 1463] New: poor performance with large block size
https://bugzilla.samba.org/show_bug.cgi?id=1463
Summary: poor performance with large block size
Product: rsync
Version: 2.6.2
Platform: x86
OS/Version: other
Status: NEW
Severity: normal
Priority: P3
Component: core
AssignedTo: wayned@samba.org
ReportedBy: wmatthews@sepaton.com
2014 Mar 26
11
[Bug 10518] New: rsync hangs (100% cpu)
https://bugzilla.samba.org/show_bug.cgi?id=10518
Summary: rsync hangs (100% cpu)
Product: rsync
Version: 3.1.1
Platform: All
OS/Version: Linux
Status: NEW
Severity: critical
Priority: P5
Component: core
AssignedTo: wayned at samba.org
ReportedBy: syzop at vulnscan.org
QAContact:
2002 Apr 19
2
out of memory in build_hash_table
I recently installed rsync 2.5.5 on both my rsync server and client being
used. I installed the latest version because I was having problems with
rsync stalling with version 2.4.6 (I read that 2.5.5 was supposed to clear
this up or at least give more appropriate errors). I am still having
problems with rsync stalling even after upgrading to 2.5.5. It only stalls
in the "/home" tree
2011 Apr 21
2
Crash copying to a zfs-fuse partition
Hi - I'm using rsync 3.0.8 on Fedora 14 x86-64 (package, not built). I get a
repeated crash trying to rsync to a particular zfs partition:
[root at xback1 diskbackup]# rsync -raHx --inplace --numeric-ids --stats --no-
whole-file --delete /xback1_back1/home_jss/20110225-000501/
/xback1_test1/home_jss/current/
Segmentation fault
[root at xback1 diskbackup]# rsync: writefd_unbuffered failed to
2011 Jul 22
2
[Bug 8315] New: hang in rsync (match.c)
https://bugzilla.samba.org/show_bug.cgi?id=8315
Summary: hang in rsync (match.c)
Product: rsync
Version: 3.0.9
Platform: All
OS/Version: All
Status: NEW
Severity: normal
Priority: P5
Component: core
AssignedTo: wayned at samba.org
ReportedBy: jeremy at jeremysanders.net
QAContact:
2006 Mar 23
1
AIX 5.1 rsync large file
Hello,
I have inherited a setup where there are 2 AIX 5.1 systems in 2 separate
sites. There are large database files that are backed from each site to
the other via rsync. Currently, it is using rsync version 2.5.4. It does
it via ssh with the options -avz. This has all been merrily plodding along
for some time. There is one file that is over 45 GB, and it started having
trouble with that
2004 Jun 19
1
[Bug 1467] Hash table generation seems to be flawed
https://bugzilla.samba.org/show_bug.cgi?id=1467
------- Additional Comments From ripper@internode.on.net 2004-06-19 02:18 -------
The way the code seems to work is as follows:
A sorted list of hashes and their block numbers are generated (targets).
They're sorted so that identical hashes are numerically next to each other.
A reverse lookup table is then generated, that (i presume) is
2008 Jun 04
0
3.0.3pre2 compile warnings under cygwin
Under cygwin 1.5.25 we get the following compile warnings
flist.c: In function `output_flist':
flist.c:2648: warning: unsigned int format, uint32_t arg (arg 4)
flist.c:2648: warning: unsigned int format, uint32_t arg (arg 4)
flist.c:2655: warning: unsigned int format, uint32_t arg (arg 5)
flist.c:2655: warning: unsigned int format, uint32_t arg (arg 5)
flist.c:2659: warning: int format,
2008 Nov 26
1
bridging - Didn't get a frame from channel
Hi,
I am having a difficulty with
getting two realtime user?s to bridge on answer. I have managed successfully to
bridge the same two users/channels via the Bridge Manager api command and
confirm that the two communicate directly bypassing the asterisk server (I
confirmed this with Wireshark).
Does anyone have some ideas? I have
put some log entries below.
I haven?t attached my
2004 Jun 11
2
reporting a bug
problem with --backup
I am doing this the old fashioned way via e-mail.
I deal with Bugzilla too much and would prefer to not open yet another bugzilla account that generates more e-mail that I dont need to sift through, especially when it looks very similar to in-house bugzilla stuff.
I am running rsync 2.6.2 as it is downloaded with only the --bwlimit patch supplied by Wayne via e-mail
2006 Jan 05
2
POP3 only - why is cur/ growing?
Hi,
I use Sarge's dovecot for a pop3 mail collection server. IMAP is not
used at all. It works well, but I would like to know why one of the
accounts also keeps a copy of all e-mail copies in the cur/ folder?
Thanks a lot,
Wally
--
Wally Winchester
wally_winchester at fastmail.fm
--
http://www.fastmail.fm - The way an email service should be
2001 Oct 31
3
2.9.9p2 and Solaris-2.8 PAM: Cannot delete credentials[7]: Permission denied
The 2 errors:
pam_setcred: error Permission denied
Cannot delete credentials[7]: Permission denied
Looks to be a major bug in the PAM module for Solaris-2.8/2.7/2.6.
Has anyone from the list (developers of OpenSSH, endusers, hackers, etc.)
came up w/ a solution? Even a temporary one?
When authenticating yourself on the same system that worked, but when
authenticating to another system failed. I
2004 Dec 23
1
searching Jonathan Baron's R Site
First, my site will be down December 27-28 because of a network
upgrade at Penn. It will also be down at least one day before
that, while I upgrade the operating system. (And another day
some time in January because of a planned power outage.)
Second, I have replaced the search engine in my R site:
http://finzi.psych.upenn.edu/
I am now using Namazu instead of HtDig. The direct link to the
2004 Jun 18
2
possible writefd_unbuffered error; what am I screwing up this time
I am trying to execute rsync manually at a "remote" server to test out --read-batch execution.
I created the batch files on another server and then rcp'ed them to the "remote" server. I had some issues of not having the
correct working directory on the remote system when I did an rsh "remote" rsync --read-batch ...... so I opened a telnet session
on the remote