similar to: How match.c hash_search works with multiple checksums that have identical tags

Displaying 20 results from an estimated 2000 matches similar to: "How match.c hash_search works with multiple checksums that have identical tags"

2004 Jan 27
1
Init array to -1 with memset()?
The match.c code has a loop that initializes an array to -1. I'm considering changing this to a memset() of 0xFF over all the array's bytes, but that depends on a system's representation of a -1 being "all bit on". Should I be anal about this and add a configure check to make sure that we're not running on some weird system where this is not true? Or should I just let
2003 Oct 05
2
Possible security hole
Maybe security related mails should be sent elsewhere? I didn't notice any so here it goes: sender.c:receive_sums() s->count = read_int(f); .. s->sums = (struct sum_buf *)malloc(sizeof(s->sums[0])*s->count); if (!s->sums) out_of_memory("receive_sums"); for (i=0; i < (int) s->count;i++) { s->sums[i].sum1 = read_int(f);
2012 Feb 01
0
timeout during hash_search
Hi, I'm experiencing timeouts during synchronization of large files. The scenario is as follows: Locally I have a large file (e.g. 20GB), which is already present at the server. When I try to resync it with: rsync --inplace --ignore-times -vvvv --compress bigfile rsync://xxx at ... The '--ignore-times' is only to force a comparison. The real world scenario would be the sync of a
2002 Mar 23
1
why variable last_i is needed in match.c rsync source ?
Hi all I see the rsync source and rsync makes hashing table and search hashing table tag_table to find the index of array struct sum_buf , which is a element of struct sum_struct. According to the source code, variable last_i is used to encourage adjacent matches allowing the RLL coding of the output to work more efficiently. Why last_i makes more efficiency? I can't understanding what
2002 Aug 05
5
[patch] read-devices
Greetings, I'd like to propose a new option to rsync, which causes it to read device files as if they were regular files. This includes pipes, character devices and block devices (I'm not sure about sockets). The main motivation is cases where you need to synchronize a large amount of data that is not available as regular files, as in the following scenarios: * Keep a copy of a block
2004 Apr 21
1
rsync-2.6.1pre-1 hang
Hi, I am running mentioned rsync version in daemon mode on a x86 machine which is a Gentoo Linux running kernel 2.4.26 and glibc 2.3.3_pre20040207. During the sync from another machine, the rsync daemon hangs; client receives no data and server waits on select() call. when this happens, rsync server process can only be killed by SIGKILL; no timeout occurs on server side. ssh is not being used for
2004 Jun 16
3
[Bug 1463] New: poor performance with large block size
https://bugzilla.samba.org/show_bug.cgi?id=1463 Summary: poor performance with large block size Product: rsync Version: 2.6.2 Platform: x86 OS/Version: other Status: NEW Severity: normal Priority: P3 Component: core AssignedTo: wayned@samba.org ReportedBy: wmatthews@sepaton.com
2014 Mar 26
11
[Bug 10518] New: rsync hangs (100% cpu)
https://bugzilla.samba.org/show_bug.cgi?id=10518 Summary: rsync hangs (100% cpu) Product: rsync Version: 3.1.1 Platform: All OS/Version: Linux Status: NEW Severity: critical Priority: P5 Component: core AssignedTo: wayned at samba.org ReportedBy: syzop at vulnscan.org QAContact:
2002 Apr 19
2
out of memory in build_hash_table
I recently installed rsync 2.5.5 on both my rsync server and client being used. I installed the latest version because I was having problems with rsync stalling with version 2.4.6 (I read that 2.5.5 was supposed to clear this up or at least give more appropriate errors). I am still having problems with rsync stalling even after upgrading to 2.5.5. It only stalls in the "/home" tree
2011 Apr 21
2
Crash copying to a zfs-fuse partition
Hi - I'm using rsync 3.0.8 on Fedora 14 x86-64 (package, not built). I get a repeated crash trying to rsync to a particular zfs partition: [root at xback1 diskbackup]# rsync -raHx --inplace --numeric-ids --stats --no- whole-file --delete /xback1_back1/home_jss/20110225-000501/ /xback1_test1/home_jss/current/ Segmentation fault [root at xback1 diskbackup]# rsync: writefd_unbuffered failed to
2011 Jul 22
2
[Bug 8315] New: hang in rsync (match.c)
https://bugzilla.samba.org/show_bug.cgi?id=8315 Summary: hang in rsync (match.c) Product: rsync Version: 3.0.9 Platform: All OS/Version: All Status: NEW Severity: normal Priority: P5 Component: core AssignedTo: wayned at samba.org ReportedBy: jeremy at jeremysanders.net QAContact:
2006 Mar 23
1
AIX 5.1 rsync large file
Hello, I have inherited a setup where there are 2 AIX 5.1 systems in 2 separate sites. There are large database files that are backed from each site to the other via rsync. Currently, it is using rsync version 2.5.4. It does it via ssh with the options -avz. This has all been merrily plodding along for some time. There is one file that is over 45 GB, and it started having trouble with that
2004 Jun 19
1
[Bug 1467] Hash table generation seems to be flawed
https://bugzilla.samba.org/show_bug.cgi?id=1467 ------- Additional Comments From ripper@internode.on.net 2004-06-19 02:18 ------- The way the code seems to work is as follows: A sorted list of hashes and their block numbers are generated (targets). They're sorted so that identical hashes are numerically next to each other. A reverse lookup table is then generated, that (i presume) is
2008 Jun 04
0
3.0.3pre2 compile warnings under cygwin
Under cygwin 1.5.25 we get the following compile warnings flist.c: In function `output_flist': flist.c:2648: warning: unsigned int format, uint32_t arg (arg 4) flist.c:2648: warning: unsigned int format, uint32_t arg (arg 4) flist.c:2655: warning: unsigned int format, uint32_t arg (arg 5) flist.c:2655: warning: unsigned int format, uint32_t arg (arg 5) flist.c:2659: warning: int format,
2008 Nov 26
1
bridging - Didn't get a frame from channel
Hi, I am having a difficulty with getting two realtime user?s to bridge on answer. I have managed successfully to bridge the same two users/channels via the Bridge Manager api command and confirm that the two communicate directly bypassing the asterisk server (I confirmed this with Wireshark). Does anyone have some ideas? I have put some log entries below. I haven?t attached my
2004 Jun 11
2
reporting a bug
problem with --backup I am doing this the old fashioned way via e-mail. I deal with Bugzilla too much and would prefer to not open yet another bugzilla account that generates more e-mail that I dont need to sift through, especially when it looks very similar to in-house bugzilla stuff. I am running rsync 2.6.2 as it is downloaded with only the --bwlimit patch supplied by Wayne via e-mail
2006 Jan 05
2
POP3 only - why is cur/ growing?
Hi, I use Sarge's dovecot for a pop3 mail collection server. IMAP is not used at all. It works well, but I would like to know why one of the accounts also keeps a copy of all e-mail copies in the cur/ folder? Thanks a lot, Wally -- Wally Winchester wally_winchester at fastmail.fm -- http://www.fastmail.fm - The way an email service should be
2001 Oct 31
3
2.9.9p2 and Solaris-2.8 PAM: Cannot delete credentials[7]: Permission denied
The 2 errors: pam_setcred: error Permission denied Cannot delete credentials[7]: Permission denied Looks to be a major bug in the PAM module for Solaris-2.8/2.7/2.6. Has anyone from the list (developers of OpenSSH, endusers, hackers, etc.) came up w/ a solution? Even a temporary one? When authenticating yourself on the same system that worked, but when authenticating to another system failed. I
2004 Dec 23
1
searching Jonathan Baron's R Site
First, my site will be down December 27-28 because of a network upgrade at Penn. It will also be down at least one day before that, while I upgrade the operating system. (And another day some time in January because of a planned power outage.) Second, I have replaced the search engine in my R site: http://finzi.psych.upenn.edu/ I am now using Namazu instead of HtDig. The direct link to the
2004 Jun 18
2
possible writefd_unbuffered error; what am I screwing up this time
I am trying to execute rsync manually at a "remote" server to test out --read-batch execution. I created the batch files on another server and then rcp'ed them to the "remote" server. I had some issues of not having the correct working directory on the remote system when I did an rsh "remote" rsync --read-batch ...... so I opened a telnet session on the remote