similar to: rsync transfers whole content when a new hardlink is created

Displaying 20 results from an estimated 60 matches similar to: "rsync transfers whole content when a new hardlink is created"

2009 Sep 20
1
Hardlink patches for sftp
Dear all, I am looking for the status on the hardlink patches that were published on this list last february by Miklos Szeredi? I'd really like to have hardlink support in sftp. This would in turn enable hardlinks in sshfs and make incremental rsync backups to remote filesystems possible. can someone tell me if these patches will be incorporated in openssh? The patches can be found under
2017 Jul 07
1
GluserFS WORM hardlink
GlusterFS WORM hard links will not be created OS is CentOS7 -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20170707/aaea8dfc/attachment.html>
2020 Jul 17
0
[OT] What is the max hardlink number for a single file on XFS
Hi list, I have a little script that uses rysnc and hardlink to perform backups. Some days ago a friend told me that rsync could crash if the hardlink limit is reached. I know (and tested) that for ext4 the max number of hardlink for a single file is 65000 but I can't get a limit on XFS. Due to the fact that I can't get good resources from google search, I tried to reach its limit
2012 Nov 08
1
Hardlink with Maildir a brief help
Hi to all, my question is: Is possible implementing SIS (with hardlink) with maildir instead of *dbox format? If yes in dovecot.conf it's only necessary the below parameters or what else? mail_attachment_dir = /var/qmail/attachments I have also acting the zlib plugin it is not a problem isn't it? Thanks in advance for any response -- */Davide Marchi /Teorema Ferrara Srl /(Tel:
2010 Jun 11
1
hardlink unlink-before-save?
Hello list On my server almost half of files are duplicated, and there are almost 100GB of data which is growing fast. I'm using weekly script to find duplicated files and replace them with hardlinks. Everything is fine as long as users aren't trying to edit and save some documents which have hardlinks. With samba 3.3 it seems that hardlinks aren't unlinked before file is saved.
2013 Dec 30
1
[Bug 10352] New: link-by-hash hardlink-collection maintenance mode
https://bugzilla.samba.org/show_bug.cgi?id=10352 Summary: link-by-hash hardlink-collection maintenance mode Product: rsync Version: 3.1.1 Platform: All OS/Version: All Status: NEW Severity: normal Priority: P5 Component: core AssignedTo: wayned at samba.org ReportedBy: jimklimov at
2012 Apr 15
0
Bug#666024: rsync --link-dest can incorrectly hardlink together destination files
[please Cc: 666024-forwarded at bugs.debian.org on any replies, thanks] Please see the attached Debian bug report, which includes a helpful bug-reproducing shell script; I've confirmed this still happens with 3.0.9. Paul Slootman On Tue 27 Mar 2012, Ian Jackson wrote: > Package: rsync > Version: 3.0.7-2 > > With rsync --link-dest, if two different source files (not hardlinked
2007 Jan 14
4
feature request, hardlink progress......
I'm copying a partition that has a bunch of hardlink based snapshots (-aPH). I think there's about 250,000 files in each backup and between 100 and 200 snapshots. Earlier today, I saw the files had completed and it was making all the hardlinks. I thought it would be "not long" but it's been making hardlinks for 12 hours (at least). There's only 36Gb in snapshot, the
2007 Apr 26
1
rsync mirroring and hardlink issues
I'm running a mirror of several repositories that are fetched using separate rsync runs. Since some of those repositories are hosting related files, I'm using the hardlink utility[1] in order to save disk space. However, I've noticed an issue that may lead to potential file metadata inconsistencies when using hardlink. Consider the following scenario: - two repositories (rep_a and
2013 Dec 19
5
[Bug 10334] New: rsync doesn't log hardlink-copies using --link-dest
https://bugzilla.samba.org/show_bug.cgi?id=10334 Summary: rsync doesn't log hardlink-copies using --link-dest Product: rsync Version: 3.0.9 Platform: All OS/Version: Linux Status: NEW Severity: normal Priority: P5 Component: core AssignedTo: wayned at samba.org ReportedBy: Kontakt at
2007 Jul 16
5
DO NOT REPLY [Bug 4793] New: link-dest hardlink does not always work well with -o -g -p
https://bugzilla.samba.org/show_bug.cgi?id=4793 Summary: link-dest hardlink does not always work well with -o -g -p Product: rsync Version: 2.6.9 Platform: PPC OS/Version: Linux Status: NEW Severity: enhancement Priority: P3 Component: core AssignedTo: wayned@samba.org
2005 Apr 22
2
hardlink bug in 2.6.4?
The background: I host one of the unofficial slackware mirrors & push to a high speed (unlisted) mirror that several other unofficials use for pulls. To save time/disk/bandwidth I use a custom script to hardlink the trees prior to pushing the updates to said unlisted mirror. The commandline (linewrap from hell, sorry): (note: username/password/hostname/modulename deliberately obfuscated)
2003 Dec 17
1
TODO hardlink reporting problem - fixed?
On Mon, 15 Dec 2003, jw schultz <jw@pegasys.ws> wrote: > OK, first pass on TODO complete. .... This hardlink bug report is nearly 21 months old... So I took a look at it using 2.5.7. See below. > BUGS --------------------------------------------------------------- > > Fix hardlink reporting 2002/03/25 > (was: There seems
2005 Jun 11
1
Feature request: rsync of device content
Hi list, I often save the content of a hardware device (a harddisk partition with a not-so-stable non-unix OS on it) on a backup disk which is connected over the network. Until now I used 'dd' to make a clone file of the device (dd if=/dev/hda2 of=disk.raw bs=2M) and then I copied or rsynced this file with the network copy. It would be nice if rsync would also be able to synchronise the
2009 Oct 15
1
DO NOT REPLY [Bug 6816] New: Delta-transfer algorithm does not reuse already transmitted identical blocks
https://bugzilla.samba.org/show_bug.cgi?id=6816 Summary: Delta-transfer algorithm does not reuse already transmitted identical blocks Product: rsync Version: 3.0.5 Platform: Other OS/Version: All Status: NEW Severity: enhancement Priority: P3 Component: core AssignedTo: wayned
2003 Dec 17
2
TODO hardlink performance optimizations
On Mon, 15 Dec 2003, jw schultz <jw@pegasys.ws> wrote: > OK, first pass on TODO complete. .... > PERFORMANCE ---------------------------------------------------------- .... > Traverse just one directory at a time > > Traverse just one directory at a time. Tridge says it's possible. > > At the moment rsync reads the whole file list into memory at the >
2008 Sep 29
5
DO NOT REPLY [Bug 5801] New: Sparse (-S) option doesn't work with new files
https://bugzilla.samba.org/show_bug.cgi?id=5801 Summary: Sparse (-S) option doesn't work with new files Product: rsync Version: 3.1.0 Platform: Other OS/Version: Linux Status: NEW Severity: normal Priority: P3 Component: core AssignedTo: wayned@samba.org ReportedBy:
2008 Nov 25
0
RFC: Proposed new web site design
Daniel Scharrer <dscharrer at gmail.com> at Nov 25, 2008 8:50 AM (MST) wrote about Re: RFC: Proposed new web site design > >The new user-centric design looks promising. Contrary to some others I >quite like the round corners and more modern look. Here are a few >observations: > >- Light gray text on a white background is not so nice. This is very hard to read. I like a
2005 Apr 06
1
rsync, --cvs-exclude option.
Hi Martin, Yes we can add "--include=*.exe", But --cvs-exclude means excluding only CVS dir right, But currently it is excluding other also. I am confused. ~~~~ static char default_cvsignore[] = /* These default ignored items come from the CVS manual. */ "RCS SCCS CVS CVS.adm RCSLOG cvslog.* tags TAGS" " .make.state .nse_depinfo *~ #* .#* ,*
2006 Apr 02
1
Make rsync exclude subversion working directories
Hi Guys, I'm backuping my home dir with rsync since 1 1/2 years, which is quite nice (I'm using the link-dir option). Now I started to use subversion and don't like that rsync is making backups of the checked-out working dirs. First I put all wd's in one folder (~/tmp) and put this on an exclude list. But now I'm use subversion on other locations as well, for example