similar to: include/exclude Problem

Displaying 20 results from an estimated 900 matches similar to: "include/exclude Problem"

2014 Mar 13
2
Fwd: C code and DLL file in the package
Hi, I am trying to compile a package, in which C code is used. When I was checking the package with the dll file under the sub-folder "SRC", a WARNING msg was returned: "Source packages should not contain undeclared executable files." When I exclued it from the folder, I got an ERROR msg: Error in library.dynam(lib, package, package.lib) : DLL 'Cs2g' not found:
2003 Nov 03
2
--move instaed of --delete
Hi, I have set up daily backup using rsync to collect the data and afterwards I use afio to write the data to tape. I do: $RSYNC -avbu --backup-dir=$PART/$HOST/old -e ssh --exclude-from $EXCLUDE root@$HOST:/ $PART/$HOST > $LOG/$HOST-$DATE.log to sync files. Since the tape capacity is limited I exclued the /$PART/$HOST/old dirs, form the afio call. These are backuped separately on weekly
2011 Nov 23
8
[PATCH 0/8] Add MD inspection support to libguestfs
This series fixes inspection in the case that fstab contains references to md devices. I've made a few changes since the previous posting, which I've summarised below. [PATCH 1/8] build: Create an MD variant of the dummy Fedora image I've double checked that no timestamp is required in the Makefile. The script will not run a second time to build fedora-md2.img. [PATCH 2/8] build:
2010 May 14
2
command line to backup my documents to external drive
I'm really confused with all the examples out there and all different types of incremental backups. I tried several scripts but cannot reduce the size of my backup folders. What I want is to backup my documents to my external drive every month and save as much disk space as possible. Lets say I have 3 backup directories in the external drive, backup03.10, backup04.10 and backup05.10. I want
2010 Apr 08
6
Mac OS X "rsync: unpack_smb_acl: sys_acl_get_info(): Unknown error: 0 (0)"
I am a developer on the LBackup project. An LBackup user recently posted a question to the mailing list asking about the following error. > "rsync: unpack_smb_acl: sys_acl_get_info(): Unknown error: 0 (0)" Link to thread : <http://www.mail-archive.com/lbackup-discussion at lists.connect.homeunix.com/msg00040.html> My understanding of this error is that when copying files via
2010 Jul 16
4
--compare-dest weirdness
Hi All, I am writing a backup program for my computer. brief outline is as follows. Running ubuntu 10.04 2 main partitions, / and /home, both ext3. 1 external usb hdd, ext3, mounted to /backups/main. once every couple of days, rsync backs up, using following command, everything worth backing up in / and /home partitions to a folder /backups/main/Full. command: "rsync -vrhRupElog
2010 Jun 22
2
few questions on rsync
Hi, I have few questions that could not find answers to in documentation. Different filesystems: Lets say I want to keep all extended attributes and everything else, so I use -A, -X, --perms etc., together with --fake-super. Now, lets say source FS supports some attributes not supported on target FS (for example, XFS extended attributes). Would this work as expected (ie. extended attributes are
2016 Jun 19
1
rsync script for snapshot backups
Am 19.06.2016 um 19:27 schrieb Simon Hobson: > Dennis Steinkamp <dennis at lightandshadow.tv> wrote: > >> i tried to create a simple rsync script that should create daily backups from a ZFS storage and put them into a timestamp folder. >> After creating the initial full backup, the following backups should only contain "new data" and the rest will be referenced
2015 Apr 17
1
Recycling directories and backup performance. Was: Re: rsync --link-dest won't link even if existing file is out of date (fwd)
How do you handle snapshotting? or do you leave that to the block/fs virtualization layer? /kc On Fri, Apr 17, 2015 at 01:35:27PM +1200, Henri Shustak said: >> Our backup procudures have provision for looking back at previous directories, but there is not much to be gained with recycled directories. Without recycling, and after a failure, the latest available backup may not have much
2009 Sep 10
2
originate sync from the daemon server
is there anything special to do this from the daemon server. I've setup the /etc/rsyncd.conf with some filesystems and I would rather originate (control) my rsyncs from this server and not from the hosts that have the data I want. ie. I want to pull not push. for instance my rsyncd.conf [www] comment = www path = /snaps/www numeric ids = true log file = /snaps/rsync/logs/www.log pid file
2008 Nov 08
2
Differential backup
Hello, I have got three folders: - /home/backup/2008-10-20 - place for differential backup - /mnt/for_backup - folder with files for backup - /home/backup/2008-10-01 - place where the last full backup is My question: is below command prepared correctly to make differential backup? rsync -avPbn --backup-dir=/home/backup/2008-10-20/ --exclude "System Volume Information" --exclude
2015 Apr 06
6
rsync --link-dest won't link even if existing file is out of date
Feature request: allow --link-dest dir to be linked to even if file exists in target. This statement from the man page is adhered to too strongly IMHO: "This option works best when copying into an empty destination hierarchy, as rsync treats existing files as definitive (so it never looks in the link-dest dirs when a destination file already exists)". I was suprised by this behaviour
2011 Jul 03
1
Brandysnap -- a new rsync-based snapshot management script
Brandysnap is an rsync-based script with a difference. Unlike dirvish, it does not assign importance to snapshots when they are created. All snapshots are created equal, and then they are managed so that the required number of old snapshots is maintained. Unlike rsnapshot, brandysnap does not get its snapshots out of sync if runs are missed, and it is very easy to configure. The link between
2014 Mar 19
1
Beating a dead horse
Sorry to do this .... AGAIN Every year or two I get stuck on this same problem involving excluding. Seems I learn how its done then 2yrs later I've totally forgotten and when I look up my notes ... this new need is just different enough that they don't apply. Here's the problem. (Simplified... and I've skipped some of the repetitive output) On remote ls A/ a/ b/ c/ d e on
2010 Jul 27
3
Getting rsync to store timing information in its logs
Is there a way to know from the rsync logs how long it took to do a backup? The only timing info,. I see is this at the end: sent 3067328 bytes received 7853035429 bytes 1187888.83 bytes/sec total size is 1559866450336 speedup is 198.55 Can I use it to figure out how long the operation took? Does the above mean it took 2.5 secs of send time and 1.8 hours of recieve time so (roughly) the
2007 Apr 25
2
Raid 1 newbie question
Hi I have a Raid 1 centos 4.4 setup and now have this /proc/mdstat output: [root at server admin]# cat /proc/mdstat Personalities : [raid1] md2 : active raid1 hdc2[1] hda2[0] 1052160 blocks [2/2] [UU] md1 : active raid1 hda3[0] 77023552 blocks [2/1] [U_] md0 : active raid1 hdc1[1] hda1[0] 104320 blocks [2/2] [UU] What happens with md1 ? My dmesg output is: [root at
2014 May 14
0
[RFC PATCH v1 16/16] drm/ttm: use rcu in core ttm
Signed-off-by: Maarten Lankhorst <maarten.lankhorst at canonical.com> --- drivers/gpu/drm/ttm/ttm_bo.c | 76 +++++++----------------------------------- 1 file changed, 13 insertions(+), 63 deletions(-) diff --git a/drivers/gpu/drm/ttm/ttm_bo.c b/drivers/gpu/drm/ttm/ttm_bo.c index 31c4a6dd722d..6fe1f4bf37ed 100644 --- a/drivers/gpu/drm/ttm/ttm_bo.c +++ b/drivers/gpu/drm/ttm/ttm_bo.c @@
2008 Nov 09
7
How to delete files older than X on backup during sync?
Hi list, I have been reading MAN pages and listarchives, but not found the answer to my question though I am sure it must be possible to acheive my wish. I wish to use rsync to create a backup BUT only keep the files for a limited period of time, EG two weeks. I have not yet been able to figure out how to do this inside rsync (while the backup is being performed) and my understanding of the
2011 Jun 23
3
Using rsync as an incremental backup
I'm using rsync to do an incremental backup of my desktop here, to a remote server as follows: #/usr/bin/bash old=$(date -d 'now - 1 week' +%Y-%m-%d) new=$(date +%Y-%m-%d) rsync -avP --delete --link-dest=../$dir /home/bakers bakers at perturb.org:/home/bakers/backup/$new/ This is actually working GREAT! The only problem is that sometimes the cronjob won't complete (internet is
2006 Aug 28
2
Cannot get simple data.frame binding.
I am stuck on a simple problem where an example works fine but the real one does not. I have a data.frame where I wish to sum up some values across the rows and create a new data.frame with some of old data.frame variables and the new summed variable. It works fine in my simple example but I am doing something wrong in the real world. In the real world I am loading a labeled data.frame. The