search for: dir3

Displaying 20 results from an estimated 91 matches for "dir3".

Did you mean: dir
2018 May 15
2
New 3.12.7 possible split-brain on replica 3
Thank you Ravi for your fast answer. As requested you will find below the "stat" and "getfattr" of one of the files and its parent directory from all three nodes of my cluster. NODE 1: File: ?/data/myvolume-private/brick/dir1/dir2/dir3/dir4/dir5/dir6/dir7/dir8/dir9/dir10/OC_DEFAULT_MODULE/filename.shareKey? Size: 0 Blocks: 38 IO Block: 131072 regular empty file Device: 23h/35d Inode: 744413 Links: 2 Access: (0644/-rw-r--r--) Uid: (20936/ UNKNOWN) Gid: (20936/ UNKNOWN) Access: 2018-05-15 08:54:20.2960488...
2007 Nov 21
2
Access control question.
...ta that about 25 users have limited access to. I only want these users to have access to a sub-set of this data, but I also only want the users to see that which they have access to. So, for example, suppose that the share looks like thus: /smbshare /smbshare/dir1 /smbshare/dir2 /smbshare/dir3 And I only want the users to see that they have access to /smbshare/dir1 and /smbshare/dir3. The way that this is currently setup is that I have symlinks from the user's home directory to /smbshare/dir1 and /smbshare/dir3. That way then the user maps their home share, they only see dir1 a...
2018 May 17
0
New 3.12.7 possible split-brain on replica 3
...5:26 PM, mabi wrote: > Thank you Ravi for your fast answer. As requested you will find below the "stat" and "getfattr" of one of the files and its parent directory from all three nodes of my cluster. > > NODE 1: > > File: ?/data/myvolume-private/brick/dir1/dir2/dir3/dir4/dir5/dir6/dir7/dir8/dir9/dir10/OC_DEFAULT_MODULE/filename.shareKey? > Size: 0 Blocks: 38 IO Block: 131072 regular empty file > Device: 23h/35d Inode: 744413 Links: 2 > Access: (0644/-rw-r--r--) Uid: (20936/ UNKNOWN) Gid: (20936/ UNKNOWN) > Access: 2018-0...
2007 Dec 09
3
OT: Rsync question
Hello All, I have an off topic question about rsync and was wondering if i can get some kind person help with it. I have two servers with each server have three same directories on them /dir1/ /dir2/ /dir3/ . How would i achieve this by using rsync? I have tried rsync -avrt --delete server_ip:/dir1/ /dir2/ /dir3/ /dir1/ /dir2/ /dir3/ this does not do anything except give errors. Someone on IRC suggested i do a loop something like this, for i in "/dir1/ /dir2/ /dir3/"; do rsync -av $i root...
2018 May 17
2
New 3.12.7 possible split-brain on replica 3
...; > Thank you Ravi for your fast answer. As requested you will find below the "stat" and "getfattr" of one of the files and its parent directory from all three nodes of my cluster. > > > > NODE 1: > > > > File: ?/data/myvolume-private/brick/dir1/dir2/dir3/dir4/dir5/dir6/dir7/dir8/dir9/dir10/OC_DEFAULT_MODULE/filename.shareKey? > > > > Size: 0 Blocks: 38 IO Block: 131072 regular empty file > > > > Device: 23h/35d Inode: 744413 Links: 2 > > > > Access: (0644/-rw-r--r--) Uid: (20936/ UNKNOWN) Gid: (20936/ UNKNOWN)...
2003 Nov 25
1
rsync-bugs and unclear semantics when copying multiple source-dirs to one target
...ossibly the most promising. You could even write a little Perl script to build the exactly correct include lists. > There is another glitch, which I will describe here: > > if you have the following directory structure (-> is softlink) > > ./dir1/dir/a > ./dir2/dir -> ../dir3/dir > ./dir3/dir/b > > and do > > rsync -av --delete dir1/ dir2/ target > > you get > > ./target/dir -> ../dir3/dir > ./dir3/dir/a > ./dir3/dir/b > > I would expect either > > Variant 1: > ./target/dir -> ../dir3/dir > ./dir3/dir/b >...
2018 May 23
0
New 3.12.7 possible split-brain on replica 3
...i for your fast answer. As requested you will find below the "stat" and "getfattr" of one of the files and its parent directory from all three nodes of my cluster. > > > > > > NODE 1: > > > > > > File: ?/data/myvolume-private/brick/dir1/dir2/dir3/dir4/dir5/dir6/dir7/dir8/dir9/dir10/OC_DEFAULT_MODULE/filename.shareKey? > > > > > > Size: 0 Blocks: 38 IO Block: 131072 regular empty file > > > > > > Device: 23h/35d Inode: 744413 Links: 2 > > > > > > Access: (0644/-rw-r--r--) Uid: (20936/...
2018 May 15
0
New 3.12.7 possible split-brain on replica 3
On 05/15/2018 12:38 PM, mabi wrote: > Dear all, > > I have upgraded my replica 3 GlusterFS cluster (and clients) last Friday from 3.12.7 to 3.12.9 in order to fix this bug but unfortunately I notice that I still have exactly the same problem as initially posted in this thread. > > It looks like this bug is not resolved as I just got right now 3 unsynched files on my arbiter node
2001 Nov 16
1
include/exclude directory question
...t to send a subset of directories sepcified in "include" arguments to a client, but I am creating all the peer and parent directories as well, although they are empty - here is basically what I'm doing. assuming I have /staging/upgrade/dir1, /staging/upgrade/dir2 and /staging/upgrade/dir3 on the source tree. --include "*/" -include upgrade/dir1/** -include upgrade/dir3/** -exclude "*" /staging/* / what I get on the client is /ugrade/dir1/* /upgrade/dir2 (an empty directory) /upgrade/dir3/* How can I get rsync to filter out and stop creating the empty dir2 dir...
2018 May 23
1
New 3.12.7 possible split-brain on replica 3
...your fast answer. As requested you will find below the "stat" and "getfattr" of one of the files and its parent directory from all three nodes of my cluster. >>>> >>>> NODE 1: >>>> >>>> File: ?/data/myvolume-private/brick/dir1/dir2/dir3/dir4/dir5/dir6/dir7/dir8/dir9/dir10/OC_DEFAULT_MODULE/filename.shareKey? >>>> >>>> Size: 0 Blocks: 38 IO Block: 131072 regular empty file >>>> >>>> Device: 23h/35d Inode: 744413 Links: 2 >>>> >>>> Access: (0644/-rw-r--r--) Uid:...
2010 Jul 16
4
--compare-dest weirdness
...he compare-dest option. it throws no error about this dir not existing, and i have tried giving it as a relative dir to the destination with same result. interestingly, i tried to see if something was wrong with my statments by doing: mkdir ~/rsynctest/dir1 mkdir ~/rsynctest/dir2 mkdir ~/rsynctest/dir3 nano ~/rsynctest/dir1/file1 (wrote the line hello world and saved) nano ~/rsynctest/dir1/file2 (wrote the line hello and saved) cp ~/rsynctest/dir1/file* ~/rsynctest/dir2/ checked md5sums of both files in both dirs to ensure they were identical the did rsync -rvu --compare-dest=/home/kevin/rsyn...
2018 May 15
2
New 3.12.7 possible split-brain on replica 3
Dear all, I have upgraded my replica 3 GlusterFS cluster (and clients) last Friday from 3.12.7 to 3.12.9 in order to fix this bug but unfortunately I notice that I still have exactly the same problem as initially posted in this thread. It looks like this bug is not resolved as I just got right now 3 unsynched files on my arbiter node like I used to do before upgrading. This problem started
2005 Dec 20
4
rsync deleting symbolic link unexpectedly
I''m running 2.6.6 on a red hat EL3 system. using the following command, rsync -rlptDvzL --stats --include=a/ --include=a/dir/ --include=a/dir/symlink/ --include=a/dir/symlink/dir2/ --include=a/dir/symlink/dir2/dir3/ --exclude=''**'' /export/stuff remote::stuff I expect the "symlink" (which is really a symbolic link to a directory) to be preserved on the remote site even if dir2/dir3 doesn''t exist. Yet, when this transfer is executed under those conditions, the remote sit...
2018 Apr 09
2
New 3.12.7 possible split-brain on replica 3
...d in the past by this mailing list a now ran a stat and getfattr on one of the problematic files on all nodes and at the end a stat on the fuse mount directly. The output is below: > > > > NODE1: > > > > STAT: > > > > File: ?/data/myvol-private/brick/dir1/dir2/dir3/dir4/dir5/dir6/dir7/dir8/dir9/dir10/dir11/dir12_Archiv/azipfiledir.zip/OC_DEFAULT_MODULE/problematicfile? > > > > Size: 0 Blocks: 38 IO Block: 131072 regular empty file > > > > Device: 23h/35d Inode: 6822549 Links: 2 > > > > Access: (0644/-rw-r--r--) Uid: (209...
2018 Apr 09
0
New 3.12.7 possible split-brain on replica 3
...trusted.afr xattrs on this brick? > > And when you say "this brick", do you mean the brick on the arbitrer node (node 3 in my case)? Sorry I should have been clearer. Yes the brick on the 3rd node. `setfattr -x trusted.afr.myvol-private-client-0 /data/myvol-private/brick/dir1/dir2/dir3/dir4/dir5/dir6/dir7/dir8/dir9/dir10/dir11/dir12_Archiv/azipfiledir.zip/OC_DEFAULT_MODULE/problematicfile` `setfattr -x trusted.afr.myvol-private-client-1 /data/myvol-private/brick/dir1/dir2/dir3/dir4/dir5/dir6/dir7/dir8/dir9/dir10/dir11/dir12_Archiv/azipfiledir.zip/OC_DEFAULT_MODULE/problematicfi...
2016 Dec 14
4
[PATCH 0/4] sysprep: Remove various backup files.
https://bugzilla.redhat.com/show_bug.cgi?id=1401320 This series contains two new operations. The second -- and least controversial -- is "passwd-backups" which removes files such as /etc/passwd-, /etc/shadow- and so on. The first one ("backup-files") searches the whole guest filesystem for any regular file which looks like an editor backup file, such as "*~" and
2013 Dec 04
1
Query on make
Greetings, I have created several scripts which needs to be packaged. I have done my groundwork on rpmbuild Let us say I have 4 directories with scripts in them dir1, dir2, dir3 and dir4 I want to create different packages which will contain the compiled code (err.. shc) pack1: dir1,dir2 pack2: dir1,dir3 pack3: dir1,dir3 I want the resulting rpm packages will be residing elsewhere (perhaps BUILDROOT). Now I am aware of using different specfiles But When and how do I co...
2003 Feb 25
2
Difference in behaviour with --backup
When I do rsync -av --backup --backup-dir=/dir1/dir2/Backup \ /dir1/dir2/dir3 machine:/dir1/dir2/dir3 I get /dir1/dir2/Backup/dir3/... /Backup/dir3/dir4/... i.e. the tree under 'dir3' (my source tree) gets created under .../Backup. This is fine. But when I do the same thing with a single file like rsync -av --backup --backup-dir=/dir1/dir2/Backup...
2018 Apr 09
2
New 3.12.7 possible split-brain on replica 3
...k? > > > > And when you say "this brick", do you mean the brick on the arbitrer node (node 3 in my case)? > > Sorry I should have been clearer. Yes the brick on the 3rd node. > > `setfattr -x trusted.afr.myvol-private-client-0 /data/myvol-private/brick/dir1/dir2/dir3/dir4/dir5/dir6/dir7/dir8/dir9/dir10/dir11/dir12_Archiv/azipfiledir.zip/OC_DEFAULT_MODULE/problematicfile` > > `setfattr -x trusted.afr.myvol-private-client-1 /data/myvol-private/brick/dir1/dir2/dir3/dir4/dir5/dir6/dir7/dir8/dir9/dir10/dir11/dir12_Archiv/azipfiledir.zip/OC_DEFAULT_MODULE/prob...
2018 Apr 09
2
New 3.12.7 possible split-brain on replica 3
As I was suggested in the past by this mailing list a now ran a stat and getfattr on one of the problematic files on all nodes and at the end a stat on the fuse mount directly. The output is below: NODE1: STAT: File: ?/data/myvol-private/brick/dir1/dir2/dir3/dir4/dir5/dir6/dir7/dir8/dir9/dir10/dir11/dir12_Archiv/azipfiledir.zip/OC_DEFAULT_MODULE/problematicfile? Size: 0 Blocks: 38 IO Block: 131072 regular empty file Device: 23h/35d Inode: 6822549 Links: 2 Access: (0644/-rw-r--r--) Uid: (20909/ UNKNOWN) Gid: (20909/ UNKNOWN) Ac...