similar to: sparse files on EC volume

Displaying 20 results from an estimated 9000 matches similar to: "sparse files on EC volume"

2017 Sep 26
2
sparse files on EC volume
Hi Xavi At this time I'm using 'plain' bricks with XFS. I'll be moving to LVM cached bricks. There is no RAID for data bricks, but I'll be using hardware RAID10 for SSD cache disks (I can use 'writeback' cache in this case). 'small file performance' is the main reason I'm looking at different options, i.e. using formated sparse files. I spent considerable
2017 Sep 26
0
sparse files on EC volume
Hi Dmitri, On 22/09/17 17:07, Dmitri Chebotarov wrote: > > Hello > > I'm running some tests to compare performance between Gluster FUSE mount > and formated sparse files (located on the same Gluster FUSE mount). > > The Gluster volume is EC (same for both tests). > > I'm seeing HUGE difference and trying to figure out why. Could you explain what hardware
2017 Sep 27
0
sparse files on EC volume
Have you done any testing with replica 2/3? IIRC my replica 2/3 tests out performed EC on smallfile workloads, it may be worth looking into if you can't get EC up to where you need it to be. -b ----- Original Message ----- > From: "Dmitri Chebotarov" <4dimach at gmail.com> > Cc: "gluster-users" <Gluster-users at gluster.org> > Sent: Tuesday,
2017 Sep 23
3
EC 1+2
Is possible to create a dispersed volume 1+2 ? (Almost the same as replica 3, the same as RAID-6) If yes, how many server I have to add in the future to expand the storage? 1 or 3? -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20170923/a702ba67/attachment.html>
2003 May 01
1
Batch Mode?
I realize batch mode is still experimental, but I was hoping there might be a workaround for a problem I am getting. I have been trying to run some tests and I get the below error when I use the --read-batch option to. I can successfully create an initial set of batch files, then a second set based upon a few modified test files from the first batch. When I first run the --read-batch option
2017 Sep 23
1
EC 1+2
Already read that. Seems that I have to use a multiple of 512, so 512*(3-2) is 512. Seems fine Il 23 set 2017 5:00 PM, "Dmitri Chebotarov" <4dimach at gmail.com> ha scritto: > Hi > > Take a look at this link (under ?Optimal volumes?), for Erasure Coded > volume optimal configuration > > http://docs.gluster.org/Administrator%20Guide/Setting%20Up%20Volumes/ >
2017 Jan 03
2
Inconsistent behavior using 3.1.2 from macOS 10.12.2 to an AFP mount
Hi, I've been attempting to use rsync 3.1.2 to copy files from a macOS 10.12.2 system to an AFP mounted share. The command I'm using is: rsync -avAX -M--fake-super ./testDir ./mnt/testDir/ This works fine and all the extended attributes are copied and readable. I then try to use rsync to copy these files back with: rsync -avAX --fake-super -M--super ./mnt/testDir/ ./testDir2/ This
2005 Mar 21
3
routes.rb question.
I have a simple program that read all HTML files from a directory and returns parts of the content and the file names which are HREFs to those file. I build the HREF string dynamicaly as PATH_TO_DOC_ROOT + "file_name". My PATH_TO_DOC_ROOT = http://127.0.0.1:3000/docman/public/docs. So, the final link that I am interested in might look like this:
2008 Oct 29
1
quota for ocfs2 do not warn a exceed for block soft-limit sometimes
Jan, since i can not cc you mail address when filing a bug on oss bugzilla,just paste the bug description here. Testing env, Kernel version:latest linus's mainline 2.6.27 Testing Nodes:test5/6 Description: Jan,i used your provided quota tools,now it works for me:),currently,i'm doing a sanity check for the quota on ocfs2. Found a strange thing when comparing with the ext3,the quota
2018 Feb 08
1
trusted.ec.dirty attribute
Hi I've got a problem on a EC volume where heal doesn't seem to work for just few files (heal info shows no progress for few days, Warning on the client mount) I ran getfattr -m . -d -e hex <brick>/<file> across all servers in the cluster and 'trusted.ec.dirty' attr is non-zero on all files which don't heal. Is it normal? Any way to correct it? It's my
2017 Sep 23
0
EC 1+2
Hi Take a look at this link (under ?Optimal volumes?), for Erasure Coded volume optimal configuration http://docs.gluster.org/Administrator%20Guide/Setting%20Up%20Volumes/ On Sat, Sep 23, 2017 at 10:01 Gandalf Corvotempesta < gandalf.corvotempesta at gmail.com> wrote: > Is possible to create a dispersed volume 1+2 ? (Almost the same as replica > 3, the same as RAID-6) > > If
2017 Nov 09
2
GlusterFS healing questions
Hi, You can set disperse.shd-max-threads to 2 or 4 in order to make heal faster. This makes my heal times 2-3x faster. Also you can play with disperse.self-heal-window-size to read more bytes at one time, but i did not test it. On Thu, Nov 9, 2017 at 4:47 PM, Xavi Hernandez <jahernan at redhat.com> wrote: > Hi Rolf, > > answers follow inline... > > On Thu, Nov 9, 2017 at
2006 Mar 23
17
Poor performance on NFS-exported ZFS volumes
I''m seeing some pretty pitiful performance using ZFS on a NFS server, with a ZFS volume exported (only with rw=host.foo.com,root=host.foo.com opts) and mounted on a Linux host running kernel 2.4.31. This linux kernel I''m working with is limited in that I can only do NFSv2 mounts... irregardless of that aspect, I''m sure something''s amiss. I mounted the zfs-based
2017 Nov 09
2
GlusterFS healing questions
Hi, We ran a test on GlusterFS 3.12.1 with erasurecoded volumes 8+2 with 10 bricks (default config,tested with 100gb, 200gb, 400gb bricksizes,10gbit nics) 1. Tests show that healing takes about double the time on healing 200gb vs 100, and abit under the double on 400gb vs 200gb bricksizes. Is this expected behaviour? In light of this would make 6,4 tb bricksizes use ~ 377 hours to heal. 100gb
2010 Feb 10
1
eval function with javascript-semantics possible in R?
Dear r-helpers, I am looking for an R-equivalent for the eval-function in javascript which can interpret a string as code on runtime, thereby allowing things like for (i in c(1:2)) { eval(items + "i") <- read.csv(eval(filename+ i), dec=","); } which would execute (with filename="testfile"): items1 <- read.csv(testfile1, dec=","); items2 <-
2017 Nov 09
0
GlusterFS healing questions
Someone on the #gluster-users irc channel said the following : "Decreasing features.locks-revocation-max-blocked to an absurdly low number is letting our distributed-disperse set heal again." Is this something to concider? Does anyone else have experience with tweaking this to speed up healing? Sent from my iPhone > On 9 Nov 2017, at 18:00, Serkan ?oban <cobanserkan at
2017 Nov 09
0
GlusterFS healing questions
Hi Rolf, answers follow inline... On Thu, Nov 9, 2017 at 3:20 PM, Rolf Larsen <rolf at jotta.no> wrote: > Hi, > > We ran a test on GlusterFS 3.12.1 with erasurecoded volumes 8+2 with 10 > bricks (default config,tested with 100gb, 200gb, 400gb bricksizes,10gbit > nics) > > 1. > Tests show that healing takes about double the time on healing 200gb vs > 100, and
2008 Oct 21
1
behavior of ALU Scheduler
Hello, I have one question about the ALU scheduler. If for example I have one UNIFY volume which is using ALU scheduler with the following config: volume unify type cluster/unify option namespace afr-ns option scheduler rr option scheduler alu # use the ALU scheduler option alu.limits.min-free-disk 3% # Don't create files one a volume with less than 5% free diskspace
2004 Feb 13
1
ACL bug
Hello, I'm using samba 3.0.2(acl) and kernel 2.4.24+acl, libacl-2.2.23. Following problem: When I create a file in an directory with extended ACLs, samba applies the "create mask" in a wrong way (IMHO). The normal behaviour of tools like chmod is that the second (middle) permission field is mapped to the "mask" ACE if the file has an extended ACL, so that the change
2017 Jun 02
1
?==?utf-8?q? Heal operation detail of EC volumes
Hi Serkan, On Thursday, June 01, 2017 21:31 CEST, Serkan ?oban <cobanserkan at gmail.com> wrote: ?>Is it possible that this matches your observations ? Yes that matches what I see. So 19 files is being in parallel by 19 SHD processes. I thought only one file is being healed at a time. Then what is the meaning of disperse.shd-max-threads parameter? If I set it to 2 then each SHD thread