similar to: Turn off readdirp

Displaying 20 results from an estimated 600 matches similar to: "Turn off readdirp"

2017 Aug 23
0
Turn off readdirp
On Wed, Aug 23, 2017 at 08:06:18PM +0430, Tahereh Fattahi wrote: > Hi > How can I turn off readdirp? > I saw some solution but they don't work. > > I tried server side > gluster volume vol performance.force-readdirp off > gluster volume vol dht.force-readdirp off > gluster volume vol perforamnce.read-ahead off > and in client side fuse mount > mount -t glusterfs
2017 May 18
1
Rebalance without changing layout
Hi Is it possible to rebalance data in gluster without changing layout? When I use rebalance with force, the layout changed. I dont want changing layout, just balancing data over layout. -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20170518/f545fec3/attachment.html>
2018 May 30
0
RDMA inline threshold?
Dear Dan, thanks for the quick reply! I actually tried restarting all processes (and even rebooting all servers), but the error persists. I can also confirm that all birck processes are running. My volume is a distrubute-only volume (not dispersed, no sharding). I also tried mounting with use_readdirp=no, because the error seems to be connected to readdirp, but this option does not change
2017 Dec 29
1
A Problem of readdir-optimize
Hi Nithya, GlusterFS version is 3.11.0, and we use the native client of GlusterFS. Please see the below information. $gluster v info vol Volume Name: vol Type: Distributed-Replicate Volume ID: d59bd014-3b8b-411a-8587-ee36d254f755 Status: Started Snapshot Count: 0 Number of Bricks: 90 x 2 = 180 Transport-type: tcp,rdma Bricks: ... Options Reconfigured: performance.force-readdirp: false
2018 May 30
2
RDMA inline threshold?
Forgot to mention, sometimes I have to do force start other volumes as well, its hard to determine which brick process is locked up from the logs. Status of volume: rhev_vms_primary Gluster process TCP Port RDMA Port Online Pid ------------------------------------------------------------------------------ Brick spidey.ib.runlevelone.lan:/gluster/brick/rhev_vms_primary
2018 Apr 30
2
Gluster rebalance taking many years
2018 Apr 30
0
Gluster rebalance taking many years
Hi, This value is an ongoing rough estimate based on the amount of data rebalance has migrated since it started. The values will cange as the rebalance progresses. A few questions: 1. How many files/dirs do you have on this volume? 2. What is the average size of the files? 3. What is the total size of the data on the volume? Can you send us the rebalance log? Thanks, Nithya On 30
2018 Apr 30
1
Gluster rebalance taking many years
I cannot calculate the number of files normally Through df -i I got the approximate number of files is 63694442 [root at CentOS-73-64-minimal ~]# df -i Filesystem Inodes IUsed IFree IUse% Mounted on /dev/md2 131981312 30901030 101080282 24% / devtmpfs 8192893 435 8192458 1% /dev tmpfs
2017 Jul 11
1
Gluster native mount is really slow compared to nfs
Hello Vijay, ? ? What do you mean exactly? What info is missing? ? PS: I already found out that for this particular test all the difference is made by :?negative-timeout=600 , when removing it, it's much much slower again. ? ? Regards Jo ? -----Original message----- From:Vijay Bellur <vbellur at redhat.com> Sent:Tue 11-07-2017 18:16 Subject:Re: [Gluster-users] Gluster native mount is
2018 May 29
2
RDMA inline threshold?
Dear all, I faced a problem with a glusterfs volume (pure distributed, _not_ dispersed) over RDMA transport. One user had a directory with a large number of files (50,000 files) and just doing an "ls" in this directory yields a "Transport endpoint not connected" error. The effect is, that "ls" only shows some files, but not all. The respective log file shows this
2017 Jun 22
0
Rebalance without changing layout
----- Original Message ----- > From: "Tahereh Fattahi" <t28.fattahi at gmail.com> > To: gluster-users at gluster.org > Sent: Friday, May 19, 2017 12:21:53 AM > Subject: [Gluster-users] Rebalance without changing layout > > Hi > Is it possible to rebalance data in gluster without changing layout? > When I use rebalance with force, the layout changed. I dont
2018 May 30
2
[ovirt-users] Re: Gluster problems, cluster performance issues
The profile seems to suggest very high latencies on the brick at ovirt1.nwfiber.com:/gluster/brick1/engine ovirt2.* shows decent numbers. Is everything OK with the brick on ovirt1? Are the bricks of engine volume on both these servers identical in terms of their config? -Krutika On Wed, May 30, 2018 at 3:07 PM, Jim Kusznir <jim at palousetech.com> wrote: > Hi: > > Thank you. I
2018 Jan 07
0
performance.readdir-ahead on volume folders not showing with ls command 3.13.1-1.el7
with performance.readdir-ahead on on the volume maked folders on mounts invisible to ls command but it will show files fine it shows folders fine with ls on bricks what am I missing? maybe some settings are incompatible guess over-tuning happened vm1:/t1 /home/t1 glusterfs
2018 May 30
1
[ovirt-users] Re: Gluster problems, cluster performance issues
I've been back at it, and still am unable to get more than one of my physical nodes to come online in ovirt, nor am I able to get more than the two gluster volumes (storage domains) to show online within ovirt. In Storage -> Volumes, they all show offline (many with one brick down, which is correct: I have one server off) However, in Storage -> domains, they all show down (although
2015 Sep 02
3
TFTP-HPA beginner needs help
Hello, I am trying to flash new firmware into a router using TFTP-HPA. I am running Ubuntu 14.04 on a flash drive and still learning how to use Linux. I have also installed XINETD.D In one of the README files that comes with TFTP-HPA it said not to call it using tptp. Here are the instructions I'm trying to follow: cd /tmp ; tftp -gl aa 192.168.0.9; sh aa I do not know what
2018 May 30
0
[ovirt-users] Re: Gluster problems, cluster performance issues
Hi all again: I'm now subscribed to gluster-users as well, so I should get any replies from that side too. At this point, I am seeing acceptable (although slower than I expect) performance much of the time, with periodic massive spikes in latency (occasionally so bad as to cause ovirt to detect a engine bad health status). Often, if I check the logs just then, I'll see those call traces
2018 May 30
0
RDMA inline threshold?
Stefan, Sounds like a brick process is not running. I have notice some strangeness in my lab when using RDMA, I often have to forcibly restart the brick process, often as in every single time I do a major operation, add a new volume, remove a volume, stop a volume, etc. gluster volume status <vol> Does any of the self heal daemons show N/A? If that's the case, try forcing a restart on
2005 Aug 17
21
HOWTO unmaintained?
Hi, more than a month ago I proposed an addition to the HOWTO to address a certain packet classification problem/bug and how to fix it (see http://mailman.ds9a.nl/pipermail/lartc/2005q3/016728.html). I never received any reaction from the HOWTO maintainers, not even when addressing them directly (see mail below). Given that a month has gone by: Is the HOWTO currently unmaintained? Regards,
2017 Jul 11
2
Extremely slow du
Hi Kashif, Thank you for your feedback! Do you have some data on the nature of performance improvement observed with 3.11 in the new setup? Adding Raghavendra and Poornima for validation of configuration and help with identifying why certain files disappeared from the mount point after enabling readdir-optimize. Regards, Vijay On 07/11/2017 11:06 AM, mohammad kashif wrote: > Hi Vijay and
2018 Feb 05
2
Very slow rsync to gluster volume UNLESS `ls` or `find` scan dir on gluster volume first
Thanks for the report Artem, Looks like the issue is about cache warming up. Specially, I suspect rsync doing a 'readdir(), stat(), file operations' loop, where as when a find or ls is issued, we get 'readdirp()' request, which contains the stat information along with entries, which also makes sure cache is up-to-date (at md-cache layer). Note that this is just a off-the memory