similar to: Is there any performance impact in setting up every gluster client as a NFS server?

Displaying 20 results from an estimated 1000 matches similar to: "Is there any performance impact in setting up every gluster client as a NFS server?"

2017 Nov 14
1
Is there any performance impact in setting up every gluster client as a NFS server?
Hi, We have around 60 hosts and each of them acts as glusterFs clients as well as server. To achieve HA, my underatanding is that we can use Ganesha NFS alone (and not Kernel NFS) and for above 3.10 versions, the HA packages are not ready yet. So, I'm thinking if I can avoid HA by making each client acts as server for itself. i.e.,the server name specified while mounting is localhost. So,
2017 Nov 14
0
What is the difference between FORGET and UNLINK fops
Hi Ben, Yes, I was referring to the same link before and I think most of the FOPS I've understood from the definitions there and I'll update if I found any corrections or new definitions. Btw, my previous underatanding about LOOKUP is wrong, its NOT dns lookup. Regards, Jeevan. On Nov 14, 2017 11:02 AM, "Ben Turner" <bturner at redhat.com> wrote: Here is some of the
2018 Jan 03
1
stale file handle on gluster NFS client when trying to remove a directory
An ESTALE error usually means the gfid could not be found. Does repeating the "rm -rf" delete the directory? Regards, Nithya On 3 January 2018 at 12:16, Jeevan Patnaik <g1patnaik at gmail.com> wrote: > Hi all, > > I haven't found any root cause or workaround for this yet. Can any one > help me in underatanding the issue? > > Regards, > Jeevan. > >
2018 Jan 03
0
stale file handle on gluster NFS client when trying to remove a directory
Hi all, I haven't found any root cause or workaround for this yet. Can any one help me in underatanding the issue? Regards, Jeevan. On Dec 21, 2017 8:20 PM, "Jeevan Patnaik" <g1patnaik at gmail.com> wrote: > Hi, > > > After running rm -rf on a directory, the files under it got deleted, but > the directory was not deleted and was showing stale file handle
2017 Aug 08
1
Slow write times to gluster disk
Soumya, its [root at mseas-data2 ~]# glusterfs --version glusterfs 3.7.11 built on Apr 27 2016 14:09:20 Repository revision: git://git.gluster.com/glusterfs.git Copyright (c) 2006-2013 Red Hat, Inc. <http://www.redhat.com/> GlusterFS comes with ABSOLUTELY NO WARRANTY. It is licensed to you under your choice of the GNU Lesser General Public License, version 3 or any later version (LGPLv3
2017 Jul 11
0
Gluster native mount is really slow compared to nfs
On 07/11/2017 08:14 AM, Jo Goossens wrote: > RE: [Gluster-users] Gluster native mount is really slow compared to nfs > > Hello Joe, > > I really appreciate your feedback, but I already tried the opcache > stuff (to not valildate at all). It improves of course then, but not > completely somehow. Still quite slow. > > I did not try the mount options yet, but I will now!
2017 Jul 11
0
Gluster native mount is really slow compared to nfs
My standard response to someone needing filesystem performance for www traffic is generally, "you're doing it wrong". https://joejulian.name/blog/optimizing-web-performance-with-glusterfs/ That said, you might also look at these mount options: attribute-timeout, entry-timeout, negative-timeout (set to some large amount of time), and fopen-keep-cache. On 07/11/2017 07:48 AM, Jo
2017 Aug 08
0
Slow write times to gluster disk
----- Original Message ----- > From: "Pat Haley" <phaley at mit.edu> > To: "Soumya Koduri" <skoduri at redhat.com>, gluster-users at gluster.org, "Pranith Kumar Karampuri" <pkarampu at redhat.com> > Cc: "Ben Turner" <bturner at redhat.com>, "Ravishankar N" <ravishankar at redhat.com>, "Raghavendra
2017 Jul 11
2
Gluster native mount is really slow compared to nfs
Hello Joe, ? ? I really appreciate your feedback, but I already tried the opcache stuff (to not valildate at all). It improves of course then, but not completely somehow. Still quite slow. ? I did not try the mount options yet, but I will now! ? ? With nfs (doesnt matter much built-in version 3 or ganesha version 4) I can even host the site perfectly fast without these extreme opcache settings.
2017 Jul 11
2
Gluster native mount is really slow compared to nfs
Hello, ? ? Here is the volume info as requested by soumya: ? #gluster volume info www ?Volume Name: www Type: Replicate Volume ID: 5d64ee36-828a-41fa-adbf-75718b954aff Status: Started Snapshot Count: 0 Number of Bricks: 1 x 3 = 3 Transport-type: tcp Bricks: Brick1: 192.168.140.41:/gluster/www Brick2: 192.168.140.42:/gluster/www Brick3: 192.168.140.43:/gluster/www Options Reconfigured:
2017 Jul 11
0
Gluster native mount is really slow compared to nfs
Hello, ? ? Here is some speedtest with a new setup we just made with gluster 3.10, there are no other differences, except glusterfs versus nfs. The nfs is about 80 times faster: ? ? root at app1:~/smallfile-master# mount -t glusterfs -o use-readdirp=no,log-level=WARNING,log-file=/var/log/glusterxxx.log 192.168.140.41:/www /var/www root at app1:~/smallfile-master# ./smallfile_cli.py ?--top
2017 Jul 11
2
Gluster native mount is really slow compared to nfs
Hello Joe, ? ? I just did a mount like this (added the bold): ? mount -t glusterfs -o attribute-timeout=600,entry-timeout=600,negative-timeout=600,fopen-keep-cache,use-readdirp=no,log-level=WARNING,log-file=/var/log/glusterxxx.log 192.168.140.41:/www /var/www ?Results: ? root at app1:~/smallfile-master# ./smallfile_cli.py ?--top /var/www/test --host-set 192.168.140.41 --threads 8 --files 5000
2017 Dec 10
0
What does gluster do when it's idle?
Hi, Also I forgot to mention there are DUMP and PORTBYBRICK calls too. I understood these too calls are to get updated on gluster configuration. But all these two calls are to a particular node only while there are total (2+1)*3= 6 nodes in my configuration. Why so? What exactly is DUMP and PORTBYBRICK? Regards Jeevan. On Dec 10, 2017 6:47 PM, "Jeevan Patnaik" <g1patnaik at
2017 Nov 13
0
What is the difference between FORGET and UNLINK fops
Filtering the brick logs in TRACE mode with rpcsvc.c does show the FOPS. >From this, I've realized that LOOKUP is actually dns lookup. This actually differs from NFS lookup operation. Please correct me if I'm wrong. Regards, Jeevan. On Nov 13, 2017 9:40 PM, "Jeevan Patnaik" <g1patnaik at gmail.com> wrote: > Hi, > > Can I get a brief description of all the
2017 Jul 14
0
Slow write times to gluster disk
Hi Soumya, I just noticed some of the notes at the bottom. In particular * Till glusterfs-3.7, gluster-NFS (gNFS) gets enabled by default. The only requirement is that kernel-NFS has to be disabled for gluster-NFS to come up. Please disable kernel-NFS server and restart glusterd to start gNFS. In case of any issues with starting gNFS server, please look at
2017 Jul 07
0
Slow write times to gluster disk
Hi All, A follow-up question. I've been looking at various pages on nfs-ganesha & gluster. Is there a version of nfs-ganesha that is recommended for use with glusterfs 3.7.11 built on Apr 27 2016 14:09:22 CentOS release 6.8 (Final) Thanks Pat On 07/05/2017 11:36 AM, Pat Haley wrote: > > Hi Soumya, > > (1) In
2017 Jul 05
2
Slow write times to gluster disk
Hi Soumya, (1) In http://mseas.mit.edu/download/phaley/GlusterUsers/TestNFSmount/ I've placed the following 2 log files etc-glusterfs-glusterd.vol.log gdata.log The first has repeated messages about nfs disconnects. The second had the <fuse_mnt_direcotry>.log name (but not much information). (2) About the gluster-NFS native server: do you know where we can find documentation on
2017 Jun 05
0
Gluster and NFS-Ganesha - cluster is down after reboot
Sorry, got sidetracked with invoicing etc. https://bitbucket.org/dismyne/gluster-ansibles/src/6df23803df43/ansible/files/?at=master <https://bitbucket.org/dismyne/gluster-ansibles/src/6df23803df43/ansible/files/?at=master> The .service files are the stuff going into SystemD, and they call the test-mounts.sh scripts. The playbook installing higher up in the directory > On 05 Jun 2017,
2017 Jul 11
2
Gluster native mount is really slow compared to nfs
Hi all, ? ? One more thing, we have 3 apps servers with the gluster on it, replicated on 3 different gluster nodes. (So the gluster nodes are app servers at the same time). We could actually almost work locally if we wouldn't need to have the same files on the 3 nodes and redundancy :) ? Initial cluster was created like this: ? gluster volume create www replica 3 transport tcp
2017 Jul 04
0
Slow write times to gluster disk
On 07/03/2017 09:01 PM, Pat Haley wrote: > > Hi Soumya, > > When I originally did the tests I ran tcpdump on the client. > > I have rerun the tests, doing tcpdump on the server > > tcpdump -i any -nnSs 0 host 172.16.1.121 -w /root/capture_nfsfail.pcap > > The results are in the same place > > http://mseas.mit.edu/download/phaley/GlusterUsers/TestNFSmount/ >