similar to: Filename completion and relative problems

Displaying 20 results from an estimated 2000 matches similar to: "Filename completion and relative problems"

2011 Jan 13
0
distribute-replicate setup GFS Client crashed
Hi there, Im running glusterfs version 3.1.0. The client crashed after sometime with below stack. 2011-01-13 08:33:49.230976] I [afr-common.c:2568:afr_notify] replicate-1: Subvolume 'distribute-1' came back up; going online. [2011-01-13 08:33:49.499909] I [afr-open.c:393:afr_openfd_sh] replicate-1: data self-heal triggered. path: /streaming/set3/work/reduce.12.1294902171.dplog.temp,
2010 Mar 15
1
Glusterfs 3.0.X crashed on Fedora 11
the glusterfs 3.0.X crashed on Fedora 12, it got buffer overflow, seems fine on Fedora 11 Name : fuse Arch : x86_64 Version : 2.8.1 Release : 4.fc12 Name : glibc Arch : x86_64 Version : 2.11.1 Release : 1 complete log: ====================================================================================================== [root at test_machine06 ~]# glusterfsd
2009 May 11
1
Problem of afr in glusterfs 2.0.0rc1
Hello: i had met the problem twice when i copy some files into the GFS space . i have five clients and two servers , when i copy files into /data which was GFS space on client A , the problem was appear. in the same path , A server can see the all files ,but B and C or D couldin't see the all files ,liks some files was missing ,but when i mount again ,the files was appear
2010 May 04
1
Posix warning : Access to ... is crossing device
I have a distributed/replicated setup with Glusterfs 3.0.2, that I'm testing on 4 servers, each with access to /mnt/gluster (which consists of all directories /mnt/data01 - data24) on each server. I'm using configs I built from volgen, but every time I access a file (via an 'ls -l') for the first time, I get all of these messages in my logs on each server: [2010-05-04 10:50:30] W
2008 Dec 10
1
df returns weird values
Hi, I'm starting to play with glusterfs, and I'm having a problem with the df output. The value seems to be wrong. (on the client) /var/mule-client$ du -sh 584K . /var/mule-client$ df -h /var/mule-client/ Filesystem Size Used Avail Use% Mounted on glusterfs 254G 209G 32G 88% /var/mule-client (on the server) /var/mule$ du -sh 584K . Is it a known
2009 Jun 26
0
Error when expand dht model volumes
HI all: I met a problem in expending dht volumes, i write in a dht storage directory untile it grew up to 90%,so i add four new volumes into the configur file. But when start again ,the data in directory some disappeared ,Why ??? Was there a special action before expending the volumes? my client cofigure file is this : volume client1 type protocol/client option transport-type
2010 Mar 04
1
[3.0.2] booster + unfsd failed
Hi list. I have been testing with glusterfs-3.0.2. glusterfs mount works well. unfsd on glusterfs mount point works well too. When using booster, unfsd realpath check failed. But ls util works well. I tried 3.0.0-git head source build but result was same. My System is Ubuntu 9.10 and using unfsd source from official gluster download site. Any comment appreciated!! - kpkim root at
2012 Mar 12
0
Data consistency with Gluster 3.2.5
I have set up a replicated, four-node gluster config for a web farm. The idea is that each web node is its own Gluster server, and will have its own copy of the entire web root locally. It then serves the cluster to itself via a mount. We're running it over dual GigE NICs bonded. The problem I am having is when we switch live traffic to nodes in the cluster, they almost immediately get
2018 May 08
1
volume start: gv01: failed: Quorum not met. Volume operation not allowed.
On 4/11/2018 11:54 AM, Alex K wrote: Hey Guy's, Returning to this topic, after disabling the the quorum: cluster.quorum-type: none cluster.server-quorum-type: none I've ran into a number of gluster errors (see below). I'm using gluster as the backend for my NFS storage. I have gluster running on two nodes, nfs01 and nfs02. It's mounted on /n on each host. The path /n is
2009 Jun 11
2
Problem with new version of GlusterFS-2.0.1 while copying.
Hi, I am having some problem with new version of GlusterFS-2.0.1 while copying using "apache" user. sudo -u apache cp -pvf zip/* test/ getting the message cp: getting attribute `trusted.glusterfs.afr.data-pending' of `zip/speccok1ma131231824637.zip': Operation not permitted `zip/speccok1ma131231824776.zip' -> `test/speccok1ma131231824776.zip' No problem while
2012 Jan 04
0
FUSE init failed
Hi, I'm having an issue using the GlusterFS native client. After doing a mount the filesystem appears mounted but any operation results in a Transport endpoint is not connected message gluster peer status and volume info don't complain. I've copied the mount log below which mentions an error at fuse_init. The kernel is based on 2.6.15 and FUSE api version is 7.3. I'm using
2009 Jun 24
0
Gluster-users Digest, Vol 14, Issue 34
HI: Was there a limit of servers which was used as storage in Gluster ? 2009-06-24 eagleeyes ???? gluster-users-request ????? 2009-06-24 03:00:42 ???? gluster-users ??? ??? Gluster-users Digest, Vol 14, Issue 34 Send Gluster-users mailing list submissions to gluster-users at gluster.org To subscribe or unsubscribe via the World Wide Web, visit
2008 Nov 20
1
My ignorance and Fuse (or glusterfs)
I have a very simple test setup of 2 servers each working as a glusterfs-server and glusterfs-client to the other in an afr capacity. The gluster-c and gluster-s both start up with no errors and are handshaking properly.. One one server, I get the expected behaviour: I touch a file in the export dir and it magically appears in the others mount point. On the other server however, the file
2008 Dec 09
1
File uploaded to webDAV server on GlusterFS AFR - ends up without xattr!
Hello list. I'm testing GlusterFS AFR mode as a solution for implementing a highly available webDAV file storage for our production environment. Whlie doing performance tests I've notticed a strange behavior: the files which are uploaded via a webDAV server, end up without extended attributes, which removes the ability to self-heal. The set up is a simple testing environment with 2
2008 Nov 04
1
fuse_setlk_cbk error
I'm building a two node cluster to run vserver systems on. I've setup glusterfs with this config: # node a volume data-posix type storage/posix option directory /export/cluster end-volume volume data1 type features/posix-locks subvolumes data-posix end-volume volume data2 type protocol/client option transport-type tcp/client option remote-host
2009 Jun 24
2
Limit of Glusterfs help
HI: Was there a limit of servers which was used as storage in Gluster ? 2009-06-24 eagleeyes ???? gluster-users-request ????? 2009-06-24 03:00:42 ???? gluster-users ??? ??? Gluster-users Digest, Vol 14, Issue 34 Send Gluster-users mailing list submissions to gluster-users at gluster.org To subscribe or unsubscribe via the World Wide Web, visit
2013 Feb 26
0
Replicated Volume Crashed
Hi, I have a gluster volume that consists of 22Bricks and includes a single folder with 3.6 Million files. Yesterday the volume crashed and turned out to be completely unresposible and I was forced to perform a hard reboot on all gluster servers because they were not able to execute a reboot command issued by the shell because they were that heavy overloaded. Each gluster server has 12 CPU cores
2009 Jun 11
2
Issue with files on glusterfs becoming unreadable.
elbert at host1:~$ dpkg -l|grep glusterfs ii glusterfs-client 1.3.8-0pre2 GlusterFS fuse client ii glusterfs-server 1.3.8-0pre2 GlusterFS fuse server ii libglusterfs0 1.3.8-0pre2 GlusterFS libraries and translator modules I have 2 hosts set up to use AFR with
2017 Jun 14
0
No NFS connection due to GlusterFS CPU load
When executing the load test with the FIO tool, execute the following job from the client When executed, the load of 2 cores is high for the CPU. Up to 100%. At that time, if another client is performing NFS mounting, the df command I can not connect NFS without coming back. The log will continue to be output below. I believe that if the CPU utilization is distributed, the load will be eliminated.
2018 Apr 09
0
volume start: gv01: failed: Quorum not met. Volume operation not allowed.
Hi, You need 3 nodes at least to have quorum enabled. In 2 node setup you need to disable quorum so as to be able to still use the volume when one of the nodes go down. On Mon, Apr 9, 2018, 09:02 TomK <tomkcpr at mdevsys.com> wrote: > Hey All, > > In a two node glusterfs setup, with one node down, can't use the second > node to mount the volume. I understand this is