similar to: 'ERROR: parsing the volfile failed' on fresh install

Displaying 20 results from an estimated 7000 matches similar to: "'ERROR: parsing the volfile failed' on fresh install"

2017 Dec 13
1
'ERROR: parsing the volfile failed' on fresh install
Hi all, I?m trying out gluster by following the Quick Start guide on two fresh installs of Ubuntu 16.04. One one node I was able to install and start gluster just fine. One the other node I am running into the following: $ sudo service glusterd start Job for glusterd.service failed because the control process exited with error code. See "systemctl status glusterd.service" and
2013 Jul 07
1
Getting ERROR: parsing the volfile failed (No such file or directory) when starting glusterd on Fedora 19
I don't get this. I am using a freshly installed copy of Fedora 19 and starting up glusterd for the first time. The goal is to have a replicated directory on two systems. But for right now, I can't even start up the glusterd daemon right out of the box. Trying to follow the Quick Start directions at http://gluster.org/community/documentation/index.php/QuickStart is, well, challenging.
2017 Oct 19
1
gluster + synology
On Wed, Oct 18, 2017 at 12:45 PM, Alex Chekholko <alex at calicolabs.com> wrote: > In theory, you can run GlusterFS on a Synology box, as it is "just a Linux > box". In practice, you might be the first person to ever try it. > Not sure if this would be the first attempt. Synology seems to bundle glusterfs in some form [1]. Regards, Vijay [1]
2017 Oct 18
0
gluster + synology
In theory, you can run GlusterFS on a Synology box, as it is "just a Linux box". In practice, you might be the first person to ever try it. On Tue, Oct 17, 2017 at 8:45 PM, Ben Mabey <ben.mabey at recursionpharma.com> wrote: > Hi, > Does anyone have any experience of using Synology NAS servers as bricks in > a gluster setup? The ops team at my work prefers Synology
2018 May 15
0
glusterfs 3.13 repo unavailable and downgrade to 3.12.9 fails
You can still get them from https://buildlogs.centos.org/centos/7/storage/x86_64/gluster-3.13/ (I don't know how much longer they'll be there. I suggest you copy them if you think you're going to need them in the future.) n 05/15/2018 04:58 AM, Davide Obbi wrote: > hi, > > i noticed that this repo for glusterfs 3.13 does not exists anymore at: > >
2018 May 15
1
[External] Re: glusterfs 3.13 repo unavailable and downgrade to 3.12.9 fails
Thanks Kaleb, any chance i can make the node working after the downgrade? thanks On Tue, May 15, 2018 at 2:02 PM, Kaleb S. KEITHLEY <kkeithle at redhat.com> wrote: > > You can still get them from > https://buildlogs.centos.org/centos/7/storage/x86_64/gluster-3.13/ > > (I don't know how much longer they'll be there. I suggest you copy them > if you think
2018 May 15
2
glusterfs 3.13 repo unavailable and downgrade to 3.12.9 fails
hi, i noticed that this repo for glusterfs 3.13 does not exists anymore at: http://mirror.centos.org/centos/7/storage/x86_64/ i knew was not going to be long term supported however the downgrade to 3.12 breaks the server node i believe the issue is with: *[2018-05-15 08:54:39.981101] E [MSGID: 101019] [xlator.c:503:xlator_init] 0-management: Initialization of volume 'management'
2018 May 15
1
[External] Re: glusterfs 3.13 repo unavailable and downgrade to 3.12.9 fails
On 05/15/2018 08:08 AM, Davide Obbi wrote: > Thanks Kaleb, > > any chance i can make the node working after the downgrade? > thanks Without knowing what doesn't work, I'll go out on a limb and guess that it's an op-version problem. Shut down your 3.13 nodes, change their op-version to one of the valid 3.12 op-versions (e.g. 31203) and restart. Then the 3.12 nodes should
2012 Oct 05
0
No subject
# gluster --version glusterfs 3.3.1 built on Oct 11 2012 22:01:05 # gluster volume info Volume Name: gdata Type: Distribute Volume ID: eccc3a90-212d-4563-ae8d-10a77758738d Status: Started Number of Bricks: 3 Transport-type: tcp Bricks: Brick1: gluster-0-0:/mseas-data-0-0 Brick2: gluster-0-1:/mseas-data-0-1 Brick3: gluster-data:/data [root at mseas-data ~]# ps -ef | grep gluster root 2783
2023 Feb 23
1
Big problems after update to 9.6
Hello, We have a cluster with two nodes, "sg" and "br", which were running GlusterFS 9.1, installed via the Ubuntu package manager. We updated the Ubuntu packages on "sg" to version 9.6, and now have big problems. The "br" node is still on version 9.1. Running "gluster volume status" on either host gives "Error : Request timed out". On
2014 Mar 04
1
glusterd service fails to start from AWS AMI
Hello all. I have a working replica 2 cluster (4 nodes) up and running happily over Amazon EC2. My end goal is to create AMIs of each machine and then quickly reproduce the same, but new, cluster from those AMIs. Essentially, I'd like a cluster "template". -Assigned original instances' Elastic IPs to new machines to reduce resolution issues. -Passwordless SSH works on initial
2023 Feb 24
1
Big problems after update to 9.6
Hi David, It seems like a network issue to me, As it's unable to connect the other node and getting timeout. Few things you can check- * Check the /etc/hosts file on both the servers and make sure it has the correct IP of the other node. * Are you binding gluster on any specific IP, which is changed after your update. * Check if you can access port 24007 from the other host. If
2017 Oct 18
0
Gluster processes remaining after stopping glusterd
On Tue, Oct 17, 2017 at 3:28 PM, ismael mondiu <mondiu at hotmail.com> wrote: > Hi, > > I noticed that when i stop my gluster server via systemctl stop glusterd > command , one glusterfs process is still up. > > Which is the correct way to stop all gluster processes in my host? > Stopping glusterd service doesn't bring down any other services than glusterd process.
2017 Aug 06
0
State: Peer Rejected (Connected)
On 2017? 08? 06? 15:59, mabi wrote: > Hi, > > I have a 3 nodes replica (including arbiter) volume with GlusterFS > 3.8.11 and this night one of my nodes (node1) had an out of memory for > some unknown reason and as such the Linux OOM killer has killed the > glusterd and glusterfs process. I restarted the glusterd process but > now that node is in "Peer Rejected"
2011 Jul 08
1
Possible to bind to multiple addresses?
I am trying to run GlusterFS on only my internal interfaces. I have setup two bricks and have a replicated volume that is started. Everything works fine when I run with no transport.socket.bind-address defined in the /etc/glusterfs/glusterd.vol file, but when I add it I get: Transport endpoint is not connected My configuration looks like this: volume management type mgmt/glusterd
2017 Oct 17
2
Gluster processes remaining after stopping glusterd
Hi, I noticed that when i stop my gluster server via systemctl stop glusterd command , one glusterfs process is still up. Which is the correct way to stop all gluster processes in my host? That's we see after run the command: *************************************************************************************************** [root at xxxxxx ~]# ps -ef | grep -i glu root 1825 1
2017 Aug 06
1
State: Peer Rejected (Connected)
Hi Ji-Hyeon, Thanks to your help I could find out the problematic file. This would be the quota file of my volume it has a different checksum on node1 whereas node2 and arbiternode have the same checksum. This is expected as I had issues which my quota file and had to fix it manually with a script (more details on this mailing list in a previous post) and I only did that on node1. So what I now
2024 Feb 05
1
Graceful shutdown doesn't stop all Gluster processes
Hello Everyone, I am using GlusterFS 9.4, and whenever we use the systemctl command to stop the Gluster server, it leaves many Gluster processes running. So, I just want to check how to shut down the Gluster server in a graceful manner. Is there any specific sequence or trick I need to follow? Currently, I am using the following command: [root at master2 ~]# systemctl stop glusterd.service
2017 Nov 08
2
Gluster clients can't see directories that exist or are created within a mounted volume, but can enter them.
On 8 November 2017 at 02:47, Sam McLeod <mailinglists at smcleod.net> wrote: > > On 6 Nov 2017, at 3:32 pm, Laura Bailey <lbailey at redhat.com> wrote: > > Do the users have permission to see/interact with the directories, in > addition to the files? > > > Yes, full access to directories and files. > Also testing using the root user. > > > On Mon,
2017 May 09
1
Empty info file preventing glusterd from starting
Hi Atin/Team, We are using gluster-3.7.6 with setup of two brick and while restart of system I have seen that the glusterd daemon is getting failed from start. At the time of analyzing the logs from etc-glusterfs.......log file I have received the below logs [2017-05-06 03:33:39.798087] I [MSGID: 100030] [glusterfsd.c:2348:main] 0-/usr/sbin/glusterd: Started running /usr/sbin/glusterd version