Displaying 20 results from an estimated 4000 matches similar to: "[3.11.2] Bricks disconnect from gluster with 0-transport: EPOLLERR"
2017 Sep 13
1
[3.11.2] Bricks disconnect from gluster with 0-transport: EPOLLERR
I ran into something like this in 3.10.4 and filed two bugs for it:
https://bugzilla.redhat.com/show_bug.cgi?id=1491059
https://bugzilla.redhat.com/show_bug.cgi?id=1491060
Please see the above bugs for full detail.
In summary, my issue was related to glusterd's pid handling of pid files
when is starts self-heal and bricks. The issues are:
a. brick pid file leaves stale pid and brick fails
2018 Mar 20
0
brick processes not starting
Hi all,
our systems have suffered a node failure in a replica three setup.
The node needed a complete reinstall. I followed the RH guide to
replace a host with the same hostname
(https://access.redhat.com/documentation/en-us/red_hat_gluster_storage/3/html/administration_guide/sect-replacing_hosts).
The machine has the same OS (CentOS 7). The new machine got a minor
version number newer gluster
2018 Mar 21
0
Brick process not starting after reinstall
Could you share the following information:
1. gluster --version
2. output of gluster volume status
3. glusterd log and all brick log files from the node where bricks didn't
come up.
On Wed, Mar 21, 2018 at 12:35 PM, Richard Neuboeck <hawk at tbi.univie.ac.at>
wrote:
> Hi all,
>
> our systems have suffered a host failure in a replica three setup.
> The host needed a
2018 Mar 21
2
Brick process not starting after reinstall
Hi all,
our systems have suffered a host failure in a replica three setup.
The host needed a complete reinstall. I followed the RH guide to
'replace a host with the same hostname'
(https://access.redhat.com/documentation/en-us/red_hat_gluster_storage/3/html/administration_guide/sect-replacing_hosts).
The machine has the same OS (CentOS 7). The new machine got a minor
version number newer
2017 Aug 21
1
Glusterd not working with systemd in redhat 7
Hi!
Please see bellow. Note that web1.dasilva.network is the address of the
local machine where one of the bricks is installed and that ties to mount.
[2017-08-20 20:30:40.359236] I [MSGID: 100030] [glusterfsd.c:2476:main]
0-/usr/sbin/glusterd: Started running /usr/sbin/glusterd version 3.11.2
(args: /usr/sbin/glusterd -p /var/run/glusterd.pid)
[2017-08-20 20:30:40.973249] I [MSGID: 106478]
2017 Aug 21
0
Glusterd not working with systemd in redhat 7
On Mon, Aug 21, 2017 at 2:49 AM, Cesar da Silva <thunderlight1 at gmail.com>
wrote:
> Hi!
> I am having same issue but I am running Ubuntu v16.04.
> It does not mount during boot, but works if I mount it manually. I am
> running the Gluster-server on the same machines (3 machines)
> Here is the /tc/fstab file
>
> /dev/sdb1 /data/gluster ext4 defaults 0 0
>
>
2017 Aug 20
2
Glusterd not working with systemd in redhat 7
Hi!
I am having same issue but I am running Ubuntu v16.04.
It does not mount during boot, but works if I mount it manually. I am
running the Gluster-server on the same machines (3 machines)
Here is the /tc/fstab file
/dev/sdb1 /data/gluster ext4 defaults 0 0
web1.dasilva.network:/www /mnt/glusterfs/www glusterfs
defaults,_netdev,log-level=debug,log-file=/var/log/gluster.log 0 0
2018 Feb 06
0
strange hostname issue on volume create command with famous Peer in Cluster state error message
Did you do gluster peer probe? Check out the documentation:
http://docs.gluster.org/en/latest/Administrator%20Guide/Storage%20Pools/
On Tue, Feb 6, 2018 at 5:01 PM, Ercan Aydo?an <ercan.aydogan at gmail.com> wrote:
> Hello,
>
> i installed glusterfs 3.11.3 version 3 nodes ubuntu 16.04 machine. All
> machines have same /etc/hosts.
>
> node1 hostname
> pri.ostechnix.lan
2017 Aug 18
1
Is transport=rdma tested with "stripe"?
On Wed, Aug 16, 2017 at 4:44 PM, Hatazaki, Takao <takao.hatazaki at hpe.com> wrote:
>> Note that "stripe" is not tested much and practically unmaintained.
>
> Ah, this was what I suspected. Understood. I'll be happy with "shard".
>
> Having said that, "stripe" works fine with transport=tcp. The failure reproduces with just 2 RDMA servers
2018 Feb 06
0
strange hostname issue on volume create command with famous Peer in Cluster state error message
I'm guessing there's something wrong w.r.t address resolution on node 1.
>From the logs it's quite clear to me that node 1 is unable to resolve the
address configured in /etc/hosts where as the other nodes do. Could you
paste the gluster peer status output from all the nodes?
Also can you please check if you're able to ping "pri.ostechnix.lan" from
node1 only? Does
2018 Feb 06
1
strange hostname issue on volume create command with famous Peer in Cluster state error message
I changed /etc/hosts
127.0.0.1 pri.ostechnix.lan pri
51.15.90.60 sec.ostechnix.lan sec
163.172.151.120 third.ostechnix.lan third
on every node matching hostname to 127.0.0.1
then
root at pri:~# apt-get purge glusterfs-server
root at pri:~# rm -rf /var/lib/glusterd/
root at pri:~# rm -rf /var/log/glusterfs/
root at pri:~# apt-get install glusterfs-server
root at pri:~#
2018 Feb 08
2
Thousands of EPOLLERR - disconnecting now
Hello
I have a large cluster in which every node is logging:
I [socket.c:2474:socket_event_handler] 0-transport: EPOLLERR -
disconnecting now
At a rate of of around 4 or 5 per second per node, which is adding up to a
lot of messages. This seems to happen while my cluster is idle.
-------------- next part --------------
An HTML attachment was scrubbed...
URL:
2017 Aug 16
0
Is transport=rdma tested with "stripe"?
> Note that "stripe" is not tested much and practically unmaintained.
Ah, this was what I suspected. Understood. I'll be happy with "shard".
Having said that, "stripe" works fine with transport=tcp. The failure reproduces with just 2 RDMA servers (with InfiniBand), one of those acts also as a client.
I looked into logs. I paste lengthy logs below with
2018 Feb 06
5
strange hostname issue on volume create command with famous Peer in Cluster state error message
Hello,
i installed glusterfs 3.11.3 version 3 nodes ubuntu 16.04 machine. All machines have same /etc/hosts.
node1 hostname
pri.ostechnix.lan
node2 hostname
sec.ostechnix.lan
node2 hostname
third.ostechnix.lan
51.15.77.14 pri.ostechnix.lan pri
51.15.90.60 sec.ostechnix.lan sec
163.172.151.120 third.ostechnix.lan third
volume create command is
root at
2018 Feb 08
0
Thousands of EPOLLERR - disconnecting now
On Thu, Feb 8, 2018 at 2:04 PM, Gino Lisignoli <glisignoli at gmail.com> wrote:
> Hello
>
> I have a large cluster in which every node is logging:
>
> I [socket.c:2474:socket_event_handler] 0-transport: EPOLLERR -
> disconnecting now
>
> At a rate of of around 4 or 5 per second per node, which is adding up to a
> lot of messages. This seems to happen while my
2017 Jun 15
1
peer probe failures
Hi,
I'm having a similar issue, were you able to solve it?
Thanks.
Hey all,
I've got a strange problem going on here. I've installed glusterfs-server
on ubuntu 16.04:
glusterfs-client/xenial,now 3.7.6-1ubuntu1 amd64 [installed,automatic]
glusterfs-common/xenial,now 3.7.6-1ubuntu1 amd64 [installed,automatic]
glusterfs-server/xenial,now 3.7.6-1ubuntu1 amd64 [installed]
I can
2018 Jan 16
0
Using the host name of the volume, its related commands can become very slow
On Mon, Jan 15, 2018 at 6:30 PM, ?? <chenxi at shudun.com> wrote:
> Using the host name of the volume, its related gluster commands can become
> very slow .For example,create,start,stop volume,nfs related commands. and
> some time And in some cases, the command will return Error : Request timed
> out
> but If using ip address to create the volume. The volume all gluster
>
2018 Feb 07
2
Ip based peer probe volume create error
On 8/02/2018 4:45 AM, Gaurav Yadav wrote:
> After seeing command history, I could see that you have 3 nodes, and
> firstly you are peer probing 51.15.90.60? and 163.172.151.120 from?
> 51.15.77.14
> So here itself you have 3 node cluster, after all this you are going
> on node 2 and again peer probing 51.15.77.14.
> ?Ideally it should work, with above steps, but due to some
2018 Jan 15
2
Using the host name of the volume, its related commands can become very slow
Using the host name of the volume, its related gluster commands can become very slow .For example,create,start,stop volume,nfs related commands. and some time And in some cases, the command will return Error : Request timed out
but If using ip address to create the volume. The volume all gluster commands are normal.
I have configured /etc/hosts correctly,Because,SSH can normally use the
2017 Sep 03
0
Glusterd proccess hangs on reboot
i usually change event threads to 4. But those logs are from a default
installation.
On Sun, Sep 3, 2017 at 9:52 PM, Ben Turner <bturner at redhat.com> wrote:
> ----- Original Message -----
>> From: "Ben Turner" <bturner at redhat.com>
>> To: "Serkan ?oban" <cobanserkan at gmail.com>
>> Cc: "Gluster Users" <gluster-users at