Nicola Petracchi
2016-May-11 09:31 UTC
[Gluster-users] Number of Bricks 1 x (2 + 1) = 3 and arbiter disappearing
Hi, thx for the quick reply I have this configuration Description: Ubuntu 14.04.4 LTS Release: 14.04 Codename: trusty Linux 3.13.0-86-generic #130-Ubuntu SMP ppa:gluster/glusterfs-3.7 ii glusterfs-client 3.7.11-ubuntu1~trusty1 amd64 clustered file-system (client package) ii glusterfs-common 3.7.11-ubuntu1~trusty1 amd64 GlusterFS common libraries and translator modules ii glusterfs-server 3.7.11-ubuntu1~trusty1 amd64 clustered file-system (server package) N 2016-05-11 10:57 GMT+02:00 Ravishankar N <ravishankar at redhat.com>:> On 05/11/2016 02:17 PM, Nicola Petracchi wrote: >> >> Hello, I have deployed a gluster configuration of this type: >> Number of Bricks: 1 x (2 + 1) = 3 >> 2 data + 1 arbiter >> >> After deploy, all machines were reporting the correct volume info >> informations: >> Volume Name: gvol0 >> Type: Replicate >> Status: Started >> Number of Bricks: 1 x (2 + 1) = 3 >> Transport-type: tcp >> Bricks: >> Brick1: pc01:/var/lib/gvol0/brick1 >> Brick2: pc02:/var/lib/gvol0/brick2 >> Brick3: pcgw:/var/lib/gvol0/brickgw (arbiter) >> >> Then, I had to restart gluster service on brick2, after that, only on >> that machine the volume info was: >> Volume Name: gvol0 >> Type: Replicate >> Status: Started >> Number of Bricks: 1 x 3 = 3 >> Transport-type: tcp >> Bricks: >> Brick1: pc01:/var/lib/gvol0/brick1 >> Brick2: pc02:/var/lib/gvol0/brick2 >> Brick3: pcgw:/var/lib/gvol0/brickgw >> others nodes were still reporting the original volume info informations. >> >> This thing appears strange to me, since I suppose all the members of a >> cluster should be consistent on output of informations of a "shared" >> resource. >> >> Btw, later on I had to reboot the arbiter server and again, on that >> server: >> Volume Name: gvol0 >> Type: Replicate >> Status: Started >> Number of Bricks: 1 x 3 = 3 >> Transport-type: tcp >> Bricks: >> Brick1: pc01:/var/lib/gvol0/brick1 >> Brick2: pc02:/var/lib/gvol0/brick2 >> Brick3: pcgw:/var/lib/gvol0/brickgw >> >> Only one node, the one never rebooted is stating that brick3 is an >> arbiter, aside this information the arbiter node is working as an >> arbiter since the data folder contains empty files and is updated with >> the others. >> >> Is this a situation to solve? an issue? > > What version of gluster are you running? This was bug that got fixed > (http://review.gluster.org/12479) in 3.7.7. I would advise you to use the > latest release 3.7.11 if you are trying out arbiter. > Hope this helps, > Ravi >> >> Considerations are appreciated. >> Regards >> N >> _______________________________________________ >> Gluster-users mailing list >> Gluster-users at gluster.org >> http://www.gluster.org/mailman/listinfo/gluster-users > > >-- Nicola Petracchi Senior Developer Skype: nicola.petracchi Mail: nicola.petracchi at digiville.it Cell: +39 338 2945627 DigiVille S.r.l. Performance & Classified Media www.digiville.it
Ravishankar N
2016-May-11 09:56 UTC
[Gluster-users] Number of Bricks 1 x (2 + 1) = 3 and arbiter disappearing
On 05/11/2016 03:01 PM, Nicola Petracchi wrote:> Hi, thx for the quick reply > > I have this configuration > Description: Ubuntu 14.04.4 LTS > Release: 14.04 > Codename: trusty > Linux 3.13.0-86-generic #130-Ubuntu SMP > > ppa:gluster/glusterfs-3.7 > > ii glusterfs-client 3.7.11-ubuntu1~trusty1 > amd64 clustered file-system (client package) > ii glusterfs-common 3.7.11-ubuntu1~trusty1 > amd64 GlusterFS common libraries and translator > modules > ii glusterfs-server 3.7.11-ubuntu1~trusty1 > amd64 clustered file-system (server package)Okay so you are running 3.7.11. 1. Was your volume created when you were running an older version? 2.What is the glusterd op-version on all nodes? (`cat /var/lib/glusterd/glusterd.info|grep operating-version`) 3.Does /var/lib/glusterd/vols/gvo10/info have 'arbiter_count=1' ? -Ravi> > N > > 2016-05-11 10:57 GMT+02:00 Ravishankar N <ravishankar at redhat.com>: >> On 05/11/2016 02:17 PM, Nicola Petracchi wrote: >>> Hello, I have deployed a gluster configuration of this type: >>> Number of Bricks: 1 x (2 + 1) = 3 >>> 2 data + 1 arbiter >>> >>> After deploy, all machines were reporting the correct volume info >>> informations: >>> Volume Name: gvol0 >>> Type: Replicate >>> Status: Started >>> Number of Bricks: 1 x (2 + 1) = 3 >>> Transport-type: tcp >>> Bricks: >>> Brick1: pc01:/var/lib/gvol0/brick1 >>> Brick2: pc02:/var/lib/gvol0/brick2 >>> Brick3: pcgw:/var/lib/gvol0/brickgw (arbiter) >>> >>> Then, I had to restart gluster service on brick2, after that, only on >>> that machine the volume info was: >>> Volume Name: gvol0 >>> Type: Replicate >>> Status: Started >>> Number of Bricks: 1 x 3 = 3 >>> Transport-type: tcp >>> Bricks: >>> Brick1: pc01:/var/lib/gvol0/brick1 >>> Brick2: pc02:/var/lib/gvol0/brick2 >>> Brick3: pcgw:/var/lib/gvol0/brickgw >>> others nodes were still reporting the original volume info informations. >>> >>> This thing appears strange to me, since I suppose all the members of a >>> cluster should be consistent on output of informations of a "shared" >>> resource. >>> >>> Btw, later on I had to reboot the arbiter server and again, on that >>> server: >>> Volume Name: gvol0 >>> Type: Replicate >>> Status: Started >>> Number of Bricks: 1 x 3 = 3 >>> Transport-type: tcp >>> Bricks: >>> Brick1: pc01:/var/lib/gvol0/brick1 >>> Brick2: pc02:/var/lib/gvol0/brick2 >>> Brick3: pcgw:/var/lib/gvol0/brickgw >>> >>> Only one node, the one never rebooted is stating that brick3 is an >>> arbiter, aside this information the arbiter node is working as an >>> arbiter since the data folder contains empty files and is updated with >>> the others. >>> >>> Is this a situation to solve? an issue? >> What version of gluster are you running? This was bug that got fixed >> (http://review.gluster.org/12479) in 3.7.7. I would advise you to use the >> latest release 3.7.11 if you are trying out arbiter. >> Hope this helps, >> Ravi >>> Considerations are appreciated. >>> Regards >>> N >>> _______________________________________________ >>> Gluster-users mailing list >>> Gluster-users at gluster.org >>> http://www.gluster.org/mailman/listinfo/gluster-users >> >> > >