Kingsley
2015-Aug-10 09:53 UTC
[Gluster-users] volume not working after yum update - gluster 3.6.3
Hi, We have a 4 way replicated volume using gluster 3.6.3 on CentOS 7. Over the weekend I did a yum update on each of the bricks in turn, but now when clients (using fuse mounts) try to access the volume, it hangs. Gluster itself wasn't updated (we've disabled that repo so that we keep to 3.6.3 for now). This was what I did: * on first brick, "yum update" * reboot brick * watch "gluster volume status" on another brick and wait for it to say all 4 bricks are online before proceeding to update the next brick I was expecting the clients might pause 30 seconds while they notice a brick is offline, but then recover. I've tried re-mounting clients, but that hasn't helped. I can't see much data in any of the log files. I've tried "gluster volume heal callrec" but it doesn't seem to have helped. What shall I do next? I've pasted some stuff below in case any of it helps. Cheers, Kingsley. [root at gluster1b-1 ~]# gluster volume info callrec Volume Name: callrec Type: Replicate Volume ID: a39830b7-eddb-4061-b381-39411274131a Status: Started Number of Bricks: 1 x 4 = 4 Transport-type: tcp Bricks: Brick1: gluster1a-1:/data/brick/callrec Brick2: gluster1b-1:/data/brick/callrec Brick3: gluster2a-1:/data/brick/callrec Brick4: gluster2b-1:/data/brick/callrec Options Reconfigured: performance.flush-behind: off [root at gluster1b-1 ~]# [root at gluster1b-1 ~]# gluster volume status callrec Status of volume: callrec Gluster process Port Online Pid ------------------------------------------------------------------------------ Brick gluster1a-1:/data/brick/callrec 49153 Y 6803 Brick gluster1b-1:/data/brick/callrec 49153 Y 2614 Brick gluster2a-1:/data/brick/callrec 49153 Y 2645 Brick gluster2b-1:/data/brick/callrec 49153 Y 4325 NFS Server on localhost 2049 Y 2769 Self-heal Daemon on localhost N/A Y 2789 NFS Server on gluster2a-1 2049 Y 2857 Self-heal Daemon on gluster2a-1 N/A Y 2814 NFS Server on 88.151.41.100 2049 Y 6833 Self-heal Daemon on 88.151.41.100 N/A Y 6824 NFS Server on gluster2b-1 2049 Y 4428 Self-heal Daemon on gluster2b-1 N/A Y 4387 Task Status of Volume callrec ------------------------------------------------------------------------------ There are no active volume tasks [root at gluster1b-1 ~]# [root at gluster1b-1 ~]# gluster volume heal callrec info Brick gluster1a-1.dns99.co.uk:/data/brick/callrec/ /to_process - Possibly undergoing heal Number of entries: 1 Brick gluster1b-1.dns99.co.uk:/data/brick/callrec/ Number of entries: 0 Brick gluster2a-1.dns99.co.uk:/data/brick/callrec/ /to_process - Possibly undergoing heal Number of entries: 1 Brick gluster2b-1.dns99.co.uk:/data/brick/callrec/ Number of entries: 0 [root at gluster1b-1 ~]#
Kingsley
2015-Aug-10 10:05 UTC
[Gluster-users] volume not working after yum update - gluster 3.6.3
Sorry for the blind panic - restarting the volume seems to have fixed it. But then my next question - why is this necessary? Surely it undermines the whole point of a high availability system? Cheers, Kingsley. On Mon, 2015-08-10 at 10:53 +0100, Kingsley wrote:> Hi, > > We have a 4 way replicated volume using gluster 3.6.3 on CentOS 7. > > Over the weekend I did a yum update on each of the bricks in turn, but > now when clients (using fuse mounts) try to access the volume, it hangs. > Gluster itself wasn't updated (we've disabled that repo so that we keep > to 3.6.3 for now). > > This was what I did: > > * on first brick, "yum update" > * reboot brick > * watch "gluster volume status" on another brick and wait for it > to say all 4 bricks are online before proceeding to update the > next brick > > I was expecting the clients might pause 30 seconds while they notice a > brick is offline, but then recover. > > I've tried re-mounting clients, but that hasn't helped. > > I can't see much data in any of the log files. > > I've tried "gluster volume heal callrec" but it doesn't seem to have > helped. > > What shall I do next? > > I've pasted some stuff below in case any of it helps. > > Cheers, > Kingsley. > > [root at gluster1b-1 ~]# gluster volume info callrec > > Volume Name: callrec > Type: Replicate > Volume ID: a39830b7-eddb-4061-b381-39411274131a > Status: Started > Number of Bricks: 1 x 4 = 4 > Transport-type: tcp > Bricks: > Brick1: gluster1a-1:/data/brick/callrec > Brick2: gluster1b-1:/data/brick/callrec > Brick3: gluster2a-1:/data/brick/callrec > Brick4: gluster2b-1:/data/brick/callrec > Options Reconfigured: > performance.flush-behind: off > [root at gluster1b-1 ~]# > > > [root at gluster1b-1 ~]# gluster volume status callrec > Status of volume: callrec > Gluster process Port Online Pid > ------------------------------------------------------------------------------ > Brick gluster1a-1:/data/brick/callrec 49153 Y 6803 > Brick gluster1b-1:/data/brick/callrec 49153 Y 2614 > Brick gluster2a-1:/data/brick/callrec 49153 Y 2645 > Brick gluster2b-1:/data/brick/callrec 49153 Y 4325 > NFS Server on localhost 2049 Y 2769 > Self-heal Daemon on localhost N/A Y 2789 > NFS Server on gluster2a-1 2049 Y 2857 > Self-heal Daemon on gluster2a-1 N/A Y 2814 > NFS Server on 88.151.41.100 2049 Y 6833 > Self-heal Daemon on 88.151.41.100 N/A Y 6824 > NFS Server on gluster2b-1 2049 Y 4428 > Self-heal Daemon on gluster2b-1 N/A Y 4387 > > Task Status of Volume callrec > ------------------------------------------------------------------------------ > There are no active volume tasks > > [root at gluster1b-1 ~]# > > > [root at gluster1b-1 ~]# gluster volume heal callrec info > Brick gluster1a-1.dns99.co.uk:/data/brick/callrec/ > /to_process - Possibly undergoing heal > > Number of entries: 1 > > Brick gluster1b-1.dns99.co.uk:/data/brick/callrec/ > Number of entries: 0 > > Brick gluster2a-1.dns99.co.uk:/data/brick/callrec/ > /to_process - Possibly undergoing heal > > Number of entries: 1 > > Brick gluster2b-1.dns99.co.uk:/data/brick/callrec/ > Number of entries: 0 > > [root at gluster1b-1 ~]# > > > _______________________________________________ > Gluster-users mailing list > Gluster-users at gluster.org > http://www.gluster.org/mailman/listinfo/gluster-users >