Steve Dainard
2016-Aug-22 22:47 UTC
[Gluster-users] Upgrade 3.7.6 -> 3.7.13 one gluster server disconnected 1 of 3 volumes
About 5 hours after upgrading gluster 3.7.6 -> 3.7.13 on Centos 7, one of my gluster servers disconnected its volume. The other two volumes this host serves were not affected. # gluster volume status storage Status of volume: storage Gluster process TCP Port RDMA Port Online Pid ------------------------------------------------------------------------------ Brick 10.0.231.50:/mnt/raid6-storage/storage 49159 0 Y 30743 Brick 10.0.231.51:/mnt/raid6-storage/storage 49159 0 Y 676 *Brick 10.0.231.52:/mnt/raid6-storage/storag**e N/A N/A N N/A * Brick 10.0.231.53:/mnt/raid6-storage/storage 49154 0 Y 10253 Brick 10.0.231.54:/mnt/raid6-storage/storage 49153 0 Y 2792 Brick 10.0.231.55:/mnt/raid6-storage/storage 49153 0 Y 13590 Brick 10.0.231.56:/mnt/raid6-storage/storage 49152 0 Y 9281 NFS Server on localhost 2049 0 Y 30775 Quota Daemon on localhost N/A N/A Y 30781 NFS Server on 10.0.231.54 2049 0 Y 2817 Quota Daemon on 10.0.231.54 N/A N/A Y 2824 NFS Server on 10.0.231.51 2049 0 Y 710 Quota Daemon on 10.0.231.51 N/A N/A Y 719 NFS Server on 10.0.231.52 2049 0 Y 9090 Quota Daemon on 10.0.231.52 N/A N/A Y 9098 NFS Server on 10.0.231.55 2049 0 Y 13611 Quota Daemon on 10.0.231.55 N/A N/A Y 13619 NFS Server on 10.0.231.56 2049 0 Y 9303 Quota Daemon on 10.0.231.56 N/A N/A Y 9310 NFS Server on 10.0.231.53 2049 0 Y 26304 Quota Daemon on 10.0.231.53 N/A N/A Y 26320 Task Status of Volume storage ------------------------------------------------------------------------------ There are no active volume tasks I see lots of logs related to trashcan (failed [file exists]), set xattrs (failed [no such file or directory]), quota (invalid arguments) in the brick logs, which I enabled as a feature after the upgrade this morning. After restarting glusterd on that host, the volume came back online. I've attached logs from that host if someone can take a look. # gluster volume info storage Volume Name: storage Type: Distribute Volume ID: 6f95525a-94d7-4174-bac4-e1a18fe010a2 Status: Started Number of Bricks: 7 Transport-type: tcp Bricks: Brick1: 10.0.231.50:/mnt/raid6-storage/storage Brick2: 10.0.231.51:/mnt/raid6-storage/storage Brick3: 10.0.231.52:/mnt/raid6-storage/storage Brick4: 10.0.231.53:/mnt/raid6-storage/storage Brick5: 10.0.231.54:/mnt/raid6-storage/storage Brick6: 10.0.231.55:/mnt/raid6-storage/storage Brick7: 10.0.231.56:/mnt/raid6-storage/storage Options Reconfigured: nfs.disable: no features.trash-max-filesize: 1GB features.trash: on features.quota-deem-statfs: on features.inode-quota: on features.quota: on performance.readdir-ahead: on # rpm -qa |grep glusterfs glusterfs-fuse-3.7.13-1.el7.x86_64 glusterfs-cli-3.7.13-1.el7.x86_64 glusterfs-3.7.13-1.el7.x86_64 glusterfs-server-3.7.13-1.el7.x86_64 glusterfs-api-3.7.13-1.el7.x86_64 glusterfs-libs-3.7.13-1.el7.x86_64 glusterfs-client-xlators-3.7.13-1.el7.x86_64 Thanks, Steve -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20160822/bfb24427/attachment-0001.html> -------------- next part -------------- A non-text attachment was scrubbed... Name: etc-glusterfs-glusterd.vol.log Type: text/x-log Size: 161664 bytes Desc: not available URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20160822/bfb24427/attachment-0002.bin> -------------- next part -------------- A non-text attachment was scrubbed... Name: mnt-raid6-storage-storage.log Type: text/x-log Size: 644099 bytes Desc: not available URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20160822/bfb24427/attachment-0003.bin>
Atin Mukherjee
2016-Aug-23 04:15 UTC
[Gluster-users] Upgrade 3.7.6 -> 3.7.13 one gluster server disconnected 1 of 3 volumes
On Tue, Aug 23, 2016 at 4:17 AM, Steve Dainard <sdainard at spd1.com> wrote:> About 5 hours after upgrading gluster 3.7.6 -> 3.7.13 on Centos 7, one of > my gluster servers disconnected its volume. The other two volumes this host > serves were not affected. > > # gluster volume status storage > Status of volume: storage > Gluster process TCP Port RDMA Port Online > Pid > ------------------------------------------------------------ > ------------------ > Brick 10.0.231.50:/mnt/raid6-storage/storage > 49159 0 Y 30743 > Brick 10.0.231.51:/mnt/raid6-storage/storage > 49159 0 Y 676 > *Brick 10.0.231.52:/mnt/raid6-storage/storag**e > N/A N/A N N/A * > Brick 10.0.231.53:/mnt/raid6-storage/storage > 49154 0 Y 10253 > Brick 10.0.231.54:/mnt/raid6-storage/storage > 49153 0 Y 2792 > Brick 10.0.231.55:/mnt/raid6-storage/storage > 49153 0 Y 13590 > Brick 10.0.231.56:/mnt/raid6-storage/storage > 49152 0 Y 9281 > NFS Server on localhost 2049 0 Y > 30775 > Quota Daemon on localhost N/A N/A Y > 30781 > NFS Server on 10.0.231.54 2049 0 Y > 2817 > Quota Daemon on 10.0.231.54 N/A N/A Y > 2824 > NFS Server on 10.0.231.51 2049 0 Y > 710 > Quota Daemon on 10.0.231.51 N/A N/A Y > 719 > NFS Server on 10.0.231.52 2049 0 Y > 9090 > Quota Daemon on 10.0.231.52 N/A N/A Y > 9098 > NFS Server on 10.0.231.55 2049 0 Y > 13611 > Quota Daemon on 10.0.231.55 N/A N/A Y > 13619 > NFS Server on 10.0.231.56 2049 0 Y > 9303 > Quota Daemon on 10.0.231.56 N/A N/A Y > 9310 > NFS Server on 10.0.231.53 2049 0 Y > 26304 > Quota Daemon on 10.0.231.53 N/A N/A Y > 26320 > > Task Status of Volume storage > ------------------------------------------------------------ > ------------------ > There are no active volume tasks > > I see lots of logs related to trashcan (failed [file exists]), set xattrs > (failed [no such file or directory]), quota (invalid arguments) in the > brick logs, which I enabled as a feature after the upgrade this morning. >Could you let us know the time (in UTC) around which this issue was seen such that we can look at the logs around that time and see if something went wrong.> After restarting glusterd on that host, the volume came back online. > > I've attached logs from that host if someone can take a look. > > # gluster volume info storage > > Volume Name: storage > Type: Distribute > Volume ID: 6f95525a-94d7-4174-bac4-e1a18fe010a2 > Status: Started > Number of Bricks: 7 > Transport-type: tcp > Bricks: > Brick1: 10.0.231.50:/mnt/raid6-storage/storage > Brick2: 10.0.231.51:/mnt/raid6-storage/storage > Brick3: 10.0.231.52:/mnt/raid6-storage/storage > Brick4: 10.0.231.53:/mnt/raid6-storage/storage > Brick5: 10.0.231.54:/mnt/raid6-storage/storage > Brick6: 10.0.231.55:/mnt/raid6-storage/storage > Brick7: 10.0.231.56:/mnt/raid6-storage/storage > Options Reconfigured: > nfs.disable: no > features.trash-max-filesize: 1GB > features.trash: on > features.quota-deem-statfs: on > features.inode-quota: on > features.quota: on > performance.readdir-ahead: on > > # rpm -qa |grep glusterfs > glusterfs-fuse-3.7.13-1.el7.x86_64 > glusterfs-cli-3.7.13-1.el7.x86_64 > glusterfs-3.7.13-1.el7.x86_64 > glusterfs-server-3.7.13-1.el7.x86_64 > glusterfs-api-3.7.13-1.el7.x86_64 > glusterfs-libs-3.7.13-1.el7.x86_64 > glusterfs-client-xlators-3.7.13-1.el7.x86_64 > > > Thanks, > Steve > > > > _______________________________________________ > Gluster-users mailing list > Gluster-users at gluster.org > http://www.gluster.org/mailman/listinfo/gluster-users >-- --Atin -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20160823/d3d2eb63/attachment.html>