Atin Mukherjee
2017-May-15 17:01 UTC
[Gluster-users] Failure while upgrading gluster to 3.10.1
On Mon, 15 May 2017 at 11:58, Pawan Alwandi <pawan at platform.sh> wrote:> Hi Atin, > > I see below error. Do I require gluster to be upgraded on all 3 hosts for > this to work? Right now I have host 1 running 3.10.1 and host 2 & 3 > running 3.6.2 > > # gluster v set all cluster.op-version 31001 > volume set: failed: Required op_version (31001) is not supported >Yes you should given 3.6 version is EOLed.> > > > On Mon, May 15, 2017 at 3:32 AM, Atin Mukherjee <amukherj at redhat.com> > wrote: > >> On Sun, 14 May 2017 at 21:43, Atin Mukherjee <amukherj at redhat.com> wrote: >> >>> Allright, I see that you haven't bumped up the op-version. Can you >>> please execute: >>> >>> gluster v set all cluster.op-version 30101 and then restart glusterd on >>> all the nodes and check the brick status? >>> >> >> s/30101/31001 >> >> >>> >>> On Sun, May 14, 2017 at 8:55 PM, Pawan Alwandi <pawan at platform.sh> >>> wrote: >>> >>>> Hello Atin, >>>> >>>> Thanks for looking at this. Below is the output you requested for. >>>> >>>> Again, I'm seeing those errors after upgrading gluster on host 1. >>>> >>>> Host 1 >>>> >>>> # cat /var/lib/glusterd/glusterd.info >>>> UUID=7f2a6e11-2a53-4ab4-9ceb-8be6a9f2d073 >>>> operating-version=30600 >>>> >>>> # cat /var/lib/glusterd/peers/* >>>> uuid=5ec54b4f-f60c-48c6-9e55-95f2bb58f633 >>>> state=3 >>>> hostname1=192.168.0.7 >>>> uuid=83e9a0b9-6bd5-483b-8516-d8928805ed95 >>>> state=3 >>>> hostname1=192.168.0.6 >>>> >>>> # gluster --version >>>> glusterfs 3.10.1 >>>> >>>> Host 2 >>>> >>>> # cat /var/lib/glusterd/glusterd.info >>>> UUID=83e9a0b9-6bd5-483b-8516-d8928805ed95 >>>> operating-version=30600 >>>> >>>> # cat /var/lib/glusterd/peers/* >>>> uuid=5ec54b4f-f60c-48c6-9e55-95f2bb58f633 >>>> state=3 >>>> hostname1=192.168.0.7 >>>> uuid=7f2a6e11-2a53-4ab4-9ceb-8be6a9f2d073 >>>> state=3 >>>> hostname1=192.168.0.5 >>>> >>>> # gluster --version >>>> glusterfs 3.6.2 built on Jan 21 2015 14:23:44 >>>> >>>> Host 3 >>>> >>>> # cat /var/lib/glusterd/glusterd.info >>>> UUID=5ec54b4f-f60c-48c6-9e55-95f2bb58f633 >>>> operating-version=30600 >>>> >>>> # cat /var/lib/glusterd/peers/* >>>> uuid=7f2a6e11-2a53-4ab4-9ceb-8be6a9f2d073 >>>> state=3 >>>> hostname1=192.168.0.5 >>>> uuid=83e9a0b9-6bd5-483b-8516-d8928805ed95 >>>> state=3 >>>> hostname1=192.168.0.6 >>>> >>>> # gluster --version >>>> glusterfs 3.6.2 built on Jan 21 2015 14:23:44 >>>> >>>> >>>> >>>> On Sat, May 13, 2017 at 6:28 PM, Atin Mukherjee <amukherj at redhat.com> >>>> wrote: >>>> >>>>> I have already asked for the following earlier: >>>>> >>>>> Can you please provide output of following from all the nodes: >>>>> >>>>> cat /var/lib/glusterd/glusterd.info >>>>> cat /var/lib/glusterd/peers/* >>>>> >>>>> On Sat, 13 May 2017 at 12:22, Pawan Alwandi <pawan at platform.sh> wrote: >>>>> >>>>>> Hello folks, >>>>>> >>>>>> Does anyone have any idea whats going on here? >>>>>> >>>>>> Thanks, >>>>>> Pawan >>>>>> >>>>>> On Wed, May 10, 2017 at 5:02 PM, Pawan Alwandi <pawan at platform.sh> >>>>>> wrote: >>>>>> >>>>>>> Hello, >>>>>>> >>>>>>> I'm trying to upgrade gluster from 3.6.2 to 3.10.1 but don't see the >>>>>>> glusterfsd and glusterfs processes coming up. >>>>>>> http://gluster.readthedocs.io/en/latest/Upgrade-Guide/upgrade_to_3.10/ >>>>>>> is the process that I'm trying to follow. >>>>>>> >>>>>>> This is a 3 node server setup with a replicated volume having >>>>>>> replica count of 3. >>>>>>> >>>>>>> Logs below: >>>>>>> >>>>>>> [2017-05-10 09:07:03.507959] I [MSGID: 100030] >>>>>>> [glusterfsd.c:2460:main] 0-/usr/sbin/glusterd: Started running >>>>>>> /usr/sbin/glusterd version 3.10.1 (args: /usr/sbin/glusterd -p >>>>>>> /var/run/glusterd.pid) >>>>>>> [2017-05-10 09:07:03.512827] I [MSGID: 106478] >>>>>>> [glusterd.c:1449:init] 0-management: Maximum allowed open file descriptors >>>>>>> set to 65536 >>>>>>> [2017-05-10 09:07:03.512855] I [MSGID: 106479] >>>>>>> [glusterd.c:1496:init] 0-management: Using /var/lib/glusterd as working >>>>>>> directory >>>>>>> [2017-05-10 09:07:03.520426] W [MSGID: 103071] >>>>>>> [rdma.c:4590:__gf_rdma_ctx_create] 0-rpc-transport/rdma: rdma_cm event >>>>>>> channel creation failed [No such device] >>>>>>> [2017-05-10 09:07:03.520452] W [MSGID: 103055] [rdma.c:4897:init] >>>>>>> 0-rdma.management: Failed to initialize IB Device >>>>>>> [2017-05-10 09:07:03.520465] W >>>>>>> [rpc-transport.c:350:rpc_transport_load] 0-rpc-transport: 'rdma' >>>>>>> initialization failed >>>>>>> [2017-05-10 09:07:03.520518] W >>>>>>> [rpcsvc.c:1661:rpcsvc_create_listener] 0-rpc-service: cannot create >>>>>>> listener, initing the transport failed >>>>>>> [2017-05-10 09:07:03.520534] E [MSGID: 106243] >>>>>>> [glusterd.c:1720:init] 0-management: creation of 1 listeners failed, >>>>>>> continuing with succeeded transport >>>>>>> [2017-05-10 09:07:04.931764] I [MSGID: 106513] >>>>>>> [glusterd-store.c:2197:glusterd_restore_op_version] 0-glusterd: retrieved >>>>>>> op-version: 30600 >>>>>>> [2017-05-10 09:07:04.964354] I [MSGID: 106544] >>>>>>> [glusterd.c:158:glusterd_uuid_init] 0-management: retrieved UUID: >>>>>>> 7f2a6e11-2a53-4ab4-9ceb-8be6a9f2d073 >>>>>>> [2017-05-10 09:07:04.993944] I [MSGID: 106498] >>>>>>> [glusterd-handler.c:3669:glusterd_friend_add_from_peerinfo] 0-management: >>>>>>> connect returned 0 >>>>>>> [2017-05-10 09:07:04.995864] I [MSGID: 106498] >>>>>>> [glusterd-handler.c:3669:glusterd_friend_add_from_peerinfo] 0-management: >>>>>>> connect returned 0 >>>>>>> [2017-05-10 09:07:04.995879] W [MSGID: 106062] >>>>>>> [glusterd-handler.c:3466:glusterd_transport_inet_options_build] 0-glusterd: >>>>>>> Failed to get tcp-user-timeout >>>>>>> [2017-05-10 09:07:04.995903] I >>>>>>> [rpc-clnt.c:1059:rpc_clnt_connection_init] 0-management: setting >>>>>>> frame-timeout to 600 >>>>>>> [2017-05-10 09:07:04.996325] I >>>>>>> [rpc-clnt.c:1059:rpc_clnt_connection_init] 0-management: setting >>>>>>> frame-timeout to 600 >>>>>>> Final graph: >>>>>>> >>>>>>> +------------------------------------------------------------------------------+ >>>>>>> 1: volume management >>>>>>> 2: type mgmt/glusterd >>>>>>> 3: option rpc-auth.auth-glusterfs on >>>>>>> 4: option rpc-auth.auth-unix on >>>>>>> 5: option rpc-auth.auth-null on >>>>>>> 6: option rpc-auth-allow-insecure on >>>>>>> 7: option transport.socket.listen-backlog 128 >>>>>>> 8: option event-threads 1 >>>>>>> 9: option ping-timeout 0 >>>>>>> 10: option transport.socket.read-fail-log off >>>>>>> 11: option transport.socket.keepalive-interval 2 >>>>>>> 12: option transport.socket.keepalive-time 10 >>>>>>> 13: option transport-type rdma >>>>>>> 14: option working-directory /var/lib/glusterd >>>>>>> 15: end-volume >>>>>>> 16: >>>>>>> >>>>>>> +------------------------------------------------------------------------------+ >>>>>>> [2017-05-10 09:07:04.996310] W [MSGID: 106062] >>>>>>> [glusterd-handler.c:3466:glusterd_transport_inet_options_build] 0-glusterd: >>>>>>> Failed to get tcp-user-timeout >>>>>>> [2017-05-10 09:07:05.000461] I [MSGID: 101190] >>>>>>> [event-epoll.c:629:event_dispatch_epoll_worker] 0-epoll: Started thread >>>>>>> with index 1 >>>>>>> [2017-05-10 09:07:05.001493] W [socket.c:593:__socket_rwv] >>>>>>> 0-management: readv on 192.168.0.7:24007 failed (No data available) >>>>>>> [2017-05-10 09:07:05.001513] I [MSGID: 106004] >>>>>>> [glusterd-handler.c:5882:__glusterd_peer_rpc_notify] 0-management: Peer >>>>>>> <192.168.0.7> (<5ec54b4f-f60c-48c6-9e55-95f2bb58f633>), in state <Peer in >>>>>>> Cluster>, h >>>>>>> as disconnected from glusterd. >>>>>>> [2017-05-10 09:07:05.001677] W >>>>>>> [glusterd-locks.c:675:glusterd_mgmt_v3_unlock] >>>>>>> (-->/usr/lib/x86_64-linux-gnu/glusterfs/3.10.1/xlator/mgmt/glusterd.so(+0x20559) >>>>>>> [0x7f0bf9d74559] -->/usr/lib/x86_64-linux-gnu >>>>>>> /glusterfs/3.10.1/xlator/mgmt/glusterd.so(+0x29cf0) [0x7f0bf9d7dcf0] >>>>>>> -->/usr/lib/x86_64-linux-gnu/glusterfs/3.10.1/xlator/mgmt/glusterd.so(+0xd5ba3) >>>>>>> [0x7f0bf9e29ba3] ) 0-management: Lock for vol shared no >>>>>>> t held >>>>>>> [2017-05-10 09:07:05.001696] W [MSGID: 106118] >>>>>>> [glusterd-handler.c:5907:__glusterd_peer_rpc_notify] 0-management: Lock not >>>>>>> released for shared >>>>>>> [2017-05-10 09:07:05.003099] E [rpc-clnt.c:365:saved_frames_unwind] >>>>>>> (--> >>>>>>> /usr/lib/x86_64-linux-gnu/libglusterfs.so.0(_gf_log_callingfn+0x13c)[0x7f0bfeeca73c] >>>>>>> (--> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(s >>>>>>> aved_frames_unwind+0x1cf)[0x7f0bfec904bf] (--> >>>>>>> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(saved_frames_destroy+0xe)[0x7f0bfec905de] >>>>>>> (--> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(rpc_clnt_connection_cleanup+0x >>>>>>> 91)[0x7f0bfec91c21] (--> >>>>>>> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(rpc_clnt_notify+0x290)[0x7f0bfec92710] >>>>>>> ))))) 0-management: forced unwinding frame type(GLUSTERD-DUMP) op(DUMP(1)) >>>>>>> called at 2017-05-10 09:0 >>>>>>> 7:05.000627 (xid=0x1) >>>>>>> [2017-05-10 09:07:05.003129] E [MSGID: 106167] >>>>>>> [glusterd-handshake.c:2181:__glusterd_peer_dump_version_cbk] 0-management: >>>>>>> Error through RPC layer, retry again later >>>>>>> [2017-05-10 09:07:05.003251] W [socket.c:593:__socket_rwv] >>>>>>> 0-management: readv on 192.168.0.6:24007 failed (No data available) >>>>>>> [2017-05-10 09:07:05.003267] I [MSGID: 106004] >>>>>>> [glusterd-handler.c:5882:__glusterd_peer_rpc_notify] 0-management: Peer >>>>>>> <192.168.0.6> (<83e9a0b9-6bd5-483b-8516-d8928805ed95>), in state <Peer in >>>>>>> Cluster>, h >>>>>>> as disconnected from glusterd. >>>>>>> [2017-05-10 09:07:05.003318] W >>>>>>> [glusterd-locks.c:675:glusterd_mgmt_v3_unlock] >>>>>>> (-->/usr/lib/x86_64-linux-gnu/glusterfs/3.10.1/xlator/mgmt/glusterd.so(+0x20559) >>>>>>> [0x7f0bf9d74559] -->/usr/lib/x86_64-linux-gnu >>>>>>> /glusterfs/3.10.1/xlator/mgmt/glusterd.so(+0x29cf0) [0x7f0bf9d7dcf0] >>>>>>> -->/usr/lib/x86_64-linux-gnu/glusterfs/3.10.1/xlator/mgmt/glusterd.so(+0xd5ba3) >>>>>>> [0x7f0bf9e29ba3] ) 0-management: Lock for vol shared no >>>>>>> t held >>>>>>> [2017-05-10 09:07:05.003329] W [MSGID: 106118] >>>>>>> [glusterd-handler.c:5907:__glusterd_peer_rpc_notify] 0-management: Lock not >>>>>>> released for shared >>>>>>> [2017-05-10 09:07:05.003457] E [rpc-clnt.c:365:saved_frames_unwind] >>>>>>> (--> >>>>>>> /usr/lib/x86_64-linux-gnu/libglusterfs.so.0(_gf_log_callingfn+0x13c)[0x7f0bfeeca73c] >>>>>>> (--> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(s >>>>>>> aved_frames_unwind+0x1cf)[0x7f0bfec904bf] (--> >>>>>>> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(saved_frames_destroy+0xe)[0x7f0bfec905de] >>>>>>> (--> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(rpc_clnt_connection_cleanup+0x >>>>>>> 91)[0x7f0bfec91c21] (--> >>>>>>> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(rpc_clnt_notify+0x290)[0x7f0bfec92710] >>>>>>> ))))) 0-management: forced unwinding frame type(GLUSTERD-DUMP) op(DUMP(1)) >>>>>>> called at 2017-05-10 09:0 >>>>>>> 7:05.001407 (xid=0x1) >>>>>>> >>>>>>> There are a bunch of errors reported but I'm not sure which is >>>>>>> signal and which ones are noise. Does anyone have any idea whats going on >>>>>>> here? >>>>>>> >>>>>>> Thanks, >>>>>>> Pawan >>>>>>> >>>>>>> _______________________________________________ >>>>>> Gluster-users mailing list >>>>>> Gluster-users at gluster.org >>>>>> http://lists.gluster.org/mailman/listinfo/gluster-users >>>>> >>>>> -- >>>>> - Atin (atinm) >>>>> >>>> >>>> >>> -- >> - Atin (atinm) >> > > --- Atin (atinm) -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20170515/9861a33a/attachment.html>
Pawan Alwandi
2017-May-17 07:15 UTC
[Gluster-users] Failure while upgrading gluster to 3.10.1
Hello Atin, I realized that these http://gluster.readthedocs.io/en/latest/Upgrade-Guide/upgrade_to_3.10/ instructions only work for upgrades from 3.7, while we are running 3.6.2. Are there instructions/suggestion you have for us to upgrade from 3.6 version? I believe upgrade from 3.6 to 3.7 and then to 3.10 would work, but I see similar errors reported when I upgraded to 3.7 too. For what its worth, I was able to set the op-version (gluster v set all cluster.op-version 30702) but that doesn't seem to help. [2017-05-17 06:48:33.700014] I [MSGID: 100030] [glusterfsd.c:2338:main] 0-/usr/sbin/glusterd: Started running /usr/sbin/glusterd version 3.7.20 (args: /usr/sbin/glusterd -p /var/run/glusterd.pid) [2017-05-17 06:48:33.703808] I [MSGID: 106478] [glusterd.c:1383:init] 0-management: Maximum allowed open file descriptors set to 65536 [2017-05-17 06:48:33.703836] I [MSGID: 106479] [glusterd.c:1432:init] 0-management: Using /var/lib/glusterd as working directory [2017-05-17 06:48:33.708866] W [MSGID: 103071] [rdma.c:4594:__gf_rdma_ctx_create] 0-rpc-transport/rdma: rdma_cm event channel creation failed [No such device] [2017-05-17 06:48:33.709011] W [MSGID: 103055] [rdma.c:4901:init] 0-rdma.management: Failed to initialize IB Device [2017-05-17 06:48:33.709033] W [rpc-transport.c:359:rpc_transport_load] 0-rpc-transport: 'rdma' initialization failed [2017-05-17 06:48:33.709088] W [rpcsvc.c:1642:rpcsvc_create_listener] 0-rpc-service: cannot create listener, initing the transport failed [2017-05-17 06:48:33.709105] E [MSGID: 106243] [glusterd.c:1656:init] 0-management: creation of 1 listeners failed, continuing with succeeded transport [2017-05-17 06:48:35.480043] I [MSGID: 106513] [glusterd-store.c:2068:glusterd_restore_op_version] 0-glusterd: retrieved op-version: 30600 [2017-05-17 06:48:35.605779] I [MSGID: 106498] [glusterd-handler.c:3640:glusterd_friend_add_from_peerinfo] 0-management: connect returned 0 [2017-05-17 06:48:35.607059] I [rpc-clnt.c:1046:rpc_clnt_connection_init] 0-management: setting frame-timeout to 600 [2017-05-17 06:48:35.607670] I [rpc-clnt.c:1046:rpc_clnt_connection_init] 0-management: setting frame-timeout to 600 [2017-05-17 06:48:35.607025] I [MSGID: 106498] [glusterd-handler.c:3640:glusterd_friend_add_from_peerinfo] 0-management: connect returned 0 [2017-05-17 06:48:35.608125] I [MSGID: 106544] [glusterd.c:159:glusterd_uuid_init] 0-management: retrieved UUID: 7f2a6e11-2a53-4ab4-9ceb-8be6a9f2d073 Final graph: +------------------------------------------------------------------------------+ 1: volume management 2: type mgmt/glusterd 3: option rpc-auth.auth-glusterfs on 4: option rpc-auth.auth-unix on 5: option rpc-auth.auth-null on 6: option rpc-auth-allow-insecure on 7: option transport.socket.listen-backlog 128 8: option event-threads 1 9: option ping-timeout 0 10: option transport.socket.read-fail-log off 11: option transport.socket.keepalive-interval 2 12: option transport.socket.keepalive-time 10 13: option transport-type rdma 14: option working-directory /var/lib/glusterd 15: end-volume 16: +------------------------------------------------------------------------------+ [2017-05-17 06:48:35.609868] I [MSGID: 101190] [event-epoll.c:632:event_dispatch_epoll_worker] 0-epoll: Started thread with index 1 [2017-05-17 06:48:35.610839] W [socket.c:596:__socket_rwv] 0-management: readv on 192.168.0.7:24007 failed (No data available) [2017-05-17 06:48:35.611907] E [rpc-clnt.c:370:saved_frames_unwind] (--> /usr/lib/x86_64-linux-gnu/libglusterfs.so.0(_gf_log_callingfn+0x1a3)[0x7fd6c2d70bb3] (--> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(saved_frames_unwind+0x1cf)[0x7fd6c2b3a2df] (--> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(saved_frames_destroy+0xe)[0x7fd6c2b3a3fe] (--> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(rpc_clnt_connection_cleanup+0x89)[0x7fd6c2b3ba39] (--> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(rpc_clnt_notify+0x160)[0x7fd6c2b3c380] ))))) 0-management: forced unwinding frame type(GLUSTERD-DUMP) op(DUMP(1)) called at 2017-05-17 06:48:35.609965 (xid=0x1) [2017-05-17 06:48:35.611928] E [MSGID: 106167] [glusterd-handshake.c:2091:__glusterd_peer_dump_version_cbk] 0-management: Error through RPC layer, retry again later [2017-05-17 06:48:35.611944] I [MSGID: 106004] [glusterd-handler.c:5201:__glusterd_peer_rpc_notify] 0-management: Peer <192.168.0.7> (<5ec54b4f-f60c-48c6-9e55-95f2bb58f633>), in state <Peer in Cluster>, has disconnected from glusterd. [2017-05-17 06:48:35.612024] W [glusterd-locks.c:681:glusterd_mgmt_v3_unlock] (-->/usr/lib/x86_64-linux-gnu/glusterfs/3.7.20/xlator/mgmt/glusterd.so(glusterd_big_locked_notify+0x4b) [0x7fd6bdc4912b] -->/usr/lib/x86_64-linux-gnu/glusterfs/3.7.20/xlator/mgmt/glusterd.so(__glusterd_peer_rpc_notify+0x160) [0x7fd6bdc52dd0] -->/usr/lib/x86_64-linux-gnu/glusterfs/3.7.20/xlator/mgmt/glusterd.so(glusterd_mgmt_v3_unlock+0x4c3) [0x7fd6bdcef1b3] ) 0-management: Lock for vol shared not held [2017-05-17 06:48:35.612039] W [MSGID: 106118] [glusterd-handler.c:5223:__glusterd_peer_rpc_notify] 0-management: Lock not released for shared [2017-05-17 06:48:35.612079] W [socket.c:596:__socket_rwv] 0-management: readv on 192.168.0.6:24007 failed (No data available) [2017-05-17 06:48:35.612179] E [rpc-clnt.c:370:saved_frames_unwind] (--> /usr/lib/x86_64-linux-gnu/libglusterfs.so.0(_gf_log_callingfn+0x1a3)[0x7fd6c2d70bb3] (--> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(saved_frames_unwind+0x1cf)[0x7fd6c2b3a2df] (--> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(saved_frames_destroy+0xe)[0x7fd6c2b3a3fe] (--> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(rpc_clnt_connection_cleanup+0x89)[0x7fd6c2b3ba39] (--> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(rpc_clnt_notify+0x160)[0x7fd6c2b3c380] ))))) 0-management: forced unwinding frame type(GLUSTERD-DUMP) op(DUMP(1)) called at 2017-05-17 06:48:35.610007 (xid=0x1) [2017-05-17 06:48:35.612197] E [MSGID: 106167] [glusterd-handshake.c:2091:__glusterd_peer_dump_version_cbk] 0-management: Error through RPC layer, retry again later [2017-05-17 06:48:35.612211] I [MSGID: 106004] [glusterd-handler.c:5201:__glusterd_peer_rpc_notify] 0-management: Peer <192.168.0.6> (<83e9a0b9-6bd5-483b-8516-d8928805ed95>), in state <Peer in Cluster>, has disconnected from glusterd. [2017-05-17 06:48:35.612292] W [glusterd-locks.c:681:glusterd_mgmt_v3_unlock] (-->/usr/lib/x86_64-linux-gnu/glusterfs/3.7.20/xlator/mgmt/glusterd.so(glusterd_big_locked_notify+0x4b) [0x7fd6bdc4912b] -->/usr/lib/x86_64-linux-gnu/glusterfs/3.7.20/xlator/mgmt/glusterd.so(__glusterd_peer_rpc_notify+0x160) [0x7fd6bdc52dd0] -->/usr/lib/x86_64-linux-gnu/glusterfs/3.7.20/xlator/mgmt/glusterd.so(glusterd_mgmt_v3_unlock+0x4c3) [0x7fd6bdcef1b3] ) 0-management: Lock for vol shared not held [2017-05-17 06:48:35.613432] W [MSGID: 106118] [glusterd-handler.c:5223:__glusterd_peer_rpc_notify] 0-management: Lock not released for shared [2017-05-17 06:48:35.614317] E [MSGID: 106170] [glusterd-handshake.c:1051:gd_validate_mgmt_hndsk_req] 0-management: Request from peer 192.168.0.6:991 has an entry in peerinfo, but uuid does not match On Mon, May 15, 2017 at 10:31 PM, Atin Mukherjee <amukherj at redhat.com> wrote:> > On Mon, 15 May 2017 at 11:58, Pawan Alwandi <pawan at platform.sh> wrote: > >> Hi Atin, >> >> I see below error. Do I require gluster to be upgraded on all 3 hosts >> for this to work? Right now I have host 1 running 3.10.1 and host 2 & 3 >> running 3.6.2 >> >> # gluster v set all cluster.op-version 31001 >> volume set: failed: Required op_version (31001) is not supported >> > > Yes you should given 3.6 version is EOLed. > >> >> >> >> On Mon, May 15, 2017 at 3:32 AM, Atin Mukherjee <amukherj at redhat.com> >> wrote: >> >>> On Sun, 14 May 2017 at 21:43, Atin Mukherjee <amukherj at redhat.com> >>> wrote: >>> >>>> Allright, I see that you haven't bumped up the op-version. Can you >>>> please execute: >>>> >>>> gluster v set all cluster.op-version 30101 and then restart glusterd >>>> on all the nodes and check the brick status? >>>> >>> >>> s/30101/31001 >>> >>> >>>> >>>> On Sun, May 14, 2017 at 8:55 PM, Pawan Alwandi <pawan at platform.sh> >>>> wrote: >>>> >>>>> Hello Atin, >>>>> >>>>> Thanks for looking at this. Below is the output you requested for. >>>>> >>>>> Again, I'm seeing those errors after upgrading gluster on host 1. >>>>> >>>>> Host 1 >>>>> >>>>> # cat /var/lib/glusterd/glusterd.info >>>>> UUID=7f2a6e11-2a53-4ab4-9ceb-8be6a9f2d073 >>>>> operating-version=30600 >>>>> >>>>> # cat /var/lib/glusterd/peers/* >>>>> uuid=5ec54b4f-f60c-48c6-9e55-95f2bb58f633 >>>>> state=3 >>>>> hostname1=192.168.0.7 >>>>> uuid=83e9a0b9-6bd5-483b-8516-d8928805ed95 >>>>> state=3 >>>>> hostname1=192.168.0.6 >>>>> >>>>> # gluster --version >>>>> glusterfs 3.10.1 >>>>> >>>>> Host 2 >>>>> >>>>> # cat /var/lib/glusterd/glusterd.info >>>>> UUID=83e9a0b9-6bd5-483b-8516-d8928805ed95 >>>>> operating-version=30600 >>>>> >>>>> # cat /var/lib/glusterd/peers/* >>>>> uuid=5ec54b4f-f60c-48c6-9e55-95f2bb58f633 >>>>> state=3 >>>>> hostname1=192.168.0.7 >>>>> uuid=7f2a6e11-2a53-4ab4-9ceb-8be6a9f2d073 >>>>> state=3 >>>>> hostname1=192.168.0.5 >>>>> >>>>> # gluster --version >>>>> glusterfs 3.6.2 built on Jan 21 2015 14:23:44 >>>>> >>>>> Host 3 >>>>> >>>>> # cat /var/lib/glusterd/glusterd.info >>>>> UUID=5ec54b4f-f60c-48c6-9e55-95f2bb58f633 >>>>> operating-version=30600 >>>>> >>>>> # cat /var/lib/glusterd/peers/* >>>>> uuid=7f2a6e11-2a53-4ab4-9ceb-8be6a9f2d073 >>>>> state=3 >>>>> hostname1=192.168.0.5 >>>>> uuid=83e9a0b9-6bd5-483b-8516-d8928805ed95 >>>>> state=3 >>>>> hostname1=192.168.0.6 >>>>> >>>>> # gluster --version >>>>> glusterfs 3.6.2 built on Jan 21 2015 14:23:44 >>>>> >>>>> >>>>> >>>>> On Sat, May 13, 2017 at 6:28 PM, Atin Mukherjee <amukherj at redhat.com> >>>>> wrote: >>>>> >>>>>> I have already asked for the following earlier: >>>>>> >>>>>> Can you please provide output of following from all the nodes: >>>>>> >>>>>> cat /var/lib/glusterd/glusterd.info >>>>>> cat /var/lib/glusterd/peers/* >>>>>> >>>>>> On Sat, 13 May 2017 at 12:22, Pawan Alwandi <pawan at platform.sh> >>>>>> wrote: >>>>>> >>>>>>> Hello folks, >>>>>>> >>>>>>> Does anyone have any idea whats going on here? >>>>>>> >>>>>>> Thanks, >>>>>>> Pawan >>>>>>> >>>>>>> On Wed, May 10, 2017 at 5:02 PM, Pawan Alwandi <pawan at platform.sh> >>>>>>> wrote: >>>>>>> >>>>>>>> Hello, >>>>>>>> >>>>>>>> I'm trying to upgrade gluster from 3.6.2 to 3.10.1 but don't see >>>>>>>> the glusterfsd and glusterfs processes coming up. >>>>>>>> http://gluster.readthedocs.io/en/latest/Upgrade-Guide/ >>>>>>>> upgrade_to_3.10/ is the process that I'm trying to follow. >>>>>>>> >>>>>>>> This is a 3 node server setup with a replicated volume having >>>>>>>> replica count of 3. >>>>>>>> >>>>>>>> Logs below: >>>>>>>> >>>>>>>> [2017-05-10 09:07:03.507959] I [MSGID: 100030] >>>>>>>> [glusterfsd.c:2460:main] 0-/usr/sbin/glusterd: Started running >>>>>>>> /usr/sbin/glusterd version 3.10.1 (args: /usr/sbin/glusterd -p >>>>>>>> /var/run/glusterd.pid) >>>>>>>> [2017-05-10 09:07:03.512827] I [MSGID: 106478] >>>>>>>> [glusterd.c:1449:init] 0-management: Maximum allowed open file descriptors >>>>>>>> set to 65536 >>>>>>>> [2017-05-10 09:07:03.512855] I [MSGID: 106479] >>>>>>>> [glusterd.c:1496:init] 0-management: Using /var/lib/glusterd as working >>>>>>>> directory >>>>>>>> [2017-05-10 09:07:03.520426] W [MSGID: 103071] >>>>>>>> [rdma.c:4590:__gf_rdma_ctx_create] 0-rpc-transport/rdma: rdma_cm >>>>>>>> event channel creation failed [No such device] >>>>>>>> [2017-05-10 09:07:03.520452] W [MSGID: 103055] [rdma.c:4897:init] >>>>>>>> 0-rdma.management: Failed to initialize IB Device >>>>>>>> [2017-05-10 09:07:03.520465] W [rpc-transport.c:350:rpc_transport_load] >>>>>>>> 0-rpc-transport: 'rdma' initialization failed >>>>>>>> [2017-05-10 09:07:03.520518] W [rpcsvc.c:1661:rpcsvc_create_listener] >>>>>>>> 0-rpc-service: cannot create listener, initing the transport failed >>>>>>>> [2017-05-10 09:07:03.520534] E [MSGID: 106243] >>>>>>>> [glusterd.c:1720:init] 0-management: creation of 1 listeners failed, >>>>>>>> continuing with succeeded transport >>>>>>>> [2017-05-10 09:07:04.931764] I [MSGID: 106513] >>>>>>>> [glusterd-store.c:2197:glusterd_restore_op_version] 0-glusterd: >>>>>>>> retrieved op-version: 30600 >>>>>>>> [2017-05-10 09:07:04.964354] I [MSGID: 106544] >>>>>>>> [glusterd.c:158:glusterd_uuid_init] 0-management: retrieved UUID: >>>>>>>> 7f2a6e11-2a53-4ab4-9ceb-8be6a9f2d073 >>>>>>>> [2017-05-10 09:07:04.993944] I [MSGID: 106498] >>>>>>>> [glusterd-handler.c:3669:glusterd_friend_add_from_peerinfo] >>>>>>>> 0-management: connect returned 0 >>>>>>>> [2017-05-10 09:07:04.995864] I [MSGID: 106498] >>>>>>>> [glusterd-handler.c:3669:glusterd_friend_add_from_peerinfo] >>>>>>>> 0-management: connect returned 0 >>>>>>>> [2017-05-10 09:07:04.995879] W [MSGID: 106062] >>>>>>>> [glusterd-handler.c:3466:glusterd_transport_inet_options_build] >>>>>>>> 0-glusterd: Failed to get tcp-user-timeout >>>>>>>> [2017-05-10 09:07:04.995903] I [rpc-clnt.c:1059:rpc_clnt_connection_init] >>>>>>>> 0-management: setting frame-timeout to 600 >>>>>>>> [2017-05-10 09:07:04.996325] I [rpc-clnt.c:1059:rpc_clnt_connection_init] >>>>>>>> 0-management: setting frame-timeout to 600 >>>>>>>> Final graph: >>>>>>>> +----------------------------------------------------------- >>>>>>>> -------------------+ >>>>>>>> 1: volume management >>>>>>>> 2: type mgmt/glusterd >>>>>>>> 3: option rpc-auth.auth-glusterfs on >>>>>>>> 4: option rpc-auth.auth-unix on >>>>>>>> 5: option rpc-auth.auth-null on >>>>>>>> 6: option rpc-auth-allow-insecure on >>>>>>>> 7: option transport.socket.listen-backlog 128 >>>>>>>> 8: option event-threads 1 >>>>>>>> 9: option ping-timeout 0 >>>>>>>> 10: option transport.socket.read-fail-log off >>>>>>>> 11: option transport.socket.keepalive-interval 2 >>>>>>>> 12: option transport.socket.keepalive-time 10 >>>>>>>> 13: option transport-type rdma >>>>>>>> 14: option working-directory /var/lib/glusterd >>>>>>>> 15: end-volume >>>>>>>> 16: >>>>>>>> +----------------------------------------------------------- >>>>>>>> -------------------+ >>>>>>>> [2017-05-10 09:07:04.996310] W [MSGID: 106062] >>>>>>>> [glusterd-handler.c:3466:glusterd_transport_inet_options_build] >>>>>>>> 0-glusterd: Failed to get tcp-user-timeout >>>>>>>> [2017-05-10 09:07:05.000461] I [MSGID: 101190] >>>>>>>> [event-epoll.c:629:event_dispatch_epoll_worker] 0-epoll: Started >>>>>>>> thread with index 1 >>>>>>>> [2017-05-10 09:07:05.001493] W [socket.c:593:__socket_rwv] >>>>>>>> 0-management: readv on 192.168.0.7:24007 failed (No data available) >>>>>>>> [2017-05-10 09:07:05.001513] I [MSGID: 106004] >>>>>>>> [glusterd-handler.c:5882:__glusterd_peer_rpc_notify] 0-management: >>>>>>>> Peer <192.168.0.7> (<5ec54b4f-f60c-48c6-9e55-95f2bb58f633>), in >>>>>>>> state <Peer in Cluster>, h >>>>>>>> as disconnected from glusterd. >>>>>>>> [2017-05-10 09:07:05.001677] W [glusterd-locks.c:675:glusterd_mgmt_v3_unlock] >>>>>>>> (-->/usr/lib/x86_64-linux-gnu/glusterfs/3.10.1/xlator/mgmt/glusterd.so(+0x20559) >>>>>>>> [0x7f0bf9d74559] -->/usr/lib/x86_64-linux-gnu >>>>>>>> /glusterfs/3.10.1/xlator/mgmt/glusterd.so(+0x29cf0) >>>>>>>> [0x7f0bf9d7dcf0] -->/usr/lib/x86_64-linux-gnu/ >>>>>>>> glusterfs/3.10.1/xlator/mgmt/glusterd.so(+0xd5ba3) >>>>>>>> [0x7f0bf9e29ba3] ) 0-management: Lock for vol shared no >>>>>>>> t held >>>>>>>> [2017-05-10 09:07:05.001696] W [MSGID: 106118] >>>>>>>> [glusterd-handler.c:5907:__glusterd_peer_rpc_notify] 0-management: >>>>>>>> Lock not released for shared >>>>>>>> [2017-05-10 09:07:05.003099] E [rpc-clnt.c:365:saved_frames_unwind] >>>>>>>> (--> /usr/lib/x86_64-linux-gnu/libglusterfs.so.0(_gf_log_ >>>>>>>> callingfn+0x13c)[0x7f0bfeeca73c] (--> /usr/lib/x86_64-linux-gnu/ >>>>>>>> libgfrpc.so.0(s >>>>>>>> aved_frames_unwind+0x1cf)[0x7f0bfec904bf] (--> >>>>>>>> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(saved_frames_destroy+0xe)[0x7f0bfec905de] >>>>>>>> (--> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(rpc_clnt_ >>>>>>>> connection_cleanup+0x >>>>>>>> 91)[0x7f0bfec91c21] (--> /usr/lib/x86_64-linux-gnu/ >>>>>>>> libgfrpc.so.0(rpc_clnt_notify+0x290)[0x7f0bfec92710] ))))) >>>>>>>> 0-management: forced unwinding frame type(GLUSTERD-DUMP) op(DUMP(1)) called >>>>>>>> at 2017-05-10 09:0 >>>>>>>> 7:05.000627 (xid=0x1) >>>>>>>> [2017-05-10 09:07:05.003129] E [MSGID: 106167] >>>>>>>> [glusterd-handshake.c:2181:__glusterd_peer_dump_version_cbk] >>>>>>>> 0-management: Error through RPC layer, retry again later >>>>>>>> [2017-05-10 09:07:05.003251] W [socket.c:593:__socket_rwv] >>>>>>>> 0-management: readv on 192.168.0.6:24007 failed (No data available) >>>>>>>> [2017-05-10 09:07:05.003267] I [MSGID: 106004] >>>>>>>> [glusterd-handler.c:5882:__glusterd_peer_rpc_notify] 0-management: >>>>>>>> Peer <192.168.0.6> (<83e9a0b9-6bd5-483b-8516-d8928805ed95>), in >>>>>>>> state <Peer in Cluster>, h >>>>>>>> as disconnected from glusterd. >>>>>>>> [2017-05-10 09:07:05.003318] W [glusterd-locks.c:675:glusterd_mgmt_v3_unlock] >>>>>>>> (-->/usr/lib/x86_64-linux-gnu/glusterfs/3.10.1/xlator/mgmt/glusterd.so(+0x20559) >>>>>>>> [0x7f0bf9d74559] -->/usr/lib/x86_64-linux-gnu >>>>>>>> /glusterfs/3.10.1/xlator/mgmt/glusterd.so(+0x29cf0) >>>>>>>> [0x7f0bf9d7dcf0] -->/usr/lib/x86_64-linux-gnu/ >>>>>>>> glusterfs/3.10.1/xlator/mgmt/glusterd.so(+0xd5ba3) >>>>>>>> [0x7f0bf9e29ba3] ) 0-management: Lock for vol shared no >>>>>>>> t held >>>>>>>> [2017-05-10 09:07:05.003329] W [MSGID: 106118] >>>>>>>> [glusterd-handler.c:5907:__glusterd_peer_rpc_notify] 0-management: >>>>>>>> Lock not released for shared >>>>>>>> [2017-05-10 09:07:05.003457] E [rpc-clnt.c:365:saved_frames_unwind] >>>>>>>> (--> /usr/lib/x86_64-linux-gnu/libglusterfs.so.0(_gf_log_ >>>>>>>> callingfn+0x13c)[0x7f0bfeeca73c] (--> /usr/lib/x86_64-linux-gnu/ >>>>>>>> libgfrpc.so.0(s >>>>>>>> aved_frames_unwind+0x1cf)[0x7f0bfec904bf] (--> >>>>>>>> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(saved_frames_destroy+0xe)[0x7f0bfec905de] >>>>>>>> (--> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(rpc_clnt_ >>>>>>>> connection_cleanup+0x >>>>>>>> 91)[0x7f0bfec91c21] (--> /usr/lib/x86_64-linux-gnu/ >>>>>>>> libgfrpc.so.0(rpc_clnt_notify+0x290)[0x7f0bfec92710] ))))) >>>>>>>> 0-management: forced unwinding frame type(GLUSTERD-DUMP) op(DUMP(1)) called >>>>>>>> at 2017-05-10 09:0 >>>>>>>> 7:05.001407 (xid=0x1) >>>>>>>> >>>>>>>> There are a bunch of errors reported but I'm not sure which is >>>>>>>> signal and which ones are noise. Does anyone have any idea whats going on >>>>>>>> here? >>>>>>>> >>>>>>>> Thanks, >>>>>>>> Pawan >>>>>>>> >>>>>>>> _______________________________________________ >>>>>>> Gluster-users mailing list >>>>>>> Gluster-users at gluster.org >>>>>>> http://lists.gluster.org/mailman/listinfo/gluster-users >>>>>> >>>>>> -- >>>>>> - Atin (atinm) >>>>>> >>>>> >>>>> >>>> -- >>> - Atin (atinm) >>> >> >> -- > - Atin (atinm) >-------------- next part -------------- An HTML attachment was scrubbed... URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20170517/5c539d22/attachment.html>