Atin Mukherjee
2017-May-14 16:13 UTC
[Gluster-users] Failure while upgrading gluster to 3.10.1
Allright, I see that you haven't bumped up the op-version. Can you please execute: gluster v set all cluster.op-version 30101 and then restart glusterd on all the nodes and check the brick status? On Sun, May 14, 2017 at 8:55 PM, Pawan Alwandi <pawan at platform.sh> wrote:> Hello Atin, > > Thanks for looking at this. Below is the output you requested for. > > Again, I'm seeing those errors after upgrading gluster on host 1. > > Host 1 > > # cat /var/lib/glusterd/glusterd.info > UUID=7f2a6e11-2a53-4ab4-9ceb-8be6a9f2d073 > operating-version=30600 > > # cat /var/lib/glusterd/peers/* > uuid=5ec54b4f-f60c-48c6-9e55-95f2bb58f633 > state=3 > hostname1=192.168.0.7 > uuid=83e9a0b9-6bd5-483b-8516-d8928805ed95 > state=3 > hostname1=192.168.0.6 > > # gluster --version > glusterfs 3.10.1 > > Host 2 > > # cat /var/lib/glusterd/glusterd.info > UUID=83e9a0b9-6bd5-483b-8516-d8928805ed95 > operating-version=30600 > > # cat /var/lib/glusterd/peers/* > uuid=5ec54b4f-f60c-48c6-9e55-95f2bb58f633 > state=3 > hostname1=192.168.0.7 > uuid=7f2a6e11-2a53-4ab4-9ceb-8be6a9f2d073 > state=3 > hostname1=192.168.0.5 > > # gluster --version > glusterfs 3.6.2 built on Jan 21 2015 14:23:44 > > Host 3 > > # cat /var/lib/glusterd/glusterd.info > UUID=5ec54b4f-f60c-48c6-9e55-95f2bb58f633 > operating-version=30600 > > # cat /var/lib/glusterd/peers/* > uuid=7f2a6e11-2a53-4ab4-9ceb-8be6a9f2d073 > state=3 > hostname1=192.168.0.5 > uuid=83e9a0b9-6bd5-483b-8516-d8928805ed95 > state=3 > hostname1=192.168.0.6 > > # gluster --version > glusterfs 3.6.2 built on Jan 21 2015 14:23:44 > > > > On Sat, May 13, 2017 at 6:28 PM, Atin Mukherjee <amukherj at redhat.com> > wrote: > >> I have already asked for the following earlier: >> >> Can you please provide output of following from all the nodes: >> >> cat /var/lib/glusterd/glusterd.info >> cat /var/lib/glusterd/peers/* >> >> On Sat, 13 May 2017 at 12:22, Pawan Alwandi <pawan at platform.sh> wrote: >> >>> Hello folks, >>> >>> Does anyone have any idea whats going on here? >>> >>> Thanks, >>> Pawan >>> >>> On Wed, May 10, 2017 at 5:02 PM, Pawan Alwandi <pawan at platform.sh> >>> wrote: >>> >>>> Hello, >>>> >>>> I'm trying to upgrade gluster from 3.6.2 to 3.10.1 but don't see the >>>> glusterfsd and glusterfs processes coming up. >>>> http://gluster.readthedocs.io/en/latest/Upgrade-Guide/upgrade_to_3.10/ >>>> is the process that I'm trying to follow. >>>> >>>> This is a 3 node server setup with a replicated volume having replica >>>> count of 3. >>>> >>>> Logs below: >>>> >>>> [2017-05-10 09:07:03.507959] I [MSGID: 100030] [glusterfsd.c:2460:main] >>>> 0-/usr/sbin/glusterd: Started running /usr/sbin/glusterd version 3.10.1 >>>> (args: /usr/sbin/glusterd -p /var/run/glusterd.pid) >>>> [2017-05-10 09:07:03.512827] I [MSGID: 106478] [glusterd.c:1449:init] >>>> 0-management: Maximum allowed open file descriptors set to 65536 >>>> [2017-05-10 09:07:03.512855] I [MSGID: 106479] [glusterd.c:1496:init] >>>> 0-management: Using /var/lib/glusterd as working directory >>>> [2017-05-10 09:07:03.520426] W [MSGID: 103071] >>>> [rdma.c:4590:__gf_rdma_ctx_create] 0-rpc-transport/rdma: rdma_cm event >>>> channel creation failed [No such device] >>>> [2017-05-10 09:07:03.520452] W [MSGID: 103055] [rdma.c:4897:init] >>>> 0-rdma.management: Failed to initialize IB Device >>>> [2017-05-10 09:07:03.520465] W [rpc-transport.c:350:rpc_transport_load] >>>> 0-rpc-transport: 'rdma' initialization failed >>>> [2017-05-10 09:07:03.520518] W [rpcsvc.c:1661:rpcsvc_create_listener] >>>> 0-rpc-service: cannot create listener, initing the transport failed >>>> [2017-05-10 09:07:03.520534] E [MSGID: 106243] [glusterd.c:1720:init] >>>> 0-management: creation of 1 listeners failed, continuing with succeeded >>>> transport >>>> [2017-05-10 09:07:04.931764] I [MSGID: 106513] >>>> [glusterd-store.c:2197:glusterd_restore_op_version] 0-glusterd: >>>> retrieved op-version: 30600 >>>> [2017-05-10 09:07:04.964354] I [MSGID: 106544] >>>> [glusterd.c:158:glusterd_uuid_init] 0-management: retrieved UUID: >>>> 7f2a6e11-2a53-4ab4-9ceb-8be6a9f2d073 >>>> [2017-05-10 09:07:04.993944] I [MSGID: 106498] >>>> [glusterd-handler.c:3669:glusterd_friend_add_from_peerinfo] >>>> 0-management: connect returned 0 >>>> [2017-05-10 09:07:04.995864] I [MSGID: 106498] >>>> [glusterd-handler.c:3669:glusterd_friend_add_from_peerinfo] >>>> 0-management: connect returned 0 >>>> [2017-05-10 09:07:04.995879] W [MSGID: 106062] >>>> [glusterd-handler.c:3466:glusterd_transport_inet_options_build] >>>> 0-glusterd: Failed to get tcp-user-timeout >>>> [2017-05-10 09:07:04.995903] I [rpc-clnt.c:1059:rpc_clnt_connection_init] >>>> 0-management: setting frame-timeout to 600 >>>> [2017-05-10 09:07:04.996325] I [rpc-clnt.c:1059:rpc_clnt_connection_init] >>>> 0-management: setting frame-timeout to 600 >>>> Final graph: >>>> +----------------------------------------------------------- >>>> -------------------+ >>>> 1: volume management >>>> 2: type mgmt/glusterd >>>> 3: option rpc-auth.auth-glusterfs on >>>> 4: option rpc-auth.auth-unix on >>>> 5: option rpc-auth.auth-null on >>>> 6: option rpc-auth-allow-insecure on >>>> 7: option transport.socket.listen-backlog 128 >>>> 8: option event-threads 1 >>>> 9: option ping-timeout 0 >>>> 10: option transport.socket.read-fail-log off >>>> 11: option transport.socket.keepalive-interval 2 >>>> 12: option transport.socket.keepalive-time 10 >>>> 13: option transport-type rdma >>>> 14: option working-directory /var/lib/glusterd >>>> 15: end-volume >>>> 16: >>>> +----------------------------------------------------------- >>>> -------------------+ >>>> [2017-05-10 09:07:04.996310] W [MSGID: 106062] >>>> [glusterd-handler.c:3466:glusterd_transport_inet_options_build] >>>> 0-glusterd: Failed to get tcp-user-timeout >>>> [2017-05-10 09:07:05.000461] I [MSGID: 101190] >>>> [event-epoll.c:629:event_dispatch_epoll_worker] 0-epoll: Started >>>> thread with index 1 >>>> [2017-05-10 09:07:05.001493] W [socket.c:593:__socket_rwv] >>>> 0-management: readv on 192.168.0.7:24007 failed (No data available) >>>> [2017-05-10 09:07:05.001513] I [MSGID: 106004] >>>> [glusterd-handler.c:5882:__glusterd_peer_rpc_notify] 0-management: >>>> Peer <192.168.0.7> (<5ec54b4f-f60c-48c6-9e55-95f2bb58f633>), in state >>>> <Peer in Cluster>, h >>>> as disconnected from glusterd. >>>> [2017-05-10 09:07:05.001677] W [glusterd-locks.c:675:glusterd_mgmt_v3_unlock] >>>> (-->/usr/lib/x86_64-linux-gnu/glusterfs/3.10.1/xlator/mgmt/glusterd.so(+0x20559) >>>> [0x7f0bf9d74559] -->/usr/lib/x86_64-linux-gnu >>>> /glusterfs/3.10.1/xlator/mgmt/glusterd.so(+0x29cf0) [0x7f0bf9d7dcf0] >>>> -->/usr/lib/x86_64-linux-gnu/glusterfs/3.10.1/xlator/mgmt/glusterd.so(+0xd5ba3) >>>> [0x7f0bf9e29ba3] ) 0-management: Lock for vol shared no >>>> t held >>>> [2017-05-10 09:07:05.001696] W [MSGID: 106118] >>>> [glusterd-handler.c:5907:__glusterd_peer_rpc_notify] 0-management: >>>> Lock not released for shared >>>> [2017-05-10 09:07:05.003099] E [rpc-clnt.c:365:saved_frames_unwind] >>>> (--> /usr/lib/x86_64-linux-gnu/libglusterfs.so.0(_gf_log_callingfn+0x13c)[0x7f0bfeeca73c] >>>> (--> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(s >>>> aved_frames_unwind+0x1cf)[0x7f0bfec904bf] (--> >>>> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(saved_frames_destroy+0xe)[0x7f0bfec905de] >>>> (--> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(rpc_clnt_connection_ >>>> cleanup+0x >>>> 91)[0x7f0bfec91c21] (--> /usr/lib/x86_64-linux-gnu/libg >>>> frpc.so.0(rpc_clnt_notify+0x290)[0x7f0bfec92710] ))))) 0-management: >>>> forced unwinding frame type(GLUSTERD-DUMP) op(DUMP(1)) called at 2017-05-10 >>>> 09:0 >>>> 7:05.000627 (xid=0x1) >>>> [2017-05-10 09:07:05.003129] E [MSGID: 106167] >>>> [glusterd-handshake.c:2181:__glusterd_peer_dump_version_cbk] >>>> 0-management: Error through RPC layer, retry again later >>>> [2017-05-10 09:07:05.003251] W [socket.c:593:__socket_rwv] >>>> 0-management: readv on 192.168.0.6:24007 failed (No data available) >>>> [2017-05-10 09:07:05.003267] I [MSGID: 106004] >>>> [glusterd-handler.c:5882:__glusterd_peer_rpc_notify] 0-management: >>>> Peer <192.168.0.6> (<83e9a0b9-6bd5-483b-8516-d8928805ed95>), in state >>>> <Peer in Cluster>, h >>>> as disconnected from glusterd. >>>> [2017-05-10 09:07:05.003318] W [glusterd-locks.c:675:glusterd_mgmt_v3_unlock] >>>> (-->/usr/lib/x86_64-linux-gnu/glusterfs/3.10.1/xlator/mgmt/glusterd.so(+0x20559) >>>> [0x7f0bf9d74559] -->/usr/lib/x86_64-linux-gnu >>>> /glusterfs/3.10.1/xlator/mgmt/glusterd.so(+0x29cf0) [0x7f0bf9d7dcf0] >>>> -->/usr/lib/x86_64-linux-gnu/glusterfs/3.10.1/xlator/mgmt/glusterd.so(+0xd5ba3) >>>> [0x7f0bf9e29ba3] ) 0-management: Lock for vol shared no >>>> t held >>>> [2017-05-10 09:07:05.003329] W [MSGID: 106118] >>>> [glusterd-handler.c:5907:__glusterd_peer_rpc_notify] 0-management: >>>> Lock not released for shared >>>> [2017-05-10 09:07:05.003457] E [rpc-clnt.c:365:saved_frames_unwind] >>>> (--> /usr/lib/x86_64-linux-gnu/libglusterfs.so.0(_gf_log_callingfn+0x13c)[0x7f0bfeeca73c] >>>> (--> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(s >>>> aved_frames_unwind+0x1cf)[0x7f0bfec904bf] (--> >>>> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(saved_frames_destroy+0xe)[0x7f0bfec905de] >>>> (--> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(rpc_clnt_connection_ >>>> cleanup+0x >>>> 91)[0x7f0bfec91c21] (--> /usr/lib/x86_64-linux-gnu/libg >>>> frpc.so.0(rpc_clnt_notify+0x290)[0x7f0bfec92710] ))))) 0-management: >>>> forced unwinding frame type(GLUSTERD-DUMP) op(DUMP(1)) called at 2017-05-10 >>>> 09:0 >>>> 7:05.001407 (xid=0x1) >>>> >>>> There are a bunch of errors reported but I'm not sure which is signal >>>> and which ones are noise. Does anyone have any idea whats going on here? >>>> >>>> Thanks, >>>> Pawan >>>> >>>> _______________________________________________ >>> Gluster-users mailing list >>> Gluster-users at gluster.org >>> http://lists.gluster.org/mailman/listinfo/gluster-users >> >> -- >> - Atin (atinm) >> > >-------------- next part -------------- An HTML attachment was scrubbed... URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20170514/e4439a25/attachment.html>
Atin Mukherjee
2017-May-14 22:02 UTC
[Gluster-users] Failure while upgrading gluster to 3.10.1
On Sun, 14 May 2017 at 21:43, Atin Mukherjee <amukherj at redhat.com> wrote:> Allright, I see that you haven't bumped up the op-version. Can you please > execute: > > gluster v set all cluster.op-version 30101 and then restart glusterd on > all the nodes and check the brick status? >s/30101/31001> > On Sun, May 14, 2017 at 8:55 PM, Pawan Alwandi <pawan at platform.sh> wrote: > >> Hello Atin, >> >> Thanks for looking at this. Below is the output you requested for. >> >> Again, I'm seeing those errors after upgrading gluster on host 1. >> >> Host 1 >> >> # cat /var/lib/glusterd/glusterd.info >> UUID=7f2a6e11-2a53-4ab4-9ceb-8be6a9f2d073 >> operating-version=30600 >> >> # cat /var/lib/glusterd/peers/* >> uuid=5ec54b4f-f60c-48c6-9e55-95f2bb58f633 >> state=3 >> hostname1=192.168.0.7 >> uuid=83e9a0b9-6bd5-483b-8516-d8928805ed95 >> state=3 >> hostname1=192.168.0.6 >> >> # gluster --version >> glusterfs 3.10.1 >> >> Host 2 >> >> # cat /var/lib/glusterd/glusterd.info >> UUID=83e9a0b9-6bd5-483b-8516-d8928805ed95 >> operating-version=30600 >> >> # cat /var/lib/glusterd/peers/* >> uuid=5ec54b4f-f60c-48c6-9e55-95f2bb58f633 >> state=3 >> hostname1=192.168.0.7 >> uuid=7f2a6e11-2a53-4ab4-9ceb-8be6a9f2d073 >> state=3 >> hostname1=192.168.0.5 >> >> # gluster --version >> glusterfs 3.6.2 built on Jan 21 2015 14:23:44 >> >> Host 3 >> >> # cat /var/lib/glusterd/glusterd.info >> UUID=5ec54b4f-f60c-48c6-9e55-95f2bb58f633 >> operating-version=30600 >> >> # cat /var/lib/glusterd/peers/* >> uuid=7f2a6e11-2a53-4ab4-9ceb-8be6a9f2d073 >> state=3 >> hostname1=192.168.0.5 >> uuid=83e9a0b9-6bd5-483b-8516-d8928805ed95 >> state=3 >> hostname1=192.168.0.6 >> >> # gluster --version >> glusterfs 3.6.2 built on Jan 21 2015 14:23:44 >> >> >> >> On Sat, May 13, 2017 at 6:28 PM, Atin Mukherjee <amukherj at redhat.com> >> wrote: >> >>> I have already asked for the following earlier: >>> >>> Can you please provide output of following from all the nodes: >>> >>> cat /var/lib/glusterd/glusterd.info >>> cat /var/lib/glusterd/peers/* >>> >>> On Sat, 13 May 2017 at 12:22, Pawan Alwandi <pawan at platform.sh> wrote: >>> >>>> Hello folks, >>>> >>>> Does anyone have any idea whats going on here? >>>> >>>> Thanks, >>>> Pawan >>>> >>>> On Wed, May 10, 2017 at 5:02 PM, Pawan Alwandi <pawan at platform.sh> >>>> wrote: >>>> >>>>> Hello, >>>>> >>>>> I'm trying to upgrade gluster from 3.6.2 to 3.10.1 but don't see the >>>>> glusterfsd and glusterfs processes coming up. >>>>> http://gluster.readthedocs.io/en/latest/Upgrade-Guide/upgrade_to_3.10/ >>>>> is the process that I'm trying to follow. >>>>> >>>>> This is a 3 node server setup with a replicated volume having replica >>>>> count of 3. >>>>> >>>>> Logs below: >>>>> >>>>> [2017-05-10 09:07:03.507959] I [MSGID: 100030] >>>>> [glusterfsd.c:2460:main] 0-/usr/sbin/glusterd: Started running >>>>> /usr/sbin/glusterd version 3.10.1 (args: /usr/sbin/glusterd -p >>>>> /var/run/glusterd.pid) >>>>> [2017-05-10 09:07:03.512827] I [MSGID: 106478] [glusterd.c:1449:init] >>>>> 0-management: Maximum allowed open file descriptors set to 65536 >>>>> [2017-05-10 09:07:03.512855] I [MSGID: 106479] [glusterd.c:1496:init] >>>>> 0-management: Using /var/lib/glusterd as working directory >>>>> [2017-05-10 09:07:03.520426] W [MSGID: 103071] >>>>> [rdma.c:4590:__gf_rdma_ctx_create] 0-rpc-transport/rdma: rdma_cm event >>>>> channel creation failed [No such device] >>>>> [2017-05-10 09:07:03.520452] W [MSGID: 103055] [rdma.c:4897:init] >>>>> 0-rdma.management: Failed to initialize IB Device >>>>> [2017-05-10 09:07:03.520465] W >>>>> [rpc-transport.c:350:rpc_transport_load] 0-rpc-transport: 'rdma' >>>>> initialization failed >>>>> [2017-05-10 09:07:03.520518] W [rpcsvc.c:1661:rpcsvc_create_listener] >>>>> 0-rpc-service: cannot create listener, initing the transport failed >>>>> [2017-05-10 09:07:03.520534] E [MSGID: 106243] [glusterd.c:1720:init] >>>>> 0-management: creation of 1 listeners failed, continuing with succeeded >>>>> transport >>>>> [2017-05-10 09:07:04.931764] I [MSGID: 106513] >>>>> [glusterd-store.c:2197:glusterd_restore_op_version] 0-glusterd: retrieved >>>>> op-version: 30600 >>>>> [2017-05-10 09:07:04.964354] I [MSGID: 106544] >>>>> [glusterd.c:158:glusterd_uuid_init] 0-management: retrieved UUID: >>>>> 7f2a6e11-2a53-4ab4-9ceb-8be6a9f2d073 >>>>> [2017-05-10 09:07:04.993944] I [MSGID: 106498] >>>>> [glusterd-handler.c:3669:glusterd_friend_add_from_peerinfo] 0-management: >>>>> connect returned 0 >>>>> [2017-05-10 09:07:04.995864] I [MSGID: 106498] >>>>> [glusterd-handler.c:3669:glusterd_friend_add_from_peerinfo] 0-management: >>>>> connect returned 0 >>>>> [2017-05-10 09:07:04.995879] W [MSGID: 106062] >>>>> [glusterd-handler.c:3466:glusterd_transport_inet_options_build] 0-glusterd: >>>>> Failed to get tcp-user-timeout >>>>> [2017-05-10 09:07:04.995903] I >>>>> [rpc-clnt.c:1059:rpc_clnt_connection_init] 0-management: setting >>>>> frame-timeout to 600 >>>>> [2017-05-10 09:07:04.996325] I >>>>> [rpc-clnt.c:1059:rpc_clnt_connection_init] 0-management: setting >>>>> frame-timeout to 600 >>>>> Final graph: >>>>> >>>>> +------------------------------------------------------------------------------+ >>>>> 1: volume management >>>>> 2: type mgmt/glusterd >>>>> 3: option rpc-auth.auth-glusterfs on >>>>> 4: option rpc-auth.auth-unix on >>>>> 5: option rpc-auth.auth-null on >>>>> 6: option rpc-auth-allow-insecure on >>>>> 7: option transport.socket.listen-backlog 128 >>>>> 8: option event-threads 1 >>>>> 9: option ping-timeout 0 >>>>> 10: option transport.socket.read-fail-log off >>>>> 11: option transport.socket.keepalive-interval 2 >>>>> 12: option transport.socket.keepalive-time 10 >>>>> 13: option transport-type rdma >>>>> 14: option working-directory /var/lib/glusterd >>>>> 15: end-volume >>>>> 16: >>>>> >>>>> +------------------------------------------------------------------------------+ >>>>> [2017-05-10 09:07:04.996310] W [MSGID: 106062] >>>>> [glusterd-handler.c:3466:glusterd_transport_inet_options_build] 0-glusterd: >>>>> Failed to get tcp-user-timeout >>>>> [2017-05-10 09:07:05.000461] I [MSGID: 101190] >>>>> [event-epoll.c:629:event_dispatch_epoll_worker] 0-epoll: Started thread >>>>> with index 1 >>>>> [2017-05-10 09:07:05.001493] W [socket.c:593:__socket_rwv] >>>>> 0-management: readv on 192.168.0.7:24007 failed (No data available) >>>>> [2017-05-10 09:07:05.001513] I [MSGID: 106004] >>>>> [glusterd-handler.c:5882:__glusterd_peer_rpc_notify] 0-management: Peer >>>>> <192.168.0.7> (<5ec54b4f-f60c-48c6-9e55-95f2bb58f633>), in state <Peer in >>>>> Cluster>, h >>>>> as disconnected from glusterd. >>>>> [2017-05-10 09:07:05.001677] W >>>>> [glusterd-locks.c:675:glusterd_mgmt_v3_unlock] >>>>> (-->/usr/lib/x86_64-linux-gnu/glusterfs/3.10.1/xlator/mgmt/glusterd.so(+0x20559) >>>>> [0x7f0bf9d74559] -->/usr/lib/x86_64-linux-gnu >>>>> /glusterfs/3.10.1/xlator/mgmt/glusterd.so(+0x29cf0) [0x7f0bf9d7dcf0] >>>>> -->/usr/lib/x86_64-linux-gnu/glusterfs/3.10.1/xlator/mgmt/glusterd.so(+0xd5ba3) >>>>> [0x7f0bf9e29ba3] ) 0-management: Lock for vol shared no >>>>> t held >>>>> [2017-05-10 09:07:05.001696] W [MSGID: 106118] >>>>> [glusterd-handler.c:5907:__glusterd_peer_rpc_notify] 0-management: Lock not >>>>> released for shared >>>>> [2017-05-10 09:07:05.003099] E [rpc-clnt.c:365:saved_frames_unwind] >>>>> (--> >>>>> /usr/lib/x86_64-linux-gnu/libglusterfs.so.0(_gf_log_callingfn+0x13c)[0x7f0bfeeca73c] >>>>> (--> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(s >>>>> aved_frames_unwind+0x1cf)[0x7f0bfec904bf] (--> >>>>> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(saved_frames_destroy+0xe)[0x7f0bfec905de] >>>>> (--> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(rpc_clnt_connection_cleanup+0x >>>>> 91)[0x7f0bfec91c21] (--> >>>>> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(rpc_clnt_notify+0x290)[0x7f0bfec92710] >>>>> ))))) 0-management: forced unwinding frame type(GLUSTERD-DUMP) op(DUMP(1)) >>>>> called at 2017-05-10 09:0 >>>>> 7:05.000627 (xid=0x1) >>>>> [2017-05-10 09:07:05.003129] E [MSGID: 106167] >>>>> [glusterd-handshake.c:2181:__glusterd_peer_dump_version_cbk] 0-management: >>>>> Error through RPC layer, retry again later >>>>> [2017-05-10 09:07:05.003251] W [socket.c:593:__socket_rwv] >>>>> 0-management: readv on 192.168.0.6:24007 failed (No data available) >>>>> [2017-05-10 09:07:05.003267] I [MSGID: 106004] >>>>> [glusterd-handler.c:5882:__glusterd_peer_rpc_notify] 0-management: Peer >>>>> <192.168.0.6> (<83e9a0b9-6bd5-483b-8516-d8928805ed95>), in state <Peer in >>>>> Cluster>, h >>>>> as disconnected from glusterd. >>>>> [2017-05-10 09:07:05.003318] W >>>>> [glusterd-locks.c:675:glusterd_mgmt_v3_unlock] >>>>> (-->/usr/lib/x86_64-linux-gnu/glusterfs/3.10.1/xlator/mgmt/glusterd.so(+0x20559) >>>>> [0x7f0bf9d74559] -->/usr/lib/x86_64-linux-gnu >>>>> /glusterfs/3.10.1/xlator/mgmt/glusterd.so(+0x29cf0) [0x7f0bf9d7dcf0] >>>>> -->/usr/lib/x86_64-linux-gnu/glusterfs/3.10.1/xlator/mgmt/glusterd.so(+0xd5ba3) >>>>> [0x7f0bf9e29ba3] ) 0-management: Lock for vol shared no >>>>> t held >>>>> [2017-05-10 09:07:05.003329] W [MSGID: 106118] >>>>> [glusterd-handler.c:5907:__glusterd_peer_rpc_notify] 0-management: Lock not >>>>> released for shared >>>>> [2017-05-10 09:07:05.003457] E [rpc-clnt.c:365:saved_frames_unwind] >>>>> (--> >>>>> /usr/lib/x86_64-linux-gnu/libglusterfs.so.0(_gf_log_callingfn+0x13c)[0x7f0bfeeca73c] >>>>> (--> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(s >>>>> aved_frames_unwind+0x1cf)[0x7f0bfec904bf] (--> >>>>> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(saved_frames_destroy+0xe)[0x7f0bfec905de] >>>>> (--> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(rpc_clnt_connection_cleanup+0x >>>>> 91)[0x7f0bfec91c21] (--> >>>>> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(rpc_clnt_notify+0x290)[0x7f0bfec92710] >>>>> ))))) 0-management: forced unwinding frame type(GLUSTERD-DUMP) op(DUMP(1)) >>>>> called at 2017-05-10 09:0 >>>>> 7:05.001407 (xid=0x1) >>>>> >>>>> There are a bunch of errors reported but I'm not sure which is signal >>>>> and which ones are noise. Does anyone have any idea whats going on here? >>>>> >>>>> Thanks, >>>>> Pawan >>>>> >>>>> _______________________________________________ >>>> Gluster-users mailing list >>>> Gluster-users at gluster.org >>>> http://lists.gluster.org/mailman/listinfo/gluster-users >>> >>> -- >>> - Atin (atinm) >>> >> >> > --- Atin (atinm) -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20170514/c3bd2a21/attachment.html>