Atin Mukherjee
2015-Dec-29 13:27 UTC
[Gluster-users] gluster volume status -> commit failed
It seems like your opCode is STATEDUMP instead of STATUS which is weird. Are you running a heterogeneous cluster? What is the last version you were running with? What's the current cluster op-version? Thanks, Atin On 12/29/2015 06:31 PM, Christophe TREFOIS wrote:> Dear all, > > I have a 3-node distribute setup with a controller of GlusterFS and > upgraded to 3.7.6 today and to CentOS 7.2. > > After the ugprade (reboot), I can start the volume fine and see mounted > volume as well on the controller. > > However, a gluster volume info <volname> results in an > > [root at stor104 glusterfs]# gluster volume status live > Commit failed on localhost. Please check the log file for more details. > > error. > > Below some information and log extracts. > > Thank you for any hints on where to start fixing this, > > Kind regards, > > ? > Christophe > > ?????? > > Here is the gluster info command for the volume. > > [root at stor104 glusterfs]# gluster volume info live > > Volume Name: live > Type: Distribute > Volume ID: 1328637d-7730-4627-8945-bbe43626d527 > Status: Started > Number of Bricks: 9 > Transport-type: tcp > Bricks: > Brick1: stor104:/zfs/brick0/brick > Brick2: stor104:/zfs/brick1/brick > Brick3: stor104:/zfs/brick2/brick > Brick4: stor106:/zfs/brick0/brick > Brick5: stor106:/zfs/brick1/brick > Brick6: stor106:/zfs/brick2/brick > Brick7: stor105:/zfs/brick0/brick > Brick8: stor105:/zfs/brick1/brick > Brick9: stor105:/zfs/brick2/brick > Options Reconfigured: > performance.io <http://performance.io>-thread-count: 8 > nfs.disable: on > performance.write-behind-window-size: 4MB > performance.client-io-threads: on > performance.cache-size: 1GB > performance.cache-refresh-timeout: 60 > performance.cache-max-file-size: 4MB > cluster.data-self-heal-algorithm: full > diagnostics.client-log-level: ERROR > diagnostics.brick-log-level: ERROR > cluster.min-free-disk: 1% > server.allow-insecure: on > > Relevant log parts when carrying out the command: > > ==> /var/log/glusterfs/cli.log <=> [2015-12-29 12:51:03.216998] I [cli.c:721:main] 0-cli: Started running > gluster with version 3.7.6 > [2015-12-29 12:51:03.226123] I > [cli-cmd-volume.c:1926:cli_check_gsync_present] 0-: geo-replication not > installed > [2015-12-29 12:51:03.226623] I [MSGID: 101190] > [event-epoll.c:632:event_dispatch_epoll_worker] 0-epoll: Started thread > with index 1 > [2015-12-29 12:51:03.226723] I [socket.c:2355:socket_event_handler] > 0-transport: disconnecting now > > ==> /var/log/glusterfs/cmd_history.log <=> [2015-12-29 12:51:03.236182] : volume status : SUCCESS > > ==> /var/log/glusterfs/etc-glusterfs-glusterd.vol.log <=> [2015-12-29 12:51:03.238562] I [MSGID: 106499] > [glusterd-handler.c:4267:__glusterd_handle_status_volume] 0-management: > Received status volume req for volume live > [2015-12-29 12:51:03.246043] E [MSGID: 106396] > [glusterd-op-sm.c:2985:_add_task_to_dict] 0-management: Statedump > operation doesn't have a task_id > [2015-12-29 12:51:03.246081] E [MSGID: 106060] > [glusterd-op-sm.c:3055:glusterd_aggregate_task_status] 0-management: > Failed to add task details to dict > [2015-12-29 12:51:03.246098] E [MSGID: 106123] > [glusterd-syncop.c:1404:gd_commit_op_phase] 0-management: Commit of > operation 'Volume Status' failed on localhost > > ==> /var/log/glusterfs/cmd_history.log <=> [2015-12-29 12:51:03.249364] : volume status : FAILED : Commit failed > on localhost. Please check the log file for more details. > > ==> /var/log/glusterfs/cli.log <=> [2015-12-29 12:51:03.249647] I [input.c:36:cli_batch] 0-: Exiting with: 0 > > All bricks are up > > [root at highlander glusterfs]# pdsh -g live 'df -h | grep brick*' > stor106: brick0 50T 33T 18T 66% /zfs/brick0 > stor106: brick1 50T 33T 18T 66% /zfs/brick1 > stor106: brick2 50T 33T 18T 66% /zfs/brick2 > stor105: brick0 40T 23T 18T 57% /zfs/brick0 > stor105: brick1 40T 23T 18T 57% /zfs/brick1 > stor105: brick2 40T 23T 18T 57% /zfs/brick2 > stor104: brick0 40T 23T 18T 57% /zfs/brick0 > stor104: brick1 40T 23T 18T 57% /zfs/brick1 > stor104: brick2 40T 23T 18T 57% /zfs/brick2 > > Package details > > [root at highlander glusterfs]# rpm -qa | grep gluster > glusterfs-client-xlators-3.7.6-1.el7.x86_64 > glusterfs-server-3.7.6-1.el7.x86_64 > samba-vfs-glusterfs-4.2.3-10.el7.x86_64 > glusterfs-3.7.6-1.el7.x86_64 > glusterfs-cli-3.7.6-1.el7.x86_64 > glusterfs-libs-3.7.6-1.el7.x86_64 > glusterfs-fuse-3.7.6-1.el7.x86_64 > glusterfs-api-3.7.6-1.el7.x86_64 > > [root at highlander glusterfs]# pdsh -g live 'rpm -qa | grep gluster' > stor105: glusterfs-libs-3.7.6-1.el7.x86_64 > stor105: glusterfs-api-3.7.6-1.el7.x86_64 > stor105: glusterfs-3.7.6-1.el7.x86_64 > stor105: glusterfs-fuse-3.7.6-1.el7.x86_64 > stor105: glusterfs-cli-3.7.6-1.el7.x86_64 > stor105: glusterfs-client-xlators-3.7.6-1.el7.x86_64 > stor105: glusterfs-server-3.7.6-1.el7.x86_64 > > stor104: glusterfs-server-3.7.6-1.el7.x86_64 > stor104: glusterfs-libs-3.7.6-1.el7.x86_64 > stor104: glusterfs-api-3.7.6-1.el7.x86_64 > stor104: glusterfs-3.7.6-1.el7.x86_64 > stor104: glusterfs-fuse-3.7.6-1.el7.x86_64 > stor104: glusterfs-cli-3.7.6-1.el7.x86_64 > stor104: glusterfs-client-xlators-3.7.6-1.el7.x86_64 > > stor106: glusterfs-3.7.6-1.el7.x86_64 > stor106: glusterfs-cli-3.7.6-1.el7.x86_64 > stor106: glusterfs-server-3.7.6-1.el7.x86_64 > stor106: glusterfs-libs-3.7.6-1.el7.x86_64 > stor106: glusterfs-client-xlators-3.7.6-1.el7.x86_64 > stor106: glusterfs-api-3.7.6-1.el7.x86_64 > stor106: glusterfs-fuse-3.7.6-1.el7.x86_64 > > More detailled logs: > > ==> /var/log/glusterfs/cli.log <=> [2015-12-29 12:57:23.520821] I [cli.c:721:main] 0-cli: Started running > gluster with version 3.7.6 > [2015-12-29 12:57:23.530898] I > [cli-cmd-volume.c:1926:cli_check_gsync_present] 0-: geo-replication not > installed > [2015-12-29 12:57:23.531844] I [MSGID: 101190] > [event-epoll.c:632:event_dispatch_epoll_worker] 0-epoll: Started thread > with index 1 > [2015-12-29 12:57:23.532004] I [socket.c:2355:socket_event_handler] > 0-transport: disconnecting now > > ==> /var/log/glusterfs/etc-glusterfs-glusterd.vol.log <=> [2015-12-29 12:57:23.534830] D [MSGID: 0] > [glusterd-op-sm.c:214:glusterd_generate_txn_id] 0-management: > Transaction_id = c2886398-03c1-4d32-924a-9f92367be85c > [2015-12-29 12:57:23.534946] D [MSGID: 0] > [glusterd-op-sm.c:311:glusterd_set_txn_opinfo] 0-management: > Successfully set opinfo for transaction ID : > c2886398-03c1-4d32-924a-9f92367be85c > [2015-12-29 12:57:23.534975] D [MSGID: 0] > [glusterd-op-sm.c:318:glusterd_set_txn_opinfo] 0-management: Returning 0 > [2015-12-29 12:57:23.535001] D [MSGID: 0] > [glusterd-syncop.c:1767:gd_sync_task_begin] 0-management: Transaction ID > : c2886398-03c1-4d32-924a-9f92367be85c > [2015-12-29 12:57:23.535031] D [MSGID: 0] > [glusterd-syncop.c:1807:gd_sync_task_begin] 0-glusterd: Failed to get > volume name > [2015-12-29 12:57:23.535078] D [MSGID: 0] > [glusterd-op-sm.c:1695:glusterd_op_stage_status_volume] 0-management: > Returning: 0 > [2015-12-29 12:57:23.535103] D [MSGID: 0] > [glusterd-op-sm.c:5515:glusterd_op_stage_validate] 0-management: OP > 18. Returning 0 > [2015-12-29 12:57:23.535555] D > [rpc-clnt-ping.c:98:rpc_clnt_remove_ping_timer_locked] (--> > /lib64/libglusterfs.so.0(_gf_log_callingfn+0x192)[0x7f410a9d5a82] (--> > /lib64/libgfrpc.so.0(rpc_clnt_remove_ping_timer_locked+0x8b)[0x7f410a7a587b] > (--> /lib64/libgfrpc.so.0(+0x13e74)[0x7f410a7a5e74] (--> > /lib64/libgfrpc.so.0(rpc_clnt_submit+0x34f)[0x7f410a7a1c4f] (--> > /usr/lib64/glusterfs/3.7.6/xlator/mgmt/glusterd.so(gd_syncop_submit_request+0x1a5)[0x7f40ff5d24b5] > ))))) 0-: 192.168.123.105:24007: ping timer event already removed > [2015-12-29 12:57:23.535935] D > [rpc-clnt-ping.c:98:rpc_clnt_remove_ping_timer_locked] (--> > /lib64/libglusterfs.so.0(_gf_log_callingfn+0x192)[0x7f410a9d5a82] (--> > /lib64/libgfrpc.so.0(rpc_clnt_remove_ping_timer_locked+0x8b)[0x7f410a7a587b] > (--> /lib64/libgfrpc.so.0(+0x13e74)[0x7f410a7a5e74] (--> > /lib64/libgfrpc.so.0(rpc_clnt_submit+0x34f)[0x7f410a7a1c4f] (--> > /usr/lib64/glusterfs/3.7.6/xlator/mgmt/glusterd.so(gd_syncop_submit_request+0x1a5)[0x7f40ff5d24b5] > ))))) 0-: 192.168.123.1:24007: ping timer event already removed > [2015-12-29 12:57:23.536279] D > [rpc-clnt-ping.c:98:rpc_clnt_remove_ping_timer_locked] (--> > /lib64/libglusterfs.so.0(_gf_log_callingfn+0x192)[0x7f410a9d5a82] (--> > /lib64/libgfrpc.so.0(rpc_clnt_remove_ping_timer_locked+0x8b)[0x7f410a7a587b] > (--> /lib64/libgfrpc.so.0(+0x13e74)[0x7f410a7a5e74] (--> > /lib64/libgfrpc.so.0(rpc_clnt_submit+0x34f)[0x7f410a7a1c4f] (--> > /usr/lib64/glusterfs/3.7.6/xlator/mgmt/glusterd.so(gd_syncop_submit_request+0x1a5)[0x7f40ff5d24b5] > ))))) 0-: 192.168.123.106:24007: ping timer event already removed > [2015-12-29 12:57:23.536358] D [MSGID: 0] > [glusterd-syncop.c:1312:gd_stage_op_phase] 0-management: Sent stage op > req for 'Volume Status' to 3 peers > [2015-12-29 12:57:23.538064] D [MSGID: 0] > [glusterd-peer-utils.c:200:glusterd_peerinfo_find_by_uuid] 0-management: > Friend found... state: Peer in Cluster > [2015-12-29 12:57:23.539059] D [logging.c:1952:_gf_msg_internal] > 0-logging-infra: Buffer overflow of a buffer whose size limit is 5. > About to flush least recently used log message to disk > The message "D [MSGID: 0] > [glusterd-peer-utils.c:200:glusterd_peerinfo_find_by_uuid] 0-management: > Friend found... state: Peer in Cluster" repeated 2 times between > [2015-12-29 12:57:23.538064] and [2015-12-29 12:57:23.538833] > [2015-12-29 12:57:23.539057] D [MSGID: 0] > [glusterd-op-sm.c:7010:glusterd_op_bricks_select] 0-management: Returning 0 > [2015-12-29 12:57:23.539163] D [MSGID: 0] > [glusterd-syncop.c:1709:gd_brick_op_phase] 0-management: Sent op req to > 0 bricks > [2015-12-29 12:57:23.539213] D [MSGID: 0] > [glusterd-op-sm.c:3360:glusterd_op_status_volume] 0-management: Returning 0 > [2015-12-29 12:57:23.539237] D [MSGID: 0] > [glusterd-op-sm.c:5642:glusterd_op_commit_perform] 0-management: Returning 0 > [2015-12-29 12:57:23.541960] D [MSGID: 0] > [glusterd-peer-utils.c:200:glusterd_peerinfo_find_by_uuid] 0-management: > Friend found... state: Peer in Cluster > [2015-12-29 12:57:23.542525] D [logging.c:1952:_gf_msg_internal] > 0-logging-infra: Buffer overflow of a buffer whose size limit is 5. > About to flush least recently used log message to disk > The message "D [MSGID: 0] > [glusterd-peer-utils.c:200:glusterd_peerinfo_find_by_uuid] 0-management: > Friend found... state: Peer in Cluster" repeated 2 times between > [2015-12-29 12:57:23.541960] and [2015-12-29 12:57:23.542186] > [2015-12-29 12:57:23.542523] D [MSGID: 0] > [glusterd-syncop.c:1449:gd_commit_op_phase] 0-management: Sent commit op > req for 'Volume Status' to 3 peers > [2015-12-29 12:57:23.542643] D [MSGID: 0] > [glusterd-op-sm.c:4487:glusterd_op_modify_op_ctx] 0-management: op_ctx > modification not required for status operation being performed > [2015-12-29 12:57:23.542680] D [MSGID: 0] > [glusterd-op-sm.c:255:glusterd_get_txn_opinfo] 0-management: > Successfully got opinfo for transaction ID : > c2886398-03c1-4d32-924a-9f92367be85c > [2015-12-29 12:57:23.542704] D [MSGID: 0] > [glusterd-op-sm.c:259:glusterd_get_txn_opinfo] 0-management: Returning 0 > [2015-12-29 12:57:23.542732] D [MSGID: 0] > [glusterd-op-sm.c:360:glusterd_clear_txn_opinfo] 0-management: > Successfully cleared opinfo for transaction ID : > c2886398-03c1-4d32-924a-9f92367be85c > [2015-12-29 12:57:23.542756] D [MSGID: 0] > [glusterd-op-sm.c:364:glusterd_clear_txn_opinfo] 0-management: Returning 0 > > ==> /var/log/glusterfs/cmd_history.log <=> [2015-12-29 12:57:23.542781] : volume status : SUCCESS > > ==> /var/log/glusterfs/etc-glusterfs-glusterd.vol.log <=> [2015-12-29 12:57:23.542847] D [MSGID: 0] > [glusterd-rpc-ops.c:205:glusterd_op_send_cli_response] 0-management: > Returning 0 > [2015-12-29 12:57:23.545697] I [MSGID: 106499] > [glusterd-handler.c:4267:__glusterd_handle_status_volume] 0-management: > Received status volume req for volume live > [2015-12-29 12:57:23.545818] D [MSGID: 0] > [glusterd-op-sm.c:214:glusterd_generate_txn_id] 0-management: > Transaction_id = fd3671e4-fa7a-4913-a6dd-b37884a3a715 > [2015-12-29 12:57:23.545872] D [MSGID: 0] > [glusterd-op-sm.c:311:glusterd_set_txn_opinfo] 0-management: > Successfully set opinfo for transaction ID : > fd3671e4-fa7a-4913-a6dd-b37884a3a715 > [2015-12-29 12:57:23.545905] D [MSGID: 0] > [glusterd-op-sm.c:318:glusterd_set_txn_opinfo] 0-management: Returning 0 > [2015-12-29 12:57:23.545926] D [MSGID: 0] > [glusterd-syncop.c:1767:gd_sync_task_begin] 0-management: Transaction ID > : fd3671e4-fa7a-4913-a6dd-b37884a3a715 > [2015-12-29 12:57:23.545978] D [MSGID: 0] > [glusterd-locks.c:562:glusterd_mgmt_v3_lock] 0-management: Trying to > acquire lock of vol live for 305c0f00-0f11-4da3-a470-50b6e6c14976 as > live_vol > [2015-12-29 12:57:23.546388] D [MSGID: 0] > [glusterd-locks.c:618:glusterd_mgmt_v3_lock] 0-management: Lock for vol > live successfully held by 305c0f00-0f11-4da3-a470-50b6e6c14976 > [2015-12-29 12:57:23.546478] D [MSGID: 0] > [glusterd-syncop.c:411:gd_syncop_mgmt_v3_lock] 0-glusterd: Returning 0 > [2015-12-29 12:57:23.549943] D [logging.c:1952:_gf_msg_internal] > 0-logging-infra: Buffer overflow of a buffer whose size limit is 5. > About to flush least recently used log message to disk > The message "D [MSGID: 0] [glusterd-syncop.c:411:gd_syncop_mgmt_v3_lock] > 0-glusterd: Returning 0" repeated 2 times between [2015-12-29 > 12:57:23.546478] and [2015-12-29 12:57:23.546565] > [2015-12-29 12:57:23.549941] D [MSGID: 0] > [glusterd-syncop.c:1205:gd_lock_op_phase] 0-management: Sent lock op req > for 'Volume Status' to 3 peers. Returning 0 > [2015-12-29 12:57:23.550063] D [MSGID: 0] > [glusterd-utils.c:1424:glusterd_volinfo_find] 0-management: Volume live > found > [2015-12-29 12:57:23.550087] D [MSGID: 0] > [glusterd-utils.c:1431:glusterd_volinfo_find] 0-management: Returning 0 > [2015-12-29 12:57:23.550121] D [MSGID: 0] > [glusterd-utils.c:1424:glusterd_volinfo_find] 0-management: Volume live > found > [2015-12-29 12:57:23.550140] D [MSGID: 0] > [glusterd-utils.c:1431:glusterd_volinfo_find] 0-management: Returning 0 > [2015-12-29 12:57:23.550167] D [MSGID: 0] > [glusterd-op-sm.c:1695:glusterd_op_stage_status_volume] 0-management: > Returning: 0 > [2015-12-29 12:57:23.550192] D [MSGID: 0] > [glusterd-op-sm.c:5515:glusterd_op_stage_validate] 0-management: OP > 18. Returning 0 > [2015-12-29 12:57:23.550309] D [MSGID: 0] > [glusterd-syncop.c:1312:gd_stage_op_phase] 0-management: Sent stage op > req for 'Volume Status' to 3 peers > [2015-12-29 12:57:23.552791] D [MSGID: 0] > [glusterd-peer-utils.c:200:glusterd_peerinfo_find_by_uuid] 0-management: > Friend found... state: Peer in Cluster > [2015-12-29 12:57:23.553395] D [logging.c:1952:_gf_msg_internal] > 0-logging-infra: Buffer overflow of a buffer whose size limit is 5. > About to flush least recently used log message to disk > The message "D [MSGID: 0] > [glusterd-peer-utils.c:200:glusterd_peerinfo_find_by_uuid] 0-management: > Friend found... state: Peer in Cluster" repeated 2 times between > [2015-12-29 12:57:23.552791] and [2015-12-29 12:57:23.553087] > [2015-12-29 12:57:23.553394] D [MSGID: 0] > [glusterd-op-sm.c:7010:glusterd_op_bricks_select] 0-management: Returning 0 > [2015-12-29 12:57:23.553499] D [MSGID: 0] > [glusterd-syncop.c:1709:gd_brick_op_phase] 0-management: Sent op req to > 0 bricks > [2015-12-29 12:57:23.553535] D [MSGID: 0] > [glusterd-utils.c:1424:glusterd_volinfo_find] 0-management: Volume live > found > [2015-12-29 12:57:23.553556] D [MSGID: 0] > [glusterd-utils.c:1431:glusterd_volinfo_find] 0-management: Returning 0 > [2015-12-29 12:57:23.553786] D [MSGID: 0] > [glusterd-snapshot-utils.c:3595:glusterd_is_snapd_enabled] 0-management: > Key features.uss not present in the dict for volume live > [2015-12-29 12:57:23.554065] E [MSGID: 106396] > [glusterd-op-sm.c:2985:_add_task_to_dict] 0-management: Statedump > operation doesn't have a task_id > [2015-12-29 12:57:23.554097] E [MSGID: 106060] > [glusterd-op-sm.c:3055:glusterd_aggregate_task_status] 0-management: > Failed to add task details to dict > [2015-12-29 12:57:23.554117] D [MSGID: 0] > [glusterd-op-sm.c:3360:glusterd_op_status_volume] 0-management: Returning -1 > [2015-12-29 12:57:23.554135] D [MSGID: 0] > [glusterd-op-sm.c:5642:glusterd_op_commit_perform] 0-management: > Returning -1 > [2015-12-29 12:57:23.554155] E [MSGID: 106123] > [glusterd-syncop.c:1404:gd_commit_op_phase] 0-management: Commit of > operation 'Volume Status' failed on localhost > [2015-12-29 12:57:23.554240] D [MSGID: 0] > [glusterd-syncop.c:510:gd_syncop_mgmt_v3_unlock] 0-glusterd: Returning 0 > [2015-12-29 12:57:23.557147] D [logging.c:1952:_gf_msg_internal] > 0-logging-infra: Buffer overflow of a buffer whose size limit is 5. > About to flush least recently used log message to disk > The message "D [MSGID: 0] > [glusterd-syncop.c:510:gd_syncop_mgmt_v3_unlock] 0-glusterd: Returning > 0" repeated 2 times between [2015-12-29 12:57:23.554240] and [2015-12-29 > 12:57:23.554319] > [2015-12-29 12:57:23.557146] D [MSGID: 0] > [glusterd-syncop.c:1558:gd_unlock_op_phase] 0-management: Sent unlock op > req for 'Volume Status' to 3 peers. Returning 0 > [2015-12-29 12:57:23.557255] D [MSGID: 0] > [glusterd-locks.c:669:glusterd_mgmt_v3_unlock] 0-management: Trying to > release lock of vol live for 305c0f00-0f11-4da3-a470-50b6e6c14976 as > live_vol > [2015-12-29 12:57:23.557286] D [MSGID: 0] > [glusterd-locks.c:714:glusterd_mgmt_v3_unlock] 0-management: Lock for > vol live successfully released > [2015-12-29 12:57:23.557312] D [MSGID: 0] > [glusterd-op-sm.c:255:glusterd_get_txn_opinfo] 0-management: > Successfully got opinfo for transaction ID : > fd3671e4-fa7a-4913-a6dd-b37884a3a715 > [2015-12-29 12:57:23.557332] D [MSGID: 0] > [glusterd-op-sm.c:259:glusterd_get_txn_opinfo] 0-management: Returning 0 > [2015-12-29 12:57:23.557356] D [MSGID: 0] > [glusterd-op-sm.c:360:glusterd_clear_txn_opinfo] 0-management: > Successfully cleared opinfo for transaction ID : > fd3671e4-fa7a-4913-a6dd-b37884a3a715 > [2015-12-29 12:57:23.557420] D [MSGID: 0] > [glusterd-op-sm.c:364:glusterd_clear_txn_opinfo] 0-management: Returning 0 > > ==> /var/log/glusterfs/cmd_history.log <=> [2015-12-29 12:57:23.557447] : volume status : FAILED : Commit failed > on localhost. Please check the log file for more details. > > Dr Christophe Trefois, Dipl.-Ing. > Technical Specialist / Post-Doc > > UNIVERSIT? DU LUXEMBOURG > > LUXEMBOURG CENTRE FOR SYSTEMS BIOMEDICINE > Campus Belval | House of Biomedicine > 6, avenue du Swing > L-4367 Belvaux > T: +352 46 66 44 6124 > F: +352 46 66 44 6949 > http://www.uni.lu/lcsb > > Facebook <https://www.facebook.com/trefex> Twitter > <https://twitter.com/Trefex> Google Plus > <https://plus.google.com/+ChristopheTrefois/> Linkedin > <https://www.linkedin.com/in/trefoischristophe> skype > <http://skype:Trefex?call> > > ---- > This message is confidential and may contain privileged information. > It is intended for the named recipient only. > If you receive it in error please notify me and permanently delete the > original message and any copies. > ---- > > > > > > _______________________________________________ > Gluster-users mailing list > Gluster-users at gluster.org > http://www.gluster.org/mailman/listinfo/gluster-users >
Christophe TREFOIS
2015-Dec-29 13:39 UTC
[Gluster-users] gluster volume status -> commit failed
Hi,> On 29 Dec 2015, at 14:27, Atin Mukherjee <amukherj at redhat.com> wrote: > > It seems like your opCode is STATEDUMP instead of STATUS which is weird. > Are you running a heterogeneous cluster?What does that mean? In principle no.> What is the last version you > were running with?I think it was 3.7.3 or 3.7.something. I?m not sure, but it was in the 3.7 repo.> What's the current cluster op-version?[root at highlander ~]# gluster volume get live cluster.op-version Option Value ------ ----- cluster.op-version 30600 Thank you,> > Thanks, > Atin > > On 12/29/2015 06:31 PM, Christophe TREFOIS wrote: >> Dear all, >> >> I have a 3-node distribute setup with a controller of GlusterFS and >> upgraded to 3.7.6 today and to CentOS 7.2. >> >> After the ugprade (reboot), I can start the volume fine and see mounted >> volume as well on the controller. >> >> However, a gluster volume info <volname> results in an >> >> [root at stor104 glusterfs]# gluster volume status live >> Commit failed on localhost. Please check the log file for more details. >> >> error. >> >> Below some information and log extracts. >> >> Thank you for any hints on where to start fixing this, >> >> Kind regards, >> >> ? >> Christophe >> >> ?????? >> >> Here is the gluster info command for the volume. >> >> [root at stor104 glusterfs]# gluster volume info live >> >> Volume Name: live >> Type: Distribute >> Volume ID: 1328637d-7730-4627-8945-bbe43626d527 >> Status: Started >> Number of Bricks: 9 >> Transport-type: tcp >> Bricks: >> Brick1: stor104:/zfs/brick0/brick >> Brick2: stor104:/zfs/brick1/brick >> Brick3: stor104:/zfs/brick2/brick >> Brick4: stor106:/zfs/brick0/brick >> Brick5: stor106:/zfs/brick1/brick >> Brick6: stor106:/zfs/brick2/brick >> Brick7: stor105:/zfs/brick0/brick >> Brick8: stor105:/zfs/brick1/brick >> Brick9: stor105:/zfs/brick2/brick >> Options Reconfigured: >> performance.io <http://performance.io>-thread-count: 8 >> nfs.disable: on >> performance.write-behind-window-size: 4MB >> performance.client-io-threads: on >> performance.cache-size: 1GB >> performance.cache-refresh-timeout: 60 >> performance.cache-max-file-size: 4MB >> cluster.data-self-heal-algorithm: full >> diagnostics.client-log-level: ERROR >> diagnostics.brick-log-level: ERROR >> cluster.min-free-disk: 1% >> server.allow-insecure: on >> >> Relevant log parts when carrying out the command: >> >> ==> /var/log/glusterfs/cli.log <=>> [2015-12-29 12:51:03.216998] I [cli.c:721:main] 0-cli: Started running >> gluster with version 3.7.6 >> [2015-12-29 12:51:03.226123] I >> [cli-cmd-volume.c:1926:cli_check_gsync_present] 0-: geo-replication not >> installed >> [2015-12-29 12:51:03.226623] I [MSGID: 101190] >> [event-epoll.c:632:event_dispatch_epoll_worker] 0-epoll: Started thread >> with index 1 >> [2015-12-29 12:51:03.226723] I [socket.c:2355:socket_event_handler] >> 0-transport: disconnecting now >> >> ==> /var/log/glusterfs/cmd_history.log <=>> [2015-12-29 12:51:03.236182] : volume status : SUCCESS >> >> ==> /var/log/glusterfs/etc-glusterfs-glusterd.vol.log <=>> [2015-12-29 12:51:03.238562] I [MSGID: 106499] >> [glusterd-handler.c:4267:__glusterd_handle_status_volume] 0-management: >> Received status volume req for volume live >> [2015-12-29 12:51:03.246043] E [MSGID: 106396] >> [glusterd-op-sm.c:2985:_add_task_to_dict] 0-management: Statedump >> operation doesn't have a task_id >> [2015-12-29 12:51:03.246081] E [MSGID: 106060] >> [glusterd-op-sm.c:3055:glusterd_aggregate_task_status] 0-management: >> Failed to add task details to dict >> [2015-12-29 12:51:03.246098] E [MSGID: 106123] >> [glusterd-syncop.c:1404:gd_commit_op_phase] 0-management: Commit of >> operation 'Volume Status' failed on localhost >> >> ==> /var/log/glusterfs/cmd_history.log <=>> [2015-12-29 12:51:03.249364] : volume status : FAILED : Commit failed >> on localhost. Please check the log file for more details. >> >> ==> /var/log/glusterfs/cli.log <=>> [2015-12-29 12:51:03.249647] I [input.c:36:cli_batch] 0-: Exiting with: 0 >> >> All bricks are up >> >> [root at highlander glusterfs]# pdsh -g live 'df -h | grep brick*' >> stor106: brick0 50T 33T 18T 66% /zfs/brick0 >> stor106: brick1 50T 33T 18T 66% /zfs/brick1 >> stor106: brick2 50T 33T 18T 66% /zfs/brick2 >> stor105: brick0 40T 23T 18T 57% /zfs/brick0 >> stor105: brick1 40T 23T 18T 57% /zfs/brick1 >> stor105: brick2 40T 23T 18T 57% /zfs/brick2 >> stor104: brick0 40T 23T 18T 57% /zfs/brick0 >> stor104: brick1 40T 23T 18T 57% /zfs/brick1 >> stor104: brick2 40T 23T 18T 57% /zfs/brick2 >> >> Package details >> >> [root at highlander glusterfs]# rpm -qa | grep gluster >> glusterfs-client-xlators-3.7.6-1.el7.x86_64 >> glusterfs-server-3.7.6-1.el7.x86_64 >> samba-vfs-glusterfs-4.2.3-10.el7.x86_64 >> glusterfs-3.7.6-1.el7.x86_64 >> glusterfs-cli-3.7.6-1.el7.x86_64 >> glusterfs-libs-3.7.6-1.el7.x86_64 >> glusterfs-fuse-3.7.6-1.el7.x86_64 >> glusterfs-api-3.7.6-1.el7.x86_64 >> >> [root at highlander glusterfs]# pdsh -g live 'rpm -qa | grep gluster' >> stor105: glusterfs-libs-3.7.6-1.el7.x86_64 >> stor105: glusterfs-api-3.7.6-1.el7.x86_64 >> stor105: glusterfs-3.7.6-1.el7.x86_64 >> stor105: glusterfs-fuse-3.7.6-1.el7.x86_64 >> stor105: glusterfs-cli-3.7.6-1.el7.x86_64 >> stor105: glusterfs-client-xlators-3.7.6-1.el7.x86_64 >> stor105: glusterfs-server-3.7.6-1.el7.x86_64 >> >> stor104: glusterfs-server-3.7.6-1.el7.x86_64 >> stor104: glusterfs-libs-3.7.6-1.el7.x86_64 >> stor104: glusterfs-api-3.7.6-1.el7.x86_64 >> stor104: glusterfs-3.7.6-1.el7.x86_64 >> stor104: glusterfs-fuse-3.7.6-1.el7.x86_64 >> stor104: glusterfs-cli-3.7.6-1.el7.x86_64 >> stor104: glusterfs-client-xlators-3.7.6-1.el7.x86_64 >> >> stor106: glusterfs-3.7.6-1.el7.x86_64 >> stor106: glusterfs-cli-3.7.6-1.el7.x86_64 >> stor106: glusterfs-server-3.7.6-1.el7.x86_64 >> stor106: glusterfs-libs-3.7.6-1.el7.x86_64 >> stor106: glusterfs-client-xlators-3.7.6-1.el7.x86_64 >> stor106: glusterfs-api-3.7.6-1.el7.x86_64 >> stor106: glusterfs-fuse-3.7.6-1.el7.x86_64 >> >> More detailled logs: >> >> ==> /var/log/glusterfs/cli.log <=>> [2015-12-29 12:57:23.520821] I [cli.c:721:main] 0-cli: Started running >> gluster with version 3.7.6 >> [2015-12-29 12:57:23.530898] I >> [cli-cmd-volume.c:1926:cli_check_gsync_present] 0-: geo-replication not >> installed >> [2015-12-29 12:57:23.531844] I [MSGID: 101190] >> [event-epoll.c:632:event_dispatch_epoll_worker] 0-epoll: Started thread >> with index 1 >> [2015-12-29 12:57:23.532004] I [socket.c:2355:socket_event_handler] >> 0-transport: disconnecting now >> >> ==> /var/log/glusterfs/etc-glusterfs-glusterd.vol.log <=>> [2015-12-29 12:57:23.534830] D [MSGID: 0] >> [glusterd-op-sm.c:214:glusterd_generate_txn_id] 0-management: >> Transaction_id = c2886398-03c1-4d32-924a-9f92367be85c >> [2015-12-29 12:57:23.534946] D [MSGID: 0] >> [glusterd-op-sm.c:311:glusterd_set_txn_opinfo] 0-management: >> Successfully set opinfo for transaction ID : >> c2886398-03c1-4d32-924a-9f92367be85c >> [2015-12-29 12:57:23.534975] D [MSGID: 0] >> [glusterd-op-sm.c:318:glusterd_set_txn_opinfo] 0-management: Returning 0 >> [2015-12-29 12:57:23.535001] D [MSGID: 0] >> [glusterd-syncop.c:1767:gd_sync_task_begin] 0-management: Transaction ID >> : c2886398-03c1-4d32-924a-9f92367be85c >> [2015-12-29 12:57:23.535031] D [MSGID: 0] >> [glusterd-syncop.c:1807:gd_sync_task_begin] 0-glusterd: Failed to get >> volume name >> [2015-12-29 12:57:23.535078] D [MSGID: 0] >> [glusterd-op-sm.c:1695:glusterd_op_stage_status_volume] 0-management: >> Returning: 0 >> [2015-12-29 12:57:23.535103] D [MSGID: 0] >> [glusterd-op-sm.c:5515:glusterd_op_stage_validate] 0-management: OP >> 18. Returning 0 >> [2015-12-29 12:57:23.535555] D >> [rpc-clnt-ping.c:98:rpc_clnt_remove_ping_timer_locked] (--> >> /lib64/libglusterfs.so.0(_gf_log_callingfn+0x192)[0x7f410a9d5a82] (--> >> /lib64/libgfrpc.so.0(rpc_clnt_remove_ping_timer_locked+0x8b)[0x7f410a7a587b] >> (--> /lib64/libgfrpc.so.0(+0x13e74)[0x7f410a7a5e74] (--> >> /lib64/libgfrpc.so.0(rpc_clnt_submit+0x34f)[0x7f410a7a1c4f] (--> >> /usr/lib64/glusterfs/3.7.6/xlator/mgmt/glusterd.so(gd_syncop_submit_request+0x1a5)[0x7f40ff5d24b5] >> ))))) 0-: 192.168.123.105:24007: ping timer event already removed >> [2015-12-29 12:57:23.535935] D >> [rpc-clnt-ping.c:98:rpc_clnt_remove_ping_timer_locked] (--> >> /lib64/libglusterfs.so.0(_gf_log_callingfn+0x192)[0x7f410a9d5a82] (--> >> /lib64/libgfrpc.so.0(rpc_clnt_remove_ping_timer_locked+0x8b)[0x7f410a7a587b] >> (--> /lib64/libgfrpc.so.0(+0x13e74)[0x7f410a7a5e74] (--> >> /lib64/libgfrpc.so.0(rpc_clnt_submit+0x34f)[0x7f410a7a1c4f] (--> >> /usr/lib64/glusterfs/3.7.6/xlator/mgmt/glusterd.so(gd_syncop_submit_request+0x1a5)[0x7f40ff5d24b5] >> ))))) 0-: 192.168.123.1:24007: ping timer event already removed >> [2015-12-29 12:57:23.536279] D >> [rpc-clnt-ping.c:98:rpc_clnt_remove_ping_timer_locked] (--> >> /lib64/libglusterfs.so.0(_gf_log_callingfn+0x192)[0x7f410a9d5a82] (--> >> /lib64/libgfrpc.so.0(rpc_clnt_remove_ping_timer_locked+0x8b)[0x7f410a7a587b] >> (--> /lib64/libgfrpc.so.0(+0x13e74)[0x7f410a7a5e74] (--> >> /lib64/libgfrpc.so.0(rpc_clnt_submit+0x34f)[0x7f410a7a1c4f] (--> >> /usr/lib64/glusterfs/3.7.6/xlator/mgmt/glusterd.so(gd_syncop_submit_request+0x1a5)[0x7f40ff5d24b5] >> ))))) 0-: 192.168.123.106:24007: ping timer event already removed >> [2015-12-29 12:57:23.536358] D [MSGID: 0] >> [glusterd-syncop.c:1312:gd_stage_op_phase] 0-management: Sent stage op >> req for 'Volume Status' to 3 peers >> [2015-12-29 12:57:23.538064] D [MSGID: 0] >> [glusterd-peer-utils.c:200:glusterd_peerinfo_find_by_uuid] 0-management: >> Friend found... state: Peer in Cluster >> [2015-12-29 12:57:23.539059] D [logging.c:1952:_gf_msg_internal] >> 0-logging-infra: Buffer overflow of a buffer whose size limit is 5. >> About to flush least recently used log message to disk >> The message "D [MSGID: 0] >> [glusterd-peer-utils.c:200:glusterd_peerinfo_find_by_uuid] 0-management: >> Friend found... state: Peer in Cluster" repeated 2 times between >> [2015-12-29 12:57:23.538064] and [2015-12-29 12:57:23.538833] >> [2015-12-29 12:57:23.539057] D [MSGID: 0] >> [glusterd-op-sm.c:7010:glusterd_op_bricks_select] 0-management: Returning 0 >> [2015-12-29 12:57:23.539163] D [MSGID: 0] >> [glusterd-syncop.c:1709:gd_brick_op_phase] 0-management: Sent op req to >> 0 bricks >> [2015-12-29 12:57:23.539213] D [MSGID: 0] >> [glusterd-op-sm.c:3360:glusterd_op_status_volume] 0-management: Returning 0 >> [2015-12-29 12:57:23.539237] D [MSGID: 0] >> [glusterd-op-sm.c:5642:glusterd_op_commit_perform] 0-management: Returning 0 >> [2015-12-29 12:57:23.541960] D [MSGID: 0] >> [glusterd-peer-utils.c:200:glusterd_peerinfo_find_by_uuid] 0-management: >> Friend found... state: Peer in Cluster >> [2015-12-29 12:57:23.542525] D [logging.c:1952:_gf_msg_internal] >> 0-logging-infra: Buffer overflow of a buffer whose size limit is 5. >> About to flush least recently used log message to disk >> The message "D [MSGID: 0] >> [glusterd-peer-utils.c:200:glusterd_peerinfo_find_by_uuid] 0-management: >> Friend found... state: Peer in Cluster" repeated 2 times between >> [2015-12-29 12:57:23.541960] and [2015-12-29 12:57:23.542186] >> [2015-12-29 12:57:23.542523] D [MSGID: 0] >> [glusterd-syncop.c:1449:gd_commit_op_phase] 0-management: Sent commit op >> req for 'Volume Status' to 3 peers >> [2015-12-29 12:57:23.542643] D [MSGID: 0] >> [glusterd-op-sm.c:4487:glusterd_op_modify_op_ctx] 0-management: op_ctx >> modification not required for status operation being performed >> [2015-12-29 12:57:23.542680] D [MSGID: 0] >> [glusterd-op-sm.c:255:glusterd_get_txn_opinfo] 0-management: >> Successfully got opinfo for transaction ID : >> c2886398-03c1-4d32-924a-9f92367be85c >> [2015-12-29 12:57:23.542704] D [MSGID: 0] >> [glusterd-op-sm.c:259:glusterd_get_txn_opinfo] 0-management: Returning 0 >> [2015-12-29 12:57:23.542732] D [MSGID: 0] >> [glusterd-op-sm.c:360:glusterd_clear_txn_opinfo] 0-management: >> Successfully cleared opinfo for transaction ID : >> c2886398-03c1-4d32-924a-9f92367be85c >> [2015-12-29 12:57:23.542756] D [MSGID: 0] >> [glusterd-op-sm.c:364:glusterd_clear_txn_opinfo] 0-management: Returning 0 >> >> ==> /var/log/glusterfs/cmd_history.log <=>> [2015-12-29 12:57:23.542781] : volume status : SUCCESS >> >> ==> /var/log/glusterfs/etc-glusterfs-glusterd.vol.log <=>> [2015-12-29 12:57:23.542847] D [MSGID: 0] >> [glusterd-rpc-ops.c:205:glusterd_op_send_cli_response] 0-management: >> Returning 0 >> [2015-12-29 12:57:23.545697] I [MSGID: 106499] >> [glusterd-handler.c:4267:__glusterd_handle_status_volume] 0-management: >> Received status volume req for volume live >> [2015-12-29 12:57:23.545818] D [MSGID: 0] >> [glusterd-op-sm.c:214:glusterd_generate_txn_id] 0-management: >> Transaction_id = fd3671e4-fa7a-4913-a6dd-b37884a3a715 >> [2015-12-29 12:57:23.545872] D [MSGID: 0] >> [glusterd-op-sm.c:311:glusterd_set_txn_opinfo] 0-management: >> Successfully set opinfo for transaction ID : >> fd3671e4-fa7a-4913-a6dd-b37884a3a715 >> [2015-12-29 12:57:23.545905] D [MSGID: 0] >> [glusterd-op-sm.c:318:glusterd_set_txn_opinfo] 0-management: Returning 0 >> [2015-12-29 12:57:23.545926] D [MSGID: 0] >> [glusterd-syncop.c:1767:gd_sync_task_begin] 0-management: Transaction ID >> : fd3671e4-fa7a-4913-a6dd-b37884a3a715 >> [2015-12-29 12:57:23.545978] D [MSGID: 0] >> [glusterd-locks.c:562:glusterd_mgmt_v3_lock] 0-management: Trying to >> acquire lock of vol live for 305c0f00-0f11-4da3-a470-50b6e6c14976 as >> live_vol >> [2015-12-29 12:57:23.546388] D [MSGID: 0] >> [glusterd-locks.c:618:glusterd_mgmt_v3_lock] 0-management: Lock for vol >> live successfully held by 305c0f00-0f11-4da3-a470-50b6e6c14976 >> [2015-12-29 12:57:23.546478] D [MSGID: 0] >> [glusterd-syncop.c:411:gd_syncop_mgmt_v3_lock] 0-glusterd: Returning 0 >> [2015-12-29 12:57:23.549943] D [logging.c:1952:_gf_msg_internal] >> 0-logging-infra: Buffer overflow of a buffer whose size limit is 5. >> About to flush least recently used log message to disk >> The message "D [MSGID: 0] [glusterd-syncop.c:411:gd_syncop_mgmt_v3_lock] >> 0-glusterd: Returning 0" repeated 2 times between [2015-12-29 >> 12:57:23.546478] and [2015-12-29 12:57:23.546565] >> [2015-12-29 12:57:23.549941] D [MSGID: 0] >> [glusterd-syncop.c:1205:gd_lock_op_phase] 0-management: Sent lock op req >> for 'Volume Status' to 3 peers. Returning 0 >> [2015-12-29 12:57:23.550063] D [MSGID: 0] >> [glusterd-utils.c:1424:glusterd_volinfo_find] 0-management: Volume live >> found >> [2015-12-29 12:57:23.550087] D [MSGID: 0] >> [glusterd-utils.c:1431:glusterd_volinfo_find] 0-management: Returning 0 >> [2015-12-29 12:57:23.550121] D [MSGID: 0] >> [glusterd-utils.c:1424:glusterd_volinfo_find] 0-management: Volume live >> found >> [2015-12-29 12:57:23.550140] D [MSGID: 0] >> [glusterd-utils.c:1431:glusterd_volinfo_find] 0-management: Returning 0 >> [2015-12-29 12:57:23.550167] D [MSGID: 0] >> [glusterd-op-sm.c:1695:glusterd_op_stage_status_volume] 0-management: >> Returning: 0 >> [2015-12-29 12:57:23.550192] D [MSGID: 0] >> [glusterd-op-sm.c:5515:glusterd_op_stage_validate] 0-management: OP >> 18. Returning 0 >> [2015-12-29 12:57:23.550309] D [MSGID: 0] >> [glusterd-syncop.c:1312:gd_stage_op_phase] 0-management: Sent stage op >> req for 'Volume Status' to 3 peers >> [2015-12-29 12:57:23.552791] D [MSGID: 0] >> [glusterd-peer-utils.c:200:glusterd_peerinfo_find_by_uuid] 0-management: >> Friend found... state: Peer in Cluster >> [2015-12-29 12:57:23.553395] D [logging.c:1952:_gf_msg_internal] >> 0-logging-infra: Buffer overflow of a buffer whose size limit is 5. >> About to flush least recently used log message to disk >> The message "D [MSGID: 0] >> [glusterd-peer-utils.c:200:glusterd_peerinfo_find_by_uuid] 0-management: >> Friend found... state: Peer in Cluster" repeated 2 times between >> [2015-12-29 12:57:23.552791] and [2015-12-29 12:57:23.553087] >> [2015-12-29 12:57:23.553394] D [MSGID: 0] >> [glusterd-op-sm.c:7010:glusterd_op_bricks_select] 0-management: Returning 0 >> [2015-12-29 12:57:23.553499] D [MSGID: 0] >> [glusterd-syncop.c:1709:gd_brick_op_phase] 0-management: Sent op req to >> 0 bricks >> [2015-12-29 12:57:23.553535] D [MSGID: 0] >> [glusterd-utils.c:1424:glusterd_volinfo_find] 0-management: Volume live >> found >> [2015-12-29 12:57:23.553556] D [MSGID: 0] >> [glusterd-utils.c:1431:glusterd_volinfo_find] 0-management: Returning 0 >> [2015-12-29 12:57:23.553786] D [MSGID: 0] >> [glusterd-snapshot-utils.c:3595:glusterd_is_snapd_enabled] 0-management: >> Key features.uss not present in the dict for volume live >> [2015-12-29 12:57:23.554065] E [MSGID: 106396] >> [glusterd-op-sm.c:2985:_add_task_to_dict] 0-management: Statedump >> operation doesn't have a task_id >> [2015-12-29 12:57:23.554097] E [MSGID: 106060] >> [glusterd-op-sm.c:3055:glusterd_aggregate_task_status] 0-management: >> Failed to add task details to dict >> [2015-12-29 12:57:23.554117] D [MSGID: 0] >> [glusterd-op-sm.c:3360:glusterd_op_status_volume] 0-management: Returning -1 >> [2015-12-29 12:57:23.554135] D [MSGID: 0] >> [glusterd-op-sm.c:5642:glusterd_op_commit_perform] 0-management: >> Returning -1 >> [2015-12-29 12:57:23.554155] E [MSGID: 106123] >> [glusterd-syncop.c:1404:gd_commit_op_phase] 0-management: Commit of >> operation 'Volume Status' failed on localhost >> [2015-12-29 12:57:23.554240] D [MSGID: 0] >> [glusterd-syncop.c:510:gd_syncop_mgmt_v3_unlock] 0-glusterd: Returning 0 >> [2015-12-29 12:57:23.557147] D [logging.c:1952:_gf_msg_internal] >> 0-logging-infra: Buffer overflow of a buffer whose size limit is 5. >> About to flush least recently used log message to disk >> The message "D [MSGID: 0] >> [glusterd-syncop.c:510:gd_syncop_mgmt_v3_unlock] 0-glusterd: Returning >> 0" repeated 2 times between [2015-12-29 12:57:23.554240] and [2015-12-29 >> 12:57:23.554319] >> [2015-12-29 12:57:23.557146] D [MSGID: 0] >> [glusterd-syncop.c:1558:gd_unlock_op_phase] 0-management: Sent unlock op >> req for 'Volume Status' to 3 peers. Returning 0 >> [2015-12-29 12:57:23.557255] D [MSGID: 0] >> [glusterd-locks.c:669:glusterd_mgmt_v3_unlock] 0-management: Trying to >> release lock of vol live for 305c0f00-0f11-4da3-a470-50b6e6c14976 as >> live_vol >> [2015-12-29 12:57:23.557286] D [MSGID: 0] >> [glusterd-locks.c:714:glusterd_mgmt_v3_unlock] 0-management: Lock for >> vol live successfully released >> [2015-12-29 12:57:23.557312] D [MSGID: 0] >> [glusterd-op-sm.c:255:glusterd_get_txn_opinfo] 0-management: >> Successfully got opinfo for transaction ID : >> fd3671e4-fa7a-4913-a6dd-b37884a3a715 >> [2015-12-29 12:57:23.557332] D [MSGID: 0] >> [glusterd-op-sm.c:259:glusterd_get_txn_opinfo] 0-management: Returning 0 >> [2015-12-29 12:57:23.557356] D [MSGID: 0] >> [glusterd-op-sm.c:360:glusterd_clear_txn_opinfo] 0-management: >> Successfully cleared opinfo for transaction ID : >> fd3671e4-fa7a-4913-a6dd-b37884a3a715 >> [2015-12-29 12:57:23.557420] D [MSGID: 0] >> [glusterd-op-sm.c:364:glusterd_clear_txn_opinfo] 0-management: Returning 0 >> >> ==> /var/log/glusterfs/cmd_history.log <=>> [2015-12-29 12:57:23.557447] : volume status : FAILED : Commit failed >> on localhost. Please check the log file for more details. >> >> Dr Christophe Trefois, Dipl.-Ing. >> Technical Specialist / Post-Doc >> >> UNIVERSIT? DU LUXEMBOURG >> >> LUXEMBOURG CENTRE FOR SYSTEMS BIOMEDICINE >> Campus Belval | House of Biomedicine >> 6, avenue du Swing >> L-4367 Belvaux >> T: +352 46 66 44 6124 >> F: +352 46 66 44 6949 >> http://www.uni.lu/lcsb >> >> Facebook <https://www.facebook.com/trefex> Twitter >> <https://twitter.com/Trefex> Google Plus >> <https://plus.google.com/+ChristopheTrefois/> Linkedin >> <https://www.linkedin.com/in/trefoischristophe> skype >> <http://skype:Trefex?call> >> >> ---- >> This message is confidential and may contain privileged information. >> It is intended for the named recipient only. >> If you receive it in error please notify me and permanently delete the >> original message and any copies. >> ---- >> >> >> >> >> >> _______________________________________________ >> Gluster-users mailing list >> Gluster-users at gluster.org >> http://www.gluster.org/mailman/listinfo/gluster-users >>