I think I have a workaround here: Remove /var/lib/glusterd/vols/<volname>/node_state.info file from all the nodes and restart GlusterD Post that you should be able to execute gluster volume status successfully. Thanks, Atin On 12/30/2015 11:36 AM, Atin Mukherjee wrote: > I have some updates: > > Actually the opCode what we check here for aggregating status is for > rebalance process (volinfo->rebal.op), not for the incoming operation, > so I was wrong with my first analysis. However having said that > rebalance op could either be GD_OP_REMOVE_BRICK or GD_OP_REBALANCE. If > its a tier volume the opCode passed is either GD_OP_DETACH_TIER or > GD_OP_TIER_MIGRATE. By no means it could be GD_OP_STATEDUMP_VOLUME. > > Although I don't see remove brick/rebalance entries in cmd_history.log. > did you perform rebalance operation in the cluster? If so was it successful? > > This seems to a incorrect op set by rebalance/remove brick command. Can > some one from DHT take a look at it? > > Thanks, > Atin > > > > On 12/30/2015 02:42 AM, Christophe TREFOIS wrote: >> Hi Atin, >> >> Please find the logs attached. >> >> Hopefully it will be useful to figure out the issue. Please let me know >> if you need anything else, or if I can help somehow. >> >> Thank you! >> >> >> >>> On 29 Dec 2015, at 17:59, Atin Mukherjee <atin.mukherjee83@xxxxxxxxx >>> <mailto:atin.mukherjee83@xxxxxxxxx>> wrote: >>> >>> I'll dive into it in detail in some time. Could you provide >>> cli/cmd_history/glusterd log files for further debugging. >>> >>> -Atin >>> Sent from one plus one >>> >>> On Dec 29, 2015 9:53 PM, "Christophe TREFOIS" >>> <christophe.trefois@xxxxxx <mailto:christophe.trefois@xxxxxx>> wrote: >>> >>> Hi Atin, >>> >>> Same issue. I restarted glusterd and glusterfsd everywhere and it >>> seems the thing is still in STATEDUMP. >>> >>> Any other pointers? >>> >>> Kind regards, >>> >>> — >>> Christophe >>> >>> >>> > On 29 Dec 2015, at 16:19, Atin Mukherjee <amukherj@xxxxxxxxxx >>> <mailto:amukherj@xxxxxxxxxx>> wrote: >>> > >>> > >>> > >>> > On 12/29/2015 07:09 PM, Christophe TREFOIS wrote: >>> >> Hi, >>> >> >>> >> >>> >>> On 29 Dec 2015, at 14:27, Atin Mukherjee <amukherj@xxxxxxxxxx >>> <mailto:amukherj@xxxxxxxxxx>> wrote: >>> >>> >>> >>> It seems like your opCode is STATEDUMP instead of STATUS which >>> is weird. >>> >>> Are you running a heterogeneous cluster? >>> >> >>> >> What does that mean? In principle no. >>> > This means when all the nodes in the cluster are not running >>> with same bits. >>> >> >>> >>> What is the last version you >>> >>> were running with? >>> >> >>> >> I think it was 3.7.3 or 3.7.something. I’m not sure, but it was >>> in the 3.7 repo. >>> >> >>> >>> What's the current cluster op-version? >>> >> >>> >> [root@highlander ~]# gluster volume get live cluster.op-version >>> >> Option Value >>> >> ------ ----- >>> >> cluster.op-version 30600 >>> > Can you bump up the op-version with 'gluster volume set all >>> > cluster.op-version 30706 and see if the problem goes away. But >>> you can >>> > only bump this up if all nodes are upgraded to 3.7.6 >>> >> >>> >> Thank you, >>> >> >>> >>> >>> >>> Thanks, >>> >>> Atin >>> >>> >>> >>> On 12/29/2015 06:31 PM, Christophe TREFOIS wrote: >>> >>>> Dear all, >>> >>>> >>> >>>> I have a 3-node distribute setup with a controller of >>> GlusterFS and >>> >>>> upgraded to 3.7.6 today and to CentOS 7.2. >>> >>>> >>> >>>> After the ugprade (reboot), I can start the volume fine and >>> see mounted >>> >>>> volume as well on the controller. >>> >>>> >>> >>>> However, a gluster volume info <volname> results in an >>> >>>> >>> >>>> [root@stor104 glusterfs]# gluster volume status live >>> >>>> Commit failed on localhost. Please check the log file for >>> more details. >>> >>>> >>> >>>> error. >>> >>>> >>> >>>> Below some information and log extracts. >>> >>>> >>> >>>> Thank you for any hints on where to start fixing this, >>> >>>> >>> >>>> Kind regards, >>> >>>> >>> >>>> — >>> >>>> Christophe >>> >>>> >>> >>>> —————— >>> >>>> >>> >>>> Here is the gluster info command for the volume. >>> >>>> >>> >>>> [root@stor104 glusterfs]# gluster volume info live >>> >>>> >>> >>>> Volume Name: live >>> >>>> Type: Distribute >>> >>>> Volume ID: 1328637d-7730-4627-8945-bbe43626d527 >>> >>>> Status: Started >>> >>>> Number of Bricks: 9 >>> >>>> Transport-type: tcp >>> >>>> Bricks: >>> >>>> Brick1: stor104:/zfs/brick0/brick >>> >>>> Brick2: stor104:/zfs/brick1/brick >>> >>>> Brick3: stor104:/zfs/brick2/brick >>> >>>> Brick4: stor106:/zfs/brick0/brick >>> >>>> Brick5: stor106:/zfs/brick1/brick >>> >>>> Brick6: stor106:/zfs/brick2/brick >>> >>>> Brick7: stor105:/zfs/brick0/brick >>> >>>> Brick8: stor105:/zfs/brick1/brick >>> >>>> Brick9: stor105:/zfs/brick2/brick >>> >>>> Options Reconfigured: >>> >>>> performance.io <http://performance.io/> >>> <http://performance.io <http://performance.io/>>-thread-count: 8 >>> >>>> nfs.disable: on >>> >>>> performance.write-behind-window-size: 4MB >>> >>>> performance.client-io-threads: on >>> >>>> performance.cache-size: 1GB >>> >>>> performance.cache-refresh-timeout: 60 >>> >>>> performance.cache-max-file-size: 4MB >>> >>>> cluster.data-self-heal-algorithm: full >>> >>>> diagnostics.client-log-level: ERROR >>> >>>> diagnostics.brick-log-level: ERROR >>> >>>> cluster.min-free-disk: 1% >>> >>>> server.allow-insecure: on >>> >>>> >>> >>>> Relevant log parts when carrying out the command: >>> >>>> >>> >>>> ==> /var/log/glusterfs/cli.log <== >>> >>>> [2015-12-29 12:51:03.216998] I [cli.c:721:main] 0-cli: >>> Started running >>> >>>> gluster with version 3.7.6 >>> >>>> [2015-12-29 12:51:03.226123] I >>> >>>> [cli-cmd-volume.c:1926:cli_check_gsync_present] 0-: >>> geo-replication not >>> >>>> installed >>> >>>> [2015-12-29 12:51:03.226623] I [MSGID: 101190] >>> >>>> [event-epoll.c:632:event_dispatch_epoll_worker] 0-epoll: >>> Started thread >>> >>>> with index 1 >>> >>>> [2015-12-29 12:51:03.226723] I >>> [socket.c:2355:socket_event_handler] >>> >>>> 0-transport: disconnecting now >>> >>>> >>> >>>> ==> /var/log/glusterfs/cmd_history.log <== >>> >>>> [2015-12-29 12:51:03.236182] : volume status : SUCCESS >>> >>>> >>> >>>> ==> /var/log/glusterfs/etc-glusterfs-glusterd.vol.log <== >>> >>>> [2015-12-29 12:51:03.238562] I [MSGID: 106499] >>> >>>> [glusterd-handler.c:4267:__glusterd_handle_status_volume] >>> 0-management: >>> >>>> Received status volume req for volume live >>> >>>> [2015-12-29 12:51:03.246043] E [MSGID: 106396] >>> >>>> [glusterd-op-sm.c:2985:_add_task_to_dict] 0-management: Statedump >>> >>>> operation doesn't have a task_id >>> >>>> [2015-12-29 12:51:03.246081] E [MSGID: 106060] >>> >>>> [glusterd-op-sm.c:3055:glusterd_aggregate_task_status] >>> 0-management: >>> >>>> Failed to add task details to dict >>> >>>> [2015-12-29 12:51:03.246098] E [MSGID: 106123] >>> >>>> [glusterd-syncop.c:1404:gd_commit_op_phase] 0-management: >>> Commit of >>> >>>> operation 'Volume Status' failed on localhost >>> >>>> >>> >>>> ==> /var/log/glusterfs/cmd_history.log <== >>> >>>> [2015-12-29 12:51:03.249364] : volume status : FAILED : >>> Commit failed >>> >>>> on localhost. Please check the log file for more details. >>> >>>> >>> >>>> ==> /var/log/glusterfs/cli.log <== >>> >>>> [2015-12-29 12:51:03.249647] I [input.c:36:cli_batch] 0-: >>> Exiting with: 0 >>> >>>> >>> >>>> All bricks are up >>> >>>> >>> >>>> [root@highlander glusterfs]# pdsh -g live 'df -h | grep brick*' >>> >>>> stor106: brick0 50T 33T 18T 66% /zfs/brick0 >>> >>>> stor106: brick1 50T 33T 18T 66% /zfs/brick1 >>> >>>> stor106: brick2 50T 33T 18T 66% /zfs/brick2 >>> >>>> stor105: brick0 40T 23T 18T 57% /zfs/brick0 >>> >>>> stor105: brick1 40T 23T 18T 57% /zfs/brick1 >>> >>>> stor105: brick2 40T 23T 18T 57% /zfs/brick2 >>> >>>> stor104: brick0 40T 23T 18T 57% /zfs/brick0 >>> >>>> stor104: brick1 40T 23T 18T 57% /zfs/brick1 >>> >>>> stor104: brick2 40T 23T 18T 57% /zfs/brick2 >>> >>>> >>> >>>> Package details >>> >>>> >>> >>>> [root@highlander glusterfs]# rpm -qa | grep gluster >>> >>>> glusterfs-client-xlators-3.7.6-1.el7.x86_64 >>> >>>> glusterfs-server-3.7.6-1.el7.x86_64 >>> >>>> samba-vfs-glusterfs-4.2.3-10.el7.x86_64 >>> >>>> glusterfs-3.7.6-1.el7.x86_64 >>> >>>> glusterfs-cli-3.7.6-1.el7.x86_64 >>> >>>> glusterfs-libs-3.7.6-1.el7.x86_64 >>> >>>> glusterfs-fuse-3.7.6-1.el7.x86_64 >>> >>>> glusterfs-api-3.7.6-1.el7.x86_64 >>> >>>> >>> >>>> [root@highlander glusterfs]# pdsh -g live 'rpm -qa | grep >>> gluster' >>> >>>> stor105: glusterfs-libs-3.7.6-1.el7.x86_64 >>> >>>> stor105: glusterfs-api-3.7.6-1.el7.x86_64 >>> >>>> stor105: glusterfs-3.7.6-1.el7.x86_64 >>> >>>> stor105: glusterfs-fuse-3.7.6-1.el7.x86_64 >>> >>>> stor105: glusterfs-cli-3.7.6-1.el7.x86_64 >>> >>>> stor105: glusterfs-client-xlators-3.7.6-1.el7.x86_64 >>> >>>> stor105: glusterfs-server-3.7.6-1.el7.x86_64 >>> >>>> >>> >>>> stor104: glusterfs-server-3.7.6-1.el7.x86_64 >>> >>>> stor104: glusterfs-libs-3.7.6-1.el7.x86_64 >>> >>>> stor104: glusterfs-api-3.7.6-1.el7.x86_64 >>> >>>> stor104: glusterfs-3.7.6-1.el7.x86_64 >>> >>>> stor104: glusterfs-fuse-3.7.6-1.el7.x86_64 >>> >>>> stor104: glusterfs-cli-3.7.6-1.el7.x86_64 >>> >>>> stor104: glusterfs-client-xlators-3.7.6-1.el7.x86_64 >>> >>>> >>> >>>> stor106: glusterfs-3.7.6-1.el7.x86_64 >>> >>>> stor106: glusterfs-cli-3.7.6-1.el7.x86_64 >>> >>>> stor106: glusterfs-server-3.7.6-1.el7.x86_64 >>> >>>> stor106: glusterfs-libs-3.7.6-1.el7.x86_64 >>> >>>> stor106: glusterfs-client-xlators-3.7.6-1.el7.x86_64 >>> >>>> stor106: glusterfs-api-3.7.6-1.el7.x86_64 >>> >>>> stor106: glusterfs-fuse-3.7.6-1.el7.x86_64 >>> >>>> >>> >>>> More detailled logs: >>> >>>> >>> >>>> ==> /var/log/glusterfs/cli.log <== >>> >>>> [2015-12-29 12:57:23.520821] I [cli.c:721:main] 0-cli: >>> Started running >>> >>>> gluster with version 3.7.6 >>> >>>> [2015-12-29 12:57:23.530898] I >>> >>>> [cli-cmd-volume.c:1926:cli_check_gsync_present] 0-: >>> geo-replication not >>> >>>> installed >>> >>>> [2015-12-29 12:57:23.531844] I [MSGID: 101190] >>> >>>> [event-epoll.c:632:event_dispatch_epoll_worker] 0-epoll: >>> Started thread >>> >>>> with index 1 >>> >>>> [2015-12-29 12:57:23.532004] I >>> [socket.c:2355:socket_event_handler] >>> >>>> 0-transport: disconnecting now >>> >>>> >>> >>>> ==> /var/log/glusterfs/etc-glusterfs-glusterd.vol.log <== >>> >>>> [2015-12-29 12:57:23.534830] D [MSGID: 0] >>> >>>> [glusterd-op-sm.c:214:glusterd_generate_txn_id] 0-management: >>> >>>> Transaction_id = c2886398-03c1-4d32-924a-9f92367be85c >>> >>>> [2015-12-29 12:57:23.534946] D [MSGID: 0] >>> >>>> [glusterd-op-sm.c:311:glusterd_set_txn_opinfo] 0-management: >>> >>>> Successfully set opinfo for transaction ID : >>> >>>> c2886398-03c1-4d32-924a-9f92367be85c >>> >>>> [2015-12-29 12:57:23.534975] D [MSGID: 0] >>> >>>> [glusterd-op-sm.c:318:glusterd_set_txn_opinfo] 0-management: >>> Returning 0 >>> >>>> [2015-12-29 12:57:23.535001] D [MSGID: 0] >>> >>>> [glusterd-syncop.c:1767:gd_sync_task_begin] 0-management: >>> Transaction ID >>> >>>> : c2886398-03c1-4d32-924a-9f92367be85c >>> >>>> [2015-12-29 12:57:23.535031] D [MSGID: 0] >>> >>>> [glusterd-syncop.c:1807:gd_sync_task_begin] 0-glusterd: >>> Failed to get >>> >>>> volume name >>> >>>> [2015-12-29 12:57:23.535078] D [MSGID: 0] >>> >>>> [glusterd-op-sm.c:1695:glusterd_op_stage_status_volume] >>> 0-management: >>> >>>> Returning: 0 >>> >>>> [2015-12-29 12:57:23.535103] D [MSGID: 0] >>> >>>> [glusterd-op-sm.c:5515:glusterd_op_stage_validate] >>> 0-management: OP = >>> >>>> 18. Returning 0 >>> >>>> [2015-12-29 12:57:23.535555] D >>> >>>> [rpc-clnt-ping.c:98:rpc_clnt_remove_ping_timer_locked] (--> >>> >>>> >>> /lib64/libglusterfs.so.0(_gf_log_callingfn+0x192)[0x7f410a9d5a82] (--> >>> >>>> >>> /lib64/libgfrpc.so.0(rpc_clnt_remove_ping_timer_locked+0x8b)[0x7f410a7a587b] >>> >>>> (--> /lib64/libgfrpc.so.0(+0x13e74)[0x7f410a7a5e74] (--> >>> >>>> /lib64/libgfrpc.so.0(rpc_clnt_submit+0x34f)[0x7f410a7a1c4f] (--> >>> >>>> >>> /usr/lib64/glusterfs/3.7.6/xlator/mgmt/glusterd.so(gd_syncop_submit_request+0x1a5)[0x7f40ff5d24b5] >>> >>>> ))))) 0-: 192.168.123.105:24007 >>> <http://192.168.123.105:24007/>: ping timer event already removed >>> >>>> [2015-12-29 12:57:23.535935] D >>> >>>> [rpc-clnt-ping.c:98:rpc_clnt_remove_ping_timer_locked] (--> >>> >>>> >>> /lib64/libglusterfs.so.0(_gf_log_callingfn+0x192)[0x7f410a9d5a82] (--> >>> >>>> >>> /lib64/libgfrpc.so.0(rpc_clnt_remove_ping_timer_locked+0x8b)[0x7f410a7a587b] >>> >>>> (--> /lib64/libgfrpc.so.0(+0x13e74)[0x7f410a7a5e74] (--> >>> >>>> /lib64/libgfrpc.so.0(rpc_clnt_submit+0x34f)[0x7f410a7a1c4f] (--> >>> >>>> >>> /usr/lib64/glusterfs/3.7.6/xlator/mgmt/glusterd.so(gd_syncop_submit_request+0x1a5)[0x7f40ff5d24b5] >>> >>>> ))))) 0-: 192.168.123.1:24007 <http://192.168.123.1:24007/>: >>> ping timer event already removed >>> >>>> [2015-12-29 12:57:23.536279] D >>> >>>> [rpc-clnt-ping.c:98:rpc_clnt_remove_ping_timer_locked] (--> >>> >>>> >>> /lib64/libglusterfs.so.0(_gf_log_callingfn+0x192)[0x7f410a9d5a82] (--> >>> >>>> >>> /lib64/libgfrpc.so.0(rpc_clnt_remove_ping_timer_locked+0x8b)[0x7f410a7a587b] >>> >>>> (--> /lib64/libgfrpc.so.0(+0x13e74)[0x7f410a7a5e74] (--> >>> >>>> /lib64/libgfrpc.so.0(rpc_clnt_submit+0x34f)[0x7f410a7a1c4f] (--> >>> >>>> >>> /usr/lib64/glusterfs/3.7.6/xlator/mgmt/glusterd.so(gd_syncop_submit_request+0x1a5)[0x7f40ff5d24b5] >>> >>>> ))))) 0-: 192.168.123.106:24007 >>> <http://192.168.123.106:24007/>: ping timer event already removed >>> >>>> [2015-12-29 12:57:23.536358] D [MSGID: 0] >>> >>>> [glusterd-syncop.c:1312:gd_stage_op_phase] 0-management: Sent >>> stage op >>> >>>> req for 'Volume Status' to 3 peers >>> >>>> [2015-12-29 12:57:23.538064] D [MSGID: 0] >>> >>>> [glusterd-peer-utils.c:200:glusterd_peerinfo_find_by_uuid] >>> 0-management: >>> >>>> Friend found... state: Peer in Cluster >>> >>>> [2015-12-29 12:57:23.539059] D [logging.c:1952:_gf_msg_internal] >>> >>>> 0-logging-infra: Buffer overflow of a buffer whose size limit >>> is 5. >>> >>>> About to flush least recently used log message to disk >>> >>>> The message "D [MSGID: 0] >>> >>>> [glusterd-peer-utils.c:200:glusterd_peerinfo_find_by_uuid] >>> 0-management: >>> >>>> Friend found... state: Peer in Cluster" repeated 2 times between >>> >>>> [2015-12-29 12:57:23.538064] and [2015-12-29 12:57:23.538833] >>> >>>> [2015-12-29 12:57:23.539057] D [MSGID: 0] >>> >>>> [glusterd-op-sm.c:7010:glusterd_op_bricks_select] >>> 0-management: Returning 0 >>> >>>> [2015-12-29 12:57:23.539163] D [MSGID: 0] >>> >>>> [glusterd-syncop.c:1709:gd_brick_op_phase] 0-management: Sent >>> op req to >>> >>>> 0 bricks >>> >>>> [2015-12-29 12:57:23.539213] D [MSGID: 0] >>> >>>> [glusterd-op-sm.c:3360:glusterd_op_status_volume] >>> 0-management: Returning 0 >>> >>>> [2015-12-29 12:57:23.539237] D [MSGID: 0] >>> >>>> [glusterd-op-sm.c:5642:glusterd_op_commit_perform] >>> 0-management: Returning 0 >>> >>>> [2015-12-29 12:57:23.541960] D [MSGID: 0] >>> >>>> [glusterd-peer-utils.c:200:glusterd_peerinfo_find_by_uuid] >>> 0-management: >>> >>>> Friend found... state: Peer in Cluster >>> >>>> [2015-12-29 12:57:23.542525] D [logging.c:1952:_gf_msg_internal] >>> >>>> 0-logging-infra: Buffer overflow of a buffer whose size limit >>> is 5. >>> >>>> About to flush least recently used log message to disk >>> >>>> The message "D [MSGID: 0] >>> >>>> [glusterd-peer-utils.c:200:glusterd_peerinfo_find_by_uuid] >>> 0-management: >>> >>>> Friend found... state: Peer in Cluster" repeated 2 times between >>> >>>> [2015-12-29 12:57:23.541960] and [2015-12-29 12:57:23.542186] >>> >>>> [2015-12-29 12:57:23.542523] D [MSGID: 0] >>> >>>> [glusterd-syncop.c:1449:gd_commit_op_phase] 0-management: >>> Sent commit op >>> >>>> req for 'Volume Status' to 3 peers >>> >>>> [2015-12-29 12:57:23.542643] D [MSGID: 0] >>> >>>> [glusterd-op-sm.c:4487:glusterd_op_modify_op_ctx] >>> 0-management: op_ctx >>> >>>> modification not required for status operation being performed >>> >>>> [2015-12-29 12:57:23.542680] D [MSGID: 0] >>> >>>> [glusterd-op-sm.c:255:glusterd_get_txn_opinfo] 0-management: >>> >>>> Successfully got opinfo for transaction ID : >>> >>>> c2886398-03c1-4d32-924a-9f92367be85c >>> >>>> [2015-12-29 12:57:23.542704] D [MSGID: 0] >>> >>>> [glusterd-op-sm.c:259:glusterd_get_txn_opinfo] 0-management: >>> Returning 0 >>> >>>> [2015-12-29 12:57:23.542732] D [MSGID: 0] >>> >>>> [glusterd-op-sm.c:360:glusterd_clear_txn_opinfo] 0-management: >>> >>>> Successfully cleared opinfo for transaction ID : >>> >>>> c2886398-03c1-4d32-924a-9f92367be85c >>> >>>> [2015-12-29 12:57:23.542756] D [MSGID: 0] >>> >>>> [glusterd-op-sm.c:364:glusterd_clear_txn_opinfo] >>> 0-management: Returning 0 >>> >>>> >>> >>>> ==> /var/log/glusterfs/cmd_history.log <== >>> >>>> [2015-12-29 12:57:23.542781] : volume status : SUCCESS >>> >>>> >>> >>>> ==> /var/log/glusterfs/etc-glusterfs-glusterd.vol.log <== >>> >>>> [2015-12-29 12:57:23.542847] D [MSGID: 0] >>> >>>> [glusterd-rpc-ops.c:205:glusterd_op_send_cli_response] >>> 0-management: >>> >>>> Returning 0 >>> >>>> [2015-12-29 12:57:23.545697] I [MSGID: 106499] >>> >>>> [glusterd-handler.c:4267:__glusterd_handle_status_volume] >>> 0-management: >>> >>>> Received status volume req for volume live >>> >>>> [2015-12-29 12:57:23.545818] D [MSGID: 0] >>> >>>> [glusterd-op-sm.c:214:glusterd_generate_txn_id] 0-management: >>> >>>> Transaction_id = fd3671e4-fa7a-4913-a6dd-b37884a3a715 >>> >>>> [2015-12-29 12:57:23.545872] D [MSGID: 0] >>> >>>> [glusterd-op-sm.c:311:glusterd_set_txn_opinfo] 0-management: >>> >>>> Successfully set opinfo for transaction ID : >>> >>>> fd3671e4-fa7a-4913-a6dd-b37884a3a715 >>> >>>> [2015-12-29 12:57:23.545905] D [MSGID: 0] >>> >>>> [glusterd-op-sm.c:318:glusterd_set_txn_opinfo] 0-management: >>> Returning 0 >>> >>>> [2015-12-29 12:57:23.545926] D [MSGID: 0] >>> >>>> [glusterd-syncop.c:1767:gd_sync_task_begin] 0-management: >>> Transaction ID >>> >>>> : fd3671e4-fa7a-4913-a6dd-b37884a3a715 >>> >>>> [2015-12-29 12:57:23.545978] D [MSGID: 0] >>> >>>> [glusterd-locks.c:562:glusterd_mgmt_v3_lock] 0-management: >>> Trying to >>> >>>> acquire lock of vol live for >>> 305c0f00-0f11-4da3-a470-50b6e6c14976 as >>> >>>> live_vol >>> >>>> [2015-12-29 12:57:23.546388] D [MSGID: 0] >>> >>>> [glusterd-locks.c:618:glusterd_mgmt_v3_lock] 0-management: >>> Lock for vol >>> >>>> live successfully held by 305c0f00-0f11-4da3-a470-50b6e6c14976 >>> >>>> [2015-12-29 12:57:23.546478] D [MSGID: 0] >>> >>>> [glusterd-syncop.c:411:gd_syncop_mgmt_v3_lock] 0-glusterd: >>> Returning 0 >>> >>>> [2015-12-29 12:57:23.549943] D [logging.c:1952:_gf_msg_internal] >>> >>>> 0-logging-infra: Buffer overflow of a buffer whose size limit >>> is 5. >>> >>>> About to flush least recently used log message to disk >>> >>>> The message "D [MSGID: 0] >>> [glusterd-syncop.c:411:gd_syncop_mgmt_v3_lock] >>> >>>> 0-glusterd: Returning 0" repeated 2 times between [2015-12-29 >>> >>>> 12:57:23.546478] and [2015-12-29 12:57:23.546565] >>> >>>> [2015-12-29 12:57:23.549941] D [MSGID: 0] >>> >>>> [glusterd-syncop.c:1205:gd_lock_op_phase] 0-management: Sent >>> lock op req >>> >>>> for 'Volume Status' to 3 peers. Returning 0 >>> >>>> [2015-12-29 12:57:23.550063] D [MSGID: 0] >>> >>>> [glusterd-utils.c:1424:glusterd_volinfo_find] 0-management: >>> Volume live >>> >>>> found >>> >>>> [2015-12-29 12:57:23.550087] D [MSGID: 0] >>> >>>> [glusterd-utils.c:1431:glusterd_volinfo_find] 0-management: >>> Returning 0 >>> >>>> [2015-12-29 12:57:23.550121] D [MSGID: 0] >>> >>>> [glusterd-utils.c:1424:glusterd_volinfo_find] 0-management: >>> Volume live >>> >>>> found >>> >>>> [2015-12-29 12:57:23.550140] D [MSGID: 0] >>> >>>> [glusterd-utils.c:1431:glusterd_volinfo_find] 0-management: >>> Returning 0 >>> >>>> [2015-12-29 12:57:23.550167] D [MSGID: 0] >>> >>>> [glusterd-op-sm.c:1695:glusterd_op_stage_status_volume] >>> 0-management: >>> >>>> Returning: 0 >>> >>>> [2015-12-29 12:57:23.550192] D [MSGID: 0] >>> >>>> [glusterd-op-sm.c:5515:glusterd_op_stage_validate] >>> 0-management: OP = >>> >>>> 18. Returning 0 >>> >>>> [2015-12-29 12:57:23.550309] D [MSGID: 0] >>> >>>> [glusterd-syncop.c:1312:gd_stage_op_phase] 0-management: Sent >>> stage op >>> >>>> req for 'Volume Status' to 3 peers >>> >>>> [2015-12-29 12:57:23.552791] D [MSGID: 0] >>> >>>> [glusterd-peer-utils.c:200:glusterd_peerinfo_find_by_uuid] >>> 0-management: >>> >>>> Friend found... state: Peer in Cluster >>> >>>> [2015-12-29 12:57:23.553395] D [logging.c:1952:_gf_msg_internal] >>> >>>> 0-logging-infra: Buffer overflow of a buffer whose size limit >>> is 5. >>> >>>> About to flush least recently used log message to disk >>> >>>> The message "D [MSGID: 0] >>> >>>> [glusterd-peer-utils.c:200:glusterd_peerinfo_find_by_uuid] >>> 0-management: >>> >>>> Friend found... state: Peer in Cluster" repeated 2 times between >>> >>>> [2015-12-29 12:57:23.552791] and [2015-12-29 12:57:23.553087] >>> >>>> [2015-12-29 12:57:23.553394] D [MSGID: 0] >>> >>>> [glusterd-op-sm.c:7010:glusterd_op_bricks_select] >>> 0-management: Returning 0 >>> >>>> [2015-12-29 12:57:23.553499] D [MSGID: 0] >>> >>>> [glusterd-syncop.c:1709:gd_brick_op_phase] 0-management: Sent >>> op req to >>> >>>> 0 bricks >>> >>>> [2015-12-29 12:57:23.553535] D [MSGID: 0] >>> >>>> [glusterd-utils.c:1424:glusterd_volinfo_find] 0-management: >>> Volume live >>> >>>> found >>> >>>> [2015-12-29 12:57:23.553556] D [MSGID: 0] >>> >>>> [glusterd-utils.c:1431:glusterd_volinfo_find] 0-management: >>> Returning 0 >>> >>>> [2015-12-29 12:57:23.553786] D [MSGID: 0] >>> >>>> [glusterd-snapshot-utils.c:3595:glusterd_is_snapd_enabled] >>> 0-management: >>> >>>> Key features.uss not present in the dict for volume live >>> >>>> [2015-12-29 12:57:23.554065] E [MSGID: 106396] >>> >>>> [glusterd-op-sm.c:2985:_add_task_to_dict] 0-management: Statedump >>> >>>> operation doesn't have a task_id >>> >>>> [2015-12-29 12:57:23.554097] E [MSGID: 106060] >>> >>>> [glusterd-op-sm.c:3055:glusterd_aggregate_task_status] >>> 0-management: >>> >>>> Failed to add task details to dict >>> >>>> [2015-12-29 12:57:23.554117] D [MSGID: 0] >>> >>>> [glusterd-op-sm.c:3360:glusterd_op_status_volume] >>> 0-management: Returning -1 >>> >>>> [2015-12-29 12:57:23.554135] D [MSGID: 0] >>> >>>> [glusterd-op-sm.c:5642:glusterd_op_commit_perform] 0-management: >>> >>>> Returning -1 >>> >>>> [2015-12-29 12:57:23.554155] E [MSGID: 106123] >>> >>>> [glusterd-syncop.c:1404:gd_commit_op_phase] 0-management: >>> Commit of >>> >>>> operation 'Volume Status' failed on localhost >>> >>>> [2015-12-29 12:57:23.554240] D [MSGID: 0] >>> >>>> [glusterd-syncop.c:510:gd_syncop_mgmt_v3_unlock] 0-glusterd: >>> Returning 0 >>> >>>> [2015-12-29 12:57:23.557147] D [logging.c:1952:_gf_msg_internal] >>> >>>> 0-logging-infra: Buffer overflow of a buffer whose size limit >>> is 5. >>> >>>> About to flush least recently used log message to disk >>> >>>> The message "D [MSGID: 0] >>> >>>> [glusterd-syncop.c:510:gd_syncop_mgmt_v3_unlock] 0-glusterd: >>> Returning >>> >>>> 0" repeated 2 times between [2015-12-29 12:57:23.554240] and >>> [2015-12-29 >>> >>>> 12:57:23.554319] >>> >>>> [2015-12-29 12:57:23.557146] D [MSGID: 0] >>> >>>> [glusterd-syncop.c:1558:gd_unlock_op_phase] 0-management: >>> Sent unlock op >>> >>>> req for 'Volume Status' to 3 peers. Returning 0 >>> >>>> [2015-12-29 12:57:23.557255] D [MSGID: 0] >>> >>>> [glusterd-locks.c:669:glusterd_mgmt_v3_unlock] 0-management: >>> Trying to >>> >>>> release lock of vol live for >>> 305c0f00-0f11-4da3-a470-50b6e6c14976 as >>> >>>> live_vol >>> >>>> [2015-12-29 12:57:23.557286] D [MSGID: 0] >>> >>>> [glusterd-locks.c:714:glusterd_mgmt_v3_unlock] 0-management: >>> Lock for >>> >>>> vol live successfully released >>> >>>> [2015-12-29 12:57:23.557312] D [MSGID: 0] >>> >>>> [glusterd-op-sm.c:255:glusterd_get_txn_opinfo] 0-management: >>> >>>> Successfully got opinfo for transaction ID : >>> >>>> fd3671e4-fa7a-4913-a6dd-b37884a3a715 >>> >>>> [2015-12-29 12:57:23.557332] D [MSGID: 0] >>> >>>> [glusterd-op-sm.c:259:glusterd_get_txn_opinfo] 0-management: >>> Returning 0 >>> >>>> [2015-12-29 12:57:23.557356] D [MSGID: 0] >>> >>>> [glusterd-op-sm.c:360:glusterd_clear_txn_opinfo] 0-management: >>> >>>> Successfully cleared opinfo for transaction ID : >>> >>>> fd3671e4-fa7a-4913-a6dd-b37884a3a715 >>> >>>> [2015-12-29 12:57:23.557420] D [MSGID: 0] >>> >>>> [glusterd-op-sm.c:364:glusterd_clear_txn_opinfo] >>> 0-management: Returning 0 >>> >>>> >>> >>>> ==> /var/log/glusterfs/cmd_history.log <== >>> >>>> [2015-12-29 12:57:23.557447] : volume status : FAILED : >>> Commit failed >>> >>>> on localhost. Please check the log file for more details. >>> >>>> >>> >>>> Dr Christophe Trefois, Dipl.-Ing. >>> >>>> Technical Specialist / Post-Doc >>> >>>> >>> >>>> UNIVERSITÉ DU LUXEMBOURG >>> >>>> >>> >>>> LUXEMBOURG CENTRE FOR SYSTEMS BIOMEDICINE >>> >>>> Campus Belval | House of Biomedicine >>> >>>> 6, avenue du Swing >>> >>>> L-4367 Belvaux >>> >>>> T: +352 46 66 44 6124 >>> >>>> F: +352 46 66 44 6949 >>> >>>> http://www.uni.lu/lcsb >>> >>>> >>> >>>> Facebook <https://www.facebook.com/trefex> Twitter >>> >>>> <https://twitter.com/Trefex> Google Plus >>> >>>> <https://plus.google.com/+ChristopheTrefois/> Linkedin >>> >>>> <https://www.linkedin.com/in/trefoischristophe> skype >>> >>>> <http://skype:Trefex?call <http://skype:trefex?call>> >>> >>>> >>> >>>> ---- >>> >>>> This message is confidential and may contain privileged >>> information. >>> >>>> It is intended for the named recipient only. >>> >>>> If you receive it in error please notify me and permanently >>> delete the >>> >>>> original message and any copies. >>> >>>> ---- >>> >>>> >>> >>>> >>> >>>> >>> >>>> >>> >>>> >>> >>>> _______________________________________________ >>> >>>> Gluster-users mailing list >>> >>>> Gluster-users@xxxxxxxxxxx <mailto:Gluster-users@xxxxxxxxxxx> >>> >>>> http://www.gluster.org/mailman/listinfo/gluster-users >>> >>>> >>> >> >>> >> >>> >>> _______________________________________________ >>> Gluster-users mailing list >>> Gluster-users@xxxxxxxxxxx <mailto:Gluster-users@xxxxxxxxxxx> >>> http://www.gluster.org/mailman/listinfo/gluster-users >>> >> > _______________________________________________ > Gluster-users mailing list > Gluster-users@xxxxxxxxxxx > http://www.gluster.org/mailman/listinfo/gluster-users > _______________________________________________ Gluster-users mailing list Gluster-users@xxxxxxxxxxx http://www.gluster.org/mailman/listinfo/gluster-users