[Gluster-users] gluster volume status -> commit failed

Atin Mukherjee amukherj at redhat.com
Tue Dec 29 13:27:37 UTC 2015


It seems like your opCode is STATEDUMP instead of STATUS which is weird.
Are you running a heterogeneous cluster? What is the last version you
were running with? What's the current cluster op-version?

Thanks,
Atin

On 12/29/2015 06:31 PM, Christophe TREFOIS wrote:
> Dear all,
> 
> I have a 3-node distribute setup with a controller of GlusterFS and
> upgraded to 3.7.6 today and to CentOS 7.2.
> 
> After the ugprade (reboot), I can start the volume fine and see mounted
> volume as well on the controller.
> 
> However, a gluster volume info <volname> results in an 
> 
> [root at stor104 glusterfs]# gluster volume status live
> Commit failed on localhost. Please check the log file for more details.
> 
> error.
> 
> Below some information and log extracts.
> 
> Thank you for any hints on where to start fixing this,
> 
> Kind regards,
> 
>> Christophe
> 
> —————— 
> 
> Here is the gluster info command for the volume.
> 
> [root at stor104 glusterfs]# gluster volume info live
> 
> Volume Name: live
> Type: Distribute
> Volume ID: 1328637d-7730-4627-8945-bbe43626d527
> Status: Started
> Number of Bricks: 9
> Transport-type: tcp
> Bricks:
> Brick1: stor104:/zfs/brick0/brick
> Brick2: stor104:/zfs/brick1/brick
> Brick3: stor104:/zfs/brick2/brick
> Brick4: stor106:/zfs/brick0/brick
> Brick5: stor106:/zfs/brick1/brick
> Brick6: stor106:/zfs/brick2/brick
> Brick7: stor105:/zfs/brick0/brick
> Brick8: stor105:/zfs/brick1/brick
> Brick9: stor105:/zfs/brick2/brick
> Options Reconfigured:
> performance.io <http://performance.io>-thread-count: 8
> nfs.disable: on
> performance.write-behind-window-size: 4MB
> performance.client-io-threads: on
> performance.cache-size: 1GB
> performance.cache-refresh-timeout: 60
> performance.cache-max-file-size: 4MB
> cluster.data-self-heal-algorithm: full
> diagnostics.client-log-level: ERROR
> diagnostics.brick-log-level: ERROR
> cluster.min-free-disk: 1%
> server.allow-insecure: on
> 
> Relevant log parts when carrying out the command:
> 
> ==> /var/log/glusterfs/cli.log <==
> [2015-12-29 12:51:03.216998] I [cli.c:721:main] 0-cli: Started running
> gluster with version 3.7.6
> [2015-12-29 12:51:03.226123] I
> [cli-cmd-volume.c:1926:cli_check_gsync_present] 0-: geo-replication not
> installed
> [2015-12-29 12:51:03.226623] I [MSGID: 101190]
> [event-epoll.c:632:event_dispatch_epoll_worker] 0-epoll: Started thread
> with index 1
> [2015-12-29 12:51:03.226723] I [socket.c:2355:socket_event_handler]
> 0-transport: disconnecting now
> 
> ==> /var/log/glusterfs/cmd_history.log <==
> [2015-12-29 12:51:03.236182]  : volume status : SUCCESS
> 
> ==> /var/log/glusterfs/etc-glusterfs-glusterd.vol.log <==
> [2015-12-29 12:51:03.238562] I [MSGID: 106499]
> [glusterd-handler.c:4267:__glusterd_handle_status_volume] 0-management:
> Received status volume req for volume live
> [2015-12-29 12:51:03.246043] E [MSGID: 106396]
> [glusterd-op-sm.c:2985:_add_task_to_dict] 0-management: Statedump
> operation doesn't have a task_id
> [2015-12-29 12:51:03.246081] E [MSGID: 106060]
> [glusterd-op-sm.c:3055:glusterd_aggregate_task_status] 0-management:
> Failed to add task details to dict
> [2015-12-29 12:51:03.246098] E [MSGID: 106123]
> [glusterd-syncop.c:1404:gd_commit_op_phase] 0-management: Commit of
> operation 'Volume Status' failed on localhost
> 
> ==> /var/log/glusterfs/cmd_history.log <==
> [2015-12-29 12:51:03.249364]  : volume status : FAILED : Commit failed
> on localhost. Please check the log file for more details.
> 
> ==> /var/log/glusterfs/cli.log <==
> [2015-12-29 12:51:03.249647] I [input.c:36:cli_batch] 0-: Exiting with: 0
> 
> All bricks are up
> 
> [root at highlander glusterfs]# pdsh -g live 'df -h | grep brick*'
> stor106: brick0           50T   33T   18T  66% /zfs/brick0
> stor106: brick1           50T   33T   18T  66% /zfs/brick1
> stor106: brick2           50T   33T   18T  66% /zfs/brick2
> stor105: brick0           40T   23T   18T  57% /zfs/brick0
> stor105: brick1           40T   23T   18T  57% /zfs/brick1
> stor105: brick2           40T   23T   18T  57% /zfs/brick2
> stor104: brick0           40T   23T   18T  57% /zfs/brick0
> stor104: brick1           40T   23T   18T  57% /zfs/brick1
> stor104: brick2           40T   23T   18T  57% /zfs/brick2
> 
> Package details
> 
> [root at highlander glusterfs]# rpm -qa | grep gluster
> glusterfs-client-xlators-3.7.6-1.el7.x86_64
> glusterfs-server-3.7.6-1.el7.x86_64
> samba-vfs-glusterfs-4.2.3-10.el7.x86_64
> glusterfs-3.7.6-1.el7.x86_64
> glusterfs-cli-3.7.6-1.el7.x86_64
> glusterfs-libs-3.7.6-1.el7.x86_64
> glusterfs-fuse-3.7.6-1.el7.x86_64
> glusterfs-api-3.7.6-1.el7.x86_64
> 
> [root at highlander glusterfs]# pdsh -g live 'rpm -qa | grep gluster'
> stor105: glusterfs-libs-3.7.6-1.el7.x86_64
> stor105: glusterfs-api-3.7.6-1.el7.x86_64
> stor105: glusterfs-3.7.6-1.el7.x86_64
> stor105: glusterfs-fuse-3.7.6-1.el7.x86_64
> stor105: glusterfs-cli-3.7.6-1.el7.x86_64
> stor105: glusterfs-client-xlators-3.7.6-1.el7.x86_64
> stor105: glusterfs-server-3.7.6-1.el7.x86_64
> 
> stor104: glusterfs-server-3.7.6-1.el7.x86_64
> stor104: glusterfs-libs-3.7.6-1.el7.x86_64
> stor104: glusterfs-api-3.7.6-1.el7.x86_64
> stor104: glusterfs-3.7.6-1.el7.x86_64
> stor104: glusterfs-fuse-3.7.6-1.el7.x86_64
> stor104: glusterfs-cli-3.7.6-1.el7.x86_64
> stor104: glusterfs-client-xlators-3.7.6-1.el7.x86_64
> 
> stor106: glusterfs-3.7.6-1.el7.x86_64
> stor106: glusterfs-cli-3.7.6-1.el7.x86_64
> stor106: glusterfs-server-3.7.6-1.el7.x86_64
> stor106: glusterfs-libs-3.7.6-1.el7.x86_64
> stor106: glusterfs-client-xlators-3.7.6-1.el7.x86_64
> stor106: glusterfs-api-3.7.6-1.el7.x86_64
> stor106: glusterfs-fuse-3.7.6-1.el7.x86_64
> 
> More detailled logs:
> 
> ==> /var/log/glusterfs/cli.log <==
> [2015-12-29 12:57:23.520821] I [cli.c:721:main] 0-cli: Started running
> gluster with version 3.7.6
> [2015-12-29 12:57:23.530898] I
> [cli-cmd-volume.c:1926:cli_check_gsync_present] 0-: geo-replication not
> installed
> [2015-12-29 12:57:23.531844] I [MSGID: 101190]
> [event-epoll.c:632:event_dispatch_epoll_worker] 0-epoll: Started thread
> with index 1
> [2015-12-29 12:57:23.532004] I [socket.c:2355:socket_event_handler]
> 0-transport: disconnecting now
> 
> ==> /var/log/glusterfs/etc-glusterfs-glusterd.vol.log <==
> [2015-12-29 12:57:23.534830] D [MSGID: 0]
> [glusterd-op-sm.c:214:glusterd_generate_txn_id] 0-management:
> Transaction_id = c2886398-03c1-4d32-924a-9f92367be85c
> [2015-12-29 12:57:23.534946] D [MSGID: 0]
> [glusterd-op-sm.c:311:glusterd_set_txn_opinfo] 0-management:
> Successfully set opinfo for transaction ID :
> c2886398-03c1-4d32-924a-9f92367be85c
> [2015-12-29 12:57:23.534975] D [MSGID: 0]
> [glusterd-op-sm.c:318:glusterd_set_txn_opinfo] 0-management: Returning 0
> [2015-12-29 12:57:23.535001] D [MSGID: 0]
> [glusterd-syncop.c:1767:gd_sync_task_begin] 0-management: Transaction ID
> : c2886398-03c1-4d32-924a-9f92367be85c
> [2015-12-29 12:57:23.535031] D [MSGID: 0]
> [glusterd-syncop.c:1807:gd_sync_task_begin] 0-glusterd: Failed to get
> volume name
> [2015-12-29 12:57:23.535078] D [MSGID: 0]
> [glusterd-op-sm.c:1695:glusterd_op_stage_status_volume] 0-management:
> Returning: 0
> [2015-12-29 12:57:23.535103] D [MSGID: 0]
> [glusterd-op-sm.c:5515:glusterd_op_stage_validate] 0-management: OP =
> 18. Returning 0
> [2015-12-29 12:57:23.535555] D
> [rpc-clnt-ping.c:98:rpc_clnt_remove_ping_timer_locked] (-->
> /lib64/libglusterfs.so.0(_gf_log_callingfn+0x192)[0x7f410a9d5a82] (-->
> /lib64/libgfrpc.so.0(rpc_clnt_remove_ping_timer_locked+0x8b)[0x7f410a7a587b]
> (--> /lib64/libgfrpc.so.0(+0x13e74)[0x7f410a7a5e74] (-->
> /lib64/libgfrpc.so.0(rpc_clnt_submit+0x34f)[0x7f410a7a1c4f] (-->
> /usr/lib64/glusterfs/3.7.6/xlator/mgmt/glusterd.so(gd_syncop_submit_request+0x1a5)[0x7f40ff5d24b5]
> ))))) 0-: 192.168.123.105:24007: ping timer event already removed
> [2015-12-29 12:57:23.535935] D
> [rpc-clnt-ping.c:98:rpc_clnt_remove_ping_timer_locked] (-->
> /lib64/libglusterfs.so.0(_gf_log_callingfn+0x192)[0x7f410a9d5a82] (-->
> /lib64/libgfrpc.so.0(rpc_clnt_remove_ping_timer_locked+0x8b)[0x7f410a7a587b]
> (--> /lib64/libgfrpc.so.0(+0x13e74)[0x7f410a7a5e74] (-->
> /lib64/libgfrpc.so.0(rpc_clnt_submit+0x34f)[0x7f410a7a1c4f] (-->
> /usr/lib64/glusterfs/3.7.6/xlator/mgmt/glusterd.so(gd_syncop_submit_request+0x1a5)[0x7f40ff5d24b5]
> ))))) 0-: 192.168.123.1:24007: ping timer event already removed
> [2015-12-29 12:57:23.536279] D
> [rpc-clnt-ping.c:98:rpc_clnt_remove_ping_timer_locked] (-->
> /lib64/libglusterfs.so.0(_gf_log_callingfn+0x192)[0x7f410a9d5a82] (-->
> /lib64/libgfrpc.so.0(rpc_clnt_remove_ping_timer_locked+0x8b)[0x7f410a7a587b]
> (--> /lib64/libgfrpc.so.0(+0x13e74)[0x7f410a7a5e74] (-->
> /lib64/libgfrpc.so.0(rpc_clnt_submit+0x34f)[0x7f410a7a1c4f] (-->
> /usr/lib64/glusterfs/3.7.6/xlator/mgmt/glusterd.so(gd_syncop_submit_request+0x1a5)[0x7f40ff5d24b5]
> ))))) 0-: 192.168.123.106:24007: ping timer event already removed
> [2015-12-29 12:57:23.536358] D [MSGID: 0]
> [glusterd-syncop.c:1312:gd_stage_op_phase] 0-management: Sent stage op
> req for 'Volume Status' to 3 peers
> [2015-12-29 12:57:23.538064] D [MSGID: 0]
> [glusterd-peer-utils.c:200:glusterd_peerinfo_find_by_uuid] 0-management:
> Friend found... state: Peer in Cluster
> [2015-12-29 12:57:23.539059] D [logging.c:1952:_gf_msg_internal]
> 0-logging-infra: Buffer overflow of a buffer whose size limit is 5.
> About to flush least recently used log message to disk
> The message "D [MSGID: 0]
> [glusterd-peer-utils.c:200:glusterd_peerinfo_find_by_uuid] 0-management:
> Friend found... state: Peer in Cluster" repeated 2 times between
> [2015-12-29 12:57:23.538064] and [2015-12-29 12:57:23.538833]
> [2015-12-29 12:57:23.539057] D [MSGID: 0]
> [glusterd-op-sm.c:7010:glusterd_op_bricks_select] 0-management: Returning 0
> [2015-12-29 12:57:23.539163] D [MSGID: 0]
> [glusterd-syncop.c:1709:gd_brick_op_phase] 0-management: Sent op req to
> 0 bricks
> [2015-12-29 12:57:23.539213] D [MSGID: 0]
> [glusterd-op-sm.c:3360:glusterd_op_status_volume] 0-management: Returning 0
> [2015-12-29 12:57:23.539237] D [MSGID: 0]
> [glusterd-op-sm.c:5642:glusterd_op_commit_perform] 0-management: Returning 0
> [2015-12-29 12:57:23.541960] D [MSGID: 0]
> [glusterd-peer-utils.c:200:glusterd_peerinfo_find_by_uuid] 0-management:
> Friend found... state: Peer in Cluster
> [2015-12-29 12:57:23.542525] D [logging.c:1952:_gf_msg_internal]
> 0-logging-infra: Buffer overflow of a buffer whose size limit is 5.
> About to flush least recently used log message to disk
> The message "D [MSGID: 0]
> [glusterd-peer-utils.c:200:glusterd_peerinfo_find_by_uuid] 0-management:
> Friend found... state: Peer in Cluster" repeated 2 times between
> [2015-12-29 12:57:23.541960] and [2015-12-29 12:57:23.542186]
> [2015-12-29 12:57:23.542523] D [MSGID: 0]
> [glusterd-syncop.c:1449:gd_commit_op_phase] 0-management: Sent commit op
> req for 'Volume Status' to 3 peers
> [2015-12-29 12:57:23.542643] D [MSGID: 0]
> [glusterd-op-sm.c:4487:glusterd_op_modify_op_ctx] 0-management: op_ctx
> modification not required for status operation being performed
> [2015-12-29 12:57:23.542680] D [MSGID: 0]
> [glusterd-op-sm.c:255:glusterd_get_txn_opinfo] 0-management:
> Successfully got opinfo for transaction ID :
> c2886398-03c1-4d32-924a-9f92367be85c
> [2015-12-29 12:57:23.542704] D [MSGID: 0]
> [glusterd-op-sm.c:259:glusterd_get_txn_opinfo] 0-management: Returning 0
> [2015-12-29 12:57:23.542732] D [MSGID: 0]
> [glusterd-op-sm.c:360:glusterd_clear_txn_opinfo] 0-management:
> Successfully cleared opinfo for transaction ID :
> c2886398-03c1-4d32-924a-9f92367be85c
> [2015-12-29 12:57:23.542756] D [MSGID: 0]
> [glusterd-op-sm.c:364:glusterd_clear_txn_opinfo] 0-management: Returning 0
> 
> ==> /var/log/glusterfs/cmd_history.log <==
> [2015-12-29 12:57:23.542781]  : volume status : SUCCESS
> 
> ==> /var/log/glusterfs/etc-glusterfs-glusterd.vol.log <==
> [2015-12-29 12:57:23.542847] D [MSGID: 0]
> [glusterd-rpc-ops.c:205:glusterd_op_send_cli_response] 0-management:
> Returning 0
> [2015-12-29 12:57:23.545697] I [MSGID: 106499]
> [glusterd-handler.c:4267:__glusterd_handle_status_volume] 0-management:
> Received status volume req for volume live
> [2015-12-29 12:57:23.545818] D [MSGID: 0]
> [glusterd-op-sm.c:214:glusterd_generate_txn_id] 0-management:
> Transaction_id = fd3671e4-fa7a-4913-a6dd-b37884a3a715
> [2015-12-29 12:57:23.545872] D [MSGID: 0]
> [glusterd-op-sm.c:311:glusterd_set_txn_opinfo] 0-management:
> Successfully set opinfo for transaction ID :
> fd3671e4-fa7a-4913-a6dd-b37884a3a715
> [2015-12-29 12:57:23.545905] D [MSGID: 0]
> [glusterd-op-sm.c:318:glusterd_set_txn_opinfo] 0-management: Returning 0
> [2015-12-29 12:57:23.545926] D [MSGID: 0]
> [glusterd-syncop.c:1767:gd_sync_task_begin] 0-management: Transaction ID
> : fd3671e4-fa7a-4913-a6dd-b37884a3a715
> [2015-12-29 12:57:23.545978] D [MSGID: 0]
> [glusterd-locks.c:562:glusterd_mgmt_v3_lock] 0-management: Trying to
> acquire lock of vol live for 305c0f00-0f11-4da3-a470-50b6e6c14976 as
> live_vol
> [2015-12-29 12:57:23.546388] D [MSGID: 0]
> [glusterd-locks.c:618:glusterd_mgmt_v3_lock] 0-management: Lock for vol
> live successfully held by 305c0f00-0f11-4da3-a470-50b6e6c14976
> [2015-12-29 12:57:23.546478] D [MSGID: 0]
> [glusterd-syncop.c:411:gd_syncop_mgmt_v3_lock] 0-glusterd: Returning 0
> [2015-12-29 12:57:23.549943] D [logging.c:1952:_gf_msg_internal]
> 0-logging-infra: Buffer overflow of a buffer whose size limit is 5.
> About to flush least recently used log message to disk
> The message "D [MSGID: 0] [glusterd-syncop.c:411:gd_syncop_mgmt_v3_lock]
> 0-glusterd: Returning 0" repeated 2 times between [2015-12-29
> 12:57:23.546478] and [2015-12-29 12:57:23.546565]
> [2015-12-29 12:57:23.549941] D [MSGID: 0]
> [glusterd-syncop.c:1205:gd_lock_op_phase] 0-management: Sent lock op req
> for 'Volume Status' to 3 peers. Returning 0
> [2015-12-29 12:57:23.550063] D [MSGID: 0]
> [glusterd-utils.c:1424:glusterd_volinfo_find] 0-management: Volume live
> found
> [2015-12-29 12:57:23.550087] D [MSGID: 0]
> [glusterd-utils.c:1431:glusterd_volinfo_find] 0-management: Returning 0
> [2015-12-29 12:57:23.550121] D [MSGID: 0]
> [glusterd-utils.c:1424:glusterd_volinfo_find] 0-management: Volume live
> found
> [2015-12-29 12:57:23.550140] D [MSGID: 0]
> [glusterd-utils.c:1431:glusterd_volinfo_find] 0-management: Returning 0
> [2015-12-29 12:57:23.550167] D [MSGID: 0]
> [glusterd-op-sm.c:1695:glusterd_op_stage_status_volume] 0-management:
> Returning: 0
> [2015-12-29 12:57:23.550192] D [MSGID: 0]
> [glusterd-op-sm.c:5515:glusterd_op_stage_validate] 0-management: OP =
> 18. Returning 0
> [2015-12-29 12:57:23.550309] D [MSGID: 0]
> [glusterd-syncop.c:1312:gd_stage_op_phase] 0-management: Sent stage op
> req for 'Volume Status' to 3 peers
> [2015-12-29 12:57:23.552791] D [MSGID: 0]
> [glusterd-peer-utils.c:200:glusterd_peerinfo_find_by_uuid] 0-management:
> Friend found... state: Peer in Cluster
> [2015-12-29 12:57:23.553395] D [logging.c:1952:_gf_msg_internal]
> 0-logging-infra: Buffer overflow of a buffer whose size limit is 5.
> About to flush least recently used log message to disk
> The message "D [MSGID: 0]
> [glusterd-peer-utils.c:200:glusterd_peerinfo_find_by_uuid] 0-management:
> Friend found... state: Peer in Cluster" repeated 2 times between
> [2015-12-29 12:57:23.552791] and [2015-12-29 12:57:23.553087]
> [2015-12-29 12:57:23.553394] D [MSGID: 0]
> [glusterd-op-sm.c:7010:glusterd_op_bricks_select] 0-management: Returning 0
> [2015-12-29 12:57:23.553499] D [MSGID: 0]
> [glusterd-syncop.c:1709:gd_brick_op_phase] 0-management: Sent op req to
> 0 bricks
> [2015-12-29 12:57:23.553535] D [MSGID: 0]
> [glusterd-utils.c:1424:glusterd_volinfo_find] 0-management: Volume live
> found
> [2015-12-29 12:57:23.553556] D [MSGID: 0]
> [glusterd-utils.c:1431:glusterd_volinfo_find] 0-management: Returning 0
> [2015-12-29 12:57:23.553786] D [MSGID: 0]
> [glusterd-snapshot-utils.c:3595:glusterd_is_snapd_enabled] 0-management:
> Key features.uss not present in the dict for volume live
> [2015-12-29 12:57:23.554065] E [MSGID: 106396]
> [glusterd-op-sm.c:2985:_add_task_to_dict] 0-management: Statedump
> operation doesn't have a task_id
> [2015-12-29 12:57:23.554097] E [MSGID: 106060]
> [glusterd-op-sm.c:3055:glusterd_aggregate_task_status] 0-management:
> Failed to add task details to dict
> [2015-12-29 12:57:23.554117] D [MSGID: 0]
> [glusterd-op-sm.c:3360:glusterd_op_status_volume] 0-management: Returning -1
> [2015-12-29 12:57:23.554135] D [MSGID: 0]
> [glusterd-op-sm.c:5642:glusterd_op_commit_perform] 0-management:
> Returning -1
> [2015-12-29 12:57:23.554155] E [MSGID: 106123]
> [glusterd-syncop.c:1404:gd_commit_op_phase] 0-management: Commit of
> operation 'Volume Status' failed on localhost
> [2015-12-29 12:57:23.554240] D [MSGID: 0]
> [glusterd-syncop.c:510:gd_syncop_mgmt_v3_unlock] 0-glusterd: Returning 0
> [2015-12-29 12:57:23.557147] D [logging.c:1952:_gf_msg_internal]
> 0-logging-infra: Buffer overflow of a buffer whose size limit is 5.
> About to flush least recently used log message to disk
> The message "D [MSGID: 0]
> [glusterd-syncop.c:510:gd_syncop_mgmt_v3_unlock] 0-glusterd: Returning
> 0" repeated 2 times between [2015-12-29 12:57:23.554240] and [2015-12-29
> 12:57:23.554319]
> [2015-12-29 12:57:23.557146] D [MSGID: 0]
> [glusterd-syncop.c:1558:gd_unlock_op_phase] 0-management: Sent unlock op
> req for 'Volume Status' to 3 peers. Returning 0
> [2015-12-29 12:57:23.557255] D [MSGID: 0]
> [glusterd-locks.c:669:glusterd_mgmt_v3_unlock] 0-management: Trying to
> release lock of vol live for 305c0f00-0f11-4da3-a470-50b6e6c14976 as
> live_vol
> [2015-12-29 12:57:23.557286] D [MSGID: 0]
> [glusterd-locks.c:714:glusterd_mgmt_v3_unlock] 0-management: Lock for
> vol live successfully released
> [2015-12-29 12:57:23.557312] D [MSGID: 0]
> [glusterd-op-sm.c:255:glusterd_get_txn_opinfo] 0-management:
> Successfully got opinfo for transaction ID :
> fd3671e4-fa7a-4913-a6dd-b37884a3a715
> [2015-12-29 12:57:23.557332] D [MSGID: 0]
> [glusterd-op-sm.c:259:glusterd_get_txn_opinfo] 0-management: Returning 0
> [2015-12-29 12:57:23.557356] D [MSGID: 0]
> [glusterd-op-sm.c:360:glusterd_clear_txn_opinfo] 0-management:
> Successfully cleared opinfo for transaction ID :
> fd3671e4-fa7a-4913-a6dd-b37884a3a715
> [2015-12-29 12:57:23.557420] D [MSGID: 0]
> [glusterd-op-sm.c:364:glusterd_clear_txn_opinfo] 0-management: Returning 0
> 
> ==> /var/log/glusterfs/cmd_history.log <==
> [2015-12-29 12:57:23.557447]  : volume status : FAILED : Commit failed
> on localhost. Please check the log file for more details.
> 
> Dr Christophe Trefois, Dipl.-Ing.  
> Technical Specialist / Post-Doc
> 
> UNIVERSITÉ DU LUXEMBOURG
> 
> LUXEMBOURG CENTRE FOR SYSTEMS BIOMEDICINE
> Campus Belval | House of Biomedicine  
> 6, avenue du Swing 
> L-4367 Belvaux  
> T: +352 46 66 44 6124 
> F: +352 46 66 44 6949  
> http://www.uni.lu/lcsb
> 
> Facebook <https://www.facebook.com/trefex>  Twitter
> <https://twitter.com/Trefex>  Google Plus
> <https://plus.google.com/+ChristopheTrefois/>  Linkedin
> <https://www.linkedin.com/in/trefoischristophe>  skype
> <http://skype:Trefex?call>
> 
> ----
> This message is confidential and may contain privileged information. 
> It is intended for the named recipient only. 
> If you receive it in error please notify me and permanently delete the
> original message and any copies. 
> ----
> 
>   
> 
> 
> 
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://www.gluster.org/mailman/listinfo/gluster-users
> 


More information about the Gluster-users mailing list