[Gluster-users] Latest glusterfs 3.8.5 server not compatible with livbirt libgfapi access
Radu Radutiu
rradutiu at gmail.com
Thu Nov 3 11:04:26 UTC 2016
Hi,
After updating glusterfs server to 3.8.5 (from Centos-gluster-3.8.repo) the
KVM virtual machines (qemu-kvm-ev-2.3.0-31) that access storage using
libgfapi are no longer able to start. The libvirt log file shows:
[2016-11-02 14:26:41.864024] I [MSGID: 104045] [glfs-master.c:91:notify]
0-gfapi: New graph 73332d32-3937-3130-2d32-3031362d3131 (0) coming up
[2016-11-02 14:26:41.864075] I [MSGID: 114020] [client.c:2356:notify]
0-testvol-client-0: parent translators are ready, attempting connect on
transport
[2016-11-02 14:26:41.882975] I [rpc-clnt.c:1947:rpc_clnt_reconfig]
0-testvol-client-0: changing port to 49152 (from 0)
[2016-11-02 14:26:41.889362] I [MSGID: 114057] [client-handshake.c:1446:
select_server_supported_programs] 0-testvol-client-0: Using Program
GlusterFS 3.3, Num (1298437), Version (330)
[2016-11-02 14:26:41.890001] I [MSGID: 114046]
[client-handshake.c:1222:client_setvolume_cbk]
0-testvol-client-0: Connected to testvol-client-0, attached to remote
volume '/data/brick1/testvol'.
[2016-11-02 14:26:41.890035] I [MSGID: 114047]
[client-handshake.c:1233:client_setvolume_cbk]
0-testvol-client-0: Server and Client lk-version numbers are not same,
reopening the fds
[2016-11-02 14:26:41.917990] I [MSGID: 114035]
[client-handshake.c:201:client_set_lk_version_cbk]
0-testvol-client-0: Server lk version = 1
[2016-11-02 14:26:41.919289] I [MSGID: 104041]
[glfs-resolve.c:885:__glfs_active_subvol]
0-testvol: switched to graph 73332d32-3937-3130-2d32-3031362d3131 (0)
[2016-11-02 14:26:41.922174] I [MSGID: 114021] [client.c:2365:notify]
0-testvol-client-0: current graph is no longer active, destroying
rpc_client
[2016-11-02 14:26:41.922269] I [MSGID: 114018]
[client.c:2280:client_rpc_notify]
0-testvol-client-0: disconnected from testvol-client-0. Client process will
keep trying to connect to glusterd until brick's port is available
[2016-11-02 14:26:41.922592] I [MSGID: 101053]
[mem-pool.c:617:mem_pool_destroy]
0-gfapi: size=84 max=1 total=1
[2016-11-02 14:26:41.923044] I [MSGID: 101053]
[mem-pool.c:617:mem_pool_destroy]
0-gfapi: size=188 max=2 total=2
[2016-11-02 14:26:41.923419] I [MSGID: 101053]
[mem-pool.c:617:mem_pool_destroy]
0-gfapi: size=140 max=2 total=2
[2016-11-02 14:26:41.923442] I [MSGID: 101053]
[mem-pool.c:617:mem_pool_destroy]
0-testvol-client-0: size=1324 max=2 total=5
[2016-11-02 14:26:41.923458] I [MSGID: 101053]
[mem-pool.c:617:mem_pool_destroy]
0-testvol-dht: size=1148 max=0 total=0
[2016-11-02 14:26:41.923546] I [MSGID: 101053]
[mem-pool.c:617:mem_pool_destroy]
0-testvol-dht: size=3380 max=2 total=5
[2016-11-02 14:26:41.923815] I [MSGID: 101053]
[mem-pool.c:617:mem_pool_destroy]
0-testvol-read-ahead: size=188 max=0 total=0
[2016-11-02 14:26:41.923832] I [MSGID: 101053]
[mem-pool.c:617:mem_pool_destroy]
0-testvol-readdir-ahead: size=60 max=0 total=0
[2016-11-02 14:26:41.923844] I [MSGID: 101053]
[mem-pool.c:617:mem_pool_destroy]
0-testvol-io-cache: size=68 max=0 total=0
[2016-11-02 14:26:41.923856] I [MSGID: 101053]
[mem-pool.c:617:mem_pool_destroy]
0-testvol-io-cache: size=252 max=1 total=3
[2016-11-02 14:26:41.923877] I [io-stats.c:3747:fini] 0-testvol: io-stats
translator unloaded
[2016-11-02 14:26:41.924191] I [MSGID: 101191]
[event-epoll.c:659:event_dispatch_epoll_worker]
0-epoll: Exited thread with index 2
[2016-11-02 14:26:41.924232] I [MSGID: 101191]
[event-epoll.c:659:event_dispatch_epoll_worker]
0-epoll: Exited thread with index 1
2016-11-02T14:26:42.825041Z qemu-kvm: -drive file=gluster://s3/testvol/c7.
img,if=none,id=drive-virtio-disk0,format=qcow2,cache=none: Could not read
L1 table: Bad file descriptor
The brick is available , runs on the same host and mounted in another
directory using fuse (to confirm that it is indeed fine).
If I downgrade the gluster server to 3.8.4 everything works fine. Anyone
has seen this or has any idea how to debug?
Regards,
Radu
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20161103/566d7b97/attachment.html>
More information about the Gluster-users
mailing list