[Gluster-users] distribute-replicate setup GFS Client crashed

Vikas R vicky.ice at gmail.com
Thu Jan 13 11:44:22 UTC 2011


Hi there,

Im running glusterfs version 3.1.0.

The client crashed after sometime with below stack.

2011-01-13 08:33:49.230976] I [afr-common.c:2568:afr_notify] replicate-1:
Subvolume 'distribute-1' came back up; going online.
[2011-01-13 08:33:49.499909] I [afr-open.c:393:afr_openfd_sh] replicate-1:
 data self-heal triggered. path:
/streaming/set3/work/reduce.12.1294902171.dplog.temp, reason: Replicate up
down flush, data lock is held
[2011-01-13 08:33:49.500500] E
[afr-self-heal-common.c:1214:sh_missing_entries_create] replicate-1: no
missing files - /streaming/set3/work/reduce.12.1294902171.dplog.temp.
proceeding to metadata check
[2011-01-13 08:33:49.501906] E [afr-common.c:110:afr_set_split_brain]
replicate-1: invalid argument: inode
[2011-01-13 08:33:49.501919] I
[afr-self-heal-common.c:1526:afr_self_heal_completion_cbk] replicate-1:
background  data self-heal completed on
/streaming/set3/work/reduce.12.1294902171.dplog.temp
[2011-01-13 08:33:49.531838] I [dht-common.c:402:dht_revalidate_cbk]
distribute-1: linkfile found in revalidate for
/streaming/set3/work/mapped/dpabort/multiple_reduce.flash_pl.2.1294901929.1.172.26.98.59.2.map.10
[2011-01-13 08:33:50.396055] W [fuse-bridge.c:2765:fuse_setlk_cbk]
glusterfs-fuse: 2230985: ERR => -1 (Invalid argument)
pending frames:
frame : type(1) op(FLUSH)
frame : type(1) op(FLUSH)
frame : type(1) op(LK)

patchset: v3.1.0
signal received: 11
time of crash: 2011-01-13 08:33:50
configuration details:
argp 1
backtrace 1
dlfcn 1
fdatasync 1
libpthread 1
llistxattr 1
setfsid 1
spinlock 1
epoll.h 1
xattr.h 1
st_atim.tv_nsec 1
package-string: glusterfs 3.1.0
[0xffffe400]
/usr/local/akamai/lib/glusterfs/3.1.0/xlator/cluster/afr.so(afr_internal_lock_finish+0x8b)[0xf6187eeb]
/usr/local/akamai/lib/glusterfs/3.1.0/xlator/cluster/afr.so(afr_post_nonblocking_inodelk_cbk+0x4f)[0xf61884ef]
/usr/local/akamai/lib/glusterfs/3.1.0/xlator/cluster/afr.so(afr_nonblocking_inodelk_cbk+0x28f)[0xf61a0dbf]
/usr/local/akamai/lib/glusterfs/3.1.0/xlator/cluster/dht.so(dht_finodelk_cbk+0x82)[0xf61c3102]
/usr/local/akamai/lib/glusterfs/3.1.0/xlator/protocol/client.so(client3_1_finodelk_cbk+0xbe)[0xf61f1c1e]
/usr/local/akamai/lib/libgfrpc.so.0(rpc_clnt_handle_reply+0xc2)[0xf7736c42]
/usr/local/akamai/lib/libgfrpc.so.0(rpc_clnt_notify+0xa2)[0xf7736e62]
/usr/local/akamai/lib/libgfrpc.so.0(rpc_transport_notify+0x35)[0xf77314c5]
/usr/local/akamai/lib/glusterfs/3.1.0/rpc-transport/socket.so(socket_event_poll_in+0x50)[0xf5f08500]
/usr/local/akamai/lib/glusterfs/3.1.0/rpc-transport/socket.so(socket_event_handler+0x15b)[0xf5f0867b]
/usr/local/akamai/lib/libglusterfs.so.0[0xf7771cff]
/usr/local/akamai/lib/libglusterfs.so.0(event_dispatch+0x21)[0xf7770a21]
glusterfsc(main+0x48c)[0x804c45c]
/lib/tls/i686/cmov/libc.so.6(__libc_start_main+0xdc)[0xf75d718c]
glusterfsc[0x804a631]

The config files are attached.

tx
Vikas
-------------- next part --------------
## file auto generated by /usr/local/bin/glusterfs-volgen (export.vol)
# Cmd line:
# $ /usr/local/bin/glusterfs-volgen --name gfs 172.24.0.68:/ghostcache/home/hsawhney/gfs/ 172.24.0.222:/ghostcache/home/hsawhney/gfs/

volume posix1
  type storage/posix
  option directory /ghostcache/gfs-export/
end-volume

volume locks1
    type features/locks
    subvolumes posix1
end-volume

#volume quota1
#    type features/quota
#    #option disk-usage-limit 100MB
#    subvolumes locks1
#end-volume

volume brickex
    type performance/io-threads
    option thread-count 4
    subvolumes locks1
end-volume

volume server-tcp
    type protocol/server
    option transport-type tcp
    option auth.addr.brickex.allow *
    option transport.socket.listen-port 6996
    option transport.socket.nodelay on
    subvolumes brickex
end-volume
-------------- next part --------------
# file auto generated by /usr/local/bin/glusterfs-volgen (mount.vol)
# Cmd line:
# $ /usr/local/bin/glusterfs-volgen --name gfs 172.24.0.68:/ghostcache/home/hsawhney/gfs/ 172.24.0.222:/ghostcache/home/hsawhney/gfs/

# TRANSPORT-TYPE tcp
volume 172.26.98.55-1
    type protocol/client
    option transport-type tcp
    option remote-host 172.26.98.55
    option transport.socket.nodelay on
    option transport.remote-port 6996
    option remote-subvolume brickex
end-volume

volume 172.26.98.56-1
    type protocol/client
    option transport-type tcp
    option remote-host 172.26.98.56
    option transport.socket.nodelay on
    option transport.remote-port 6996
    option remote-subvolume brickex
end-volume

volume 172.26.98.57-1
    type protocol/client
    option transport-type tcp
    option remote-host 172.26.98.57
    option transport.socket.nodelay on
    option transport.remote-port 6996
    option remote-subvolume brickex
end-volume

volume 172.26.98.59-1
    type protocol/client
    option transport-type tcp
    option remote-host 172.26.98.59
    option transport.socket.nodelay on
    option transport.remote-port 6996
    option remote-subvolume brickex
end-volume

#volume 172.26.98.61-1
#    type protocol/client
#    option transport-type tcp
#    option remote-host 172.26.98.61
#    option transport.socket.nodelay on
#    option transport.remote-port 6996
#    option remote-subvolume brickex
#end-volume

#volume 172.26.98.62-1
#    type protocol/client
#    option remote-host 172.26.98.62
#    option transport.socket.nodelay on
#    option transport.remote-port 6996
#    option remote-subvolume brickex
#end-volume

volume distribute-1
    type cluster/dht
    subvolumes 172.26.98.55-1 172.26.98.56-1
end-volume

volume distribute-2
    type cluster/dht
    subvolumes 172.26.98.57-1 172.26.98.59-1
end-volume

#volume distribute-3
#    type cluster/dht
#    subvolumes 172.26.98.61-1 172.26.98.62-1
#end-volume

volume replicate-1
    type cluster/afr
    option lookup-unhashed yes
    subvolumes distribute-1 distribute-2
    #subvolumes distribute-1 distribute-2 distribute-3
end-volume

#volume stripe
#    type cluster/stripe
#    option block-size 1MB
#    subvolumes replicate-1 replicate-2 replicate-3
#end-volume

volume writebehind
    type performance/write-behind
    option cache-size 4MB
    subvolumes replicate-1
end-volume

volume io-cache
  type performance/io-cache
  option cache-size 64MB             # default is 32MB
  #option priority *.h:3,*.html:2,*:1 # default is '*:0'
  option cache-timeout 2             # default is 1 second
  subvolumes writebehind
end-volume

volume stat-prefetch
  type performance/stat-prefetch
  subvolumes io-cache
end-volume







More information about the Gluster-users mailing list