[Gluster-devel] glusterfs 1.3.1 tla472 crash
NovA
av.nova at gmail.com
Thu Sep 6 14:50:48 UTC 2007
Hello everybody!
Today I've tried to switch from tla patch 184 to the 1.3.1 release.
But client crashes with core, when writing something to imported unify
volume. Here there are the backtrace and my specs:
------------
Using host libthread_db library "/lib64/libthread_db.so.1".
Core was generated by `[glusterfs]
'.
Program terminated with signal 8, Arithmetic exception.
#0 0x00002aaaab569af2 in get_stats_free_disk (this=0x7fffa19e8420) at alu.c:71
in alu.c
#0 0x00002aaaab569af2 in get_stats_free_disk (this=0x7fffa19e8420) at alu.c:71
#1 0x00002aaaab56a21d in update_stat_array_cbk (frame=0x65d2c0,
cookie=0x6147e0, xl=<value optimized out>, op_ret=0,
op_errno=<value optimized out>, trav_stats=0x7fffa19e8420) at alu.c:483
#2 0x00002aaaaaaaf9d5 in client_stats_cbk (frame=0x6560e0,
args=<value optimized out>) at client-protocol.c:3974
#3 0x00002aaaaaab2c33 in notify (this=0x614800, event=<value
optimized out>, data=0x62cd80) at client-protocol.c:4409
#4 0x00002aac092eeb22 in sys_epoll_iteration (ctx=<value optimized
out>) at epoll.c:53
#5 0x000000000040348b in main (argc=6, argv=0x7fffa19e8728) at glusterfs.c:388
------------
--------- server-head.vol ------
# Namespace
volume brick-ns
type storage/posix # POSIX FS translator
option directory /mnt/glusterfs-ns # Export this directory
end-volume
# Data
volume disk
type storage/posix # POSIX FS translator
option directory /mnt/hd # Export this directory
end-volume
volume locks
type features/posix-locks
option mandatory on
subvolumes disk
end-volume
volume brick #iothreads can give performance a boost
type performance/io-threads
option thread-count 8
subvolumes locks
end-volume
volume server
type protocol/server
option transport-type tcp/server # For TCP/IP transport
# option bind-address 192.168.1.10 # Default is to listen on all interfaces
# option listen-port 6996 # Default is 6996
# option client-volume-filename /etc/glusterfs/client.vol
subvolumes brick brick-ns
option auth.ip.brick.allow 10.1.0.* # Allow access to "brick" volume
option auth.ip.brick-ns.allow 10.1.0.* # Allow access to "brick-ns" volume
end-volume
---------------------------------------
-------- server-node.vol -----------
volume disk
type storage/posix
option directory /mnt/hd
end-volume
volume locks
type features/posix-locks
option mandatory on
subvolumes disk
end-volume
volume brick
type performance/io-threads
subvolumes locks
end-volume
volume server
type protocol/server
option transport-type tcp/server # For TCP/IP transport
# option bind-address 192.168.1.10 # Default is to listen on all interfaces
# option listen-port 6996 # Default is 6996
option client-volume-filename /etc/glusterfs/client.vol
subvolumes brick
option auth.ip.brick.allow 10.1.0.* # Allow access to "brick" volume
end-volume
-----------------------------------
------------ client.vol -----------
### Remote subvolumes
volume c0
type protocol/client
option transport-type tcp/client # for TCP/IP transport
option remote-host 10.1.0.1 # IP address of the remote brick
# option remote-port 6996 # default server port is 6996
option remote-subvolume brick # name of the remote volume
end-volume
volume c-ns # namespace
type protocol/client
option transport-type tcp/client
option remote-host 10.1.0.1
option remote-subvolume brick-ns
end-volume
volume c1
type protocol/client
option transport-type tcp/client # for TCP/IP transport
option remote-host 10.1.0.2 # IP address of the remote brick
option remote-subvolume brick # name of the remote volume
end-volume
[ ...skipped... ]
volume c10
type protocol/client
option transport-type tcp/client # for TCP/IP transport
option remote-host 10.1.0.11 # IP address of the remote brick
option remote-subvolume brick # name of the remote volume
end-volume
volume bricks
type cluster/unify
option namespace c-ns # this will not be storage child of unify.
subvolumes c0 c1 c2 c3 c4 c5 c6 c7 c8 c9 c10
option scheduler alu
option alu.limits.min-free-disk 5% #%
option alu.limits.max-open-files 10000
option alu.order
disk-usage:read-usage:write-usage:open-files-usage:disk-speed-usage
option alu.disk-usage.entry-threshold 2GB
option alu.disk-usage.exit-threshold 128MB
option alu.open-files-usage.entry-threshold 1024
option alu.open-files-usage.exit-threshold 32
option alu.read-usage.entry-threshold 20 #%
option alu.read-usage.exit-threshold 4 #%
option alu.write-usage.entry-threshold 20 #%
option alu.write-usage.exit-threshold 4 #%
option alu.disk-speed-usage.entry-threshold 0
option alu.disk-speed-usage.exit-threshold 0
option alu.stat-refresh.interval 10sec
option alu.stat-refresh.num-file-create 10
end-volume
#volume debug
# type debug/trace
# subvolumes bricks
#end-volume
volume threads
type performance/io-threads
option thread-count 11
subvolumes bricks
end-volume
volume wb
type performance/write-behind
option aggregate-size 1MB
subvolumes threads
end-volume
---------------------------------
With best regards,
Andrey
More information about the Gluster-devel
mailing list