[Gluster-devel] glusterfs 1.3.1 tla472 crash

NovA av.nova at gmail.com
Thu Sep 6 14:50:48 UTC 2007


Hello everybody!

Today I've tried to switch from tla patch 184 to the 1.3.1 release.
But client crashes with core, when writing something to imported unify
volume. Here there are the backtrace and my specs:

------------
Using host libthread_db library "/lib64/libthread_db.so.1".
Core was generated by `[glusterfs]
                             '.
Program terminated with signal 8, Arithmetic exception.
#0  0x00002aaaab569af2 in get_stats_free_disk (this=0x7fffa19e8420) at alu.c:71
	in alu.c
#0  0x00002aaaab569af2 in get_stats_free_disk (this=0x7fffa19e8420) at alu.c:71
#1  0x00002aaaab56a21d in update_stat_array_cbk (frame=0x65d2c0,
cookie=0x6147e0, xl=<value optimized out>, op_ret=0,
    op_errno=<value optimized out>, trav_stats=0x7fffa19e8420) at alu.c:483
#2  0x00002aaaaaaaf9d5 in client_stats_cbk (frame=0x6560e0,
args=<value optimized out>) at client-protocol.c:3974
#3  0x00002aaaaaab2c33 in notify (this=0x614800, event=<value
optimized out>, data=0x62cd80) at client-protocol.c:4409
#4  0x00002aac092eeb22 in sys_epoll_iteration (ctx=<value optimized
out>) at epoll.c:53
#5  0x000000000040348b in main (argc=6, argv=0x7fffa19e8728) at glusterfs.c:388
------------


--------- server-head.vol ------
# Namespace
volume brick-ns
  type storage/posix                  # POSIX FS translator
  option directory /mnt/glusterfs-ns  # Export this directory
end-volume

# Data
volume disk
  type storage/posix              # POSIX FS translator
  option directory /mnt/hd        # Export this directory
end-volume

volume locks
  type features/posix-locks
  option mandatory on
  subvolumes disk
end-volume

volume brick    #iothreads can give performance a boost
  type performance/io-threads
  option thread-count 8
  subvolumes locks
end-volume

volume server
  type protocol/server
  option transport-type tcp/server     # For TCP/IP transport
# option bind-address 192.168.1.10     # Default is to listen on all interfaces
# option listen-port 6996              # Default is 6996
# option client-volume-filename /etc/glusterfs/client.vol
  subvolumes brick brick-ns
  option auth.ip.brick.allow 10.1.0.*     # Allow access to "brick" volume
  option auth.ip.brick-ns.allow 10.1.0.*  # Allow access to "brick-ns" volume
end-volume
---------------------------------------

-------- server-node.vol -----------
volume disk
  type storage/posix
  option directory /mnt/hd
end-volume

volume locks
  type features/posix-locks
  option mandatory on
  subvolumes disk
end-volume

volume brick
  type performance/io-threads
  subvolumes locks
end-volume

volume server
  type protocol/server
  option transport-type tcp/server     # For TCP/IP transport
# option bind-address 192.168.1.10     # Default is to listen on all interfaces
# option listen-port 6996              # Default is 6996
  option client-volume-filename /etc/glusterfs/client.vol
  subvolumes brick
  option auth.ip.brick.allow 10.1.0.*  # Allow access to "brick" volume
end-volume
-----------------------------------

------------ client.vol -----------
### Remote subvolumes
volume c0
  type protocol/client
  option transport-type tcp/client     # for TCP/IP transport
  option remote-host 10.1.0.1          # IP address of the remote brick
# option remote-port 6996              # default server port is 6996
  option remote-subvolume brick        # name of the remote volume
end-volume

volume c-ns  # namespace
  type protocol/client
  option transport-type tcp/client
  option remote-host 10.1.0.1
  option remote-subvolume brick-ns
end-volume

volume c1
  type protocol/client
  option transport-type tcp/client     # for TCP/IP transport
  option remote-host 10.1.0.2          # IP address of the remote brick
  option remote-subvolume brick        # name of the remote volume
end-volume

[ ...skipped... ]

volume c10
  type protocol/client
  option transport-type tcp/client     # for TCP/IP transport
  option remote-host 10.1.0.11         # IP address of the remote brick
  option remote-subvolume brick        # name of the remote volume
end-volume

volume bricks
  type cluster/unify
  option namespace c-ns    # this will not be storage child of unify.
  subvolumes c0 c1 c2 c3 c4 c5 c6 c7 c8 c9 c10

  option scheduler alu
    option alu.limits.min-free-disk  5% #%
    option alu.limits.max-open-files 10000
    option alu.order
disk-usage:read-usage:write-usage:open-files-usage:disk-speed-usage
    option alu.disk-usage.entry-threshold 2GB
    option alu.disk-usage.exit-threshold  128MB
    option alu.open-files-usage.entry-threshold 1024
    option alu.open-files-usage.exit-threshold 32
    option alu.read-usage.entry-threshold 20 #%
    option alu.read-usage.exit-threshold 4 #%
    option alu.write-usage.entry-threshold 20 #%
    option alu.write-usage.exit-threshold 4 #%
    option alu.disk-speed-usage.entry-threshold 0
    option alu.disk-speed-usage.exit-threshold 0
    option alu.stat-refresh.interval 10sec
    option alu.stat-refresh.num-file-create 10
end-volume
				
#volume debug
#    type debug/trace
#    subvolumes bricks
#end-volume

volume threads
  type performance/io-threads
  option thread-count 11
  subvolumes bricks
end-volume

volume wb
  type performance/write-behind
  option aggregate-size 1MB
  subvolumes threads
end-volume
---------------------------------

With best regards,
  Andrey





More information about the Gluster-devel mailing list