[Gluster-devel] GlusterFS /home and KDE
NovA
av.nova at gmail.com
Mon May 7 12:09:09 UTC 2007
Avati,
I've updated GlusterFS to patch-161 half an hour ago. The bug is
still there. :(
As I've found out, it's related with posix/locks, not io-threads. If
posix/locks is turned off in server spec, then KDE starts flawlessly.
Otherwise glusterfsd dies with the following backtrace log:
--------
[May 07 15:33:04] [CRITICAL/common-utils.c:215/gf_print_trace()]
debug-backtrace:Got signal (11), printing backtrace
[May 07 15:33:04] [CRITICAL/common-utils.c:217/gf_print_trace()]
debug-backtrace:/usr/lib64/libglusterfs.so.0(gf_print_trace+0x21)
[0x2b25736faf71]
[May 07 15:33:04] [CRITICAL/common-utils.c:217/gf_print_trace()]
debug-backtrace:/lib64/libc.so.6 [0x2b25741725b0]
[May 07 15:33:04] [CRITICAL/common-utils.c:217/gf_print_trace()]
debug-backtrace:/usr/lib64/glusterfs/1.3.0-pre3/xlator/features/posix-locks.so
[0x2aaaaacb47b7]
[May 07 15:33:04] [CRITICAL/common-utils.c:217/gf_print_trace()]
debug-backtrace:/usr/lib64/glusterfs/1.3.0-pre3/xlator/protocol/server.so
[0x2aaaaaebf4fb]
[May 07 15:33:04] [CRITICAL/common-utils.c:217/gf_print_trace()]
debug-backtrace:/usr/lib64/glusterfs/1.3.0-pre3/xlator/protocol/server.so
[0x2aaaaaeba694]
[May 07 15:33:04] [CRITICAL/common-utils.c:217/gf_print_trace()]
debug-backtrace:/usr/lib64/libglusterfs.so.0(sys_epoll_iteration+0xd4)
[0x2b25736fc7e4]
[May 07 15:33:04] [CRITICAL/common-utils.c:217/gf_print_trace()]
debug-backtrace:[glusterfsd] [0x40164c]
[May 07 15:33:04] [CRITICAL/common-utils.c:217/gf_print_trace()]
debug-backtrace:/lib64/libc.so.6(__libc_start_main+0xf4)
[0x2b257415fae4]
[May 07 15:33:04] [CRITICAL/common-utils.c:217/gf_print_trace()]
debug-backtrace:[glusterfsd] [0x401149]
---------
My server.vol spec contains:
--------
volume disk
type storage/posix # POSIX FS translator
option directory /mnt/hd # Export this directory
end-volume
#volume brick
# type features/posix-locks
# subvolumes disk
#end-volume
volume brick
type performance/io-threads
option thread-count 8
subvolumes disk
end-volume
### Add network serving capability to above brick
volume server
type protocol/server
option transport-type tcp/server # For TCP/IP transport
# option bind-address 192.168.1.10 # Default is to listen on all interfaces
# option listen-port 6996 # Default is 6996
option client-volume-filename /etc/glusterfs/client.vol
subvolumes brick
option auth.ip.brick.allow 10.1.0.* # Allow access to "brick" volume
end-volume
-------
With best regards,
Andrey
2007/5/5, Anand Avati <avati at zresearch.com>:
> Andrey,
> can you please confirm the bug with the latest TLA checkout? few
> fixes have been committed in various places along with io-threads.
>
More information about the Gluster-devel
mailing list