[Gluster-devel] GlusterFS crash on on node reboot
Anand Avati
avati at zresearch.com
Thu Jul 19 06:17:44 UTC 2007
Hello,
can you please paste a little more of the logs (a few lines before the
segfault log begins) ?
thanks!
avati
2007/7/19, Insomniac <dcs at digitalcybersoft.com>:
>
> My situation is that I have two nodes, both running client/server. If I
> reboot one of the nodes, then attempt to do a stat (or any other action)
> on the mounted volume glusterfsd locks up for a bit, and then crashes.
> The debug output is as follows:
>
> 2007-07-18 16:15:35 C [common-utils.c:208:gf_print_trace]
> debug-backtrace: Got signal (11), printing backtrace
> 2007-07-18 16:15:35 C [common-utils.c:210:gf_print_trace]
> debug-backtrace: /usr/local/lib/libglusterfs.so.0(gf_print_trace+0x2d)
> [0x2a890d]
> 2007-07-18 16:15:35 C [common-utils.c:210:gf_print_trace]
> debug-backtrace: [0x18f420]
> 2007-07-18 16:15:35 C [common-utils.c:210:gf_print_trace]
> debug-backtrace: /usr/local/lib/libglusterfs.so.0(dict_ref+0x29)
> [0x2a1f89]
> 2007-07-18 16:15:35 C [common-utils.c:210:gf_print_trace]
> debug-backtrace: /usr/local/lib/glusterfs/1.3.0-pre5.4/xlator/protocol/server.so
> [0xb04be8]
> 2007-07-18 16:15:35 C [common-utils.c:210:gf_print_trace]
> debug-backtrace: /usr/local/lib/glusterfs/1.3.0-pre5.4/xlator/protocol/server.so
> [0xb06ac2]
> 2007-07-18 16:15:35 C [common-utils.c:210:gf_print_trace]
> debug-backtrace: /usr/local/lib/glusterfs/1.3.0-pre5.4
> /xlator/performance/io-threads.so [0x8b4ae5]
> 2007-07-18 16:15:35 C [common-utils.c:210:gf_print_trace]
> debug-backtrace: /usr/local/lib/glusterfs/1.3.0-pre5.4/xlator/cluster/unify.so
> [0xbab62c]
> 2007-07-18 16:15:35 C [common-utils.c:210:gf_print_trace]
> debug-backtrace: /usr/local/lib/glusterfs/1.3.0-pre5.4/xlator/cluster/afr.so
> [0x4524c2]
> 2007-07-18 16:15:35 C [common-utils.c:210:gf_print_trace]
> debug-backtrace: /usr/local/lib/glusterfs/1.3.0-pre5.4/xlator/protocol/client.so
> [0x64a215]
> 2007-07-18 16:15:35 C [common-utils.c:210:gf_print_trace]
> debug-backtrace: /usr/local/lib/glusterfs/1.3.0-pre5.4/xlator/protocol/client.so(notify+0x446)
> [0x647ef6]
> 2007-07-18 16:15:35 C [common-utils.c:210:gf_print_trace]
> debug-backtrace: /usr/local/lib/libglusterfs.so.0(transport_notify+0x37)
> [0x2a9e17]
> 2007-07-18 16:15:35 C [common-utils.c:210:gf_print_trace]
> debug-backtrace: /usr/local/lib/libglusterfs.so.0(sys_epoll_iteration
> +0xd9) [0x2aa939]
> 2007-07-18 16:15:35 C [common-utils.c:210:gf_print_trace]
> debug-backtrace: /usr/local/lib/libglusterfs.so.0(poll_iteration+0x1d)
> [0x2a9eed]
> 2007-07-18 16:15:35 C [common-utils.c:210:gf_print_trace]
> debug-backtrace: [glusterfsd] [0x8049250]
> 2007-07-18 16:15:35 C [common-utils.c:210:gf_print_trace]
> debug-backtrace: /lib/libc.so.6(__libc_start_main+0xe0) [0x4748df70]
> 2007-07-18 16:15:35 C [common-utils.c:210:gf_print_trace]
> debug-backtrace: [glusterfsd] [0x8048c31]
>
>
> My server config is as follows:
>
> volume domainfs-ds
> type storage/posix # POSIX FS translator
> option directory /export/domainfs # Export this
> directory
> end-volume
>
> volume domainfs-ns
> type storage/posix # POSIX FS translator
> option directory /home/domainfs-ns # Export this
> directory
> end-volume
>
> volume domainfs-loadb1-ds
> type protocol/client
> option transport-type tcp/client
> option remote-host 192.168.0.2
> option remote-subvolume domainfs-ds
> end-volume
>
> volume domainfs-loadb1-ns
> type protocol/client
> option transport-type tcp/client
> option remote-host 192.168.0.2
> option remote-subvolume domainfs-ns
> end-volume
>
> volume domainfs-loadb2-ds
> type protocol/client
> option transport-type tcp/client
> option remote-host 192.168.0.3
> option remote-subvolume domainfs-ds
> end-volume
>
> volume domainfs-loadb2-ns
> type protocol/client
> option transport-type tcp/client
> option remote-host 192.168.0.3
> option remote-subvolume domainfs-ns
> end-volume
>
> volume domainfs-ns-afr
> type cluster/afr
> # There appears to be a bug with AFR and Local Posix Volumes.
> # To get around this we pretend the local volume is remote
> with an extra client volume named domainfs-santa1-ns.
> # subvolumes domainfs-ns domainfs-santa2-ns
> domainfs-santa3-ns
> subvolumes domainfs-loadb1-ns domainfs-loadb2-ns
> option replicate *:2
> end-volume
>
> volume domainfs-ds-afr
> type cluster/afr
> # There appears to be a bug with AFR and Local Posix Volumes.
> # To get around this we pretend the local volume is remote
> with an extra client volume named domainfs-santa1-ds.
> subvolumes domainfs-loadb1-ds domainfs-loadb2-ds
> option replicate *:2
> end-volume
>
> volume domainfs-unify
> type cluster/unify
> subvolumes domainfs-ds-afr
> option namespace domainfs-ns-afr
> option scheduler rr
> end-volume
>
> volume domainfs
> type performance/io-threads
> option thread-count 8
> option cache-size 64MB
> subvolumes domainfs-unify
> end-volume
>
> ### Add network serving capability to above brick.
> volume server
> type protocol/server
> option transport-type tcp/server # For TCP/IP transport
> subvolumes domainfs
> option auth.ip.domainfs-ds.allow 192.168.0.*,127.0.0.1 # Allow
> access to "brick" volume
> option auth.ip.domainfs-ns.allow 192.168.0.*,127.0.0.1 # Allow
> access to "brick" volume
> option auth.ip.domainfs.allow * # Allow access to "brick" volume
> end-volume
>
>
>
> My client config:
>
> ### Add client feature and attach to remote subvolume
> volume loadb
> type protocol/client
> option transport-type tcp/client # for TCP/IP transport
> option remote-host 192.168.0.3 # IP address of the remote
> brick (points to localhost on both box's)
> option remote-subvolume domainfs # name of the remote volume
> end-volume
>
> # volume loadb2
> # type protocol/client
> # option transport-type tcp/client # for TCP/IP transport
> # option remote-host 192.168.0.3 # IP address of the remote
> brick
> # option remote-subvolume domainfs # name of the remote
> volume
> # end-volume
>
> ### Add writeback feature
> volume writeback
> type performance/write-behind
> option aggregate-size 131072 # unit in bytes
> subvolumes loadb
> end-volume
>
> ### Add readahead feature
> volume readahead
> type performance/read-ahead
> option page-size 65536 # unit in bytes
> option page-count 16 # cache per file = (page-count x
> page-size)
> subvolumes writeback
> end-volume
>
>
>
>
> _______________________________________________
> Gluster-devel mailing list
> Gluster-devel at nongnu.org
> http://lists.nongnu.org/mailman/listinfo/gluster-devel
>
--
Anand V. Avati
More information about the Gluster-devel
mailing list