[Gluster-users] glusterfsd crashed
Anand Avati
avati at gluster.com
Wed Apr 1 16:10:00 UTC 2009
Greg,
can you describe the sequence of operations performed on the
mountpoint before you faced this?
Avati
On Wed, Apr 1, 2009 at 9:06 PM, Greg <greg at easyflirt.com> wrote:
> Hi,
>
> I've just upgraded to 2.0.0-rc7, and add read-ahead + io-cache translators
> on server side, and got this error on both servers :
>
> 2009-04-01 17:26:21 N [server-protocol.c:7513:mop_setvolume] server:
> accepted client from 10.0.8.3:954
> 2009-04-01 17:26:21 N [server-protocol.c:7513:mop_setvolume] server:
> accepted client from 10.0.8.3:953
> pending frames:
> <D4>J5Y<DB>^?
> patchset: 4e5c297d7c3480d0d3ab1c0c2a184c6a4fb801ef
> signal received: 11
> configuration details:argp 1
> backtrace 1
> bdb->cursor->get 1
> db.h 1
> dlfcn 1
> fdatasync 1
> libpthread 1
> llistxattr 1
> setfsid 1
> spinlock 1
> epoll.h 1
> xattr.h 1
> st_atim.tv_nsec 1
> package-string: glusterfs 2.0.0rc7
> /lib/libc.so.6[0x7fdb58be9f60]
> /lib/libpthread.so.0(pthread_spin_lock+0x0)[0x7fdb58f16630]
> /usr/lib/libglusterfs.so.0(fd_ref+0x1a)[0x7fdb5935139a]
> /usr/lib/libglusterfs.so.0(fop_flush_stub+0x4a)[0x7fdb5934b69a]
> /usr/lib/glusterfs/2.0.0rc7/xlator/performance/io-threads.so(iot_flush+0x28)[0x7fdb5859a848]
> /usr/lib/libglusterfs.so.0(default_flush+0xaa)[0x7fdb59340eba]
> /usr/lib/glusterfs/2.0.0rc7/xlator/performance/write-behind.so(wb_flush+0x268)[0x7fdb5818a088]
> /usr/lib/glusterfs/2.0.0rc7/xlator/performance/read-ahead.so(ra_flush+0xe0)[0x7fdb57f7fb30]
> /usr/lib/glusterfs/2.0.0rc7/xlator/protocol/server.so(server_release+0xf9)[0x7fdb57d69e79]
> /usr/lib/glusterfs/2.0.0rc7/xlator/protocol/server.so(protocol_server_pollin+0xa6)[0x7fdb57d647d6]
> /usr/lib/glusterfs/2.0.0rc7/xlator/protocol/server.so(notify+0x38)[0x7fdb57d64818]
> /usr/lib/glusterfs/2.0.0rc7/transport/socket.so(socket_event_handler+0xe0)[0x7fdb57b57b80]
> /usr/lib/libglusterfs.so.0[0x7fdb593531ef]
> /usr/sbin/glusterfsd(main+0xa81)[0x403a21]
> /lib/libc.so.6(__libc_start_main+0xe6)[0x7fdb58bd61a6]
> /usr/sbin/glusterfsd[0x402519]
> ---------
>
>
> Configuration:
> # file: /etc/glusterfs/glusterfsd.vol
>
> #
> # Volumes
> #
> volume media-small
> type storage/posix
> option directory /var/local/glusterfs/media_small
> end-volume
>
> volume media-medium
> type storage/posix
> option directory /var/local/glusterfs/media_medium
> end-volume
>
> # Lock posix
> volume media-small-locks
> type features/posix-locks
> option mandatory-locks on
> subvolumes media-small
> # subvolumes trash # enable this if you need trash can support (NOTE:
> not present in 1.3.0-pre5+ releases)
> end-volume
>
> volume media-medium-locks
> type features/posix-locks
> option mandatory-locks on
> subvolumes media-medium
> # subvolumes trash # enable this if you need trash can support (NOTE:
> not present in 1.3.0-pre5+ releases)
> end-volume
>
>
> #
> # Performance
> #
> volume media-small-iot
> type performance/io-threads
> subvolumes media-small-locks
> option thread-count 4 # default value is 1
> end-volume
>
> volume media-small-ioc
> type performance/io-cache
> option cache-size 128MB # default is 32MB
> option page-size 128KB # default is 128KB
> subvolumes media-small-iot
> end-volume
>
> volume media-small-wb
> type performance/write-behind
> option flush-behind on # default is off
> subvolumes media-small-ioc
> end-volume
>
> volume media-small-ra
> type performance/read-ahead
> subvolumes media-small-wb
> option page-size 256KB # default is 256KB
> option page-count 4 # default is 2 - cache per file =
> (page-count x page-size)
> option force-atime-update no # defalut is 'no'
> end-volume
>
>
> volume media-medium-iot
> type performance/io-threads
> subvolumes media-medium-locks
> option thread-count 4 # default value is 1
> end-volume
>
> volume media-medium-ioc
> type performance/io-cache
> option cache-size 128MB # default is 32MB
> option page-size 128KB # default is 128KB
> subvolumes media-medium-iot
> end-volume
>
> volume media-medium-wb
> type performance/write-behind
> option flush-behind on # default is off
> subvolumes media-medium-ioc
> end-volume
>
> volume media-medium-ra
> type performance/read-ahead
> subvolumes media-medium-wb
> option page-size 256KB # default is 256KB
> option page-count 4 # default is 2 - cache per file =
> (page-count x page-size)
> option force-atime-update no # defalut is 'no'
> end-volume
>
>
>
>
> #
> # Serveur
> #
> volume server
> type protocol/server
> option transport-type tcp/server
> option auth.addr.media-small-ra.allow 10.0.*.*
> option auth.addr.media-medium-ra.allow 10.0.*.*
> # Autoconfiguration, e.g. :
> # glusterfs -l /tmp/glusterfs.log --server=filer-04 ./Cache
> option client-volume-filename /etc/glusterfs/glusterfs.vol
> subvolumes media-small-ra media-medium-ra # volumes exportés
> end-volume
>
> # file: /etc/glusterfs/glusterfs.vol
>
>
> #
> # Clients
> #
> volume media-small-filer-04
> type protocol/client
> option transport-type tcp/client
> option remote-host filer-04.local
> option remote-subvolume media-small-ra
> end-volume
>
> volume media-small-filer-05
> type protocol/client
> option transport-type tcp/client
> option remote-host filer-05.local
> option remote-subvolume media-small-ra
> end-volume
>
> volume media-medium-filer-04
> type protocol/client
> option transport-type tcp/client
> option remote-host filer-04.local
> option remote-subvolume media-medium-ra
> end-volume
>
> volume media-medium-filer-05
> type protocol/client
> option transport-type tcp/client
> option remote-host filer-05.local
> option remote-subvolume media-medium-ra
> end-volume
>
>
> #
> # Volume principal
> #
> volume afr-small
> # AFR has been renamed to "Replicate" for simplicity.
> type cluster/replicate
> # Il faut mettre le serveur avec le moins d'espace disque en 1er :
> # "When doing a "df -h" on a client, the AVAILABLE disk space will
> display the maximum disk space of the first AFR sub volume defined in the
> spec file. So if you have two servers with 50 gigs and 100 gigs of free disk
> space, and the server with 100 gigs is listed first, then you will see 100
> gigs available even though one server only has 50 gigs free. "
> subvolumes media-small-filer-04 media-small-filer-05
> end-volume
>
> volume afr-medium
> # AFR has been renamed to "Replicate" for simplicity.
> type cluster/replicate
> subvolumes media-medium-filer-04 media-medium-filer-05
> end-volume
>
> #
> # Performance
> #
> volume iot-small
> type performance/io-threads
> option thread-count 8 # default is 1
> subvolumes afr-small
> end-volume
>
> volume readahead-small
> type performance/read-ahead
> subvolumes iot-small
> option page-size 1MB # default is 256KB
> option page-count 4 # default is 2 - cache per file =
> (page-count x page-size)
> option force-atime-update no # defalut is 'no'
> end-volume
>
> volume iocache-small
> type performance/io-cache
> option cache-size 64MB # default is 32MB
> option page-size 256KB # default is 128KB
> subvolumes readahead-small
> end-volume
>
> volume wb-small
> type performance/write-behind
> option window-size 1MB # max 4MB
> option flush-behind on # default is off
> subvolumes iocache-small
> end-volume
>
> volume iot-medium
> type performance/io-threads
> option thread-count 8 # default is 1
> subvolumes afr-medium
> end-volume
>
> volume readahead-medium
> type performance/read-ahead
> subvolumes iot-medium
> option page-size 1MB # default is 256KB
> option page-count 4 # default is 2 - cache per file =
> (page-count x page-size)
> option force-atime-update no # defalut is 'no'
> end-volume
>
> volume iocache-medium
> type performance/io-cache
> option cache-size 64MB # default is 32MB
> option page-size 256KB # default is 128KB
> subvolumes readahead-medium
> end-volume
>
> volume wb-medium
> type performance/write-behind
> option window-size 1MB # max 4MB
> option flush-behind on # default is off
> subvolumes iocache-medium
> end-volume
>
> Clients mount afr-* subvolumes, not with performance translators. I'm
> testing performance translators on servers side.
>
> Regards,
>
> --
> Greg
>
>
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://zresearch.com/cgi-bin/mailman/listinfo/gluster-users
>
More information about the Gluster-users
mailing list