[Gluster-devel] GlusterFS crash on on node reboot

Anand Avati avati at zresearch.com
Thu Jul 19 20:16:45 UTC 2007


Hello,
  the bug you hit seems to be most likely the same bug which Urban faced
too. Please update yourself t the latest patchset on TLA and see if you
still face the issue (the fix for Urban's bug is in there)

thanks,
avati

2007/7/19, Anand Avati <avati at zresearch.com>:
>
> Hello,
>   can you please paste a little more of the logs (a few lines before the
> segfault log begins) ?
>
> thanks!
> avati
>
> 2007/7/19, Insomniac < dcs at digitalcybersoft.com>:
> >
> > My situation is that I have two nodes, both running client/server. If I
> > reboot one of the nodes, then attempt to do a stat (or any other action)
> > on the mounted volume glusterfsd locks up for a bit, and then crashes.
> > The debug output is as follows:
> >
> > 2007-07-18 16:15:35 C [common-utils.c :208:gf_print_trace]
> > debug-backtrace: Got signal (11), printing backtrace
> > 2007-07-18 16:15:35 C [common-utils.c:210:gf_print_trace]
> > debug-backtrace: /usr/local/lib/libglusterfs.so.0(gf_print_trace+0x2d)
> > [0x2a890d]
> > 2007-07-18 16:15:35 C [common-utils.c:210:gf_print_trace]
> > debug-backtrace: [0x18f420]
> > 2007-07-18 16:15:35 C [common-utils.c:210:gf_print_trace]
> > debug-backtrace: /usr/local/lib/libglusterfs.so.0(dict_ref+0x29)
> > [0x2a1f89]
> > 2007-07-18 16:15:35 C [common-utils.c:210:gf_print_trace]
> > debug-backtrace: /usr/local/lib/glusterfs/1.3.0-pre5.4/xlator/protocol/server.so
> > [0xb04be8]
> > 2007-07-18 16:15:35 C [common-utils.c:210:gf_print_trace]
> > debug-backtrace: /usr/local/lib/glusterfs/1.3.0-pre5.4/xlator/protocol/server.so
> > [0xb06ac2]
> > 2007-07-18 16:15:35 C [common-utils.c:210:gf_print_trace]
> > debug-backtrace: /usr/local/lib/glusterfs/1.3.0-pre5.4/xlator/performance/io-
> > threads.so [0x8b4ae5]
> > 2007-07-18 16:15:35 C [common-utils.c:210:gf_print_trace]
> > debug-backtrace: /usr/local/lib/glusterfs/1.3.0-pre5.4/xlator/cluster/unify.so
> > [0xbab62c]
> > 2007-07-18 16:15:35 C [common-utils.c:210:gf_print_trace]
> > debug-backtrace: /usr/local/lib/glusterfs/1.3.0-pre5.4/xlator/cluster/afr.so
> > [0x4524c2]
> > 2007-07-18 16:15:35 C [common-utils.c:210:gf_print_trace]
> > debug-backtrace: /usr/local/lib/glusterfs/1.3.0-pre5.4/xlator/protocol/client.so
> > [0x64a215]
> > 2007-07-18 16:15:35 C [common-utils.c:210:gf_print_trace]
> > debug-backtrace: /usr/local/lib/glusterfs/1.3.0-pre5.4/xlator/protocol/client.so(notify+0x446)
> > [0x647ef6]
> > 2007-07-18 16:15:35 C [common-utils.c:210:gf_print_trace]
> > debug-backtrace: /usr/local/lib/libglusterfs.so.0(transport_notify+0x37)
> > [0x2a9e17]
> > 2007-07-18 16:15:35 C [common-utils.c:210:gf_print_trace]
> > debug-backtrace: /usr/local/lib/libglusterfs.so.0(sys_epoll_iteration
> > +0xd9) [0x2aa939]
> > 2007-07-18 16:15:35 C [common-utils.c:210:gf_print_trace]
> > debug-backtrace: /usr/local/lib/libglusterfs.so.0(poll_iteration+0x1d)
> > [0x2a9eed]
> > 2007-07-18 16:15:35 C [common-utils.c:210:gf_print_trace]
> > debug-backtrace: [glusterfsd] [0x8049250]
> > 2007-07-18 16:15:35 C [common-utils.c:210:gf_print_trace]
> > debug-backtrace: /lib/libc.so.6(__libc_start_main+0xe0) [0x4748df70]
> > 2007-07-18 16:15:35 C [common-utils.c:210:gf_print_trace]
> > debug-backtrace: [glusterfsd] [0x8048c31]
> >
> >
> > My server config is as follows:
> >
> >    volume domainfs-ds
> >            type storage/posix                   # POSIX FS translator
> >            option directory /export/domainfs        # Export this
> > directory
> >    end-volume
> >
> >    volume domainfs-ns
> >            type storage/posix                   # POSIX FS translator
> >            option directory /home/domainfs-ns        # Export this
> > directory
> >    end-volume
> >
> >    volume domainfs-loadb1-ds
> >            type protocol/client
> >            option transport-type tcp/client
> >            option remote-host 192.168.0.2
> >            option remote-subvolume domainfs-ds
> >    end-volume
> >
> >    volume domainfs-loadb1-ns
> >            type protocol/client
> >            option transport-type tcp/client
> >            option remote-host 192.168.0.2
> >            option remote-subvolume domainfs-ns
> >    end-volume
> >
> >    volume domainfs-loadb2-ds
> >            type protocol/client
> >            option transport-type tcp/client
> >            option remote-host 192.168.0.3
> >            option remote-subvolume domainfs-ds
> >    end-volume
> >
> >    volume domainfs-loadb2-ns
> >            type protocol/client
> >            option transport-type tcp/client
> >            option remote-host 192.168.0.3
> >            option remote-subvolume domainfs-ns
> >    end-volume
> >
> >    volume domainfs-ns-afr
> >            type cluster/afr
> >            # There appears to be a bug with AFR and Local Posix Volumes.
> >            # To get around this we pretend the local volume is remote
> > with an extra client volume named domainfs-santa1-ns.
> >            # subvolumes domainfs-ns domainfs-santa2-ns
> > domainfs-santa3-ns
> >            subvolumes domainfs-loadb1-ns domainfs-loadb2-ns
> >            option replicate *:2
> >    end-volume
> >
> >    volume domainfs-ds-afr
> >            type cluster/afr
> >            # There appears to be a bug with AFR and Local Posix Volumes.
> >            # To get around this we pretend the local volume is remote
> > with an extra client volume named domainfs-santa1-ds.
> >            subvolumes domainfs-loadb1-ds domainfs-loadb2-ds
> >            option replicate *:2
> >    end-volume
> >
> >    volume domainfs-unify
> >            type cluster/unify
> >            subvolumes domainfs-ds-afr
> >            option namespace domainfs-ns-afr
> >            option scheduler rr
> >    end-volume
> >
> >    volume domainfs
> >            type performance/io-threads
> >            option thread-count 8
> >            option cache-size 64MB
> >            subvolumes domainfs-unify
> >    end-volume
> >
> >    ### Add network serving capability to above brick.
> >    volume server
> >      type protocol/server
> >      option transport-type tcp/server     # For TCP/IP transport
> >      subvolumes domainfs
> >      option auth.ip.domainfs-ds.allow 192.168.0.*,127.0.0.1 # Allow
> > access to "brick" volume
> >      option auth.ip.domainfs-ns.allow 192.168.0.* ,127.0.0.1 # Allow
> > access to "brick" volume
> >      option auth.ip.domainfs.allow * # Allow access to "brick" volume
> >    end-volume
> >
> >
> >
> > My client config:
> >
> >    ### Add client feature and attach to remote subvolume
> >    volume loadb
> >      type protocol/client
> >      option transport-type tcp/client     # for TCP/IP transport
> >      option remote-host 192.168.0.3       # IP address of the remote
> > brick (points to localhost on both box's)
> >      option remote-subvolume domainfs        # name of the remote volume
> >    end-volume
> >
> > #   volume loadb2
> > #     type protocol/client
> > #     option transport-type tcp/client     # for TCP/IP transport
> > #     option remote-host 192.168.0.3       # IP address of the remote
> > brick
> > #     option remote-subvolume domainfs        # name of the remote
> > volume
> > #   end-volume
> >
> >    ### Add writeback feature
> >    volume writeback
> >      type performance/write-behind
> >      option aggregate-size 131072 # unit in bytes
> >      subvolumes loadb
> >    end-volume
> >
> >    ### Add readahead feature
> >    volume readahead
> >      type performance/read-ahead
> >      option page-size 65536     # unit in bytes
> >      option page-count 16       # cache per file  = (page-count x
> > page-size)
> >      subvolumes writeback
> >    end-volume
> >
> >
> >
> >
> > _______________________________________________
> > Gluster-devel mailing list
> > Gluster-devel at nongnu.org
> > http://lists.nongnu.org/mailman/listinfo/gluster-devel
> >
>
>
>
> --
> Anand V. Avati




-- 
Anand V. Avati



More information about the Gluster-devel mailing list