[Gluster-users] Error : gluster2.0.3rc1 with fuse2.8 inkernel2.6.30 , help !!!!!
Harshavardhana
harsha at gluster.com
Mon Jul 6 09:34:56 UTC 2009
If you can see fuse version on client 2 is different than fuse version of
client 1. Install the fuse -2.7.x release and recompile glusterfs on client
2 machine. It should work.
Regards
--
Harshavardhana
Z Research Inc http://www.zresearch.com/
2009/7/6 eagleeyes <eagleeyes at 126.com>
> How could i remove one of fuse ?
>
>
> client 1
> 2.6.30
> gfs1:~ # dmesg | grep -i fuse
> fuse init (API version 7.11)
> gfs1:~ # grep -i FUSE_MINOR_VERSION /usr/local/include/fuse/fuse_common.h
> #define FUSE_MINOR_VERSION 7
>
> #define FUSE_VERSION FUSE_MAKE_VERSION(FUSE_MAJOR_VERSION, FUSE_MINOR_VERSION)
> # undef FUSE_MINOR_VERSION
> # define FUSE_MINOR_VERSION 5
> # define FUSE_MINOR_VERSION 4
> # define FUSE_MINOR_VERSION 1
> # define FUSE_MINOR_VERSION 1
>
>
> client 2
> 2.6.16.21-0.8-smp
> linux-2ca1:~ # dmesg | grep -i fuse
> fuse: module not supported by Novell, setting U taint flag.
> fuse init (API version 7.6)
>
> linux-2ca1:~ # grep -i FUSE_MINOR_VERSION /usr/local/include/fuse/fuse_common.h
> #define FUSE_MINOR_VERSION 8
>
> #define FUSE_VERSION FUSE_MAKE_VERSION(FUSE_MAJOR_VERSION, FUSE_MINOR_VERSION)
> # undef FUSE_MINOR_VERSION
> # define FUSE_MINOR_VERSION 5
> # define FUSE_MINOR_VERSION 4
> # define FUSE_MINOR_VERSION 1
> # define FUSE_MINOR_VERSION 1
>
>
>
>
>
>
> 2009-07-06
> ------------------------------
> eagleeyes
> ------------------------------
> *发件人:* Harshavardhana
> *发送时间:* 2009-07-06 17:18:34
> *收件人:* eagleeyes
> *抄送:* Anand Avati; gluster-users
> *主题:* Re: [Gluster-users] Error : gluster2.0.3rc1 with fuse2.8
> inkernel2.6.30 ,help !!!!!
> Eagleyes,
>
> I think you are using glusterfs with two different versions of fuse API
> versions. API versions for 2.6.30 kernel are not compatible with 2.6.16-21
> version. I would suggest you to use same fuse API versions for glusterfs.
> Can i have a few details
>
> 1. dmesg | grep -i fuse (on each clients)
> 2. grep -i FUSE_MINOR_VERSION /usr/include/fuse/fuse_common.h (on each
> clients)
>
> Regards
> --
> Harshavardhana
> Z Research Inc http://www.zresearch.com/
>
>
> On Mon, Jul 6, 2009 at 12:31 PM, eagleeyes <eagleeyes at 126.com> wrote:
>
>> HI
>>
>> 1. I use gluster2.0.3rc2 with fuse init (API version 7.11) in SUSE
>> sp10 ,kernel 2.6.30.
>> There were some error log :
>> pending frames:
>> frame : type(1) op(WRITE)
>> frame : type(1) op(READ)
>> frame : type(1) op(READ)
>> frame : type(1) op(READ)
>> patchset: 65524f58b29f0b813549412ba6422711a505f5d8
>> signal received: 11
>> configuration details:argp 1
>> backtrace 1
>> dlfcn 1
>> fdatasync 1
>> libpthread 1
>> llistxattr 1
>> setfsid 1
>> spinlock 1
>> epoll.h 1
>> xattr.h 1
>> st_atim.tv_nsec 1
>> package-string: glusterfs 2.0.3rc2
>> [0xffffe400]
>> /usr/local/lib/libfuse.so.2(fuse_session_process+0x26)[0xb752fb56]
>> /lib/glusterfs/2.0.3rc2/xlator/mount/fuse.so[0xb755de25]
>> /lib/libpthread.so.0[0xb7f0d2ab]
>> /lib/libc.so.6(__clone+0x5e)[0xb7ea4a4e]
>> ---------
>> 2. Use glusterfs 2.0.3rc2 with fuse init (API version 7.6) in suse
>> sp10, kernel 2.6.16.21-0.8-smp ,
>> when i expanded dht volumes from four to six ,then i "rm *" in gluster
>> directory , there were some error :
>>
>> [2009-07-06 22:56:23] W [fuse-bridge.c:921:fuse_unlink_cbk] glusterfs-fuse: 1636: RMDIR() /scheduler => -1 (No such file or directory)
>>
>> [2009-07-06 22:56:23] W [fuse-bridge.c:921:fuse_unlink_cbk] glusterfs-fuse: 1643: RMDIR() /transport => -1 (No such file or directory)
>>
>> [2009-07-06 22:56:23] W [fuse-bridge.c:921:fuse_unlink_cbk] glusterfs-fuse: 1655: RMDIR() /xlators/cluster => -1 (No such file or directory)
>>
>> [2009-07-06 22:56:23] W [fuse-bridge.c:921:fuse_unlink_cbk] glusterfs-fuse: 1666: RMDIR() /xlators/debug => -1 (No such file or directory)
>> [2009-07-06 22:56:23] W [fuse-bridge.c:921:fuse_unlink_cbk] glusterfs-fuse: 1677: RMDIR() /xlators/mount => -1 (No such file or directory)
>>
>>
>> and new files didn't write into the new volumes after expansion .
>>
>>
>>
>>
>> 2009-07-06
>> ------------------------------
>> eagleeyes
>> ------------------------------
>> *发件人:* Anand Avati
>> *发送时间:* 2009-07-06 12:09:13
>> *收件人:* eagleeyes
>> *抄送:* gluster-users
>> *主题:* Re: [Gluster-users] Error : gluster2.0.3rc1 with fuse2.8 in
>> kernel2.6.30 ,help !!!!!
>> Please use 2.0.3 stable, or upgrade to the next rc2 until then. This
>> has been fixed in rc2.
>> Avati
>> On Mon, Jul 6, 2009 at 8:31 AM, eagleeyes<eagleeyes at 126.com> wrote:
>> > HI
>> > I use gluster2.0.3rc1 with fuse 2.8 in kernel
>>
>> > 2.6.30(SUSE Linux Enterprise Server 10 SP1 with kernel 2.6.30 ) . the mount
>> > message was :
>> >
>> > /dev/hda4 on /data type reiserfs (rw,user_xattr)
>>
>> > glusterfs-client.vol.dht on /home type fuse.glusterfs (rw,allow_other,default_permissions,max_read=131072)
>> >
>> >
>> >
>>
>> > There was some error when i "touce 111" in gluster directory ,the error was
>> > :
>> > /home: Transport endpoint is not connected
>> >
>> > pending frames:
>> > patchset: e0db4ff890b591a58332994e37ce6db2bf430213
>> > signal received: 11
>> > configuration details:argp 1
>> > backtrace 1
>> > dlfcn 1
>> > fdatasync 1
>> > libpthread 1
>> > llistxattr 1
>> > setfsid 1
>> > spinlock 1
>> > epoll.h 1
>> > xattr.h 1
>> > st_atim.tv_nsec 1
>> > package-string: glusterfs 2.0.3rc1
>> > [0xffffe400]
>> > /lib/glusterfs/2.0.3rc1/xlator/mount/fuse.so[0xb75c6288]
>>
>> > /lib/glusterfs/2.0.3rc1/xlator/performance/write-behind.so(wb_create_cbk+0xa7)[0xb75ccad7]
>>
>> > /lib/glusterfs/2.0.3rc1/xlator/performance/io-cache.so(ioc_create_cbk+0xde)[0xb7fbe8ae]
>>
>> > /lib/glusterfs/2.0.3rc1/xlator/performance/read-ahead.so(ra_create_cbk+0x167)[0xb7fc78b7]
>>
>> > /lib/glusterfs/2.0.3rc1/xlator/cluster/dht.so(dht_create_cbk+0xf7)[0xb75e25b7]
>>
>> > /lib/glusterfs/2.0.3rc1/xlator/protocol/client.so(client_create_cbk+0x2ad)[0xb76004ad]
>>
>> > /lib/glusterfs/2.0.3rc1/xlator/protocol/client.so(protocol_client_interpret+0x1ef)[0xb75ef8ff]
>>
>> > /lib/glusterfs/2.0.3rc1/xlator/protocol/client.so(protocol_client_pollin+0xcf)[0xb75efaef]
>>
>> > /lib/glusterfs/2.0.3rc1/xlator/protocol/client.so(notify+0x1ec)[0xb75f6ddc]
>>
>> > /lib/glusterfs/2.0.3rc1/transport/socket.so(socket_event_poll_in+0x3b)[0xb75b775b]
>>
>> > /lib/glusterfs/2.0.3rc1/transport/socket.so(socket_event_handler+0xae)[0xb75b7b8e]
>> > /lib/libglusterfs.so.0[0xb7facbda]
>> > /lib/libglusterfs.so.0(event_dispatch+0x21)[0xb7fabac1]
>> > glusterfs(main+0xc2e)[0x804b6ae]
>> > /lib/libc.so.6(__libc_start_main+0xdc)[0xb7e6087c]
>> > glusterfs[0x8049c11]
>> > ---------
>> >
>> > the server configuration
>> >
>> > gfs1:/ # cat /etc/glusterfs/glusterfsd-sever.vol
>> > volume posix1
>> > type storage/posix # POSIX FS translator
>> > option directory /data/data1 # Export this directory
>> > end-volume
>> > volume posix2
>> > type storage/posix # POSIX FS translator
>> > option directory /data/data2 # Export this directory
>> > end-volume
>> > volume posix3
>> > type storage/posix # POSIX FS translator
>> > option directory /data/data3 # Export this directory
>> > end-volume
>> > volume posix4
>> > type storage/posix # POSIX FS translator
>> > option directory /data/data4 # Export this directory
>> > end-volume
>> > volume posix5
>> > type storage/posix # POSIX FS translator
>> > option directory /data/data5 # Export this directory
>> > end-volume
>> > volume posix6
>> > type storage/posix # POSIX FS translator
>> > option directory /data/data6 # Export this directory
>> > end-volume
>> > volume posix7
>> > type storage/posix # POSIX FS translator
>> > option directory /data/data7 # Export this directory
>> > end-volume
>> > volume posix8
>> > type storage/posix # POSIX FS translator
>> > option directory /data/data8 # Export this directory
>> > end-volume
>> > volume brick1
>> > type features/posix-locks
>>
>> > option mandatory-locks on # enables mandatory locking on all files
>> > subvolumes posix1
>> > end-volume
>> > volume brick2
>> > type features/posix-locks
>>
>> > option mandatory-locks on # enables mandatory locking on all files
>> > subvolumes posix2
>> > end-volume
>> > volume brick3
>> > type features/posix-locks
>>
>> > option mandatory-locks on # enables mandatory locking on all files
>> > subvolumes posix3
>> > end-volume
>> > volume brick4
>> > type features/posix-locks
>>
>> > option mandatory-locks on # enables mandatory locking on all files
>> > subvolumes posix4
>> > end-volume
>> > volume brick5
>> > type features/posix-locks
>>
>> > option mandatory-locks on # enables mandatory locking on all files
>> > subvolumes posix5
>> > end-volume
>> > volume brick6
>> > type features/posix-locks
>>
>> > option mandatory-locks on # enables mandatory locking on all files
>> > subvolumes posix6
>> > end-volume
>> > volume brick7
>> > type features/posix-locks
>>
>> > option mandatory-locks on # enables mandatory locking on all files
>> > subvolumes posix7
>> > end-volume
>> > volume brick8
>> > type features/posix-locks
>>
>> > option mandatory-locks on # enables mandatory locking on all files
>> > subvolumes posix8
>> > end-volume
>> > ### Add network serving capability to above brick.
>> > volume server
>> > type protocol/server
>> > option transport-type tcp
>>
>> > option transport.socket.bind-address 172.20.92.240 # Default is to listen on all interfaces
>>
>> > option transport.socket.listen-port 6996 # Default is 6996
>> > subvolumes brick1 brick2 brick3 brick4
>> > option auth.addr.brick1.allow * # Allow access to "brick" volume
>> > option auth.addr.brick2.allow * # Allow access to "brick" volume
>> > option auth.addr.brick3.allow * # Allow access to "brick" volume
>> > option auth.addr.brick4.allow * # Allow access to "brick" volume
>> > option auth.addr.brick5.allow * # Allow access to "brick" volume
>> > option auth.addr.brick6.allow * # Allow access to "brick" volume
>> > option auth.addr.brick7.allow * # Allow access to "brick" volume
>> > option auth.addr.brick8.allow * # Allow access to "brick" volume
>> > end-volume
>> >
>> > the client configuration:
>> >
>> > gfs1:/ # cat /etc/glusterfs/glusterfs-client.vol.dht
>> > volume client1
>> > type protocol/client
>> > option transport-type tcp
>>
>> > option remote-host 172.20.92.240 # IP address of the remote brick2
>> > option remote-port 6996
>> > option remote-subvolume brick1 # name of the remote volume
>> > end-volume
>> > volume client2
>> > type protocol/client
>> > option transport-type tcp
>>
>> > option remote-host 172.20.92.240 # IP address of the remote brick2
>> > option remote-port 6996
>> > #option transport-timeout 10 # seconds to wait for a reply
>> > option remote-subvolume brick2 # name of the remote volume
>> > end-volume
>> > volume client3
>> > type protocol/client
>> > option transport-type tcp
>>
>> > option remote-host 172.20.92.240 # IP address of the remote brick2
>> > option remote-port 6996
>> > #option transport-timeout 10 # seconds to wait for a reply
>> > option remote-subvolume brick3 # name of the remote volume
>> > end-volume
>> > volume client4
>> > type protocol/client
>> > option transport-type tcp
>>
>> > option remote-host 172.20.92.240 # IP address of the remote brick2
>> > option remote-port 6996
>> > #option transport-timeout 10 # seconds to wait for a reply
>> > option remote-subvolume brick4 # name of the remote volume
>> > end-volume
>> > volume client5
>> > type protocol/client
>> > option transport-type tcp
>>
>> > option remote-host 172.20.92.240 # IP address of the remote brick2
>> > option remote-port 6996
>> > #option transport-timeout 10 # seconds to wait for a reply
>> > option remote-subvolume brick1 # name of the remote volume
>> > end-volume
>> > volume client6
>> > type protocol/client
>> > option transport-type tcp
>>
>> > option remote-host 172.20.92.240 # IP address of the remote brick2
>> > option remote-port 6996
>> > #option transport-timeout 10 # seconds to wait for a reply
>> > option remote-subvolume brick2 # name of the remote volume
>> > end-volume
>> > volume client7
>> > type protocol/client
>> > option transport-type tcp
>>
>> > option remote-host 172.20.92.240 # IP address of the remote brick2
>> > option remote-port 6996
>> > #option transport-timeout 10 # seconds to wait for a reply
>> > option remote-subvolume brick3 # name of the remote volume
>> > end-volume
>> > volume client8
>> > type protocol/client
>> > option transport-type tcp
>>
>> > option remote-host 172.20.92.240 # IP address of the remote brick2
>> > option remote-port 6996
>> > #option transport-timeout 10 # seconds to wait for a reply
>> > option remote-subvolume brick4 # name of the remote volume
>> > end-volume
>> > #volume afr3
>> > # type cluster/afr
>> > # subvolumes client3 client6
>> > #end-volume
>> > volume dht
>> > type cluster/dht
>> > option lookup-unhashed yes
>> > subvolumes client1 client2 client3 client4
>> > end-volume
>> >
>> > Could you help me ?
>> >
>> >
>> >
>> > 2009-07-06
>> > ________________________________
>> > eagleeyes
>> > ________________________________
>> > 发件人: Sachidananda
>> > 发送时间: 2009-07-04 11:39:03
>> > 收件人: eagleeyes
>> > 抄送: gluster-users
>> > 主题: Re: [Gluster-users] HELP : Files lost after DHT expansion
>> > Hi,
>> > eagleeyes wrote:
>>
>> > > When i update to gluster2.0.3 ,after dht expansion ,double directorys
>> > > appear in the gluster directory ,why ?
>> > >
>> > > client configure
>> > > volume dht
>> > > type cluster/dht
>> > > option lookup-unhashed yes
>> > > option min-free-disk 10%
>> > > subvolumes client1 client2 client3 client4 client5 client6 client7
>> > client8
>> > > #subvolumes client1 client2 client3 client4
>> > > end-volume
>> > >
>> > >
>> > Can you please send us your server/client volume files?
>> > --
>> > Sachidananda.
>> > _______________________________________________
>> > Gluster-users mailing list
>> > Gluster-users at gluster.org
>> > http://zresearch.com/cgi-bin/mailman/listinfo/gluster-users
>> >
>> >
>>
>> _______________________________________________
>> Gluster-users mailing list
>> Gluster-users at gluster.org
>> http://zresearch.com/cgi-bin/mailman/listinfo/gluster-users
>>
>>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://supercolony.gluster.org/pipermail/gluster-users/attachments/20090706/21eec691/attachment.html>
More information about the Gluster-users
mailing list