[Gluster-devel] Files not available in all clients immediately

Claudio Cuqui claudio at c3systems.com.br
Wed Mar 19 02:00:21 UTC 2008


Amar,

I replaced server and all clients with 710. Using this version I was 
able to mount glusterfs server . But, the original problem remains:

2008-03-18 22:54:40 W [fuse-bridge.c:402:fuse_entry_cbk] glusterfs-fuse: 
8728: /QHc9Gi1lXRnURud3MmGz3tyjSgLfmT5M => 204460 Rehashing because 
st_nlink less than dentry maps
2008-03-18 22:54:40 W [fuse-bridge.c:402:fuse_entry_cbk] glusterfs-fuse: 
8735: /pnnLJ3EstrnDQQHjQtRJd5VwjKye2Bhn => 204462 Rehashing because 
st_nlink less than dentry maps
2008-03-18 22:54:40 W [fuse-bridge.c:402:fuse_entry_cbk] glusterfs-fuse: 
8742: /ysIRd6Pss7Z6W7GP16z0UQ96y4VRHTIJ => 204463 Rehashing because 
st_nlink less than dentry maps
2008-03-18 22:55:34 W [fuse-bridge.c:402:fuse_entry_cbk] glusterfs-fuse: 
9599: /9BFa3KJ9cpkbai6AzXaMIborxLYao2lt => 204435 Rehashing because 
st_nlink less than dentry maps
2008-03-18 22:55:52 W [fuse-bridge.c:402:fuse_entry_cbk] glusterfs-fuse: 
9827: /KBHHPXIUFzJvje2sWk9QqzEqoAOLpGwf => 204495 Rehashing because 
st_nlink less than dentry maps
2008-03-18 22:55:52 W [fuse-bridge.c:402:fuse_entry_cbk] glusterfs-fuse: 
9833: /51MOR3FyFCsQPQuBqaL0NnrSYDRjcwj6 => 204518 Rehashing because 
st_nlink less than dentry maps
2008-03-18 22:56:01 W [fuse-bridge.c:402:fuse_entry_cbk] glusterfs-fuse: 
9927: /Lqnkrr4420JJyZ2XJqGkqsgR9Q0UeX6T => 204466 Rehashing because 
st_nlink less than dentry maps
2008-03-18 22:56:01 W [fuse-bridge.c:402:fuse_entry_cbk] glusterfs-fuse: 
9934: /T4XxoQpSxMELBBIwpC3Whme5t8yYMRJa => 204471 Rehashing because 
st_nlink less than dentry maps
2008-03-18 22:56:01 W [fuse-bridge.c:402:fuse_entry_cbk] glusterfs-fuse: 
9948: /YC0qJs9fREgAE9UFkGiFGgkAa1bNYHt5 => 204488 Rehashing because 
st_nlink less than dentry maps
2008-03-18 22:56:01 W [fuse-bridge.c:402:fuse_entry_cbk] glusterfs-fuse: 
9955: /2EzNMzA38tj3fCLGJkrzPBeU49vleWKM => 204489 Rehashing because 
st_nlink less than dentry maps
2008-03-18 22:56:23 W [fuse-bridge.c:402:fuse_entry_cbk] glusterfs-fuse: 
10323: /gV2xJR4hsHZj4olgRRlUFJTzcSIRym3a => 204448 Rehashing because 
st_nlink less than dentry maps
2008-03-18 22:56:24 W [fuse-bridge.c:402:fuse_entry_cbk] glusterfs-fuse: 
10341: /eQEObb9IFV4VLzpvXHkxm90mjW85ih5j => 204416 Rehashing because 
st_nlink less than dentry maps
2008-03-18 22:56:24 W [fuse-bridge.c:402:fuse_entry_cbk] glusterfs-fuse: 
10347: /737k73PiweusfPf4CN6SPTXmRVGeBC5O => 204421 Rehashing because 
st_nlink less than dentry maps
2008-03-18 22:56:24 W [fuse-bridge.c:402:fuse_entry_cbk] glusterfs-fuse: 
10357: /FVqYNAeqgDYIxzahNybKtk0gVw6TX1kr => 204497 Rehashing because 
st_nlink less than dentry maps
2008-03-18 22:56:24 W [fuse-bridge.c:402:fuse_entry_cbk] glusterfs-fuse: 
10364: /WKW7SNvGeRRGzl7USIIrQwsrKKDtIu69 => 204498 Rehashing because 
st_nlink less than dentry maps
2008-03-18 22:56:28 W [fuse-bridge.c:402:fuse_entry_cbk] glusterfs-fuse: 
10402: /fWTxnBDfas3pth57dA4dwGYgfrOYWB2G => 204398 Rehashing because 
st_nlink less than dentry maps
2008-03-18 22:56:35 W [fuse-bridge.c:402:fuse_entry_cbk] glusterfs-fuse: 
11919: /Q5eQvZg1E8L3Uaj0X1bDD5B9LAGRwevB => 204486 Rehashing because 
st_nlink less than dentry maps

Regards,

Claudio


Amar S. Tumballi wrote:
> Hi Claudio,
>  I made a fix for that bug and patch-710 should work fine for you. You 
> can just upgrade the client machine to make a quick test.
>
> Regards,
> Amar
>
> On Tue, Mar 18, 2008 at 5:37 PM, Amar S. Tumballi <amar at zresearch.com 
> <mailto:amar at zresearch.com>> wrote:
>
>     Nope, thats the latest. But this should be fixed soon (at office
>     time IST).
>     sorry for the inconvenience.
>
>     -amar
>
>
>     On Tue, Mar 18, 2008 at 5:07 PM, Claudio Cuqui
>     <claudio at c3systems.com.br <mailto:claudio at c3systems.com.br>> wrote:
>
>         Hi Avati,
>
>         I tried, but it don´t even allow me to start it:
>
>         TLA Repo Revision: glusterfs--mainline--2.5--patch-709
>         Time : 2008-03-18 20:52:32
>         Signal Number : 11
>
>         /C3Systems/gluster/bin/sbin/glusterfs -f
>         /C3Systems/gluster/bin/etc/glusterfs/glusterfs-client.vol -l
>         /C3Systems/gluster/bin-patch709/var/log/glusterfs/glusterfs.log -L
>         WARNING /C3Systems/data/domains/webmail.pop.com.br/attachments
>         volume fuse
>          type mount/fuse
>          option direct-io-mode 1
>          option entry-timeout 1
>          option attr-timeout 1
>          option mount-point
>         /C3Systems/data/domains/webmail.pop.com.br/attachments
>          subvolumes iocache
>         end-volume
>
>         volume iocache
>          type performance/io-cache
>          option page-count 2
>          option page-size 256KB
>          subvolumes readahead
>         end-volume
>
>         volume readahead
>          type performance/read-ahead
>          option page-count 2
>          option page-size 1MB
>          subvolumes client
>         end-volume
>
>         volume client
>          type protocol/client
>          option remote-subvolume attachments
>          option remote-host 200.175.8.85 <http://200.175.8.85>
>          option transport-type tcp/client
>         end-volume
>
>         frame : type(1) op(34)
>
>         /lib64/libc.so.6[0x3edca300c0]
>         /lib64/libc.so.6(strcmp+0x0)[0x3edca75bd0]
>         /C3Systems/gluster/bin-patch709/lib/glusterfs/1.3.8/xlator/mount/fuse.so[0x2aaaab302937]
>         /C3Systems/gluster/bin-patch709/lib/glusterfs/1.3.8/xlator/mount/fuse.so[0x2aaaab302b42]
>         /C3Systems/gluster/bin-patch709/lib/glusterfs/1.3.8/xlator/performance/io-cache.so(ioc_lookup_cbk+0x67)[0x2aaaab0f6557]
>         /C3Systems/gluster/bin-patch709/lib/libglusterfs.so.0[0x2aaaaaab8344]
>         /C3Systems/gluster/bin-patch709/lib/glusterfs/1.3.8/xlator/protocol/client.so(client_lookup_cbk+0x1b3)[0x2aaaaace93a3]
>         /C3Systems/gluster/bin-patch709/lib/glusterfs/1.3.8/xlator/protocol/client.so(notify+0x8fc)[0x2aaaaace273c]
>         /C3Systems/gluster/bin-patch709/lib/libglusterfs.so.0(sys_epoll_iteration+0xc0)[0x2aaaaaabdb90]
>         /C3Systems/gluster/bin-patch709/lib/libglusterfs.so.0(poll_iteration+0x75)[0x2aaaaaabd095]
>         [glusterfs](main+0x658)[0x4026b8]
>         /lib64/libc.so.6(__libc_start_main+0xf4)[0x3edca1d8a4]
>         [glusterfs][0x401b89]
>         ---------
>
>         Is there any other release that I should try ?
>
>         Regards,
>
>         Cuqui
>
>         Anand Avati wrote:
>         > Claudio,
>         >  Can you try with glusterfs--mainline--2.5--patch-709 ? A
>         similar
>         > issue is addressed in that revision. We are interested to
>         know if that
>         > solves your issue as well.
>         >
>         > thanks,
>         >
>         > avati
>         >
>         > 2008/3/19, Claudio Cuqui <claudio at c3systems.com.br
>         <mailto:claudio at c3systems.com.br>
>         > <mailto:claudio at c3systems.com.br
>         <mailto:claudio at c3systems.com.br>>>:
>         >
>         >     Hi there !
>         >
>         >     We are using gluster on an environment with multiple
>         webservers
>         >     and load
>         >     balancer, where we have only one server and multiple
>         clients (6).
>         >     All servers are running Fedora Core 6 X86_64 with kernel
>         >     2.6.22.14-72.fc6 (with exactly same packages installed
>         in all server).
>         >     The gluster version used is 1.3.8pre2 + 2.7.2glfs8 (both
>         compiled
>         >     locally). The underlying FS is reiserfs mounted with the
>         follow
>         >     options
>         >     rw,noatime,nodiratime,notail. This filesystem has almost
>         4 thousand
>         >     files from 2k - 10Mb in size. We are using gluster to
>         export this
>         >     filesystem for all other webservers. Below the config
>         file used by
>         >     gluster server:
>         >
>         >     ### Export volume "brick" with the contents of
>         "/home/export"
>         >     directory.
>         >     volume attachments-nl
>         >       type storage/posix                   # POSIX FS translator
>         >       option directory
>         >     /C3Systems/data/domains/webmail.pop.com.br/attachments
>         >     end-volume
>         >
>         >     volume attachments
>         >       type features/posix-locks
>         >       subvolumes attachments-nl
>         >       option mandatory on
>         >     end-volume
>         >
>         >
>         >     ### Add network serving capability to above brick.
>         >     volume server
>         >       type protocol/server
>         >       option transport-type tcp/server     # For TCP/IP
>         transport
>         >       option client-volume-filename
>         >     /C3Systems/gluster/bin/etc/glusterfs/glusterfs-client.vol
>         >       subvolumes attachments-nl attachments
>         >       option auth.ip.attachments-nl.allow * # Allow access to
>         >     "attachments-nl" volume
>         >       option auth.ip.attachments.allow * # Allow access to
>         >     "attachments" volume
>         >     end-volume
>         >
>         >     The problem happen when the LB sent the post (the
>         uploaded file)
>         >     to one
>         >     webserver and than the next post goes to other webserver
>          that try to
>         >     access the same file. When this happen, the other client
>         got these
>         >     messages:
>         >
>         >     PHP Warning:
>         >    
>         fopen(/C3Systems/data/domains/c3systems.com.br/attachments/27gBgFQSIiOLDEo7AvxlpsFkqZw9jdnZ):
>         >     failed to open stream: File Not Found.
>         >     PHP Warning:
>         >    
>         unlink(/C3Systems/data/domains/c3systems.com.br/attachments/5Dech7jNxjORZ2cZ9IAbR7kmgmgn2vTE):
>         >     File Not Found.
>         >
>         >     LB is using RoundRobin to distribute the load between
>         servers.
>         >
>         >     Below, you can find the gluster configuration file used
>         by all
>         >     clients:
>         >
>         >     ### file: client-volume.spec.sample
>         >
>         >     ##############################################
>         >     ###  GlusterFS Client Volume Specification  ##
>         >     ##############################################
>         >
>         >     #### CONFIG FILE RULES:
>         >     ### "#" is comment character.
>         >     ### - Config file is case sensitive
>         >     ### - Options within a volume block can be in any order.
>         >     ### - Spaces or tabs are used as delimitter within a line.
>         >     ### - Each option should end within a line.
>         >     ### - Missing or commented fields will assume default
>         values.
>         >     ### - Blank/commented lines are allowed.
>         >     ### - Sub-volumes should already be defined above before
>         referring.
>         >
>         >     ### Add client feature and attach to remote subvolume
>         >     volume client
>         >       type protocol/client
>         >       option transport-type tcp/client     # for TCP/IP
>         transport
>         >     # option ib-verbs-work-request-send-size  1048576
>         >     # option ib-verbs-work-request-send-count 16
>         >     # option ib-verbs-work-request-recv-size  1048576
>         >     # option ib-verbs-work-request-recv-count 16
>         >     # option transport-type ib-sdp/client  # for Infiniband
>         transport
>         >     # option transport-type ib-verbs/client # for ib-verbs
>         transport
>         >       option remote-host 1.2.3.4 <http://1.2.3.4>
>         <http://1.2.3.4>      # IP address of
>         >     the remote brick
>         >     # option remote-port 6996              # default server
>         port is 6996
>         >
>         >     # option transport-timeout 30          # seconds to wait
>         for a reply
>         >                                            # from server for
>         each request
>         >       option remote-subvolume attachments  # name of the
>         remote volume
>         >     end-volume
>         >
>         >     ### Add readahead feature
>         >     volume readahead
>         >       type performance/read-ahead
>         >       option page-size 1MB     # unit in bytes
>         >       option page-count 2       # cache per file  =
>         (page-count x
>         >     page-size)
>         >       subvolumes client
>         >     end-volume
>         >
>         >     ### Add IO-Cache feature
>         >     volume iocache
>         >       type performance/io-cache
>         >       option page-size 256KB
>         >       option page-count 2
>         >       subvolumes readahead
>         >     end-volume
>         >
>         >     ### Add writeback feature
>         >     #volume writeback
>         >     #  type performance/write-behind
>         >     #  option aggregate-size 1MB
>         >     #  option flush-behind off
>         >     #  subvolumes iocache
>         >     #end-volume
>         >
>         >     When I do the test manually, everything goes fine. What
>         I think is
>         >     happening is that gluster isn´t having enough time to
>         sync all clients
>         >     before clients trying to access the files (those servers
>         are very busy
>         >     ones.....they receive millions of requests per day).
>         >
>         >     Is this configuration appropriate for this situation ? a
>         bug ? a
>         >     feature
>         >     ;-) ? Is there any option like the sync used in NFS that
>         I can use in
>         >     order guarantee that when the file is write down, all
>         the clients
>         >     already
>         >     have it ?
>         >
>         >     TIA,
>         >
>         >     Claudio Cuqui
>         >
>         >
>         >
>         >
>         >
>         >     _______________________________________________
>         >     Gluster-devel mailing list
>         >     Gluster-devel at nongnu.org
>         <mailto:Gluster-devel at nongnu.org>
>         <mailto:Gluster-devel at nongnu.org
>         <mailto:Gluster-devel at nongnu.org>>
>         >     http://lists.nongnu.org/mailman/listinfo/gluster-devel
>         >
>         >
>         >
>         >
>         > --
>         > If I traveled to the end of the rainbow
>         > As Dame Fortune did intend,
>         > Murphy would be there to tell me
>         > The pot's at the other end.
>
>         _______________________________________________
>         Gluster-devel mailing list
>         Gluster-devel at nongnu.org <mailto:Gluster-devel at nongnu.org>
>         http://lists.nongnu.org/mailman/listinfo/gluster-devel
>
>
>
>
>     -- 
>     Amar Tumballi
>     Gluster/GlusterFS Hacker
>     [bulde on #gluster/irc.gnu.org]
>     http://www.zresearch.com - Commoditizing Supercomputing and
>     Superstorage! 
>
>
>
>
> -- 
> Amar Tumballi
> Gluster/GlusterFS Hacker
> [bulde on #gluster/irc.gnu.org]
> http://www.zresearch.com - Commoditizing Supercomputing and Superstorage! 




More information about the Gluster-devel mailing list