[Gluster-users] mnt-mail.log

Bryan McGuire bmcguire at newnet66.org
Wed May 12 01:45:44 UTC 2010


Done..... but now have this error in /var/log/glusterfs/mnt-mail.log

[2010-05-11 20:40:21] E [quick-read.c:2194:init] quickread: FATAL:  
volume (quickread) not configured with exactly one child
[2010-05-11 20:40:21] E [xlator.c:839:xlator_init_rec] quickread:  
Initialization of volume 'quickread' failed, review your volfile again
[2010-05-11 20:40:21] E [glusterfsd.c:591:_xlator_graph_init]  
glusterfs: initializing translator failed
[2010-05-11 20:40:21] E [glusterfsd.c:1394:main] glusterfs: translator  
initialization failed.  exiting

I did change the option max-file-size but when I received the errors I  
put it back to 64kb.

New glusterfs.vol

## file auto generated by /bin/glusterfs-volgen (mount.vol)
# Cmd line:
# $ /bin/glusterfs-volgen --name repstore1 --raid 1 192.168.1.15:/fs  
192.168.1.16:/fs

# RAID 1
# TRANSPORT-TYPE tcp
volume 192.168.1.16-1
     type protocol/client
     option transport-type tcp
     option remote-host 192.168.1.16
     option transport.socket.nodelay on
     option transport.remote-port 6996
     option remote-subvolume brick1
end-volume

volume 192.168.1.15-1
     type protocol/client
     option transport-type tcp
     option remote-host 192.168.1.15
     option transport.socket.nodelay on
     option transport.remote-port 6996
     option remote-subvolume brick1
end-volume

volume mirror-0
     type cluster/replicate
     subvolumes 192.168.1.15-1 192.168.1.16-1
end-volume

#volume readahead
#    type performance/read-ahead
#    option page-count 4
#    subvolumes mirror-0
#end-volume

#volume iocache
#    type performance/io-cache
#    option cache-size `echo $(( $(grep 'MemTotal' /proc/meminfo | sed  
's/[^0-9]//g') / 5120 ))`MB
#    option cache-timeout 1
#    subvolumes readahead
#end-volume

volume quickread
     type performance/quick-read
     option cache-timeout 10
     option max-file-size 64kB
#    subvolumes iocache
end-volume

volume writebehind
     type performance/write-behind
     option cache-size 4MB
     subvolumes quickread
end-volume

volume statprefetch
     type performance/stat-prefetch
     subvolumes writebehind
end-volume




Bryan McGuire


On May 11, 2010, at 7:37 PM, Craig Carl wrote:

> Bryan -
>   Your server vol file isn't perfect for large rsync operations. The  
> changes I'm recommending will improve your rsync performance if you  
> are moving a lot of small files. Please backup the current file  
> before making any changes. You should comment out "readahead" and  
> "iocache". In the quickread section change the "option cache- 
> timeout" to 10 and change the "max-file-size" to the size of the  
> largest file of which you have many, rounded up to the nearest  
> factor of 4.
>   After you have made the changes across all the storage nodes  
> please restart Gluster and measure the throughput again.
>
> Thanks,
>
> Craig
>
>
>
> ----- Original Message -----
> From: "Bryan McGuire" <bmcguire at newnet66.org>
> To: "Craig Carl" <craig at gluster.com>
> Cc: gluster-users at gluster.org
> Sent: Tuesday, May 11, 2010 2:31:48 PM GMT -08:00 US/Canada Pacific
> Subject: Re: [Gluster-users] mnt-mail.log
>
> Here they are,
>
>
> For msvr1 - 192.168.1.15
>
> ## file auto generated by /bin/glusterfs-volgen (export.vol)
> # Cmd line:
> # $ /bin/glusterfs-volgen --name repstore1 --raid 1 192.168.1.15:/fs
> 192.168.1.16:/fs
>
> volume posix1
>   type storage/posix
>   option directory /fs
> end-volume
>
> volume locks1
>     type features/locks
>     subvolumes posix1
> end-volume
>
> volume brick1
>     type performance/io-threads
>     option thread-count 8
>     subvolumes locks1
> end-volume
>
> volume server-tcp
>     type protocol/server
>     option transport-type tcp
>     option auth.addr.brick1.allow *
>     option transport.socket.listen-port 6996
>     option transport.socket.nodelay on
>     subvolumes brick1
> end-volume
>
>
> ## file auto generated by /bin/glusterfs-volgen (mount.vol)
> # Cmd line:
> # $ /bin/glusterfs-volgen --name repstore1 --raid 1 192.168.1.15:/fs
> 192.168.1.16:/fs
>
> # RAID 1
> # TRANSPORT-TYPE tcp
> volume 192.168.1.16-1
>     type protocol/client
>     option transport-type tcp
>     option remote-host 192.168.1.16
>     option transport.socket.nodelay on
>     option transport.remote-port 6996
>     option remote-subvolume brick1
> end-volume
>
> volume 192.168.1.15-1
>     type protocol/client
>     option transport-type tcp
>     option remote-host 192.168.1.15
>     option transport.socket.nodelay on
>     option transport.remote-port 6996
>     option remote-subvolume brick1
> end-volume
>
> volume mirror-0
>     type cluster/replicate
>     subvolumes 192.168.1.15-1 192.168.1.16-1
> end-volume
>
> volume readahead
>     type performance/read-ahead
>     option page-count 4
>     subvolumes mirror-0
> end-volume
>
> volume iocache
>     type performance/io-cache
>     option cache-size `echo $(( $(grep 'MemTotal' /proc/meminfo | sed
> 's/[^0-9]//g') / 5120 ))`MB
>     option cache-timeout 1
>     subvolumes readahead
> end-volume
>
> volume quickread
>     type performance/quick-read
>     option cache-timeout 1
>     option max-file-size 64kB
>     subvolumes iocache
> end-volume
>
> volume writebehind
>     type performance/write-behind
>     option cache-size 4MB
>     subvolumes quickread
> end-volume
>
> volume statprefetch
>     type performance/stat-prefetch
>     subvolumes writebehind
> end-volume
>
>
> For msvr2 192.168.1.16
>
> ## file auto generated by /bin/glusterfs-volgen (export.vol)
> # Cmd line:
> # $ /bin/glusterfs-volgen --name repstore1 --raid 1 192.168.1.15:/fs
> 192.168.1.16:/fs
>
> volume posix1
>   type storage/posix
>   option directory /fs
> end-volume
>
> volume locks1
>     type features/locks
>     subvolumes posix1
> end-volume
>
> volume brick1
>     type performance/io-threads
>     option thread-count 8
>     subvolumes locks1
> end-volume
>
> volume server-tcp
>     type protocol/server
>     option transport-type tcp
>     option auth.addr.brick1.allow *
>     option transport.socket.listen-port 6996
>     option transport.socket.nodelay on
>     subvolumes brick1
> end-volume
>
>
> ## file auto generated by /bin/glusterfs-volgen (mount.vol)
> # Cmd line:
> # $ /bin/glusterfs-volgen --name repstore1 --raid 1 192.168.1.15:/fs
> 192.168.1.16:/fs
>
> # RAID 1
> # TRANSPORT-TYPE tcp
> volume 192.168.1.16-1
>     type protocol/client
>     option transport-type tcp
>     option remote-host 192.168.1.16
>     option transport.socket.nodelay on
>     option transport.remote-port 6996
>     option remote-subvolume brick1
> end-volume
>
> volume 192.168.1.15-1
>     type protocol/client
>     option transport-type tcp
>     option remote-host 192.168.1.15
>     option transport.socket.nodelay on
>     option transport.remote-port 6996
>     option remote-subvolume brick1
> end-volume
>
> volume mirror-0
>     type cluster/replicate
>     subvolumes 192.168.1.15-1 192.168.1.16-1
> end-volume
>
> volume readahead
>     type performance/read-ahead
>     option page-count 4
>     subvolumes mirror-0
> end-volume
>
> volume iocache
>     type performance/io-cache
>     option cache-size `echo $(( $(grep 'MemTotal' /proc/meminfo | sed
> 's/[^0-9]//g') / 5120 ))`MB
>     option cache-timeout 1
>     subvolumes readahead
> end-volume
>
> volume quickread
>     type performance/quick-read
>     option cache-timeout 1
>     option max-file-size 64kB
>     subvolumes iocache
> end-volume
>
> volume writebehind
>     type performance/write-behind
>     option cache-size 4MB
>     subvolumes quickread
> end-volume
>
> volume statprefetch
>     type performance/stat-prefetch
>     subvolumes writebehind
> end-volume
>
>
>
>
>
>
> Bryan McGuire
> Senior Network Engineer
> NewNet 66
>
> 918.231.8063
> bmcguire at newnet66.org
>
>
>
>
>
> On May 11, 2010, at 4:26 PM, Craig Carl wrote:
>
>> Bryan -
>>  Can you send your client and server vol files?
>>
>> Thanks,
>>
>> Craig
>>
>> -- 
>> Craig Carl
>> Sales Engineer
>> Gluster, Inc.
>> Cell - (408) 829-9953 (California, USA)
>> Office - (408) 770-1884
>> Gtalk - craig.carl at gmail.com
>> Twitter - @gluster
>>
>> ----- Original Message -----
>> From: "Bryan McGuire" <bmcguire at newnet66.org>
>> To: gluster-users at gluster.org
>> Sent: Tuesday, May 11, 2010 2:12:13 PM GMT -08:00 US/Canada Pacific
>> Subject: [Gluster-users] mnt-mail.log
>>
>> Hello,
>>
>> I have Glusterfs 3.0.4 setup in a two node replication. It appears to
>> be working just fine. Although I am using rsync to move over 350 Gig
>> of email files and the process is very slow.
>>
>> I have noticed the following in the file /var/log/gluserfs/mntl-
>> mail.log....... could someone explain what the lines mean. Thanks
>>
>> [2010-05-11 15:41:51] W [fuse-bridge.c:491:fuse_entry_cbk] glusterfs-
>> fuse: LOOKUP(/_outgoing/retry/201005100854105298-1273614110_8.tmp)
>> inode (ptr=0xa235c70, ino=808124434, gen=5468694309383486383) found
>> conflict (ptr=0x2aaaea26c290, ino=808124434, gen=5468694309383486383)
>> [2010-05-11 15:46:53] W [fuse-bridge.c:491:fuse_entry_cbk] glusterfs-
>> fuse: LOOKUP(/_outgoing/retry/201005101016464462-1273614395_8.tmp)
>> inode (ptr=0x2aaaf07f5550, ino=808124438, gen=5468694309383486385)
>> found conflict (ptr=0x82e4420, ino=808124438,  
>> gen=5468694309383486385)
>> [2010-05-11 15:46:53] W [fuse-bridge.c:491:fuse_entry_cbk] glusterfs-
>> fuse: LOOKUP(/_outgoing/retry/201005100830599960-1273614395_8.tmp)
>> inode (ptr=0x2aaac01da520, ino=808124430, gen=5468694309383486381)
>> found conflict (ptr=0x60d7a90, ino=808124430,  
>> gen=5468694309383486381)
>> [2010-05-11 15:46:53] W [fuse-bridge.c:491:fuse_entry_cbk] glusterfs-
>> fuse: LOOKUP(/_outgoing/retry/201005101417175132-1273614396_8.tmp)
>> inode (ptr=0x2aaaf07f5550, ino=808124446, gen=5468694309383486389)
>> found conflict (ptr=0x8eb16e0, ino=808124446,  
>> gen=5468694309383486389)
>> [2010-05-11 15:51:53] W [fuse-bridge.c:491:fuse_entry_cbk] glusterfs-
>> fuse: LOOKUP(/_outgoing/retry/201005100749045904-1273614665_8.tmp)
>> inode (ptr=0x1ec11ee0, ino=808124420, gen=5468694309383486379) found
>> conflict (ptr=0x2aaaea26bd30, ino=808124420, gen=5468694309383486379)
>>
>>
>>
>> Bryan
>>
>> _______________________________________________
>> Gluster-users mailing list
>> Gluster-users at gluster.org
>> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
>
>
> -- 
>
>
>
> Craig Carl
>
>
>
> Join us for a Webinar on May 26;
> Case Studies: Deploying Open Source Storage Clouds
> Sales Engineer
> Gluster, Inc.
> Cell - (408) 829-9953 (California, USA)
> Office - (408) 770-1884
> Gtalk - craig.carl at gmail.com
> Twitter - @gluster
> http://www.gluster.com/files/installation-demo/demo.html




More information about the Gluster-users mailing list