[Gluster-devel] no element to prune

Pooya Woodcock pooya at packetcloud.net
Tue Jul 10 19:10:18 UTC 2007


Here are my configs...

# SERVER 1
volume brick
         type storage/posix
         option directory /GlusterFS
end-volume

volume locks
         type features/posix-locks
         subvolumes brick
end-volume

volume ns
         type storage/posix
         option directory /GlusterFS-NS
end-volume

volume brick-afr
         type storage/posix
         option directory /GlusterFS-AFR
end-volume

volume locks-afr
   type features/posix-locks
   subvolumes brick-afr
end-volume


volume server
         type protocol/server
         option transport-type tcp/server
         option bind-address 10.175.175.134
         option listen-port 6996
         subvolumes locks locks-afr
         option auth.ip.ns.allow 10.175.175.*
         option auth.ip.locks.allow 10.175.175.*
         option auth.ip.locks-afr.allow 10.175.175.*
end-volume

volume trace
         type debug/trace
         subvolumes server
         option debug on
end-volume

# SERVER 2-4
same as above without the namespace.

# CLIENT
volume ns
type protocol/client
option transport-type tcp/client
option remote-host 10.175.175.134
option remote-subvolume ns
end-volume

volume brick1
type protocol/client
option transport-type tcp/client
option remote-host 10.175.175.134
option remote-port 6996
option remote-subvolume locks
end-volume

volume brick1-afr
type protocol/client
option transport-type tcp/client
option remote-host 10.175.175.135
option remote-port 6996
option remote-subvolume locks-afr
end-volume

volume brick2
type protocol/client
option transport-type tcp/client
option remote-host 10.175.175.135
option remote-port 6996
option remote-subvolume locks
end-volume

volume brick2-afr
type protocol/client
option transport-type tcp/client
option remote-host 10.175.175.136
option remote-port 6996
option remote-subvolume locks-afr
end-volume

volume brick3
type protocol/client
option transport-type tcp/client
option remote-host 10.175.175.136
option remote-port 6996
option remote-subvolume locks
end-volume

volume brick3-afr
type protocol/client
option transport-type tcp/client
option remote-host 10.175.175.137
option remote-port 6996
option remote-subvolume locks-afr
end-volume

volume brick4
type protocol/client
option transport-type tcp/client
option remote-host 10.175.175.137
option remote-port 6996
option remote-subvolume locks
end-volume

volume brick4-afr
type protocol/client
option transport-type tcp/client
option remote-host 10.175.175.134
option remote-port 6996
option remote-subvolume locks-afr
end-volume

volume afr1
type cluster/afr
subvolumes brick1 brick1-afr
option replicate *:2
end-volume

volume afr2
type cluster/afr
subvolumes brick2 brick2-afr
option replicate *:2
end-volume

volume afr3
type cluster/afr
subvolumes brick3 brick3-afr
option replicate *:2
end-volume

volume afr4
type cluster/afr
subvolumes brick4 brick4-afr
option replicate *:2
end-volume

volume unify1
type cluster/unify
subvolumes afr1 afr2 afr3 afr4
   option namespace ns
   option readdir-force-success on
   option scheduler rr
   option rr.limits.min-free-disk 10GB
end-volume

volume iothreads
         type performance/io-threads
         option thread-count 8
         subvolumes unify1
end-volume

volume readahead
         type performance/read-ahead
         option page-size 131072
         option page-count 16
         subvolumes iothreads
end-volume

#volume io-cache
#       type performance/io-cache
#       option page-count 16
#       option page-size 1MB
#       subvolumes readahead
#end-volume




On Jul 10, 2007, at 12:05 PM, Pooya Woodcock wrote:

>
> Hi everyone, been a while since I last posted since I've been  
> running very table in the 2.4 branch. Just recently I took the  
> plunge--- I am running mainline 2.5 latest TLA as for an hour ago  
> (294?)
>
> 2007-07-10 12:01:11 D [inode.c:332:__passive_inode] brick-afr/ 
> inode: passivating inode(34735742), lru=56/1000
> 2007-07-10 12:01:11 D [server-protocol.c:639:server_inode_prune]  
> locks-afr: no element to prune
> 2007-07-10 12:01:11 D [server-protocol.c:639:server_inode_prune]  
> locks-afr: no element to prune
> 2007-07-10 12:01:11 D [server-protocol.c:639:server_inode_prune]  
> locks-afr: no element to prune
> 2007-07-10 12:01:11 D [server-protocol.c:639:server_inode_prune]  
> locks-afr: no element to prune
> 2007-07-10 12:01:11 D [server-protocol.c:639:server_inode_prune]  
> locks-afr: no element to prune
> 2007-07-10 12:01:12 D [inode.c:302:__active_inode] brick-afr/inode:  
> activating inode(34685086), lru=55/1000
> 2007-07-10 12:01:12 D [server-protocol.c:639:server_inode_prune]  
> locks-afr: no element to prune
> 2007-07-10 12:01:12 D [inode.c:332:__passive_inode] brick-afr/ 
> inode: passivating inode(34685086), lru=56/1000
> 2007-07-10 12:01:12 D [inode.c:302:__active_inode] brick/inode:  
> activating inode(11862174), lru=55/1000
> 2007-07-10 12:01:12 D [server-protocol.c:639:server_inode_prune]  
> locks: no element to prune
> 2007-07-10 12:01:12 D [inode.c:332:__passive_inode] brick/inode:  
> passivating inode(11862174), lru=56/1000
> 2007-07-10 12:01:12 D [inode.c:302:__active_inode] brick/inode:  
> activating inode(11912029), lru=55/1000
>
> There are lots of "no element to prune"s on my daemon log files.  
> Should I worry about this?
>
> [root at node04 ~]# tail /var/log/glusterfsd.log
> 2007-07-10 12:03:45 D [inode.c:332:__passive_inode] brick-afr/ 
> inode: passivating inode(10552395), lru=57/1000
> 2007-07-10 12:03:45 D [inode.c:302:__active_inode] brick-afr/inode:  
> activating inode(10552395), lru=56/1000
> 2007-07-10 12:03:45 D [server-protocol.c:639:server_inode_prune]  
> locks-afr: no element to prune
> 2007-07-10 12:03:45 D [inode.c:302:__active_inode] brick-afr/inode:  
> activating inode(10552520), lru=55/1000
> 2007-07-10 12:03:45 D [server-protocol.c:639:server_inode_prune]  
> locks-afr: no element to prune
> 2007-07-10 12:03:45 D [inode.c:332:__passive_inode] brick-afr/ 
> inode: passivating inode(10552520), lru=56/1000
> 2007-07-10 12:03:45 D [inode.c:302:__active_inode] brick/inode:  
> activating inode(12141234), lru=56/1000
> 2007-07-10 12:03:45 D [server-protocol.c:639:server_inode_prune]  
> locks: no element to prune
> 2007-07-10 12:03:45 D [inode.c:332:__passive_inode] brick/inode:  
> passivating inode(12141234), lru=57/1000
> 2007-07-10 12:03:45 D [server-protocol.c:639:server_inode_prune]  
> locks: no element to prune
>
>
> _______________________________________________
> Gluster-devel mailing list
> Gluster-devel at nongnu.org
> http://lists.nongnu.org/mailman/listinfo/gluster-devel






More information about the Gluster-devel mailing list