[Gluster-devel] Handling huge number of file read requests
Amrik Singh
asingh at ideeinc.com
Fri May 4 14:26:24 UTC 2007
Hi Anand,
Please find attached the config files. This is the configuration for a
setup with a single brick.
thanks
### file: client-volume.spec
##############################################
### GlusterFS Client Volume Specification ##
##############################################
#### CONFIG FILE RULES:
### Add client feature and attach to remote subvolume
volume client
type protocol/client
option transport-type tcp/client # for TCP/IP transport
# option transport-type ib-sdp/client # for Infiniband transport
option remote-host 192.168.10.254 # IP address of the remote brick
# option remote-port 6996 # default server port is 6996
option remote-subvolume brick # name of the remote volume
end-volume
### Add writeback feature
volume writeback
type performance/write-back
option aggregate-size 131072 # unit in bytes
subvolumes client
end-volume
### Add readahead feature
volume readahead
type performance/read-ahead
option page-size 65536 # unit in bytes
option page-count 16 # cache per file = (page-count x page-size)
subvolumes writeback
end-volume
### If you are not concerned about performance of interactive commands
### like "ls -l", you wouln't need this translator.
# volume statprefetch
# type performance/stat-prefetch
# option cache-seconds 2 # cache expires in 2 seconds
# subvolumes readahead # add "stat-prefetch" feature to "readahead"
volume
# end-volume
### file: server-volume.spec.sample
##############################################
### GlusterFS Server Volume Specification ##
##############################################
#### CONFIG FILE RULES:
### Export volume "brick" with the contents of "/home/EMS" directory.
volume brick
type storage/posix # POSIX FS translator
option directory /home/EMS # Export this directory
end-volume
### Add network serving capability to above brick.
volume server
type protocol/server
option transport-type tcp/server # For TCP/IP transport
# option transport-type ib-sdp/server # For Infiniband transport
# option bind-address 192.168.1.10 # Default is to listen on all
interfaces
# option listen-port 6996 # Default is 6996
# option client-volume-filename /etc/glusterfs/glusterfs-client.vol
subvolumes brick
# NOTE: Access to any volume through protocol/server is denied by
# default. You need to explicitly grant access through # "auth"
# option.
option auth.ip.brick.allow 192.168.* # Allow access to "brick" volume
end-volume
Amrik
Anand Avati wrote:
> Amrik,
> can you send me your config files of both server and client?
>
> regards,
> avati
>
> On 5/4/07, Amrik Singh <asingh at ideeinc.com> wrote:
>> Hi Guys,
>>
>> We are hoping that glusterfs would help us in the particular problem
>> that we are facing with our cluster. We have a visual search application
>> that runs on a cluster with around 300 processors. These compute nodes
>> run a search for images that are hosted on an NFS server. In certain
>> circumstances all these compute nodes are sending requests for query
>> images at extremely high rates (20-40 images per second). When 300 nodes
>> send 20-40 requests per second for these images, the NFS server just
>> can't cope with it and we start seeing a lot of retransmissions and a
>> very high wait time on the server as well as on the nodes. The images
>> are sized at around 2MB each.
>>
>> With the current application we are not in a position where we can
>> quickly change the way things are being done so we are looking for a
>> file system that can handle this kind of situation. We tried glusterfs
>> with the default settings but we did not see any improvement. Is there a
>> way to tune glusterfs to handle this kind of situation.
>>
>> I can provide more details about our setup as needed.
>>
>>
>> thanks
>>
>> --
>> Amrik Singh
>> Idée Inc.
>> http://www.ideeinc.com
>>
>>
>>
>>
>> _______________________________________________
>> Gluster-devel mailing list
>> Gluster-devel at nongnu.org
>> http://lists.nongnu.org/mailman/listinfo/gluster-devel
>>
>
>
More information about the Gluster-devel
mailing list