[Gluster-users] samba export: transport endpoint not connected

Ate Poorthuis atepoorthuis at gmail.com
Fri May 1 15:01:57 UTC 2009


Forgot to add that I am using Debian 5.0.1 and glusterfs 2.0.0 final, not
using the gluster patched fuse.

On Fri, May 1, 2009 at 4:47 PM, Ate Poorthuis <atepoorthuis at gmail.com>wrote:

> I have a fully working gluster setup (2 distributed nodes that are
> replicated on the server side to 2 other nodes). When exporting the
> mountpoint as samba share, the gigE link is fully saturated on writing over
> samba. However, when reading, performance is seriously reduced.
>
> Client bandwidth:
>
>   |         iface                   Rx                   Tx
> Total
>
> ==============================================================================
>                lo:           0.00 KB/s            0.00 KB/s            0.00
> KB/s
>              eth0:         255.29 KB/s         5546.58 KB/s         5801.86
> KB/s
>              eth1:       16890.96 KB/s          389.40 KB/s        17280.36
> KB/s
>
> ------------------------------------------------------------------------------
>             total:       17146.25 KB/s         5935.98 KB/s        23082.23
> KB/s
>
> And the following errors shows up in the logs:
>
> 2009-05-01 16:31:25 E [fuse-bridge.c:2332:fuse_getlk_cbk] glusterfs-fuse:
> 73272: ERR => -1 (Transport endpoint is not connected)
> 2009-05-01 16:31:25 E [fuse-bridge.c:2332:fuse_getlk_cbk] glusterfs-fuse:
> 73289: ERR => -1 (Transport endpoint is not connected)
> 2009-05-01 16:31:25 E [fuse-bridge.c:2332:fuse_getlk_cbk] glusterfs-fuse:
> 73306: ERR => -1 (Transport endpoint is not connected)
> 2009-05-01 16:31:25 E [fuse-bridge.c:2332:fuse_getlk_cbk] glusterfs-fuse:
> 73323: ERR => -1 (Transport endpoint is not connected)
>
> I hope someone can provide any pointers on how to solve this. Vol files can
> be found below.
>
> Thanks a lot,
>
> Ate
>
> Server .vol
> --------
> volume posix1
>   type storage/posix                   # POSIX FS translator
>   option directory /srv/export/gfs1/        # Export this directory
> end-volume
>
> volume locks1
>  type features/locks
>  subvolumes posix1
> end-volume
>
> volume read-ahead1
>  type performance/read-ahead
>  option page-count 4
>  subvolumes locks1
> end-volume
>
> volume write-behind1
>  type performance/write-behind
>  subvolumes read-ahead1
> end-volume
>
> volume afr_52
>    type protocol/client
>    option transport-type tcp/client
>    option remote-host 192.168.5.52
>    option remote-subvolume write-behind1
> end-volume
>
> volume afr
>   type cluster/replicate
>   subvolumes write-behind1 afr_52
> end-volume
>
> ### Add network serving capability to above brick.
> volume server
>   type protocol/server
>   option transport-type tcp
>   subvolumes write-behind1 afr
>   option auth.addr.write-behind1.allow * # Allow access to "brick" volume
> option auth.addr.afr.allow * # Allow access to "brick" volume
> end-volume
>
>
> Client .vol
> -----------
> volume brick1_51
>   type protocol/client
>   option transport-type tcp
>   option remote-host 192.168.5.51         # IP address of the remote brick
>   option remote-subvolume afr        # name of the remote volume
> end-volume
> volume brick1_101
>   type protocol/client
>   option transport-type tcp
>   option remote-host 192.168.5.101         # IP address of the remote brick
>   option remote-subvolume afr        # name of the remote volume
> end-volume
>
> volume ha1
>   type /testing/cluster/ha
>   subvolumes brick1_51 brick1_101
> end-volume
>
> volume brick1_52
>   type protocol/client
>   option transport-type tcp
>   option remote-host 192.168.5.52         # IP address of the remote brick
>   option remote-subvolume afr        # name of the remote volume
> end-volume
>
> volume brick1_102
>   type protocol/client
>   option transport-type tcp
>   option remote-host 192.168.5.102         # IP address of the remote brick
>   option remote-subvolume afr        # name of the remote volume
> end-volume
>
> volume ha2
>   type /testing/cluster/ha
>   subvolumes brick1_52 brick1_102
> end-volume
>
> volume bricks
>   type cluster/distribute
>   option min-free-disk 5%
>   subvolumes ha1 ha2
> end-volume
>
> ### Add readahead feature
> volume readahead
>   type performance/read-ahead
>   option page-size 1MB     # unit in bytes
>   option page-count 2       # cache per file  = (page-count x page-size)
>   subvolumes bricks
> end-volume
> ### Add writeback feature
> volume writeback
>   type performance/write-behind
>   option flush-behind off
>   subvolumes readahead
> end-volume
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://supercolony.gluster.org/pipermail/gluster-users/attachments/20090501/533f7f9a/attachment.html>


More information about the Gluster-users mailing list