[Gluster-users] GlusterFS v3.1.5 Stable Configuration
Mohit Anchlia
mohitanchlia at gmail.com
Mon Jul 18 18:59:12 UTC 2011
On Mon, Jul 18, 2011 at 10:53 AM, Remi Broemeling <remi at goclio.com> wrote:
> Hi,
>
> We've been using GlusterFS to manage shared files across a number of hosts
> in the past few months and have ran into a few problems -- basically one
> every month, roughly. The problems are occasionally extremely difficult to
> track down to GlusterFS, as they often masquerade as something else in the
> application log files that we have. The problems have been one instance of
> split-brain and then a number of instances of "stuck" files (i.e. any stat
> calls would block for an hour and then timeout with an error) as well as a
> couple instances of "ghost" files (remove the file, but GlusterFS continues
> to show it for a little while until the cache times out).
>
> We do not place a large amount of load on GlusterFS, and don't have any
> significant performance issues to deal with. With that in mind, the core
> question of this e-mail is: "How can I modify our configuration to be the
> absolute most stable (problem free) that it can be, even if it means
> sacrificing performance?" In sum, I don't have any particular performance
It depends on kind of bugs or issues you are encountering. There might
be solution for some bugs and may not be for others.
> concerns at this moment, but the GlusterFS bugs that we encounter are quite
> problematic -- so I'm willing to entertain any suggested stability
> improvement, even if it has a negative impact on performance (I suspect that
> the answer here is just "turn off all performance-enhancing gluster
> caching", but I wanted to validate that is actually true before going so
> far). Thus please suggest anything that could be done to improve the
> stability of our setup -- as an aside, I think that this would be an
> advantageous thing to add to the FAQ. Right now the FAQ contains
> information for performance tuning, but not for stability tuning.
>
> Thanks for any help that you can give/suggestions that you can make.
>
> Here are the details of our environment:
>
> OS: RHEL5
> GlusterFS Version: 3.1.5
> Mount method: glusterfsd/FUSE
> GlusterFS Servers: web01, web02
> GlusterFS Clients: web01, web02, dj01, dj02
>
> $ sudo gluster volume info
>
> Volume Name: shared-application-data
> Type: Replicate
> Status: Started
> Number of Bricks: 2
> Transport-type: tcp
> Bricks:
> Brick1: web01:/var/glusterfs/bricks/shared
> Brick2: web02:/var/glusterfs/bricks/shared
> Options Reconfigured:
> network.ping-timeout: 5
> nfs.disable: on
>
> Configuration File Contents:
> /etc/glusterd/vols/shared-application-data/shared-application-data-fuse.vol
> volume shared-application-data-client-0
> type protocol/client
> option remote-host web01
> option remote-subvolume /var/glusterfs/bricks/shared
> option transport-type tcp
> option ping-timeout 5
> end-volume
>
> volume shared-application-data-client-1
> type protocol/client
> option remote-host web02
> option remote-subvolume /var/glusterfs/bricks/shared
> option transport-type tcp
> option ping-timeout 5
> end-volume
>
> volume shared-application-data-replicate-0
> type cluster/replicate
> subvolumes shared-application-data-client-0
> shared-application-data-client-1
> end-volume
>
> volume shared-application-data-write-behind
> type performance/write-behind
> subvolumes shared-application-data-replicate-0
> end-volume
>
> volume shared-application-data-read-ahead
> type performance/read-ahead
> subvolumes shared-application-data-write-behind
> end-volume
>
> volume shared-application-data-io-cache
> type performance/io-cache
> subvolumes shared-application-data-read-ahead
> end-volume
>
> volume shared-application-data-quick-read
> type performance/quick-read
> subvolumes shared-application-data-io-cache
> end-volume
>
> volume shared-application-data-stat-prefetch
> type performance/stat-prefetch
> subvolumes shared-application-data-quick-read
> end-volume
>
> volume shared-application-data
> type debug/io-stats
> subvolumes shared-application-data-stat-prefetch
> end-volume
>
> /etc/glusterfs/glusterd.vol
> volume management
> type mgmt/glusterd
> option working-directory /etc/glusterd
> option transport-type socket,rdma
> option transport.socket.keepalive-time 10
> option transport.socket.keepalive-interval 2
> end-volume
>
> --
> Remi Broemeling
> System Administrator
> Clio - Practice Management Simplified
> 1-888-858-2546 x(2^5) | remi at goclio.com
> www.goclio.com | blog | twitter | facebook
>
> ____
> _⌠ oo ⌡_
> (_ _)
> | |
> ⌡_⌡⌡_⌡
>
>
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
>
>
More information about the Gluster-users
mailing list