[Gluster-users] Gluster 2.0.2 locking up issues
Daniel Jordan Bambach
dan at lateral.net
Thu Jun 18 10:53:59 UTC 2009
Willdo, though I recently added in those lines to help be explicit
about behaviour (I had no options set before at all, leaving it to the
default of 16 threads). I will remove and specify the default of 16 to
see if that helps.
Im adding:
volume trace
type debug/trace
subvolumes cache
end-volume
to both sides now as well, so next time (if any) it locks up perhaps
there will be some more info.
Thanks Shehjar
On 18 Jun 2009, at 11:26, Shehjar Tikoo wrote:
> Daniel Jordan Bambach wrote:
>> I'm experiencing various locking up issues ranging from Gluster
>> locking up ( 'ls'ing the mount hangs ), to the whole machine
>> locking up under load.
>> My current config is below (two servers, afring)
>> I would love to be able to get to the bottom of this, because it
>> seems very strange that we should see erratic behaviour on such a
>> simple setup.
>> There is approx 12Gb of files, and to stress test (and heal) i run
>> ls -alR on the mount. This will run for a while and eventually lock
>> up Gluster, and occasionally the machine. I have found that in some
>> cases killing Gluster and re-mounting does not solve the problem
>> (in that perhaps both servers have entered a locked state in some
>> way).
>> Im finding it very hard to collect and debug information of any
>> use, as there is no crashlog, no errors in the volume log.
>> Can anyone suggest what I migth be able to do to extract more
>> information as to what is occuring at lock-up time?
>> volume posix
>> type storage/posix
>> option directory /home/export
>> end-volume
>> volume locks
>> type features/locks
>> subvolumes posix
>> end-volume
>> volume brick
>> type performance/io-threads
>> subvolumes locks
>> option autoscaling on
>> option min-threads 8
>> option max-threads 32
>> end-volume
> I see that the max-threads will never exceed 32 which is
> a reasonable valueand should work fine in most cases but considering
> some of the other reports we've been getting, could you please try
> again
> but without the autoscaling turned on?
>
> It is off by default, so you can simply set the number of threads
> you need by:
>
> option thread-count <COUNT>
>
> ...instead of the three "option" lines above.
>
> Thanks
> Shehjar
>
>
>> volume server
>> type protocol/server
>> option transport-type tcp
>> option auth.addr.brick.allow *
>> subvolumes brick
>> end-volume
>> volume latsrv2
>> type protocol/client
>> option transport-type tcp
>> option remote-host latsrv2
>> option remote-subvolume brick
>> end-volume
>> volume afr
>> type cluster/replicate
>> subvolumes brick latsrv2
>> option read-subvolume brick
>> end-volume
>> volume writebehind
>> type performance/write-behind
>> option cache-size 2MB
>> subvolumes afr
>> end-volume
>> volume cache
>> type performance/io-cache
>> option cache-size 32MB
>> option priority *.pyc:4,*.html:3,*.php:2,*:1
>> option cache-timeout 5
>> subvolumes writebehind
>> end-volume
>> _______________________________________________
>> Gluster-users mailing list
>> Gluster-users at gluster.org
>> http://zresearch.com/cgi-bin/mailman/listinfo/gluster-users
>
>
More information about the Gluster-users
mailing list