[Gluster-users] Gluster 2.0.2 locking up issues

Daniel Jordan Bambach dan at lateral.net
Thu Jun 18 10:53:59 UTC 2009

Willdo, though I recently added in those lines to help be explicit  
about behaviour (I had no options set before at all, leaving it to the  
default of 16 threads). I will remove and specify the default of 16 to  
see if that helps.

Im adding:

volume trace
   type debug/trace
   subvolumes cache

to both sides now as well, so next time (if any) it locks up perhaps  
there will be some more info.

Thanks Shehjar

On 18 Jun 2009, at 11:26, Shehjar Tikoo wrote:

> Daniel Jordan Bambach wrote:
>> I'm experiencing various locking up issues ranging from Gluster  
>> locking up ( 'ls'ing the mount hangs ), to the whole machine  
>> locking up under load.
>> My current config is below (two servers, afring)
>> I would love to be able to get to the bottom of this, because it  
>> seems very strange that we should see erratic behaviour on such a  
>> simple setup.
>> There is approx 12Gb of files, and to stress test (and heal) i run  
>> ls -alR on the mount. This will run for a while and eventually lock  
>> up Gluster, and occasionally the machine. I have found that in some  
>> cases killing Gluster and re-mounting does not solve the problem  
>> (in that perhaps both servers have entered a locked state in some  
>> way).
>> Im finding it very hard to collect and debug information of any  
>> use, as there is no crashlog, no errors in the volume log.
>> Can anyone suggest what I migth be able to do to extract more  
>> information as to what is occuring at lock-up time?
>> volume posix
>> type storage/posix
>> option directory /home/export
>> end-volume
>> volume locks
>>  type features/locks
>>  subvolumes posix
>> end-volume
>> volume brick
>> type performance/io-threads
>> subvolumes locks
>> option autoscaling on
>> option min-threads 8
>> option max-threads 32
>> end-volume
> I see that the max-threads will never exceed 32 which is
> a reasonable valueand should work fine in most cases but considering
> some of the other reports we've been getting, could you please try  
> again
> but without the autoscaling turned on?
> It is off by default, so you can simply set the number of threads
> you need by:
> option thread-count <COUNT>
> ...instead of the three "option" lines above.
> Thanks
> Shehjar
>> volume server
>> type protocol/server
>> option transport-type tcp
>> option auth.addr.brick.allow *
>> subvolumes brick
>> end-volume
>> volume latsrv2
>> type protocol/client
>> option transport-type tcp
>> option remote-host latsrv2
>> option remote-subvolume brick
>> end-volume
>> volume afr
>>  type cluster/replicate
>>  subvolumes brick latsrv2
>>  option read-subvolume brick
>> end-volume
>> volume writebehind
>>  type performance/write-behind
>>  option cache-size 2MB
>>  subvolumes afr
>> end-volume
>> volume cache
>>  type performance/io-cache
>>  option cache-size 32MB
>>  option priority *.pyc:4,*.html:3,*.php:2,*:1
>>  option cache-timeout 5
>>  subvolumes writebehind
>> end-volume
>> _______________________________________________
>> Gluster-users mailing list
>> Gluster-users at gluster.org
>> http://zresearch.com/cgi-bin/mailman/listinfo/gluster-users

More information about the Gluster-users mailing list