[Gluster-devel] glusterfs v1.3.8 client segfaulting in io-cache

Dan Parsons dparsons at nyip.net
Thu May 8 16:03:45 UTC 2008


Amar, any update on this issue?


Dan Parsons


On May 6, 2008, at 11:21 PM, Dan Parsons wrote:

> Amar, quick question. I've switched to readahead but really wish I  
> could use io-cache. How likely do you think changing block-size from  
> 128kb to 1MB (the same as what stripe uses, based on your advice)  
> would fix the crash issue?
>
>
> Dan Parsons
>
>
> On May 6, 2008, at 12:43 PM, Amar S. Tumballi wrote:
>
>> Mostly give the same block-size (and page-size) in stripe and io- 
>> cache. Just
>> for checking it. But currently you can fall back to read-ahead.
>>
>> Regards,
>> Amar
>>
>> On Tue, May 6, 2008 at 12:38 PM, Dan Parsons <dparsons at nyip.net>  
>> wrote:
>>
>>> Ah, so it's not something I'm doing wrong? Do you think changing
>>> cache-size back to 32MB will prevent the problem from happening?
>>>
>>> Perhaps I should switch to readahead until fix?
>>>
>>>
>>> Dan Parsons
>>>
>>>
>>>
>>> On May 6, 2008, at 12:37 PM, Amar S. Tumballi wrote:
>>>
>>> Thanks for the bug report, We will get back to you in another 2-3  
>>> days
>>>> about it. Mostly with a fix :)
>>>>
>>>> Regards,
>>>> Amar
>>>>
>>>> On Tue, May 6, 2008 at 10:14 AM, Dan Parsons <dparsons at nyip.net>  
>>>> wrote:
>>>> Oh, one more useful bit of information, I see lines like the  
>>>> below a lot
>>>> in glusterfs log files, what do they mean?
>>>>
>>>> 2008-05-05 21:20:11 W [fuse-bridge.c:402:fuse_entry_cbk]  
>>>> glusterfs-fuse:
>>>> 18054459: (34)
>>>> /bio/data/fast-hmmsearch-all/tmpDCex3b_fast-hmmsearch-all_job/ 
>>>> result.tigrfam.TIGR02736.hmmhits
>>>> => 610503040 Rehashing because st_nlink less than dentry maps
>>>>
>>>> Dan Parsons
>>>>
>>>>
>>>>
>>>> On May 6, 2008, at 10:13 AM, Dan Parsons wrote:
>>>>
>>>> I'm experiencing a glusterfs client crash, signal 11, under the  
>>>> io-cache
>>>> xlator. This is on our bioinformatics cluster- the crash happened  
>>>> on 2 out
>>>> of 33 machines. I've verified the hardware stability of the  
>>>> machines.
>>>>
>>>> Running v1.3.8 built May 5th, 2008 from latest downloadable  
>>>> version.
>>>>
>>>> Here is the crash message:
>>>>
>>>> [0xffffe420]
>>>>
>>>> /usr/local/lib/glusterfs/1.3.8/xlator/performance/io- 
>>>> cache.so(ioc_page_wakeup+0x67)[0xb76c5f67]
>>>>
>>>> /usr/local/lib/glusterfs/1.3.8/xlator/performance/io- 
>>>> cache.so(ioc_inode_wakeup+0xb2)[0xb76c6902]
>>>>
>>>> /usr/local/lib/glusterfs/1.3.8/xlator/performance/io- 
>>>> cache.so(ioc_cache_validate_cbk+0xae)[0xb76c1e5e]
>>>>
>>>> /usr/local/lib/glusterfs/1.3.8/xlator/cluster/ 
>>>> stripe.so(stripe_stack_unwind_buf_cbk+0x98)[0xb76cd038]
>>>>
>>>> /usr/local/lib/glusterfs/1.3.8/xlator/protocol/ 
>>>> client.so(client_fstat_cbk+0xcc)[0xb76dd13c]
>>>>
>>>> /usr/local/lib/glusterfs/1.3.8/xlator/protocol/client.so(notify 
>>>> +0xa97)[0xb76db117]
>>>> /usr/local/lib/libglusterfs.so.0(transport_notify+0x38)[0xb7efe978]
>>>> /usr/local/lib/libglusterfs.so.0(sys_epoll_iteration+0xd6) 
>>>> [0xb7eff906]
>>>> /usr/local/lib/libglusterfs.so.0(poll_iteration+0x98)[0xb7efeb28]
>>>> [glusterfs](main+0x85e)[0x804a14e]
>>>> /lib/libc.so.6(__libc_start_main+0xdc)[0x7b1dec]
>>>> [glusterfs][0x8049391]
>>>>
>>>> And here is my config file. The only thing I can think of is  
>>>> maybe my
>>>> cache-size is too big. I want a lot of cache, we have big files,  
>>>> and the
>>>> boxes have the RAM. Anyway, below is the config. If you see any  
>>>> problems
>>>> with it, please let me know. There are no errors on the  
>>>> glusterfsd servers,
>>>> except for an EOF from the machines where glusterfs client  
>>>> segfaulted.
>>>>
>>>> volume fuse
>>>> type mount/fuse
>>>> option direct-io-mode 1
>>>> option entry-timeout 1
>>>> option attr-timeout 1
>>>> option mount-point /glusterfs
>>>> subvolumes ioc
>>>> end-volume
>>>>
>>>> volume ioc
>>>> type performance/io-cache
>>>> option priority *.psiblast:3,*.seq:2,*:1
>>>> option force-revalidate-timeout 5
>>>> option cache-size 1200MB
>>>> option page-size 128KB
>>>> subvolumes stripe0
>>>> end-volume
>>>>
>>>> volume stripe0
>>>> type cluster/stripe
>>>> option alu.disk-usage.exit-threshold 100MB
>>>> option alu.disk-usage.entry-threshold 2GB
>>>> option alu.write-usage.exit-threshold 4%
>>>> option alu.write-usage.entry-threshold 20%
>>>> option alu.read-usage.exit-threshold 4%
>>>> option alu.read-usage.entry-threshold 20%
>>>> option alu.order read-usage:write-usage:disk-usage
>>>> option scheduler alu
>>>> option block-size *:1MB
>>>> subvolumes distfs01 distfs02 distfs03 distfs04
>>>> end-volume
>>>>
>>>> volume distfs04
>>>> type protocol/client
>>>> option remote-subvolume brick
>>>> option remote-host 10.8.101.54
>>>> option transport-type tcp/client
>>>> end-volume
>>>>
>>>> volume distfs03
>>>> type protocol/client
>>>> option remote-subvolume brick
>>>> option remote-host 10.8.101.53
>>>> option transport-type tcp/client
>>>> end-volume
>>>>
>>>> volume distfs02
>>>> type protocol/client
>>>> option remote-subvolume brick
>>>> option remote-host 10.8.101.52
>>>> option transport-type tcp/client
>>>> end-volume
>>>>
>>>> volume distfs01
>>>> type protocol/client
>>>> option remote-subvolume brick
>>>> option remote-host 10.8.101.51
>>>> option transport-type tcp/client
>>>> end-volume
>>>>
>>>>
>>>> Dan Parsons
>>>>
>>>>
>>>>
>>>>
>>>> _______________________________________________
>>>> Gluster-devel mailing list
>>>> Gluster-devel at nongnu.org
>>>> http://lists.nongnu.org/mailman/listinfo/gluster-devel
>>>>
>>>>
>>>>
>>>>
>>>> _______________________________________________
>>>> Gluster-devel mailing list
>>>> Gluster-devel at nongnu.org
>>>> http://lists.nongnu.org/mailman/listinfo/gluster-devel
>>>>
>>>>
>>>>
>>>> --
>>>> Amar Tumballi
>>>> Gluster/GlusterFS Hacker
>>>> [bulde on #gluster/irc.gnu.org]
>>>> http://www.zresearch.com - Commoditizing Super Storage!
>>>>
>>>
>>>
>>>
>>
>>
>> -- 
>> Amar Tumballi
>> Gluster/GlusterFS Hacker
>> [bulde on #gluster/irc.gnu.org]
>> http://www.zresearch.com - Commoditizing Super Storage!
>
>
>
> _______________________________________________
> Gluster-devel mailing list
> Gluster-devel at nongnu.org
> http://lists.nongnu.org/mailman/listinfo/gluster-devel
>






More information about the Gluster-devel mailing list