[Gluster-users] high memory usage of mount

Tamas Papp tompos at martos.bme.hu
Fri Aug 8 07:48:20 UTC 2014


Poornima, would it help, if I give you access via teamviewer?

Thanks,
tamas

On 08/08/2014 09:47 AM, Tamas Papp wrote:
> hi Poornima,
>
> The volume size is 25TB but only 11TB is used.
> It's mostly read and less writing.
>
>
> There are 6 gluster nodes (distributed), each mounted on itself and 
> shared via smb a couple of netatalk clients. I have this issue only on 
> this particular node.
>
> Typical file size varies between a few bytes and 50MB, but mostly 
> under 1MB.
> There are a lot of temp files (created ... deleted and so on).
>
> I have no idea, how much files there are, butI will try to find out.
>
>
> Thanks,
> tamas
>
>
> On 08/08/2014 09:34 AM, Poornima Gurusiddaiah wrote:
>>  From the statedump, it looks like the iobufs are not leaking any more.
>> Inode and dentry have huge hot counts, but it is expected if large 
>> number
>> of files are present and also depends on the kernel parameter 'VFS 
>> cache pressure'.
>> Unable to identify which resource is leaking.
>>
>> Can you provide the workload(data size, number of files, operations) 
>> that is leading to memory leak?
>> This will help us reproduce and debug.
>>
>> Regards,
>> Poornima
>>
>> ----- Original Message -----
>> From: "Tamas Papp" <tompos at martos.bme.hu>
>> To: "Pranith Kumar Karampuri" <pkarampu at redhat.com>, "Poornima 
>> Gurusiddaiah" <pgurusid at redhat.com>
>> Cc: Gluster-users at gluster.org
>> Sent: Wednesday, August 6, 2014 5:59:15 PM
>> Subject: Re: [Gluster-users] high memory usage of mount
>>
>> Yes, I did.
>> I have to do it at least once per day.
>>
>> Currently:
>>
>> $ free
>>                total       used       free     shared    buffers cached
>> Mem:      16422548   16047536     375012          0        320 256884
>> -/+ buffers/cache:   15790332     632216
>> Swap:      5859324    3841584    2017740
>>
>>
>> http://rtfm.co.hu/glusterdump.24405.dump.1407327928.zip
>>
>> Thanks,
>> tamas
>>
>> On 08/06/2014 02:22 PM, Pranith Kumar Karampuri wrote:
>>> You may have to remount the volume so that the already leaked memory
>>> is reclaimed by the system. If you still see the leaks, please provide
>>> the updated statedumps.
>>>
>>> Pranith
>>>
>>> On 08/05/2014 07:30 PM, Tamas Papp wrote:
>>>> Just an update, the settings below did not help for me.
>>>>
>>>> Current settings:
>>>>
>>>> Volume Name: w-vol
>>>> Type: Distribute
>>>> Volume ID: 89e31546-cc2e-4a27-a448-17befda04726
>>>> Status: Started
>>>> Number of Bricks: 5
>>>> Transport-type: tcp
>>>> Bricks:
>>>> Brick1: gl0:/mnt/brick1/export
>>>> Brick2: gl1:/mnt/brick1/export
>>>> Brick3: gl2:/mnt/brick1/export
>>>> Brick4: gl3:/mnt/brick1/export
>>>> Brick5: gl4:/mnt/brick1/export
>>>> Options Reconfigured:
>>>> nfs.mount-udp: on
>>>> nfs.addr-namelookup: off
>>>> nfs.ports-insecure: on
>>>> nfs.port: 2049
>>>> cluster.stripe-coalesce: on
>>>> nfs.disable: off
>>>> performance.flush-behind: on
>>>> performance.io-thread-count: 64
>>>> performance.quick-read: off
>>>> performance.stat-prefetch: on
>>>> performance.io-cache: off
>>>> performance.write-behind: on
>>>> performance.read-ahead: on
>>>> performance.write-behind-window-size: 4MB
>>>> performance.cache-refresh-timeout: 1
>>>> performance.cache-size: 4GB
>>>> network.frame-timeout: 60
>>>> performance.cache-max-file-size: 1GB
>>>>
>>>>
>>>> Cheers,
>>>> tamas
>>>>
>>>> On 08/04/2014 09:22 AM, Tamas Papp wrote:
>>>>> hi Poornima,
>>>>>
>>>>> I don't really have any advice, how you could reproduce this issue
>>>>> also I don't have coredump (the process killed after oom issue).
>>>>>
>>>>> I will see, what can I do.
>>>>>
>>>>>
>>>>> I set the two settings you wrote.
>>>>>
>>>>>
>>>>> Cheers,
>>>>> tamas
>>>>>
>>>>> On 08/04/2014 08:36 AM, Poornima Gurusiddaiah wrote:
>>>>>> Hi,
>>>>>>
>>>>>>   From the statedump it is evident that the iobufs are leaking.
>>>>>> Also the hot count of the pool-name=w-vol-io-cache:rbthash_entry_t
>>>>>> is 10053, implies io-cache xlator could be the cause of the leak.
>>>>>>   From the logs, it looks like, quick-read performance xlator is
>>>>>> calling iobuf_free with NULL pointers, implies quick-read could be
>>>>>> leaking iobufs as well.
>>>>>>
>>>>>> As a temperory solution, could you disable io-cache and/or
>>>>>> quick-read and see if the leak still persists?
>>>>>>
>>>>>> $gluster volume set io-cache off
>>>>>> $gluster volume set quick-read off
>>>>>>
>>>>>> This may reduce the performance to certain extent.
>>>>>>
>>>>>> For further debugging, could you provide the core dump or steps to
>>>>>> reproduce if avaiable?
>>>>>>
>>>>>> Regards,
>>>>>> Poornima
>>>>>>
>>>>>> ----- Original Message -----
>>>>>> From: "Tamas Papp" <tompos at martos.bme.hu>
>>>>>> To: "Poornima Gurusiddaiah" <pgurusid at redhat.com>
>>>>>> Cc: Gluster-users at gluster.org
>>>>>> Sent: Sunday, August 3, 2014 10:33:17 PM
>>>>>> Subject: Re: [Gluster-users] high memory usage of mount
>>>>>>
>>>>>>
>>>>>> On 07/31/2014 09:17 AM, Tamas Papp wrote:
>>>>>>> On 07/31/2014 09:02 AM, Poornima Gurusiddaiah wrote:
>>>>>>>> Hi,
>>>>>>> hi,
>>>>>>>
>>>>>>>> Can you provide the statedump of the process, it can be 
>>>>>>>> obtained as
>>>>>>>> follows:
>>>>>>>> $ gluster --print-statedumpdir  #create this directory if it 
>>>>>>>> doesn't
>>>>>>>> exist.
>>>>>>>> $ kill -USR1 <pid-of-glusterfs-process> #generates state dump.
>>>>>>> http://rtfm.co.hu/glusterdump.2464.dump.1406790562.zip
>>>>>>>
>>>>>>>> Also, xporting Gluster via Samba-VFS-plugin method is preferred 
>>>>>>>> over
>>>>>>>> Fuse mount export. For more details refer to:
>>>>>>>> http://lalatendumohanty.wordpress.com/2014/02/11/using-glusterfs-with-samba-and-samba-vfs-plugin-for-glusterfs-on-fedora-20/ 
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>> When I tried it about half year ago it didn't work properly. 
>>>>>>> Clients
>>>>>>> lost mounts, access errors etc.
>>>>>>>
>>>>>>> But I will give it a try, though it's not included in ubuntu's 
>>>>>>> samba
>>>>>>> AFAIK.
>>>>>>>
>>>>>>>
>>>>>>> Thank you,
>>>>>>> tamas
>>>>>>>
>>>>>>> ps. I forget to mention, I can see this issue only one node. The 
>>>>>>> rest
>>>>>>> of nodes are fine.
>>>>>> hi Poornima,
>>>>>>
>>>>>> Do you have  idea, what's going on here?
>>>>>>
>>>>>> Thanks,
>>>>>> tamas
>>>>> _______________________________________________
>>>>> Gluster-users mailing list
>>>>> Gluster-users at gluster.org
>>>>> http://supercolony.gluster.org/mailman/listinfo/gluster-users
>>>> _______________________________________________
>>>> Gluster-users mailing list
>>>> Gluster-users at gluster.org
>>>> http://supercolony.gluster.org/mailman/listinfo/gluster-users
>
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://supercolony.gluster.org/mailman/listinfo/gluster-users



More information about the Gluster-users mailing list