[Gluster-users] Double counting of quota

Alessandro De Salvo Alessandro.DeSalvo at roma1.infn.it
Tue Jun 9 07:58:43 UTC 2015


Hi Vijiay,
it’s the latest one, I guess:

# gluster --version
glusterfs 3.7.1 built on Jun  1 2015 17:53:10
Repository revision: git://git.gluster.com/glusterfs.git
Copyright (c) 2006-2011 Gluster Inc. <http://www.gluster.com>
GlusterFS comes with ABSOLUTELY NO WARRANTY.
You may redistribute copies of GlusterFS under the terms of the GNU General Public License.

Many thanks,

	Alessandro

> Il giorno 09/giu/2015, alle ore 09:08, Vijaikumar M <vmallika at redhat.com> ha scritto:
> 
> Hi Alessandro,
> 
> We have recently fixed one issues related to below warning message.
> Please provide the output of 'gluster --version', we will check if the issue is the same.
> 
> Thanks,
> Vijay
> 
> 
> On Monday 08 June 2015 06:43 PM, Alessandro De Salvo wrote:
>> OK, many thanks Rajesh.
>> I just wanted to add that I see a lot of warnings in the logs like the following:
>> 
>> [2015-06-08 13:13:10.365633] W [marker-quota.c:3162:mq_initiate_quota_task] 0-atlas-data-01-marker: inode ctx get failed, aborting quota txn
>> 
>> I’m not sure if this is a bug (related or not to the one you mention) or if it is normal and harmless.
>> Thanks,
>> 
>> 
>>         Alessandro
>> 
>> 
>>> Il giorno 08/giu/2015, alle ore 10:39, Rajesh kumar Reddy Mekala <rmekala at redhat.com <mailto:rmekala at redhat.com>> ha scritto:
>>> 
>>> We have open bug 1227724 for the similar problem 
>>> 
>>> Thanks,
>>> Rajesh
>>> 
>>> On 06/08/2015 12:08 PM, Vijaikumar M wrote:
>>>> Hi Alessandro,
>>>> 
>>>> Please provide the test-case, so that we can try to re-create this problem in-house?
>>>> 
>>>> Thanks,
>>>> Vijay
>>>> 
>>>> On Saturday 06 June 2015 05:59 AM, Alessandro De Salvo wrote:
>>>>> Hi,
>>>>> just to answer to myself, it really seems the temp files from rsync are the culprit, it seems that their size are summed up to the real contents of the directories I’m synchronizing, or in other terms their size is not removed from the used size after they are removed. I suppose this is someway connected to the error on removexattr I’m seeing. The temporary solution I’ve found is to use rsync with the option to write the temp files to /tmp, but it would be very interesting to understand why this is happening.
>>>>> Cheers,
>>>>> 
>>>>> 	Alessandro
>>>>> 
>>>>>> Il giorno 06/giu/2015, alle ore 01:19, Alessandro De Salvo <Alessandro.DeSalvo at roma1.infn.it> <mailto:Alessandro.DeSalvo at roma1.infn.it> ha scritto:
>>>>>> 
>>>>>> Hi,
>>>>>> I currently have two brick with replica 2 on the same machine, pointing to different disks of a connected SAN.
>>>>>> The volume itself is fine:
>>>>>> 
>>>>>> # gluster volume info atlas-home-01
>>>>>> 
>>>>>> Volume Name: atlas-home-01
>>>>>> Type: Replicate
>>>>>> Volume ID: 660db960-31b8-4341-b917-e8b43070148b
>>>>>> Status: Started
>>>>>> Number of Bricks: 1 x 2 = 2
>>>>>> Transport-type: tcp
>>>>>> Bricks:
>>>>>> Brick1: host1:/bricks/atlas/home02/data
>>>>>> Brick2: host2:/bricks/atlas/home01/data
>>>>>> Options Reconfigured:
>>>>>> performance.write-behind-window-size: 4MB
>>>>>> performance.io-thread-count: 32
>>>>>> performance.readdir-ahead: on
>>>>>> server.allow-insecure: on
>>>>>> nfs.disable: true
>>>>>> features.quota: on
>>>>>> features.inode-quota: on
>>>>>> 
>>>>>> 
>>>>>> However, when I set a quota on a dir of the volume the size show is twice the physical size of the actual dir:
>>>>>> 
>>>>>> # gluster volume quota atlas-home-01 list /user1
>>>>>>                  Path                   Hard-limit Soft-limit   Used  Available  Soft-limit exceeded? Hard-limit exceeded?
>>>>>> ---------------------------------------------------------------------------------------------------------------------------
>>>>>> /user1                                    4.0GB       80%       3.2GB 853.4MB              No                   No
>>>>>> 
>>>>>> # du -sh /storage/atlas/home/user1
>>>>>> 1.6G    /storage/atlas/home/user1
>>>>>> 
>>>>>> If I remove one of the bricks the quota shows the correct value.
>>>>>> Is there any double counting in case the bricks are on the same machine?
>>>>>> Also, I see a lot of errors in the logs like the following:
>>>>>> 
>>>>>> [2015-06-05 21:59:27.450407] E [posix-handle.c:157:posix_make_ancestryfromgfid] 0-atlas-home-01-posix: could not read the link from the gfid handle /bricks/atlas/home01/data/.glusterfs/be/e5/bee5e2b8-c639-4539-a483-96c19cd889eb (No such file or directory)
>>>>>> 
>>>>>> and also
>>>>>> 
>>>>>> [2015-06-05 22:52:01.112070] E [marker-quota.c:2363:mq_mark_dirty] 0-atlas-home-01-marker: failed to get inode ctx for /user1/file1
>>>>>> 
>>>>>> When running rsync I also see the following errors:
>>>>>> 
>>>>>> [2015-06-05 23:06:22.203968] E [marker-quota.c:2601:mq_remove_contri] 0-atlas-home-01-marker: removexattr trusted.glusterfs.quota.fddf31ba-7f1d-4ba8-a5ad-2ebd6e4030f3.contri failed for /user1/..bashrc.O4kekp: No data available
>>>>>> 
>>>>>> Those files are the temp files of rsync, I’m not sure why the throw errors in glusterfs.
>>>>>> Any help?
>>>>>> Thanks,
>>>>>> 
>>>>>> 	Alessandro
>>>>>> 
>>>>>> 
>>>>>> _______________________________________________
>>>>>> Gluster-users mailing list
>>>>>> Gluster-users at gluster.org <mailto:Gluster-users at gluster.org>
>>>>>> http://www.gluster.org/mailman/listinfo/gluster-users <http://www.gluster.org/mailman/listinfo/gluster-users>
>>>>> 
>>>>> 
>>>>> _______________________________________________
>>>>> Gluster-users mailing list
>>>>> Gluster-users at gluster.org <mailto:Gluster-users at gluster.org>
>>>>> http://www.gluster.org/mailman/listinfo/gluster-users <http://www.gluster.org/mailman/listinfo/gluster-users>
>>>> 
>>>> 
>>>> _______________________________________________
>>>> Gluster-users mailing list
>>>> Gluster-users at gluster.org <mailto:Gluster-users at gluster.org>
>>>> http://www.gluster.org/mailman/listinfo/gluster-users <http://www.gluster.org/mailman/listinfo/gluster-users>
>> 
> 

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20150609/3bd9b26e/attachment.html>
-------------- next part --------------
A non-text attachment was scrubbed...
Name: smime.p7s
Type: application/pkcs7-signature
Size: 1770 bytes
Desc: not available
URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20150609/3bd9b26e/attachment.p7s>


More information about the Gluster-users mailing list