[Gluster-users] dht_layout_dir_mismatch after simultaneously mkdir

kenji kondo kkay.jp at gmail.com
Thu Dec 6 14:37:26 UTC 2012


Hello Avati and all,
I sent the log of getfattr, and gfid mismatch is found.
Is this solution a difficult?
And I'm interested in also the mechanism of gfid mismatch.

Best regards,
Kondo


2012/12/4 kenji kondo <kkay.jp at gmail.com>

> Hello Avati,
>
> I got the log as below.
> We can see that one gfid (g13) differs from others.
> I sometimes found same situation so far.
> Best regards
> Kondo
> ------------------------
>
> backend/ai/anchor/3:
>
> g10:trusted.gfid=0xd1366c6a09c345e8bab26b7953ac6405
>
> g11:trusted.gfid=0xd1366c6a09c345e8bab26b7953ac6405
>
> g12:trusted.gfid=0xd1366c6a09c345e8bab26b7953ac6405
>
> g13:trusted.gfid=0x93c1e7a2797a4b888e17d9b6582da365
>
> g14:trusted.gfid=0xd1366c6a09c345e8bab26b7953ac6405
>
> g15:trusted.gfid=0xd1366c6a09c345e8bab26b7953ac6405
>
> g16:trusted.gfid=0xd1366c6a09c345e8bab26b7953ac6405
>
> g17:trusted.gfid=0xd1366c6a09c345e8bab26b7953ac6405
>
> g18:trusted.gfid=0xd1366c6a09c345e8bab26b7953ac6405
>
> g10:trusted.glusterfs.dht=0x00000001000000003ffffffe5ffffffc
>
> g11:trusted.glusterfs.dht=0x00000001000000005ffffffd7ffffffb
>
> g12:trusted.glusterfs.dht=0x00000001000000007ffffffc9ffffffa
>
> g13:trusted.glusterfs.dht=0x000000010000000000000000ffffffff
>
> g14:trusted.glusterfs.dht=0x00000001000000009ffffffbbffffff9
>
> g15:trusted.glusterfs.dht=0x0000000100000000bffffffadffffff8
>
> g16:trusted.glusterfs.dht=0x0000000100000000dffffff9ffffffff
>
> g17:trusted.glusterfs.dht=0x0000000100000000000000001ffffffe
>
> g18:trusted.glusterfs.dht=0x00000001000000001fffffff3ffffffd
>
>
>
> /backend/ai/anchor:
>
> g10:trusted.gfid=0x89a33935d5da4acd85586e075eca221d
>
> g11:trusted.gfid=0x89a33935d5da4acd85586e075eca221d
>
> g12:trusted.gfid=0x89a33935d5da4acd85586e075eca221d
>
> g13:trusted.gfid=0x6ee644c825dd4806b2680bfeb985a27d
>
> g14:trusted.gfid=0x89a33935d5da4acd85586e075eca221d
>
> g15:trusted.gfid=0x89a33935d5da4acd85586e075eca221d
>
> g16:trusted.gfid=0x89a33935d5da4acd85586e075eca221d
>
> g17:trusted.gfid=0x89a33935d5da4acd85586e075eca221d
>
> g18:trusted.gfid=0x89a33935d5da4acd85586e075eca221d
>
> g10:trusted.glusterfs.dht=0x00000001000000005555555471c71c6f
>
> g11:trusted.glusterfs.dht=0x000000010000000071c71c708e38e38b
>
> g12:trusted.glusterfs.dht=0x00000001000000008e38e38caaaaaaa7
>
> g13:trusted.glusterfs.dht=0x0000000100000000aaaaaaa8c71c71c3
>
> g14:trusted.glusterfs.dht=0x0000000100000000c71c71c4e38e38df
>
> g15:trusted.glusterfs.dht=0x0000000100000000e38e38e0ffffffff
>
> g16:trusted.glusterfs.dht=0x0000000100000000000000001c71c71b
>
> g17:trusted.glusterfs.dht=0x00000001000000001c71c71c38e38e37
>
> g18:trusted.glusterfs.dht=0x000000010000000038e38e3855555553
>
>
>
> /backend/ai
>
> g10.gfid:trusted.gfid=0x258ba47b30ee41a984ceea1a491b1669
>
> g11.gfid:trusted.gfid=0x258ba47b30ee41a984ceea1a491b1669
>
> g12.gfid:trusted.gfid=0x258ba47b30ee41a984ceea1a491b1669
>
> g13.gfid:trusted.gfid=0x258ba47b30ee41a984ceea1a491b1669
>
> g14.gfid:trusted.gfid=0x258ba47b30ee41a984ceea1a491b1669
>
> g15.gfid:trusted.gfid=0x258ba47b30ee41a984ceea1a491b1669
>
> g16.gfid:trusted.gfid=0x258ba47b30ee41a984ceea1a491b1669
>
> g17.gfid:trusted.gfid=0x258ba47b30ee41a984ceea1a491b1669
>
> g18.gfid:trusted.gfid=0x258ba47b30ee41a984ceea1a491b1669
>
> g10.gfid:trusted.glusterfs.dht=0x00000001000000008e38e38caaaaaaa7
>
> g11.gfid:trusted.glusterfs.dht=0x0000000100000000aaaaaaa8c71c71c3
>
> g12.gfid:trusted.glusterfs.dht=0x0000000100000000c71c71c4e38e38df
>
> g13.gfid:trusted.glusterfs.dht=0x0000000100000000e38e38e0ffffffff
>
> g14.gfid:trusted.glusterfs.dht=0x0000000100000000000000001c71c71b
>
> g15.gfid:trusted.glusterfs.dht=0x00000001000000001c71c71c38e38e37
>
> g16.gfid:trusted.glusterfs.dht=0x000000010000000038e38e3855555553
>
> g17.gfid:trusted.glusterfs.dht=0x00000001000000005555555471c71c6f
>
> g18.gfid:trusted.glusterfs.dht=0x000000010000000071c71c708e38e38b
>
>
>
>
>
> On 2012/12/04, at 9:20, Anand Avati <anand.avati at gmail.com> wrote:
>
>
>
> On Mon, Dec 3, 2012 at 5:14 AM, kenji kondo <kkay.jp at gmail.com> wrote:
>
>> Hello Avati, thank you for your reply.
>>
>> I tried to test your suggestion in 3.2.7, but I could not test in 3.3.x
>> because I don't have.
>> For the results, unfortunately new similar problems occurred as following:
>>
>> gluster> volume set vol22 performance.stat-prefetch off
>> gluster> volume info vol22
>>
>> Volume Name: vol22
>> Type: Distribute
>> Status: Started
>> Number of Bricks: 9
>> Transport-type: tcp
>> Bricks:
>> Brick1: gluster10:/export22/brick
>> Brick2: gluster11:/export22/brick
>> Brick3: gluster12:/export22/brick
>> Brick4: gluster13:/export22/brick
>> Brick5: gluster14:/export22/brick
>> Brick6: gluster15:/export22/brick
>> Brick7: gluster16:/export22/brick
>> Brick8: gluster17:/export22/brick
>> Brick9: gluster18:/export22/brick
>> Options Reconfigured:
>> performance.stat-prefetch: off
>>
>>
>> After this setting, I ran same simulation in the vol22.
>> But strange directories were made, that I could not remove the some
>> directories as below:
>>
>> $ ls -a ai/anchor/3
>> .  ..
>>
>> $ rmdir ai/anchor/3
>> rmdir: ai/anchor/3: No such file or directory
>>
>>
>> then I found error messages:
>>
>> [2012-12-03 18:08:14.816313] E [client3_1-fops.c:2228:client3_1_lookup_cbk]
>> 0-vol22-client-3: remote operation failed: Stale NFS file handle
>> [2012-12-03 18:08:14.817196] W [dht-common.c:178:dht_lookup_dir_cbk]
>> 0-vol22-dht: /test1130/ai/anchor: gfid different on vol22-client-4
>> [2012-12-03 18:08:14.817258] W [dht-common.c:178:dht_lookup_dir_cbk]
>> 0-vol22-dht: /test1130/ai/anchor: gfid different on vol22-client-0
>> [2012-12-03 18:08:14.817322] W [dht-common.c:178:dht_lookup_dir_cbk]
>> 0-vol22-dht: /test1130/ai/anchor: gfid different on vol22-client-2
>> [2012-12-03 18:08:14.817367] W [dht-common.c:178:dht_lookup_dir_cbk]
>> 0-vol22-dht: /test1130/ai/anchor: gfid different on vol22-client-1
>> [2012-12-03 18:08:14.817398] W [dht-common.c:178:dht_lookup_dir_cbk]
>> 0-vol22-dht: /test1130/ai/anchor: gfid different on vol22-client-5
>> [2012-12-03 18:08:14.817430] W [dht-common.c:178:dht_lookup_dir_cbk]
>> 0-vol22-dht: /test1130/ai/anchor: gfid different on vol22-client-8
>> [2012-12-03 18:08:14.817460] W [dht-common.c:178:dht_lookup_dir_cbk]
>> 0-vol22-dht: /test1130/ai/anchor: gfid different on vol22-client-7
>> [2012-12-03 18:08:14.817506] W [dht-common.c:178:dht_lookup_dir_cbk]
>> 0-vol22-dht: /test1130/ai/anchor: gfid different on vol22-client-6
>> [2012-12-03 18:08:14.818865] E [client3_1-fops.c:2132:client3_1_opendir_cbk]
>> 0-vol22-client-3: remote operation failed: No such file or directory
>> [2012-12-03 18:08:14.819198] W [fuse-bridge.c:1016:fuse_unlink_cbk]
>> 0-glusterfs-fuse: 1684950: RMDIR()
>> /test1130/ai/anchor/3 => -1 (No such file or directory)
>>
>>
>> And, I found strange dht with getfattr command,
>>
>> $ sudo getfattr -d -m '.*' -n trusted.glusterfs.pathinfo ai/anchor/3
>>
>>
>> trusted.glusterfs.pathinfo="(vol22-dht-layout (vol22-client-0 1073741822
>> 1610612732) (vol22-client-1 1610612733 2147483643) (vol22-client-2 2147483644
>> 2684354554)
>>
>>  (vol22-client-3 0 0)
>>
>> (vol22-client-4 2684354555 3221225465)
>> (vol22-client-5 3221225466 3758096376) (vol22-client-6 3758096377 4294967295)
>> (vol22-client-7 0 536870910) (vol22-client-8 536870911 1073741821))"
>>
>>
>> vol22-client-3 is 0 to 0?: this would be incorrect.
>>
>> Above problem is found in all of our clients.
>> I expect that the problems relate to exclusive control for mkdir, but
>> I don't enough understand this phenomenon.
>>
>> Can you have some idea?
>> I will try.
>>
>>
> In such a state, can you get the backend xattr dumps of ai/anchor/ and
> ai/anchor/3/ with the following commands run on ALL servers -
>
> sh# getfattr -d -e hex -m . /backend/ai/anchor/
> sh# getfattr -d -e hex -m . /backend/ai/anchor/3/
>
> Thanks,
> Avati
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://supercolony.gluster.org/pipermail/gluster-users/attachments/20121206/0e6f4dfc/attachment.html>


More information about the Gluster-users mailing list