[Gluster-users] Deleting large files on sharded volume hangs and doesn't delete shards

Walter Deignan WDeignan at uline.com
Wed May 24 18:02:43 UTC 2017


Sorry for the slow response. Your hunch was right. It seems to be a 
problem between tiering and sharding.

I untiered the volume and the symptom vanished. I then deleted and 
recreated the volume entirely (without tiering) in order to cleanup the 
orphaned shards.

-Walter Deignan
-Uline IT, Systems Architect



From:   Nithya Balachandran <nbalacha at redhat.com>
To:     Walter Deignan <WDeignan at uline.com>
Cc:     gluster-users <gluster-users at gluster.org>
Date:   05/17/2017 10:17 PM
Subject:        Re: [Gluster-users] Deleting large files on sharded volume 
hangs and doesn't delete shards



I don't think we have tested shards with a tiered volume.  Do you see such 
issues on non-tiered sharded volumes?

Regards,
Nithya

On 18 May 2017 at 00:51, Walter Deignan <WDeignan at uline.com> wrote:
I have a reproducible issue where attempting to delete a file large enough 
to have been sharded hangs. I can't kill the 'rm' command and eventually 
am forced to reboot the client (which in this case is also part of the 
gluster cluster). After the node finishes rebooting I can see that while 
the file front-end is gone, the back-end shards are still present. 

Is this a known issue? Any way to get around it? 

---------------------------------------------- 

[root at dc-vihi19 ~]# gluster volume info gv0 

Volume Name: gv0 
Type: Tier 
Volume ID: d42e366f-381d-4787-bcc5-cb6770cb7d58 
Status: Started 
Snapshot Count: 0 
Number of Bricks: 24 
Transport-type: tcp 
Hot Tier : 
Hot Tier Type : Distributed-Replicate 
Number of Bricks: 4 x 2 = 8 
Brick1: dc-vihi71:/gluster/bricks/brick4/data 
Brick2: dc-vihi19:/gluster/bricks/brick4/data 
Brick3: dc-vihi70:/gluster/bricks/brick4/data 
Brick4: dc-vihi19:/gluster/bricks/brick3/data 
Brick5: dc-vihi71:/gluster/bricks/brick3/data 
Brick6: dc-vihi19:/gluster/bricks/brick2/data 
Brick7: dc-vihi70:/gluster/bricks/brick3/data 
Brick8: dc-vihi19:/gluster/bricks/brick1/data 
Cold Tier: 
Cold Tier Type : Distributed-Replicate 
Number of Bricks: 8 x 2 = 16 
Brick9: dc-vihi19:/gluster/bricks/brick5/data 
Brick10: dc-vihi70:/gluster/bricks/brick1/data 
Brick11: dc-vihi19:/gluster/bricks/brick6/data 
Brick12: dc-vihi71:/gluster/bricks/brick1/data 
Brick13: dc-vihi19:/gluster/bricks/brick7/data 
Brick14: dc-vihi70:/gluster/bricks/brick2/data 
Brick15: dc-vihi19:/gluster/bricks/brick8/data 
Brick16: dc-vihi71:/gluster/bricks/brick2/data 
Brick17: dc-vihi19:/gluster/bricks/brick9/data 
Brick18: dc-vihi70:/gluster/bricks/brick5/data 
Brick19: dc-vihi19:/gluster/bricks/brick10/data 
Brick20: dc-vihi71:/gluster/bricks/brick5/data 
Brick21: dc-vihi19:/gluster/bricks/brick11/data 
Brick22: dc-vihi70:/gluster/bricks/brick6/data 
Brick23: dc-vihi19:/gluster/bricks/brick12/data 
Brick24: dc-vihi71:/gluster/bricks/brick6/data 
Options Reconfigured: 
nfs.disable: on 
transport.address-family: inet 
features.ctr-enabled: on 
cluster.tier-mode: cache 
features.shard: on 
features.shard-block-size: 512MB 
network.ping-timeout: 5 
cluster.server-quorum-ratio: 51% 

[root at dc-vihi19 temp]# ls -lh 
total 26G 
-rw-rw-rw-. 1 root root 31G May 17 10:38 win7.qcow2 
[root at dc-vihi19 temp]# getfattr -n glusterfs.gfid.string win7.qcow2 
# file: win7.qcow2 
glusterfs.gfid.string="7f4a0fea-72c0-41e4-97a5-6297be0a9142" 

[root at dc-vihi19 temp]# rm win7.qcow2 
rm: remove regular file âwin7.qcow2â? y 

*Process hangs and can't be killed. A reboot later...* 

login as: root 
Authenticating with public key "rsa-key-20170510" 
Last login: Wed May 17 14:04:29 2017 from ****** 
[root at dc-vihi19 ~]# find /gluster/bricks -name 
"7f4a0fea-72c0-41e4-97a5-6297be0a9142*" 
/gluster/bricks/brick1/data/.shard/7f4a0fea-72c0-41e4-97a5-6297be0a9142.23 

/gluster/bricks/brick1/data/.shard/7f4a0fea-72c0-41e4-97a5-6297be0a9142.35 

/gluster/bricks/brick2/data/.shard/7f4a0fea-72c0-41e4-97a5-6297be0a9142.52 

/gluster/bricks/brick2/data/.shard/7f4a0fea-72c0-41e4-97a5-6297be0a9142.29 

/gluster/bricks/brick2/data/.shard/7f4a0fea-72c0-41e4-97a5-6297be0a9142.22 

/gluster/bricks/brick2/data/.shard/7f4a0fea-72c0-41e4-97a5-6297be0a9142.24 


and so on... 


-Walter Deignan
-Uline IT, Systems Architect
_______________________________________________
Gluster-users mailing list
Gluster-users at gluster.org
http://lists.gluster.org/mailman/listinfo/gluster-users


-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20170524/3b399ab0/attachment.html>


More information about the Gluster-users mailing list