[Gluster-users] heaps split-brains during back-transfert

Vijaikumar M vmallika at redhat.com
Tue Aug 4 03:54:14 UTC 2015


Adding Raghavendra.G for RDMA issue...


Hi Geoffrey,

Please find my comments in-line..

Thanks,
Vijay


On Monday 03 August 2015 09:15 PM, Geoffrey Letessier wrote:
> Hi Vijay,
>
> Yes of course, i sent my email after making some tests and checks and 
> the result was still wrong (even after a couple of hours/1day after 
> having forced the start of every bricks) … until i decided to do a 
> « du » on every quota path. Now, all seems to ~OK as you can read below:
> # gluster volume quota vol_home list
>                   Path     Hard-limit 
> Soft-limit   Used  Available  Soft-limit exceeded? Hard-limit exceeded?
> ---------------------------------------------------------------------------------------------------------------------------
> /simlab_team       5.0TB       80%       1.2TB   3.8TB              No 
>                 No
> /amyloid_team       7.0TB       80%       4.9TB   2.1TB            
>   No                 No
> /amyloid_team/nguyen       3.5TB       80%       2.0TB   1.5TB        
>       No                 No
> /sacquin_team       10.0TB       80%      55.3GB   9.9TB            
>   No                 No
> /baaden_team       20.0TB       80%      11.5TB   8.5TB            
>   No                 No
> /derreumaux_team       5.0TB       80%       2.2TB   2.8TB            
>   No                 No
> /sterpone_team       14.0TB       80%       9.3TB   4.7TB            
>   No                 No
> /admin_team       1.0TB       80%      15.8GB 1008.2GB              No 
>                 No
> # for path in $(gluster volume quota vol_home list|awk 'NR>2 {print 
> $1}'); do pdsh -w storage[1,3] "du -sh 
> /export/brick_home/brick{1,2}/data$path"; done
> storage1: 219G/export/brick_home/brick1/data/simlab_team
> storage3: 334G/export/brick_home/brick1/data/simlab_team
> storage1: 307G/export/brick_home/brick2/data/simlab_team
> storage3: 327G/export/brick_home/brick2/data/simlab_team
> storage1: 1,2T/export/brick_home/brick1/data/amyloid_team
> storage3: 1,2T/export/brick_home/brick1/data/amyloid_team
> storage1: 1,2T/export/brick_home/brick2/data/amyloid_team
> storage3: 1,2T/export/brick_home/brick2/data/amyloid_team
> storage1: 505G/export/brick_home/brick1/data/amyloid_team/nguyen
> storage1: 483G/export/brick_home/brick2/data/amyloid_team/nguyen
> storage3: 508G/export/brick_home/brick1/data/amyloid_team/nguyen
> storage3: 503G/export/brick_home/brick2/data/amyloid_team/nguyen
> storage3: 16G/export/brick_home/brick1/data/sacquin_team
> storage1: 14G/export/brick_home/brick1/data/sacquin_team
> storage3: 13G/export/brick_home/brick2/data/sacquin_team
> storage1: 13G/export/brick_home/brick2/data/sacquin_team
> storage1: 3,2T/export/brick_home/brick1/data/baaden_team
> storage1: 2,8T/export/brick_home/brick2/data/baaden_team
> storage3: 2,9T/export/brick_home/brick1/data/baaden_team
> storage3: 2,7T/export/brick_home/brick2/data/baaden_team
> storage3: 588G/export/brick_home/brick1/data/derreumaux_team
> storage1: 566G/export/brick_home/brick1/data/derreumaux_team
> storage1: 563G/export/brick_home/brick2/data/derreumaux_team
> storage3: 610G/export/brick_home/brick2/data/derreumaux_team
> storage3: 2,5T/export/brick_home/brick1/data/sterpone_team
> storage1: 2,7T/export/brick_home/brick1/data/sterpone_team
> storage3: 2,4T/export/brick_home/brick2/data/sterpone_team
> storage1: 2,4T/export/brick_home/brick2/data/sterpone_team
> storage3: 519M/export/brick_home/brick1/data/admin_team
> storage1: 11G/export/brick_home/brick1/data/admin_team
> storage3: 974M/export/brick_home/brick2/data/admin_team
> storage1: 4,0G/export/brick_home/brick2/data/admin_team
>
> In short:
> simlab_team: ~1.2TB
> amyloid_team: ~4.8TB
> amyloid_team/nguyen: ~2TB
> sacquin_team: ~56GB
> baaden_team: ~11.6TB
> derreumaux_team: 2.3TB
> sterpone_team: ~10TB
> admin_team: ~16.5GB
>
> There’s still some difference but it’s globally quite correct (except 
> for sterpone_team quota defined).
>
> But, I also noticed something strange: here are the result of every 
> « du » i did to force the « recompute » of the quota size (on the 
> glusterfs mount point):
> # du -sh /home/simlab_team/
> 1,2T    /home/simlab_team/
> # du -sh /home/amyloid_team/
> 4,7T    /home/amyloid_team/
> # du -sh /home/sacquin_team/
> 56G     /home/sacquin_team/
> # du -sh /home/baaden_team/
> 12T     /home/baaden_team/
> # du -sh /home/derreumaux_team/
> 2,3T    /home/derreumaux_team/
> # du -sh /home/sterpone_team/
> 9,9T    /home/sterpone_team/
>
> As you can above, I dont understand why the quota size computed by 
> quota daemon is different than a "du", especially concerning the quota 
> size of /sterpone_team
>
du command can round-off the values, could you check the values with 'du 
-sk'?



> Now, concerning all hangs i met, can you provide me the brand of your 
> infiniband interconnect? From my side, we use QLogic -maybe the 
> problem takes its origin here (Intel/Qlogic and Mellanox are quite 
> different).
>
>
> Concerning the brick logs, I just noticed I have a lot of error on one 
> of my brick logs and the file take around 5GB. Here is an extract:
> # tail -30l /var/log/glusterfs/bricks/export-brick_home-brick1-data.log
> [2015-08-03 15:32:37.408204] E [dict.c:1418:dict_copy_with_ref] 
> (-->/usr/lib64/glusterfs/3.7.3/xlator/protocol/server.so(server_resolve_inode+0x60) 
> [0x7f021c6f7410] 
> -->/usr/lib64/glusterfs/3.7.3/xlator/protocol/server.so(resolve_gfid+0x88) 
> [0x7f021c6f7188] 
> -->/usr/lib64/libglusterfs.so.0(dict_copy_with_ref+0xa4) 
> [0x7f0229cba674] ) 0-dict: invalid argument: dict [Argument invalide]
> [2015-08-03 15:32:37.410017] E [dict.c:1418:dict_copy_with_ref] 
> (-->/usr/lib64/glusterfs/3.7.3/xlator/protocol/server.so(server_resolve_inode+0x60) 
> [0x7f021c6f7410] 
> -->/usr/lib64/glusterfs/3.7.3/xlator/protocol/server.so(resolve_gfid+0x88) 
> [0x7f021c6f7188] 
> -->/usr/lib64/libglusterfs.so.0(dict_copy_with_ref+0xa4) 
> [0x7f0229cba674] ) 0-dict: invalid argument: dict [Argument invalide]
> [2015-08-03 15:32:37.410689] E [dict.c:1418:dict_copy_with_ref] 
> (-->/usr/lib64/glusterfs/3.7.3/xlator/protocol/server.so(server_resolve_inode+0x60) 
> [0x7f021c6f7410] 
> -->/usr/lib64/glusterfs/3.7.3/xlator/protocol/server.so(resolve_gfid+0x88) 
> [0x7f021c6f7188] 
> -->/usr/lib64/libglusterfs.so.0(dict_copy_with_ref+0xa4) 
> [0x7f0229cba674] ) 0-dict: invalid argument: dict [Argument invalide]
> [2015-08-03 15:32:37.410860] E [dict.c:1418:dict_copy_with_ref] 
> (-->/usr/lib64/glusterfs/3.7.3/xlator/protocol/server.so(server_resolve_inode+0x60) 
> [0x7f021c6f7410] 
> -->/usr/lib64/glusterfs/3.7.3/xlator/protocol/server.so(resolve_gfid+0x88) 
> [0x7f021c6f7188] 
> -->/usr/lib64/libglusterfs.so.0(dict_copy_with_ref+0xa4) 
> [0x7f0229cba674] ) 0-dict: invalid argument: dict [Argument invalide]
> [2015-08-03 15:32:37.412638] E [dict.c:1418:dict_copy_with_ref] 
> (-->/usr/lib64/glusterfs/3.7.3/xlator/protocol/server.so(server_resolve_inode+0x60) 
> [0x7f021c6f7410] 
> -->/usr/lib64/glusterfs/3.7.3/xlator/protocol/server.so(resolve_gfid+0x88) 
> [0x7f021c6f7188] 
> -->/usr/lib64/libglusterfs.so.0(dict_copy_with_ref+0xa4) 
> [0x7f0229cba674] ) 0-dict: invalid argument: dict [Argument invalide]
> [2015-08-03 15:32:37.413435] E [dict.c:1418:dict_copy_with_ref] 
> (-->/usr/lib64/glusterfs/3.7.3/xlator/protocol/server.so(server_resolve_inode+0x60) 
> [0x7f021c6f7410] 
> -->/usr/lib64/glusterfs/3.7.3/xlator/protocol/server.so(resolve_gfid+0x88) 
> [0x7f021c6f7188] 
> -->/usr/lib64/libglusterfs.so.0(dict_copy_with_ref+0xa4) 
> [0x7f0229cba674] ) 0-dict: invalid argument: dict [Argument invalide]
> [2015-08-03 15:32:37.413640] E [dict.c:1418:dict_copy_with_ref] 
> (-->/usr/lib64/glusterfs/3.7.3/xlator/protocol/server.so(server_resolve_inode+0x60) 
> [0x7f021c6f7410] 
> -->/usr/lib64/glusterfs/3.7.3/xlator/protocol/server.so(resolve_gfid+0x88) 
> [0x7f021c6f7188] 
> -->/usr/lib64/libglusterfs.so.0(dict_copy_with_ref+0xa4) 
> [0x7f0229cba674] ) 0-dict: invalid argument: dict [Argument invalide]
> [2015-08-03 15:32:37.415325] E [dict.c:1418:dict_copy_with_ref] 
> (-->/usr/lib64/glusterfs/3.7.3/xlator/protocol/server.so(server_resolve_inode+0x60) 
> [0x7f021c6f7410] 
> -->/usr/lib64/glusterfs/3.7.3/xlator/protocol/server.so(resolve_gfid+0x88) 
> [0x7f021c6f7188] 
> -->/usr/lib64/libglusterfs.so.0(dict_copy_with_ref+0xa4) 
> [0x7f0229cba674] ) 0-dict: invalid argument: dict [Argument invalide]
> [2015-08-03 15:32:37.416102] E [dict.c:1418:dict_copy_with_ref] 
> (-->/usr/lib64/glusterfs/3.7.3/xlator/protocol/server.so(server_resolve_inode+0x60) 
> [0x7f021c6f7410] 
> -->/usr/lib64/glusterfs/3.7.3/xlator/protocol/server.so(resolve_gfid+0x88) 
> [0x7f021c6f7188] 
> -->/usr/lib64/libglusterfs.so.0(dict_copy_with_ref+0xa4) 
> [0x7f0229cba674] ) 0-dict: invalid argument: dict [Argument invalide]
> [2015-08-03 15:32:37.416308] E [dict.c:1418:dict_copy_with_ref] 
> (-->/usr/lib64/glusterfs/3.7.3/xlator/protocol/server.so(server_resolve_inode+0x60) 
> [0x7f021c6f7410] 
> -->/usr/lib64/glusterfs/3.7.3/xlator/protocol/server.so(resolve_gfid+0x88) 
> [0x7f021c6f7188] 
> -->/usr/lib64/libglusterfs.so.0(dict_copy_with_ref+0xa4) 
> [0x7f0229cba674] ) 0-dict: invalid argument: dict [Argument invalide]
> [2015-08-03 15:32:37.418025] E [dict.c:1418:dict_copy_with_ref] 
> (-->/usr/lib64/glusterfs/3.7.3/xlator/protocol/server.so(server_resolve_inode+0x60) 
> [0x7f021c6f7410] 
> -->/usr/lib64/glusterfs/3.7.3/xlator/protocol/server.so(resolve_gfid+0x88) 
> [0x7f021c6f7188] 
> -->/usr/lib64/libglusterfs.so.0(dict_copy_with_ref+0xa4) 
> [0x7f0229cba674] ) 0-dict: invalid argument: dict [Argument invalide]
> [2015-08-03 15:32:37.418799] E [dict.c:1418:dict_copy_with_ref] 
> (-->/usr/lib64/glusterfs/3.7.3/xlator/protocol/server.so(server_resolve_inode+0x60) 
> [0x7f021c6f7410] 
> -->/usr/lib64/glusterfs/3.7.3/xlator/protocol/server.so(resolve_gfid+0x88) 
> [0x7f021c6f7188] 
> -->/usr/lib64/libglusterfs.so.0(dict_copy_with_ref+0xa4) 
> [0x7f0229cba674] ) 0-dict: invalid argument: dict [Argument invalide]
> [2015-08-03 15:32:37.419001] E [dict.c:1418:dict_copy_with_ref] 
> (-->/usr/lib64/glusterfs/3.7.3/xlator/protocol/server.so(server_resolve_inode+0x60) 
> [0x7f021c6f7410] 
> -->/usr/lib64/glusterfs/3.7.3/xlator/protocol/server.so(resolve_gfid+0x88) 
> [0x7f021c6f7188] 
> -->/usr/lib64/libglusterfs.so.0(dict_copy_with_ref+0xa4) 
> [0x7f0229cba674] ) 0-dict: invalid argument: dict [Argument invalide]
> [2015-08-03 15:32:37.420681] E [dict.c:1418:dict_copy_with_ref] 
> (-->/usr/lib64/glusterfs/3.7.3/xlator/protocol/server.so(server_resolve_inode+0x60) 
> [0x7f021c6f7410] 
> -->/usr/lib64/glusterfs/3.7.3/xlator/protocol/server.so(resolve_gfid+0x88) 
> [0x7f021c6f7188] 
> -->/usr/lib64/libglusterfs.so.0(dict_copy_with_ref+0xa4) 
> [0x7f0229cba674] ) 0-dict: invalid argument: dict [Argument invalide]
> [2015-08-03 15:32:37.421416] E [dict.c:1418:dict_copy_with_ref] 
> (-->/usr/lib64/glusterfs/3.7.3/xlator/protocol/server.so(server_resolve_inode+0x60) 
> [0x7f021c6f7410] 
> -->/usr/lib64/glusterfs/3.7.3/xlator/protocol/server.so(resolve_gfid+0x88) 
> [0x7f021c6f7188] 
> -->/usr/lib64/libglusterfs.so.0(dict_copy_with_ref+0xa4) 
> [0x7f0229cba674] ) 0-dict: invalid argument: dict [Argument invalide]
> [2015-08-03 15:32:37.421607] E [dict.c:1418:dict_copy_with_ref] 
> (-->/usr/lib64/glusterfs/3.7.3/xlator/protocol/server.so(server_resolve_inode+0x60) 
> [0x7f021c6f7410] 
> -->/usr/lib64/glusterfs/3.7.3/xlator/protocol/server.so(resolve_gfid+0x88) 
> [0x7f021c6f7188] 
> -->/usr/lib64/libglusterfs.so.0(dict_copy_with_ref+0xa4) 
> [0x7f0229cba674] ) 0-dict: invalid argument: dict [Argument invalide]
> [2015-08-03 15:32:37.423208] E [dict.c:1418:dict_copy_with_ref] 
> (-->/usr/lib64/glusterfs/3.7.3/xlator/protocol/server.so(server_resolve_inode+0x60) 
> [0x7f021c6f7410] 
> -->/usr/lib64/glusterfs/3.7.3/xlator/protocol/server.so(resolve_gfid+0x88) 
> [0x7f021c6f7188] 
> -->/usr/lib64/libglusterfs.so.0(dict_copy_with_ref+0xa4) 
> [0x7f0229cba674] ) 0-dict: invalid argument: dict [Argument invalide]
> [2015-08-03 15:32:37.423882] E [dict.c:1418:dict_copy_with_ref] 
> (-->/usr/lib64/glusterfs/3.7.3/xlator/protocol/server.so(server_resolve_inode+0x60) 
> [0x7f021c6f7410] 
> -->/usr/lib64/glusterfs/3.7.3/xlator/protocol/server.so(resolve_gfid+0x88) 
> [0x7f021c6f7188] 
> -->/usr/lib64/libglusterfs.so.0(dict_copy_with_ref+0xa4) 
> [0x7f0229cba674] ) 0-dict: invalid argument: dict [Argument invalide]
> [2015-08-03 15:32:37.424089] E [dict.c:1418:dict_copy_with_ref] 
> (-->/usr/lib64/glusterfs/3.7.3/xlator/protocol/server.so(server_resolve_inode+0x60) 
> [0x7f021c6f7410] 
> -->/usr/lib64/glusterfs/3.7.3/xlator/protocol/server.so(resolve_gfid+0x88) 
> [0x7f021c6f7188] 
> -->/usr/lib64/libglusterfs.so.0(dict_copy_with_ref+0xa4) 
> [0x7f0229cba674] ) 0-dict: invalid argument: dict [Argument invalide]
> [2015-08-03 15:32:37.425863] E [dict.c:1418:dict_copy_with_ref] 
> (-->/usr/lib64/glusterfs/3.7.3/xlator/protocol/server.so(server_resolve_inode+0x60) 
> [0x7f021c6f7410] 
> -->/usr/lib64/glusterfs/3.7.3/xlator/protocol/server.so(resolve_gfid+0x88) 
> [0x7f021c6f7188] 
> -->/usr/lib64/libglusterfs.so.0(dict_copy_with_ref+0xa4) 
> [0x7f0229cba674] ) 0-dict: invalid argument: dict [Argument invalide]
> [2015-08-03 15:32:37.426581] E [dict.c:1418:dict_copy_with_ref] 
> (-->/usr/lib64/glusterfs/3.7.3/xlator/protocol/server.so(server_resolve_inode+0x60) 
> [0x7f021c6f7410] 
> -->/usr/lib64/glusterfs/3.7.3/xlator/protocol/server.so(resolve_gfid+0x88) 
> [0x7f021c6f7188] 
> -->/usr/lib64/libglusterfs.so.0(dict_copy_with_ref+0xa4) 
> [0x7f0229cba674] ) 0-dict: invalid argument: dict [Argument invalide]
> [2015-08-03 15:32:37.426790] E [dict.c:1418:dict_copy_with_ref] 
> (-->/usr/lib64/glusterfs/3.7.3/xlator/protocol/server.so(server_resolve_inode+0x60) 
> [0x7f021c6f7410] 
> -->/usr/lib64/glusterfs/3.7.3/xlator/protocol/server.so(resolve_gfid+0x88) 
> [0x7f021c6f7188] 
> -->/usr/lib64/libglusterfs.so.0(dict_copy_with_ref+0xa4) 
> [0x7f0229cba674] ) 0-dict: invalid argument: dict [Argument invalide]
> [2015-08-03 15:32:37.428438] E [dict.c:1418:dict_copy_with_ref] 
> (-->/usr/lib64/glusterfs/3.7.3/xlator/protocol/server.so(server_resolve_inode+0x60) 
> [0x7f021c6f7410] 
> -->/usr/lib64/glusterfs/3.7.3/xlator/protocol/server.so(resolve_gfid+0x88) 
> [0x7f021c6f7188] 
> -->/usr/lib64/libglusterfs.so.0(dict_copy_with_ref+0xa4) 
> [0x7f0229cba674] ) 0-dict: invalid argument: dict [Argument invalide]
> [2015-08-03 15:32:37.429133] E [dict.c:1418:dict_copy_with_ref] 
> (-->/usr/lib64/glusterfs/3.7.3/xlator/protocol/server.so(server_resolve_inode+0x60) 
> [0x7f021c6f7410] 
> -->/usr/lib64/glusterfs/3.7.3/xlator/protocol/server.so(resolve_gfid+0x88) 
> [0x7f021c6f7188] 
> -->/usr/lib64/libglusterfs.so.0(dict_copy_with_ref+0xa4) 
> [0x7f0229cba674] ) 0-dict: invalid argument: dict [Argument invalide]
> [2015-08-03 15:32:37.429325] E [dict.c:1418:dict_copy_with_ref] 
> (-->/usr/lib64/glusterfs/3.7.3/xlator/protocol/server.so(server_resolve_inode+0x60) 
> [0x7f021c6f7410] 
> -->/usr/lib64/glusterfs/3.7.3/xlator/protocol/server.so(resolve_gfid+0x88) 
> [0x7f021c6f7188] 
> -->/usr/lib64/libglusterfs.so.0(dict_copy_with_ref+0xa4) 
> [0x7f0229cba674] ) 0-dict: invalid argument: dict [Argument invalide]
> The message "W [MSGID: 120003] [quota.c:759:quota_build_ancestry_cbk] 
> 0-vol_home-quota: parent is NULL [Argument invalide]" repeated 9016 
> times between [2015-08-03 15:31:55.379522] and [2015-08-03 
> 15:32:00.997113]
> [2015-08-03 15:32:37.442244] I [MSGID: 115036] 
> [server.c:545:server_rpc_notify] 0-vol_home-server: disconnecting 
> connection from lucifer.lbt.ibpc.fr 
> <http://lucifer.lbt.ibpc.fr>-21153-2015/08/03-15:31:23:33181-vol_home-client-0-0-0
> [2015-08-03 15:32:37.442286] I [MSGID: 101055] 
> [client_t.c:419:gf_client_unref] 0-vol_home-server: Shutting down 
> connection lucifer.lbt.ibpc.fr 
> <http://lucifer.lbt.ibpc.fr>-21153-2015/08/03-15:31:23:33181-vol_home-client-0-0-0
> The message "E [MSGID: 113104] 
> [posix-handle.c:154:posix_make_ancestryfromgfid] 0-vol_home-posix: 
> could not read the link from the gfid handle 
> /export/brick_home/brick1/data/.glusterfs/19/b6/19b67130-b409-4666-9237-2661241a8847 
> [Aucun fichier ou dossier de ce type]" repeated 755 times between 
> [2015-08-03 15:31:25.553801] and [2015-08-03 15:31:43.528305]
> The message "E [MSGID: 113104] 
> [posix-handle.c:154:posix_make_ancestryfromgfid] 0-vol_home-posix: 
> could not read the link from the gfid handle 
> /export/brick_home/brick1/data/.glusterfs/81/5a/815acde3-7f47-410b-9131-e8d75c71a5bd 
> [Aucun fichier ou dossier de ce type]" repeated 8147 times between 
> [2015-08-03 15:31:25.521255] and [2015-08-03 15:31:53.593932]
> Do you have an idea where this issue come from and what I have to do 
> to fix it?
We will investigate on this issue and update you soon on the same.




>
> # grep -rc "\] E \[" 
> /var/log/glusterfs/bricks/export-brick_home-brick{1,2}-data.log
> /var/log/glusterfs/bricks/export-brick_home-brick1-data.log:11038933
> /var/log/glusterfs/bricks/export-brick_home-brick2-data.log:243
>
> FYI I updated GlusterFS to the latest version (v3.7.3) 2 days ago.
>
> Thanks in advance for the next answers. and thanks for all your help 
> (all the support team).
> Best,
> Geoffrey
>
> ------------------------------------------------------
> Geoffrey Letessier
> Responsable informatique & ingénieur système
> UPR 9080 - CNRS - Laboratoire de Biochimie Théorique
> Institut de Biologie Physico-Chimique
> 13, rue Pierre et Marie Curie - 75005 Paris
> Tel: 01 58 41 50 93 - eMail: geoffrey.letessier at ibpc.fr 
> <mailto:geoffrey.letessier at ibpc.fr>
>
> Le 3 août 2015 à 08:51, Vijaikumar M <vmallika at redhat.com 
> <mailto:vmallika at redhat.com>> a écrit :
>
>> Hi Geoffrey,
>>
>> Please find my comments in-line.
>>
>>
>> On Saturday 01 August 2015 04:10 AM, Geoffrey Letessier wrote:
>>> Hello,
>>>
>>> As Krutika said, I resolved with success all split-brains (more than 
>>> 3450) appeared after the first data transfert from one backup server 
>>> to my new and fresh volume but…
>>>
>>> The following step to validate my new volume was to enable the quota 
>>> on it; and now, more than one day after this activation, all the 
>>> results are still completely wrong:
>>> Example:
>>> # df -h /home/sterpone_team
>>> Filesystem            Size Used Avail Use% Mounted on
>>> ib-storage1:vol_home.tcp
>>>                        14T 3,3T   11T  24% /home
>>> # pdsh -w storage[1,3] du -sh 
>>> /export/brick_home/brick{1,2}/data/sterpone_team
>>> storage3: 2,5T/export/brick_home/brick1/data/sterpone_team
>>> storage3: 2,4T/export/brick_home/brick2/data/sterpone_team
>>> storage1: 2,7T/export/brick_home/brick1/data/sterpone_team
>>> storage1: 2,4T/export/brick_home/brick2/data/sterpone_team
>>> As you can read, all data for this account is around 10TB and quota 
>>> displays only 3.3TB used.
>>>
>>> Worse:
>>> # pdsh -w storage[1,3] du -sh 
>>> /export/brick_home/brick{1,2}/data/baaden_team
>>> storage3: 2,9T/export/brick_home/brick1/data/baaden_team
>>> storage3: 2,7T/export/brick_home/brick2/data/baaden_team
>>> storage1: 3,2T/export/brick_home/brick1/data/baaden_team
>>> storage1: 2,8T/export/brick_home/brick2/data/baaden_team
>>> # df -h /home/baaden_team/
>>> Filesystem            Size Used Avail Use% Mounted on
>>> ib-storage1:vol_home.tcp
>>>                        20T 786G   20T   4% /home
>>> # gluster volume quota vol_home list /baaden_team
>>>                   Path             Hard-limit Soft-limit   Used  
>>> Available Soft-limit exceeded? Hard-limit exceeded?
>>> ---------------------------------------------------------------------------------------------------------------------------
>>> /baaden_team               20.0TB       80%     785.6GB  19.2TB     
>>>         No                   No
>>> This account is around 11.6TB and quota detects only 786GB used…
>>>
>> As you mentioned below, some of the bricks were down. 'quota list' 
>> will only show the aggregated value of online bricks, Could you 
>> please check the 'quota list' when all the bricks are up and running?
>> I suspect quota initiate might not have completed because of brick down.
>>
>>> Can someone help me to fix it -knowing if I've previously updated 
>>> GlusterFS from 3.5.3 to 3.7.2 it was exactly to solve a similar 
>>> trouble…
>>>
>>> For information, in quotad log file:
>>> [2015-07-31 22:13:00.574361] I [MSGID: 114047] 
>>> [client-handshake.c:1225:client_setvolume_cbk] 0-vol_home-client-7: 
>>> Server and Client lk-version numbers are not same, reopening the fds
>>> [2015-07-31 22:13:00.574507] I [MSGID: 114035] 
>>> [client-handshake.c:193:client_set_lk_version_cbk] 
>>> 0-vol_home-client-7: Server lk version = 1
>>>
>>> is there any causal connection (client/server version conflict)?
>>>
>>> Here what i noticed on 
>>> my /var/log/glusterfs/quota-mount-vol_home.log file:
>>> … <same kind of lines>
>>> [2015-07-31 21:26:15.247269] I [rpc-clnt.c:1819:rpc_clnt_reconfig] 
>>> 0-vol_home-client-5: changing port to 49162 (from 0)
>>> [2015-07-31 21:26:15.250272] E [socket.c:2332:socket_connect_finish] 
>>> 0-vol_home-client-5: connection to 10.0.4.2:49162 failed (Connexion 
>>> refusée)
>>> [2015-07-31 21:26:19.250545] I [rpc-clnt.c:1819:rpc_clnt_reconfig] 
>>> 0-vol_home-client-5: changing port to 49162 (from 0)
>>> [2015-07-31 21:26:19.253643] E [socket.c:2332:socket_connect_finish] 
>>> 0-vol_home-client-5: connection to 10.0.4.2:49162 failed (Connexion 
>>> refusée)
>>> … <same kind of lines>
>>>
>> Connection refused is because brick is down.
>>
>>> <A few minutes after:> OK, this was due to one brick which was down. 
>>> It’s strange: since I have updated GlusteFS to 3.7.x I notice a lot 
>>> of bricks which go down, sometimes a few moment after starting the 
>>> volume, sometime after a couple of days/weeks… What never happened 
>>> with GlusterFS version 3.3.1 and 3.5.3.
>>>
>> Could please provide brick log? We will check the log on this issue, 
>> once this issue is fixed, we can initiate quota healing again.
>>
>>
>>> Now, I need to stop-and-start the volume because I notice again some 
>>> hangs with "gluster volume quota … ", "df", etc. One more time, i’ve 
>>> never noticed this kind of hangs with previous versions of GlusterFS 
>>> I used; is it "expected"?
>>
>> From you previous mail we tried re-creating hang problem, however it 
>> was not re-creating.
>>
>>
>>
>>> One more time: thank you very much by advance.
>>> Geoffrey
>>>
>>> ------------------------------------------------------
>>> Geoffrey Letessier
>>> Responsable informatique & ingénieur système
>>> UPR 9080 - CNRS - Laboratoire de Biochimie Théorique
>>> Institut de Biologie Physico-Chimique
>>> 13, rue Pierre et Marie Curie - 75005 Paris
>>> Tel: 01 58 41 50 93 - eMail: geoffrey.letessier at ibpc.fr 
>>> <mailto:geoffrey.letessier at ibpc.fr>
>>>
>>> Le 31 juil. 2015 à 11:26, Niels de Vos <ndevos at redhat.com 
>>> <mailto:ndevos at redhat.com>> a écrit :
>>>
>>>> On Wed, Jul 29, 2015 at 12:44:38AM +0200, Geoffrey Letessier wrote:
>>>>> OK, thank you Niels for this explanation. Now, this makes sense.
>>>>>
>>>>> And concerning all split-brains appeared during the 
>>>>> back-transfert, do you have an idea where is this coming from?
>>>>
>>>> Sorry, no, I dont know how that is happening in your environment. I'll
>>>> try to find someone that understands more about it and can help you 
>>>> with
>>>> that.
>>>>
>>>> Niels
>>>>
>>>>>
>>>>> Best,
>>>>> Geoffrey
>>>>> ------------------------------------------------------
>>>>> Geoffrey Letessier
>>>>> Responsable informatique & ingénieur système
>>>>> UPR 9080 - CNRS - Laboratoire de Biochimie Théorique
>>>>> Institut de Biologie Physico-Chimique
>>>>> 13, rue Pierre et Marie Curie - 75005 Paris
>>>>> Tel: 01 58 41 50 93 - eMail: geoffrey.letessier at ibpc.fr 
>>>>> <mailto:geoffrey.letessier at ibpc.fr>
>>>>>
>>>>> Le 29 juil. 2015 à 00:02, Niels de Vos <ndevos at redhat.com 
>>>>> <mailto:ndevos at redhat.com>> a écrit :
>>>>>
>>>>>> On Tue, Jul 28, 2015 at 03:46:37PM +0200, Geoffrey Letessier wrote:
>>>>>>> Hi,
>>>>>>>
>>>>>>> In addition of all split brains reported, is it normal to notice
>>>>>>> thousands and thousands (several tens nay hundreds of thousands)
>>>>>>> broken symlinks browsing the .glusterfs directory on each brick?
>>>>>>
>>>>>> Yes, I think it is normal. A symlink points to a particular filename,
>>>>>> possibly in a different directory. If the target file is located on a
>>>>>> different brick, the symlink points to a non-local file.
>>>>>>
>>>>>> Consider this example with two bricks in a distributed volume:
>>>>>> - file: README
>>>>>> - symlink: IMPORTANT -> README
>>>>>>
>>>>>> When the distribution algorithm is done, README 'hashes' to 
>>>>>> brick-A. The
>>>>>> symlink 'hashes' to brick-B. This means that README will be 
>>>>>> localed on
>>>>>> brick-A, and the symlink with name IMPORTANT would be located on
>>>>>> brick-B. Because README is not on the same brick as IMPORTANT, the
>>>>>> symlink points to the non-existing file README on brick-B.
>>>>>>
>>>>>> However, when a Gluster client reads the target of symlink IMPORTANT,
>>>>>> the Gluster client calculate the location of README and will know 
>>>>>> that
>>>>>> README can be found on brick-A.
>>>>>>
>>>>>> I hope that makes sense?
>>>>>>
>>>>>> Niels
>>>>>>
>>>>>>
>>>>>>> For the moment, i just synchronized one remote directory (around 
>>>>>>> 30TB
>>>>>>> and a few million files) into my new volume. No other operations on
>>>>>>> files on this volume has yet been done.
>>>>>>> How can I fix it? Can I delete these dead-symlinks? How can I 
>>>>>>> fix all
>>>>>>> my split-brains?
>>>>>>>
>>>>>>> Here is an example of a ls:
>>>>>>> [root at cl-storage3 ~]# cd 
>>>>>>> /export/brick_home/brick1/data/.glusterfs/7b/d2/
>>>>>>> [root at cl-storage3 d2]# ll
>>>>>>> total 8,7M
>>>>>>>    13706 drwx------   2 root      root            8,0K 26 juil. 
>>>>>>> 17:22 .
>>>>>>> 2147483784 drwx------ 258 root      root            8,0K 20 
>>>>>>> juil. 23:07 ..
>>>>>>> 2148444137 -rwxrwxrwx   2 baaden    baaden_team     173K 22 mai 
>>>>>>>    2008 7bd200dd-1774-4395-9065-605ae30ec18b
>>>>>>>  1559384 -rw-rw-r--   2 tarus     amyloid_team    4,3K 19 juin 
>>>>>>>   2013 7bd2155c-7a05-4edc-ae77-35ed7e16afbc
>>>>>>>   287295 lrwxrwxrwx   1 root      root              58 20 juil. 
>>>>>>> 23:38 7bd2370a-100b-411e-89a4-d184da9f0f88 -> 
>>>>>>> ../../a7/59/a759de6f-cdf5-43dd-809a-baf81d103bf7/prop-base
>>>>>>> 2149090201 -rw-rw-r--   2 tarus     amyloid_team     76K  8 mars 
>>>>>>>   2014 7bd2497f-d24b-4b19-a1c5-80a4956e56a1
>>>>>>> 2148561174 -rw-r--r--   2 tran      derreumaux_team  575 14 
>>>>>>> févr. 07:54 7bd25db0-67f5-43e5-a56a-52cf8c4c60dd
>>>>>>>  1303943 -rw-r--r--   2 tran      derreumaux_team  576 10 févr. 
>>>>>>> 06:06 7bd25e97-18be-4faf-b122-5868582b4fd8
>>>>>>>  1308607 -rw-r--r--   2 tran      derreumaux_team 414K 16 juin 
>>>>>>>  11:05 7bd2618f-950a-4365-a753-723597ef29f5
>>>>>>>    45745 -rw-r--r--   2 letessier admin_team       585  5 janv. 
>>>>>>>  2012 7bd265c7-e204-4ee8-8717-e4a0c393fb0f
>>>>>>> 2148144918 -rw-rw-r--   2 tarus     amyloid_team    107K 28 
>>>>>>> févr.  2014 7bd26c5b-d48a-481a-9ca6-2dc27768b5ad
>>>>>>>    13705 -rw-rw-r--   2 tarus     amyloid_team     25K  4 juin 
>>>>>>>   2014 7bd27e4c-46ba-4f21-a766-389bfa52fd78
>>>>>>>  1633627 -rw-rw-r--   2 tarus     amyloid_team     75K 12 mars 
>>>>>>>   2014 7bd28631-90af-4c16-8ff0-c3d46d5026c6
>>>>>>>  1329165 -rw-r--r--   2 tran      derreumaux_team  175 15 juin 
>>>>>>>  23:40 7bd2957e-a239-4110-b3d8-b4926c7f060b
>>>>>>>   797803 lrwxrwxrwx   2 baaden    baaden_team       26  2 avril 
>>>>>>>  2007 7bd29933-1c80-4c6b-ae48-e64e4da874cb -> 
>>>>>>> ../divided/a7/2a7o.pdb1.gz
>>>>>>>  1532463 -rw-rw-rw-   2 baaden    baaden_team     1,8M  2 nov. 
>>>>>>>   2009 7bd29d70-aeb4-4eca-ac55-fae2d46ba911
>>>>>>>  1411112 -rw-r--r--   2 sterpone  sterpone_team   3,1K  2 mai 
>>>>>>>    2012 7bd2a5eb-62a4-47fc-b149-31e10bd3c33d
>>>>>>> 2148865896 -rw-r--r--   2 tran      derreumaux_team 2,1M 15 juin 
>>>>>>>  23:46 7bd2ae9c-18ca-471f-a54a-6e4aec5aea89
>>>>>>> 2148762578 -rw-rw-r--   2 tarus     amyloid_team    154K 11 mars 
>>>>>>>   2014 7bd2b7d7-7745-4842-b7b4-400791c1d149
>>>>>>>   149216 -rw-r--r--   2 vamparys  sacquin_team    241K 17 mai 
>>>>>>>    2013 7bd2ba98-6a42-40ea-87ea-acb607d73cb5
>>>>>>> 2148977923 -rwxr-xr-x   2 murail    baaden_team      23K 18 juin 
>>>>>>>   2012 7bd2cf57-19e7-451c-885d-fd02fd988d43
>>>>>>>  1176623 -rw-rw-r--   2 tarus     amyloid_team    227K  8 mars 
>>>>>>>   2014 7bd2d92c-7ec8-4af8-9043-49d1908a99dc
>>>>>>>  1172122 lrwxrwxrwx   2 sterpone  sterpone_team     61 17 avril 
>>>>>>> 12:49 7bd2d96e-e925-45f0-a26a-56b95c084122 -> 
>>>>>>> ../../../../../src/libs/ck-libs/ParFUM-Tops-Dev/ParFUM_TOPS.h
>>>>>>>  1385933 -rw-r--r--   2 tran      derreumaux_team 2,9M 16 juin 
>>>>>>>  05:29 7bd2df54-17d2-4644-96b7-f8925a67ec1e
>>>>>>>   745899 lrwxrwxrwx   1 root      root              58 22 juil. 
>>>>>>> 09:50 7bd2df83-ce58-4a17-aca8-a32b71e953d4 -> 
>>>>>>> ../../5c/39/5c39010f-fa77-49df-8df6-8d72cf74fd64/model_009
>>>>>>> 2149100186 -rw-rw-r--   2 tarus     amyloid_team    494K 17 mars 
>>>>>>>   2014 7bd2e865-a2f4-4d90-ab29-dccebe2e3440
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> Best.
>>>>>>> Geoffrey
>>>>>>> ------------------------------------------------------
>>>>>>> Geoffrey Letessier
>>>>>>> Responsable informatique & ingénieur système
>>>>>>> UPR 9080 - CNRS - Laboratoire de Biochimie Théorique
>>>>>>> Institut de Biologie Physico-Chimique
>>>>>>> 13, rue Pierre et Marie Curie - 75005 Paris
>>>>>>> Tel: 01 58 41 50 93 - eMail: geoffrey.letessier at ibpc.fr 
>>>>>>> <mailto:geoffrey.letessier at ibpc.fr>
>>>>>>>
>>>>>>> Le 27 juil. 2015 à 22:57, Geoffrey Letessier 
>>>>>>> <geoffrey.letessier at cnrs.fr <mailto:geoffrey.letessier at cnrs.fr>> 
>>>>>>> a écrit :
>>>>>>>
>>>>>>>> Dears,
>>>>>>>>
>>>>>>>> For a couple of weeks (more than one month), our computing 
>>>>>>>> production is stopped due to several -but amazing- troubles 
>>>>>>>> with GlusterFS.
>>>>>>>>
>>>>>>>> After having noticed a big problem with incorrect quota size 
>>>>>>>> accounted for many many files, i decided under the guidance of 
>>>>>>>> Gluster team support to upgrade my storage cluster from version 
>>>>>>>> 3.5.3 to the latest (3.7.2-3) because these bugs are 
>>>>>>>> theoretically fixed in this branch. Now, since i’ve done this 
>>>>>>>> upgrade, it’s the amazing mess and i cannot restart the production.
>>>>>>>> Indeed :
>>>>>>>> 1 - RDMA protocol is not working and hang my system / shell 
>>>>>>>> commands; only TCP protocol (over Infiniband) is more or less 
>>>>>>>> operational   - it’s not a blocking point but…
>>>>>>>> 2 - read/write performance relatively low
>>>>>>>> 3 - thousands split-brains are appeared.
>>>>>>>>
>>>>>>>> So, for the moment, i believe GlusterFS 3.7 is not actually 
>>>>>>>> production ready.
>>>>>>>>
>>>>>>>> Concerning the third point: after having destroy all my volumes 
>>>>>>>> (RAID re-init, new partition, GlusterFS volumes, etc.), 
>>>>>>>> recreate the main one, I tried to back-transfert my data from 
>>>>>>>> archive/backup server info this new volume and I note a lot of 
>>>>>>>> errors in my mount log file, as your can read in this extract:
>>>>>>>> [2015-07-26 22:35:16.962815] I 
>>>>>>>> [afr-self-heal-entry.c:565:afr_selfheal_entry_do] 
>>>>>>>> 0-vol_home-replicate-0: performing entry selfheal on 
>>>>>>>> 865083fa-984e-44bd-aacf-b8195789d9e0
>>>>>>>> [2015-07-26 22:35:16.965896] E 
>>>>>>>> [afr-self-heal-entry.c:249:afr_selfheal_detect_gfid_and_type_mismatch] 
>>>>>>>> 0-vol_home-replicate-0: Gfid mismatch detected for 
>>>>>>>> <865083fa-984e-44bd-aacf-b8195789d9e0/job.pbs>, 
>>>>>>>> e944d444-66c5-40a4-9603-7c190ad86013 on vol_home-client-1 and 
>>>>>>>> 820f9bcc-a0f6-40e0-bcec-28a76b4195ea on vol_home-client-0. 
>>>>>>>> Skipping conservative merge on the file.
>>>>>>>> [2015-07-26 22:35:16.975206] I 
>>>>>>>> [afr-self-heal-entry.c:565:afr_selfheal_entry_do] 
>>>>>>>> 0-vol_home-replicate-0: performing entry selfheal on 
>>>>>>>> 29382d8d-c507-4d2e-b74d-dbdcb791ca65
>>>>>>>> [2015-07-26 22:35:28.719935] E 
>>>>>>>> [afr-self-heal-entry.c:249:afr_selfheal_detect_gfid_and_type_mismatch] 
>>>>>>>> 0-vol_home-replicate-0: Gfid mismatch detected for 
>>>>>>>> <29382d8d-c507-4d2e-b74d-dbdcb791ca65/res_1BVK_r_u_1IBR_l_u_Cond.1IBR_l_u.1BVK_r_u.UB.global.dat.txt>, 
>>>>>>>> 951c5ffb-ca38-4630-93f3-8e4119ab0bd8 on vol_home-client-1 and 
>>>>>>>> 5ae663ca-e896-4b92-8ec5-5b15422ab861 on vol_home-client-0. 
>>>>>>>> Skipping conservative merge on the file.
>>>>>>>> [2015-07-26 22:35:29.764891] I 
>>>>>>>> [afr-self-heal-entry.c:565:afr_selfheal_entry_do] 
>>>>>>>> 0-vol_home-replicate-0: performing entry selfheal on 
>>>>>>>> 865083fa-984e-44bd-aacf-b8195789d9e0
>>>>>>>> [2015-07-26 22:35:29.768339] E 
>>>>>>>> [afr-self-heal-entry.c:249:afr_selfheal_detect_gfid_and_type_mismatch] 
>>>>>>>> 0-vol_home-replicate-0: Gfid mismatch detected for 
>>>>>>>> <865083fa-984e-44bd-aacf-b8195789d9e0/job.pbs>, 
>>>>>>>> e944d444-66c5-40a4-9603-7c190ad86013 on vol_home-client-1 and 
>>>>>>>> 820f9bcc-a0f6-40e0-bcec-28a76b4195ea on vol_home-client-0. 
>>>>>>>> Skipping conservative merge on the file.
>>>>>>>> [2015-07-26 22:35:29.775037] I 
>>>>>>>> [afr-self-heal-entry.c:565:afr_selfheal_entry_do] 
>>>>>>>> 0-vol_home-replicate-0: performing entry selfheal on 
>>>>>>>> 29382d8d-c507-4d2e-b74d-dbdcb791ca65
>>>>>>>> [2015-07-26 22:35:29.776857] E 
>>>>>>>> [afr-self-heal-entry.c:249:afr_selfheal_detect_gfid_and_type_mismatch] 
>>>>>>>> 0-vol_home-replicate-0: Gfid mismatch detected for 
>>>>>>>> <29382d8d-c507-4d2e-b74d-dbdcb791ca65/res_1BVK_r_u_1IBR_l_u_Cond.1IBR_l_u.1BVK_r_u.UB.global.dat.txt>, 
>>>>>>>> 951c5ffb-ca38-4630-93f3-8e4119ab0bd8 on vol_home-client-1 and 
>>>>>>>> 5ae663ca-e896-4b92-8ec5-5b15422ab861 on vol_home-client-0. 
>>>>>>>> Skipping conservative merge on the file.
>>>>>>>> [2015-07-26 22:35:29.800535] W [MSGID: 108008] 
>>>>>>>> [afr-self-heal-name.c:353:afr_selfheal_name_gfid_mismatch_check] 0-vol_home-replicate-0: 
>>>>>>>> GFID mismatch for 
>>>>>>>> <gfid:29382d8d-c507-4d2e-b74d-dbdcb791ca65>/res_1BVK_r_u_1IBR_l_u_Cond.1IBR_l_u.1BVK_r_u.UB.global.dat.txt 
>>>>>>>> 951c5ffb-ca38-4630-93f3-8e4119ab0bd8 on vol_home-client-1 and 
>>>>>>>> 5ae663ca-e896-4b92-8ec5-5b15422ab861 on vol_home-client-0
>>>>>>>>
>>>>>>>> And when I try to browse some folders (still in mount log file):
>>>>>>>> [2015-07-27 09:00:19.005763] I 
>>>>>>>> [afr-self-heal-entry.c:565:afr_selfheal_entry_do] 
>>>>>>>> 0-vol_home-replicate-0: performing entry selfheal on 
>>>>>>>> 2ac27442-8be0-4985-b48f-3328a86a6686
>>>>>>>> [2015-07-27 09:00:22.322316] E 
>>>>>>>> [afr-self-heal-entry.c:249:afr_selfheal_detect_gfid_and_type_mismatch] 
>>>>>>>> 0-vol_home-replicate-0: Gfid mismatch detected for 
>>>>>>>> <2ac27442-8be0-4985-b48f-3328a86a6686/md0012588.gro>, 
>>>>>>>> 9c635868-054b-4a13-b974-0ba562991586 on vol_home-client-1 and 
>>>>>>>> 1943175c-b336-4b33-aa1c-74a1c51f17b9 on vol_home-client-0. 
>>>>>>>> Skipping conservative merge on the file.
>>>>>>>> [2015-07-27 09:00:23.008771] I 
>>>>>>>> [afr-self-heal-entry.c:565:afr_selfheal_entry_do] 
>>>>>>>> 0-vol_home-replicate-0: performing entry selfheal on 
>>>>>>>> 2ac27442-8be0-4985-b48f-3328a86a6686
>>>>>>>> [2015-07-27 08:59:50.359187] W [MSGID: 108008] 
>>>>>>>> [afr-self-heal-name.c:353:afr_selfheal_name_gfid_mismatch_check] 0-vol_home-replicate-0: 
>>>>>>>> GFID mismatch for 
>>>>>>>> <gfid:2ac27442-8be0-4985-b48f-3328a86a6686>/md0012588.gro 
>>>>>>>> 9c635868-054b-4a13-b974-0ba562991586 on vol_home-client-1 and 
>>>>>>>> 1943175c-b336-4b33-aa1c-74a1c51f17b9 on vol_home-client-0
>>>>>>>> [2015-07-27 09:00:02.500419] W [MSGID: 108008] 
>>>>>>>> [afr-self-heal-name.c:353:afr_selfheal_name_gfid_mismatch_check] 0-vol_home-replicate-0: 
>>>>>>>> GFID mismatch for 
>>>>>>>> <gfid:2ac27442-8be0-4985-b48f-3328a86a6686>/md0012590.gro 
>>>>>>>> b22aec09-2be3-41ea-a976-7b8d0e6f61f0 on vol_home-client-1 and 
>>>>>>>> ec100f9e-ec48-4b29-b75e-a50ec6245de6 on vol_home-client-0
>>>>>>>> [2015-07-27 09:00:02.506925] W [MSGID: 108008] 
>>>>>>>> [afr-self-heal-name.c:353:afr_selfheal_name_gfid_mismatch_check] 0-vol_home-replicate-0: 
>>>>>>>> GFID mismatch for 
>>>>>>>> <gfid:2ac27442-8be0-4985-b48f-3328a86a6686>/md0009059.gro 
>>>>>>>> 0485c093-11ca-4829-b705-e259668ebd8c on vol_home-client-1 and 
>>>>>>>> e83a492b-7f8c-4b32-a76e-343f984142fe on vol_home-client-0
>>>>>>>> [2015-07-27 09:00:23.001121] W [MSGID: 108008] 
>>>>>>>> [afr-read-txn.c:241:afr_read_txn] 0-vol_home-replicate-0: 
>>>>>>>> Unreadable subvolume -1 found with event generation 2. 
>>>>>>>> (Possible split-brain)
>>>>>>>> [2015-07-27 09:00:26.231262] E 
>>>>>>>> [afr-self-heal-entry.c:249:afr_selfheal_detect_gfid_and_type_mismatch] 
>>>>>>>> 0-vol_home-replicate-0: Gfid mismatch detected for 
>>>>>>>> <2ac27442-8be0-4985-b48f-3328a86a6686/md0012588.gro>, 
>>>>>>>> 9c635868-054b-4a13-b974-0ba562991586 on vol_home-client-1 and 
>>>>>>>> 1943175c-b336-4b33-aa1c-74a1c51f17b9 on vol_home-client-0. 
>>>>>>>> Skipping conservative merge on the file.
>>>>>>>>
>>>>>>>> And, above all, browsing folder I get a lot of input/ouput errors.
>>>>>>>>
>>>>>>>> Currently I have 6.2M inodes and roughly 30TB in my "new" volume.
>>>>>>>>
>>>>>>>> For the moment, Quota is disable to increase the IO performance 
>>>>>>>> during the back-transfert…
>>>>>>>>
>>>>>>>> Your can also find in attachments:
>>>>>>>> - an "ls" result
>>>>>>>> - a split-brain research result
>>>>>>>> - the volume information and status
>>>>>>>> - a complete volume heal info
>>>>>>>>
>>>>>>>> Hoping this can help your to help me to fix all my problems and 
>>>>>>>> reopen the computing production.
>>>>>>>>
>>>>>>>> Thanks in advance,
>>>>>>>> Geoffrey
>>>>>>>>
>>>>>>>> PS: « Erreur d’Entrée/Sortie » = « Input / Output Error »
>>>>>>>> ------------------------------------------------------
>>>>>>>> Geoffrey Letessier
>>>>>>>> Responsable informatique & ingénieur système
>>>>>>>> UPR 9080 - CNRS - Laboratoire de Biochimie Théorique
>>>>>>>> Institut de Biologie Physico-Chimique
>>>>>>>> 13, rue Pierre et Marie Curie - 75005 Paris
>>>>>>>> Tel: 01 58 41 50 93 - eMail: geoffrey.letessier at ibpc.fr 
>>>>>>>> <mailto:geoffrey.letessier at ibpc.fr>
>>>>>>>>
>>>>>>>> <ls_example.txt>
>>>>>>>> <split_brain__20150725.txt>
>>>>>>>> <vol_home_healinfo.txt>
>>>>>>>> <vol_home_info.txt>
>>>>>>>> <vol_home_status.txt>
>>>>>>>
>>>>>
>>>
>>
>

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20150804/b446d306/attachment.html>


More information about the Gluster-users mailing list