[Gluster-users] Input/output error when trying to access a file on client
Alessandro Ipe
Alessandro.Ipe at meteo.be
Thu Mar 12 11:45:08 UTC 2015
Hi,
Actually, my gluster volume is distribute-replicate so I should provide the attributes on
all the bricks. Here they are:
1. # file: data/glusterfs/md1/brick1/root
trusted.afr.md1-client-0=0sAAAAAAAAAAAAAAAA
trusted.afr.md1-client-1=0sAAAAAAAAAAAAAAAA
trusted.gfid=0s3DmMvSq0QOyf7T1ZN2VPSw==
trusted.glusterfs.dht=0sAAAAAQAAAACqqqqq/////w==
2. # file: data/glusterfs/md1/brick1/root
trusted.afr.md1-client-0=0sAAAAAAAAAAAAAAAA
trusted.afr.md1-client-1=0sAAAAAAAAAAAAAAAA
trusted.gfid=0s3DmMvSq0QOyf7T1ZN2VPSw==
trusted.glusterfs.dht=0sAAAAAQAAAACqqqqq/////w==
3. # file: data/glusterfs/md1/brick1/root
trusted.afr.md1-client-2=0sAAAAAAAAAAAAAAAA
trusted.afr.md1-client-3=0sAAAAAAAAAAEAAAAA
trusted.gfid=0s3DmMvSq0QOyf7T1ZN2VPSw==
trusted.glusterfs.dht=0sAAAAAQAAAAAAAAAAVVVVVA==
4. # file: data/glusterfs/md1/brick1/root
trusted.afr.md1-client-2=0sAAAAAAAAAAEAAAAA
trusted.afr.md1-client-3=0sAAAAAAAAAAAAAAAA
trusted.gfid=0s3DmMvSq0QOyf7T1ZN2VPSw==
trusted.glusterfs.dht=0sAAAAAQAAAAAAAAAAVVVVVA==
5. # file: data/glusterfs/md1/brick1/root
trusted.afr.md1-client-4=0sAAAAAAAAAAAAAAAA
trusted.afr.md1-client-5=0sAAAAAAAAAAEAAAAA
trusted.gfid=0s3DmMvSq0QOyf7T1ZN2VPSw==
trusted.glusterfs.dht=0sAAAAAQAAAABVVVVVqqqqqQ==
6. # file: data/glusterfs/md1/brick1/root
trusted.afr.md1-client-4=0sAAAAAAAAAAEAAAAA
trusted.afr.md1-client-5=0sAAAAAAAAAAAAAAAA
trusted.gfid=0s3DmMvSq0QOyf7T1ZN2VPSw==
trusted.glusterfs.dht=0sAAAAAQAAAABVVVVVqqqqqQ==
so it seems in fact that there are discrepancies between 3-4 and 5-6 (replicate pairs).
A.
On Thursday 12 March 2015 11:33:00 Alessandro Ipe wrote:
Hi,
"gluster volume heal md1 info split-brain" returns approximatively 2000 files (already
divided by 2
due to replicate volume). So manually repairing each split-brain is unfeasable. Before
scripting some
procedure, I need to be sure that I will not harm further the gluster system.
Moreover, I noticed that the messages printed in the logs are all about directories,
e.g.
[2015-03-12 10:06:53.423856] E [afr-self-heal-
common.c:233:afr_sh_print_split_brain_log] 0-md1-replicate-1: Unable to self-heal
contents of '/root' (possible split-brain). Please delete the file from all but the preferred
subvolume.- Pending matrix: [ [ 0 1 ] [ 1 0 ] ]
[2015-03-12 10:06:53.424005] E [afr-self-heal-
common.c:233:afr_sh_print_split_brain_log] 0-md1-replicate-2: Unable to self-heal
contents of '/root' (possible split-brain). Please delete the file from all but the preferred
subvolume.- Pending matrix: [ [ 0 1 ] [ 1 0 ] ]
[2015-03-12 10:06:53.424110] E [afr-self-heal-
common.c:2868:afr_log_self_heal_completion_status] 0-md1-replicate-1: metadata
self heal failed, on /root
[2015-03-12 10:06:53.424290] E [afr-self-heal-
common.c:2868:afr_log_self_heal_completion_status] 0-md1-replicate-2: metadata
self heal failed, on /root
Getting the attributes of that directory on each brick gives me for the first
# file: data/glusterfs/md1/brick1/root
trusted.afr.md1-client-0=0sAAAAAAAAAAAAAAAA
trusted.afr.md1-client-1=0sAAAAAAAAAAAAAAAA
trusted.gfid=0s3DmMvSq0QOyf7T1ZN2VPSw==
trusted.glusterfs.dht=0sAAAAAQAAAACqqqqq/////w==
and for the second
# file: data/glusterfs/md1/brick1/root
trusted.afr.md1-client-0=0sAAAAAAAAAAAAAAAA
trusted.afr.md1-client-1=0sAAAAAAAAAAAAAAAA
trusted.gfid=0s3DmMvSq0QOyf7T1ZN2VPSw==
trusted.glusterfs.dht=0sAAAAAQAAAACqqqqq/////w==
so it seems that there are both rigorously identical. However, according to your split -
brain tutorial,
none of them has 0x000000000000000000000000. What 0sAAAAAAAAAAAAAAAA
means
in fact ?
Should I change both attributes on each directory to 0x000000000000000000000000
?
Many thanks,
A.
On Wednesday 11 March 2015 08:02:56 Krutika Dhananjay wrote:
Hi,
Have you gone
through https://github.com/gluster/glusterfs/blob/master/doc/debugging/split-brain.md[1] ?
If not, could you go through that once and try the steps given there? Do let us know if
something is not clear in the doc.
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20150312/1d6b300b/attachment.html>
More information about the Gluster-users
mailing list