[Gluster-users] Input/output error when trying to access a file on client
Alessandro Ipe
Alessandro.Ipe at meteo.be
Thu Mar 12 12:12:21 UTC 2015
Hi,
Doing splitmount localhost md1 . and ls -l gives me
total 8
drwxr-xr-x 12 root root 426 Jan 19 11:04 r1
drwxr-xr-x 12 root root 426 Jan 19 11:04 r2
-rw------- 1 root root 2840 Mar 12 12:08 tmp7TLytQ
-rw------- 1 root root 2840 Mar 12 12:08 tmptI3gv_
Doing ls -l r1/root/bash_cmd/ gives me
total 5
-rwxr-xr-x 1 root root 212 Nov 21 17:50 ira
-rwxr-xr-x 1 root root 2311 Nov 21 17:50 listing
drwxr-xr-x 2 root root 52 Jan 19 11:24 mbl
-rwxr-xr-x 1 root root 1210 Nov 21 17:50 viewhdf
while doing ls -l r1/root/bash_cmd/mbl/ gives me
ls: cannot access r1/root/bash_cmd/mbl/mbl.c: Software caused connection abort
ls: reading directory r1/root/bash_cmd/mbl/: Transport endpoint is not connected
total 0
?????????? ? ? ? ? ? mbl.c
A.
On Wednesday 11 March 2015 07:52:11 Joe Julian wrote:
http://joejulian.name/blog/glusterfs-split-brain-recovery-made-easy/[1]
On March 11, 2015 4:24:09 AM PDT, Alessandro Ipe <Alessandro.Ipe at meteo.be>
wrote:
Well, it is even worse. Now when doing a "ls -R" on the volume results in a lot of
[2015-03-11 11:18:31.957505] E [afr-self-heal-
common.c:233:afr_sh_print_split_brain_log] 0-md1-replicate-2: Unable to self-heal
contents of '/library' (possible split-brain). Please delete the file from all but the
preferred subvolume.- Pending matrix: [ [ 0 2 ] [ 1 0 ] ][2015-03-11 11:18:31.957692]
E [afr-self-heal-common.c:2868:afr_log_self_heal_completion_status] 0-md1-
replicate-2: metadata self heal failed, on /library
I am desperate...
A.
On Wednesday 11 March 2015 12:05:33 you wrote:
Hi, When trying to access a file on a gluster client (through fuse), I get an
"Input/output error" message. Getting the attributes for the file gives me for the first
brick # file: data/glusterfs/md1/brick1/kvm/hail/hail_home.qcow2 trusted.afr.md1-
client-2=0sAAAAAAAAAAAAAAAA trusted.afr.md1-client-3=0sAAABdAAAAAAAAAAA
trusted.gfid=0sOCFPGCdrQ9uyq2yTTPCKqQ== while for the second (replicate)
brick # file: data/glusterfs/md1/brick1/kvm/hail/hail_home.qcow2 trusted.afr.md1-
client-2=0sAAABJAAAAAAAAAAA trusted.afr.md1-client-3=0sAAAAAAAAAAAAAAAA
trusted.gfid=0sOCFPGCdrQ9uyq2yTTPCKqQ== It seems that I have a split-brain.
How can I solve this issue by resetting the attributes, please ? Thanks, Alessandro.
================== gluster volume info md1 Volume Name: md1 Type:
Distributed-Replicate Volume ID: 6da4b915-1def-4df4-a41c-2f3300ebf16b Status:
Started Number of Bricks: 3 x 2 = 6 Transport-type: tcp Bricks: Brick1:
tsunami1:/data/glusterfs/md1/brick1
Brick2: tsunami2:/data/glusterfs/md1/brick1 Brick3:
tsunami3:/data/glusterfs/md1/brick1 Brick4: tsunami4:/data/glusterfs/md1/brick1
Brick5: tsunami5:/data/glusterfs/md1/brick1 Brick6:
tsunami6:/data/glusterfs/md1/brick1 Options Reconfigured: server.allow-insecure: on
cluster.read-hash-mode: 2 features.quota: off performance.write-behind: on
performance.write-behind-window-size: 4MB performance.flush-behind: off
performance.io[2]-thread-count: 64 performance.cache-size: 512MB nfs.disable: on
cluster.lookup-unhashed: off
--------------------
http://www.gluster.org/mailman/listinfo/gluster-users[3]
--
Dr. Ir. Alessandro Ipe
Department of Observations Tel. +32 2 373 06 31
Remote Sensing from Space Fax. +32 2 374 67 88
Royal Meteorological Institute
Avenue Circulaire 3 Email:
B-1180 Brussels Belgium Alessandro.Ipe at meteo.be
Web: http://gerb.oma.be
--------
[1] http://joejulian.name/blog/glusterfs-split-brain-recovery-made-easy/
[2] http://performance.io
[3] http://www.gluster.org/mailman/listinfo/gluster-users
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20150312/993d0c22/attachment.html>
More information about the Gluster-users
mailing list