[Gluster-users] "Input/output error" on mkdir for PPC64 based client

Niels de Vos ndevos at redhat.com
Thu Sep 21 14:16:06 UTC 2017


On Tue, Sep 19, 2017 at 04:10:26PM -0500, Walter Deignan wrote:
> I recently compiled the 3.10-5 client from source on a few PPC64 systems 
> running RHEL 7.3. They are mounting a Gluster volume which is hosted on 
> more traditional x86 servers.
> 
> Everything seems to be working properly except for creating new 
> directories from the PPC64 clients. The mkdir command gives a 
> "Input/output error" and for the first few minutes the new directory is 
> inaccessible. I checked the backend bricks and confirmed the directory was 
> created properly on all of them. After waiting for 2-5 minutes the 
> directory magically becomes accessible.
> 
> This inaccessible directory issue only appears from the client which 
> created it. When creating the directory from client #1 I can immediately 
> see it with no errors from client #2.
> 
> Using a pre-compiled 3.10-5 package on an x86 client doesn't show the 
> issue.
> 
> I poked around bugzilla but couldn't seem to find anything which matches 
> this.

Maybe https://bugzilla.redhat.com/show_bug.cgi?id=951903 ?
Some of the details have also been captured in
https://github.com/gluster/glusterfs/issues/203

Capturing a tcpdump and opening it up with Wireshark may help in
identifying if GFID's are mixed up. Possibly some are also mentioned in
different logs, but that might be more difficult to find.

HTH,
Niels


> 
> [root at mqdev1 hafsdev1_gv0]# ls -lh
> total 8.0K
> drwxrwxr-x. 4 mqm  mqm  4.0K Sep 19 15:47 data
> drwxr-xr-x. 2 root root 4.0K Sep 19 15:47 testdir
> [root at mqdev1 hafsdev1_gv0]# mkdir testdir2
> mkdir: cannot create directory ?testdir2?: Input/output error
> [root at mqdev1 hafsdev1_gv0]# ls
> ls: cannot access testdir2: No such file or directory
> data  testdir  testdir2
> [root at mqdev1 hafsdev1_gv0]# ls -lht
> ls: cannot access testdir2: No such file or directory
> total 8.0K
> drwxr-xr-x. 2 root root 4.0K Sep 19 15:47 testdir
> drwxrwxr-x. 4 mqm  mqm  4.0K Sep 19 15:47 data
> d?????????? ? ?    ?       ?            ? testdir2
> [root at mqdev1 hafsdev1_gv0]# cd testdir2
> -bash: cd: testdir2: No such file or directory
> 
> *Wait a few minutes...*
> 
> [root at mqdev1 hafsdev1_gv0]# ls -lht
> total 12K
> drwxr-xr-x. 2 root root 4.0K Sep 19 15:50 testdir2
> drwxr-xr-x. 2 root root 4.0K Sep 19 15:47 testdir
> drwxrwxr-x. 4 mqm  mqm  4.0K Sep 19 15:47 data
> [root at mqdev1 hafsdev1_gv0]#
> 
> My volume config...
> 
> [root at dc-hafsdev1a bricks]# gluster volume info
> 
> Volume Name: gv0
> Type: Replicate
> Volume ID: a2d37705-05cb-4700-8ed8-2cb89376faf0
> Status: Started
> Snapshot Count: 0
> Number of Bricks: 1 x 3 = 3
> Transport-type: tcp
> Bricks:
> Brick1: dc-hafsdev1a.ulinedm.com:/gluster/bricks/brick1/data
> Brick2: dc-hafsdev1b.ulinedm.com:/gluster/bricks/brick1/data
> Brick3: dc-hafsdev1c.ulinedm.com:/gluster/bricks/brick1/data
> Options Reconfigured:
> nfs.disable: on
> transport.address-family: inet
> network.ping-timeout: 2
> features.bitrot: on
> features.scrub: Active
> cluster.server-quorum-ratio: 51%
> 
> -Walter Deignan
> -Uline IT, Systems Architect

> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://lists.gluster.org/mailman/listinfo/gluster-users



More information about the Gluster-users mailing list