On Tue, Sep 19, 2017 at 04:10:26PM -0500, Walter Deignan wrote: > I recently compiled the 3.10-5 client from source on a few PPC64 systems > running RHEL 7.3. They are mounting a Gluster volume which is hosted on > more traditional x86 servers. > > Everything seems to be working properly except for creating new > directories from the PPC64 clients. The mkdir command gives a > "Input/output error" and for the first few minutes the new directory is > inaccessible. I checked the backend bricks and confirmed the directory was > created properly on all of them. After waiting for 2-5 minutes the > directory magically becomes accessible. > > This inaccessible directory issue only appears from the client which > created it. When creating the directory from client #1 I can immediately > see it with no errors from client #2. > > Using a pre-compiled 3.10-5 package on an x86 client doesn't show the > issue. > > I poked around bugzilla but couldn't seem to find anything which matches > this. Maybe https://bugzilla.redhat.com/show_bug.cgi?id=951903 ? Some of the details have also been captured in https://github.com/gluster/glusterfs/issues/203 Capturing a tcpdump and opening it up with Wireshark may help in identifying if GFID's are mixed up. Possibly some are also mentioned in different logs, but that might be more difficult to find. HTH, Niels > > [root@mqdev1 hafsdev1_gv0]# ls -lh > total 8.0K > drwxrwxr-x. 4 mqm mqm 4.0K Sep 19 15:47 data > drwxr-xr-x. 2 root root 4.0K Sep 19 15:47 testdir > [root@mqdev1 hafsdev1_gv0]# mkdir testdir2 > mkdir: cannot create directory ?testdir2?: Input/output error > [root@mqdev1 hafsdev1_gv0]# ls > ls: cannot access testdir2: No such file or directory > data testdir testdir2 > [root@mqdev1 hafsdev1_gv0]# ls -lht > ls: cannot access testdir2: No such file or directory > total 8.0K > drwxr-xr-x. 2 root root 4.0K Sep 19 15:47 testdir > drwxrwxr-x. 4 mqm mqm 4.0K Sep 19 15:47 data > d?????????? ? ? ? ? ? testdir2 > [root@mqdev1 hafsdev1_gv0]# cd testdir2 > -bash: cd: testdir2: No such file or directory > > *Wait a few minutes...* > > [root@mqdev1 hafsdev1_gv0]# ls -lht > total 12K > drwxr-xr-x. 2 root root 4.0K Sep 19 15:50 testdir2 > drwxr-xr-x. 2 root root 4.0K Sep 19 15:47 testdir > drwxrwxr-x. 4 mqm mqm 4.0K Sep 19 15:47 data > [root@mqdev1 hafsdev1_gv0]# > > My volume config... > > [root@dc-hafsdev1a bricks]# gluster volume info > > Volume Name: gv0 > Type: Replicate > Volume ID: a2d37705-05cb-4700-8ed8-2cb89376faf0 > Status: Started > Snapshot Count: 0 > Number of Bricks: 1 x 3 = 3 > Transport-type: tcp > Bricks: > Brick1: dc-hafsdev1a.ulinedm.com:/gluster/bricks/brick1/data > Brick2: dc-hafsdev1b.ulinedm.com:/gluster/bricks/brick1/data > Brick3: dc-hafsdev1c.ulinedm.com:/gluster/bricks/brick1/data > Options Reconfigured: > nfs.disable: on > transport.address-family: inet > network.ping-timeout: 2 > features.bitrot: on > features.scrub: Active > cluster.server-quorum-ratio: 51% > > -Walter Deignan > -Uline IT, Systems Architect > _______________________________________________ > Gluster-users mailing list > Gluster-users@xxxxxxxxxxx > http://lists.gluster.org/mailman/listinfo/gluster-users _______________________________________________ Gluster-users mailing list Gluster-users@xxxxxxxxxxx http://lists.gluster.org/mailman/listinfo/gluster-users