Hello Avati and all, I sent the log of getfattr, and gfid mismatch is found. Is this solution a difficult? And I'm interested in also the mechanism of gfid mismatch. Best regards, Kondo 2012/12/4 kenji kondo <kkay.jp at gmail.com> > Hello Avati, > > I got the log as below. > We can see that one gfid (g13) differs from others. > I sometimes found same situation so far. > Best regards > Kondo > ------------------------ > > backend/ai/anchor/3: > > g10:trusted.gfid=0xd1366c6a09c345e8bab26b7953ac6405 > > g11:trusted.gfid=0xd1366c6a09c345e8bab26b7953ac6405 > > g12:trusted.gfid=0xd1366c6a09c345e8bab26b7953ac6405 > > g13:trusted.gfid=0x93c1e7a2797a4b888e17d9b6582da365 > > g14:trusted.gfid=0xd1366c6a09c345e8bab26b7953ac6405 > > g15:trusted.gfid=0xd1366c6a09c345e8bab26b7953ac6405 > > g16:trusted.gfid=0xd1366c6a09c345e8bab26b7953ac6405 > > g17:trusted.gfid=0xd1366c6a09c345e8bab26b7953ac6405 > > g18:trusted.gfid=0xd1366c6a09c345e8bab26b7953ac6405 > > g10:trusted.glusterfs.dht=0x00000001000000003ffffffe5ffffffc > > g11:trusted.glusterfs.dht=0x00000001000000005ffffffd7ffffffb > > g12:trusted.glusterfs.dht=0x00000001000000007ffffffc9ffffffa > > g13:trusted.glusterfs.dht=0x000000010000000000000000ffffffff > > g14:trusted.glusterfs.dht=0x00000001000000009ffffffbbffffff9 > > g15:trusted.glusterfs.dht=0x0000000100000000bffffffadffffff8 > > g16:trusted.glusterfs.dht=0x0000000100000000dffffff9ffffffff > > g17:trusted.glusterfs.dht=0x0000000100000000000000001ffffffe > > g18:trusted.glusterfs.dht=0x00000001000000001fffffff3ffffffd > > > > /backend/ai/anchor: > > g10:trusted.gfid=0x89a33935d5da4acd85586e075eca221d > > g11:trusted.gfid=0x89a33935d5da4acd85586e075eca221d > > g12:trusted.gfid=0x89a33935d5da4acd85586e075eca221d > > g13:trusted.gfid=0x6ee644c825dd4806b2680bfeb985a27d > > g14:trusted.gfid=0x89a33935d5da4acd85586e075eca221d > > g15:trusted.gfid=0x89a33935d5da4acd85586e075eca221d > > g16:trusted.gfid=0x89a33935d5da4acd85586e075eca221d > > g17:trusted.gfid=0x89a33935d5da4acd85586e075eca221d > > g18:trusted.gfid=0x89a33935d5da4acd85586e075eca221d > > g10:trusted.glusterfs.dht=0x00000001000000005555555471c71c6f > > g11:trusted.glusterfs.dht=0x000000010000000071c71c708e38e38b > > g12:trusted.glusterfs.dht=0x00000001000000008e38e38caaaaaaa7 > > g13:trusted.glusterfs.dht=0x0000000100000000aaaaaaa8c71c71c3 > > g14:trusted.glusterfs.dht=0x0000000100000000c71c71c4e38e38df > > g15:trusted.glusterfs.dht=0x0000000100000000e38e38e0ffffffff > > g16:trusted.glusterfs.dht=0x0000000100000000000000001c71c71b > > g17:trusted.glusterfs.dht=0x00000001000000001c71c71c38e38e37 > > g18:trusted.glusterfs.dht=0x000000010000000038e38e3855555553 > > > > /backend/ai > > g10.gfid:trusted.gfid=0x258ba47b30ee41a984ceea1a491b1669 > > g11.gfid:trusted.gfid=0x258ba47b30ee41a984ceea1a491b1669 > > g12.gfid:trusted.gfid=0x258ba47b30ee41a984ceea1a491b1669 > > g13.gfid:trusted.gfid=0x258ba47b30ee41a984ceea1a491b1669 > > g14.gfid:trusted.gfid=0x258ba47b30ee41a984ceea1a491b1669 > > g15.gfid:trusted.gfid=0x258ba47b30ee41a984ceea1a491b1669 > > g16.gfid:trusted.gfid=0x258ba47b30ee41a984ceea1a491b1669 > > g17.gfid:trusted.gfid=0x258ba47b30ee41a984ceea1a491b1669 > > g18.gfid:trusted.gfid=0x258ba47b30ee41a984ceea1a491b1669 > > g10.gfid:trusted.glusterfs.dht=0x00000001000000008e38e38caaaaaaa7 > > g11.gfid:trusted.glusterfs.dht=0x0000000100000000aaaaaaa8c71c71c3 > > g12.gfid:trusted.glusterfs.dht=0x0000000100000000c71c71c4e38e38df > > g13.gfid:trusted.glusterfs.dht=0x0000000100000000e38e38e0ffffffff > > g14.gfid:trusted.glusterfs.dht=0x0000000100000000000000001c71c71b > > g15.gfid:trusted.glusterfs.dht=0x00000001000000001c71c71c38e38e37 > > g16.gfid:trusted.glusterfs.dht=0x000000010000000038e38e3855555553 > > g17.gfid:trusted.glusterfs.dht=0x00000001000000005555555471c71c6f > > g18.gfid:trusted.glusterfs.dht=0x000000010000000071c71c708e38e38b > > > > > > On 2012/12/04, at 9:20, Anand Avati <anand.avati at gmail.com> wrote: > > > > On Mon, Dec 3, 2012 at 5:14 AM, kenji kondo <kkay.jp at gmail.com> wrote: > >> Hello Avati, thank you for your reply. >> >> I tried to test your suggestion in 3.2.7, but I could not test in 3.3.x >> because I don't have. >> For the results, unfortunately new similar problems occurred as following: >> >> gluster> volume set vol22 performance.stat-prefetch off >> gluster> volume info vol22 >> >> Volume Name: vol22 >> Type: Distribute >> Status: Started >> Number of Bricks: 9 >> Transport-type: tcp >> Bricks: >> Brick1: gluster10:/export22/brick >> Brick2: gluster11:/export22/brick >> Brick3: gluster12:/export22/brick >> Brick4: gluster13:/export22/brick >> Brick5: gluster14:/export22/brick >> Brick6: gluster15:/export22/brick >> Brick7: gluster16:/export22/brick >> Brick8: gluster17:/export22/brick >> Brick9: gluster18:/export22/brick >> Options Reconfigured: >> performance.stat-prefetch: off >> >> >> After this setting, I ran same simulation in the vol22. >> But strange directories were made, that I could not remove the some >> directories as below: >> >> $ ls -a ai/anchor/3 >> . .. >> >> $ rmdir ai/anchor/3 >> rmdir: ai/anchor/3: No such file or directory >> >> >> then I found error messages: >> >> [2012-12-03 18:08:14.816313] E [client3_1-fops.c:2228:client3_1_lookup_cbk] >> 0-vol22-client-3: remote operation failed: Stale NFS file handle >> [2012-12-03 18:08:14.817196] W [dht-common.c:178:dht_lookup_dir_cbk] >> 0-vol22-dht: /test1130/ai/anchor: gfid different on vol22-client-4 >> [2012-12-03 18:08:14.817258] W [dht-common.c:178:dht_lookup_dir_cbk] >> 0-vol22-dht: /test1130/ai/anchor: gfid different on vol22-client-0 >> [2012-12-03 18:08:14.817322] W [dht-common.c:178:dht_lookup_dir_cbk] >> 0-vol22-dht: /test1130/ai/anchor: gfid different on vol22-client-2 >> [2012-12-03 18:08:14.817367] W [dht-common.c:178:dht_lookup_dir_cbk] >> 0-vol22-dht: /test1130/ai/anchor: gfid different on vol22-client-1 >> [2012-12-03 18:08:14.817398] W [dht-common.c:178:dht_lookup_dir_cbk] >> 0-vol22-dht: /test1130/ai/anchor: gfid different on vol22-client-5 >> [2012-12-03 18:08:14.817430] W [dht-common.c:178:dht_lookup_dir_cbk] >> 0-vol22-dht: /test1130/ai/anchor: gfid different on vol22-client-8 >> [2012-12-03 18:08:14.817460] W [dht-common.c:178:dht_lookup_dir_cbk] >> 0-vol22-dht: /test1130/ai/anchor: gfid different on vol22-client-7 >> [2012-12-03 18:08:14.817506] W [dht-common.c:178:dht_lookup_dir_cbk] >> 0-vol22-dht: /test1130/ai/anchor: gfid different on vol22-client-6 >> [2012-12-03 18:08:14.818865] E [client3_1-fops.c:2132:client3_1_opendir_cbk] >> 0-vol22-client-3: remote operation failed: No such file or directory >> [2012-12-03 18:08:14.819198] W [fuse-bridge.c:1016:fuse_unlink_cbk] >> 0-glusterfs-fuse: 1684950: RMDIR() >> /test1130/ai/anchor/3 => -1 (No such file or directory) >> >> >> And, I found strange dht with getfattr command, >> >> $ sudo getfattr -d -m '.*' -n trusted.glusterfs.pathinfo ai/anchor/3 >> >> >> trusted.glusterfs.pathinfo="(vol22-dht-layout (vol22-client-0 1073741822 >> 1610612732) (vol22-client-1 1610612733 2147483643) (vol22-client-2 2147483644 >> 2684354554) >> >> (vol22-client-3 0 0) >> >> (vol22-client-4 2684354555 3221225465) >> (vol22-client-5 3221225466 3758096376) (vol22-client-6 3758096377 4294967295) >> (vol22-client-7 0 536870910) (vol22-client-8 536870911 1073741821))" >> >> >> vol22-client-3 is 0 to 0?: this would be incorrect. >> >> Above problem is found in all of our clients. >> I expect that the problems relate to exclusive control for mkdir, but >> I don't enough understand this phenomenon. >> >> Can you have some idea? >> I will try. >> >> > In such a state, can you get the backend xattr dumps of ai/anchor/ and > ai/anchor/3/ with the following commands run on ALL servers - > > sh# getfattr -d -e hex -m . /backend/ai/anchor/ > sh# getfattr -d -e hex -m . /backend/ai/anchor/3/ > > Thanks, > Avati > > -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://supercolony.gluster.org/pipermail/gluster-users/attachments/20121206/0e6f4dfc/attachment-0001.html>