Martin, The output suggests that there are 2 replicas per 1 volume. So it should be present on only 2 bricks. Why is the file present in 4 bricks?. It should either be present on pserver12&13 or pserver3 & 5. I am not sure why you are expecting it to be there on 4 bricks. Am I missing any info here?. Pranith ----- Original Message ----- From: "Martin Schenker" <martin.schenker at profitbricks.com> To: gluster-users at gluster.org Sent: Wednesday, May 18, 2011 2:23:09 PM Subject: Re: Client and server file "view", different results?! Client can't see the right file. Here is another occurrence: The file 20819 is shown twice, different timestamps and attributes. 0 filesize on pserver3, outdated on pserver5, just 12&13 seems to be in sync. So what's going on? 0 root at de-dc1-c1-pserver13:~ # ls -al /opt/profitbricks/storage/images/2078/ebb83b05-3a83-9d18-ad8f-8542864da6ef/h dd-images/2081* -rwxrwx--- 1 libvirt-qemu kvm 53687091200 May 18 08:44 /opt/profitbricks/storage/images/2078/ebb83b05-3a83-9d18-ad8f-8542864da6ef/h dd-images/20819 -rwxrwx--- 1 libvirt-qemu kvm 53687091200 May 18 08:44 /opt/profitbricks/storage/images/2078/ebb83b05-3a83-9d18-ad8f-8542864da6ef/h dd-images/20819 0 root at de-dc1-c1-pserver3:~ # find /mnt/gluster/brick?/ -name 20819 | xargs -i ls -al {} -rwxrwx--- 1 libvirt-qemu vcb 0 May 14 17:00 /mnt/gluster/brick0/storage/images/2078/ebb83b05-3a83-9d18-ad8f-8542864da6ef /hdd-images/20819 0 root at de-dc1-c1-pserver3:~ # getfattr -dm - /mnt/gluster/brick0/storage/images/2078/ebb83b05-3a83-9d18-ad8f-8542864da6ef /hdd-images/20819 getfattr: Removing leading '/' from absolute path names # file: mnt/gluster/brick0/storage/images/2078/ebb83b05-3a83-9d18-ad8f-8542864da6ef/ hdd-images/20819 trusted.gfid=0sa5/rvjUUQ3ibSf32O3izOw== 0 root at pserver5:~ # find /mnt/gluster/brick?/ -name 20819 | xargs -i ls -al {} -rwxrwx--- 1 libvirt-qemu vcb 53687091200 May 14 17:00 /mnt/gluster/brick0/storage/images/2078/ebb83b05-3a83-9d18-ad8f-8542864da6ef /hdd-images/20819 0 root at pserver5:~ # getfattr -dm - /mnt/gluster/brick0/storage/images/2078/ebb83b05-3a83-9d18-ad8f-8542864da6ef /hdd-images/20819 getfattr: Removing leading '/' from absolute path names # file: mnt/gluster/brick0/storage/images/2078/ebb83b05-3a83-9d18-ad8f-8542864da6ef/ hdd-images/20819 trusted.afr.storage0-client-0=0sAAAAAgAAAAIAAAAA trusted.afr.storage0-client-1=0sAAAAAAAAAAAAAAAA trusted.gfid=0sa5/rvjUUQ3ibSf32O3izOw== 0 root at pserver12:~ # find /mnt/gluster/brick?/ -name 20819 | xargs -i ls -al {} -rwxrwx--- 1 libvirt-qemu kvm 53687091200 May 18 08:41 /mnt/gluster/brick1/storage/images/2078/ebb83b05-3a83-9d18-ad8f-8542864da6ef /hdd-images/20819 0 root at pserver12:~ # getfattr -dm - /mnt/gluster/brick1/storage/images/2078/ebb83b05-3a83-9d18-ad8f-8542864da6ef /hdd-images/20819 getfattr: Removing leading '/' from absolute path names # file: mnt/gluster/brick1/storage/images/2078/ebb83b05-3a83-9d18-ad8f-8542864da6ef/ hdd-images/20819 trusted.afr.storage0-client-6=0sAAAAAAAAAAAAAAAA trusted.afr.storage0-client-7=0sAAAAAAAAAAAAAAAA trusted.gfid=0sa5/rvjUUQ3ibSf32O3izOw== trusted.glusterfs.dht.linkto="storage0-replicate-0 0 root at de-dc1-c1-pserver13:~ # find /mnt/gluster/brick?/ -name 20819 | xargs -i ls -al {} -rwxrwx--- 1 libvirt-qemu kvm 53687091200 May 18 08:39 /mnt/gluster/brick1/storage/images/2078/ebb83b05-3a83-9d18-ad8f-8542864da6ef /hdd-images/20819 0 root at de-dc1-c1-pserver13:~ # getfattr -dm - /mnt/gluster/brick1/storage/images/2078/ebb83b05-3a83-9d18-ad8f-8542864da6ef /hdd-images/20819 getfattr: Removing leading '/' from absolute path names # file: mnt/gluster/brick1/storage/images/2078/ebb83b05-3a83-9d18-ad8f-8542864da6ef/ hdd-images/20819 trusted.afr.storage0-client-6=0sAAAAAAAAAAAAAAAA trusted.afr.storage0-client-7=0sAAAAAAAAAAAAAAAA trusted.gfid=0sa5/rvjUUQ3ibSf32O3izOw== trusted.glusterfs.dht.linkto="storage0-replicate-0 Only entrance in log file on pserver5, no references in the other three logs/servers: 0 root at pserver5:~ # grep 20819 /var/log/glusterfs/opt-profitbricks-storage.log [2011-05-17 20:37:30.52535] I [client-handshake.c:407:client3_1_reopen_cbk] 0-storage0-client-7: reopen on /images/2078/ebb83b05-3a83-9d18-ad8f-8542864da6ef/hdd-images/20819 succeeded (remote-fd = 6) [2011-05-17 20:37:34.824934] I [afr-open.c:435:afr_openfd_sh] 0-storage0-replicate-3: data self-heal triggered. path: /images/2078/ebb83b05-3a83-9d18-ad8f-8542864da6ef/hdd-images/20819, reason: Replicate up down flush, data lock is held [2011-05-17 20:37:34.825557] E [afr-self-heal-common.c:1214:sh_missing_entries_create] 0-storage0-replicate-3: no missing files - /images/2078/ebb83b05-3a83-9d18-ad8f-8542864da6ef/hdd-images/20819. proceeding to metadata check [2011-05-17 21:08:59.241203] I [afr-self-heal-algorithm.c:526:sh_diff_loop_driver_done] 0-storage0-replicate-3: diff self-heal on /images/2078/ebb83b05-3a83-9d18-ad8f-8542864da6ef/hdd-images/20819: 6 blocks of 409600 were different (0.00%) [2011-05-17 21:08:59.275873] I [afr-self-heal-common.c:1527:afr_self_heal_completion_cbk] 0-storage0-replicate-3: background data self-heal completed on /images/2078/ebb83b05-3a83-9d18-ad8f-8542864da6ef/hdd-images/20819 _______________________________________________ Gluster-users mailing list Gluster-users at gluster.org http://gluster.org/cgi-bin/mailman/listinfo/gluster-users