Hi there, after reporting some trouble with group access permissions, http://gluster.org/pipermail/gluster-users/2011-May/007619.html (which still persist, btw.) things get worse and worse with each day. Now, we see a lot of duplicate files (again, only fuse-clients here), access permissions are reset on a random and totally annoying basis. Files are empty from time to time and become: -rwxrws--x 1 user1 group2 594 2011-02-04 18:43 preprocessing128.m -rwxrws--x 1 user1 group2 594 2011-02-04 18:43 preprocessing128.m -rwxrws--x 1 user2 group2 531 2011-03-03 10:47 result_11.mat ------S--T 1 root group2 0 2011-04-14 07:57 result_11.mat -rwxrws--x 1 user1 group2 11069 2010-12-02 14:53 trigger.odt -rwxrws--x 1 user1 group2 11069 2010-12-02 14:53 trigger.odt where group2 are secondary groups. How come that there are these empty and duplicate files? Again, this listing is from the fuse-mount Could it be that version 3.2.0 is totally borked? Btw.: From time to time, these permissions as well as which duplicate files one sees change in a random manner. I followed various hints on configuring and deconfiguring options Went from: root at store02:/var/log/glusterfs# gluster volume info store Volume Name: store Type: Distributed-Replicate Status: Started Number of Bricks: 5 x 2 = 10 Transport-type: tcp Bricks: Brick1: store01-i:/srv/store01 Brick2: pvmserv01-i:/srv/store01 Brick3: pvmserv02-i:/srv/store01 Brick4: store02-i:/srv/store03 Brick5: store02-i:/srv/store01 Brick6: store01-i:/srv/store02 Brick7: store02-i:/srv/store04 Brick8: store02-i:/srv/store05 Brick9: store02-i:/srv/store06 Brick10: store02-i:/srv/store02 Options Reconfigured: nfs.disable: on auth.allow: 127.0.0.1,10.10.* performance.cache-size: 1024Mb performance.write-behind-window-size: 64Mb performance.io-thread-count: 32 diagnostics.dump-fd-stats: off diagnostics.brick-log-level: WARNING diagnostics.client-log-level: WARNING performance.stat-prefetch: off diagnostics.latency-measurement: off performance.flush-behind: off performance.quick-read: disable to: Volume Name: store Type: Distributed-Replicate Status: Started Number of Bricks: 5 x 2 = 10 Transport-type: tcp Bricks: Brick1: store01-i:/srv/store01 Brick2: pvmserv01-i:/srv/store01 Brick3: pvmserv02-i:/srv/store01 Brick4: store02-i:/srv/store03 Brick5: store02-i:/srv/store01 Brick6: store01-i:/srv/store02 Brick7: store02-i:/srv/store04 Brick8: store02-i:/srv/store05 Brick9: store02-i:/srv/store06 Brick10: store02-i:/srv/store02 Options Reconfigured: auth.allow: 127.0.0.1,10.10.* nothing helped. Currently our only option seems to be to go away from glusterfs to some other filesystem which would be a bitter decission. Thanks for any help, udo. -- Institute of Cognitive Science - System Administration Team Albrechtstrasse 28 - 49076 Osnabrueck - Germany Tel: +49-541-969-3362 - Fax: +49-541-969-3361 https://doc.ikw.uni-osnabrueck.de