I am seeing the directory size to be different here. Let me confirm if we are checking extra for size to be same also (for directories it will not be needed). In that case, this log makes sense, but surely that is a false positive. -Amar On Thu, Apr 28, 2011 at 9:44 PM, Mohit Anchlia <mohitanchlia at gmail.com>wrote: > Yes they are the same. It looks like this problem appears only when I > use -I 256 when creating mkfs. Why would that be? > > [root at dsdb1 ~]# ls -ltr /data/ > total 5128 > drwx------ 2 root root 16384 Apr 27 16:57 lost+found > drwxr-xr-x 30003 root root 4562944 Apr 27 17:15 mnt-stress > drwxr-xr-x 30003 root root 598016 Apr 27 17:15 gluster > [root at dsdb1 ~]# ls -ltr /data1/ > total 572 > drwx------ 2 root root 16384 Apr 27 16:59 lost+found > drwxr-xr-x 30003 root root 561152 Apr 27 17:15 gluster > > [root at dsdb2 ~]# ls -ltr /data > total 588 > drwx------ 2 root root 16384 Apr 27 16:52 lost+found > drwxr-xr-x 2 root root 4096 Apr 27 17:09 mnt-stress > drwxr-xr-x 30003 root root 573440 Apr 27 17:15 gluster > [root at dsdb2 ~]# ls -ltr /data1 > total 592 > drwx------ 2 root root 16384 Apr 27 16:54 lost+found > drwxr-xr-x 30003 root root 581632 Apr 27 17:15 gluster > > > On Wed, Apr 27, 2011 at 11:18 PM, Amar Tumballi <amar at gluster.com> wrote: > >> > >> [2011-04-27 17:11:29.13142] E > >> [afr-self-heal-metadata.c:524:afr_sh_metadata_fix] > >> 0-stress-volume-replicate-0: Unable to self-heal permissions/ownership > >> of '/' (possible split-brain). Please fix the file on all backend > >> volumes > >> > >> Can someone please help me reason for this problem? > >> > >> gluster volume info all > >> > >> Volume Name: stress-volume > >> Type: Distributed-Replicate > >> Status: Started > >> Number of Bricks: 8 x 2 = 16 > >> Transport-type: tcp > >> Bricks: > >> Brick1: dsdb1:/data/gluster > >> Brick2: dsdb2:/data/gluster > > > > Did you check the permission/ownership of these exports? Please make sure > > that they are same. > > Regards, > > Amar > > > >> > >> Brick3: dsdb3:/data/gluster > >> Brick4: dsdb4:/data/gluster > >> Brick5: dsdb5:/data/gluster > >> Brick6: dsdb6:/data/gluster > >> Brick7: dslg1:/data/gluster > >> Brick8: dslg2:/data/gluster > >> Brick9: dsdb1:/data1/gluster > >> Brick10: dsdb2:/data1/gluster > >> Brick11: dsdb3:/data1/gluster > >> Brick12: dsdb4:/data1/gluster > >> Brick13: dsdb5:/data1/gluster > >> Brick14: dsdb6:/data1/gluster > >> Brick15: dslg1:/data1/gluster > >> Brick16: dslg2:/data1/gluster > > > > > > > -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://gluster.org/pipermail/gluster-users/attachments/20110428/31e4ee7b/attachment.htm>