To start self-healing data execute : "gluster volume heal datastore1 full" . To monitor self-heal completion status exeucte : "gluster volume heal datastore1 info". The number of entries under each brick should be 0. When the number of entries count becomes 0 the self-heal is completed. Other way to trigger self-heal is to create a new fuse mount and execute "find . | xargs stat" from the mount point. -Shwetha On 11/27/2013 01:07 PM, gandalf istari wrote: > Thank you so much, > > It seems to be working the directories are now created but still > empty. I suppose this will take a while to sync 44GB > > The only change I have done was: leave out nas-02-data for your command. > > whats the best way now to monitor the sync process ? > > > > You couldn't force start the volume because the brick > "nas-02-data:/datastore" doesn't have the "trusted.glusterfs.volume-id" . > > From nas-02 node execute : > > 1. setfattr -n trusted.glusterfs.volume-id -v > 0xfdff519085ef4cba9056a6bbbd8d6863 nas-02-data:/datastore > > 2. gluster volume start datastore1 force. > > > -Shwetha > On 11/27/2013 12:23 PM, gandalf istari wrote: >> Hi Shwetha, >> >> [root at nas-01 ~]# getfattr -d -e hex -m . /datastore >> >> getfattr: Removing leading '/' from absolute path names >> >> # file: datastore >> >> trusted.afr.datastore1-client-0=0x000000000000000000000000 >> >> trusted.afr.datastore1-client-1=0x000000000000000000000000 >> >> trusted.gfid=0x00000000000000000000000000000001 >> >> trusted.glusterfs.dht=0x000000010000000000000000ffffffff >> >> trusted.glusterfs.volume-id=0xfdff519085ef4cba9056a6bbbd8d6863 >> >> >> >> [root at nas-02 ~]# getfattr -d -e hex -m . /datastore >> >> getfattr: Removing leading '/' from absolute path names >> >> # file: datastore >> >> security.selinux=0x73797374656d5f753a6f626a6563745f723a6574635f72756e74696d655f743a733000 >> >> I don't understand what happened . >> >> >> gr >> >> Patrick >> >> >> Hi Gandalf, >> >> can you run the following command on the brick path? >> >> "getfattr -d -e hex -m . /datastore" on both "nas-01-data" and >> "nas-02-data" nodes. >> >> This will let us know whether there is "trusted.glusterfs.volume-id" >> set. >> >> -Shwetha >> >> On 11/26/2013 07:36 PM, gandalf istari wrote: >>> hi thanks for the quick answer. >>> >>> I'm running glusterfs 3.4.1 >>> >>> [root at nas-02 datastore]# gluster volume start datastore1 force >>> >>> volume start: datastore1: failed: Failed to get extended attribute >>> trusted.glusterfs.volume-id for brick dir /datastore. Reason : No >>> data available >>> >>> It seems that the .gluster directory is missing for some reason. >>> >>> >>> volume replace-brick datastore1 nas-01-data:/datastorenas-02-data:/datastore >>> commit force >>> To rebuild/replace the missing brick ? >>> I'm quite new with glusterfs >>> Thanks >>> >>> >>> >>> >>> >>> >>> On 26/11/13 12:47, gandalf istari wrote: >>>> Hi have setup a two node replication glusterfs. After the initial >>>> installation the "master" node was put into the datacenter and >>>> after two week we moved the second one also to the datacenter. >>>> >>>> But the sync has not started yet. >>>> >>>> On the "master" >>>> >>>> gluster> volume info all >>>> >>>> Volume Name: datastore1 >>>> >>>> Type: Replicate >>>> >>>> Volume ID: fdff5190-85ef-4cba-9056-a6bbbd8d6863 >>>> >>>> Status: Started >>>> >>>> Number of Bricks: 1 x 2 = 2 >>>> >>>> Transport-type: tcp >>>> >>>> Bricks: >>>> >>>> Brick1: nas-01-data:/datastore >>>> >>>> Brick2: nas-02-data:/datastore >>>> >>>> gluster> peer status >>>> >>>> Number of Peers: 1 >>>> >>>> >>>> Hostname: nas-02-data >>>> >>>> Uuid: 71df9f86-a87b-481d-896c-c0d4ab679cfa >>>> >>>> State: Peer in Cluster (Connected) >>>> >>>> >>>> On the "slave" >>>> >>>> gluster> peer status >>>> >>>> Number of Peers: 1 >>>> >>>> Hostname: 192.168.70.6 >>>> >>>> Uuid: 97ef0154-ad7b-402a-b0cb-22be09134a3c >>>> >>>> State: Peer in Cluster (Connected) >>>> >>>> >>>> gluster> volume status all >>>> >>>> Status of volume: datastore1 >>>> >>>> Gluster processPortOnlinePid >>>> >>>> ------------------------------------------------------------------------------ >>>> >>>> Brick nas-01-data:/datastore49152Y2130 >>>> >>>> Brick nas-02-data:/datastoreN/ANN/A >>>> >>>> NFS Server on localhost2049Y8064 >>>> >>>> Self-heal Daemon on localhostN/AY8073 >>>> >>>> NFS Server on 192.168.70.62049Y3379 >>>> >>>> Self-heal Daemon on 192.168.70.6N/AY3384 >>>> >>> Which version of glusterfs are you running? >>> >>> volume status suggests that the second brick >>> (nas-02-data:/datastore) is not running. >>> >>> Can you run "gluster volume start <volname> force" in any of these >>> two nodes and try again? >>> Then you would also required to run `find . | xargs stat` on the >>> mountpoint of the volume. That should trigger the self heal. >>>> >>>> There are no active volume tasks >>>> >>>> >>>> I would like to run on the "slave" gluster volume sync nas-01-data >>>> datastore1 >>>> >>> BTW, There is no concept of "master" and "slave" in afr >>> (replication). However there is concept of "master volume" and >>> "slave volume" in gluster geo-replication. >>>> >>>> But then the virtual machines hosted will be unavailible is there >>>> another way to start the replication ? >>>> >>>> >>>> Thanks >>>> >>>> >>>> >>>> >>>> >>>> >>>> _______________________________________________ >>>> Gluster-users mailing list >>>> Gluster-users at gluster.org <mailto:Gluster-users at gluster.org> >>>> http://supercolony.gluster.org/mailman/listinfo/gluster-users >>> >>> >>> >>> _______________________________________________ >>> Gluster-users mailing list >>> Gluster-users at gluster.org <mailto:Gluster-users at gluster.org> >>> http://supercolony.gluster.org/mailman/listinfo/gluster-users >> >> >> > > -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://supercolony.gluster.org/pipermail/gluster-users/attachments/20131127/b374d2f4/attachment.html>