Hi Gandalf, can you run the following command on the brick path? "getfattr -d -e hex -m . /datastore" on both "nas-01-data" and "nas-02-data" nodes. This will let us know whether there is "trusted.glusterfs.volume-id" set. -Shwetha On 11/26/2013 07:36 PM, gandalf istari wrote: > hi thanks for the quick answer. > > I'm running glusterfs 3.4.1 > > [root at nas-02 datastore]# gluster volume start datastore1 force > > volume start: datastore1: failed: Failed to get extended attribute > trusted.glusterfs.volume-id for brick dir /datastore. Reason : No data > available > > It seems that the .gluster directory is missing for some reason. > > > volume replace-brick datastore1 nas-01-data:/datastorenas-02-data:/datastore > commit force > To rebuild/replace the missing brick ? > I'm quite new with glusterfs > Thanks > > > > > > > On 26/11/13 12:47, gandalf istari wrote: >> Hi have setup a two node replication glusterfs. After the initial >> installation the "master" node was put into the datacenter and after >> two week we moved the second one also to the datacenter. >> >> But the sync has not started yet. >> >> On the "master" >> >> gluster> volume info all >> >> Volume Name: datastore1 >> >> Type: Replicate >> >> Volume ID: fdff5190-85ef-4cba-9056-a6bbbd8d6863 >> >> Status: Started >> >> Number of Bricks: 1 x 2 = 2 >> >> Transport-type: tcp >> >> Bricks: >> >> Brick1: nas-01-data:/datastore >> >> Brick2: nas-02-data:/datastore >> >> gluster> peer status >> >> Number of Peers: 1 >> >> >> Hostname: nas-02-data >> >> Uuid: 71df9f86-a87b-481d-896c-c0d4ab679cfa >> >> State: Peer in Cluster (Connected) >> >> >> On the "slave" >> >> gluster> peer status >> >> Number of Peers: 1 >> >> Hostname: 192.168.70.6 >> >> Uuid: 97ef0154-ad7b-402a-b0cb-22be09134a3c >> >> State: Peer in Cluster (Connected) >> >> >> gluster> volume status all >> >> Status of volume: datastore1 >> >> Gluster processPortOnlinePid >> >> ------------------------------------------------------------------------------ >> >> Brick nas-01-data:/datastore49152Y2130 >> >> Brick nas-02-data:/datastoreN/ANN/A >> >> NFS Server on localhost2049Y8064 >> >> Self-heal Daemon on localhostN/AY8073 >> >> NFS Server on 192.168.70.62049Y3379 >> >> Self-heal Daemon on 192.168.70.6N/AY3384 >> > Which version of glusterfs are you running? > > volume status suggests that the second brick (nas-02-data:/datastore) > is not running. > > Can you run "gluster volume start <volname> force" in any of these two > nodes and try again? > Then you would also required to run `find . | xargs stat` on the > mountpoint of the volume. That should trigger the self heal. >> >> There are no active volume tasks >> >> >> I would like to run on the "slave" gluster volume sync nas-01-data >> datastore1 >> > BTW, There is no concept of "master" and "slave" in afr (replication). > However there is concept of "master volume" and "slave volume" in > gluster geo-replication. >> >> But then the virtual machines hosted will be unavailible is there >> another way to start the replication ? >> >> >> Thanks >> >> >> >> >> >> >> _______________________________________________ >> Gluster-users mailing list >> Gluster-users at gluster.org <mailto:Gluster-users at gluster.org> >> http://supercolony.gluster.org/mailman/listinfo/gluster-users > > > > _______________________________________________ > Gluster-users mailing list > Gluster-users at gluster.org > http://supercolony.gluster.org/mailman/listinfo/gluster-users -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://supercolony.gluster.org/pipermail/gluster-users/attachments/20131127/6607f9ab/attachment.html>