You couldn't force start the volume because the brick "nas-02-data:/datastore" doesn't have the "trusted.glusterfs.volume-id" . From nas-02 node execute : 1. setfattr -n trusted.glusterfs.volume-id -v 0xfdff519085ef4cba9056a6bbbd8d6863 nas-02-data:/datastore 2. gluster volume start datastore1 force. -Shwetha On 11/27/2013 12:23 PM, gandalf istari wrote: > Hi Shwetha, > > [root at nas-01 ~]# getfattr -d -e hex -m . /datastore > > getfattr: Removing leading '/' from absolute path names > > # file: datastore > > trusted.afr.datastore1-client-0=0x000000000000000000000000 > > trusted.afr.datastore1-client-1=0x000000000000000000000000 > > trusted.gfid=0x00000000000000000000000000000001 > > trusted.glusterfs.dht=0x000000010000000000000000ffffffff > > trusted.glusterfs.volume-id=0xfdff519085ef4cba9056a6bbbd8d6863 > > > > [root at nas-02 ~]# getfattr -d -e hex -m . /datastore > > getfattr: Removing leading '/' from absolute path names > > # file: datastore > > security.selinux=0x73797374656d5f753a6f626a6563745f723a6574635f72756e74696d655f743a733000 > > I don't understand what happened . > > > gr > > Patrick > > > Hi Gandalf, > > can you run the following command on the brick path? > > "getfattr -d -e hex -m . /datastore" on both "nas-01-data" and > "nas-02-data" nodes. > > This will let us know whether there is "trusted.glusterfs.volume-id" set. > > -Shwetha > > On 11/26/2013 07:36 PM, gandalf istari wrote: >> hi thanks for the quick answer. >> >> I'm running glusterfs 3.4.1 >> >> [root at nas-02 datastore]# gluster volume start datastore1 force >> >> volume start: datastore1: failed: Failed to get extended attribute >> trusted.glusterfs.volume-id for brick dir /datastore. Reason : No >> data available >> >> It seems that the .gluster directory is missing for some reason. >> >> >> volume replace-brick datastore1 nas-01-data:/datastorenas-02-data:/datastore >> commit force >> To rebuild/replace the missing brick ? >> I'm quite new with glusterfs >> Thanks >> >> >> >> >> >> >> On 26/11/13 12:47, gandalf istari wrote: >>> Hi have setup a two node replication glusterfs. After the initial >>> installation the "master" node was put into the datacenter and after >>> two week we moved the second one also to the datacenter. >>> >>> But the sync has not started yet. >>> >>> On the "master" >>> >>> gluster> volume info all >>> >>> Volume Name: datastore1 >>> >>> Type: Replicate >>> >>> Volume ID: fdff5190-85ef-4cba-9056-a6bbbd8d6863 >>> >>> Status: Started >>> >>> Number of Bricks: 1 x 2 = 2 >>> >>> Transport-type: tcp >>> >>> Bricks: >>> >>> Brick1: nas-01-data:/datastore >>> >>> Brick2: nas-02-data:/datastore >>> >>> gluster> peer status >>> >>> Number of Peers: 1 >>> >>> >>> Hostname: nas-02-data >>> >>> Uuid: 71df9f86-a87b-481d-896c-c0d4ab679cfa >>> >>> State: Peer in Cluster (Connected) >>> >>> >>> On the "slave" >>> >>> gluster> peer status >>> >>> Number of Peers: 1 >>> >>> Hostname: 192.168.70.6 >>> >>> Uuid: 97ef0154-ad7b-402a-b0cb-22be09134a3c >>> >>> State: Peer in Cluster (Connected) >>> >>> >>> gluster> volume status all >>> >>> Status of volume: datastore1 >>> >>> Gluster processPortOnlinePid >>> >>> ------------------------------------------------------------------------------ >>> >>> Brick nas-01-data:/datastore49152Y2130 >>> >>> Brick nas-02-data:/datastoreN/ANN/A >>> >>> NFS Server on localhost2049Y8064 >>> >>> Self-heal Daemon on localhostN/AY8073 >>> >>> NFS Server on 192.168.70.62049Y3379 >>> >>> Self-heal Daemon on 192.168.70.6N/AY3384 >>> >> Which version of glusterfs are you running? >> >> volume status suggests that the second brick (nas-02-data:/datastore) >> is not running. >> >> Can you run "gluster volume start <volname> force" in any of these >> two nodes and try again? >> Then you would also required to run `find . | xargs stat` on the >> mountpoint of the volume. That should trigger the self heal. >>> >>> There are no active volume tasks >>> >>> >>> I would like to run on the "slave" gluster volume sync nas-01-data >>> datastore1 >>> >> BTW, There is no concept of "master" and "slave" in afr >> (replication). However there is concept of "master volume" and "slave >> volume" in gluster geo-replication. >>> >>> But then the virtual machines hosted will be unavailible is there >>> another way to start the replication ? >>> >>> >>> Thanks >>> >>> >>> >>> >>> >>> >>> _______________________________________________ >>> Gluster-users mailing list >>> Gluster-users at gluster.org <mailto:Gluster-users at gluster.org> >>> http://supercolony.gluster.org/mailman/listinfo/gluster-users >> >> >> >> _______________________________________________ >> Gluster-users mailing list >> Gluster-users at gluster.org <mailto:Gluster-users at gluster.org> >> http://supercolony.gluster.org/mailman/listinfo/gluster-users > > > -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://supercolony.gluster.org/pipermail/gluster-users/attachments/20131127/4ce6ed2e/attachment.html>