thanks Shwetha. you solved my problem! --------------------------- ??? 010 5881 3749 ????? ????????? On 2013?11?28?, at ??5:47, shwetha <spandura at redhat.com> wrote: > Since the "trusted.glusterfs.volume-id" is not set on the bricks , the volume start will fail. > > 1) Execute : "getfattr -e hex -n trusted.glusterfs.volume-id /opt/gluster_data/eccp_glance" from any of the node which the brick process running. > > You will get the hex value of the "trusted.glusterfs.volume-id" extended attribute . Set this value on the bricks created (On the nodes where you deleted the bricks) > > 2) setfattr -n trusted.glusterfs.volume-id -v <hex_value_of_the_trusted_glusterfs_volume-id> /opt/gluster_data/eccp_glance > > 3) gluster volume start <volume_name> force : To restart the brick process > > 4) gluster volume status <volume_name> : Check all the brick process are started. > > 5) gluster volume heal <volume_name> full : To trigger self-heal on to the removed bricks. > > Shwetha > On 11/28/2013 02:59 PM, weiyuanke wrote: >> hi Shwetha, >> >> >> command "gluster volume start eccp_glance force" on the other node gives following: >> >> >> with cli.log >> >> >> on the damaged node?gluster volume start eccp_glance force gives: >> >> >> >> >> >> --------------------------- >> ??? >> 010 5881 3749 >> ????? ????????? >> >> >> >> >> >> On 2013?11?28?, at ??4:50, shwetha <spandura at redhat.com> wrote: >> >>> 1) create the brick directory "/opt/gluster_data/eccp_glance" on the nodes where you deleted the directories. >>> >>> 2) From any of the storage node execute : >>> gluster volume start <volume_name> force : To restart the brick process >>> gluster volume status <volume_name> : Check all the brick process are started. >>> gluster volume heal <volume_name> full : To trigger self-heal on to the removed bricks. >>> -Shwetha >>> >>> On 11/28/2013 02:09 PM, ??? wrote: >>>> hi all, >>>> >>>> I accidentally removed the brick directory of a volume on one node, the replica for this volume is 2. >>>> >>>> now the situation is , there is no corresponding glusterfsd process on this node, and 'glusterfs volume status' shows that the brick is offline, like this: >>>> Brick 192.168.64.11:/opt/gluster_data/eccp_glance N/A Y 2513 >>>> Brick 192.168.64.12:/opt/gluster_data/eccp_glance 49161 Y 2542 >>>> Brick 192.168.64.17:/opt/gluster_data/eccp_glance 49164 Y 2537 >>>> Brick 192.168.64.18:/opt/gluster_data/eccp_glance 49154 Y 4978 >>>> Brick 192.168.64.29:/opt/gluster_data/eccp_glance N/A N N/A >>>> Brick 192.168.64.30:/opt/gluster_data/eccp_glance 49154 Y 4072 >>>> Brick 192.168.64.25:/opt/gluster_data/eccp_glance 49155 Y 11975 >>>> Brick 192.168.64.26:/opt/gluster_data/eccp_glance 49155 Y 17947 >>>> Brick 192.168.64.13:/opt/gluster_data/eccp_glance 49154 Y 26045 >>>> Brick 192.168.64.14:/opt/gluster_data/eccp_glance 49154 Y 22143 >>>> >>>> >>>> so are there ways to bring this brick back to normal? >>>> >>>> thanks! >>>> >>>> >>>> ----------------------------------------------------------------- >>>> ??? >>>> ????? ????????? >>>> >>>> >>>> >>>> _______________________________________________ >>>> Gluster-users mailing list >>>> Gluster-users at gluster.org >>>> http://supercolony.gluster.org/mailman/listinfo/gluster-users >>> >>> >> > > -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://supercolony.gluster.org/pipermail/gluster-users/attachments/20131128/a9b21bb4/attachment.html>