Just started trying gluster, to decide if we want to put it into production.
Replicated, distributed volume, two servers, 20 bricks per server:
Wouldn't you know it, within a week or two of pulling the hardware together and getting gluster installed and configured, a disk dies. Note the dead process for brick15 on server storinator2.[root@storinator1 ~]# gluster volume status gv0Status of volume: gv0Gluster process TCP Port RDMA Port Online Pid------------------------------------------------------------------------------Brick storinator1:/export/brick1/gv0 49153 0 Y 2554Brick storinator2:/export/brick1/gv0 49153 0 Y 9686Brick storinator1:/export/brick2/gv0 49154 0 Y 2562Brick storinator2:/export/brick2/gv0 49154 0 Y 9708Brick storinator1:/export/brick3/gv0 49155 0 Y 2568Brick storinator2:/export/brick3/gv0 49155 0 Y 9692Brick storinator1:/export/brick4/gv0 49156 0 Y 2574Brick storinator2:/export/brick4/gv0 49156 0 Y 9765Brick storinator1:/export/brick5/gv0 49173 0 Y 16901Brick storinator2:/export/brick5/gv0 49173 0 Y 9727Brick storinator1:/export/brick6/gv0 49174 0 Y 16920Brick storinator2:/export/brick6/gv0 49174 0 Y 9733Brick storinator1:/export/brick7/gv0 49175 0 Y 16939Brick storinator2:/export/brick7/gv0 49175 0 Y 9739Brick storinator1:/export/brick8/gv0 49176 0 Y 16958Brick storinator2:/export/brick8/gv0 49176 0 Y 9703Brick storinator1:/export/brick9/gv0 49177 0 Y 16977Brick storinator2:/export/brick9/gv0 49177 0 Y 9713Brick storinator1:/export/brick10/gv0 49178 0 Y 16996Brick storinator2:/export/brick10/gv0 49178 0 Y 9718Brick storinator1:/export/brick11/gv0 49179 0 Y 17015Brick storinator2:/export/brick11/gv0 49179 0 Y 9746Brick storinator1:/export/brick12/gv0 49180 0 Y 17034Brick storinator2:/export/brick12/gv0 49180 0 Y 9792Brick storinator1:/export/brick13/gv0 49181 0 Y 17053Brick storinator2:/export/brick13/gv0 49181 0 Y 9755Brick storinator1:/export/brick14/gv0 49182 0 Y 17072Brick storinator2:/export/brick14/gv0 49182 0 Y 9767Brick storinator1:/export/brick15/gv0 49183 0 Y 17091Brick storinator2:/export/brick15/gv0 N/A N/A N N/ABrick storinator1:/export/brick16/gv0 49184 0 Y 17110Brick storinator2:/export/brick16/gv0 49184 0 Y 9791Brick storinator1:/export/brick17/gv0 49185 0 Y 17129Brick storinator2:/export/brick17/gv0 49185 0 Y 9756Brick storinator1:/export/brick18/gv0 49186 0 Y 17148Brick storinator2:/export/brick18/gv0 49186 0 Y 9766Brick storinator1:/export/brick19/gv0 49187 0 Y 17167Brick storinator2:/export/brick19/gv0 49187 0 Y 9745Brick storinator1:/export/brick20/gv0 49188 0 Y 17186Brick storinator2:/export/brick20/gv0 49188 0 Y 9783NFS Server on localhost 2049 0 Y 17206Self-heal Daemon on localhost N/A N/A Y 17214NFS Server on storinator2 2049 0 Y 9657Self-heal Daemon on storinator2 N/A N/A Y 9677Task Status of Volume gv0------------------------------------------------------------------------------Task : RebalanceID : 28c733e9-d618-44fc-873f-405d3b29a609Status : completed
I would like to remove (not replace) the failed brick (and its replica). (I don't have a spare disk handy, and there's plenty of room on the other bricks.) But gluster doesn't seem to want to remove a brick if the brick is dead:
[root@storinator1 ~]# gluster volume remove-brick gv0 storinator{1..2}:/export/brick15/gv0 startvolume remove-brick start: failed: Staging failed on storinator2. Error: Found stopped brick storinator2:/export/brick15/gv0
Thanks in advance for any help you can give.
phil
_______________________________________________ Gluster-users mailing list Gluster-users@xxxxxxxxxxx http://www.gluster.org/mailman/listinfo/gluster-users