rhev/gluster questions , possibly restoring from a failed node

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



In my lab, one of my RAID cards started acting up and took one of my three gluster nodes offline (two nodes with data and an arbiter node). I'm hoping it's simply the backplane, but during that time troubleshooting and waiting for parts, the hypervisors was fenced. Since the firewall was replaced and now several VMs are not starting correctly, fsck, scandisk and xfs_repair on the problematic VMs are running, but It's been 2 days and xfs_repair hasn't moved an inch. 

When the gluster volume has an offline brick, should it be slow? 
If it's suppose to be slow, does it make sense to remove the bricks and re-add them when the failed node is back?
Is the VM unresponsiveness an issue of the deteriorated gluster volume?

Lastly, in the event that the VMs are totally gone, I need to restore from the failed node, I was going to simply copy/scp the contents from the gluster directory onto the running nodes through glusterfs, is there a better way of restoring, overwriting the data on the good nodes?

Thanks,

Dan
_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
http://lists.gluster.org/mailman/listinfo/gluster-users

[Index of Archives]     [Gluster Development]     [Linux Filesytems Development]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux