On 11/16/2017 12:54 PM, Daniel Berteaud
wrote:
Le 15/11/2017 à 09:45, Ravishankar N a écrit :
If it is only the brick that is faulty on the bad node, but
everything else is fine, like glusterd running, the node being a
part of the trusted storage pool etc, you could just kill the
brick first and do step-13 in "10.6.2. Replacing a Host Machine
with the Same Hostname", (the mkdir of non-existent dir,
followed by setfattr of non-existent key) of https://access.redhat.com/documentation/en-US/Red_Hat_Storage/3.1/pdf/Administration_Guide/Red_Hat_Storage-3.1-Administration_Guide-en-US.pdf,
then restart the brick by restarting glusterd on that node. Read
10.5 and 10.6 sections in the doc to get a better understanding
of replacing bricks.
Thanks, I'll try that.
Any way in this situation to check which file will be healed from
which brick before reconnecting ? Using some getfattr tricks ?
Yes, there are afr xattrs that determine the heal direction for each
file. The good copy will have non-zero trusted.afr* xattrs that
blame the bad one and heal will happen from good to bad. If both
bricks have attrs blaming the other, then the file is in
split-brain.
-Ravi
Regards, Daniel
|
_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
http://lists.gluster.org/mailman/listinfo/gluster-users