The version on the server of this specific mount is 3.7.11. The client is running version 3.4.2.
There is more to that. This client is actually mounting to volumes where the other server is running 3.4.2 as well. What’s your advice, update that other server to 3.7.11 (or higher) first? Of start
with the client update?
Van: Mohammed
Rafi K C [mailto:rkavunga@xxxxxxxxxx]
Verzonden: vrijdag 24 februari 2017 07:02
Aan: Graaf, Max de; gluster-users@xxxxxxxxxxx
Onderwerp: Re: self heal failed, on /
Hi,
We have a 4 node glusterfs setup that seems to be running without any problems. We can’t find any problems with replication or whatever.
We also have 4 machines running the glusterfs client. On all 4 machines we see the following error in the logs at random moments:
[2017-02-23 00:04:33.168778] I [afr-self-heal-common.c:2869:afr_log_self_heal_completion_status] 0-aab-replicate-0: metadata self heal is successfully completed, metadata self heal from source aab-client-0 to aab-client-1,
aab-client-2, aab-client-3, metadata - Pending matrix: [ [ 0 0 0 0 ] [ 0 0 0 0 ] [ 0 0 0 0 ] [ 0 0 0 0 ] ], on /
[2017-02-23 00:09:34.431089] E [afr-self-heal-common.c:2869:afr_log_self_heal_completion_status] 0-aab-replicate-0: metadata self heal failed, on /
[2017-02-23 00:14:34.948975] I [afr-self-heal-common.c:2869:afr_log_self_heal_completion_status] 0-aab-replicate-0: metadata self heal is successfully completed, metadata self heal from source aab-client-0 to aab-client-1,
aab-client-2, aab-client-3, metadata - Pending matrix: [ [ 0 0 0 0 ] [ 0 0 0 0 ] [ 0 0 0 0 ] [ 0 0 0 0 ] ], on /
The content within the glusterfs filesystems is rather static with only minor changes on it. This “self heal failed” is printed randomly in the logs on the glusterfs client. It’s printed even at moment where nothing
has changed within the glusterfs filesystem. When it is printed, its never on multiple servers at the same time. What we also don’t understand : the error indicates self heal failed on root “/”. In the root of this glusterfs mount there only 2 folders and
no files are ever written at the root level.
Any thoughts?
>From the logs, It looks like an older version of gluster , probably 3.5 . Please confirm your glusterfs version. The version is pretty old and it may be moved End of Life. And this is AFR v1 , where the latest stable version runs with AFRv2.
So I would suggest you to upgrade to a later version may be 3.8 .
If you still want to go with this version, I can give it a try. Let me know the version, volume info and volume status. Still I will suggest to upgrade ;)
Regards
Rafi KC
_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
http://lists.gluster.org/mailman/listinfo/gluster-users