Re: self heal failed, on /

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 





On 02/24/2017 01:25 PM, max.degraaf@xxxxxxx wrote:

Any advice on the sequence of updating? Server of client first?

 

I assume it’s a simple in-place update where configuration is preserved. Right?


Configuration will be preserved. I don't know actual procedure for rolling upgrade (other than servers first with one after other). May be somebody else in this ML can provide you more info or just a google search will find you some blogs.


Regards
Rafi KC


 

 

Van: Mohammed Rafi K C [mailto:rkavunga@xxxxxxxxxx]
Verzonden: vrijdag 24 februari 2017 08:49
Aan: Graaf, Max de; gluster-users@xxxxxxxxxxx
Onderwerp: Re: self heal failed, on /

 

 

 

On 02/24/2017 11:47 AM, max.degraaf@xxxxxxx wrote:

The version on the server of this specific mount is 3.7.11. The client is running version 3.4.2.


It is always better to have everything in one version, all clients and all servers. In this case there is huge gap between the versions, 3.7 and 3.4 .

An additional thing is the code running on 3.4 is replicaV1 code and on 3.7 it v2, meaning there is huge difference to the logic of replication/healing. So I recommend to keep all the gluster instances to the same version


~Rafi



 

There is more to that. This client is actually mounting to volumes where the other server is running 3.4.2 as well. What’s your advice, update that other server to 3.7.11 (or higher) first? Of start with the client update?

 

Van: Mohammed Rafi K C [mailto:rkavunga@xxxxxxxxxx]
Verzonden: vrijdag 24 februari 2017 07:02
Aan: Graaf, Max de; gluster-users@xxxxxxxxxxx
Onderwerp: Re: self heal failed, on /

 

 

 

On 02/23/2017 12:18 PM, max.degraaf@xxxxxxx wrote:

Hi,

 

We have a 4 node glusterfs setup that seems to be running without any problems. We can’t find any problems with replication or whatever.

 

We also have 4 machines running the glusterfs client. On all 4 machines we see the following error in the logs at random moments:

 

[2017-02-23 00:04:33.168778] I [afr-self-heal-common.c:2869:afr_log_self_heal_completion_status] 0-aab-replicate-0:  metadata self heal  is successfully completed,   metadata self heal from source aab-client-0 to aab-client-1,  aab-client-2,  aab-client-3,  metadata - Pending matrix:  [ [ 0 0 0 0 ] [ 0 0 0 0 ] [ 0 0 0 0 ] [ 0 0 0 0 ] ], on /

[2017-02-23 00:09:34.431089] E [afr-self-heal-common.c:2869:afr_log_self_heal_completion_status] 0-aab-replicate-0:  metadata self heal  failed,   on /

[2017-02-23 00:14:34.948975] I [afr-self-heal-common.c:2869:afr_log_self_heal_completion_status] 0-aab-replicate-0:  metadata self heal  is successfully completed,   metadata self heal from source aab-client-0 to aab-client-1,  aab-client-2,  aab-client-3,  metadata - Pending matrix:  [ [ 0 0 0 0 ] [ 0 0 0 0 ] [ 0 0 0 0 ] [ 0 0 0 0 ] ], on /

 

The content within the glusterfs filesystems is rather static with only minor changes on it. This “self heal  failed” is printed randomly in the logs on the glusterfs client. It’s printed even at moment where nothing has changed within the glusterfs filesystem. When it is printed, its never on multiple servers at the same time. What we also don’t understand : the error indicates self heal failed on root “/”. In the root of this glusterfs mount there only 2 folders and no files are ever written at the root level.

 

Any thoughts?


From the logs, It looks like an older version of gluster , probably 3.5 . Please confirm your glusterfs version. The version is pretty old and it may be moved End of Life. And this is AFR v1 , where the latest stable version runs with AFRv2.

So I would suggest you to upgrade to a later version may be 3.8 .

If you still want to go with this version, I can give it a try. Let me know the version, volume info and volume status. Still I will suggest to upgrade ;)


Regards
Rafi KC





 





_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
http://lists.gluster.org/mailman/listinfo/gluster-users

 

 


_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
http://lists.gluster.org/mailman/listinfo/gluster-users

[Index of Archives]     [Gluster Development]     [Linux Filesytems Development]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux