I am slightly confused you say you have image file corruption but then you say the qemu-img check says there is no corruption. If what you mean is that you see I/O errors during a heal this is likely to be due to io starvation, something that is a well know issue.
There is work happening to improve this in version 3.8:
On 19 May 2016 at 09:58, Kevin Lemonnier <lemonnierk@xxxxxxxxx> wrote:
That's a different problem then, I have corruption without removing or adding bricks,
as mentionned. Might be two separate issue
On Thu, May 19, 2016 at 11:25:34PM +1000, Lindsay Mathieson wrote:
> On 19/05/2016 12:17 AM, Lindsay Mathieson wrote:
>
> One thought - since the VM's are active while the brick is
> removed/re-added, could it be the shards that are written while the
> brick is added that are the reverse healing shards?
>
> I tested by:
>
> - removing brick 3
>
> - erasing brick 3
>
> - closing down all VM's
>
> - adding new brick 3
>
> - waiting until heal number reached its max and started decreasing
>
> There were no reverse heals
>
> - Started the VM's backup. No real issues there though one showed IO
> errors, presumably due to shards being locked as they were healed.
>
> - VM's started ok, no reverse heals were noted and eventually Brick 3 was
> fully healed. The VM's do not appear to be corrupted.
>
> So it would appear the problem is adding a brick while the volume is being
> written to.
>
> Cheers,
>
> --
> Lindsay Mathieson
> _______________________________________________
> Gluster-users mailing list
> Gluster-users@xxxxxxxxxxx
> http://www.gluster.org/mailman/listinfo/gluster-users
--
Kevin Lemonnier
PGP Fingerprint : 89A5 2283 04A0 E6E9 0111
_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
http://www.gluster.org/mailman/listinfo/gluster-users
_______________________________________________ Gluster-users mailing list Gluster-users@xxxxxxxxxxx http://www.gluster.org/mailman/listinfo/gluster-users