Hi Ben,
we are not using EC pool on that cluster.
OSD out behavior almost stopped when we solved memory issues (less memory allocated to OSD's).
Now we are not working on that cluster anymore so we have no other info about that problem.
Jan
On 20/07/2020 07.59, Benoît Knecht wrote:
Hi Jan,
Jan Pekař wrote:
Also I'm concerned, that this OSD restart caused data degradation and recovery - cluster should be clean immediately after OSD up when no
client was uploading/modifying data during my tests.
We're experiencing the same thing on our 14.2.10 cluster. After marking an OSD out, if it's briefly marked down (due to the missed heartbeats or because the daemon was manually restarted) the PGs that were still mapped on it disappear all at once, and we get degraded objects as a result.
In our case, those PGs belong to an EC pool, and we use the PG balancer in upmap mode, so we have a few upmapped PGs on that OSD. Is that the case for you too?
We're going to run some tests to try and better understand what's going on there, but we welcome any feedback in the meantime.
Cheers,
--
Ben
--
============
Ing. Jan Pekař
jan.pekar@xxxxxxxxx
----
Imatic | Jagellonská 14 | Praha 3 | 130 00
http://www.imatic.cz | +420326555326
============
--
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx