Hi all,
We've 8 osd hosts, 4 in room 1 and 4 in room2.
A pool with size = 3 using following crush map is created, to cater for room failure.
1.for each object, there are always 2 replicas in one room and 1 replica in other room making size=3. But we can't control which room has 1 or 2 replicas.
2.in case an osd host fails, ceph will assign remaining osds to the same PG to hold replicas on the failed osd host. Selection is based on crush rule of the pool, thus maintaining the same failure domain - won't make all replicas in the same room.
3.in case of entire room with 1 replica fails, the pool will remain degraded but won't do any replica relocation.
4. in case of entire room with 2 replicas fails, ceph will make use of osds in the surviving room and making 2 replicas. Pool will not be writeable before all objects are made 2 copies (unless we make pool size=4?). Then when recovery is complete, pool
will remain in degraded state until the failed room recover.
Is our understanding correct? Thanks a lot.
Will do some simulation later to verify.
Regards,
/stwong
|
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com