Hi! I have fresh ceph cluster. 12 host and 3 osd on each host (one - hdd and two - ssd). Each host located in own rack. I make such crush configuration on fresh ceph installation: sudo ceph osd crush add-bucket R-26-3-1 rack sudo ceph osd crush add-bucket R-26-3-2 rack sudo ceph osd crush add-bucket R-26-4-1 rack sudo ceph osd crush add-bucket R-26-4-2 rack [...] sudo ceph osd crush add-bucket R-26-8-1 rack sudo ceph osd crush add-bucket R-26-8-2 rack sudo ceph osd crush move R-26-3-1 root=default [...] sudo ceph osd crush move R-26-8-2 root=default sudo ceph osd crush move S-26-3-1-1 rack=R-26-3-1 [...] sudo ceph osd crush move S-26-8-2-1 rack=R-26-8-2 sudo ceph osd crush rule create-replicated hddreplrule default rack hdd sudo ceph osd pool create rbd 256 256 replicated hddreplrule sudo ceph osd pool set rbd size 3 sudo ceph osd pool set rbd min_size 2 osd tree look like: ID CLASS WEIGHT TYPE NAME STATUS REWEIGHT PRI-AFF -1 117.36346 root default -2 9.78029 rack R-26-3-1 -27 9.78029 host S-26-3-1-1 0 hdd 9.32390 osd.0 up 1.00000 1.00000 1 ssd 0.22820 osd.1 up 1.00000 1.00000 2 ssd 0.22820 osd.2 up 1.00000 1.00000 -3 9.78029 rack R-26-3-2 -43 9.78029 host S-26-3-2-1 3 hdd 9.32390 osd.3 up 1.00000 1.00000 4 ssd 0.22820 osd.4 up 1.00000 1.00000 5 ssd 0.22820 osd.5 up 1.00000 1.00000 [...] Now write some data to rbd pool and shutdown one node. cluster: id: 9000d700-8529-4d38-b9f5-24d6079429a2 health: HEALTH_WARN 3 osds down 1 host (3 osds) down 1 rack (3 osds) down Degraded data redundancy: 1223/12300 objects degraded (9.943%), 74 pgs degraded, 74 pgs undersized And ceph does not try to repair pool. Why? WBR, Fyodor. _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com