Hello Stefan, The status was "HEALTH_OK" before i ran those commands. I removed the OSDs with: ceph osd out osd.10 ceph auth del osd.10 systemctl stop ceph-osd@10 ceph osd rm 10 umount /var/lib/ceph/osd/ceph-10 ceph osd crush remove osd.10 dd if=/dev/zero of=/dev/sdc ceph osd out osd.9 ceph auth del osd.9 systemctl stop ceph-osd@9 ceph osd rm 9 umount /var/lib/ceph/osd/ceph-9 ceph osd crush remove osd.9 root@ceph01:~# ceph osd crush rule dump [ { "rule_id": 0, "rule_name": "replicated_ruleset", "ruleset": 0, "type": 1, "min_size": 1, "max_size": 10, "steps": [ { "op": "take", "item": -1, "item_name": "default" }, { "op": "chooseleaf_firstn", "num": 0, "type": "host" }, { "op": "emit" } ] } ] Thanks, Mario On Sun, Dec 29, 2019 at 2:16 PM Stefan Kooman <stefan@xxxxxx> wrote: > > Quoting Ml Ml (mliebherr99@xxxxxxxxxxxxxx): > > Hello List, > > i have size = 3 and min_size = 2 with 3 Nodes. > > That's good. > > > > > > > I replaced two osds on node ceph01 and ran into "HEALTH_ERR". > > My problem: it waits for the backfilling process? > > Why did i run into HEALTH_ERR? I thought all data will be available on > > at least one more node. or even two: > > How did you replace them? Did you first set them "out" and waited for > the data to be repicated elsewhere before you removed them? > > It *might* be because your CRUSH rule set is replicating over "OSD" and > not host. What does a "ceph osd crush rule dump" shows? > > Gr. Stefan > > -- > | BIT BV https://www.bit.nl/ Kamer van Koophandel 09090351 > | GPG: 0xD14839C6 +31 318 648 688 / info@xxxxxx _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com