On Wed, Mar 12, 2025 at 9:18 PM Janne Johansson <icepic.dz@xxxxxxxxx> wrote: > > Den ons 12 mars 2025 kl 11:41 skrev Kai Stian Olstad <ceph+list@xxxxxxxxxx>: > > > > Say we have 10 host with 10 OSDs and the failure domain is host. > > > > host0 osd 0 to 9 > > host1 osd 10 to 19 > > host2 osd 20 to 29 > > host3 osd 30 to 39 > > host4 osd 40 to 49 > > host5 osd 50 to 59 > > host6 osd 60 to 69 > > host7 osd 70 to 79 > > host8 osd 80 to 89 > > host9 osd 90 to 99 > > > > A pool have EC 4+2 and PG 2.1 has UP and ACTING the following OSDs > > 0,10,20,30,40 and 50. > > > > I need to take 3 of them, 0, 10 and 30, out, is it safe to run out on all 3 > > OSDs at the same time with "ceph osd out 0 10 20" or do I need to take one > > after the other out? > > It is not safe. You have no control of where ALL the PGs place their > pieces, so if you take out 3 OSDs immediately, by random chance, there > will be one PG for which osd 0,10 and 30 holds a piece for it, and now > EC can't piece together a 4+2 PG, since it only has 3 pieces, where it > needs 4 in order to rebuild the data. Or counted inversely, EC X+2 > means at the very most 2 drives can be lost before you lose data. It > will also turn READONLY while rebuilding the two missing PGs, and turn > RW again when at least one piece is recreated, but in terms of "can I > remove 3 drives on EC 4+2" the answer is no. > > What you can do is lower the weight of them to 0.0 and ceph will start > move PGs off them into the others, and when they have 0 PGs, you can > purge them and stop the OSD process and remove the disk. > > But you must let ceph move the data off the OSDs first. Please note that the original post talks about not yanking OSD drives physically, but about running the "ceph osd out" command. This command is exactly equivalent to reweighing the OSD to zero. -- Alexander Patrakov _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx