Hi,
how many OSDs do you have in total? Can you share your osd tree, please?
You could check the unit.meta file on each OSD host to see which
service it refers to and simply change it according to the service you
intend to keep:
host1:~ # grep -r service_name
/var/lib/ceph/543967bc-e586-32b8-bd2c-2d8b8b168f02/osd*
/var/lib/ceph/543967bc-e586-32b8-bd2c-2d8b8b168f02/osd.14/unit.meta:
"service_name": "osd.osd-hdd-ssd-mix",
/var/lib/ceph/543967bc-e586-32b8-bd2c-2d8b8b168f02/osd.15/unit.meta:
"service_name": "osd.osd-hdd-ssd-mix",
host1:~# ceph orch ls osd
NAME PORTS RUNNING REFRESHED AGE PLACEMENT
osd.osd-hdd-ssd-mix 16 3m ago 2w
host1;host2;host3;host4;host5;host6;host7;host8
After you restart the OSDs they should show up correctly in the orch
ls output, then you should be able to remove the unused specs. I'm
just not sure if that works the same way in Octopus, it's already EOL
and I don't have an Octopus cluster at hand to verify. :-)
Regards,
Eugen
Zitat von Jeffrey Turmelle <jefft@xxxxxxxxxxxxxxxx>:
Running on Octopus:
While attempting to install a bunch of new OSDs on multiple hosts, I
ran some ceph orchestrator commands to install them, such as
ceph orch apply osd --all-available-devices
ceph orch apply osd -I HDD_drive_group.yaml
I assumed these were just helper processes, and they would be
short-lived. In fact, they didn’t actually work and I ended up
installing each drive by hand like this:
ceph orch daemon add osd ceph4.iri.columbia.edu:/dev/sdag
However, now I have these services running:
# ceph orch ls --service-type=osd
NAME RUNNING REFRESHED AGE PLACEMENT
IMAGE NAME IMAGE ID
osd.HDD_drive_group 2/2 7m ago 6w
ceph[456].iri.columbia.edu docker.io/ceph/ceph:v15 2cf504fded39
osd.None 54/0 7m ago - <unmanaged>
docker.io/ceph/ceph:v15 2cf504fded39
osd.all-available-devices 1/0 7m ago - <unmanaged>
docker.io/ceph/ceph:v15 2cf504fded39
I’m certain none of these actually created any of my running OSD
daemons, but I’m not sure if it’s ok to remove them.
For example:
ceph orch daemon rm osd.all-available-devices
ceph orch daemon rm osd.HDD_drive_group
ceph orch daemon rm osd.None
Does anyone have any insight to this? I can just leave them there,
they don’t seem to be doing anything, but on the other hand, I don’t
want any new devices to be automatically loaded or any other
unintended consequences of these.
Thanks for any guidance,
Jeff Turmelle
International Research Institute for Climate & Society
<https://iri.columbia.edu/>
The Climate School <https://climate.columbia.edu/> at Columbia
University <https://columbia.edu/>
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx