Thanks again Eugen. Looking at my command history it does look like I did execute the migration but didn't commit it. I wasn't surprised to see it in the trash based on the doc you mentioned, I only tried the restore as a desperate measure to clean up my mess. It doesn't help that I messed around like this, including with hexedit :O. I should have reached out before messing around. I'll proceed with the migrate/re-create and report back. I'm just crossing my fingers that I'll be allowed to delete the pool. It's a lesson to me to take more care of my wee cluster. Cheers, Rhys ------- Original Message ------- On Wednesday, October 11th, 2023 at 7:54 AM, Eugen Block <eblock@xxxxxx> wrote: > Hi, > > I just re-read the docs on rbd migration [1], haven't done that in a > while, and it states the following: > > > Note that the source image will be moved to the RBD trash to avoid > > mistaken usage during the migration process > > > So it was expected that your source image was in the trash during the > migration, no need to restore. According to your history you also ran > the "execute" command, do you remember if ran successfully as well? > Did you "execute" after the prepare command completed? But you also > state that the target image isn't there anymore, so it's hard to tell > what exactly happened here. I'm not sure how to continue from here, > maybe migrating/re-creating is the only way now. > > [1] https://docs.ceph.com/en/quincy/rbd/rbd-live-migration/ > > Zitat von Rhys Goodwin rhys.goodwin@xxxxxxxxx: > > > Thanks Eugen. > > > > root@hcn03:~# rbd status infra-pool/sophosbuild > > 2023-10-10T09:44:21.234+0000 7f1675c524c0 -1 librbd::Migration: > > open_images: failed to open destination image images/65d188c5f5a34: > > (2) No such file or directory > > rbd: getting migration status failed: (2) No such file or directory > > Watchers: none > > > > I've checked over the other pools again, but they only contain > > Openstack images. There are only 42 images in total across all > > pools. In fact, the "infra-pool" pool only has 3 images, including > > the faulty one. So migrating/re-creating is not a big deal. It's > > more just that I'd like to learn more about how to resolve such > > issues, if possible. > > > > Good call on the history. I found this smoking gun with: 'history > > |grep "rbd migration": > > rbd migration prepare infra-pool/sophosbuild images/sophosbuild > > rbd migration execute images/sophosbuild > > > > But images/sophosbuild is definitely not there anymore, and not in > > the trash. It looks like I was missing the commit. > > > > Kind regards, > > Rhys > > > > ------- Original Message ------- > > > > Eugen Block Wrote: > > > > Hi, there are a couple of things I would check before migrating all > > images. What's the current 'rbd status infra-pool/sophosbuild'? You > > probably don't have an infinite number of pools so I would also > > check if any of the other pools contains an image with the same > > name, just in case you wanted to keep its original name and only > > change the pool. Even if you don't have the terminal output, maybe > > you find some of the commands in the history? > > _______________________________________________ > > ceph-users mailing list -- ceph-users@xxxxxxx > > To unsubscribe send an email to ceph-users-leave@xxxxxxx > > > > _______________________________________________ > ceph-users mailing list -- ceph-users@xxxxxxx > To unsubscribe send an email to ceph-users-leave@xxxxxxx _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx