On Thu, Apr 18, 2019 at 3:01 PM Sergei Genchev <sgenchev@xxxxxxxxx> wrote: > > Thank you Alfredo > I did not have any reasons to keep volumes around. > I tried using ceph-volume to zap these stores, but none of the command worked, including yours 'ceph-volume lvm zap osvg-sdd-db/2ukzAx-g9pZ-IyxU-Sp9h-fHv2-INNY-1vTpvz' If you do not want to keep them around you would need to use --destroy and use the lv path as input: ceph-volume lvm zap --destroy osvg-sdd-db/2ukzAx-g9pZ-IyxU-Sp9h-fHv2-INNY-1vTpvz > > I ended up manually removing LUKS volumes and then deleting LVM LV, VG, and PV > > cryptsetup remove /dev/mapper/AeV0iG-odWF-NRPE-1bVK-0mxH-OgHL-fneTzr > cryptsetup remove /dev/mapper/2ukzAx-g9pZ-IyxU-Sp9h-fHv2-INNY-1vTpvz > lvremove /dev/ceph-f4efa78f-a467-4214-b550-81653da1c9bd/osd-block-097d59be-bbe6-493a-b785-48b259d2ff35 > sgdisk -Z /dev/sdd > > # ceph-volume lvm zap osvg-sdd-db/2ukzAx-g9pZ-IyxU-Sp9h-fHv2-INNY-1vTpvz > Running command: /usr/sbin/cryptsetup status /dev/mapper/ > --> Zapping: osvg-sdd-db/2ukzAx-g9pZ-IyxU-Sp9h-fHv2-INNY-1vTpvz > Running command: /usr/sbin/wipefs --all osvg-sdd-db/2ukzAx-g9pZ-IyxU-Sp9h-fHv2-INNY-1vTpvz > stderr: wipefs: error: osvg-sdd-db/2ukzAx-g9pZ-IyxU-Sp9h-fHv2-INNY-1vTpvz: probing initialization failed: No such file or directory > --> RuntimeError: command returned non-zero exit status: 1 In this case, you removed the LV so the wipefs failed because that LV no longer exists. Do you have output on how it failed before? > > > On Thu, Apr 18, 2019 at 10:10 AM Alfredo Deza <adeza@xxxxxxxxxx> wrote: >> >> On Thu, Apr 18, 2019 at 10:55 AM Sergei Genchev <sgenchev@xxxxxxxxx> wrote: >> > >> > Hello, >> > I have a server with 18 disks, and 17 OSD daemons configured. One of the OSD daemons failed to deploy with ceph-deploy. The reason for failing is unimportant at this point, I believe it was race condition, as I was running ceph-deploy inside while loop for all disks in this server. >> > Now I have two left over LVM dmcrypded volumes that I am not sure how clean up. The command that failed and did not quite clean up after itself was: >> > ceph-deploy osd create --bluestore --dmcrypt --data /dev/sdd --block-db osvg/sdd-db ${SERVERNAME} >> > >> > # lsblk >> > ....... >> > sdd 8:48 0 7.3T 0 disk >> > └─ceph--f4efa78f--a467--4214--b550--81653da1c9bd-osd--block--097d59be--bbe6--493a--b785--48b259d2ff35 >> > 253:32 0 7.3T 0 lvm >> > └─AeV0iG-odWF-NRPE-1bVK-0mxH-OgHL-fneTzr 253:33 0 7.3T 0 crypt >> > >> > sds 65:32 0 223.5G 0 disk >> > ├─sds1 65:33 0 512M 0 part /boot >> > └─sds2 65:34 0 223G 0 part >> > ....... >> > ├─osvg-sdd--db 253:8 0 8G 0 lvm >> > │ └─2ukzAx-g9pZ-IyxU-Sp9h-fHv2-INNY-1vTpvz 253:34 0 8G 0 crypt >> > >> > # ceph-volume inventory /dev/sdd >> > >> > ====== Device report /dev/sdd ====== >> > >> > available False >> > rejected reasons locked >> > path /dev/sdd >> > scheduler mode deadline >> > rotational 1 >> > vendor SEAGATE >> > human readable size 7.28 TB >> > sas address 0x5000c500a6b1d581 >> > removable 0 >> > model ST8000NM0185 >> > ro 0 >> > --- Logical Volume --- >> > cluster name ceph >> > name osd-block-097d59be-bbe6-493a-b785-48b259d2ff35 >> > osd id 39 >> > cluster fsid 8e7a3953-7647-4133-9b9a-7f4a2e2b7da7 >> > type block >> > block uuid AeV0iG-odWF-NRPE-1bVK-0mxH-OgHL-fneTzr >> > osd fsid 097d59be-bbe6-493a-b785-48b259d2ff35 >> > >> > I was trying to run >> > ceph-volume lvm zap --destroy /dev/sdd but it errored out. Osd id on this volume is the same as on next drive, /dev/sde, and osd.39 daemon is running. This command was trying to zap running osd. >> > >> > What is the proper way to clean both data and block db volumes, so I can rerun ceph-deploy again, and add them to the pool? >> > >> >> Do you want to keep the LVs around or you want to complete get rid of >> them? If you are passing /dev/sdd to 'zap' you are telling the tool to >> destroy everything that is in there, regardless of who owns it >> (including running >> OSDs). >> >> If you want to keep LVs around then you can omit the --destroy flag >> and pass the LVs as input, or if using a recent enough version you can >> use --osd-fsid to zap: >> >> ceph-volume lvm zap osvg-sdd-db/2ukzAx-g9pZ-IyxU-Sp9h-fHv2-INNY-1vTpvz >> >> If you don't want the LVs around you can add --destroy, but use the LV >> as input (not the device) >> >> > Thank you! >> > >> > >> > >> > _______________________________________________ >> > ceph-users mailing list >> > ceph-users@xxxxxxxxxxxxxx >> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com