On Thu, Aug 16, 2018 at 9:00 PM, Robert Stanford <rstanford8896@xxxxxxxxx> wrote: > > I am following the steps to my filestore journal with a bluestore journal > (http://docs.ceph.com/docs/mimic/rados/operations/bluestore-migration/). It > is broken at ceph-volume lvm create. Here is my error: > > --> Zapping successful for: /dev/sdc > Preparing sdc > Running command: /bin/ceph-authtool --gen-print-key > Running command: /bin/ceph --cluster ceph --name client.bootstrap-osd > --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring osd tree -f json > Running command: /bin/ceph --cluster ceph --name client.bootstrap-osd > --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring -i - osd new > ff523216-350d-4ca0-9022-0c17662c2c3b 10 > Running command: vgcreate --force --yes > ceph-459b4fbe-e3c4-4f28-b58e-3496bf3ea95a /dev/sdc > stdout: Physical volume "/dev/sdc" successfully created. > stdout: Volume group "ceph-459b4fbe-e3c4-4f28-b58e-3496bf3ea95a" > successfully created > Running command: lvcreate --yes -l 100%FREE -n > osd-block-ff523216-350d-4ca0-9022-0c17662c2c3b > ceph-459b4fbe-e3c4-4f28-b58e-3496bf3ea95a > stdout: Logical volume "osd-block-ff523216-350d-4ca0-9022-0c17662c2c3b" > created. > --> blkid could not detect a PARTUUID for device: sdb > --> Was unable to complete a new OSD, will rollback changes > --> OSD will be destroyed, keeping the ID because it was provided with > --osd-id > Running command: ceph osd destroy osd.10 --yes-i-really-mean-it > stderr: destroyed osd.10 > --> RuntimeError: unable to use device > > Note that SDB is the SSD journal. It has been zapped prior. I can't see what the actual command you used is, but I am guessing you did something like: ceph-volume lvm prepare --filestore --data /dev/sdb --journal /dev/sdb Which is not possible. There are a few ways you can do this (see: http://docs.ceph.com/docs/master/ceph-volume/lvm/prepare/#filestore ) With a raw device and a pre-created partition (must have a PARTUUID): ceph-volume lvm prepare --data /dev/sdb --journal /dev/sdc1 With LVs: ceph-volume lvm prepare --data vg/my-data --journal vg/my-journal With an LV for data and a partition: ceph-volume lvm prepare --data vg/my-data --journal /dev/sdc1 > > What is going wrong, and how can I fix it? > > Thank you > R > > > _______________________________________________ > ceph-users mailing list > ceph-users@xxxxxxxxxxxxxx > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com