> ceph-volume lvm create --osd-id 0 --bluestore --data /dev/sdc --block.db /dev/sdb --block.wal /dev/sdb
Likely the command you want will end up being this after you create a partition on the SSD for the db/wal.
`ceph-volume lvm create --osd-id 0 --bluestore --data /dev/sdc --block.db /dev/sdb1`
On Fri, Aug 17, 2018 at 10:24 AM Robert Stanford <rstanford8896@xxxxxxxxx> wrote:
ThanksThat is the command context I've found on the web. Is it wrong?I was using the ceph-volume create command, which I understand combines the prepare and activate functions.
ceph-volume lvm create --osd-id 0 --bluestore --data /dev/sdc --block.db /dev/sdb --block.wal /dev/sdbR_______________________________________________On Fri, Aug 17, 2018 at 5:55 AM Alfredo Deza <adeza@xxxxxxxxxx> wrote:On Thu, Aug 16, 2018 at 9:00 PM, Robert Stanford
<rstanford8896@xxxxxxxxx> wrote:
>
> I am following the steps to my filestore journal with a bluestore journal
> (http://docs.ceph.com/docs/mimic/rados/operations/bluestore-migration/). It
> is broken at ceph-volume lvm create. Here is my error:
>
> --> Zapping successful for: /dev/sdc
> Preparing sdc
> Running command: /bin/ceph-authtool --gen-print-key
> Running command: /bin/ceph --cluster ceph --name client.bootstrap-osd
> --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring osd tree -f json
> Running command: /bin/ceph --cluster ceph --name client.bootstrap-osd
> --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring -i - osd new
> ff523216-350d-4ca0-9022-0c17662c2c3b 10
> Running command: vgcreate --force --yes
> ceph-459b4fbe-e3c4-4f28-b58e-3496bf3ea95a /dev/sdc
> stdout: Physical volume "/dev/sdc" successfully created.
> stdout: Volume group "ceph-459b4fbe-e3c4-4f28-b58e-3496bf3ea95a"
> successfully created
> Running command: lvcreate --yes -l 100%FREE -n
> osd-block-ff523216-350d-4ca0-9022-0c17662c2c3b
> ceph-459b4fbe-e3c4-4f28-b58e-3496bf3ea95a
> stdout: Logical volume "osd-block-ff523216-350d-4ca0-9022-0c17662c2c3b"
> created.
> --> blkid could not detect a PARTUUID for device: sdb
> --> Was unable to complete a new OSD, will rollback changes
> --> OSD will be destroyed, keeping the ID because it was provided with
> --osd-id
> Running command: ceph osd destroy osd.10 --yes-i-really-mean-it
> stderr: destroyed osd.10
> --> RuntimeError: unable to use device
>
> Note that SDB is the SSD journal. It has been zapped prior.
I can't see what the actual command you used is, but I am guessing you
did something like:
ceph-volume lvm prepare --filestore --data /dev/sdb --journal /dev/sdb
Which is not possible. There are a few ways you can do this (see:
http://docs.ceph.com/docs/master/ceph-volume/lvm/prepare/#filestore )
With a raw device and a pre-created partition (must have a PARTUUID):
ceph-volume lvm prepare --data /dev/sdb --journal /dev/sdc1
With LVs:
ceph-volume lvm prepare --data vg/my-data --journal vg/my-journal
With an LV for data and a partition:
ceph-volume lvm prepare --data vg/my-data --journal /dev/sdc1
>
> What is going wrong, and how can I fix it?
>
> Thank you
> R
>
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@xxxxxxxxxxxxxx
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com