On Fri, Aug 17, 2018 at 9:48 AM David Turner <drakonstein@xxxxxxxxx> wrote:
> ceph-volume lvm create --osd-id 0 --bluestore --data /dev/sdc --block.db /dev/sdb --block.wal /dev/sdbThat command can't work... You're telling it to use the entire /dev/sdb device for the db and then again to do it for the wal, but you can only use the entire device once. There are 2 things wrong with that. First, if you're putting db and wal on the same device you do not need to specify the wal. Second if you are actually intending to use a partition on /dev/sdb instead of the entire block device for this single OSD, then you need to manually create a partition for it and supply that partition to the --block.db command.Likely the command you want will end up being this after you create a partition on the SSD for the db/wal.`ceph-volume lvm create --osd-id 0 --bluestore --data /dev/sdc --block.db /dev/sdb1`On Fri, Aug 17, 2018 at 10:24 AM Robert Stanford <rstanford8896@xxxxxxxxx> wrote:ThanksThat is the command context I've found on the web. Is it wrong?I was using the ceph-volume create command, which I understand combines the prepare and activate functions.
ceph-volume lvm create --osd-id 0 --bluestore --data /dev/sdc --block.db /dev/sdb --block.wal /dev/sdbR_______________________________________________On Fri, Aug 17, 2018 at 5:55 AM Alfredo Deza <adeza@xxxxxxxxxx> wrote:On Thu, Aug 16, 2018 at 9:00 PM, Robert Stanford
<rstanford8896@xxxxxxxxx> wrote:
>
> I am following the steps to my filestore journal with a bluestore journal
> (http://docs.ceph.com/docs/mimic/rados/operations/bluestore-migration/). It
> is broken at ceph-volume lvm create. Here is my error:
>
> --> Zapping successful for: /dev/sdc
> Preparing sdc
> Running command: /bin/ceph-authtool --gen-print-key
> Running command: /bin/ceph --cluster ceph --name client.bootstrap-osd
> --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring osd tree -f json
> Running command: /bin/ceph --cluster ceph --name client.bootstrap-osd
> --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring -i - osd new
> ff523216-350d-4ca0-9022-0c17662c2c3b 10
> Running command: vgcreate --force --yes
> ceph-459b4fbe-e3c4-4f28-b58e-3496bf3ea95a /dev/sdc
> stdout: Physical volume "/dev/sdc" successfully created.
> stdout: Volume group "ceph-459b4fbe-e3c4-4f28-b58e-3496bf3ea95a"
> successfully created
> Running command: lvcreate --yes -l 100%FREE -n
> osd-block-ff523216-350d-4ca0-9022-0c17662c2c3b
> ceph-459b4fbe-e3c4-4f28-b58e-3496bf3ea95a
> stdout: Logical volume "osd-block-ff523216-350d-4ca0-9022-0c17662c2c3b"
> created.
> --> blkid could not detect a PARTUUID for device: sdb
> --> Was unable to complete a new OSD, will rollback changes
> --> OSD will be destroyed, keeping the ID because it was provided with
> --osd-id
> Running command: ceph osd destroy osd.10 --yes-i-really-mean-it
> stderr: destroyed osd.10
> --> RuntimeError: unable to use device
>
> Note that SDB is the SSD journal. It has been zapped prior.
I can't see what the actual command you used is, but I am guessing you
did something like:
ceph-volume lvm prepare --filestore --data /dev/sdb --journal /dev/sdb
Which is not possible. There are a few ways you can do this (see:
http://docs.ceph.com/docs/master/ceph-volume/lvm/prepare/#filestore )
With a raw device and a pre-created partition (must have a PARTUUID):
ceph-volume lvm prepare --data /dev/sdb --journal /dev/sdc1
With LVs:
ceph-volume lvm prepare --data vg/my-data --journal vg/my-journal
With an LV for data and a partition:
ceph-volume lvm prepare --data vg/my-data --journal /dev/sdc1
>
> What is going wrong, and how can I fix it?
>
> Thank you
> R
>
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@xxxxxxxxxxxxxx
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com