On Fri, Aug 17, 2018 at 10:24 AM, Robert Stanford <rstanford8896@xxxxxxxxx> wrote: > > I was using the ceph-volume create command, which I understand combines the > prepare and activate functions. > > ceph-volume lvm create --osd-id 0 --bluestore --data /dev/sdc --block.db > /dev/sdb --block.wal /dev/sdb > > That is the command context I've found on the web. Is it wrong? It is very wrong :( If this was coming from our docs, it needs to be fixed because it will never work. If you really want to place both block.db and block.wal on /dev/sdb, you will need to create one LV for each. ceph-volume will not do this for you. And then you can pass those newly created LVs like: ceph-volume lvm create --osd-id 0 --bluestore --data /dev/sdc --block.db sdb-vg/block-lv --block.wal sdb-vg/wal-lv > > Thanks > R > > On Fri, Aug 17, 2018 at 5:55 AM Alfredo Deza <adeza@xxxxxxxxxx> wrote: >> >> On Thu, Aug 16, 2018 at 9:00 PM, Robert Stanford >> <rstanford8896@xxxxxxxxx> wrote: >> > >> > I am following the steps to my filestore journal with a bluestore >> > journal >> > (http://docs.ceph.com/docs/mimic/rados/operations/bluestore-migration/). >> > It >> > is broken at ceph-volume lvm create. Here is my error: >> > >> > --> Zapping successful for: /dev/sdc >> > Preparing sdc >> > Running command: /bin/ceph-authtool --gen-print-key >> > Running command: /bin/ceph --cluster ceph --name client.bootstrap-osd >> > --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring osd tree -f json >> > Running command: /bin/ceph --cluster ceph --name client.bootstrap-osd >> > --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring -i - osd new >> > ff523216-350d-4ca0-9022-0c17662c2c3b 10 >> > Running command: vgcreate --force --yes >> > ceph-459b4fbe-e3c4-4f28-b58e-3496bf3ea95a /dev/sdc >> > stdout: Physical volume "/dev/sdc" successfully created. >> > stdout: Volume group "ceph-459b4fbe-e3c4-4f28-b58e-3496bf3ea95a" >> > successfully created >> > Running command: lvcreate --yes -l 100%FREE -n >> > osd-block-ff523216-350d-4ca0-9022-0c17662c2c3b >> > ceph-459b4fbe-e3c4-4f28-b58e-3496bf3ea95a >> > stdout: Logical volume "osd-block-ff523216-350d-4ca0-9022-0c17662c2c3b" >> > created. >> > --> blkid could not detect a PARTUUID for device: sdb >> > --> Was unable to complete a new OSD, will rollback changes >> > --> OSD will be destroyed, keeping the ID because it was provided with >> > --osd-id >> > Running command: ceph osd destroy osd.10 --yes-i-really-mean-it >> > stderr: destroyed osd.10 >> > --> RuntimeError: unable to use device >> > >> > Note that SDB is the SSD journal. It has been zapped prior. >> >> I can't see what the actual command you used is, but I am guessing you >> did something like: >> >> ceph-volume lvm prepare --filestore --data /dev/sdb --journal /dev/sdb >> >> Which is not possible. There are a few ways you can do this (see: >> http://docs.ceph.com/docs/master/ceph-volume/lvm/prepare/#filestore ) >> >> With a raw device and a pre-created partition (must have a PARTUUID): >> >> ceph-volume lvm prepare --data /dev/sdb --journal /dev/sdc1 >> >> With LVs: >> >> ceph-volume lvm prepare --data vg/my-data --journal vg/my-journal >> >> With an LV for data and a partition: >> >> ceph-volume lvm prepare --data vg/my-data --journal /dev/sdc1 >> >> > >> > What is going wrong, and how can I fix it? >> > >> > Thank you >> > R >> > >> > >> > _______________________________________________ >> > ceph-users mailing list >> > ceph-users@xxxxxxxxxxxxxx >> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >> > _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com