Re: BlueStore upgrade steps broken

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 




 What's more, I was planning on using this single journal device (SSD) for 4 OSDs.  With filestore I simply told each OSD to use this drive, sdb, on the command line, and it would create a new partition on that drive every time I created an OSD.  I thought it would be the same for BlueStore.  So that begs the question, how does one set up an SSD to hold journals for multiple OSDs, both db and wal?  Searching has yielded nothing.

 R


On Fri, Aug 17, 2018 at 9:48 AM David Turner <drakonstein@xxxxxxxxx> wrote:
ceph-volume lvm create --osd-id 0 --bluestore --data /dev/sdc --block.db /dev/sdb --block.wal /dev/sdb

That command can't work... You're telling it to use the entire /dev/sdb device for the db and then again to do it for the wal, but you can only use the entire device once.  There are 2 things wrong with that.  First, if you're putting db and wal on the same device you do not need to specify the wal.  Second if you are actually intending to use a partition on /dev/sdb instead of the entire block device for this single OSD, then you need to manually create a partition for it and supply that partition to the --block.db command.

Likely the command you want will end up being this after you create a partition on the SSD for the db/wal.
`ceph-volume lvm create --osd-id 0 --bluestore --data /dev/sdc --block.db /dev/sdb1`

On Fri, Aug 17, 2018 at 10:24 AM Robert Stanford <rstanford8896@xxxxxxxxx> wrote:

 I was using the ceph-volume create command, which I understand combines the prepare and activate functions.

ceph-volume lvm create --osd-id 0 --bluestore --data /dev/sdc --block.db /dev/sdb --block.wal /dev/sdb

 That is the command context I've found on the web.  Is it wrong?

 Thanks
R

On Fri, Aug 17, 2018 at 5:55 AM Alfredo Deza <adeza@xxxxxxxxxx> wrote:
On Thu, Aug 16, 2018 at 9:00 PM, Robert Stanford
<rstanford8896@xxxxxxxxx> wrote:
>
>  I am following the steps to my filestore journal with a bluestore journal
> (http://docs.ceph.com/docs/mimic/rados/operations/bluestore-migration/).  It
> is broken at ceph-volume lvm create.  Here is my error:
>
> --> Zapping successful for: /dev/sdc
> Preparing sdc
> Running command: /bin/ceph-authtool --gen-print-key
> Running command: /bin/ceph --cluster ceph --name client.bootstrap-osd
> --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring osd tree -f json
> Running command: /bin/ceph --cluster ceph --name client.bootstrap-osd
> --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring -i - osd new
> ff523216-350d-4ca0-9022-0c17662c2c3b 10
> Running command: vgcreate --force --yes
> ceph-459b4fbe-e3c4-4f28-b58e-3496bf3ea95a /dev/sdc
>  stdout: Physical volume "/dev/sdc" successfully created.
>  stdout: Volume group "ceph-459b4fbe-e3c4-4f28-b58e-3496bf3ea95a"
> successfully created
> Running command: lvcreate --yes -l 100%FREE -n
> osd-block-ff523216-350d-4ca0-9022-0c17662c2c3b
> ceph-459b4fbe-e3c4-4f28-b58e-3496bf3ea95a
>  stdout: Logical volume "osd-block-ff523216-350d-4ca0-9022-0c17662c2c3b"
> created.
> --> blkid could not detect a PARTUUID for device: sdb
> --> Was unable to complete a new OSD, will rollback changes
> --> OSD will be destroyed, keeping the ID because it was provided with
> --osd-id
> Running command: ceph osd destroy osd.10 --yes-i-really-mean-it
>  stderr: destroyed osd.10
> -->  RuntimeError: unable to use device
>
>  Note that SDB is the SSD journal.  It has been zapped prior.

I can't see what the actual command you used is, but I am guessing you
did something like:

ceph-volume lvm prepare --filestore --data /dev/sdb --journal /dev/sdb

Which is not possible. There are a few ways you can do this (see:
http://docs.ceph.com/docs/master/ceph-volume/lvm/prepare/#filestore )

With a raw device and a pre-created partition (must have a PARTUUID):

    ceph-volume lvm prepare --data /dev/sdb --journal /dev/sdc1

With LVs:

    ceph-volume lvm prepare --data vg/my-data --journal vg/my-journal

With an LV for data and a partition:

    ceph-volume lvm prepare --data vg/my-data --journal /dev/sdc1

>
>  What is going wrong, and how can I fix it?
>
>  Thank you
>  R
>
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@xxxxxxxxxxxxxx
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux