Re: BlueStore upgrade steps broken

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



The WAL will choose the fastest device available.  If you do not specify the WAL or DB, both will reside on the data disk.  If you specify the DB, then the WAL will reside with the DB.  If you specify WAL without DB, the DB will be on the data disk.  This data and more is included at the top of the URL page that Alfredo linked you to.  The docs have a lot of good information like this in the actual text between sections of commands and config options.  I often forget to read those sections myself, but most of my questions are usually answered there.

On Fri, Aug 17, 2018 at 12:51 PM Robert Stanford <rstanford8896@xxxxxxxxx> wrote:

 This is helpful, thanks.  Since the example is only for block.db, does that imply that the wal should (can efficiently) live on the same disk as data?

 R

On Fri, Aug 17, 2018 at 10:50 AM Alfredo Deza <adeza@xxxxxxxxxx> wrote:
On Fri, Aug 17, 2018 at 11:47 AM, Robert Stanford
<rstanford8896@xxxxxxxxx> wrote:
>
>  What's more, I was planning on using this single journal device (SSD) for 4
> OSDs.  With filestore I simply told each OSD to use this drive, sdb, on the
> command line, and it would create a new partition on that drive every time I
> created an OSD.  I thought it would be the same for BlueStore.  So that begs
> the question, how does one set up an SSD to hold journals for multiple OSDs,
> both db and wal?  Searching has yielded nothing.

We are working on expanding the tooling to this for you, but until
then, it is up to the user to create the LVs manually.

This section might help out a bit on what you would need (for block.db):

http://docs.ceph.com/docs/master/rados/configuration/bluestore-config-ref/#block-and-block-db
>
>  R
>
>
> On Fri, Aug 17, 2018 at 9:48 AM David Turner <drakonstein@xxxxxxxxx> wrote:
>>
>> > ceph-volume lvm create --osd-id 0 --bluestore --data /dev/sdc --block.db
>> > /dev/sdb --block.wal /dev/sdb
>>
>> That command can't work... You're telling it to use the entire /dev/sdb
>> device for the db and then again to do it for the wal, but you can only use
>> the entire device once.  There are 2 things wrong with that.  First, if
>> you're putting db and wal on the same device you do not need to specify the
>> wal.  Second if you are actually intending to use a partition on /dev/sdb
>> instead of the entire block device for this single OSD, then you need to
>> manually create a partition for it and supply that partition to the
>> --block.db command.
>>
>> Likely the command you want will end up being this after you create a
>> partition on the SSD for the db/wal.
>> `ceph-volume lvm create --osd-id 0 --bluestore --data /dev/sdc --block.db
>> /dev/sdb1`
>>
>> On Fri, Aug 17, 2018 at 10:24 AM Robert Stanford <rstanford8896@xxxxxxxxx>
>> wrote:
>>>
>>>
>>>  I was using the ceph-volume create command, which I understand combines
>>> the prepare and activate functions.
>>>
>>> ceph-volume lvm create --osd-id 0 --bluestore --data /dev/sdc --block.db
>>> /dev/sdb --block.wal /dev/sdb
>>>
>>>  That is the command context I've found on the web.  Is it wrong?
>>>
>>>  Thanks
>>> R
>>>
>>> On Fri, Aug 17, 2018 at 5:55 AM Alfredo Deza <adeza@xxxxxxxxxx> wrote:
>>>>
>>>> On Thu, Aug 16, 2018 at 9:00 PM, Robert Stanford
>>>> <rstanford8896@xxxxxxxxx> wrote:
>>>> >
>>>> >  I am following the steps to my filestore journal with a bluestore
>>>> > journal
>>>> >
>>>> > (http://docs.ceph.com/docs/mimic/rados/operations/bluestore-migration/).  It
>>>> > is broken at ceph-volume lvm create.  Here is my error:
>>>> >
>>>> > --> Zapping successful for: /dev/sdc
>>>> > Preparing sdc
>>>> > Running command: /bin/ceph-authtool --gen-print-key
>>>> > Running command: /bin/ceph --cluster ceph --name client.bootstrap-osd
>>>> > --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring osd tree -f json
>>>> > Running command: /bin/ceph --cluster ceph --name client.bootstrap-osd
>>>> > --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring -i - osd new
>>>> > ff523216-350d-4ca0-9022-0c17662c2c3b 10
>>>> > Running command: vgcreate --force --yes
>>>> > ceph-459b4fbe-e3c4-4f28-b58e-3496bf3ea95a /dev/sdc
>>>> >  stdout: Physical volume "/dev/sdc" successfully created.
>>>> >  stdout: Volume group "ceph-459b4fbe-e3c4-4f28-b58e-3496bf3ea95a"
>>>> > successfully created
>>>> > Running command: lvcreate --yes -l 100%FREE -n
>>>> > osd-block-ff523216-350d-4ca0-9022-0c17662c2c3b
>>>> > ceph-459b4fbe-e3c4-4f28-b58e-3496bf3ea95a
>>>> >  stdout: Logical volume
>>>> > "osd-block-ff523216-350d-4ca0-9022-0c17662c2c3b"
>>>> > created.
>>>> > --> blkid could not detect a PARTUUID for device: sdb
>>>> > --> Was unable to complete a new OSD, will rollback changes
>>>> > --> OSD will be destroyed, keeping the ID because it was provided with
>>>> > --osd-id
>>>> > Running command: ceph osd destroy osd.10 --yes-i-really-mean-it
>>>> >  stderr: destroyed osd.10
>>>> > -->  RuntimeError: unable to use device
>>>> >
>>>> >  Note that SDB is the SSD journal.  It has been zapped prior.
>>>>
>>>> I can't see what the actual command you used is, but I am guessing you
>>>> did something like:
>>>>
>>>> ceph-volume lvm prepare --filestore --data /dev/sdb --journal /dev/sdb
>>>>
>>>> Which is not possible. There are a few ways you can do this (see:
>>>> http://docs.ceph.com/docs/master/ceph-volume/lvm/prepare/#filestore )
>>>>
>>>> With a raw device and a pre-created partition (must have a PARTUUID):
>>>>
>>>>     ceph-volume lvm prepare --data /dev/sdb --journal /dev/sdc1
>>>>
>>>> With LVs:
>>>>
>>>>     ceph-volume lvm prepare --data vg/my-data --journal vg/my-journal
>>>>
>>>> With an LV for data and a partition:
>>>>
>>>>     ceph-volume lvm prepare --data vg/my-data --journal /dev/sdc1
>>>>
>>>> >
>>>> >  What is going wrong, and how can I fix it?
>>>> >
>>>> >  Thank you
>>>> >  R
>>>> >
>>>> >
>>>> > _______________________________________________
>>>> > ceph-users mailing list
>>>> > ceph-users@xxxxxxxxxxxxxx
>>>> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>> >
>>>
>>> _______________________________________________
>>> ceph-users mailing list
>>> ceph-users@xxxxxxxxxxxxxx
>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux