Re: ceph-disk improvements

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Fri, Apr 1, 2016 at 11:36 AM, Sage Weil <sweil@xxxxxxxxxx> wrote:
> Hi all,
>
> There are a couple of looming features for ceph-disk:
>
> 1- Support for additional devices when using BlueStore.  There can be up
> to three: the main device, a WAL/journal device (small, ~128MB, ideally
> NVRAM), and a fast metadata device (as big as you have available; will be
> used for internal metadata).
>
> 2- Support for setting up dm-cache, bcache, and/or FlashCache underneath
> filestore or bluestore.
>
> The current syntax of
>
>  ceph-disk prepare [--dmcrypt] [--bluestore] DATADEV [JOURNALDEV]
>
> isn't terribly expressive.  For example, the journal device size is set
> via a config option, not on the command line.  For bluestore, the metadata
> device will probably want/need explicit user input so they can ensure it's
> 1/Nth of their SSD (if they have N HDDs to each SSD).
>
> And if we put dmcache in there, that partition will need to be sized too.

Sebastien's suggestion of allowing plugins for ceph-disk is ideal
here, because it would allow to enable extra functionality
(and possibly at a faster release pace) without interfering with the
current syntax.

Reusing your examples, a "bluestore" plugin could be a sub-command:

    ceph-disk bluestore prepare [...]

Device size, extra flags or overriding options would be clearly
separated because of the subcommand. This would be the same
case for dm-cache, bcache, or whatever comes next.

>
> Another consideration is that right now we don't play nice with LVM at
> all.  Should we?  dm-cache is usually used in conjunction with LVM
> (although it doesn't have to be).  Does LVM provide value?  Like, the
> ability for users to add a second SSD to a box and migrate cache, wal, or
> journal partitions around?

One of the problematic corners of ceph-disk is that it tries to be
helpful by trying to predict accurately sizes and partitions
to make it simpler for a user. I would love to see ceph-disk be less
flexible here and require actual full devices for an OSD and a
separate
device for a Journal, while starting to deprecate journal collocation
and directory-osd.

Going back to the plugin idea, LVM support could be enabled by a
separate plugin and ceph-disk could stay lean.

>
> I'm interested in hearing feedback on requirements, approaches, and
> interfaces before we go too far down the road...
>
> Thanks!
> sage
> --
> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> the body of a message to majordomo@xxxxxxxxxxxxxxx
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html



[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux