Thanks! I'll start looking into rebuilding my roles once 12.2.2 is out then. On 28 November 2017 at 13:37, Alfredo Deza <adeza@xxxxxxxxxx> wrote: > On Tue, Nov 28, 2017 at 7:22 AM, Andreas Calminder > <andreas.calminder@xxxxxxxxxx> wrote: >>> For the `simple` sub-command there is no prepare/activate, it is just >>> a way of taking over management of an already deployed OSD. For *new* >>> OSDs, yes, we are implying that we are going only with Logical Volumes >>> for data devices. It is a bit more flexible for Journals, block.db, >>> and block.wal as those >>> can be either logical volumes or GPT partitions (ceph-volume will not >>> create these for you). >> >> Ok, so if I understand this correctly, for future one-device-per-osd >> setups I would create a volume group per device before handing it over >> to ceph-volume, to get the "same" functionality as ceph-disk. I >> understand the flexibility aspect of this, my setup will have an extra >> step setting up lvm for my osd devices which is fine. > > If you don't require any special configuration for your logical volume > and don't mind a naive LV handling, then ceph-volume can create > the logical volume for you from either a partition or a device (for > data), although it will still require a GPT partition for Journals, > block.wal, and block.db > > For example: > > ceph-volume lvm create --data /path/to/device > > Would create a new volume group with the device and then produce a > single LV from it. > >> Apologies if I >> missed the information, but is it possible to get command output as >> json, something like "ceph-disk list --format json" since it's quite >> helpful while setting up stuff through ansible > > Yes, this is implemented in both "pretty" and JSON formats: > http://docs.ceph.com/docs/master/ceph-volume/lvm/list/#ceph-volume-lvm-list >> >> Thanks, >> Andreas >> >> On 28 November 2017 at 12:47, Alfredo Deza <adeza@xxxxxxxxxx> wrote: >>> On Tue, Nov 28, 2017 at 1:56 AM, Andreas Calminder >>> <andreas.calminder@xxxxxxxxxx> wrote: >>>> Hello, >>>> Thanks for the heads-up. As someone who's currently maintaining a >>>> Jewel cluster and are in the process of setting up a shiny new >>>> Luminous cluster and writing Ansible roles in the process to make >>>> setup reproducible. I immediately proceeded to look into ceph-volume >>>> and I've some questions/concerns, mainly due to my own setup, which is >>>> one osd per device, simple. >>>> >>>> Running ceph-volume in Luminous 12.2.1 suggests there's only the lvm >>>> subcommand available and the man-page only covers lvm. The online >>>> documentation http://docs.ceph.com/docs/master/ceph-volume/ lists >>>> simple however it's lacking some of the ceph-disk commands, like >>>> 'prepare' which seems crucial in the 'simple' scenario. Does the >>>> ceph-disk deprecation imply that lvm is mandatory for using devices >>>> with ceph or is just the documentation and tool features lagging >>>> behind, I.E the 'simple' parts will be added well in time for Mimic >>>> and during the Luminous lifecycle? Or am I missing something? >>> >>> In your case, all your existing OSDs will be able to be managed by >>> `ceph-volume` once scanned and the information persisted. So anything >>> from Jewel should still work. For 12.2.1 you are right, that command >>> is not yet available, it will be present in 12.2.2 >>> >>> For the `simple` sub-command there is no prepare/activate, it is just >>> a way of taking over management of an already deployed OSD. For *new* >>> OSDs, yes, we are implying that we are going only with Logical Volumes >>> for data devices. It is a bit more flexible for Journals, block.db, >>> and block.wal as those >>> can be either logical volumes or GPT partitions (ceph-volume will not >>> create these for you). >>> >>>> >>>> Best regards, >>>> Andreas >>>> >>>> On 27 November 2017 at 14:36, Alfredo Deza <adeza@xxxxxxxxxx> wrote: >>>>> For the upcoming Luminous release (12.2.2), ceph-disk will be >>>>> officially in 'deprecated' mode (bug fixes only). A large banner with >>>>> deprecation information has been added, which will try to raise >>>>> awareness. >>>>> >>>>> We are strongly suggesting using ceph-volume for new (and old) OSD >>>>> deployments. The only current exceptions to this are encrypted OSDs >>>>> and FreeBSD systems >>>>> >>>>> Encryption support is planned and will be coming soon to ceph-volume. >>>>> >>>>> A few items to consider: >>>>> >>>>> * ceph-disk is expected to be fully removed by the Mimic release >>>>> * Existing OSDs are supported by ceph-volume. They can be "taken over" [0] >>>>> * ceph-ansible already fully supports ceph-volume and will soon default to it >>>>> * ceph-deploy support is planned and should be fully implemented soon >>>>> >>>>> >>>>> [0] http://docs.ceph.com/docs/master/ceph-volume/simple/ >>>>> -- >>>>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in >>>>> the body of a message to majordomo@xxxxxxxxxxxxxxx >>>>> More majordomo info at http://vger.kernel.org/majordomo-info.html _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com