On Tue, Jan 9, 2018 at 1:35 PM, Reed Dier <reed.dier@xxxxxxxxxxx> wrote: > I would just like to mirror what Dan van der Ster’s sentiments are. > > As someone attempting to move an OSD to bluestore, with limited/no LVM > experience, it is a completely different beast and complexity level compared > to the ceph-disk/filestore days. > > ceph-deploy was a very simple tool that did exactly what I was looking to > do, but now we have deprecated ceph-disk halfway into a release, ceph-deploy > doesn’t appear to fully support ceph-volume, which is now the official way > to manage OSDs moving forward. ceph-deploy now fully supports ceph-volume, we should get a release soon > > My ceph-volume create statement ‘succeeded’ but the OSD doesn’t start, so > now I am trying to zap the disk to try to recreate the OSD, and the zap is > failing as Dan’s did. I would encourage you to open a ticket in the tracker so that we can improve on what failed for you http://tracker.ceph.com/projects/ceph-volume/issues/new ceph-volume keeps thorough logs in /var/log/ceph/ceph-volume.log and /var/log/ceph/ceph-volume-systemd.log If you create a ticket, please make sure to add all the output and steps that you can > > And yes, I was able to get it zapped using the lvremove, vgremove, pvremove > commands, but that is not obvious to someone who hasn’t used LVM extensively > for storage management before. > > I also want to mirror Dan’s sentiments about the unnecessary complexity > imposed on what I expect is the default use case of an entire disk being > used. I can’t see anything more than the ‘entire disk’ method being the > largest use case for users of ceph, especially the smaller clusters trying > to maximize hardware/spend. We don't take lightly the introduction of LVM here. The new tool is addressing several insurmountable issues with how ceph-disk operated. Although using an entire disk might be easier in the use case you are in, it is certainly not the only thing we have to support, so then again, we can't reliably decide what strategy would be best to destroy that volume, or group, or if the PV should be destroyed as well. The 'zap' sub-command will allow that lv to be reused for an OSD and that should work. Again, if it isn't sufficient, we really do need more information and a ticket in the tracker is the best way. > > Just wanted to piggy back this thread to echo Dan’s frustration. > > Thanks, > > Reed > > On Jan 8, 2018, at 10:41 AM, Alfredo Deza <adeza@xxxxxxxxxx> wrote: > > On Mon, Jan 8, 2018 at 10:53 AM, Dan van der Ster <dan@xxxxxxxxxxxxxx> > wrote: > > On Mon, Jan 8, 2018 at 4:37 PM, Alfredo Deza <adeza@xxxxxxxxxx> wrote: > > On Thu, Dec 21, 2017 at 11:35 AM, Stefan Kooman <stefan@xxxxxx> wrote: > > Quoting Dan van der Ster (dan@xxxxxxxxxxxxxx): > > Thanks Stefan. But isn't there also some vgremove or lvremove magic > that needs to bring down these /dev/dm-... devices I have? > > > Ah, you want to clean up properly before that. Sure: > > lvremove -f <volume_group>/<logical_volume> > vgremove <volume_group> > pvremove /dev/ceph-device (should wipe labels) > > So ideally there should be a ceph-volume lvm destroy / zap option that > takes care of this: > > 1) Properly remove LV/VG/PV as shown above > 2) wipefs to get rid of LVM signatures > 3) dd zeroes to get rid of signatures that might still be there > > > ceph-volume does have a 'zap' subcommand, but it does not remove > logical volumes or groups. It is intended to leave those in place for > re-use. It uses wipefs, but > not in a way that would end up removing LVM signatures. > > Docs for zap are at: http://docs.ceph.com/docs/master/ceph-volume/lvm/zap/ > > The reason for not attempting removal is that an LV might not be a > 1-to-1 device to volume group. It is being suggested here to "vgremove > <volume_group>" > but what if the group has several other LVs that should not get > removed? Similarly, what if the logical volume is not a single PV but > many? > > We believe that these operations should be up to the administrator > with better context as to what goes where and what (if anything) > really needs to be removed > from LVM. > > > Maybe I'm missing something, but aren't most (almost all?) use-cases just > > ceph-volume lvm create /dev/<thewholedisk> > > > No > > > ? Or do you expect most deployments to do something more complicated with > lvm? > > > Yes, we do. For example dmcache, which to ceph-volume looks like a > plain logical volume, but it can be vary on how it is implemented > behind the scenes > > In that above whole-disk case, I think it would be useful to have a > very simple cmd to tear down whatever ceph-volume created, so that > ceph admins don't need to reverse engineer what ceph-volume is doing > with lvm. > > > Right, that would work if that was the only supported way of dealing > with lvm. We aren't imposing this, we added it as a convenience if a > user did not want > to deal with lvm at all. LVM has a plethora of ways to create an LV, > and we don't want to either restrict users to our view of LVM or > attempt to understand all the many different > ways that may be and assume some behavior is desired (like removing a VG) > > > Otherwise, perhaps it would be useful to document the expected normal > lifecycle of an lvm osd: create, failure / replacement handling, > decommissioning. > > Cheers, Dan > > > > > > Gr. Stefan > > -- > | BIT BV http://www.bit.nl/ Kamer van Koophandel 09090351 > | GPG: 0xD14839C6 +31 318 648 688 / info@xxxxxx > _______________________________________________ > ceph-users mailing list > ceph-users@xxxxxxxxxxxxxx > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > _______________________________________________ > ceph-users mailing list > ceph-users@xxxxxxxxxxxxxx > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com