On Thu, 19 Oct 2017, Dan van der Ster wrote: > Hi Wido, > > Unexpected crush location changes can indeed be quite nasty. > > With this in mind, I wonder if a crush lock would be useful. > > ceph osd set nocrushchange > > With that flag set, osds could still go in and out, but crush > move/add/remove/etc..., also tunables changes, would be blocked. The problem I see with this is that it would prevent new OSD additions or other changes.. and if you went to unset the flag in order to allow a new node to be adjusted or brought online then you might get an avalance of blocked changes. I think what we actually want is a more targetted variation of osd_crush_update_on_start that only updates teh location if it has never been set (i.e., it is a new osd). Like, osd_crush_update_on_create. Then it's left to the admin to move OSDs? > On Mon, Oct 16, 2017 at 2:02 PM, Wido den Hollander <wido@xxxxxxxx> wrote: > > Hi, > > > > I completely overread this, but I just found out that osd_crush_location_hook was renamed to crush_location_hook in the new config style. > > > > When upgrading from Jewel to Luminous without touching your configuration OSDs will move to the default CRUSH location due to the hook not being executed. > > > > Was this an oversight with Luminous or was it intentional? The implications are an oversight.. I didn't think about customized hooks that would get reverted if the config option wasn't changed. Otherwise the item in the release notes would have read more like a warning: * The `osd crush location` config option is no longer supported. Please update your ceph.conf to use the `crush location` option instead. Sorry about that! sage -- To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html