On Tue, Sep 12, 2017 at 4:45 PM, Two Spirit <twospirit6905@xxxxxxxxx> wrote: > Hi Sage, > > I tried to install an encrypted OSD, but I found the documentation to > be light and not finding any examples. I was able to prepare > (ceph-deploy osd prepare --dmcrypt <host>:<relative_device_name>', but > I couldn't activate it. What OS Version is this ? and What is the error? I dont see any issues in the nightly runs with CentOS/Ubuntu. Have you tried osd create --bluestore --dmcrypt <host>:<block> ? I found a tracker page that listed you as one > of the owners http://tracker.ceph.com/projects/ceph/wiki/Osd_-_simple_ceph-mon_dm-crypt_key_management > that seemed to imply that it was in the 'planning' phase. I was > wondering if this is in planning phase, or if it is fully supported > now ready to begin testing. > > I have a requirement to place complete disks in encrypted containers > (including partition tables. kpartx accesses the encrypted partitions. > some versions of Ubuntu I have to do this manually, but newer versions > seem to automatically do this). I haven't absorbed all the ceph > dmcrypt architecture yet, but I did not feel comfortable that the keys > were available by querying ceph. Not saying I have any idea how to do > it better in a ceph environment. Can you accept a network accessible > keyfile? > > I thought I might be able to get some testing by creating my own > encrypted container outside of ceph, which was accessible via > /dev/mapper/lukscontainer. I tried 'ceph-deploy osd prepare > luksosd:/dev/mapper/lukscontainer' and it didn't like that. I saw docs > using absolute references (/dev/sdb) so I thought > /dev/mapper/lukscontainer should be usable. ceph-deploy osd prepare > --dmcrypt luksosd:sdb seemed to work -- maybe it didn't like the fact > the path was an extra level deeper. Where is the encrypted ceph data > partition to activate? > > As a temporary workaround, I was thinking I could create a /dev/mdXXX > device using mdadm forcing a single disk raid. This would keep key > management outside of ceph for now. and be able to run the relative > path format 'ceph-deploy osd prepare luksosd:mdXXX'. In theory, this > should work and supported by ceph, right? > -- > To unsubscribe from this list: send the line "unsubscribe ceph-devel" in > the body of a message to majordomo@xxxxxxxxxxxxxxx > More majordomo info at http://vger.kernel.org/majordomo-info.html -- To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html