> Op 22 oktober 2017 om 18:45 schreef Sean Sullivan <lookcrabs@xxxxxxxxx>: > > > On freshly installed ubuntu 16.04 servers with the HWE kernel selected > (4.10). I can not use ceph-deploy or ceph-disk to provision osd. > > > whenever I try I get the following:: > > ceph-disk -v prepare --dmcrypt --dmcrypt-key-dir /etc/ceph/dmcrypt-keys > --bluestore --cluster ceph --fs-type xfs -- /dev/sdy > command: Running command: /usr/bin/ceph-osd --cluster=ceph > --show-config-value=fsid > get_dm_uuid: get_dm_uuid /dev/sdy uuid path is /sys/dev/block/65:128/dm/uuid > set_type: Will colocate block with data on /dev/sdy > command: Running command: /usr/bin/ceph-conf --cluster=ceph --name=osd. > --lookup bluestore_block_size > [command: Running command: /usr/bin/ceph-conf --cluster=ceph --name=osd. > --lookup bluestore_block_db_size > command: Running command: /usr/bin/ceph-conf --cluster=ceph --name=osd. > --lookup bluestore_block_size > command: Running command: /usr/bin/ceph-conf --cluster=ceph --name=osd. > --lookup bluestore_block_wal_size > get_dm_uuid: get_dm_uuid /dev/sdy uuid path is /sys/dev/block/65:128/dm/uuid > get_dm_uuid: get_dm_uuid /dev/sdy uuid path is /sys/dev/block/65:128/dm/uuid > get_dm_uuid: get_dm_uuid /dev/sdy uuid path is /sys/dev/block/65:128/dm/uuid > Traceback (most recent call last): > File "/usr/sbin/ceph-disk", line 9, in <module> > load_entry_point('ceph-disk==1.0.0', 'console_scripts', 'ceph-disk')() > File "/usr/lib/python2.7/dist-packages/ceph_disk/main.py", line 5704, in > run > main(sys.argv[1:]) > File "/usr/lib/python2.7/dist-packages/ceph_disk/main.py", line 5655, in > main > args.func(args) > File "/usr/lib/python2.7/dist-packages/ceph_disk/main.py", line 2091, in > main > Prepare.factory(args).prepare() > File "/usr/lib/python2.7/dist-packages/ceph_disk/main.py", line 2080, in > prepare > self._prepare() > File "/usr/lib/python2.7/dist-packages/ceph_disk/main.py", line 2154, in > _prepare > self.lockbox.prepare() > File "/usr/lib/python2.7/dist-packages/ceph_disk/main.py", line 2842, in > prepare > verify_not_in_use(self.args.lockbox, check_partitions=True) > File "/usr/lib/python2.7/dist-packages/ceph_disk/main.py", line 950, in > verify_not_in_use > raise Error('Device is mounted', partition) > ceph_disk.main.Error: Error: Device is mounted: /dev/sdy5 > > unmounting the disk does not seem to help either. I'm assuming something is > triggering too early but i'm not sure how to delay or figure that out. > > has anyone deployed on xenial with the 4.10 kernel? Am I missing something > important? Yes I have without any issues, I've did: $ ceph-disk prepare /dev/sdb Luminous default to BlueStore and that worked just fine. Yes, this is with a 4.10 HWE kernel from Ubuntu 16.04. Wido > _______________________________________________ > ceph-users mailing list > ceph-users@xxxxxxxxxxxxxx > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com