I want to do some test in container with ceph, but after i install ceph, i find osd doesn't be installed successfully: [root@f3c3a6580a0f ~]$ ceph -s cluster: id: 20f00ca7-0de9-449e-a996-3d53aec4ed16 health: HEALTH_WARN Reduced data availability: 128 pgs inactive Degraded data redundancy: 128 pgs unclean services: mon: 2 daemons, quorum f3c3a6580a0f,2fca4942f691 mgr: f3c3a6580a0f(active), standbys: 2fca4942f691 mds: cephfs-1/1/1 up {0=2fca4942f691=up:creating}, 1 up:standby osd: 0 osds: 0 up, 0 in data: pools: 2 pools, 128 pgs objects: 0 objects, 0 bytes usage: 0 kB used, 0 kB / 0 kB avail pgs: 100.000% pgs unknown 128 unknown I only found below error from install log: Traceback (most recent call last): File "/usr/sbin/ceph-disk", line 9, in <module> load_entry_point('ceph-disk==1.0.0', 'console_scripts', 'ceph-disk')() File "/usr/lib/python2.7/site-packages/ceph_disk/main.py", line 5704, in run main(sys.argv[1:]) File "/usr/lib/python2.7/site-packages/ceph_disk/main.py", line 5657, in main main_catch(args.func, args) File "/usr/lib/python2.7/site-packages/ceph_disk/main.py", line 5682, in main_catch func(args) File "/usr/lib/python2.7/site-packages/ceph_disk/main.py", line 4617, in main_list main_list_protected(args) File "/usr/lib/python2.7/site-packages/ceph_disk/main.py", line 4621, in main_list_protected devices = list_devices() File "/usr/lib/python2.7/site-packages/ceph_disk/main.py", line 4525, in list_devices partmap = list_all_partitions() File "/usr/lib/python2.7/site-packages/ceph_disk/main.py", line 778, in list_all_partitions dev_part_list[name] = list_partitions(get_dev_path(name)) File "/usr/lib/python2.7/site-packages/ceph_disk/main.py", line 791, in list_partitions if is_mpath(dev): File "/usr/lib/python2.7/site-packages/ceph_disk/main.py", line 626, in is_mpath uuid = get_dm_uuid(dev) File "/usr/lib/python2.7/site-packages/ceph_disk/main.py", line 611, in get_dm_uuid uuid_path = os.path.join(block_path(dev), 'dm', 'uuid') File "/usr/lib/python2.7/site-packages/ceph_disk/main.py", line 605, in block_path rdev = os.stat(path).st_rdev OSError: [Errno 2] No such file or directory: '/dev/dm-6' but next info confused me: [ceph_deploy.osd][DEBUG ] Host 172.17.0.4 is now ready for osd use. In 172.17.0.4, `ceph-disk list` show above error too. Since start container in order, i can call `ceph-disk list` in the last one without above error. I know ceph-deploy would call ceph-disk, but above error do not make whole install process exit and in docker, the dm type device is for containers self. Is there any possible to skip dm device and make ceph install osd succeed indeed? -- Best Regards Dai Xiang _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com