Hi, recently I tried to upgrade from 0.57 to 0.67.3, hit the changes of mon protocol, and so I updated all of the 5 mons. After upgrading the mon, (and during the debugging of other problems,) I removed and created the mon filesystem from scratch. OSDMap and Crushmap seems to be blown by this; I expected at that time that they are easily recovered. Now, I tried to recover but couldn't make osdmap remember any OSDs. Would somebody help me, please ? osd dump shows: % ceph osd dump epoch 5 fsid 6ef8e7bb-745c-4e20-8946-8808b9843380 created 2013-09-11 18:36:36.731905 modified 2013-09-12 00:53:46.375623 flags pool 0 'data' rep size 2 min_size 1 crush_ruleset 0 object_hash rjenkins pg_num 64 pgp_num 64 last_change 1 owner 0 crash_replay_interval 45 pool 1 'metadata' rep size 2 min_size 1 crush_ruleset 1 object_hash rjenkins pg_num 64 pgp_num 64 last_change 1 owner 0 pool 2 'rbd' rep size 2 min_size 1 crush_ruleset 2 object_hash rjenkins pg_num 64 pgp_num 64 last_change 1 owner 0 max_osd 16 (nothing further) % It seems that crushmap is set correctly (attached in the end of this e-mail). Although I turned off authx temporarily, I set osds in the ceph auth list (manually by ceph auth add command). installed auth entries: osd.0 key: AQCMcvZQIA6XNRAAbVshyOevLfNUI0SIkVEAzQ== caps: [mon] allow rwx caps: [osd] allow * osd.1 key: AQB/c/ZQKAf8OBAAIs9sHZlAjNoIn7eE2bbggg== caps: [mon] allow rwx caps: [osd] allow * : : # ceph osd in 3 osd.3 does not exist. # /etc/init.d/ceph -a start osd.3 === osd.3 === Mounting Btrfs on host-4:/ceph/data/osd-disk Scanning for Btrfs filesystems Error ENOENT: osd.3 does not exist. create it before updating the crush map Starting Ceph osd.3 on host-4... starting osd.3 at :/0 osd_data /ceph/data/osd-disk /ceph/data/osd-journal # /ceph/bin/ceph status cluster 6ef8e7bb-745c-4e20-8946-8808b9843380 health HEALTH_ERR 192 pgs stuck inactive; 192 pgs stuck unclean; no osds monmap e1: 5 mons at {0=A.B.C.D:6789/0,1=A.B.C.D:6789/0,2=A.B.C.D:6789/0,3=A.B.C.D:6789/0,4=A.B.C.D:6789/0}, election epoch 22, quorum 0,1,2,3,4 0,1,2,3,4 osdmap e5: 0 osds: 0 up, 0 in pgmap v6: 192 pgs: 192 creating; 0 bytes data, 0 KB used, 0 KB / 0 KB avail mdsmap e7: 1/1/1 up {0=1=up:creating}, 4 up:standby How can I recover the osd map ? Thanks ! Yasu (Crushmap follows.) # begin crush map # devices device 0 osd.0 device 1 osd.1 device 2 osd.2 device 3 osd.3 device 4 osd.4 device 5 osd.5 device 6 osd.6 device 7 osd.7 device 8 osd.8 device 9 osd.9 device 10 osd.10 device 11 osd.11 device 12 osd.12 device 13 osd.13 device 14 osd.14 device 15 osd.15 # types type 0 device type 1 host type 2 rack type 3 row type 4 room type 5 datacenter type 6 root # buckets host host-1 { id -1 # do not change unnecessarily # weight 1.000 alg straw hash 0 # rjenkins1 item osd.0 weight 1.000 } host host-2 { id -2 # do not change unnecessarily # weight 1.000 alg straw hash 0 # rjenkins1 item osd.1 weight 1.000 } host host-3 { id -3 # do not change unnecessarily # weight 1.000 alg straw hash 0 # rjenkins1 item osd.2 weight 1.000 } host host-4 { id -4 # do not change unnecessarily # weight 1.000 alg straw hash 0 # rjenkins1 item osd.3 weight 1.000 } host host-5 { id -5 # do not change unnecessarily # weight 1.000 alg straw hash 0 # rjenkins1 item osd.4 weight 1.000 } host host-6 { id -6 # do not change unnecessarily # weight 1.000 alg straw hash 0 # rjenkins1 item osd.5 weight 1.000 } host host-7 { id -7 # do not change unnecessarily # weight 1.000 alg straw hash 0 # rjenkins1 item osd.6 weight 1.000 } host host-8 { id -8 # do not change unnecessarily # weight 1.000 alg straw hash 0 # rjenkins1 item osd.7 weight 1.000 } host host-9 { id -9 # do not change unnecessarily # weight 1.000 alg straw hash 0 # rjenkins1 item osd.8 weight 1.000 } host host-10 { id -10 # do not change unnecessarily # weight 1.000 alg straw hash 0 # rjenkins1 item osd.9 weight 1.000 } host host-11 { id -11 # do not change unnecessarily # weight 1.000 alg straw hash 0 # rjenkins1 item osd.10 weight 1.000 } host host-12 { id -12 # do not change unnecessarily # weight 1.000 alg straw hash 0 # rjenkins1 item osd.11 weight 1.000 } host host-13 { id -13 # do not change unnecessarily # weight 1.000 alg straw hash 0 # rjenkins1 item osd.12 weight 1.000 } host host-14 { id -14 # do not change unnecessarily # weight 1.000 alg straw hash 0 # rjenkins1 item osd.13 weight 1.000 } host host-15 { id -15 # do not change unnecessarily # weight 1.000 alg straw hash 0 # rjenkins1 item osd.14 weight 1.000 } host host-16 { id -16 # do not change unnecessarily # weight 1.000 alg straw hash 0 # rjenkins1 item osd.15 weight 1.000 } rack rack0 { id -17 # do not change unnecessarily # weight 16.000 alg straw hash 0 # rjenkins1 item host-1 weight 1.000 item host-2 weight 1.000 item host-3 weight 1.000 item host-4 weight 1.000 item host-5 weight 1.000 item host-6 weight 1.000 item host-7 weight 1.000 item host-8 weight 1.000 item host-9 weight 1.000 item host-10 weight 1.000 item host-11 weight 1.000 item host-12 weight 1.000 item host-13 weight 1.000 item host-14 weight 1.000 item host-15 weight 1.000 item host-16 weight 1.000 } row row0 { id -18 # do not change unnecessarily # weight 16.000 alg straw hash 0 # rjenkins1 item rack0 weight 16.000 } room room0 { id -19 # do not change unnecessarily # weight 16.000 alg straw hash 0 # rjenkins1 item row0 weight 16.000 } datacenter datacenter0 { id -20 # do not change unnecessarily # weight 16.000 alg straw hash 0 # rjenkins1 item room0 weight 16.000 } root default { id -21 # do not change unnecessarily # weight 16.000 alg straw hash 0 # rjenkins1 item datacenter0 weight 16.000 } # rules rule data { ruleset 0 type replicated min_size 1 max_size 10 step take default step chooseleaf firstn 0 type host step emit } rule metadata { ruleset 1 type replicated min_size 1 max_size 10 step take default step chooseleaf firstn 0 type host step emit } rule rbd { ruleset 2 type replicated min_size 1 max_size 10 step take default step chooseleaf firstn 0 type host step emit } # end crush map _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com