Hi all, I’m testing Ceph Luminous 12.2.1 installed with ceph ansible. Doing some failover tests I noticed that when I kill an osd or and hosts Ceph doesn’t recover automatically remaining in this state until I bring OSDs or host back online. I’ve 3 pools volumes, cephfs_data and cephfs_metadata with size 3 and min_size 1. Is there something I’m missing ? Below some cluster info. Thank you all Regards Matteo cluster: id: ab7cb890-ee21-484e-9290-14b9e5e85125 health: HEALTH_WARN 3 osds down Degraded data redundancy: 2842/73686 objects degraded (3.857%), 318 pgs unclean, 318 pgs degraded, 318 pgs undersized services: mon: 3 daemons, quorum controller001,controller002,controller003 mgr: controller001(active), standbys: controller002, controller003 mds: cephfs-1/1/1 up {0=controller002=up:active}, 2 up:standby osd: 77 osds: 74 up, 77 in data: pools: 3 pools, 4112 pgs objects: 36843 objects, 142 GB usage: 470 GB used, 139 TB / 140 TB avail pgs: 2842/73686 objects degraded (3.857%) 3794 active+clean 318 active+undersized+degraded ID CLASS WEIGHT TYPE NAME STATUS REWEIGHT PRI-AFF -1 140.02425 root default -9 20.00346 host storage001 0 hdd 1.81850 osd.0 up 1.00000 1.00000 6 hdd 1.81850 osd.6 up 1.00000 1.00000 8 hdd 1.81850 osd.8 up 1.00000 1.00000 11 hdd 1.81850 osd.11 up 1.00000 1.00000 14 hdd 1.81850 osd.14 up 1.00000 1.00000 18 hdd 1.81850 osd.18 up 1.00000 1.00000 24 hdd 1.81850 osd.24 up 1.00000 1.00000 28 hdd 1.81850 osd.28 up 1.00000 1.00000 33 hdd 1.81850 osd.33 up 1.00000 1.00000 40 hdd 1.81850 osd.40 up 1.00000 1.00000 45 hdd 1.81850 osd.45 up 1.00000 1.00000 -7 20.00346 host storage002 1 hdd 1.81850 osd.1 up 1.00000 1.00000 5 hdd 1.81850 osd.5 up 1.00000 1.00000 9 hdd 1.81850 osd.9 up 1.00000 1.00000 21 hdd 1.81850 osd.21 up 1.00000 1.00000 22 hdd 1.81850 osd.22 up 1.00000 1.00000 23 hdd 1.81850 osd.23 up 1.00000 1.00000 35 hdd 1.81850 osd.35 up 1.00000 1.00000 36 hdd 1.81850 osd.36 up 1.00000 1.00000 38 hdd 1.81850 osd.38 up 1.00000 1.00000 42 hdd 1.81850 osd.42 up 1.00000 1.00000 49 hdd 1.81850 osd.49 up 1.00000 1.00000 -11 20.00346 host storage003 27 hdd 1.81850 osd.27 up 1.00000 1.00000 31 hdd 1.81850 osd.31 up 1.00000 1.00000 32 hdd 1.81850 osd.32 up 1.00000 1.00000 37 hdd 1.81850 osd.37 up 1.00000 1.00000 44 hdd 1.81850 osd.44 up 1.00000 1.00000 46 hdd 1.81850 osd.46 up 1.00000 1.00000 48 hdd 1.81850 osd.48 up 1.00000 1.00000 53 hdd 1.81850 osd.53 up 1.00000 1.00000 54 hdd 1.81850 osd.54 up 1.00000 1.00000 56 hdd 1.81850 osd.56 up 1.00000 1.00000 59 hdd 1.81850 osd.59 up 1.00000 1.00000 -3 20.00346 host storage004 2 hdd 1.81850 osd.2 up 1.00000 1.00000 4 hdd 1.81850 osd.4 up 1.00000 1.00000 10 hdd 1.81850 osd.10 up 1.00000 1.00000 16 hdd 1.81850 osd.16 up 1.00000 1.00000 17 hdd 1.81850 osd.17 up 1.00000 1.00000 19 hdd 1.81850 osd.19 up 1.00000 1.00000 26 hdd 1.81850 osd.26 up 1.00000 1.00000 29 hdd 1.81850 osd.29 up 1.00000 1.00000 39 hdd 1.81850 osd.39 up 1.00000 1.00000 43 hdd 1.81850 osd.43 up 1.00000 1.00000 50 hdd 1.81850 osd.50 up 1.00000 1.00000 -5 20.00346 host storage005 3 hdd 1.81850 osd.3 up 1.00000 1.00000 7 hdd 1.81850 osd.7 up 1.00000 1.00000 12 hdd 1.81850 osd.12 up 1.00000 1.00000 13 hdd 1.81850 osd.13 up 1.00000 1.00000 15 hdd 1.81850 osd.15 up 1.00000 1.00000 20 hdd 1.81850 osd.20 up 1.00000 1.00000 25 hdd 1.81850 osd.25 up 1.00000 1.00000 30 hdd 1.81850 osd.30 up 1.00000 1.00000 34 hdd 1.81850 osd.34 up 1.00000 1.00000 41 hdd 1.81850 osd.41 up 1.00000 1.00000 47 hdd 1.81850 osd.47 up 1.00000 1.00000 -13 20.00346 host storage006 51 hdd 1.81850 osd.51 up 1.00000 1.00000 55 hdd 1.81850 osd.55 up 1.00000 1.00000 58 hdd 1.81850 osd.58 up 1.00000 1.00000 61 hdd 1.81850 osd.61 up 1.00000 1.00000 63 hdd 1.81850 osd.63 up 1.00000 1.00000 65 hdd 1.81850 osd.65 up 1.00000 1.00000 66 hdd 1.81850 osd.66 up 1.00000 1.00000 69 hdd 1.81850 osd.69 up 1.00000 1.00000 71 hdd 1.81850 osd.71 up 1.00000 1.00000 73 hdd 1.81850 osd.73 up 1.00000 1.00000 75 hdd 1.81850 osd.75 up 1.00000 1.00000 -15 20.00346 host storage007 52 hdd 1.81850 osd.52 up 1.00000 1.00000 57 hdd 1.81850 osd.57 up 1.00000 1.00000 60 hdd 1.81850 osd.60 up 1.00000 1.00000 62 hdd 1.81850 osd.62 up 1.00000 1.00000 64 hdd 1.81850 osd.64 up 1.00000 1.00000 67 hdd 1.81850 osd.67 up 1.00000 1.00000 68 hdd 1.81850 osd.68 up 1.00000 1.00000 70 hdd 1.81850 osd.70 up 1.00000 1.00000 72 hdd 1.81850 osd.72 up 1.00000 1.00000 74 hdd 1.81850 osd.74 up 1.00000 1.00000 76 hdd 1.81850 osd.76 up 1.00000 1.00000 # begin crush map tunable choose_local_tries 0 tunable choose_local_fallback_tries 0 tunable choose_total_tries 50 tunable chooseleaf_descend_once 1 tunable chooseleaf_vary_r 1 tunable chooseleaf_stable 1 tunable straw_calc_version 1 tunable allowed_bucket_algs 54 # devices device 0 osd.0 class hdd device 1 osd.1 class hdd device 2 osd.2 class hdd device 3 osd.3 class hdd device 4 osd.4 class hdd device 5 osd.5 class hdd device 6 osd.6 class hdd device 7 osd.7 class hdd device 8 osd.8 class hdd device 9 osd.9 class hdd device 10 osd.10 class hdd device 11 osd.11 class hdd device 12 osd.12 class hdd device 13 osd.13 class hdd device 14 osd.14 class hdd device 15 osd.15 class hdd device 16 osd.16 class hdd device 17 osd.17 class hdd device 18 osd.18 class hdd device 19 osd.19 class hdd device 20 osd.20 class hdd device 21 osd.21 class hdd device 22 osd.22 class hdd device 23 osd.23 class hdd device 24 osd.24 class hdd device 25 osd.25 class hdd device 26 osd.26 class hdd device 27 osd.27 class hdd device 28 osd.28 class hdd device 29 osd.29 class hdd device 30 osd.30 class hdd device 31 osd.31 class hdd device 32 osd.32 class hdd device 33 osd.33 class hdd device 34 osd.34 class hdd device 35 osd.35 class hdd device 36 osd.36 class hdd device 37 osd.37 class hdd device 38 osd.38 class hdd device 39 osd.39 class hdd device 40 osd.40 class hdd device 41 osd.41 class hdd device 42 osd.42 class hdd device 43 osd.43 class hdd device 44 osd.44 class hdd device 45 osd.45 class hdd device 46 osd.46 class hdd device 47 osd.47 class hdd device 48 osd.48 class hdd device 49 osd.49 class hdd device 50 osd.50 class hdd device 51 osd.51 class hdd device 52 osd.52 class hdd device 53 osd.53 class hdd device 54 osd.54 class hdd device 55 osd.55 class hdd device 56 osd.56 class hdd device 57 osd.57 class hdd device 58 osd.58 class hdd device 59 osd.59 class hdd device 60 osd.60 class hdd device 61 osd.61 class hdd device 62 osd.62 class hdd device 63 osd.63 class hdd device 64 osd.64 class hdd device 65 osd.65 class hdd device 66 osd.66 class hdd device 67 osd.67 class hdd device 68 osd.68 class hdd device 69 osd.69 class hdd device 70 osd.70 class hdd device 71 osd.71 class hdd device 72 osd.72 class hdd device 73 osd.73 class hdd device 74 osd.74 class hdd device 75 osd.75 class hdd device 76 osd.76 class hdd # types type 0 osd type 1 host type 2 chassis type 3 rack type 4 row type 5 pdu type 6 pod type 7 room type 8 datacenter type 9 region type 10 root # buckets host storage004 { id -3 # do not change unnecessarily id -4 class hdd # do not change unnecessarily # weight 20.003 alg straw2 hash 0 # rjenkins1 item osd.2 weight 1.818 item osd.4 weight 1.818 item osd.10 weight 1.818 item osd.16 weight 1.818 item osd.17 weight 1.818 item osd.19 weight 1.818 item osd.26 weight 1.818 item osd.29 weight 1.818 item osd.39 weight 1.818 item osd.43 weight 1.818 item osd.50 weight 1.818 } host storage005 { id -5 # do not change unnecessarily id -6 class hdd # do not change unnecessarily # weight 20.003 alg straw2 hash 0 # rjenkins1 item osd.3 weight 1.818 item osd.7 weight 1.818 item osd.12 weight 1.818 item osd.13 weight 1.818 item osd.15 weight 1.818 item osd.20 weight 1.818 item osd.25 weight 1.818 item osd.30 weight 1.818 item osd.34 weight 1.818 item osd.41 weight 1.818 item osd.47 weight 1.818 } host storage002 { id -7 # do not change unnecessarily id -8 class hdd # do not change unnecessarily # weight 20.003 alg straw2 hash 0 # rjenkins1 item osd.1 weight 1.818 item osd.5 weight 1.818 item osd.9 weight 1.818 item osd.21 weight 1.818 item osd.22 weight 1.818 item osd.23 weight 1.818 item osd.35 weight 1.818 item osd.36 weight 1.818 item osd.38 weight 1.818 item osd.42 weight 1.818 item osd.49 weight 1.818 } host storage001 { id -9 # do not change unnecessarily id -10 class hdd # do not change unnecessarily # weight 20.003 alg straw2 hash 0 # rjenkins1 item osd.0 weight 1.818 item osd.6 weight 1.818 item osd.8 weight 1.818 item osd.11 weight 1.818 item osd.14 weight 1.818 item osd.18 weight 1.818 item osd.24 weight 1.818 item osd.28 weight 1.818 item osd.33 weight 1.818 item osd.40 weight 1.818 item osd.45 weight 1.818 } host storage003 { id -11 # do not change unnecessarily id -12 class hdd # do not change unnecessarily # weight 20.003 alg straw2 hash 0 # rjenkins1 item osd.27 weight 1.818 item osd.31 weight 1.818 item osd.32 weight 1.818 item osd.37 weight 1.818 item osd.44 weight 1.818 item osd.46 weight 1.818 item osd.48 weight 1.818 item osd.54 weight 1.818 item osd.53 weight 1.818 item osd.59 weight 1.818 item osd.56 weight 1.818 } host storage006 { id -13 # do not change unnecessarily id -14 class hdd # do not change unnecessarily # weight 20.003 alg straw2 hash 0 # rjenkins1 item osd.51 weight 1.818 item osd.55 weight 1.818 item osd.58 weight 1.818 item osd.61 weight 1.818 item osd.63 weight 1.818 item osd.65 weight 1.818 item osd.66 weight 1.818 item osd.69 weight 1.818 item osd.71 weight 1.818 item osd.73 weight 1.818 item osd.75 weight 1.818 } host storage007 { id -15 # do not change unnecessarily id -16 class hdd # do not change unnecessarily # weight 20.003 alg straw2 hash 0 # rjenkins1 item osd.52 weight 1.818 item osd.57 weight 1.818 item osd.60 weight 1.818 item osd.62 weight 1.818 item osd.64 weight 1.818 item osd.67 weight 1.818 item osd.70 weight 1.818 item osd.68 weight 1.818 item osd.72 weight 1.818 item osd.74 weight 1.818 item osd.76 weight 1.818 } root default { id -1 # do not change unnecessarily id -2 class hdd # do not change unnecessarily # weight 140.024 alg straw2 hash 0 # rjenkins1 item storage004 weight 20.003 item storage005 weight 20.003 item storage002 weight 20.003 item storage001 weight 20.003 item storage003 weight 20.003 item storage006 weight 20.003 item storage007 weight 20.003 } # rules rule replicated_rule { id 0 type replicated min_size 1 max_size 10 step take default step chooseleaf firstn 0 type host step emit } # end crush map _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com