For this pool I have set EC 3+2 (so in total I have 5 nodes) which one was temporarily removed, but maybe this was the problem? On Thu, May 27, 2021 at 3:51 PM Rok Jaklič <rjaklic@xxxxxxxxx> wrote: > Hi, thanks for quick reply > > root@ctplmon1:~# ceph pg dump pgs_brief | grep undersized > dumped pgs_brief > 9.5 active+undersized+degraded [72,85,54,120,2147483647] > 72 [72,85,54,120,2147483647] 72 > 9.6 active+undersized+degraded [101,47,113,74,2147483647] > 101 [101,47,113,74,2147483647] 101 > 9.2 active+undersized+degraded [86,118,74,2147483647,49] > 86 [86,118,74,2147483647,49] 86 > 9.d active+undersized+degraded [49,136,83,90,2147483647] > 49 [49,136,83,90,2147483647] 49 > 9.f active+undersized+degraded [55,103,81,128,2147483647] > 55 [55,103,81,128,2147483647] 55 > 9.18 active+undersized+degraded [115,50,61,89,2147483647] > 115 [115,50,61,89,2147483647] 115 > 9.1d active+undersized+degraded [61,90,31,2147483647,125] > 61 [61,90,31,2147483647,125] 61 > 9.10 active+undersized+degraded [46,2147483647,71,86,122] > 46 [46,2147483647,71,86,122] 46 > 9.17 active+undersized+degraded [60,95,114,2147483647,48] > 60 [60,95,114,2147483647,48] 60 > 9.15 active+undersized+degraded [121,76,30,101,2147483647] > 121 [121,76,30,101,2147483647] 121 > root@ctplmon1:~# ceph osd tree > ID CLASS WEIGHT TYPE NAME STATUS REWEIGHT PRI-AFF > -1 764.11981 root default > -3 152.82378 host ctplosd1 > 0 hdd 5.45798 osd.0 down 0 1.00000 > 1 hdd 5.45799 osd.1 down 0 1.00000 > 2 hdd 5.45799 osd.2 down 0 1.00000 > 3 hdd 5.45799 osd.3 down 0 1.00000 > 4 hdd 5.45799 osd.4 down 0 1.00000 > 5 hdd 5.45799 osd.5 down 0 1.00000 > 6 hdd 5.45799 osd.6 down 0 1.00000 > 7 hdd 5.45799 osd.7 down 0 1.00000 > 8 hdd 5.45799 osd.8 down 0 1.00000 > 9 hdd 5.45799 osd.9 down 0 1.00000 > 10 hdd 5.45799 osd.10 down 0 1.00000 > 11 hdd 5.45799 osd.11 down 0 1.00000 > 12 hdd 5.45799 osd.12 down 0 1.00000 > 13 hdd 5.45799 osd.13 down 0 1.00000 > 14 hdd 5.45799 osd.14 down 0 1.00000 > 15 hdd 5.45799 osd.15 down 0 1.00000 > 16 hdd 5.45799 osd.16 down 0 1.00000 > 17 hdd 5.45799 osd.17 down 0 1.00000 > 18 hdd 5.45799 osd.18 down 0 1.00000 > 19 hdd 5.45799 osd.19 down 0 1.00000 > 20 hdd 5.45799 osd.20 down 0 1.00000 > 21 hdd 5.45799 osd.21 down 0 1.00000 > 22 hdd 5.45799 osd.22 down 0 1.00000 > 23 hdd 5.45799 osd.23 down 0 1.00000 > 24 hdd 5.45799 osd.24 down 0 1.00000 > 25 hdd 5.45799 osd.25 down 0 1.00000 > 26 hdd 5.45799 osd.26 down 0 1.00000 > 27 hdd 5.45799 osd.27 down 0 1.00000 > -11 152.82401 host ctplosd5 > 112 hdd 5.45799 osd.112 up 1.00000 1.00000 > 113 hdd 5.45799 osd.113 up 1.00000 1.00000 > 114 hdd 5.45799 osd.114 up 1.00000 1.00000 > 115 hdd 5.45799 osd.115 up 1.00000 1.00000 > 116 hdd 5.45799 osd.116 up 1.00000 1.00000 > 117 hdd 5.45799 osd.117 up 1.00000 1.00000 > 118 hdd 5.45799 osd.118 up 1.00000 1.00000 > 119 hdd 5.45799 osd.119 up 1.00000 1.00000 > 120 hdd 5.45799 osd.120 up 1.00000 1.00000 > 121 hdd 5.45799 osd.121 up 1.00000 1.00000 > 122 hdd 5.45799 osd.122 up 1.00000 1.00000 > 123 hdd 5.45799 osd.123 up 1.00000 1.00000 > 124 hdd 5.45799 osd.124 up 1.00000 1.00000 > 125 hdd 5.45799 osd.125 up 1.00000 1.00000 > 126 hdd 5.45799 osd.126 up 1.00000 1.00000 > 127 hdd 5.45799 osd.127 up 1.00000 1.00000 > 128 hdd 5.45799 osd.128 up 1.00000 1.00000 > 129 hdd 5.45799 osd.129 up 1.00000 1.00000 > 130 hdd 5.45799 osd.130 up 1.00000 1.00000 > 131 hdd 5.45799 osd.131 up 1.00000 1.00000 > 132 hdd 5.45799 osd.132 up 1.00000 1.00000 > 133 hdd 5.45799 osd.133 up 1.00000 1.00000 > 134 hdd 5.45799 osd.134 up 1.00000 1.00000 > 135 hdd 5.45799 osd.135 up 1.00000 1.00000 > 136 hdd 5.45799 osd.136 up 1.00000 1.00000 > 137 hdd 5.45799 osd.137 up 1.00000 1.00000 > 138 hdd 5.45799 osd.138 up 1.00000 1.00000 > 139 hdd 5.45799 osd.139 up 1.00000 1.00000 > -7 152.82401 host ctplosd6 > 57 hdd 5.45799 osd.57 up 1.00000 1.00000 > 58 hdd 5.45799 osd.58 up 1.00000 1.00000 > 59 hdd 5.45799 osd.59 up 1.00000 1.00000 > 60 hdd 5.45799 osd.60 up 1.00000 1.00000 > 61 hdd 5.45799 osd.61 up 1.00000 1.00000 > 62 hdd 5.45799 osd.62 up 1.00000 1.00000 > 63 hdd 5.45799 osd.63 up 1.00000 1.00000 > 64 hdd 5.45799 osd.64 up 1.00000 1.00000 > 65 hdd 5.45799 osd.65 up 1.00000 1.00000 > 66 hdd 5.45799 osd.66 up 1.00000 1.00000 > 67 hdd 5.45799 osd.67 up 1.00000 1.00000 > 68 hdd 5.45799 osd.68 up 1.00000 1.00000 > 69 hdd 5.45799 osd.69 up 1.00000 1.00000 > 70 hdd 5.45799 osd.70 up 1.00000 1.00000 > 71 hdd 5.45799 osd.71 up 1.00000 1.00000 > 72 hdd 5.45799 osd.72 up 1.00000 1.00000 > 73 hdd 5.45799 osd.73 up 1.00000 1.00000 > 74 hdd 5.45799 osd.74 up 1.00000 1.00000 > 75 hdd 5.45799 osd.75 up 1.00000 1.00000 > 76 hdd 5.45799 osd.76 up 1.00000 1.00000 > 77 hdd 5.45799 osd.77 up 1.00000 1.00000 > 78 hdd 5.45799 osd.78 up 1.00000 1.00000 > 79 hdd 5.45799 osd.79 up 1.00000 1.00000 > 80 hdd 5.45799 osd.80 up 1.00000 1.00000 > 81 hdd 5.45799 osd.81 up 1.00000 1.00000 > 82 hdd 5.45799 osd.82 up 1.00000 1.00000 > 83 hdd 5.45799 osd.83 up 1.00000 1.00000 > 84 hdd 5.45799 osd.84 up 1.00000 1.00000 > -5 152.82401 host ctplosd7 > 28 hdd 5.45799 osd.28 up 1.00000 1.00000 > 29 hdd 5.45799 osd.29 up 1.00000 1.00000 > 30 hdd 5.45799 osd.30 up 1.00000 1.00000 > 31 hdd 5.45799 osd.31 up 1.00000 1.00000 > 32 hdd 5.45799 osd.32 up 1.00000 1.00000 > 33 hdd 5.45799 osd.33 up 1.00000 1.00000 > 34 hdd 5.45799 osd.34 up 1.00000 1.00000 > 35 hdd 5.45799 osd.35 up 1.00000 1.00000 > 36 hdd 5.45799 osd.36 up 1.00000 1.00000 > 37 hdd 5.45799 osd.37 up 1.00000 1.00000 > 38 hdd 5.45799 osd.38 up 1.00000 1.00000 > 39 hdd 5.45799 osd.39 up 1.00000 1.00000 > 40 hdd 5.45799 osd.40 up 1.00000 1.00000 > 41 hdd 5.45799 osd.41 up 1.00000 1.00000 > 42 hdd 5.45799 osd.42 up 1.00000 1.00000 > 43 hdd 5.45799 osd.43 up 1.00000 1.00000 > 44 hdd 5.45799 osd.44 up 1.00000 1.00000 > 45 hdd 5.45799 osd.45 up 1.00000 1.00000 > 46 hdd 5.45799 osd.46 up 1.00000 1.00000 > 47 hdd 5.45799 osd.47 up 1.00000 1.00000 > 48 hdd 5.45799 osd.48 up 1.00000 1.00000 > 49 hdd 5.45799 osd.49 up 1.00000 1.00000 > 50 hdd 5.45799 osd.50 up 1.00000 1.00000 > 51 hdd 5.45799 osd.51 up 1.00000 1.00000 > 52 hdd 5.45799 osd.52 up 1.00000 1.00000 > 53 hdd 5.45799 osd.53 up 1.00000 1.00000 > 54 hdd 5.45799 osd.54 up 1.00000 1.00000 > 55 hdd 5.45799 osd.55 up 1.00000 1.00000 > -9 152.82401 host ctplosd8 > 56 hdd 5.45799 osd.56 up 1.00000 1.00000 > 85 hdd 5.45799 osd.85 up 1.00000 1.00000 > 86 hdd 5.45799 osd.86 up 1.00000 1.00000 > 87 hdd 5.45799 osd.87 up 1.00000 1.00000 > 88 hdd 5.45799 osd.88 up 1.00000 1.00000 > 89 hdd 5.45799 osd.89 up 1.00000 1.00000 > 90 hdd 5.45799 osd.90 up 1.00000 1.00000 > 91 hdd 5.45799 osd.91 up 1.00000 1.00000 > 92 hdd 5.45799 osd.92 up 1.00000 1.00000 > 93 hdd 5.45799 osd.93 up 1.00000 1.00000 > 94 hdd 5.45799 osd.94 up 1.00000 1.00000 > 95 hdd 5.45799 osd.95 up 1.00000 1.00000 > 96 hdd 5.45799 osd.96 up 1.00000 1.00000 > 97 hdd 5.45799 osd.97 up 1.00000 1.00000 > 98 hdd 5.45799 osd.98 up 1.00000 1.00000 > 99 hdd 5.45799 osd.99 up 1.00000 1.00000 > 100 hdd 5.45799 osd.100 up 1.00000 1.00000 > 101 hdd 5.45799 osd.101 up 1.00000 1.00000 > 102 hdd 5.45799 osd.102 up 1.00000 1.00000 > 103 hdd 5.45799 osd.103 up 1.00000 1.00000 > 104 hdd 5.45799 osd.104 up 1.00000 1.00000 > 105 hdd 5.45799 osd.105 up 1.00000 1.00000 > 106 hdd 5.45799 osd.106 up 1.00000 1.00000 > 107 hdd 5.45799 osd.107 up 1.00000 1.00000 > 108 hdd 5.45799 osd.108 up 1.00000 1.00000 > 109 hdd 5.45799 osd.109 up 1.00000 1.00000 > 110 hdd 5.45799 osd.110 up 1.00000 1.00000 > 111 hdd 5.45799 osd.111 up 1.00000 1.00000 > root@ctplmon1:~# ceph osd pool ls detail > pool 9 'default.rgw.buckets.data' erasure profile ec-32-profile size 5 > min_size 4 crush_rule 1 object_hash rjenkins pg_num 32 pgp_num 32 > autoscale_mode on last_change 128267 lfor 0/127784/127779 flags > hashpspool,ec_overwrites stripe_width 12288 application rgw > > ---- > > The affected pool is pool number 9 and host is ctplosd1. This is the host > I removed at the first place (to reinstall OS) and now I added this host > back to the cluster, but osds on this host cannot be brought back to up > state for some reason, even though osd processes are running on the host. > > Kind regards, > Rok > > > > > > On Thu, May 27, 2021 at 3:32 PM Eugen Block <eblock@xxxxxx> wrote: > >> Hi, >> >> this sounds like your crush rule(s) for one or more pools can't place >> the PGs because the host is missing. Please share >> >> ceph pg dump pgs_brief | grep undersized >> ceph osd tree >> ceph osd pool ls detail >> >> and the crush rule(s) for the affected pool(s). >> >> >> Zitat von Rok Jaklič <rjaklic@xxxxxxxxx>: >> >> > Hi, >> > >> > I have removed one node, but now ceph seems to stuck in: >> > Degraded data redundancy: 67/2393 objects degraded (2.800%), 12 pgs >> > degraded, 12 pgs undersized >> > >> > How to "force" rebalancing? Or should I just wait a little bit more? >> > >> > Kind regards, >> > rok >> > _______________________________________________ >> > ceph-users mailing list -- ceph-users@xxxxxxx >> > To unsubscribe send an email to ceph-users-leave@xxxxxxx >> >> >> _______________________________________________ >> ceph-users mailing list -- ceph-users@xxxxxxx >> To unsubscribe send an email to ceph-users-leave@xxxxxxx >> > _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx