Hi, thanks for quick reply
root@ctplmon1:~# ceph pg dump pgs_brief | grep undersized
dumped pgs_brief
9.5 active+undersized+degraded [72,85,54,120,2147483647]
72 [72,85,54,120,2147483647] 72
9.6 active+undersized+degraded [101,47,113,74,2147483647]
101 [101,47,113,74,2147483647] 101
9.2 active+undersized+degraded [86,118,74,2147483647,49]
86 [86,118,74,2147483647,49] 86
9.d active+undersized+degraded [49,136,83,90,2147483647]
49 [49,136,83,90,2147483647] 49
9.f active+undersized+degraded [55,103,81,128,2147483647]
55 [55,103,81,128,2147483647] 55
9.18 active+undersized+degraded [115,50,61,89,2147483647]
115 [115,50,61,89,2147483647] 115
9.1d active+undersized+degraded [61,90,31,2147483647,125]
61 [61,90,31,2147483647,125] 61
9.10 active+undersized+degraded [46,2147483647,71,86,122]
46 [46,2147483647,71,86,122] 46
9.17 active+undersized+degraded [60,95,114,2147483647,48]
60 [60,95,114,2147483647,48] 60
9.15 active+undersized+degraded [121,76,30,101,2147483647]
121 [121,76,30,101,2147483647] 121
root@ctplmon1:~# ceph osd tree
ID CLASS WEIGHT TYPE NAME STATUS REWEIGHT PRI-AFF
-1 764.11981 root default
-3 152.82378 host ctplosd1
0 hdd 5.45798 osd.0 down 0 1.00000
1 hdd 5.45799 osd.1 down 0 1.00000
2 hdd 5.45799 osd.2 down 0 1.00000
3 hdd 5.45799 osd.3 down 0 1.00000
4 hdd 5.45799 osd.4 down 0 1.00000
5 hdd 5.45799 osd.5 down 0 1.00000
6 hdd 5.45799 osd.6 down 0 1.00000
7 hdd 5.45799 osd.7 down 0 1.00000
8 hdd 5.45799 osd.8 down 0 1.00000
9 hdd 5.45799 osd.9 down 0 1.00000
10 hdd 5.45799 osd.10 down 0 1.00000
11 hdd 5.45799 osd.11 down 0 1.00000
12 hdd 5.45799 osd.12 down 0 1.00000
13 hdd 5.45799 osd.13 down 0 1.00000
14 hdd 5.45799 osd.14 down 0 1.00000
15 hdd 5.45799 osd.15 down 0 1.00000
16 hdd 5.45799 osd.16 down 0 1.00000
17 hdd 5.45799 osd.17 down 0 1.00000
18 hdd 5.45799 osd.18 down 0 1.00000
19 hdd 5.45799 osd.19 down 0 1.00000
20 hdd 5.45799 osd.20 down 0 1.00000
21 hdd 5.45799 osd.21 down 0 1.00000
22 hdd 5.45799 osd.22 down 0 1.00000
23 hdd 5.45799 osd.23 down 0 1.00000
24 hdd 5.45799 osd.24 down 0 1.00000
25 hdd 5.45799 osd.25 down 0 1.00000
26 hdd 5.45799 osd.26 down 0 1.00000
27 hdd 5.45799 osd.27 down 0 1.00000
-11 152.82401 host ctplosd5
112 hdd 5.45799 osd.112 up 1.00000 1.00000
113 hdd 5.45799 osd.113 up 1.00000 1.00000
114 hdd 5.45799 osd.114 up 1.00000 1.00000
115 hdd 5.45799 osd.115 up 1.00000 1.00000
116 hdd 5.45799 osd.116 up 1.00000 1.00000
117 hdd 5.45799 osd.117 up 1.00000 1.00000
118 hdd 5.45799 osd.118 up 1.00000 1.00000
119 hdd 5.45799 osd.119 up 1.00000 1.00000
120 hdd 5.45799 osd.120 up 1.00000 1.00000
121 hdd 5.45799 osd.121 up 1.00000 1.00000
122 hdd 5.45799 osd.122 up 1.00000 1.00000
123 hdd 5.45799 osd.123 up 1.00000 1.00000
124 hdd 5.45799 osd.124 up 1.00000 1.00000
125 hdd 5.45799 osd.125 up 1.00000 1.00000
126 hdd 5.45799 osd.126 up 1.00000 1.00000
127 hdd 5.45799 osd.127 up 1.00000 1.00000
128 hdd 5.45799 osd.128 up 1.00000 1.00000
129 hdd 5.45799 osd.129 up 1.00000 1.00000
130 hdd 5.45799 osd.130 up 1.00000 1.00000
131 hdd 5.45799 osd.131 up 1.00000 1.00000
132 hdd 5.45799 osd.132 up 1.00000 1.00000
133 hdd 5.45799 osd.133 up 1.00000 1.00000
134 hdd 5.45799 osd.134 up 1.00000 1.00000
135 hdd 5.45799 osd.135 up 1.00000 1.00000
136 hdd 5.45799 osd.136 up 1.00000 1.00000
137 hdd 5.45799 osd.137 up 1.00000 1.00000
138 hdd 5.45799 osd.138 up 1.00000 1.00000
139 hdd 5.45799 osd.139 up 1.00000 1.00000
-7 152.82401 host ctplosd6
57 hdd 5.45799 osd.57 up 1.00000 1.00000
58 hdd 5.45799 osd.58 up 1.00000 1.00000
59 hdd 5.45799 osd.59 up 1.00000 1.00000
60 hdd 5.45799 osd.60 up 1.00000 1.00000
61 hdd 5.45799 osd.61 up 1.00000 1.00000
62 hdd 5.45799 osd.62 up 1.00000 1.00000
63 hdd 5.45799 osd.63 up 1.00000 1.00000
64 hdd 5.45799 osd.64 up 1.00000 1.00000
65 hdd 5.45799 osd.65 up 1.00000 1.00000
66 hdd 5.45799 osd.66 up 1.00000 1.00000
67 hdd 5.45799 osd.67 up 1.00000 1.00000
68 hdd 5.45799 osd.68 up 1.00000 1.00000
69 hdd 5.45799 osd.69 up 1.00000 1.00000
70 hdd 5.45799 osd.70 up 1.00000 1.00000
71 hdd 5.45799 osd.71 up 1.00000 1.00000
72 hdd 5.45799 osd.72 up 1.00000 1.00000
73 hdd 5.45799 osd.73 up 1.00000 1.00000
74 hdd 5.45799 osd.74 up 1.00000 1.00000
75 hdd 5.45799 osd.75 up 1.00000 1.00000
76 hdd 5.45799 osd.76 up 1.00000 1.00000
77 hdd 5.45799 osd.77 up 1.00000 1.00000
78 hdd 5.45799 osd.78 up 1.00000 1.00000
79 hdd 5.45799 osd.79 up 1.00000 1.00000
80 hdd 5.45799 osd.80 up 1.00000 1.00000
81 hdd 5.45799 osd.81 up 1.00000 1.00000
82 hdd 5.45799 osd.82 up 1.00000 1.00000
83 hdd 5.45799 osd.83 up 1.00000 1.00000
84 hdd 5.45799 osd.84 up 1.00000 1.00000
-5 152.82401 host ctplosd7
28 hdd 5.45799 osd.28 up 1.00000 1.00000
29 hdd 5.45799 osd.29 up 1.00000 1.00000
30 hdd 5.45799 osd.30 up 1.00000 1.00000
31 hdd 5.45799 osd.31 up 1.00000 1.00000
32 hdd 5.45799 osd.32 up 1.00000 1.00000
33 hdd 5.45799 osd.33 up 1.00000 1.00000
34 hdd 5.45799 osd.34 up 1.00000 1.00000
35 hdd 5.45799 osd.35 up 1.00000 1.00000
36 hdd 5.45799 osd.36 up 1.00000 1.00000
37 hdd 5.45799 osd.37 up 1.00000 1.00000
38 hdd 5.45799 osd.38 up 1.00000 1.00000
39 hdd 5.45799 osd.39 up 1.00000 1.00000
40 hdd 5.45799 osd.40 up 1.00000 1.00000
41 hdd 5.45799 osd.41 up 1.00000 1.00000
42 hdd 5.45799 osd.42 up 1.00000 1.00000
43 hdd 5.45799 osd.43 up 1.00000 1.00000
44 hdd 5.45799 osd.44 up 1.00000 1.00000
45 hdd 5.45799 osd.45 up 1.00000 1.00000
46 hdd 5.45799 osd.46 up 1.00000 1.00000
47 hdd 5.45799 osd.47 up 1.00000 1.00000
48 hdd 5.45799 osd.48 up 1.00000 1.00000
49 hdd 5.45799 osd.49 up 1.00000 1.00000
50 hdd 5.45799 osd.50 up 1.00000 1.00000
51 hdd 5.45799 osd.51 up 1.00000 1.00000
52 hdd 5.45799 osd.52 up 1.00000 1.00000
53 hdd 5.45799 osd.53 up 1.00000 1.00000
54 hdd 5.45799 osd.54 up 1.00000 1.00000
55 hdd 5.45799 osd.55 up 1.00000 1.00000
-9 152.82401 host ctplosd8
56 hdd 5.45799 osd.56 up 1.00000 1.00000
85 hdd 5.45799 osd.85 up 1.00000 1.00000
86 hdd 5.45799 osd.86 up 1.00000 1.00000
87 hdd 5.45799 osd.87 up 1.00000 1.00000
88 hdd 5.45799 osd.88 up 1.00000 1.00000
89 hdd 5.45799 osd.89 up 1.00000 1.00000
90 hdd 5.45799 osd.90 up 1.00000 1.00000
91 hdd 5.45799 osd.91 up 1.00000 1.00000
92 hdd 5.45799 osd.92 up 1.00000 1.00000
93 hdd 5.45799 osd.93 up 1.00000 1.00000
94 hdd 5.45799 osd.94 up 1.00000 1.00000
95 hdd 5.45799 osd.95 up 1.00000 1.00000
96 hdd 5.45799 osd.96 up 1.00000 1.00000
97 hdd 5.45799 osd.97 up 1.00000 1.00000
98 hdd 5.45799 osd.98 up 1.00000 1.00000
99 hdd 5.45799 osd.99 up 1.00000 1.00000
100 hdd 5.45799 osd.100 up 1.00000 1.00000
101 hdd 5.45799 osd.101 up 1.00000 1.00000
102 hdd 5.45799 osd.102 up 1.00000 1.00000
103 hdd 5.45799 osd.103 up 1.00000 1.00000
104 hdd 5.45799 osd.104 up 1.00000 1.00000
105 hdd 5.45799 osd.105 up 1.00000 1.00000
106 hdd 5.45799 osd.106 up 1.00000 1.00000
107 hdd 5.45799 osd.107 up 1.00000 1.00000
108 hdd 5.45799 osd.108 up 1.00000 1.00000
109 hdd 5.45799 osd.109 up 1.00000 1.00000
110 hdd 5.45799 osd.110 up 1.00000 1.00000
111 hdd 5.45799 osd.111 up 1.00000 1.00000
root@ctplmon1:~# ceph osd pool ls detail
pool 9 'default.rgw.buckets.data' erasure profile ec-32-profile size 5
min_size 4 crush_rule 1 object_hash rjenkins pg_num 32 pgp_num 32
autoscale_mode on last_change 128267 lfor 0/127784/127779 flags
hashpspool,ec_overwrites stripe_width 12288 application rgw
----
The affected pool is pool number 9 and host is ctplosd1. This is the host
I removed at the first place (to reinstall OS) and now I added this host
back to the cluster, but osds on this host cannot be brought back to up
state for some reason, even though osd processes are running on the host.
Kind regards,
Rok
On Thu, May 27, 2021 at 3:32 PM Eugen Block <eblock@xxxxxx> wrote:
Hi,
this sounds like your crush rule(s) for one or more pools can't place
the PGs because the host is missing. Please share
ceph pg dump pgs_brief | grep undersized
ceph osd tree
ceph osd pool ls detail
and the crush rule(s) for the affected pool(s).
Zitat von Rok Jaklič <rjaklic@xxxxxxxxx>:
> Hi,
>
> I have removed one node, but now ceph seems to stuck in:
> Degraded data redundancy: 67/2393 objects degraded (2.800%), 12 pgs
> degraded, 12 pgs undersized
>
> How to "force" rebalancing? Or should I just wait a little bit more?
>
> Kind regards,
> rok
> _______________________________________________
> ceph-users mailing list -- ceph-users@xxxxxxx
> To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx