Re: rebalancing after node more

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Yes, if your pool requires 5 chunks and you only have 5 hosts (with failure domain host) your PGs become undersized when a host fails and won't recover until the OSDs come back. Which ceph version is this?


Zitat von Rok Jaklič <rjaklic@xxxxxxxxx>:

For this pool I have set EC 3+2 (so in total I have 5 nodes) which one was
temporarily removed, but maybe this was the problem?

On Thu, May 27, 2021 at 3:51 PM Rok Jaklič <rjaklic@xxxxxxxxx> wrote:

Hi, thanks for quick reply

root@ctplmon1:~# ceph pg dump pgs_brief | grep undersized
dumped pgs_brief
9.5      active+undersized+degraded   [72,85,54,120,2147483647]
72   [72,85,54,120,2147483647]              72
9.6      active+undersized+degraded  [101,47,113,74,2147483647]
 101  [101,47,113,74,2147483647]             101
9.2      active+undersized+degraded   [86,118,74,2147483647,49]
86   [86,118,74,2147483647,49]              86
9.d      active+undersized+degraded   [49,136,83,90,2147483647]
49   [49,136,83,90,2147483647]              49
9.f      active+undersized+degraded  [55,103,81,128,2147483647]
55  [55,103,81,128,2147483647]              55
9.18     active+undersized+degraded   [115,50,61,89,2147483647]
 115   [115,50,61,89,2147483647]             115
9.1d     active+undersized+degraded   [61,90,31,2147483647,125]
61   [61,90,31,2147483647,125]              61
9.10     active+undersized+degraded   [46,2147483647,71,86,122]
46   [46,2147483647,71,86,122]              46
9.17     active+undersized+degraded   [60,95,114,2147483647,48]
60   [60,95,114,2147483647,48]              60
9.15     active+undersized+degraded  [121,76,30,101,2147483647]
 121  [121,76,30,101,2147483647]             121
root@ctplmon1:~# ceph osd tree
ID   CLASS  WEIGHT     TYPE NAME          STATUS  REWEIGHT  PRI-AFF
 -1         764.11981  root default
 -3         152.82378      host ctplosd1
  0    hdd    5.45798          osd.0        down         0  1.00000
  1    hdd    5.45799          osd.1        down         0  1.00000
  2    hdd    5.45799          osd.2        down         0  1.00000
  3    hdd    5.45799          osd.3        down         0  1.00000
  4    hdd    5.45799          osd.4        down         0  1.00000
  5    hdd    5.45799          osd.5        down         0  1.00000
  6    hdd    5.45799          osd.6        down         0  1.00000
  7    hdd    5.45799          osd.7        down         0  1.00000
  8    hdd    5.45799          osd.8        down         0  1.00000
  9    hdd    5.45799          osd.9        down         0  1.00000
 10    hdd    5.45799          osd.10       down         0  1.00000
 11    hdd    5.45799          osd.11       down         0  1.00000
 12    hdd    5.45799          osd.12       down         0  1.00000
 13    hdd    5.45799          osd.13       down         0  1.00000
 14    hdd    5.45799          osd.14       down         0  1.00000
 15    hdd    5.45799          osd.15       down         0  1.00000
 16    hdd    5.45799          osd.16       down         0  1.00000
 17    hdd    5.45799          osd.17       down         0  1.00000
 18    hdd    5.45799          osd.18       down         0  1.00000
 19    hdd    5.45799          osd.19       down         0  1.00000
 20    hdd    5.45799          osd.20       down         0  1.00000
 21    hdd    5.45799          osd.21       down         0  1.00000
 22    hdd    5.45799          osd.22       down         0  1.00000
 23    hdd    5.45799          osd.23       down         0  1.00000
 24    hdd    5.45799          osd.24       down         0  1.00000
 25    hdd    5.45799          osd.25       down         0  1.00000
 26    hdd    5.45799          osd.26       down         0  1.00000
 27    hdd    5.45799          osd.27       down         0  1.00000
-11         152.82401      host ctplosd5
112    hdd    5.45799          osd.112        up   1.00000  1.00000
113    hdd    5.45799          osd.113        up   1.00000  1.00000
114    hdd    5.45799          osd.114        up   1.00000  1.00000
115    hdd    5.45799          osd.115        up   1.00000  1.00000
116    hdd    5.45799          osd.116        up   1.00000  1.00000
117    hdd    5.45799          osd.117        up   1.00000  1.00000
118    hdd    5.45799          osd.118        up   1.00000  1.00000
119    hdd    5.45799          osd.119        up   1.00000  1.00000
120    hdd    5.45799          osd.120        up   1.00000  1.00000
121    hdd    5.45799          osd.121        up   1.00000  1.00000
122    hdd    5.45799          osd.122        up   1.00000  1.00000
123    hdd    5.45799          osd.123        up   1.00000  1.00000
124    hdd    5.45799          osd.124        up   1.00000  1.00000
125    hdd    5.45799          osd.125        up   1.00000  1.00000
126    hdd    5.45799          osd.126        up   1.00000  1.00000
127    hdd    5.45799          osd.127        up   1.00000  1.00000
128    hdd    5.45799          osd.128        up   1.00000  1.00000
129    hdd    5.45799          osd.129        up   1.00000  1.00000
130    hdd    5.45799          osd.130        up   1.00000  1.00000
131    hdd    5.45799          osd.131        up   1.00000  1.00000
132    hdd    5.45799          osd.132        up   1.00000  1.00000
133    hdd    5.45799          osd.133        up   1.00000  1.00000
134    hdd    5.45799          osd.134        up   1.00000  1.00000
135    hdd    5.45799          osd.135        up   1.00000  1.00000
136    hdd    5.45799          osd.136        up   1.00000  1.00000
137    hdd    5.45799          osd.137        up   1.00000  1.00000
138    hdd    5.45799          osd.138        up   1.00000  1.00000
139    hdd    5.45799          osd.139        up   1.00000  1.00000
 -7         152.82401      host ctplosd6
 57    hdd    5.45799          osd.57         up   1.00000  1.00000
 58    hdd    5.45799          osd.58         up   1.00000  1.00000
 59    hdd    5.45799          osd.59         up   1.00000  1.00000
 60    hdd    5.45799          osd.60         up   1.00000  1.00000
 61    hdd    5.45799          osd.61         up   1.00000  1.00000
 62    hdd    5.45799          osd.62         up   1.00000  1.00000
 63    hdd    5.45799          osd.63         up   1.00000  1.00000
 64    hdd    5.45799          osd.64         up   1.00000  1.00000
 65    hdd    5.45799          osd.65         up   1.00000  1.00000
 66    hdd    5.45799          osd.66         up   1.00000  1.00000
 67    hdd    5.45799          osd.67         up   1.00000  1.00000
 68    hdd    5.45799          osd.68         up   1.00000  1.00000
 69    hdd    5.45799          osd.69         up   1.00000  1.00000
 70    hdd    5.45799          osd.70         up   1.00000  1.00000
 71    hdd    5.45799          osd.71         up   1.00000  1.00000
 72    hdd    5.45799          osd.72         up   1.00000  1.00000
 73    hdd    5.45799          osd.73         up   1.00000  1.00000
 74    hdd    5.45799          osd.74         up   1.00000  1.00000
 75    hdd    5.45799          osd.75         up   1.00000  1.00000
 76    hdd    5.45799          osd.76         up   1.00000  1.00000
 77    hdd    5.45799          osd.77         up   1.00000  1.00000
 78    hdd    5.45799          osd.78         up   1.00000  1.00000
 79    hdd    5.45799          osd.79         up   1.00000  1.00000
 80    hdd    5.45799          osd.80         up   1.00000  1.00000
 81    hdd    5.45799          osd.81         up   1.00000  1.00000
 82    hdd    5.45799          osd.82         up   1.00000  1.00000
 83    hdd    5.45799          osd.83         up   1.00000  1.00000
 84    hdd    5.45799          osd.84         up   1.00000  1.00000
 -5         152.82401      host ctplosd7
 28    hdd    5.45799          osd.28         up   1.00000  1.00000
 29    hdd    5.45799          osd.29         up   1.00000  1.00000
 30    hdd    5.45799          osd.30         up   1.00000  1.00000
 31    hdd    5.45799          osd.31         up   1.00000  1.00000
 32    hdd    5.45799          osd.32         up   1.00000  1.00000
 33    hdd    5.45799          osd.33         up   1.00000  1.00000
 34    hdd    5.45799          osd.34         up   1.00000  1.00000
 35    hdd    5.45799          osd.35         up   1.00000  1.00000
 36    hdd    5.45799          osd.36         up   1.00000  1.00000
 37    hdd    5.45799          osd.37         up   1.00000  1.00000
 38    hdd    5.45799          osd.38         up   1.00000  1.00000
 39    hdd    5.45799          osd.39         up   1.00000  1.00000
 40    hdd    5.45799          osd.40         up   1.00000  1.00000
 41    hdd    5.45799          osd.41         up   1.00000  1.00000
 42    hdd    5.45799          osd.42         up   1.00000  1.00000
 43    hdd    5.45799          osd.43         up   1.00000  1.00000
 44    hdd    5.45799          osd.44         up   1.00000  1.00000
 45    hdd    5.45799          osd.45         up   1.00000  1.00000
 46    hdd    5.45799          osd.46         up   1.00000  1.00000
 47    hdd    5.45799          osd.47         up   1.00000  1.00000
 48    hdd    5.45799          osd.48         up   1.00000  1.00000
 49    hdd    5.45799          osd.49         up   1.00000  1.00000
 50    hdd    5.45799          osd.50         up   1.00000  1.00000
 51    hdd    5.45799          osd.51         up   1.00000  1.00000
 52    hdd    5.45799          osd.52         up   1.00000  1.00000
 53    hdd    5.45799          osd.53         up   1.00000  1.00000
 54    hdd    5.45799          osd.54         up   1.00000  1.00000
 55    hdd    5.45799          osd.55         up   1.00000  1.00000
 -9         152.82401      host ctplosd8
 56    hdd    5.45799          osd.56         up   1.00000  1.00000
 85    hdd    5.45799          osd.85         up   1.00000  1.00000
 86    hdd    5.45799          osd.86         up   1.00000  1.00000
 87    hdd    5.45799          osd.87         up   1.00000  1.00000
 88    hdd    5.45799          osd.88         up   1.00000  1.00000
 89    hdd    5.45799          osd.89         up   1.00000  1.00000
 90    hdd    5.45799          osd.90         up   1.00000  1.00000
 91    hdd    5.45799          osd.91         up   1.00000  1.00000
 92    hdd    5.45799          osd.92         up   1.00000  1.00000
 93    hdd    5.45799          osd.93         up   1.00000  1.00000
 94    hdd    5.45799          osd.94         up   1.00000  1.00000
 95    hdd    5.45799          osd.95         up   1.00000  1.00000
 96    hdd    5.45799          osd.96         up   1.00000  1.00000
 97    hdd    5.45799          osd.97         up   1.00000  1.00000
 98    hdd    5.45799          osd.98         up   1.00000  1.00000
 99    hdd    5.45799          osd.99         up   1.00000  1.00000
100    hdd    5.45799          osd.100        up   1.00000  1.00000
101    hdd    5.45799          osd.101        up   1.00000  1.00000
102    hdd    5.45799          osd.102        up   1.00000  1.00000
103    hdd    5.45799          osd.103        up   1.00000  1.00000
104    hdd    5.45799          osd.104        up   1.00000  1.00000
105    hdd    5.45799          osd.105        up   1.00000  1.00000
106    hdd    5.45799          osd.106        up   1.00000  1.00000
107    hdd    5.45799          osd.107        up   1.00000  1.00000
108    hdd    5.45799          osd.108        up   1.00000  1.00000
109    hdd    5.45799          osd.109        up   1.00000  1.00000
110    hdd    5.45799          osd.110        up   1.00000  1.00000
111    hdd    5.45799          osd.111        up   1.00000  1.00000
root@ctplmon1:~# ceph osd pool ls detail
pool 9 'default.rgw.buckets.data' erasure profile ec-32-profile size 5
min_size 4 crush_rule 1 object_hash rjenkins pg_num 32 pgp_num 32
autoscale_mode on last_change 128267 lfor 0/127784/127779 flags
hashpspool,ec_overwrites stripe_width 12288 application rgw

----

The affected pool is pool number 9 and host is ctplosd1. This is the host
I removed at the first place (to reinstall OS) and now I added this host
back to the cluster, but osds on this host cannot be brought back to up
state for some reason, even though osd processes are running on the host.

Kind regards,
Rok





On Thu, May 27, 2021 at 3:32 PM Eugen Block <eblock@xxxxxx> wrote:

Hi,

this sounds like your crush rule(s) for one or more pools can't place
the PGs because the host is missing. Please share

ceph pg dump pgs_brief | grep undersized
ceph osd tree
ceph osd pool ls detail

and the crush rule(s) for the affected pool(s).


Zitat von Rok Jaklič <rjaklic@xxxxxxxxx>:

> Hi,
>
> I have removed one node, but now ceph seems to stuck in:
> Degraded data redundancy: 67/2393 objects degraded (2.800%), 12 pgs
> degraded, 12 pgs undersized
>
> How to "force" rebalancing? Or should I just wait a little bit more?
>
> Kind regards,
> rok
> _______________________________________________
> ceph-users mailing list -- ceph-users@xxxxxxx
> To unsubscribe send an email to ceph-users-leave@xxxxxxx


_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux