Re: rebalancing after node more

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



For this pool I have set EC 3+2 (so in total I have 5 nodes) which one was
temporarily removed, but maybe this was the problem?

On Thu, May 27, 2021 at 3:51 PM Rok Jaklič <rjaklic@xxxxxxxxx> wrote:

> Hi, thanks for quick reply
>
> root@ctplmon1:~# ceph pg dump pgs_brief | grep undersized
> dumped pgs_brief
> 9.5      active+undersized+degraded   [72,85,54,120,2147483647]
> 72   [72,85,54,120,2147483647]              72
> 9.6      active+undersized+degraded  [101,47,113,74,2147483647]
>  101  [101,47,113,74,2147483647]             101
> 9.2      active+undersized+degraded   [86,118,74,2147483647,49]
> 86   [86,118,74,2147483647,49]              86
> 9.d      active+undersized+degraded   [49,136,83,90,2147483647]
> 49   [49,136,83,90,2147483647]              49
> 9.f      active+undersized+degraded  [55,103,81,128,2147483647]
> 55  [55,103,81,128,2147483647]              55
> 9.18     active+undersized+degraded   [115,50,61,89,2147483647]
>  115   [115,50,61,89,2147483647]             115
> 9.1d     active+undersized+degraded   [61,90,31,2147483647,125]
> 61   [61,90,31,2147483647,125]              61
> 9.10     active+undersized+degraded   [46,2147483647,71,86,122]
> 46   [46,2147483647,71,86,122]              46
> 9.17     active+undersized+degraded   [60,95,114,2147483647,48]
> 60   [60,95,114,2147483647,48]              60
> 9.15     active+undersized+degraded  [121,76,30,101,2147483647]
>  121  [121,76,30,101,2147483647]             121
> root@ctplmon1:~# ceph osd tree
> ID   CLASS  WEIGHT     TYPE NAME          STATUS  REWEIGHT  PRI-AFF
>  -1         764.11981  root default
>  -3         152.82378      host ctplosd1
>   0    hdd    5.45798          osd.0        down         0  1.00000
>   1    hdd    5.45799          osd.1        down         0  1.00000
>   2    hdd    5.45799          osd.2        down         0  1.00000
>   3    hdd    5.45799          osd.3        down         0  1.00000
>   4    hdd    5.45799          osd.4        down         0  1.00000
>   5    hdd    5.45799          osd.5        down         0  1.00000
>   6    hdd    5.45799          osd.6        down         0  1.00000
>   7    hdd    5.45799          osd.7        down         0  1.00000
>   8    hdd    5.45799          osd.8        down         0  1.00000
>   9    hdd    5.45799          osd.9        down         0  1.00000
>  10    hdd    5.45799          osd.10       down         0  1.00000
>  11    hdd    5.45799          osd.11       down         0  1.00000
>  12    hdd    5.45799          osd.12       down         0  1.00000
>  13    hdd    5.45799          osd.13       down         0  1.00000
>  14    hdd    5.45799          osd.14       down         0  1.00000
>  15    hdd    5.45799          osd.15       down         0  1.00000
>  16    hdd    5.45799          osd.16       down         0  1.00000
>  17    hdd    5.45799          osd.17       down         0  1.00000
>  18    hdd    5.45799          osd.18       down         0  1.00000
>  19    hdd    5.45799          osd.19       down         0  1.00000
>  20    hdd    5.45799          osd.20       down         0  1.00000
>  21    hdd    5.45799          osd.21       down         0  1.00000
>  22    hdd    5.45799          osd.22       down         0  1.00000
>  23    hdd    5.45799          osd.23       down         0  1.00000
>  24    hdd    5.45799          osd.24       down         0  1.00000
>  25    hdd    5.45799          osd.25       down         0  1.00000
>  26    hdd    5.45799          osd.26       down         0  1.00000
>  27    hdd    5.45799          osd.27       down         0  1.00000
> -11         152.82401      host ctplosd5
> 112    hdd    5.45799          osd.112        up   1.00000  1.00000
> 113    hdd    5.45799          osd.113        up   1.00000  1.00000
> 114    hdd    5.45799          osd.114        up   1.00000  1.00000
> 115    hdd    5.45799          osd.115        up   1.00000  1.00000
> 116    hdd    5.45799          osd.116        up   1.00000  1.00000
> 117    hdd    5.45799          osd.117        up   1.00000  1.00000
> 118    hdd    5.45799          osd.118        up   1.00000  1.00000
> 119    hdd    5.45799          osd.119        up   1.00000  1.00000
> 120    hdd    5.45799          osd.120        up   1.00000  1.00000
> 121    hdd    5.45799          osd.121        up   1.00000  1.00000
> 122    hdd    5.45799          osd.122        up   1.00000  1.00000
> 123    hdd    5.45799          osd.123        up   1.00000  1.00000
> 124    hdd    5.45799          osd.124        up   1.00000  1.00000
> 125    hdd    5.45799          osd.125        up   1.00000  1.00000
> 126    hdd    5.45799          osd.126        up   1.00000  1.00000
> 127    hdd    5.45799          osd.127        up   1.00000  1.00000
> 128    hdd    5.45799          osd.128        up   1.00000  1.00000
> 129    hdd    5.45799          osd.129        up   1.00000  1.00000
> 130    hdd    5.45799          osd.130        up   1.00000  1.00000
> 131    hdd    5.45799          osd.131        up   1.00000  1.00000
> 132    hdd    5.45799          osd.132        up   1.00000  1.00000
> 133    hdd    5.45799          osd.133        up   1.00000  1.00000
> 134    hdd    5.45799          osd.134        up   1.00000  1.00000
> 135    hdd    5.45799          osd.135        up   1.00000  1.00000
> 136    hdd    5.45799          osd.136        up   1.00000  1.00000
> 137    hdd    5.45799          osd.137        up   1.00000  1.00000
> 138    hdd    5.45799          osd.138        up   1.00000  1.00000
> 139    hdd    5.45799          osd.139        up   1.00000  1.00000
>  -7         152.82401      host ctplosd6
>  57    hdd    5.45799          osd.57         up   1.00000  1.00000
>  58    hdd    5.45799          osd.58         up   1.00000  1.00000
>  59    hdd    5.45799          osd.59         up   1.00000  1.00000
>  60    hdd    5.45799          osd.60         up   1.00000  1.00000
>  61    hdd    5.45799          osd.61         up   1.00000  1.00000
>  62    hdd    5.45799          osd.62         up   1.00000  1.00000
>  63    hdd    5.45799          osd.63         up   1.00000  1.00000
>  64    hdd    5.45799          osd.64         up   1.00000  1.00000
>  65    hdd    5.45799          osd.65         up   1.00000  1.00000
>  66    hdd    5.45799          osd.66         up   1.00000  1.00000
>  67    hdd    5.45799          osd.67         up   1.00000  1.00000
>  68    hdd    5.45799          osd.68         up   1.00000  1.00000
>  69    hdd    5.45799          osd.69         up   1.00000  1.00000
>  70    hdd    5.45799          osd.70         up   1.00000  1.00000
>  71    hdd    5.45799          osd.71         up   1.00000  1.00000
>  72    hdd    5.45799          osd.72         up   1.00000  1.00000
>  73    hdd    5.45799          osd.73         up   1.00000  1.00000
>  74    hdd    5.45799          osd.74         up   1.00000  1.00000
>  75    hdd    5.45799          osd.75         up   1.00000  1.00000
>  76    hdd    5.45799          osd.76         up   1.00000  1.00000
>  77    hdd    5.45799          osd.77         up   1.00000  1.00000
>  78    hdd    5.45799          osd.78         up   1.00000  1.00000
>  79    hdd    5.45799          osd.79         up   1.00000  1.00000
>  80    hdd    5.45799          osd.80         up   1.00000  1.00000
>  81    hdd    5.45799          osd.81         up   1.00000  1.00000
>  82    hdd    5.45799          osd.82         up   1.00000  1.00000
>  83    hdd    5.45799          osd.83         up   1.00000  1.00000
>  84    hdd    5.45799          osd.84         up   1.00000  1.00000
>  -5         152.82401      host ctplosd7
>  28    hdd    5.45799          osd.28         up   1.00000  1.00000
>  29    hdd    5.45799          osd.29         up   1.00000  1.00000
>  30    hdd    5.45799          osd.30         up   1.00000  1.00000
>  31    hdd    5.45799          osd.31         up   1.00000  1.00000
>  32    hdd    5.45799          osd.32         up   1.00000  1.00000
>  33    hdd    5.45799          osd.33         up   1.00000  1.00000
>  34    hdd    5.45799          osd.34         up   1.00000  1.00000
>  35    hdd    5.45799          osd.35         up   1.00000  1.00000
>  36    hdd    5.45799          osd.36         up   1.00000  1.00000
>  37    hdd    5.45799          osd.37         up   1.00000  1.00000
>  38    hdd    5.45799          osd.38         up   1.00000  1.00000
>  39    hdd    5.45799          osd.39         up   1.00000  1.00000
>  40    hdd    5.45799          osd.40         up   1.00000  1.00000
>  41    hdd    5.45799          osd.41         up   1.00000  1.00000
>  42    hdd    5.45799          osd.42         up   1.00000  1.00000
>  43    hdd    5.45799          osd.43         up   1.00000  1.00000
>  44    hdd    5.45799          osd.44         up   1.00000  1.00000
>  45    hdd    5.45799          osd.45         up   1.00000  1.00000
>  46    hdd    5.45799          osd.46         up   1.00000  1.00000
>  47    hdd    5.45799          osd.47         up   1.00000  1.00000
>  48    hdd    5.45799          osd.48         up   1.00000  1.00000
>  49    hdd    5.45799          osd.49         up   1.00000  1.00000
>  50    hdd    5.45799          osd.50         up   1.00000  1.00000
>  51    hdd    5.45799          osd.51         up   1.00000  1.00000
>  52    hdd    5.45799          osd.52         up   1.00000  1.00000
>  53    hdd    5.45799          osd.53         up   1.00000  1.00000
>  54    hdd    5.45799          osd.54         up   1.00000  1.00000
>  55    hdd    5.45799          osd.55         up   1.00000  1.00000
>  -9         152.82401      host ctplosd8
>  56    hdd    5.45799          osd.56         up   1.00000  1.00000
>  85    hdd    5.45799          osd.85         up   1.00000  1.00000
>  86    hdd    5.45799          osd.86         up   1.00000  1.00000
>  87    hdd    5.45799          osd.87         up   1.00000  1.00000
>  88    hdd    5.45799          osd.88         up   1.00000  1.00000
>  89    hdd    5.45799          osd.89         up   1.00000  1.00000
>  90    hdd    5.45799          osd.90         up   1.00000  1.00000
>  91    hdd    5.45799          osd.91         up   1.00000  1.00000
>  92    hdd    5.45799          osd.92         up   1.00000  1.00000
>  93    hdd    5.45799          osd.93         up   1.00000  1.00000
>  94    hdd    5.45799          osd.94         up   1.00000  1.00000
>  95    hdd    5.45799          osd.95         up   1.00000  1.00000
>  96    hdd    5.45799          osd.96         up   1.00000  1.00000
>  97    hdd    5.45799          osd.97         up   1.00000  1.00000
>  98    hdd    5.45799          osd.98         up   1.00000  1.00000
>  99    hdd    5.45799          osd.99         up   1.00000  1.00000
> 100    hdd    5.45799          osd.100        up   1.00000  1.00000
> 101    hdd    5.45799          osd.101        up   1.00000  1.00000
> 102    hdd    5.45799          osd.102        up   1.00000  1.00000
> 103    hdd    5.45799          osd.103        up   1.00000  1.00000
> 104    hdd    5.45799          osd.104        up   1.00000  1.00000
> 105    hdd    5.45799          osd.105        up   1.00000  1.00000
> 106    hdd    5.45799          osd.106        up   1.00000  1.00000
> 107    hdd    5.45799          osd.107        up   1.00000  1.00000
> 108    hdd    5.45799          osd.108        up   1.00000  1.00000
> 109    hdd    5.45799          osd.109        up   1.00000  1.00000
> 110    hdd    5.45799          osd.110        up   1.00000  1.00000
> 111    hdd    5.45799          osd.111        up   1.00000  1.00000
> root@ctplmon1:~# ceph osd pool ls detail
> pool 9 'default.rgw.buckets.data' erasure profile ec-32-profile size 5
> min_size 4 crush_rule 1 object_hash rjenkins pg_num 32 pgp_num 32
> autoscale_mode on last_change 128267 lfor 0/127784/127779 flags
> hashpspool,ec_overwrites stripe_width 12288 application rgw
>
> ----
>
> The affected pool is pool number 9 and host is ctplosd1. This is the host
> I removed at the first place (to reinstall OS) and now I added this host
> back to the cluster, but osds on this host cannot be brought back to up
> state for some reason, even though osd processes are running on the host.
>
> Kind regards,
> Rok
>
>
>
>
>
> On Thu, May 27, 2021 at 3:32 PM Eugen Block <eblock@xxxxxx> wrote:
>
>> Hi,
>>
>> this sounds like your crush rule(s) for one or more pools can't place
>> the PGs because the host is missing. Please share
>>
>> ceph pg dump pgs_brief | grep undersized
>> ceph osd tree
>> ceph osd pool ls detail
>>
>> and the crush rule(s) for the affected pool(s).
>>
>>
>> Zitat von Rok Jaklič <rjaklic@xxxxxxxxx>:
>>
>> > Hi,
>> >
>> > I have removed one node, but now ceph seems to stuck in:
>> > Degraded data redundancy: 67/2393 objects degraded (2.800%), 12 pgs
>> > degraded, 12 pgs undersized
>> >
>> > How to "force" rebalancing? Or should I just wait a little bit more?
>> >
>> > Kind regards,
>> > rok
>> > _______________________________________________
>> > ceph-users mailing list -- ceph-users@xxxxxxx
>> > To unsubscribe send an email to ceph-users-leave@xxxxxxx
>>
>>
>> _______________________________________________
>> ceph-users mailing list -- ceph-users@xxxxxxx
>> To unsubscribe send an email to ceph-users-leave@xxxxxxx
>>
>
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux