Re: OSDs not balanced

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



The balancer will operate on all pools unless otherwise specified.

Josh

On Mon, Mar 4, 2024 at 1:12 PM Cedric <yipikai7@xxxxxxxxx> wrote:
>
> Did the balancer has enabled pools ? "ceph balancer pool ls"
>
> Actually I am wondering if the balancer do something when no pools are
> added.
>
>
>
> On Mon, Mar 4, 2024, 11:30 Ml Ml <mliebherr99@xxxxxxxxxxxxxx> wrote:
>
> > Hello,
> >
> > i wonder why my autobalancer is not working here:
> >
> > root@ceph01:~# ceph -s
> >   cluster:
> >     id:     5436dd5d-83d4-4dc8-a93b-60ab5db145df
> >     health: HEALTH_ERR
> >             1 backfillfull osd(s)
> >             1 full osd(s)
> >             1 nearfull osd(s)
> >             4 pool(s) full
> >
> > => osd.17 was too full (92% or something like that)
> >
> > root@ceph01:~# ceph osd df tree
> > ID   CLASS  WEIGHT     REWEIGHT  SIZE     ... %USE  ... PGS TYPE NAME
> > -25         209.50084         -  213 TiB  ... 69.56 ...   - datacenter
> > xxx-dc-root
> > -19          84.59369         -   86 TiB  ... 56.97 ...   -     rack
> > RZ1.Reihe4.R10
> >  -3          35.49313         -   37 TiB  ... 57.88 ...   -         host
> > ceph02
> >   2    hdd    1.70000   1.00000  1.7 TiB  ... 58.77 ...  44
> >  osd.2
> >   3    hdd    1.00000   1.00000  2.7 TiB  ... 22.14 ...  25
> >  osd.3
> >   7    hdd    2.50000   1.00000  2.7 TiB  ... 58.84 ...  70
> >  osd.7
> >   9    hdd    9.50000   1.00000  9.5 TiB  ... 63.07 ... 268
> >  osd.9
> >  13    hdd    2.67029   1.00000  2.7 TiB  ... 53.59 ...  65
> >  osd.13
> >  16    hdd    2.89999   1.00000  2.7 TiB  ... 59.35 ...  71
> >  osd.16
> >  19    hdd    1.70000   1.00000  1.7 TiB  ... 48.98 ...  37
> >  osd.19
> >  23    hdd    2.38419   1.00000  2.4 TiB  ... 59.33 ...  64
> >  osd.23
> >  24    hdd    1.39999   1.00000  1.7 TiB  ... 51.23 ...  39
> >  osd.24
> >  28    hdd    3.63869   1.00000  3.6 TiB  ... 64.17 ... 104
> >  osd.28
> >  31    hdd    2.70000   1.00000  2.7 TiB  ... 64.73 ...  76
> >  osd.31
> >  32    hdd    3.39999   1.00000  3.3 TiB  ... 67.28 ... 101
> >  osd.32
> >  -9          22.88817         -   23 TiB  ... 56.96 ...   -         host
> > ceph06
> >  35    hdd    7.15259   1.00000  7.2 TiB  ... 55.71 ... 182
> >  osd.35
> >  36    hdd    5.24519   1.00000  5.2 TiB  ... 53.75 ... 128
> >  osd.36
> >  45    hdd    5.24519   1.00000  5.2 TiB  ... 60.91 ... 144
> >  osd.45
> >  48    hdd    5.24519   1.00000  5.2 TiB  ... 57.94 ... 139
> >  osd.48
> > -17          26.21239         -   26 TiB  ... 55.67 ...   -         host
> > ceph08
> >  37    hdd    6.67569   1.00000  6.7 TiB  ... 58.17 ... 174
> >  osd.37
> >  40    hdd    9.53670   1.00000  9.5 TiB  ... 58.54 ... 250
> >  osd.40
> >  46    hdd    5.00000   1.00000  5.0 TiB  ... 52.39 ... 116
> >  osd.46
> >  47    hdd    5.00000   1.00000  5.0 TiB  ... 50.05 ... 112
> >  osd.47
> > -20          59.11053         -   60 TiB  ... 82.47 ...   -     rack
> > RZ1.Reihe4.R9
> >  -4          23.09996         -   24 TiB  ... 79.92 ...   -         host
> > ceph03
> >   5    hdd    1.70000   0.75006  1.7 TiB  ... 87.24 ...  66
> >  osd.5
> >   6    hdd    1.70000   0.44998  1.7 TiB  ... 47.30 ...  36
> >  osd.6
> >  10    hdd    2.70000   0.85004  2.7 TiB  ... 83.23 ... 100
> >  osd.10
> >  15    hdd    2.70000   0.75006  2.7 TiB  ... 74.26 ...  88
> >  osd.15
> >  17    hdd    0.50000   0.85004  1.6 TiB  ... 91.44 ...  67
> >  osd.17
> >  20    hdd    2.00000   0.85004  1.7 TiB  ... 88.41 ...  68
> >  osd.20
> >  21    hdd    2.79999   0.75006  2.7 TiB  ... 77.25 ...  91
> >  osd.21
> >  25    hdd    1.70000   0.90002  1.7 TiB  ... 78.31 ...  60
> >  osd.25
> >  26    hdd    2.70000   1.00000  2.7 TiB  ... 82.75 ...  99
> >  osd.26
> >  27    hdd    2.70000   0.90002  2.7 TiB  ... 84.26 ... 101
> >  osd.27
> >  63    hdd    1.89999   0.90002  1.7 TiB  ... 84.15 ...  65
> >  osd.63
> > -13          36.01057         -   36 TiB  ... 84.12 ...   -         host
> > ceph05
> >  11    hdd    7.15259   0.90002  7.2 TiB  ... 85.45 ... 273
> >  osd.11
> >  39    hdd    7.20000   0.85004  7.2 TiB  ... 80.90 ... 257
> >  osd.39
> >  41    hdd    7.20000   0.75006  7.2 TiB  ... 74.95 ... 239
> >  osd.41
> >  42    hdd    9.00000   1.00000  9.5 TiB  ... 92.00 ... 392
> >  osd.42
> >  43    hdd    5.45799   1.00000  5.5 TiB  ... 84.84 ... 207
> >  osd.43
> > -21          65.79662         -   66 TiB  ... 74.29 ...   -     rack
> > RZ3.Reihe3.R10
> >  -2          28.49664         -   29 TiB  ... 74.79 ...   -         host
> > ceph01
> >   0    hdd    2.70000   1.00000  2.7 TiB  ... 73.82 ...  88
> >  osd.0
> >   1    hdd    3.63869   1.00000  3.6 TiB  ... 73.47 ... 121
> >  osd.1
> >   4    hdd    2.70000   1.00000  2.7 TiB  ... 74.63 ...  89
> >  osd.4
> >   8    hdd    2.70000   1.00000  2.7 TiB  ... 77.10 ...  92
> >  osd.8
> >  12    hdd    2.70000   1.00000  2.7 TiB  ... 78.76 ...  94
> >  osd.12
> >  14    hdd    5.45799   1.00000  5.5 TiB  ... 78.86 ... 193
> >  osd.14
> >  18    hdd    1.89999   1.00000  2.7 TiB  ... 63.79 ...  76
> >  osd.18
> >  22    hdd    1.70000   1.00000  1.7 TiB  ... 74.85 ...  57
> >  osd.22
> >  30    hdd    1.70000   1.00000  1.7 TiB  ... 76.34 ...  59
> >  osd.30
> >  64    hdd    3.29999   1.00000  3.3 TiB  ... 73.48 ... 110
> >  osd.64
> > -11          12.39999         -   12 TiB  ... 73.40 ...   -         host
> > ceph04
> >  34    hdd    5.20000   1.00000  5.2 TiB  ... 72.81 ... 171
> >  osd.34
> >  44    hdd    7.20000   1.00000  7.2 TiB  ... 73.83 ... 236
> >  osd.44
> > -15          24.89998         -   25 TiB  ... 74.15 ...   -         host
> > ceph07
> >  66    hdd    7.20000   1.00000  7.2 TiB  ... 74.07 ... 236
> >  osd.66
> >  67    hdd    7.20000   1.00000  7.2 TiB  ... 73.74 ... 236
> >  osd.67
> >  68    hdd    3.29999   1.00000  3.3 TiB  ... 72.99 ... 110
> >  osd.68
> >  69    hdd    7.20000   1.00000  7.2 TiB  ... 75.18 ... 241
> >  osd.69
> >  -1                 0         -      0 B  ...     0 ...   - root default
> >                           TOTAL  213 TiB  ... 69.56
> >
> > root@ceph01:~# ceph balancer status
> > {
> >     "active": true,
> >     "last_optimize_duration": "0:00:00.256761",
> >     "last_optimize_started": "Mon Mar  4 10:25:10 2024",
> >     "mode": "upmap",
> >     "no_optimization_needed": true,
> >     "optimize_result": "Unable to find further optimization, or
> > pool(s) pg_num is decreasing, or distribution is already perfect",
> >     "plans": []
> > }
> >
> >
> > Where as reweight-by-utilization would change quite a bit!:
> >
> > root@ceph01:~# ceph osd test-reweight-by-utilization 110 .5 10
> > moved 260 / 6627 (3.92334%)
> > avg 127.442
> > stddev 81.345 -> 78.169 (expected baseline 11.18)
> > min osd.17 with 24 -> 16 pgs (0.188321 -> 0.125547 * mean)
> > max osd.42 with 401 -> 320 pgs (3.14652 -> 2.51094 * mean)
> >
> > oload 110
> > max_change 0.5
> > max_change_osds 10
> > average_utilization 0.6956
> > overload_utilization 0.7652
> > osd.42 weight 1.0000 -> 0.7561
> > osd.6 weight 0.4500 -> 0.6616
> > osd.17 weight 0.8500 -> 0.6466
> > osd.20 weight 0.8500 -> 0.6688
> > osd.5 weight 0.7501 -> 0.6004
> > osd.11 weight 0.9000 -> 0.7326
> > osd.43 weight 1.0000 -> 0.8199
> > osd.27 weight 0.9000 -> 0.7430
> > osd.63 weight 0.9000 -> 0.7440
> > osd.10 weight 0.8500 -> 0.7104
> > no change
> >
> >
> >
> > root@ceph01:~# ceph versions
> > {
> >     "mon": {
> >         "ceph version 17.2.7
> > (b12291d110049b2f35e32e0de30d70e9a4c060d2) quincy (stable)": 3
> >     },
> >     "mgr": {
> >         "ceph version 17.2.7
> > (b12291d110049b2f35e32e0de30d70e9a4c060d2) quincy (stable)": 2
> >     },
> >     "osd": {
> >         "ceph version 17.2.7
> > (b12291d110049b2f35e32e0de30d70e9a4c060d2) quincy (stable)": 52
> >     },
> >     "mds": {
> >         "ceph version 17.2.7
> > (b12291d110049b2f35e32e0de30d70e9a4c060d2) quincy (stable)": 4
> >     },
> >     "overall": {
> >         "ceph version 17.2.7
> > (b12291d110049b2f35e32e0de30d70e9a4c060d2) quincy (stable)": 61
> >     }
> > }
> >
> >
> > Cheers,
> > Michael
> > _______________________________________________
> > ceph-users mailing list -- ceph-users@xxxxxxx
> > To unsubscribe send an email to ceph-users-leave@xxxxxxx
> >
> _______________________________________________
> ceph-users mailing list -- ceph-users@xxxxxxx
> To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux