Did the balancer has enabled pools ? "ceph balancer pool ls" Actually I am wondering if the balancer do something when no pools are added. On Mon, Mar 4, 2024, 11:30 Ml Ml <mliebherr99@xxxxxxxxxxxxxx> wrote: > Hello, > > i wonder why my autobalancer is not working here: > > root@ceph01:~# ceph -s > cluster: > id: 5436dd5d-83d4-4dc8-a93b-60ab5db145df > health: HEALTH_ERR > 1 backfillfull osd(s) > 1 full osd(s) > 1 nearfull osd(s) > 4 pool(s) full > > => osd.17 was too full (92% or something like that) > > root@ceph01:~# ceph osd df tree > ID CLASS WEIGHT REWEIGHT SIZE ... %USE ... PGS TYPE NAME > -25 209.50084 - 213 TiB ... 69.56 ... - datacenter > xxx-dc-root > -19 84.59369 - 86 TiB ... 56.97 ... - rack > RZ1.Reihe4.R10 > -3 35.49313 - 37 TiB ... 57.88 ... - host > ceph02 > 2 hdd 1.70000 1.00000 1.7 TiB ... 58.77 ... 44 > osd.2 > 3 hdd 1.00000 1.00000 2.7 TiB ... 22.14 ... 25 > osd.3 > 7 hdd 2.50000 1.00000 2.7 TiB ... 58.84 ... 70 > osd.7 > 9 hdd 9.50000 1.00000 9.5 TiB ... 63.07 ... 268 > osd.9 > 13 hdd 2.67029 1.00000 2.7 TiB ... 53.59 ... 65 > osd.13 > 16 hdd 2.89999 1.00000 2.7 TiB ... 59.35 ... 71 > osd.16 > 19 hdd 1.70000 1.00000 1.7 TiB ... 48.98 ... 37 > osd.19 > 23 hdd 2.38419 1.00000 2.4 TiB ... 59.33 ... 64 > osd.23 > 24 hdd 1.39999 1.00000 1.7 TiB ... 51.23 ... 39 > osd.24 > 28 hdd 3.63869 1.00000 3.6 TiB ... 64.17 ... 104 > osd.28 > 31 hdd 2.70000 1.00000 2.7 TiB ... 64.73 ... 76 > osd.31 > 32 hdd 3.39999 1.00000 3.3 TiB ... 67.28 ... 101 > osd.32 > -9 22.88817 - 23 TiB ... 56.96 ... - host > ceph06 > 35 hdd 7.15259 1.00000 7.2 TiB ... 55.71 ... 182 > osd.35 > 36 hdd 5.24519 1.00000 5.2 TiB ... 53.75 ... 128 > osd.36 > 45 hdd 5.24519 1.00000 5.2 TiB ... 60.91 ... 144 > osd.45 > 48 hdd 5.24519 1.00000 5.2 TiB ... 57.94 ... 139 > osd.48 > -17 26.21239 - 26 TiB ... 55.67 ... - host > ceph08 > 37 hdd 6.67569 1.00000 6.7 TiB ... 58.17 ... 174 > osd.37 > 40 hdd 9.53670 1.00000 9.5 TiB ... 58.54 ... 250 > osd.40 > 46 hdd 5.00000 1.00000 5.0 TiB ... 52.39 ... 116 > osd.46 > 47 hdd 5.00000 1.00000 5.0 TiB ... 50.05 ... 112 > osd.47 > -20 59.11053 - 60 TiB ... 82.47 ... - rack > RZ1.Reihe4.R9 > -4 23.09996 - 24 TiB ... 79.92 ... - host > ceph03 > 5 hdd 1.70000 0.75006 1.7 TiB ... 87.24 ... 66 > osd.5 > 6 hdd 1.70000 0.44998 1.7 TiB ... 47.30 ... 36 > osd.6 > 10 hdd 2.70000 0.85004 2.7 TiB ... 83.23 ... 100 > osd.10 > 15 hdd 2.70000 0.75006 2.7 TiB ... 74.26 ... 88 > osd.15 > 17 hdd 0.50000 0.85004 1.6 TiB ... 91.44 ... 67 > osd.17 > 20 hdd 2.00000 0.85004 1.7 TiB ... 88.41 ... 68 > osd.20 > 21 hdd 2.79999 0.75006 2.7 TiB ... 77.25 ... 91 > osd.21 > 25 hdd 1.70000 0.90002 1.7 TiB ... 78.31 ... 60 > osd.25 > 26 hdd 2.70000 1.00000 2.7 TiB ... 82.75 ... 99 > osd.26 > 27 hdd 2.70000 0.90002 2.7 TiB ... 84.26 ... 101 > osd.27 > 63 hdd 1.89999 0.90002 1.7 TiB ... 84.15 ... 65 > osd.63 > -13 36.01057 - 36 TiB ... 84.12 ... - host > ceph05 > 11 hdd 7.15259 0.90002 7.2 TiB ... 85.45 ... 273 > osd.11 > 39 hdd 7.20000 0.85004 7.2 TiB ... 80.90 ... 257 > osd.39 > 41 hdd 7.20000 0.75006 7.2 TiB ... 74.95 ... 239 > osd.41 > 42 hdd 9.00000 1.00000 9.5 TiB ... 92.00 ... 392 > osd.42 > 43 hdd 5.45799 1.00000 5.5 TiB ... 84.84 ... 207 > osd.43 > -21 65.79662 - 66 TiB ... 74.29 ... - rack > RZ3.Reihe3.R10 > -2 28.49664 - 29 TiB ... 74.79 ... - host > ceph01 > 0 hdd 2.70000 1.00000 2.7 TiB ... 73.82 ... 88 > osd.0 > 1 hdd 3.63869 1.00000 3.6 TiB ... 73.47 ... 121 > osd.1 > 4 hdd 2.70000 1.00000 2.7 TiB ... 74.63 ... 89 > osd.4 > 8 hdd 2.70000 1.00000 2.7 TiB ... 77.10 ... 92 > osd.8 > 12 hdd 2.70000 1.00000 2.7 TiB ... 78.76 ... 94 > osd.12 > 14 hdd 5.45799 1.00000 5.5 TiB ... 78.86 ... 193 > osd.14 > 18 hdd 1.89999 1.00000 2.7 TiB ... 63.79 ... 76 > osd.18 > 22 hdd 1.70000 1.00000 1.7 TiB ... 74.85 ... 57 > osd.22 > 30 hdd 1.70000 1.00000 1.7 TiB ... 76.34 ... 59 > osd.30 > 64 hdd 3.29999 1.00000 3.3 TiB ... 73.48 ... 110 > osd.64 > -11 12.39999 - 12 TiB ... 73.40 ... - host > ceph04 > 34 hdd 5.20000 1.00000 5.2 TiB ... 72.81 ... 171 > osd.34 > 44 hdd 7.20000 1.00000 7.2 TiB ... 73.83 ... 236 > osd.44 > -15 24.89998 - 25 TiB ... 74.15 ... - host > ceph07 > 66 hdd 7.20000 1.00000 7.2 TiB ... 74.07 ... 236 > osd.66 > 67 hdd 7.20000 1.00000 7.2 TiB ... 73.74 ... 236 > osd.67 > 68 hdd 3.29999 1.00000 3.3 TiB ... 72.99 ... 110 > osd.68 > 69 hdd 7.20000 1.00000 7.2 TiB ... 75.18 ... 241 > osd.69 > -1 0 - 0 B ... 0 ... - root default > TOTAL 213 TiB ... 69.56 > > root@ceph01:~# ceph balancer status > { > "active": true, > "last_optimize_duration": "0:00:00.256761", > "last_optimize_started": "Mon Mar 4 10:25:10 2024", > "mode": "upmap", > "no_optimization_needed": true, > "optimize_result": "Unable to find further optimization, or > pool(s) pg_num is decreasing, or distribution is already perfect", > "plans": [] > } > > > Where as reweight-by-utilization would change quite a bit!: > > root@ceph01:~# ceph osd test-reweight-by-utilization 110 .5 10 > moved 260 / 6627 (3.92334%) > avg 127.442 > stddev 81.345 -> 78.169 (expected baseline 11.18) > min osd.17 with 24 -> 16 pgs (0.188321 -> 0.125547 * mean) > max osd.42 with 401 -> 320 pgs (3.14652 -> 2.51094 * mean) > > oload 110 > max_change 0.5 > max_change_osds 10 > average_utilization 0.6956 > overload_utilization 0.7652 > osd.42 weight 1.0000 -> 0.7561 > osd.6 weight 0.4500 -> 0.6616 > osd.17 weight 0.8500 -> 0.6466 > osd.20 weight 0.8500 -> 0.6688 > osd.5 weight 0.7501 -> 0.6004 > osd.11 weight 0.9000 -> 0.7326 > osd.43 weight 1.0000 -> 0.8199 > osd.27 weight 0.9000 -> 0.7430 > osd.63 weight 0.9000 -> 0.7440 > osd.10 weight 0.8500 -> 0.7104 > no change > > > > root@ceph01:~# ceph versions > { > "mon": { > "ceph version 17.2.7 > (b12291d110049b2f35e32e0de30d70e9a4c060d2) quincy (stable)": 3 > }, > "mgr": { > "ceph version 17.2.7 > (b12291d110049b2f35e32e0de30d70e9a4c060d2) quincy (stable)": 2 > }, > "osd": { > "ceph version 17.2.7 > (b12291d110049b2f35e32e0de30d70e9a4c060d2) quincy (stable)": 52 > }, > "mds": { > "ceph version 17.2.7 > (b12291d110049b2f35e32e0de30d70e9a4c060d2) quincy (stable)": 4 > }, > "overall": { > "ceph version 17.2.7 > (b12291d110049b2f35e32e0de30d70e9a4c060d2) quincy (stable)": 61 > } > } > > > Cheers, > Michael > _______________________________________________ > ceph-users mailing list -- ceph-users@xxxxxxx > To unsubscribe send an email to ceph-users-leave@xxxxxxx > _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx