Hi Dan, I've set the autoscaler to warn, but it actually does not warn for now. So not touching it for now. this is what the log says in minute intervals: 2021-03-15 13:51:00.970 7f307d5fd700 4 mgr get_config get_config key: mgr/balancer/active 2021-03-15 13:51:00.970 7f307d5fd700 4 mgr get_config get_config key: mgr/balancer/sleep_interval 2021-03-15 13:51:00.970 7f307d5fd700 4 mgr get_config get_config key: mgr/balancer/begin_time 2021-03-15 13:51:00.970 7f307d5fd700 4 mgr get_config get_config key: mgr/balancer/end_time 2021-03-15 13:51:00.970 7f307d5fd700 4 mgr get_config get_config key: mgr/balancer/begin_weekday 2021-03-15 13:51:00.970 7f307d5fd700 4 mgr get_config get_config key: mgr/balancer/end_weekday 2021-03-15 13:51:00.971 7f307d5fd700 4 mgr get_config get_config key: mgr/balancer/pool_ids 2021-03-15 13:51:01.203 7f307d5fd700 4 mgr[balancer] Optimize plan auto_2021-03-15_13:51:00 2021-03-15 13:51:01.203 7f307d5fd700 4 mgr get_config get_config key: mgr/balancer/mode 2021-03-15 13:51:01.203 7f307d5fd700 4 mgr[balancer] Mode upmap, max misplaced 0.050000 2021-03-15 13:51:01.203 7f307d5fd700 4 mgr[balancer] do_upmap 2021-03-15 13:51:01.203 7f307d5fd700 4 mgr get_config get_config key: mgr/balancer/upmap_max_iterations 2021-03-15 13:51:01.203 7f307d5fd700 4 mgr get_config get_config key: mgr/balancer/upmap_max_deviation 2021-03-15 13:51:01.203 7f307d5fd700 4 mgr[balancer] pools ['eu-msg-1.rgw.data.root', 'eu-msg-1.rgw.buckets.non-ec', 'eu-central-1.rgw.users.keys', 'eu-central-1.rgw.gc', 'eu-central-1.rgw.buckets.data', 'eu-central-1.rgw.users.email', 'eu-msg-1.rgw.gc', 'eu-central-1.rgw.usage', 'eu-msg-1.rgw.users.keys', 'eu-central-1.rgw.buckets.index', 'rbd', 'eu-msg-1.rgw.log', 'whitespace-again-2021-03-10_2', 'eu-msg-1.rgw.buckets.index', 'eu-msg-1.rgw.meta', 'eu-central-1.rgw.log', 'default.rgw.gc', 'eu-central-1.rgw.buckets.non-ec', 'eu-msg-1.rgw.usage', 'whitespace-again-2021-03-10', 'fra-1.rgw.meta', 'eu-central-1.rgw.users.uid', 'eu-msg-1.rgw.users.email', 'fra-1.rgw.control', 'eu-msg-1.rgw.users.uid', 'eu-msg-1.rgw.control', '.rgw.root', 'eu-msg-1.rgw.buckets.data', 'default.rgw.control', 'fra-1.rgw.log', 'default.rgw.data.root', 'whitespace-again-2021-03-10_3', 'default.rgw.log', 'eu-central-1.rgw.meta', 'eu-central-1.rgw.data.root', 'default.rgw.users.uid', 'eu-central-1.rgw.control'] 2021-03-15 13:51:01.224 7f307d5fd700 4 mgr[balancer] prepared 0/10 changes Am Mo., 15. März 2021 um 14:15 Uhr schrieb Dan van der Ster < dan@xxxxxxxxxxxxxx>: > I suggest to just disable the autoscaler until your balancing is > understood. > > What does your active mgr log say (with debug_mgr 4/5), grep balancer > /var/log/ceph/ceph-mgr.*.log > > -- Dan > > On Mon, Mar 15, 2021 at 1:47 PM Boris Behrens <bb@xxxxxxxxx> wrote: > > > > Hi, > > this unfortunally did not solve my problem. I still have some OSDs that > fill up to 85% > > > > According to the logging, the autoscaler might want to add more PGs to > one Bucken and reduce almost all other buckets to 32. > > 2021-03-15 12:19:58.825 7f307f601700 4 mgr[pg_autoscaler] Pool > 'eu-central-1.rgw.buckets.data' root_id -1 using 0.705080476146 of space, > bias 1.0, pg target 1974.22533321 quantized to 2048 (current 1024) > > > > Why the balancing does not happen is still nebulous to me. > > > > > > > > Am Sa., 13. März 2021 um 16:37 Uhr schrieb Dan van der Ster < > dan@xxxxxxxxxxxxxx>: > >> > >> OK > >> Btw, you might need to fail to a new mgr... I'm not sure if the current > active will read that new config. > >> > >> .. dan > >> > >> > >> On Sat, Mar 13, 2021, 4:36 PM Boris Behrens <bb@xxxxxxxxx> wrote: > >>> > >>> Hi, > >>> > >>> ok thanks. I just changed the value and rewighted everything back to > 1. Now I let it sync the weekend and check how it will be on monday. > >>> We tried to have the systems total storage balanced as possible. New > systems will be with 8TB disks but for the exiting ones we added 16TB to > offset the 4TB disks and we needed a lot of storage fast, because of a DC > move. If you have any recommendations I would be happy to hear them. > >>> > >>> Cheers > >>> Boris > >>> > >>> Am Sa., 13. März 2021 um 16:20 Uhr schrieb Dan van der Ster < > dan@xxxxxxxxxxxxxx>: > >>>> > >>>> Thanks. > >>>> > >>>> Decreasing the max deviation to 2 or 1 should help in your case. This > option controls when the balancer stops trying to move PGs around -- by > default it stops when the deviation from the mean is 5. Yes this is too > large IMO -- all of our clusters have this set to 1. > >>>> > >>>> And given that you have some OSDs with more than 200 PGs, you > definitely shouldn't increase the num PGs. > >>>> > >>>> But anyway with your mixed device sizes it might be challenging to > make a perfectly uniform distribution. Give it a try with 1 though, and let > us know how it goes. > >>>> > >>>> .. Dan > >>>> > >>>> > >>>> > >>>> > >>>> > >>>> On Sat, Mar 13, 2021, 4:11 PM Boris Behrens <bb@xxxxxxxxx> wrote: > >>>>> > >>>>> Hi Dan, > >>>>> > >>>>> upmap_max_deviation is default (5) in our cluster. Is 1 the > recommended deviation? > >>>>> > >>>>> I added the whole ceph osd df tree, (I need to remove some OSDs and > readd them as bluestore with SSD, so 69, 73 and 82 are a bit off now. I > also reweighted to try to get the %USE mitigated). > >>>>> > >>>>> I will increase the mgr debugging to see what is the problem. > >>>>> > >>>>> [root@s3db1 ~]# ceph osd df tree > >>>>> ID CLASS WEIGHT REWEIGHT SIZE RAW USE DATA OMAP META > AVAIL %USE VAR PGS STATUS TYPE NAME > >>>>> -1 673.54224 - 659 TiB 491 TiB 464 TiB 96 GiB 1.2 TiB > 168 TiB 74.57 1.00 - root default > >>>>> -2 58.30331 - 44 TiB 22 TiB 17 TiB 5.7 GiB 38 > GiB 22 TiB 49.82 0.67 - host s3db1 > >>>>> 23 hdd 14.65039 1.00000 15 TiB 1.8 TiB 1.7 TiB 156 MiB 4.4 > GiB 13 TiB 12.50 0.17 101 up osd.23 > >>>>> 69 hdd 14.55269 0 0 B 0 B 0 B 0 B 0 > B 0 B 0 0 11 up osd.69 > >>>>> 73 hdd 14.55269 1.00000 15 TiB 10 TiB 10 TiB 6.1 MiB 33 GiB > 4.2 TiB 71.15 0.95 107 up osd.73 > >>>>> 79 hdd 3.63689 1.00000 3.6 TiB 2.9 TiB 747 GiB 2.0 GiB 0 B > 747 GiB 79.94 1.07 52 up osd.79 > >>>>> 80 hdd 3.63689 1.00000 3.6 TiB 2.6 TiB 1.0 TiB 1.9 GiB 0 B > 1.0 TiB 71.61 0.96 58 up osd.80 > >>>>> 81 hdd 3.63689 1.00000 3.6 TiB 2.2 TiB 1.5 TiB 1.1 GiB 0 B > 1.5 TiB 60.07 0.81 55 up osd.81 > >>>>> 82 hdd 3.63689 1.00000 3.6 TiB 1.9 TiB 1.7 TiB 536 MiB 0 B > 1.7 TiB 52.68 0.71 30 up osd.82 > >>>>> -11 50.94173 - 51 TiB 38 TiB 38 TiB 3.7 GiB 100 > GiB 13 TiB 74.69 1.00 - host s3db10 > >>>>> 63 hdd 7.27739 1.00000 7.3 TiB 5.5 TiB 5.5 TiB 616 MiB 14 GiB > 1.7 TiB 76.04 1.02 92 up osd.63 > >>>>> 64 hdd 7.27739 1.00000 7.3 TiB 5.5 TiB 5.5 TiB 820 MiB 15 GiB > 1.8 TiB 75.54 1.01 101 up osd.64 > >>>>> 65 hdd 7.27739 1.00000 7.3 TiB 5.3 TiB 5.3 TiB 109 MiB 14 GiB > 2.0 TiB 73.17 0.98 105 up osd.65 > >>>>> 66 hdd 7.27739 1.00000 7.3 TiB 5.8 TiB 5.8 TiB 423 MiB 15 GiB > 1.4 TiB 80.38 1.08 98 up osd.66 > >>>>> 67 hdd 7.27739 1.00000 7.3 TiB 5.1 TiB 5.1 TiB 572 MiB 14 GiB > 2.2 TiB 70.10 0.94 100 up osd.67 > >>>>> 68 hdd 7.27739 1.00000 7.3 TiB 5.3 TiB 5.3 TiB 630 MiB 13 GiB > 2.0 TiB 72.88 0.98 107 up osd.68 > >>>>> 70 hdd 7.27739 1.00000 7.3 TiB 5.4 TiB 5.4 TiB 648 MiB 14 GiB > 1.8 TiB 74.73 1.00 102 up osd.70 > >>>>> -12 50.99052 - 51 TiB 39 TiB 39 TiB 2.9 GiB 99 > GiB 12 TiB 77.24 1.04 - host s3db11 > >>>>> 46 hdd 7.27739 1.00000 7.3 TiB 5.7 TiB 5.7 TiB 102 MiB 15 GiB > 1.5 TiB 78.91 1.06 97 up osd.46 > >>>>> 47 hdd 7.27739 1.00000 7.3 TiB 5.2 TiB 5.2 TiB 61 MiB 13 GiB > 2.1 TiB 71.47 0.96 96 up osd.47 > >>>>> 48 hdd 7.27739 1.00000 7.3 TiB 6.1 TiB 6.1 TiB 853 MiB 15 GiB > 1.2 TiB 83.46 1.12 109 up osd.48 > >>>>> 49 hdd 7.27739 1.00000 7.3 TiB 5.7 TiB 5.7 TiB 708 MiB 15 GiB > 1.5 TiB 78.96 1.06 98 up osd.49 > >>>>> 50 hdd 7.27739 1.00000 7.3 TiB 5.9 TiB 5.8 TiB 472 MiB 15 GiB > 1.4 TiB 80.40 1.08 102 up osd.50 > >>>>> 51 hdd 7.27739 1.00000 7.3 TiB 5.9 TiB 5.9 TiB 729 MiB 15 GiB > 1.3 TiB 81.70 1.10 110 up osd.51 > >>>>> 72 hdd 7.32619 1.00000 7.3 TiB 4.8 TiB 4.8 TiB 91 MiB 12 GiB > 2.5 TiB 65.82 0.88 89 up osd.72 > >>>>> -37 58.55478 - 59 TiB 46 TiB 46 TiB 5.0 GiB 124 > GiB 12 TiB 79.04 1.06 - host s3db12 > >>>>> 19 hdd 3.68750 1.00000 3.7 TiB 3.1 TiB 3.1 TiB 462 MiB 8.2 GiB > 559 GiB 85.18 1.14 55 up osd.19 > >>>>> 71 hdd 3.68750 1.00000 3.7 TiB 2.9 TiB 2.8 TiB 3.9 MiB 7.8 GiB > 825 GiB 78.14 1.05 50 up osd.71 > >>>>> 75 hdd 3.68750 1.00000 3.7 TiB 3.1 TiB 3.1 TiB 576 MiB 8.3 GiB > 555 GiB 85.29 1.14 57 up osd.75 > >>>>> 76 hdd 3.68750 1.00000 3.7 TiB 3.2 TiB 3.1 TiB 239 MiB 9.3 GiB > 501 GiB 86.73 1.16 50 up osd.76 > >>>>> 77 hdd 14.60159 1.00000 15 TiB 11 TiB 11 TiB 880 MiB 30 GiB > 3.6 TiB 75.57 1.01 202 up osd.77 > >>>>> 78 hdd 14.60159 1.00000 15 TiB 11 TiB 11 TiB 1.0 GiB 30 GiB > 3.4 TiB 76.65 1.03 196 up osd.78 > >>>>> 83 hdd 14.60159 1.00000 15 TiB 12 TiB 12 TiB 1.8 GiB 31 GiB > 2.9 TiB 80.04 1.07 223 up osd.83 > >>>>> -3 58.49872 - 58 TiB 43 TiB 38 TiB 8.1 GiB 91 > GiB 16 TiB 73.15 0.98 - host s3db2 > >>>>> 1 hdd 14.65039 1.00000 15 TiB 11 TiB 11 TiB 3.1 GiB 38 GiB > 3.6 TiB 75.52 1.01 194 up osd.1 > >>>>> 3 hdd 3.63689 1.00000 3.6 TiB 2.2 TiB 1.4 TiB 418 MiB 0 B > 1.4 TiB 60.94 0.82 52 up osd.3 > >>>>> 4 hdd 3.63689 0.89999 3.6 TiB 3.2 TiB 401 GiB 845 MiB 0 B > 401 GiB 89.23 1.20 53 up osd.4 > >>>>> 5 hdd 3.63689 1.00000 3.6 TiB 2.3 TiB 1.3 TiB 437 MiB 0 B > 1.3 TiB 62.88 0.84 51 up osd.5 > >>>>> 6 hdd 3.63689 1.00000 3.6 TiB 2.0 TiB 1.7 TiB 1.8 GiB 0 B > 1.7 TiB 54.51 0.73 47 up osd.6 > >>>>> 7 hdd 14.65039 1.00000 15 TiB 11 TiB 11 TiB 493 MiB 26 GiB > 3.8 TiB 73.90 0.99 185 up osd.7 > >>>>> 74 hdd 14.65039 1.00000 15 TiB 11 TiB 11 TiB 1.1 GiB 27 GiB > 3.5 TiB 76.27 1.02 208 up osd.74 > >>>>> -4 58.49872 - 58 TiB 43 TiB 37 TiB 33 GiB 86 > GiB 15 TiB 74.05 0.99 - host s3db3 > >>>>> 2 hdd 14.65039 1.00000 15 TiB 11 TiB 11 TiB 850 MiB 26 GiB > 4.0 TiB 72.78 0.98 203 up osd.2 > >>>>> 9 hdd 14.65039 1.00000 15 TiB 11 TiB 11 TiB 8.3 GiB 33 GiB > 3.6 TiB 75.62 1.01 189 up osd.9 > >>>>> 10 hdd 14.65039 1.00000 15 TiB 11 TiB 11 TiB 663 MiB 28 GiB > 3.5 TiB 76.34 1.02 211 up osd.10 > >>>>> 12 hdd 3.63689 1.00000 3.6 TiB 2.4 TiB 1.2 TiB 633 MiB 0 B > 1.2 TiB 66.22 0.89 44 up osd.12 > >>>>> 13 hdd 3.63689 1.00000 3.6 TiB 2.9 TiB 720 GiB 2.3 GiB 0 B > 720 GiB 80.66 1.08 66 up osd.13 > >>>>> 14 hdd 3.63689 1.00000 3.6 TiB 3.1 TiB 552 GiB 18 GiB 0 B > 552 GiB 85.18 1.14 60 up osd.14 > >>>>> 15 hdd 3.63689 1.00000 3.6 TiB 2.0 TiB 1.7 TiB 2.1 GiB 0 B > 1.7 TiB 53.72 0.72 44 up osd.15 > >>>>> -5 58.49872 - 58 TiB 45 TiB 37 TiB 7.2 GiB 99 > GiB 14 TiB 76.37 1.02 - host s3db4 > >>>>> 11 hdd 14.65039 1.00000 15 TiB 12 TiB 12 TiB 897 MiB 28 GiB > 2.8 TiB 81.15 1.09 205 up osd.11 > >>>>> 17 hdd 14.65039 1.00000 15 TiB 11 TiB 11 TiB 1.2 GiB 27 GiB > 3.6 TiB 75.38 1.01 211 up osd.17 > >>>>> 18 hdd 14.65039 1.00000 15 TiB 11 TiB 11 TiB 965 MiB 44 GiB > 4.0 TiB 72.86 0.98 188 up osd.18 > >>>>> 20 hdd 3.63689 1.00000 3.6 TiB 2.9 TiB 796 GiB 529 MiB 0 B > 796 GiB 78.63 1.05 66 up osd.20 > >>>>> 21 hdd 3.63689 1.00000 3.6 TiB 2.6 TiB 1.1 TiB 2.1 GiB 0 B > 1.1 TiB 70.32 0.94 47 up osd.21 > >>>>> 22 hdd 3.63689 1.00000 3.6 TiB 2.9 TiB 802 GiB 882 MiB 0 B > 802 GiB 78.47 1.05 58 up osd.22 > >>>>> 24 hdd 3.63689 1.00000 3.6 TiB 2.8 TiB 856 GiB 645 MiB 0 B > 856 GiB 77.01 1.03 47 up osd.24 > >>>>> -6 58.89636 - 59 TiB 44 TiB 44 TiB 2.4 GiB 111 > GiB 15 TiB 75.22 1.01 - host s3db5 > >>>>> 0 hdd 3.73630 1.00000 3.7 TiB 2.4 TiB 2.3 TiB 70 MiB 6.6 GiB > 1.3 TiB 65.00 0.87 48 up osd.0 > >>>>> 25 hdd 3.73630 1.00000 3.7 TiB 2.4 TiB 2.3 TiB 5.3 MiB 6.6 GiB > 1.4 TiB 63.86 0.86 41 up osd.25 > >>>>> 26 hdd 3.73630 1.00000 3.7 TiB 2.9 TiB 2.8 TiB 181 MiB 7.6 GiB > 862 GiB 77.47 1.04 48 up osd.26 > >>>>> 27 hdd 3.73630 1.00000 3.7 TiB 2.3 TiB 2.2 TiB 7.0 MiB 6.1 GiB > 1.5 TiB 61.00 0.82 48 up osd.27 > >>>>> 28 hdd 14.65039 1.00000 15 TiB 12 TiB 12 TiB 937 MiB 30 GiB > 2.8 TiB 81.19 1.09 203 up osd.28 > >>>>> 29 hdd 14.65039 1.00000 15 TiB 11 TiB 11 TiB 536 MiB 26 GiB > 3.8 TiB 73.95 0.99 200 up osd.29 > >>>>> 30 hdd 14.65039 1.00000 15 TiB 12 TiB 11 TiB 744 MiB 28 GiB > 3.1 TiB 79.07 1.06 207 up osd.30 > >>>>> -7 58.89636 - 59 TiB 44 TiB 44 TiB 14 GiB 122 > GiB 14 TiB 75.41 1.01 - host s3db6 > >>>>> 32 hdd 3.73630 1.00000 3.7 TiB 3.1 TiB 3.0 TiB 16 MiB 8.2 GiB > 622 GiB 83.74 1.12 65 up osd.32 > >>>>> 33 hdd 3.73630 0.79999 3.7 TiB 3.0 TiB 2.9 TiB 14 MiB 8.1 GiB > 740 GiB 80.67 1.08 52 up osd.33 > >>>>> 34 hdd 3.73630 0.79999 3.7 TiB 2.9 TiB 2.8 TiB 449 MiB 7.7 GiB > 877 GiB 77.08 1.03 52 up osd.34 > >>>>> 35 hdd 3.73630 0.79999 3.7 TiB 2.3 TiB 2.2 TiB 133 MiB 7.0 GiB > 1.4 TiB 62.18 0.83 42 up osd.35 > >>>>> 36 hdd 14.65039 1.00000 15 TiB 11 TiB 11 TiB 544 MiB 26 GiB > 4.0 TiB 72.98 0.98 220 up osd.36 > >>>>> 37 hdd 14.65039 1.00000 15 TiB 11 TiB 11 TiB 11 GiB 38 GiB > 3.6 TiB 75.30 1.01 200 up osd.37 > >>>>> 38 hdd 14.65039 1.00000 15 TiB 11 TiB 11 TiB 1.2 GiB 28 GiB > 3.3 TiB 77.43 1.04 217 up osd.38 > >>>>> -8 58.89636 - 59 TiB 47 TiB 46 TiB 3.9 GiB 116 > GiB 12 TiB 78.98 1.06 - host s3db7 > >>>>> 39 hdd 3.73630 1.00000 3.7 TiB 3.2 TiB 3.2 TiB 19 MiB 8.5 GiB > 499 GiB 86.96 1.17 43 up osd.39 > >>>>> 40 hdd 3.73630 1.00000 3.7 TiB 2.6 TiB 2.5 TiB 144 MiB 7.0 GiB > 1.2 TiB 68.33 0.92 39 up osd.40 > >>>>> 41 hdd 3.73630 1.00000 3.7 TiB 3.0 TiB 2.9 TiB 218 MiB 7.9 GiB > 732 GiB 80.86 1.08 64 up osd.41 > >>>>> 42 hdd 3.73630 1.00000 3.7 TiB 2.5 TiB 2.4 TiB 594 MiB 7.0 GiB > 1.2 TiB 67.97 0.91 50 up osd.42 > >>>>> 43 hdd 14.65039 1.00000 15 TiB 12 TiB 12 TiB 564 MiB 28 GiB > 2.9 TiB 80.32 1.08 213 up osd.43 > >>>>> 44 hdd 14.65039 1.00000 15 TiB 12 TiB 11 TiB 1.3 GiB 28 GiB > 3.1 TiB 78.59 1.05 198 up osd.44 > >>>>> 45 hdd 14.65039 1.00000 15 TiB 12 TiB 12 TiB 1.2 GiB 30 GiB > 2.8 TiB 81.05 1.09 214 up osd.45 > >>>>> -9 51.28331 - 51 TiB 41 TiB 41 TiB 4.9 GiB 108 > GiB 10 TiB 79.75 1.07 - host s3db8 > >>>>> 8 hdd 7.32619 1.00000 7.3 TiB 5.8 TiB 5.8 TiB 472 MiB 15 GiB > 1.5 TiB 79.68 1.07 99 up osd.8 > >>>>> 16 hdd 7.32619 1.00000 7.3 TiB 5.9 TiB 5.8 TiB 785 MiB 15 GiB > 1.4 TiB 80.25 1.08 97 up osd.16 > >>>>> 31 hdd 7.32619 1.00000 7.3 TiB 5.5 TiB 5.5 TiB 438 MiB 14 GiB > 1.8 TiB 75.36 1.01 87 up osd.31 > >>>>> 52 hdd 7.32619 1.00000 7.3 TiB 5.7 TiB 5.7 TiB 844 MiB 15 GiB > 1.6 TiB 78.19 1.05 113 up osd.52 > >>>>> 53 hdd 7.32619 1.00000 7.3 TiB 6.2 TiB 6.1 TiB 792 MiB 18 GiB > 1.1 TiB 84.46 1.13 109 up osd.53 > >>>>> 54 hdd 7.32619 1.00000 7.3 TiB 5.6 TiB 5.6 TiB 959 MiB 15 GiB > 1.7 TiB 76.73 1.03 115 up osd.54 > >>>>> 55 hdd 7.32619 1.00000 7.3 TiB 6.1 TiB 6.1 TiB 699 MiB 16 GiB > 1.2 TiB 83.56 1.12 122 up osd.55 > >>>>> -10 51.28331 - 51 TiB 39 TiB 39 TiB 4.7 GiB 100 > GiB 12 TiB 76.05 1.02 - host s3db9 > >>>>> 56 hdd 7.32619 1.00000 7.3 TiB 5.2 TiB 5.2 TiB 840 MiB 13 GiB > 2.1 TiB 71.06 0.95 105 up osd.56 > >>>>> 57 hdd 7.32619 1.00000 7.3 TiB 6.1 TiB 6.0 TiB 1.0 GiB 16 GiB > 1.2 TiB 83.17 1.12 102 up osd.57 > >>>>> 58 hdd 7.32619 1.00000 7.3 TiB 6.0 TiB 5.9 TiB 43 MiB 15 GiB > 1.4 TiB 81.56 1.09 105 up osd.58 > >>>>> 59 hdd 7.32619 1.00000 7.3 TiB 5.9 TiB 5.9 TiB 429 MiB 15 GiB > 1.4 TiB 80.64 1.08 94 up osd.59 > >>>>> 60 hdd 7.32619 1.00000 7.3 TiB 5.4 TiB 5.3 TiB 226 MiB 14 GiB > 2.0 TiB 73.25 0.98 101 up osd.60 > >>>>> 61 hdd 7.32619 1.00000 7.3 TiB 4.8 TiB 4.8 TiB 1.1 GiB 12 GiB > 2.5 TiB 65.84 0.88 103 up osd.61 > >>>>> 62 hdd 7.32619 1.00000 7.3 TiB 5.6 TiB 5.6 TiB 1.0 GiB 15 GiB > 1.7 TiB 76.83 1.03 126 up osd.62 > >>>>> TOTAL 674 TiB 501 TiB 473 TiB 96 GiB 1.2 TiB > 173 TiB 74.57 > >>>>> MIN/MAX VAR: 0.17/1.20 STDDEV: 10.25 > >>>>> > >>>>> > >>>>> > >>>>> Am Sa., 13. März 2021 um 15:57 Uhr schrieb Dan van der Ster < > dan@xxxxxxxxxxxxxx>: > >>>>>> > >>>>>> No, increasing num PGs won't help substantially. > >>>>>> > >>>>>> Can you share the entire output of ceph osd df tree ? > >>>>>> > >>>>>> Did you already set > >>>>>> > >>>>>> ceph config set mgr mgr/balancer/upmap_max_deviation 1 > >>>>>> > >>>>>> > >>>>>> ?? > >>>>>> And I recommend debug_mgr 4/5 so you can see some basic upmap > balancer logging. > >>>>>> > >>>>>> .. Dan > >>>>>> > >>>>>> > >>>>>> > >>>>>> > >>>>>> > >>>>>> > >>>>>> On Sat, Mar 13, 2021, 3:49 PM Boris Behrens <bb@xxxxxxxxx> wrote: > >>>>>>> > >>>>>>> Hello people, > >>>>>>> > >>>>>>> I am still struggeling with the balancer > >>>>>>> (https://www.mail-archive.com/ceph-users@xxxxxxx/msg09124.html) > >>>>>>> Now I've read some more and might think that I do not have enough > PGs. > >>>>>>> Currently I have 84OSDs and 1024PGs for the main pool (3008 > total). I > >>>>>>> have the autoscaler enabled, but I doesn't tell me to increase the > >>>>>>> PGs. > >>>>>>> > >>>>>>> What do you think? > >>>>>>> > >>>>>>> -- > >>>>>>> Die Selbsthilfegruppe "UTF-8-Probleme" trifft sich diesmal > abweichend > >>>>>>> im groüen Saal. > >>>>>>> _______________________________________________ > >>>>>>> ceph-users mailing list -- ceph-users@xxxxxxx > >>>>>>> To unsubscribe send an email to ceph-users-leave@xxxxxxx > >>>>> > >>>>> > >>>>> > >>>>> -- > >>>>> Die Selbsthilfegruppe "UTF-8-Probleme" trifft sich diesmal > abweichend im groüen Saal. > >>> > >>> > >>> > >>> -- > >>> Die Selbsthilfegruppe "UTF-8-Probleme" trifft sich diesmal abweichend > im groüen Saal. > > > > > > > > -- > > Die Selbsthilfegruppe "UTF-8-Probleme" trifft sich diesmal abweichend im > groüen Saal. > -- Die Selbsthilfegruppe "UTF-8-Probleme" trifft sich diesmal abweichend im groüen Saal. _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx