Re: should I increase the amount of PGs?

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



I raised the backfillfull_ratio to .91 to see what happens, now I am
waiting. Some OSDs were around 89-91%, some are around 50-60%
The pgp_num is on 1946 since one week. I think this will solve itself, when
the cluster becomes a bit more tidy.

Am Di., 30. März 2021 um 15:23 Uhr schrieb Dan van der Ster <
dan@xxxxxxxxxxxxxx>:

> You started with 1024 PGs, and are splitting to 2048.
> Currently there are 1946 PGs used .. so it is nearly there at the goal.
>
> You need to watch that value 1946 and see if it increases slowly. If
> it does not increase, then those backfill_toofull PGs are probably
> splitting PGs, and they are blocked by not having enough free space.
>
> To solve that free space problem, you could either increase the
> backfillfull_ratio like we discussed earlier, or add capacity.
> I prefer the former, if the OSDs are just over the 90% default limit.
>
> -- dan
>
> On Tue, Mar 30, 2021 at 3:18 PM Boris Behrens <bb@xxxxxxxxx> wrote:
> >
> > The output from ceph osd pool ls detail tell me nothing, except that the
> pgp_num is not where it should be. Can you help me to read the output? How
> do I estimate how long the split will take?
> >
> > [root@s3db1 ~]# ceph status
> >   cluster:
> >     id:     dca79fff-ffd0-58f4-1cff-82a2feea05f4
> >     health: HEALTH_WARN
> >             noscrub,nodeep-scrub flag(s) set
> >             10 backfillfull osd(s)
> >             19 nearfull osd(s)
> >             37 pool(s) backfillfull
> >             BlueFS spillover detected on 1 OSD(s)
> >             13 large omap objects
> >             Low space hindering backfill (add storage if this doesn't
> resolve itself): 234 pgs backfill_toofull
> > ...
> >   data:
> >     pools:   37 pools, 4032 pgs
> >     objects: 121.40M objects, 199 TiB
> >     usage:   627 TiB used, 169 TiB / 795 TiB avail
> >     pgs:     45263471/364213596 objects misplaced (12.428%)
> >              3719 active+clean
> >              209  active+remapped+backfill_wait+backfill_toofull
> >              59   active+remapped+backfill_wait
> >              24   active+remapped+backfill_toofull
> >              20   active+remapped+backfilling
> >              1    active+remapped+forced_backfill+backfill_toofull
> >
> >   io:
> >     client:   8.4 MiB/s rd, 127 MiB/s wr, 208 op/s rd, 163 op/s wr
> >     recovery: 276 MiB/s, 164 objects/s
> >
> > [root@s3db1 ~]# ceph osd pool ls detail
> > ...
> > pool 10 'eu-central-1.rgw.buckets.index' replicated size 3 min_size 1
> crush_rule 0 object_hash rjenkins pg_num 64 pgp_num 64 autoscale_mode warn
> last_change 320966 lfor 0/193276/306366 flags hashpspool,backfillfull
> stripe_width 0 application rgw
> > pool 11 'eu-central-1.rgw.buckets.data' replicated size 3 min_size 2
> crush_rule 0 object_hash rjenkins pg_num 2048 pgp_num 1946 pgp_num_target
> 2048 autoscale_mode warn last_change 320966 lfor 0/263549/317774 flags
> hashpspool,backfillfull stripe_width 0 application rgw
> > ...
> >
> > Am Di., 30. März 2021 um 15:07 Uhr schrieb Dan van der Ster <
> dan@xxxxxxxxxxxxxx>:
> >>
> >> It would be safe to turn off the balancer, yes go ahead.
> >>
> >> To know if adding more hardware will help, we need to see how much
> >> longer this current splitting should take. This will help:
> >>
> >>     ceph status
> >>     ceph osd pool ls detail
> >>
> >> -- dan
> >>
> >> On Tue, Mar 30, 2021 at 3:00 PM Boris Behrens <bb@xxxxxxxxx> wrote:
> >> >
> >> > I would think due to splitting, because the balancer doesn't refuses
> it's work, because to many misplaced objects.
> >> > I also think to turn it off for now, so it doesn't begin it's work at
> 5% missplaced objects.
> >> >
> >> > Would adding more hardware help? We wanted to insert another OSD node
> with 7x8TB disks anyway, but postponed it due to the rebalancing.
> >> >
> >> > Am Di., 30. März 2021 um 14:23 Uhr schrieb Dan van der Ster <
> dan@xxxxxxxxxxxxxx>:
> >> >>
> >> >> Are those PGs backfilling due to splitting or due to balancing?
> >> >> If it's the former, I don't think there's a way to pause them with
> >> >> upmap or any other trick.
> >> >>
> >> >> -- dan
> >> >>
> >> >> On Tue, Mar 30, 2021 at 2:07 PM Boris Behrens <bb@xxxxxxxxx> wrote:
> >> >> >
> >> >> > One week later the ceph is still balancing.
> >> >> > What worries me like hell is the %USE on a lot of those OSDs. Does
> ceph
> >> >> > resolv this on it's own? We are currently down to 5TB space in the
> cluster.
> >> >> > Rebalancing single OSDs doesn't work well and it increases the
> "missplaced
> >> >> > objects".
> >> >> >
> >> >> > I thought about letting upmap do some rebalancing. Anyone know if
> this is a
> >> >> > good idea? Or if I should bite my nails an wait as I am the
> headache of my
> >> >> > life.
> >> >> > [root@s3db1 ~]# ceph osd getmap -o om; osdmaptool om --upmap
> out.txt
> >> >> > --upmap-pool eu-central-1.rgw.buckets.data --upmap-max 10; cat
> out.txt
> >> >> > got osdmap epoch 321975
> >> >> > osdmaptool: osdmap file 'om'
> >> >> > writing upmap command output to: out.txt
> >> >> > checking for upmap cleanups
> >> >> > upmap, max-count 10, max deviation 5
> >> >> >  limiting to pools eu-central-1.rgw.buckets.data ([11])
> >> >> > pools eu-central-1.rgw.buckets.data
> >> >> > prepared 10/10 changes
> >> >> > ceph osd rm-pg-upmap-items 11.209
> >> >> > ceph osd rm-pg-upmap-items 11.253
> >> >> > ceph osd pg-upmap-items 11.7f 79 88
> >> >> > ceph osd pg-upmap-items 11.fc 53 31 105 78
> >> >> > ceph osd pg-upmap-items 11.1d8 84 50
> >> >> > ceph osd pg-upmap-items 11.47f 94 86
> >> >> > ceph osd pg-upmap-items 11.49c 44 71
> >> >> > ceph osd pg-upmap-items 11.553 74 50
> >> >> > ceph osd pg-upmap-items 11.6c3 66 63
> >> >> > ceph osd pg-upmap-items 11.7ad 43 50
> >> >> >
> >> >> > ID  CLASS WEIGHT    REWEIGHT SIZE    RAW USE DATA     OMAP     META
> >> >> >  AVAIL    %USE  VAR  PGS STATUS TYPE NAME
> >> >> >  -1       795.42548        - 795 TiB 626 TiB  587 TiB   82 GiB 1.4
> TiB  170
> >> >> > TiB 78.64 1.00   -        root default
> >> >> >  56   hdd   7.32619  1.00000 7.3 TiB 6.4 TiB  6.4 TiB  684 MiB  16
> GiB  910
> >> >> > GiB 87.87 1.12 129     up         osd.56
> >> >> >  67   hdd   7.27739  1.00000 7.3 TiB 6.4 TiB  6.4 TiB  582 MiB  16
> GiB  865
> >> >> > GiB 88.40 1.12 115     up         osd.67
> >> >> >  79   hdd   3.63689  1.00000 3.6 TiB 3.2 TiB  432 GiB  1.9 GiB
>  0 B  432
> >> >> > GiB 88.40 1.12  63     up         osd.79
> >> >> >  53   hdd   7.32619  1.00000 7.3 TiB 6.5 TiB  6.4 TiB  971 MiB  22
> GiB  864
> >> >> > GiB 88.48 1.13 114     up         osd.53
> >> >> >  51   hdd   7.27739  1.00000 7.3 TiB 6.5 TiB  6.4 TiB  734 MiB  15
> GiB  837
> >> >> > GiB 88.77 1.13 120     up         osd.51
> >> >> >  73   hdd  14.55269  1.00000  15 TiB  13 TiB   13 TiB  1.8 GiB  39
> GiB  1.6
> >> >> > TiB 88.97 1.13 246     up         osd.73
> >> >> >  55   hdd   7.32619  1.00000 7.3 TiB 6.5 TiB  6.5 TiB  259 MiB  15
> GiB  825
> >> >> > GiB 89.01 1.13 118     up         osd.55
> >> >> >  70   hdd   7.27739  1.00000 7.3 TiB 6.5 TiB  6.5 TiB  291 MiB  16
> GiB  787
> >> >> > GiB 89.44 1.14 119     up         osd.70
> >> >> >  42   hdd   3.73630  1.00000 3.7 TiB 3.4 TiB  3.3 TiB  685 MiB 8.2
> GiB  374
> >> >> > GiB 90.23 1.15  60     up         osd.42
> >> >> >  94   hdd   3.63869  1.00000 3.6 TiB 3.3 TiB  3.3 TiB  132 MiB 7.7
> GiB  345
> >> >> > GiB 90.75 1.15  64     up         osd.94
> >> >> >  25   hdd   3.73630  1.00000 3.7 TiB 3.4 TiB  3.3 TiB  3.2 MiB 8.1
> GiB  352
> >> >> > GiB 90.79 1.15  53     up         osd.25
> >> >> >  31   hdd   7.32619  1.00000 7.3 TiB 6.7 TiB  6.6 TiB  223 MiB  15
> GiB  690
> >> >> > GiB 90.80 1.15 117     up         osd.31
> >> >> >  84   hdd   7.52150  1.00000 7.5 TiB 6.8 TiB  6.6 TiB  159 MiB  16
> GiB  699
> >> >> > GiB 90.93 1.16 121     up         osd.84
> >> >> >  82   hdd   3.63689  1.00000 3.6 TiB 3.3 TiB  332 GiB  1.0 GiB
>  0 B  332
> >> >> > GiB 91.08 1.16  59     up         osd.82
> >> >> >  89   hdd   7.52150  1.00000 7.5 TiB 6.9 TiB  6.6 TiB  400 MiB  15
> GiB  670
> >> >> > GiB 91.29 1.16 126     up         osd.89
> >> >> >  33   hdd   3.73630  1.00000 3.7 TiB 3.4 TiB  3.3 TiB  382 MiB 8.6
> GiB  327
> >> >> > GiB 91.46 1.16  66     up         osd.33
> >> >> >  90   hdd   7.52150  1.00000 7.5 TiB 6.9 TiB  6.6 TiB  338 MiB  15
> GiB  658
> >> >> > GiB 91.46 1.16 112     up         osd.90
> >> >> > 105   hdd   3.63869  0.89999 3.6 TiB 3.3 TiB  3.3 TiB  206 MiB 8.1
> GiB  301
> >> >> > GiB 91.91 1.17  56     up         osd.105
> >> >> >  66   hdd   7.27739  0.95000 7.3 TiB 6.7 TiB  6.7 TiB  322 MiB  16
> GiB  548
> >> >> > GiB 92.64 1.18 121     up         osd.66
> >> >> >  46   hdd   7.27739  1.00000 7.3 TiB 6.8 TiB  6.7 TiB  316 MiB  16
> GiB  536
> >> >> > GiB 92.81 1.18 119     up         osd.46
> >> >> >
> >> >> > Am Di., 23. März 2021 um 19:59 Uhr schrieb Boris Behrens <
> bb@xxxxxxxxx>:
> >> >> >
> >> >> > > Good point. Thanks for the hint. I changed it for all OSDs from
> 5 to 1
> >> >> > > *crossing finger*
> >> >> > >
> >> >> > > Am Di., 23. März 2021 um 19:45 Uhr schrieb Dan van der Ster <
> >> >> > > dan@xxxxxxxxxxxxxx>:
> >> >> > >
> >> >> > >> I see. When splitting PGs, the OSDs will increase is used space
> >> >> > >> temporarily to make room for the new PGs.
> >> >> > >> When going from 1024->2048 PGs, that means that half of the
> objects from
> >> >> > >> each PG will be copied to a new PG, and then the previous PGs
> will have
> >> >> > >> those objects deleted.
> >> >> > >>
> >> >> > >> Make sure osd_max_backfills is set to 1, so that not too many
> PGs are
> >> >> > >> moving concurrently.
> >> >> > >>
> >> >> > >>
> >> >> > >>
> >> >> > >> On Tue, Mar 23, 2021, 7:39 PM Boris Behrens <bb@xxxxxxxxx>
> wrote:
> >> >> > >>
> >> >> > >>> Thank you.
> >> >> > >>> Currently I do not have any full OSDs (all <90%) but I keep
> this in mind.
> >> >> > >>> What worries me is the ever increasing %USE metric (it went up
> from
> >> >> > >>> around 72% to 75% in three hours). It looks like there is
> comming a lot of
> >> >> > >>> data (there comes barely new data at the moment), but I think
> this might
> >> >> > >>> have to do with my "let's try to increase the PGs to 2048". I
> hope that
> >> >> > >>> ceph begins to split the old PGs into new ones and removes the
> old PGs.
> >> >> > >>>
> >> >> > >>> ID  CLASS WEIGHT    REWEIGHT SIZE    RAW USE DATA    OMAP
> META
> >> >> > >>> AVAIL    %USE  VAR  PGS STATUS TYPE NAME
> >> >> > >>>  -1       795.42548        - 795 TiB 597 TiB 556 TiB  88 GiB
> 1.4 TiB
> >> >> > >>>  198 TiB 75.12 1.00   -        root default
> >> >> > >>>
> >> >> > >>> Am Di., 23. März 2021 um 19:21 Uhr schrieb Dan van der Ster <
> >> >> > >>> dan@xxxxxxxxxxxxxx>:
> >> >> > >>>
> >> >> > >>>> While you're watching things, if an OSD is getting too close
> for
> >> >> > >>>> comfort to the full ratio, you can temporarily increase it,
> e.g.
> >> >> > >>>>     ceph osd set-full-ratio 0.96
> >> >> > >>>>
> >> >> > >>>> But don't set that too high -- you can really break an OSD if
> it gets
> >> >> > >>>> 100% full (and then can't delete objects or whatever...)
> >> >> > >>>>
> >> >> > >>>> -- dan
> >> >> > >>>>
> >> >> > >>>> On Tue, Mar 23, 2021 at 7:17 PM Boris Behrens <bb@xxxxxxxxx>
> wrote:
> >> >> > >>>> >
> >> >> > >>>> > Ok, then I will try to reweight the most filled OSDs to .95
> and see
> >> >> > >>>> if this helps.
> >> >> > >>>> >
> >> >> > >>>> > Am Di., 23. März 2021 um 19:13 Uhr schrieb Dan van der Ster
> <
> >> >> > >>>> dan@xxxxxxxxxxxxxx>:
> >> >> > >>>> >>
> >> >> > >>>> >> Data goes to *all* PGs uniformly.
> >> >> > >>>> >> Max_avail is limited by the available space on the most
> full OSD --
> >> >> > >>>> >> you should pay close attention to those and make sure they
> are moving
> >> >> > >>>> >> in the right direction (decreasing!)
> >> >> > >>>> >>
> >> >> > >>>> >> Another point -- IMHO you should aim to get all PGs
> active+clean
> >> >> > >>>> >> before you add yet another batch of new disks. While there
> are PGs
> >> >> > >>>> >> backfilling, your osdmaps are accumulating on the mons and
> osds --
> >> >> > >>>> >> this itself will start to use a lot of space, and
> active+clean is the
> >> >> > >>>> >> only way to trim the old maps.
> >> >> > >>>> >>
> >> >> > >>>> >> -- dan
> >> >> > >>>> >>
> >> >> > >>>> >> On Tue, Mar 23, 2021 at 7:05 PM Boris Behrens <
> bb@xxxxxxxxx> wrote:
> >> >> > >>>> >> >
> >> >> > >>>> >> > So,
> >> >> > >>>> >> > doing nothing and wait for the ceph to recover?
> >> >> > >>>> >> >
> >> >> > >>>> >> > In theory there should be enough disk space (more disks
> arriving
> >> >> > >>>> tomorrow), but I fear that there might be an issue, when the
> backups get
> >> >> > >>>> exported over night to this s3. Currently the max_avail
> lingers around 13TB
> >> >> > >>>> and I hope, that the data will go to other PGs than the ones
> that are
> >> >> > >>>> currently on filled OSDs.
> >> >> > >>>> >> >
> >> >> > >>>> >> >
> >> >> > >>>> >> >
> >> >> > >>>> >> > Am Di., 23. März 2021 um 18:58 Uhr schrieb Dan van der
> Ster <
> >> >> > >>>> dan@xxxxxxxxxxxxxx>:
> >> >> > >>>> >> >>
> >> >> > >>>> >> >> Hi,
> >> >> > >>>> >> >>
> >> >> > >>>> >> >> backfill_toofull is not a bad thing when the cluster is
> really
> >> >> > >>>> full
> >> >> > >>>> >> >> like yours. You should expect some of the most full
> OSDs to
> >> >> > >>>> eventually
> >> >> > >>>> >> >> start decreasing in usage, as the PGs are moved to the
> new OSDs.
> >> >> > >>>> Those
> >> >> > >>>> >> >> backfill_toofull states should then resolve themselves
> as the OSD
> >> >> > >>>> >> >> usage flattens out.
> >> >> > >>>> >> >> Keep an eye on the usage of the backfill_full and
> nearfull OSDs
> >> >> > >>>> though
> >> >> > >>>> >> >> -- if they do eventually go above the full_ratio (95%
> by default),
> >> >> > >>>> >> >> then writes to those OSDs would stop.
> >> >> > >>>> >> >>
> >> >> > >>>> >> >> But if on the other hand you're suffering from lots of
> slow ops or
> >> >> > >>>> >> >> anything else visible to your users, then you could try
> to take
> >> >> > >>>> some
> >> >> > >>>> >> >> actions to slow down the rebalancing. Just let us know
> if that's
> >> >> > >>>> the
> >> >> > >>>> >> >> case and we can see about changing osd_max_backfills,
> some
> >> >> > >>>> weights or
> >> >> > >>>> >> >> maybe using the upmap-remapped tool.
> >> >> > >>>> >> >>
> >> >> > >>>> >> >> -- Dan
> >> >> > >>>> >> >>
> >> >> > >>>> >> >> On Tue, Mar 23, 2021 at 6:07 PM Boris Behrens <
> bb@xxxxxxxxx>
> >> >> > >>>> wrote:
> >> >> > >>>> >> >> >
> >> >> > >>>> >> >> > Ok, I should have listened to you :)
> >> >> > >>>> >> >> >
> >> >> > >>>> >> >> > In the last week we added more storage but the issue
> got worse
> >> >> > >>>> instead.
> >> >> > >>>> >> >> > Today I realized that the PGs were up to 90GB (bytes
> column in
> >> >> > >>>> ceph pg ls said 95705749636), and the balance kept mentioning
> the 2048 PGs
> >> >> > >>>> for this pool. We were at 72% utilization (ceph osd df tree,
> first line)
> >> >> > >>>> for our cluster and I increased the PGs to 2048.
> >> >> > >>>> >> >> >
> >> >> > >>>> >> >> > Now I am in a world of trouble.
> >> >> > >>>> >> >> > The space in the cluster went down, I am at 45%
> misplaced
> >> >> > >>>> objects, and we already added 20x4TB disks just to not go
> down completly.
> >> >> > >>>> >> >> >
> >> >> > >>>> >> >> > The utilization is still going up and the overall
> free space in
> >> >> > >>>> the cluster seems to go down. This is what my ceph status
> looks like and
> >> >> > >>>> now I really need help to get that thing back to normal:
> >> >> > >>>> >> >> > [root@s3db1 ~]# ceph status
> >> >> > >>>> >> >> >   cluster:
> >> >> > >>>> >> >> >     id:     dca79fff-ffd0-58f4-1cff-82a2feea05f4
> >> >> > >>>> >> >> >     health: HEALTH_WARN
> >> >> > >>>> >> >> >             4 backfillfull osd(s)
> >> >> > >>>> >> >> >             17 nearfull osd(s)
> >> >> > >>>> >> >> >             37 pool(s) backfillfull
> >> >> > >>>> >> >> >             13 large omap objects
> >> >> > >>>> >> >> >             Low space hindering backfill (add storage
> if this
> >> >> > >>>> doesn't resolve itself): 570 pgs backfill_toofull
> >> >> > >>>> >> >> >
> >> >> > >>>> >> >> >   services:
> >> >> > >>>> >> >> >     mon: 3 daemons, quorum
> >> >> > >>>> ceph-s3-mon1,ceph-s3-mon2,ceph-s3-mon3 (age 44m)
> >> >> > >>>> >> >> >     mgr: ceph-mgr2(active, since 15m), standbys:
> ceph-mgr3,
> >> >> > >>>> ceph-mgr1
> >> >> > >>>> >> >> >     mds:  3 up:standby
> >> >> > >>>> >> >> >     osd: 110 osds: 110 up (since 28m), 110 in (since
> 28m); 1535
> >> >> > >>>> remapped pgs
> >> >> > >>>> >> >> >     rgw: 3 daemons active (eu-central-1, eu-msg-1,
> eu-secure-1)
> >> >> > >>>> >> >> >
> >> >> > >>>> >> >> >   task status:
> >> >> > >>>> >> >> >
> >> >> > >>>> >> >> >   data:
> >> >> > >>>> >> >> >     pools:   37 pools, 4032 pgs
> >> >> > >>>> >> >> >     objects: 116.23M objects, 182 TiB
> >> >> > >>>> >> >> >     usage:   589 TiB used, 206 TiB / 795 TiB avail
> >> >> > >>>> >> >> >     pgs:     160918554/348689415 objects misplaced
> (46.150%)
> >> >> > >>>> >> >> >              2497 active+clean
> >> >> > >>>> >> >> >              779  active+remapped+backfill_wait
> >> >> > >>>> >> >> >              538
> active+remapped+backfill_wait+backfill_toofull
> >> >> > >>>> >> >> >              186  active+remapped+backfilling
> >> >> > >>>> >> >> >              32   active+remapped+backfill_toofull
> >> >> > >>>> >> >> >
> >> >> > >>>> >> >> >   io:
> >> >> > >>>> >> >> >     client:   27 MiB/s rd, 69 MiB/s wr, 497 op/s rd,
> 153 op/s wr
> >> >> > >>>> >> >> >     recovery: 1.5 GiB/s, 922 objects/s
> >> >> > >>>> >> >> >
> >> >> > >>>> >> >> > Am Di., 16. März 2021 um 09:34 Uhr schrieb Boris
> Behrens <
> >> >> > >>>> bb@xxxxxxxxx>:
> >> >> > >>>> >> >> >>
> >> >> > >>>> >> >> >> Hi Dan,
> >> >> > >>>> >> >> >>
> >> >> > >>>> >> >> >> my EC profile look very "default" to me.
> >> >> > >>>> >> >> >> [root@s3db1 ~]# ceph osd erasure-code-profile ls
> >> >> > >>>> >> >> >> default
> >> >> > >>>> >> >> >> [root@s3db1 ~]# ceph osd erasure-code-profile get
> default
> >> >> > >>>> >> >> >> k=2
> >> >> > >>>> >> >> >> m=1
> >> >> > >>>> >> >> >> plugin=jerasure
> >> >> > >>>> >> >> >> technique=reed_sol_van
> >> >> > >>>> >> >> >>
> >> >> > >>>> >> >> >> I don't understand the ouput, but the balancing get
> worse over
> >> >> > >>>> night:
> >> >> > >>>> >> >> >>
> >> >> > >>>> >> >> >> [root@s3db1 ~]#
> ceph-scripts/tools/ceph-pool-pg-distribution
> >> >> > >>>> 11
> >> >> > >>>> >> >> >> Searching for PGs in pools: ['11']
> >> >> > >>>> >> >> >> Summary: 1024 PGs on 84 osds
> >> >> > >>>> >> >> >>
> >> >> > >>>> >> >> >> Num OSDs with X PGs:
> >> >> > >>>> >> >> >> 15: 8
> >> >> > >>>> >> >> >> 16: 7
> >> >> > >>>> >> >> >> 17: 6
> >> >> > >>>> >> >> >> 18: 10
> >> >> > >>>> >> >> >> 19: 1
> >> >> > >>>> >> >> >> 32: 10
> >> >> > >>>> >> >> >> 33: 4
> >> >> > >>>> >> >> >> 34: 6
> >> >> > >>>> >> >> >> 35: 8
> >> >> > >>>> >> >> >> 65: 5
> >> >> > >>>> >> >> >> 66: 5
> >> >> > >>>> >> >> >> 67: 4
> >> >> > >>>> >> >> >> 68: 10
> >> >> > >>>> >> >> >> [root@s3db1 ~]# ceph-scripts/tools/ceph-pg-histogram
> >> >> > >>>> --normalize --pool=11
> >> >> > >>>> >> >> >> # NumSamples = 84; Min = 4.12; Max = 5.09
> >> >> > >>>> >> >> >> # Mean = 4.553355; Variance = 0.052415; SD =
> 0.228942; Median
> >> >> > >>>> 4.561608
> >> >> > >>>> >> >> >> # each ∎ represents a count of 1
> >> >> > >>>> >> >> >>     4.1244 -     4.2205 [     8]: ∎∎∎∎∎∎∎∎
> >> >> > >>>> >> >> >>     4.2205 -     4.3166 [     6]: ∎∎∎∎∎∎
> >> >> > >>>> >> >> >>     4.3166 -     4.4127 [    11]: ∎∎∎∎∎∎∎∎∎∎∎
> >> >> > >>>> >> >> >>     4.4127 -     4.5087 [    10]: ∎∎∎∎∎∎∎∎∎∎
> >> >> > >>>> >> >> >>     4.5087 -     4.6048 [    11]: ∎∎∎∎∎∎∎∎∎∎∎
> >> >> > >>>> >> >> >>     4.6048 -     4.7009 [    19]: ∎∎∎∎∎∎∎∎∎∎∎∎∎∎∎∎∎∎∎
> >> >> > >>>> >> >> >>     4.7009 -     4.7970 [     6]: ∎∎∎∎∎∎
> >> >> > >>>> >> >> >>     4.7970 -     4.8931 [     8]: ∎∎∎∎∎∎∎∎
> >> >> > >>>> >> >> >>     4.8931 -     4.9892 [     4]: ∎∎∎∎
> >> >> > >>>> >> >> >>     4.9892 -     5.0852 [     1]: ∎
> >> >> > >>>> >> >> >> [root@s3db1 ~]# ceph osd df tree | sort -nk 17 |
> tail
> >> >> > >>>> >> >> >>  14   hdd   3.63689  1.00000 3.6 TiB 2.9 TiB 724
> GiB   19 GiB
> >> >> > >>>>    0 B 724 GiB 80.56 1.07  56     up         osd.14
> >> >> > >>>> >> >> >>  19   hdd   3.68750  1.00000 3.7 TiB 3.0 TiB 2.9
> TiB  466 MiB
> >> >> > >>>> 7.9 GiB 708 GiB 81.25 1.08  53     up         osd.19
> >> >> > >>>> >> >> >>   4   hdd   3.63689  1.00000 3.6 TiB 3.0 TiB 698
> GiB  703 MiB
> >> >> > >>>>    0 B 698 GiB 81.27 1.08  48     up         osd.4
> >> >> > >>>> >> >> >>  24   hdd   3.63689  1.00000 3.6 TiB 3.0 TiB 695
> GiB  640 MiB
> >> >> > >>>>    0 B 695 GiB 81.34 1.08  46     up         osd.24
> >> >> > >>>> >> >> >>  75   hdd   3.68750  1.00000 3.7 TiB 3.0 TiB 2.9
> TiB  440 MiB
> >> >> > >>>> 8.1 GiB 704 GiB 81.35 1.08  48     up         osd.75
> >> >> > >>>> >> >> >>  71   hdd   3.68750  1.00000 3.7 TiB 3.0 TiB 3.0
> TiB  7.5 MiB
> >> >> > >>>> 8.0 GiB 663 GiB 82.44 1.09  47     up         osd.71
> >> >> > >>>> >> >> >>  76   hdd   3.68750  1.00000 3.7 TiB 3.1 TiB 3.0
> TiB  251 MiB
> >> >> > >>>> 9.0 GiB 617 GiB 83.65 1.11  50     up         osd.76
> >> >> > >>>> >> >> >>  33   hdd   3.73630  1.00000 3.7 TiB 3.1 TiB 3.0
> TiB  399 MiB
> >> >> > >>>> 8.1 GiB 618 GiB 83.85 1.11  55     up         osd.33
> >> >> > >>>> >> >> >>  35   hdd   3.73630  1.00000 3.7 TiB 3.1 TiB 3.0
> TiB  317 MiB
> >> >> > >>>> 8.8 GiB 617 GiB 83.87 1.11  50     up         osd.35
> >> >> > >>>> >> >> >>  34   hdd   3.73630  1.00000 3.7 TiB 3.2 TiB 3.1
> TiB  451 MiB
> >> >> > >>>> 8.7 GiB 545 GiB 85.75 1.14  54     up         osd.34
> >> >> > >>>> >> >> >>
> >> >> > >>>> >> >> >> Am Mo., 15. März 2021 um 17:23 Uhr schrieb Dan van
> der Ster <
> >> >> > >>>> dan@xxxxxxxxxxxxxx>:
> >> >> > >>>> >> >> >>>
> >> >> > >>>> >> >> >>> Hi,
> >> >> > >>>> >> >> >>>
> >> >> > >>>> >> >> >>> How wide are your EC profiles? If they are really
> wide, you
> >> >> > >>>> might be
> >> >> > >>>> >> >> >>> reaching the limits of what is physically possible.
> Also, I'm
> >> >> > >>>> not sure
> >> >> > >>>> >> >> >>> that upmap in 14.2.11 is very smart about
> *improving*
> >> >> > >>>> existing upmap
> >> >> > >>>> >> >> >>> rules for a given PG, in the case that a PG already
> has an
> >> >> > >>>> upmap-items
> >> >> > >>>> >> >> >>> entry but it would help the distribution to add
> more mapping
> >> >> > >>>> pairs to
> >> >> > >>>> >> >> >>> that entry. What this means, is that it might
> sometimes be
> >> >> > >>>> useful to
> >> >> > >>>> >> >> >>> randomly remove some upmap entries and see if the
> balancer
> >> >> > >>>> does a
> >> >> > >>>> >> >> >>> better job when it replaces them.
> >> >> > >>>> >> >> >>>
> >> >> > >>>> >> >> >>> But before you do that, I re-remembered that
> looking at the
> >> >> > >>>> total PG
> >> >> > >>>> >> >> >>> numbers is not useful -- you need to check the PGs
> per OSD
> >> >> > >>>> for the
> >> >> > >>>> >> >> >>> eu-central-1.rgw.buckets.data pool only.
> >> >> > >>>> >> >> >>>
> >> >> > >>>> >> >> >>> We have a couple tools that can help with this:
> >> >> > >>>> >> >> >>>
> >> >> > >>>> >> >> >>> 1. To see the PGs per OSD for a given pool:
> >> >> > >>>> >> >> >>>
> >> >> > >>>>
> https://github.com/cernceph/ceph-scripts/blob/master/tools/ceph-pool-pg-distribution
> >> >> > >>>> >> >> >>>
> >> >> > >>>> >> >> >>>     E.g.: ./ceph-pool-pg-distribution 11  # to see
> the
> >> >> > >>>> distribution of
> >> >> > >>>> >> >> >>> your eu-central-1.rgw.buckets.data pool.
> >> >> > >>>> >> >> >>>
> >> >> > >>>> >> >> >>>     The output looks like this on my well balanced
> clusters:
> >> >> > >>>> >> >> >>>
> >> >> > >>>> >> >> >>> # ceph-scripts/tools/ceph-pool-pg-distribution 15
> >> >> > >>>> >> >> >>> Searching for PGs in pools: ['15']
> >> >> > >>>> >> >> >>> Summary: 256 pgs on 56 osds
> >> >> > >>>> >> >> >>>
> >> >> > >>>> >> >> >>> Num OSDs with X PGs:
> >> >> > >>>> >> >> >>>  13: 16
> >> >> > >>>> >> >> >>>  14: 40
> >> >> > >>>> >> >> >>>
> >> >> > >>>> >> >> >>>     You should expect a trimodal for your cluster.
> >> >> > >>>> >> >> >>>
> >> >> > >>>> >> >> >>> 2. You can also use another script from that repo
> to see the
> >> >> > >>>> PGs per
> >> >> > >>>> >> >> >>> OSD normalized to crush weight:
> >> >> > >>>> >> >> >>>     ceph-scripts/tools/ceph-pg-histogram
> --normalize --pool=15
> >> >> > >>>> >> >> >>>
> >> >> > >>>> >> >> >>>    This might explain what is going wrong.
> >> >> > >>>> >> >> >>>
> >> >> > >>>> >> >> >>> Cheers, Dan
> >> >> > >>>> >> >> >>>
> >> >> > >>>> >> >> >>>
> >> >> > >>>> >> >> >>> On Mon, Mar 15, 2021 at 3:04 PM Boris Behrens <
> bb@xxxxxxxxx>
> >> >> > >>>> wrote:
> >> >> > >>>> >> >> >>> >
> >> >> > >>>> >> >> >>> > Absolutly:
> >> >> > >>>> >> >> >>> > [root@s3db1 ~]# ceph osd df tree
> >> >> > >>>> >> >> >>> > ID  CLASS WEIGHT    REWEIGHT SIZE    RAW USE
> DATA     OMAP
> >> >> > >>>>    META    AVAIL    %USE  VAR  PGS STATUS TYPE NAME
> >> >> > >>>> >> >> >>> >  -1       673.54224        - 674 TiB 496 TiB  468
> TiB   97
> >> >> > >>>> GiB 1.2 TiB  177 TiB 73.67 1.00   -        root default
> >> >> > >>>> >> >> >>> >  -2        58.30331        -  58 TiB  42 TiB   38
> TiB  9.2
> >> >> > >>>> GiB  99 GiB   16 TiB 72.88 0.99   -            host s3db1
> >> >> > >>>> >> >> >>> >  23   hdd  14.65039  1.00000  15 TiB  11 TiB   11
> TiB  714
> >> >> > >>>> MiB  25 GiB  3.7 TiB 74.87 1.02 194     up         osd.23
> >> >> > >>>> >> >> >>> >  69   hdd  14.55269  1.00000  15 TiB  11 TiB   11
> TiB  1.6
> >> >> > >>>> GiB  40 GiB  3.4 TiB 76.32 1.04 199     up         osd.69
> >> >> > >>>> >> >> >>> >  73   hdd  14.55269  1.00000  15 TiB  11 TiB   11
> TiB  1.3
> >> >> > >>>> GiB  34 GiB  3.8 TiB 74.15 1.01 203     up         osd.73
> >> >> > >>>> >> >> >>> >  79   hdd   3.63689  1.00000 3.6 TiB 2.4 TiB  1.3
> TiB  1.8
> >> >> > >>>> GiB     0 B  1.3 TiB 65.44 0.89  47     up         osd.79
> >> >> > >>>> >> >> >>> >  80   hdd   3.63689  1.00000 3.6 TiB 2.4 TiB  1.3
> TiB  2.2
> >> >> > >>>> GiB     0 B  1.3 TiB 65.34 0.89  48     up         osd.80
> >> >> > >>>> >> >> >>> >  81   hdd   3.63689  1.00000 3.6 TiB 2.4 TiB  1.3
> TiB  1.1
> >> >> > >>>> GiB     0 B  1.3 TiB 65.38 0.89  47     up         osd.81
> >> >> > >>>> >> >> >>> >  82   hdd   3.63689  1.00000 3.6 TiB 2.5 TiB  1.1
> TiB  619
> >> >> > >>>> MiB     0 B  1.1 TiB 68.46 0.93  41     up         osd.82
> >> >> > >>>> >> >> >>> > -11        50.94173        -  51 TiB  37 TiB   37
> TiB  3.5
> >> >> > >>>> GiB  98 GiB   14 TiB 71.90 0.98   -            host s3db10
> >> >> > >>>> >> >> >>> >  63   hdd   7.27739  1.00000 7.3 TiB 5.3 TiB  5.3
> TiB  647
> >> >> > >>>> MiB  14 GiB  2.0 TiB 72.72 0.99  94     up         osd.63
> >> >> > >>>> >> >> >>> >  64   hdd   7.27739  1.00000 7.3 TiB 5.3 TiB  5.2
> TiB  668
> >> >> > >>>> MiB  14 GiB  2.0 TiB 72.23 0.98  93     up         osd.64
> >> >> > >>>> >> >> >>> >  65   hdd   7.27739  1.00000 7.3 TiB 5.2 TiB  5.2
> TiB  227
> >> >> > >>>> MiB  14 GiB  2.1 TiB 71.16 0.97 100     up         osd.65
> >> >> > >>>> >> >> >>> >  66   hdd   7.27739  1.00000 7.3 TiB 5.4 TiB  5.4
> TiB  313
> >> >> > >>>> MiB  14 GiB  1.9 TiB 74.25 1.01  92     up         osd.66
> >> >> > >>>> >> >> >>> >  67   hdd   7.27739  1.00000 7.3 TiB 5.1 TiB  5.1
> TiB  584
> >> >> > >>>> MiB  14 GiB  2.1 TiB 70.63 0.96  96     up         osd.67
> >> >> > >>>> >> >> >>> >  68   hdd   7.27739  1.00000 7.3 TiB 5.2 TiB  5.2
> TiB  720
> >> >> > >>>> MiB  14 GiB  2.1 TiB 71.72 0.97 101     up         osd.68
> >> >> > >>>> >> >> >>> >  70   hdd   7.27739  1.00000 7.3 TiB 5.1 TiB  5.1
> TiB  425
> >> >> > >>>> MiB  14 GiB  2.1 TiB 70.59 0.96  97     up         osd.70
> >> >> > >>>> >> >> >>> > -12        50.99052        -  51 TiB  38 TiB   37
> TiB  2.1
> >> >> > >>>> GiB  97 GiB   13 TiB 73.77 1.00   -            host s3db11
> >> >> > >>>> >> >> >>> >  46   hdd   7.27739  1.00000 7.3 TiB 5.6 TiB  5.6
> TiB  229
> >> >> > >>>> MiB  14 GiB  1.7 TiB 77.05 1.05  97     up         osd.46
> >> >> > >>>> >> >> >>> >  47   hdd   7.27739  1.00000 7.3 TiB 5.1 TiB  5.1
> TiB  159
> >> >> > >>>> MiB  13 GiB  2.2 TiB 70.00 0.95  89     up         osd.47
> >> >> > >>>> >> >> >>> >  48   hdd   7.27739  1.00000 7.3 TiB 5.2 TiB  5.2
> TiB  279
> >> >> > >>>> MiB  14 GiB  2.1 TiB 71.82 0.97  98     up         osd.48
> >> >> > >>>> >> >> >>> >  49   hdd   7.27739  1.00000 7.3 TiB 5.5 TiB  5.4
> TiB  276
> >> >> > >>>> MiB  14 GiB  1.8 TiB 74.90 1.02  95     up         osd.49
> >> >> > >>>> >> >> >>> >  50   hdd   7.27739  1.00000 7.3 TiB 5.2 TiB  5.2
> TiB  336
> >> >> > >>>> MiB  14 GiB  2.0 TiB 72.13 0.98  93     up         osd.50
> >> >> > >>>> >> >> >>> >  51   hdd   7.27739  1.00000 7.3 TiB 5.7 TiB  5.6
> TiB  728
> >> >> > >>>> MiB  15 GiB  1.6 TiB 77.76 1.06  98     up         osd.51
> >> >> > >>>> >> >> >>> >  72   hdd   7.32619  1.00000 7.3 TiB 5.3 TiB  5.3
> TiB  147
> >> >> > >>>> MiB  13 GiB  2.0 TiB 72.75 0.99  95     up         osd.72
> >> >> > >>>> >> >> >>> > -37        58.55478        -  59 TiB  44 TiB   44
> TiB  4.4
> >> >> > >>>> GiB 122 GiB   15 TiB 75.20 1.02   -            host s3db12
> >> >> > >>>> >> >> >>> >  19   hdd   3.68750  1.00000 3.7 TiB 2.9 TiB  2.9
> TiB  454
> >> >> > >>>> MiB 8.2 GiB  780 GiB 79.35 1.08  53     up         osd.19
> >> >> > >>>> >> >> >>> >  71   hdd   3.68750  1.00000 3.7 TiB 3.0 TiB  2.9
> TiB  7.1
> >> >> > >>>> MiB 8.0 GiB  734 GiB 80.56 1.09  47     up         osd.71
> >> >> > >>>> >> >> >>> >  75   hdd   3.68750  1.00000 3.7 TiB 2.9 TiB  2.9
> TiB  439
> >> >> > >>>> MiB 8.2 GiB  777 GiB 79.43 1.08  48     up         osd.75
> >> >> > >>>> >> >> >>> >  76   hdd   3.68750  1.00000 3.7 TiB 3.0 TiB  3.0
> TiB  241
> >> >> > >>>> MiB 8.9 GiB  688 GiB 81.77 1.11  50     up         osd.76
> >> >> > >>>> >> >> >>> >  77   hdd  14.60159  1.00000  15 TiB  11 TiB   11
> TiB  880
> >> >> > >>>> MiB  30 GiB  3.6 TiB 75.44 1.02 201     up         osd.77
> >> >> > >>>> >> >> >>> >  78   hdd  14.60159  1.00000  15 TiB  10 TiB   10
> TiB 1015
> >> >> > >>>> MiB  28 GiB  4.2 TiB 71.26 0.97 193     up         osd.78
> >> >> > >>>> >> >> >>> >  83   hdd  14.60159  1.00000  15 TiB  11 TiB   11
> TiB  1.4
> >> >> > >>>> GiB  30 GiB  3.8 TiB 73.76 1.00 203     up         osd.83
> >> >> > >>>> >> >> >>> >  -3        58.49872        -  58 TiB  42 TiB   36
> TiB  8.2
> >> >> > >>>> GiB  89 GiB   17 TiB 71.71 0.97   -            host s3db2
> >> >> > >>>> >> >> >>> >   1   hdd  14.65039  1.00000  15 TiB  11 TiB   11
> TiB  3.2
> >> >> > >>>> GiB  37 GiB  3.7 TiB 74.58 1.01 196     up         osd.1
> >> >> > >>>> >> >> >>> >   3   hdd   3.63689  1.00000 3.6 TiB 2.3 TiB  1.3
> TiB  566
> >> >> > >>>> MiB     0 B  1.3 TiB 64.11 0.87  50     up         osd.3
> >> >> > >>>> >> >> >>> >   4   hdd   3.63689  1.00000 3.6 TiB 2.9 TiB  771
> GiB  695
> >> >> > >>>> MiB     0 B  771 GiB 79.30 1.08  48     up         osd.4
> >> >> > >>>> >> >> >>> >   5   hdd   3.63689  1.00000 3.6 TiB 2.4 TiB  1.2
> TiB  482
> >> >> > >>>> MiB     0 B  1.2 TiB 66.51 0.90  49     up         osd.5
> >> >> > >>>> >> >> >>> >   6   hdd   3.63689  1.00000 3.6 TiB 2.3 TiB  1.3
> TiB  1.8
> >> >> > >>>> GiB     0 B  1.3 TiB 64.00 0.87  42     up         osd.6
> >> >> > >>>> >> >> >>> >   7   hdd  14.65039  1.00000  15 TiB  11 TiB   11
> TiB  639
> >> >> > >>>> MiB  26 GiB  4.0 TiB 72.44 0.98 192     up         osd.7
> >> >> > >>>> >> >> >>> >  74   hdd  14.65039  1.00000  15 TiB  10 TiB   10
> TiB  907
> >> >> > >>>> MiB  26 GiB  4.2 TiB 71.32 0.97 193     up         osd.74
> >> >> > >>>> >> >> >>> >  -4        58.49872        -  58 TiB  43 TiB   36
> TiB   34
> >> >> > >>>> GiB  85 GiB   16 TiB 72.69 0.99   -            host s3db3
> >> >> > >>>> >> >> >>> >   2   hdd  14.65039  1.00000  15 TiB  11 TiB   11
> TiB  980
> >> >> > >>>> MiB  26 GiB  3.8 TiB 74.36 1.01 203     up         osd.2
> >> >> > >>>> >> >> >>> >   9   hdd  14.65039  1.00000  15 TiB  11 TiB   11
> TiB  8.4
> >> >> > >>>> GiB  33 GiB  3.9 TiB 73.51 1.00 186     up         osd.9
> >> >> > >>>> >> >> >>> >  10   hdd  14.65039  1.00000  15 TiB  10 TiB   10
> TiB  650
> >> >> > >>>> MiB  26 GiB  4.2 TiB 71.64 0.97 201     up         osd.10
> >> >> > >>>> >> >> >>> >  12   hdd   3.63689  1.00000 3.6 TiB 2.3 TiB  1.3
> TiB  754
> >> >> > >>>> MiB     0 B  1.3 TiB 64.17 0.87  44     up         osd.12
> >> >> > >>>> >> >> >>> >  13   hdd   3.63689  1.00000 3.6 TiB 2.8 TiB  813
> GiB  2.4
> >> >> > >>>> GiB     0 B  813 GiB 78.17 1.06  58     up         osd.13
> >> >> > >>>> >> >> >>> >  14   hdd   3.63689  1.00000 3.6 TiB 2.9 TiB  797
> GiB   19
> >> >> > >>>> GiB     0 B  797 GiB 78.60 1.07  56     up         osd.14
> >> >> > >>>> >> >> >>> >  15   hdd   3.63689  1.00000 3.6 TiB 2.3 TiB  1.3
> TiB  2.2
> >> >> > >>>> GiB     0 B  1.3 TiB 63.96 0.87  41     up         osd.15
> >> >> > >>>> >> >> >>> >  -5        58.49872        -  58 TiB  43 TiB   36
> TiB  6.7
> >> >> > >>>> GiB  97 GiB   15 TiB 74.04 1.01   -            host s3db4
> >> >> > >>>> >> >> >>> >  11   hdd  14.65039  1.00000  15 TiB  11 TiB   11
> TiB  940
> >> >> > >>>> MiB  26 GiB  4.0 TiB 72.49 0.98 196     up         osd.11
> >> >> > >>>> >> >> >>> >  17   hdd  14.65039  1.00000  15 TiB  11 TiB   11
> TiB 1022
> >> >> > >>>> MiB  26 GiB  3.6 TiB 75.23 1.02 204     up         osd.17
> >> >> > >>>> >> >> >>> >  18   hdd  14.65039  1.00000  15 TiB  11 TiB   11
> TiB  945
> >> >> > >>>> MiB  45 GiB  3.8 TiB 74.16 1.01 193     up         osd.18
> >> >> > >>>> >> >> >>> >  20   hdd   3.63689  1.00000 3.6 TiB 2.6 TiB 1020
> GiB  596
> >> >> > >>>> MiB     0 B 1020 GiB 72.62 0.99  57     up         osd.20
> >> >> > >>>> >> >> >>> >  21   hdd   3.63689  1.00000 3.6 TiB 2.6 TiB 1023
> GiB  1.9
> >> >> > >>>> GiB     0 B 1023 GiB 72.54 0.98  41     up         osd.21
> >> >> > >>>> >> >> >>> >  22   hdd   3.63689  1.00000 3.6 TiB 2.6 TiB 1023
> GiB  797
> >> >> > >>>> MiB     0 B 1023 GiB 72.54 0.98  53     up         osd.22
> >> >> > >>>> >> >> >>> >  24   hdd   3.63689  1.00000 3.6 TiB 2.9 TiB  766
> GiB  618
> >> >> > >>>> MiB     0 B  766 GiB 79.42 1.08  46     up         osd.24
> >> >> > >>>> >> >> >>> >  -6        58.89636        -  59 TiB  43 TiB   43
> TiB  3.0
> >> >> > >>>> GiB 108 GiB   16 TiB 73.40 1.00   -            host s3db5
> >> >> > >>>> >> >> >>> >   0   hdd   3.73630  1.00000 3.7 TiB 2.7 TiB  2.6
> TiB   92
> >> >> > >>>> MiB 7.2 GiB  1.1 TiB 71.16 0.97  45     up         osd.0
> >> >> > >>>> >> >> >>> >  25   hdd   3.73630  1.00000 3.7 TiB 2.7 TiB  2.6
> TiB  2.4
> >> >> > >>>> MiB 7.3 GiB  1.1 TiB 71.23 0.97  41     up         osd.25
> >> >> > >>>> >> >> >>> >  26   hdd   3.73630  1.00000 3.7 TiB 2.8 TiB  2.7
> TiB  181
> >> >> > >>>> MiB 7.6 GiB  935 GiB 75.57 1.03  45     up         osd.26
> >> >> > >>>> >> >> >>> >  27   hdd   3.73630  1.00000 3.7 TiB 2.7 TiB  2.6
> TiB  5.1
> >> >> > >>>> MiB 7.0 GiB  1.1 TiB 71.20 0.97  47     up         osd.27
> >> >> > >>>> >> >> >>> >  28   hdd  14.65039  1.00000  15 TiB  11 TiB   11
> TiB  977
> >> >> > >>>> MiB  26 GiB  3.8 TiB 73.85 1.00 197     up         osd.28
> >> >> > >>>> >> >> >>> >  29   hdd  14.65039  1.00000  15 TiB  11 TiB   10
> TiB  872
> >> >> > >>>> MiB  26 GiB  4.1 TiB 71.98 0.98 196     up         osd.29
> >> >> > >>>> >> >> >>> >  30   hdd  14.65039  1.00000  15 TiB  11 TiB   11
> TiB  943
> >> >> > >>>> MiB  27 GiB  3.6 TiB 75.51 1.03 202     up         osd.30
> >> >> > >>>> >> >> >>> >  -7        58.89636        -  59 TiB  44 TiB   43
> TiB   13
> >> >> > >>>> GiB 122 GiB   15 TiB 74.97 1.02   -            host s3db6
> >> >> > >>>> >> >> >>> >  32   hdd   3.73630  1.00000 3.7 TiB 2.8 TiB  2.7
> TiB   27
> >> >> > >>>> MiB 7.6 GiB  940 GiB 75.42 1.02  55     up         osd.32
> >> >> > >>>> >> >> >>> >  33   hdd   3.73630  1.00000 3.7 TiB 3.1 TiB  3.0
> TiB  376
> >> >> > >>>> MiB 8.2 GiB  691 GiB 81.94 1.11  55     up         osd.33
> >> >> > >>>> >> >> >>> >  34   hdd   3.73630  1.00000 3.7 TiB 3.1 TiB  3.0
> TiB  450
> >> >> > >>>> MiB 8.5 GiB  620 GiB 83.79 1.14  54     up         osd.34
> >> >> > >>>> >> >> >>> >  35   hdd   3.73630  1.00000 3.7 TiB 3.1 TiB  3.0
> TiB  316
> >> >> > >>>> MiB 8.4 GiB  690 GiB 81.98 1.11  50     up         osd.35
> >> >> > >>>> >> >> >>> >  36   hdd  14.65039  1.00000  15 TiB  11 TiB   10
> TiB  489
> >> >> > >>>> MiB  25 GiB  4.1 TiB 71.69 0.97 208     up         osd.36
> >> >> > >>>> >> >> >>> >  37   hdd  14.65039  1.00000  15 TiB  11 TiB   11
> TiB   11
> >> >> > >>>> GiB  38 GiB  4.0 TiB 72.41 0.98 195     up         osd.37
> >> >> > >>>> >> >> >>> >  38   hdd  14.65039  1.00000  15 TiB  11 TiB   11
> TiB  1.1
> >> >> > >>>> GiB  26 GiB  3.7 TiB 74.88 1.02 204     up         osd.38
> >> >> > >>>> >> >> >>> >  -8        58.89636        -  59 TiB  44 TiB   43
> TiB  3.8
> >> >> > >>>> GiB 111 GiB   15 TiB 74.16 1.01   -            host s3db7
> >> >> > >>>> >> >> >>> >  39   hdd   3.73630  1.00000 3.7 TiB 2.8 TiB  2.7
> TiB   19
> >> >> > >>>> MiB 7.5 GiB  936 GiB 75.54 1.03  39     up         osd.39
> >> >> > >>>> >> >> >>> >  40   hdd   3.73630  1.00000 3.7 TiB 2.6 TiB  2.5
> TiB  144
> >> >> > >>>> MiB 7.1 GiB  1.1 TiB 69.87 0.95  39     up         osd.40
> >> >> > >>>> >> >> >>> >  41   hdd   3.73630  1.00000 3.7 TiB 2.7 TiB  2.7
> TiB  219
> >> >> > >>>> MiB 7.6 GiB 1011 GiB 73.57 1.00  55     up         osd.41
> >> >> > >>>> >> >> >>> >  42   hdd   3.73630  1.00000 3.7 TiB 2.6 TiB  2.5
> TiB  593
> >> >> > >>>> MiB 7.1 GiB  1.1 TiB 70.02 0.95  47     up         osd.42
> >> >> > >>>> >> >> >>> >  43   hdd  14.65039  1.00000  15 TiB  11 TiB   11
> TiB  500
> >> >> > >>>> MiB  27 GiB  3.7 TiB 74.67 1.01 204     up         osd.43
> >> >> > >>>> >> >> >>> >  44   hdd  14.65039  1.00000  15 TiB  11 TiB   11
> TiB  1.1
> >> >> > >>>> GiB  27 GiB  3.7 TiB 74.62 1.01 193     up         osd.44
> >> >> > >>>> >> >> >>> >  45   hdd  14.65039  1.00000  15 TiB  11 TiB   11
> TiB  1.2
> >> >> > >>>> GiB  29 GiB  3.6 TiB 75.16 1.02 204     up         osd.45
> >> >> > >>>> >> >> >>> >  -9        51.28331        -  51 TiB  39 TiB   39
> TiB  4.9
> >> >> > >>>> GiB 107 GiB   12 TiB 76.50 1.04   -            host s3db8
> >> >> > >>>> >> >> >>> >   8   hdd   7.32619  1.00000 7.3 TiB 5.6 TiB  5.5
> TiB  474
> >> >> > >>>> MiB  14 GiB  1.7 TiB 76.37 1.04  98     up         osd.8
> >> >> > >>>> >> >> >>> >  16   hdd   7.32619  1.00000 7.3 TiB 5.7 TiB  5.7
> TiB  783
> >> >> > >>>> MiB  15 GiB  1.6 TiB 78.39 1.06 100     up         osd.16
> >> >> > >>>> >> >> >>> >  31   hdd   7.32619  1.00000 7.3 TiB 5.7 TiB  5.6
> TiB  441
> >> >> > >>>> MiB  14 GiB  1.6 TiB 77.70 1.05  91     up         osd.31
> >> >> > >>>> >> >> >>> >  52   hdd   7.32619  1.00000 7.3 TiB 5.6 TiB  5.5
> TiB  939
> >> >> > >>>> MiB  14 GiB  1.7 TiB 76.29 1.04 102     up         osd.52
> >> >> > >>>> >> >> >>> >  53   hdd   7.32619  1.00000 7.3 TiB 5.4 TiB  5.4
> TiB  848
> >> >> > >>>> MiB  18 GiB  1.9 TiB 74.30 1.01  98     up         osd.53
> >> >> > >>>> >> >> >>> >  54   hdd   7.32619  1.00000 7.3 TiB 5.6 TiB  5.6
> TiB  1.0
> >> >> > >>>> GiB  16 GiB  1.7 TiB 76.99 1.05 106     up         osd.54
> >> >> > >>>> >> >> >>> >  55   hdd   7.32619  1.00000 7.3 TiB 5.5 TiB  5.5
> TiB  460
> >> >> > >>>> MiB  15 GiB  1.8 TiB 75.46 1.02 105     up         osd.55
> >> >> > >>>> >> >> >>> > -10        51.28331        -  51 TiB  37 TiB   37
> TiB  3.8
> >> >> > >>>> GiB  96 GiB   14 TiB 72.77 0.99   -            host s3db9
> >> >> > >>>> >> >> >>> >  56   hdd   7.32619  1.00000 7.3 TiB 5.2 TiB  5.2
> TiB  846
> >> >> > >>>> MiB  13 GiB  2.1 TiB 71.16 0.97 104     up         osd.56
> >> >> > >>>> >> >> >>> >  57   hdd   7.32619  1.00000 7.3 TiB 5.6 TiB  5.6
> TiB  513
> >> >> > >>>> MiB  15 GiB  1.7 TiB 76.53 1.04  96     up         osd.57
> >> >> > >>>> >> >> >>> >  58   hdd   7.32619  1.00000 7.3 TiB 5.2 TiB  5.2
> TiB  604
> >> >> > >>>> MiB  13 GiB  2.1 TiB 71.23 0.97  98     up         osd.58
> >> >> > >>>> >> >> >>> >  59   hdd   7.32619  1.00000 7.3 TiB 5.1 TiB  5.1
> TiB  414
> >> >> > >>>> MiB  13 GiB  2.2 TiB 70.03 0.95  88     up         osd.59
> >> >> > >>>> >> >> >>> >  60   hdd   7.32619  1.00000 7.3 TiB 5.5 TiB  5.5
> TiB  227
> >> >> > >>>> MiB  14 GiB  1.8 TiB 75.54 1.03  97     up         osd.60
> >> >> > >>>> >> >> >>> >  61   hdd   7.32619  1.00000 7.3 TiB 5.1 TiB  5.1
> TiB  456
> >> >> > >>>> MiB  13 GiB  2.2 TiB 70.01 0.95  95     up         osd.61
> >> >> > >>>> >> >> >>> >  62   hdd   7.32619  1.00000 7.3 TiB 5.5 TiB  5.4
> TiB  843
> >> >> > >>>> MiB  14 GiB  1.8 TiB 74.93 1.02 110     up         osd.62
> >> >> > >>>> >> >> >>> >                        TOTAL 674 TiB 496 TiB  468
> TiB   97
> >> >> > >>>> GiB 1.2 TiB  177 TiB 73.67
> >> >> > >>>> >> >> >>> > MIN/MAX VAR: 0.87/1.14  STDDEV: 4.22
> >> >> > >>>> >> >> >>> >
> >> >> > >>>> >> >> >>> > Am Mo., 15. März 2021 um 15:02 Uhr schrieb Dan
> van der Ster
> >> >> > >>>> <dan@xxxxxxxxxxxxxx>:
> >> >> > >>>> >> >> >>> >>
> >> >> > >>>> >> >> >>> >> OK thanks. Indeed "prepared 0/10 changes" means
> it thinks
> >> >> > >>>> things are balanced.
> >> >> > >>>> >> >> >>> >> Could you again share the full ceph osd df tree?
> >> >> > >>>> >> >> >>> >>
> >> >> > >>>> >> >> >>> >> On Mon, Mar 15, 2021 at 2:54 PM Boris Behrens <
> >> >> > >>>> bb@xxxxxxxxx> wrote:
> >> >> > >>>> >> >> >>> >> >
> >> >> > >>>> >> >> >>> >> > Hi Dan,
> >> >> > >>>> >> >> >>> >> >
> >> >> > >>>> >> >> >>> >> > I've set the autoscaler to warn, but it
> actually does
> >> >> > >>>> not warn for now. So not touching it for now.
> >> >> > >>>> >> >> >>> >> >
> >> >> > >>>> >> >> >>> >> > this is what the log says in minute intervals:
> >> >> > >>>> >> >> >>> >> > 2021-03-15 13:51:00.970 7f307d5fd700  4 mgr
> get_config
> >> >> > >>>> get_config key: mgr/balancer/active
> >> >> > >>>> >> >> >>> >> > 2021-03-15 13:51:00.970 7f307d5fd700  4 mgr
> get_config
> >> >> > >>>> get_config key: mgr/balancer/sleep_interval
> >> >> > >>>> >> >> >>> >> > 2021-03-15 13:51:00.970 7f307d5fd700  4 mgr
> get_config
> >> >> > >>>> get_config key: mgr/balancer/begin_time
> >> >> > >>>> >> >> >>> >> > 2021-03-15 13:51:00.970 7f307d5fd700  4 mgr
> get_config
> >> >> > >>>> get_config key: mgr/balancer/end_time
> >> >> > >>>> >> >> >>> >> > 2021-03-15 13:51:00.970 7f307d5fd700  4 mgr
> get_config
> >> >> > >>>> get_config key: mgr/balancer/begin_weekday
> >> >> > >>>> >> >> >>> >> > 2021-03-15 13:51:00.970 7f307d5fd700  4 mgr
> get_config
> >> >> > >>>> get_config key: mgr/balancer/end_weekday
> >> >> > >>>> >> >> >>> >> > 2021-03-15 13:51:00.971 7f307d5fd700  4 mgr
> get_config
> >> >> > >>>> get_config key: mgr/balancer/pool_ids
> >> >> > >>>> >> >> >>> >> > 2021-03-15 13:51:01.203 7f307d5fd700  4
> mgr[balancer]
> >> >> > >>>> Optimize plan auto_2021-03-15_13:51:00
> >> >> > >>>> >> >> >>> >> > 2021-03-15 13:51:01.203 7f307d5fd700  4 mgr
> get_config
> >> >> > >>>> get_config key: mgr/balancer/mode
> >> >> > >>>> >> >> >>> >> > 2021-03-15 13:51:01.203 7f307d5fd700  4
> mgr[balancer]
> >> >> > >>>> Mode upmap, max misplaced 0.050000
> >> >> > >>>> >> >> >>> >> > 2021-03-15 13:51:01.203 7f307d5fd700  4
> mgr[balancer]
> >> >> > >>>> do_upmap
> >> >> > >>>> >> >> >>> >> > 2021-03-15 13:51:01.203 7f307d5fd700  4 mgr
> get_config
> >> >> > >>>> get_config key: mgr/balancer/upmap_max_iterations
> >> >> > >>>> >> >> >>> >> > 2021-03-15 13:51:01.203 7f307d5fd700  4 mgr
> get_config
> >> >> > >>>> get_config key: mgr/balancer/upmap_max_deviation
> >> >> > >>>> >> >> >>> >> > 2021-03-15 13:51:01.203 7f307d5fd700  4
> mgr[balancer]
> >> >> > >>>> pools ['eu-msg-1.rgw.data.root',
> 'eu-msg-1.rgw.buckets.non-ec',
> >> >> > >>>> 'eu-central-1.rgw.users.keys', 'eu-central-1.rgw.gc',
> >> >> > >>>> 'eu-central-1.rgw.buckets.data',
> 'eu-central-1.rgw.users.email',
> >> >> > >>>> 'eu-msg-1.rgw.gc', 'eu-central-1.rgw.usage',
> 'eu-msg-1.rgw.users.keys',
> >> >> > >>>> 'eu-central-1.rgw.buckets.index', 'rbd', 'eu-msg-1.rgw.log',
> >> >> > >>>> 'whitespace-again-2021-03-10_2', 'eu-msg-1.rgw.buckets.index',
> >> >> > >>>> 'eu-msg-1.rgw.meta', 'eu-central-1.rgw.log', 'default.rgw.gc',
> >> >> > >>>> 'eu-central-1.rgw.buckets.non-ec', 'eu-msg-1.rgw.usage',
> >> >> > >>>> 'whitespace-again-2021-03-10', 'fra-1.rgw.meta',
> >> >> > >>>> 'eu-central-1.rgw.users.uid', 'eu-msg-1.rgw.users.email',
> >> >> > >>>> 'fra-1.rgw.control', 'eu-msg-1.rgw.users.uid',
> 'eu-msg-1.rgw.control',
> >> >> > >>>> '.rgw.root', 'eu-msg-1.rgw.buckets.data',
> 'default.rgw.control',
> >> >> > >>>> 'fra-1.rgw.log', 'default.rgw.data.root',
> 'whitespace-again-2021-03-10_3',
> >> >> > >>>> 'default.rgw.log', 'eu-central-1.rgw.meta',
> 'eu-central-1.rgw.data.root',
> >> >> > >>>> 'default.rgw.users.uid', 'eu-central-1.rgw.control']
> >> >> > >>>> >> >> >>> >> > 2021-03-15 13:51:01.224 7f307d5fd700  4
> mgr[balancer]
> >> >> > >>>> prepared 0/10 changes
> >> >> > >>>> >> >> >>> >> >
> >> >> > >>>> >> >> >>> >> > Am Mo., 15. März 2021 um 14:15 Uhr schrieb Dan
> van der
> >> >> > >>>> Ster <dan@xxxxxxxxxxxxxx>:
> >> >> > >>>> >> >> >>> >> >>
> >> >> > >>>> >> >> >>> >> >> I suggest to just disable the autoscaler
> until your
> >> >> > >>>> balancing is understood.
> >> >> > >>>> >> >> >>> >> >>
> >> >> > >>>> >> >> >>> >> >> What does your active mgr log say (with
> debug_mgr 4/5),
> >> >> > >>>> grep balancer
> >> >> > >>>> >> >> >>> >> >> /var/log/ceph/ceph-mgr.*.log
> >> >> > >>>> >> >> >>> >> >>
> >> >> > >>>> >> >> >>> >> >> -- Dan
> >> >> > >>>> >> >> >>> >> >>
> >> >> > >>>> >> >> >>> >> >> On Mon, Mar 15, 2021 at 1:47 PM Boris Behrens
> <
> >> >> > >>>> bb@xxxxxxxxx> wrote:
> >> >> > >>>> >> >> >>> >> >> >
> >> >> > >>>> >> >> >>> >> >> > Hi,
> >> >> > >>>> >> >> >>> >> >> > this unfortunally did not solve my problem.
> I still
> >> >> > >>>> have some OSDs that fill up to 85%
> >> >> > >>>> >> >> >>> >> >> >
> >> >> > >>>> >> >> >>> >> >> > According to the logging, the autoscaler
> might want
> >> >> > >>>> to add more PGs to one Bucken and reduce almost all other
> buckets to 32.
> >> >> > >>>> >> >> >>> >> >> > 2021-03-15 12:19:58.825 7f307f601700  4
> >> >> > >>>> mgr[pg_autoscaler] Pool 'eu-central-1.rgw.buckets.data'
> root_id -1 using
> >> >> > >>>> 0.705080476146 of space, bias 1.0, pg target 1974.22533321
> quantized to
> >> >> > >>>> 2048 (current 1024)
> >> >> > >>>> >> >> >>> >> >> >
> >> >> > >>>> >> >> >>> >> >> > Why the balancing does not happen is still
> nebulous
> >> >> > >>>> to me.
> >> >> > >>>> >> >> >>> >> >> >
> >> >> > >>>> >> >> >>> >> >> >
> >> >> > >>>> >> >> >>> >> >> >
> >> >> > >>>> >> >> >>> >> >> > Am Sa., 13. März 2021 um 16:37 Uhr schrieb
> Dan van
> >> >> > >>>> der Ster <dan@xxxxxxxxxxxxxx>:
> >> >> > >>>> >> >> >>> >> >> >>
> >> >> > >>>> >> >> >>> >> >> >> OK
> >> >> > >>>> >> >> >>> >> >> >> Btw, you might need to fail to a new
> mgr... I'm not
> >> >> > >>>> sure if the current active will read that new config.
> >> >> > >>>> >> >> >>> >> >> >>
> >> >> > >>>> >> >> >>> >> >> >> .. dan
> >> >> > >>>> >> >> >>> >> >> >>
> >> >> > >>>> >> >> >>> >> >> >>
> >> >> > >>>> >> >> >>> >> >> >> On Sat, Mar 13, 2021, 4:36 PM Boris
> Behrens <
> >> >> > >>>> bb@xxxxxxxxx> wrote:
> >> >> > >>>> >> >> >>> >> >> >>>
> >> >> > >>>> >> >> >>> >> >> >>> Hi,
> >> >> > >>>> >> >> >>> >> >> >>>
> >> >> > >>>> >> >> >>> >> >> >>> ok thanks. I just changed the value and
> rewighted
> >> >> > >>>> everything back to 1. Now I let it sync the weekend and check
> how it will
> >> >> > >>>> be on monday.
> >> >> > >>>> >> >> >>> >> >> >>> We tried to have the systems total
> storage balanced
> >> >> > >>>> as possible. New systems will be with 8TB disks but for the
> exiting ones we
> >> >> > >>>> added 16TB to offset the 4TB disks and we needed a lot of
> storage fast,
> >> >> > >>>> because of a DC move. If you have any recommendations I would
> be happy to
> >> >> > >>>> hear them.
> >> >> > >>>> >> >> >>> >> >> >>>
> >> >> > >>>> >> >> >>> >> >> >>> Cheers
> >> >> > >>>> >> >> >>> >> >> >>>  Boris
> >> >> > >>>> >> >> >>> >> >> >>>
> >> >> > >>>> >> >> >>> >> >> >>> Am Sa., 13. März 2021 um 16:20 Uhr
> schrieb Dan van
> >> >> > >>>> der Ster <dan@xxxxxxxxxxxxxx>:
> >> >> > >>>> >> >> >>> >> >> >>>>
> >> >> > >>>> >> >> >>> >> >> >>>> Thanks.
> >> >> > >>>> >> >> >>> >> >> >>>>
> >> >> > >>>> >> >> >>> >> >> >>>> Decreasing the max deviation to 2 or 1
> should help
> >> >> > >>>> in your case. This option controls when the balancer stops
> trying to move
> >> >> > >>>> PGs around -- by default it stops when the deviation from the
> mean is 5.
> >> >> > >>>> Yes this is too large IMO -- all of our clusters have this
> set to 1.
> >> >> > >>>> >> >> >>> >> >> >>>>
> >> >> > >>>> >> >> >>> >> >> >>>> And given that you have some OSDs with
> more than
> >> >> > >>>> 200 PGs, you definitely shouldn't increase the num PGs.
> >> >> > >>>> >> >> >>> >> >> >>>>
> >> >> > >>>> >> >> >>> >> >> >>>> But anyway with your mixed device sizes
> it might
> >> >> > >>>> be challenging to make a perfectly uniform distribution. Give
> it a try with
> >> >> > >>>> 1 though, and let us know how it goes.
> >> >> > >>>> >> >> >>> >> >> >>>>
> >> >> > >>>> >> >> >>> >> >> >>>> .. Dan
> >> >> > >>>> >> >> >>> >> >> >>>>
> >> >> > >>>> >> >> >>> >> >> >>>>
> >> >> > >>>> >> >> >>> >> >> >>>>
> >> >> > >>>> >> >> >>> >> >> >>>>
> >> >> > >>>> >> >> >>> >> >> >>>>
> >> >> > >>>> >> >> >>> >> >> >>>> On Sat, Mar 13, 2021, 4:11 PM Boris
> Behrens <
> >> >> > >>>> bb@xxxxxxxxx> wrote:
> >> >> > >>>> >> >> >>> >> >> >>>>>
> >> >> > >>>> >> >> >>> >> >> >>>>> Hi Dan,
> >> >> > >>>> >> >> >>> >> >> >>>>>
> >> >> > >>>> >> >> >>> >> >> >>>>> upmap_max_deviation is default (5) in
> our
> >> >> > >>>> cluster. Is 1 the recommended deviation?
> >> >> > >>>> >> >> >>> >> >> >>>>>
> >> >> > >>>> >> >> >>> >> >> >>>>> I added the whole ceph osd df tree, (I
> need to
> >> >> > >>>> remove some OSDs and readd them as bluestore with SSD, so 69,
> 73 and 82 are
> >> >> > >>>> a bit off now. I also reweighted to try to get the %USE
> mitigated).
> >> >> > >>>> >> >> >>> >> >> >>>>>
> >> >> > >>>> >> >> >>> >> >> >>>>> I will increase the mgr debugging to
> see what is
> >> >> > >>>> the problem.
> >> >> > >>>> >> >> >>> >> >> >>>>>
> >> >> > >>>> >> >> >>> >> >> >>>>> [root@s3db1 ~]# ceph osd df tree
> >> >> > >>>> >> >> >>> >> >> >>>>> ID  CLASS WEIGHT    REWEIGHT SIZE
> RAW USE
> >> >> > >>>> DATA    OMAP    META    AVAIL   %USE  VAR  PGS STATUS TYPE
> NAME
> >> >> > >>>> >> >> >>> >> >> >>>>>  -1       673.54224        - 659 TiB
> 491 TiB 464
> >> >> > >>>> TiB  96 GiB 1.2 TiB 168 TiB 74.57 1.00   -        root default
> >> >> > >>>> >> >> >>> >> >> >>>>>  -2        58.30331        -  44 TiB
> 22 TiB  17
> >> >> > >>>> TiB 5.7 GiB  38 GiB  22 TiB 49.82 0.67   -            host
> s3db1
> >> >> > >>>> >> >> >>> >> >> >>>>>  23   hdd  14.65039  1.00000  15 TiB
> 1.8 TiB 1.7
> >> >> > >>>> TiB 156 MiB 4.4 GiB  13 TiB 12.50 0.17 101     up
>  osd.23
> >> >> > >>>> >> >> >>> >> >> >>>>>  69   hdd  14.55269        0     0 B
>  0 B
> >> >> > >>>>  0 B     0 B     0 B     0 B     0    0  11     up
>  osd.69
> >> >> > >>>> >> >> >>> >> >> >>>>>  73   hdd  14.55269  1.00000  15 TiB
> 10 TiB  10
> >> >> > >>>> TiB 6.1 MiB  33 GiB 4.2 TiB 71.15 0.95 107     up
>  osd.73
> >> >> > >>>> >> >> >>> >> >> >>>>>  79   hdd   3.63689  1.00000 3.6 TiB
> 2.9 TiB 747
> >> >> > >>>> GiB 2.0 GiB     0 B 747 GiB 79.94 1.07  52     up
>  osd.79
> >> >> > >>>> >> >> >>> >> >> >>>>>  80   hdd   3.63689  1.00000 3.6 TiB
> 2.6 TiB 1.0
> >> >> > >>>> TiB 1.9 GiB     0 B 1.0 TiB 71.61 0.96  58     up
>  osd.80
> >> >> > >>>> >> >> >>> >> >> >>>>>  81   hdd   3.63689  1.00000 3.6 TiB
> 2.2 TiB 1.5
> >> >> > >>>> TiB 1.1 GiB     0 B 1.5 TiB 60.07 0.81  55     up
>  osd.81
> >> >> > >>>> >> >> >>> >> >> >>>>>  82   hdd   3.63689  1.00000 3.6 TiB
> 1.9 TiB 1.7
> >> >> > >>>> TiB 536 MiB     0 B 1.7 TiB 52.68 0.71  30     up
>  osd.82
> >> >> > >>>> >> >> >>> >> >> >>>>> -11        50.94173        -  51 TiB
> 38 TiB  38
> >> >> > >>>> TiB 3.7 GiB 100 GiB  13 TiB 74.69 1.00   -            host
> s3db10
> >> >> > >>>> >> >> >>> >> >> >>>>>  63   hdd   7.27739  1.00000 7.3 TiB
> 5.5 TiB 5.5
> >> >> > >>>> TiB 616 MiB  14 GiB 1.7 TiB 76.04 1.02  92     up
>  osd.63
> >> >> > >>>> >> >> >>> >> >> >>>>>  64   hdd   7.27739  1.00000 7.3 TiB
> 5.5 TiB 5.5
> >> >> > >>>> TiB 820 MiB  15 GiB 1.8 TiB 75.54 1.01 101     up
>  osd.64
> >> >> > >>>> >> >> >>> >> >> >>>>>  65   hdd   7.27739  1.00000 7.3 TiB
> 5.3 TiB 5.3
> >> >> > >>>> TiB 109 MiB  14 GiB 2.0 TiB 73.17 0.98 105     up
>  osd.65
> >> >> > >>>> >> >> >>> >> >> >>>>>  66   hdd   7.27739  1.00000 7.3 TiB
> 5.8 TiB 5.8
> >> >> > >>>> TiB 423 MiB  15 GiB 1.4 TiB 80.38 1.08  98     up
>  osd.66
> >> >> > >>>> >> >> >>> >> >> >>>>>  67   hdd   7.27739  1.00000 7.3 TiB
> 5.1 TiB 5.1
> >> >> > >>>> TiB 572 MiB  14 GiB 2.2 TiB 70.10 0.94 100     up
>  osd.67
> >> >> > >>>> >> >> >>> >> >> >>>>>  68   hdd   7.27739  1.00000 7.3 TiB
> 5.3 TiB 5.3
> >> >> > >>>> TiB 630 MiB  13 GiB 2.0 TiB 72.88 0.98 107     up
>  osd.68
> >> >> > >>>> >> >> >>> >> >> >>>>>  70   hdd   7.27739  1.00000 7.3 TiB
> 5.4 TiB 5.4
> >> >> > >>>> TiB 648 MiB  14 GiB 1.8 TiB 74.73 1.00 102     up
>  osd.70
> >> >> > >>>> >> >> >>> >> >> >>>>> -12        50.99052        -  51 TiB
> 39 TiB  39
> >> >> > >>>> TiB 2.9 GiB  99 GiB  12 TiB 77.24 1.04   -            host
> s3db11
> >> >> > >>>> >> >> >>> >> >> >>>>>  46   hdd   7.27739  1.00000 7.3 TiB
> 5.7 TiB 5.7
> >> >> > >>>> TiB 102 MiB  15 GiB 1.5 TiB 78.91 1.06  97     up
>  osd.46
> >> >> > >>>> >> >> >>> >> >> >>>>>  47   hdd   7.27739  1.00000 7.3 TiB
> 5.2 TiB 5.2
> >> >> > >>>> TiB  61 MiB  13 GiB 2.1 TiB 71.47 0.96  96     up
>  osd.47
> >> >> > >>>> >> >> >>> >> >> >>>>>  48   hdd   7.27739  1.00000 7.3 TiB
> 6.1 TiB 6.1
> >> >> > >>>> TiB 853 MiB  15 GiB 1.2 TiB 83.46 1.12 109     up
>  osd.48
> >> >> > >>>> >> >> >>> >> >> >>>>>  49   hdd   7.27739  1.00000 7.3 TiB
> 5.7 TiB 5.7
> >> >> > >>>> TiB 708 MiB  15 GiB 1.5 TiB 78.96 1.06  98     up
>  osd.49
> >> >> > >>>> >> >> >>> >> >> >>>>>  50   hdd   7.27739  1.00000 7.3 TiB
> 5.9 TiB 5.8
> >> >> > >>>> TiB 472 MiB  15 GiB 1.4 TiB 80.40 1.08 102     up
>  osd.50
> >> >> > >>>> >> >> >>> >> >> >>>>>  51   hdd   7.27739  1.00000 7.3 TiB
> 5.9 TiB 5.9
> >> >> > >>>> TiB 729 MiB  15 GiB 1.3 TiB 81.70 1.10 110     up
>  osd.51
> >> >> > >>>> >> >> >>> >> >> >>>>>  72   hdd   7.32619  1.00000 7.3 TiB
> 4.8 TiB 4.8
> >> >> > >>>> TiB  91 MiB  12 GiB 2.5 TiB 65.82 0.88  89     up
>  osd.72
> >> >> > >>>> >> >> >>> >> >> >>>>> -37        58.55478        -  59 TiB
> 46 TiB  46
> >> >> > >>>> TiB 5.0 GiB 124 GiB  12 TiB 79.04 1.06   -            host
> s3db12
> >> >> > >>>> >> >> >>> >> >> >>>>>  19   hdd   3.68750  1.00000 3.7 TiB
> 3.1 TiB 3.1
> >> >> > >>>> TiB 462 MiB 8.2 GiB 559 GiB 85.18 1.14  55     up
>  osd.19
> >> >> > >>>> >> >> >>> >> >> >>>>>  71   hdd   3.68750  1.00000 3.7 TiB
> 2.9 TiB 2.8
> >> >> > >>>> TiB 3.9 MiB 7.8 GiB 825 GiB 78.14 1.05  50     up
>  osd.71
> >> >> > >>>> >> >> >>> >> >> >>>>>  75   hdd   3.68750  1.00000 3.7 TiB
> 3.1 TiB 3.1
> >> >> > >>>> TiB 576 MiB 8.3 GiB 555 GiB 85.29 1.14  57     up
>  osd.75
> >> >> > >>>> >> >> >>> >> >> >>>>>  76   hdd   3.68750  1.00000 3.7 TiB
> 3.2 TiB 3.1
> >> >> > >>>> TiB 239 MiB 9.3 GiB 501 GiB 86.73 1.16  50     up
>  osd.76
> >> >> > >>>> >> >> >>> >> >> >>>>>  77   hdd  14.60159  1.00000  15 TiB
> 11 TiB  11
> >> >> > >>>> TiB 880 MiB  30 GiB 3.6 TiB 75.57 1.01 202     up
>  osd.77
> >> >> > >>>> >> >> >>> >> >> >>>>>  78   hdd  14.60159  1.00000  15 TiB
> 11 TiB  11
> >> >> > >>>> TiB 1.0 GiB  30 GiB 3.4 TiB 76.65 1.03 196     up
>  osd.78
> >> >> > >>>> >> >> >>> >> >> >>>>>  83   hdd  14.60159  1.00000  15 TiB
> 12 TiB  12
> >> >> > >>>> TiB 1.8 GiB  31 GiB 2.9 TiB 80.04 1.07 223     up
>  osd.83
> >> >> > >>>> >> >> >>> >> >> >>>>>  -3        58.49872        -  58 TiB
> 43 TiB  38
> >> >> > >>>> TiB 8.1 GiB  91 GiB  16 TiB 73.15 0.98   -            host
> s3db2
> >> >> > >>>> >> >> >>> >> >> >>>>>   1   hdd  14.65039  1.00000  15 TiB
> 11 TiB  11
> >> >> > >>>> TiB 3.1 GiB  38 GiB 3.6 TiB 75.52 1.01 194     up
>  osd.1
> >> >> > >>>> >> >> >>> >> >> >>>>>   3   hdd   3.63689  1.00000 3.6 TiB
> 2.2 TiB 1.4
> >> >> > >>>> TiB 418 MiB     0 B 1.4 TiB 60.94 0.82  52     up
>  osd.3
> >> >> > >>>> >> >> >>> >> >> >>>>>   4   hdd   3.63689  0.89999 3.6 TiB
> 3.2 TiB 401
> >> >> > >>>> GiB 845 MiB     0 B 401 GiB 89.23 1.20  53     up
>  osd.4
> >> >> > >>>> >> >> >>> >> >> >>>>>   5   hdd   3.63689  1.00000 3.6 TiB
> 2.3 TiB 1.3
> >> >> > >>>> TiB 437 MiB     0 B 1.3 TiB 62.88 0.84  51     up
>  osd.5
> >> >> > >>>> >> >> >>> >> >> >>>>>   6   hdd   3.63689  1.00000 3.6 TiB
> 2.0 TiB 1.7
> >> >> > >>>> TiB 1.8 GiB     0 B 1.7 TiB 54.51 0.73  47     up
>  osd.6
> >> >> > >>>> >> >> >>> >> >> >>>>>   7   hdd  14.65039  1.00000  15 TiB
> 11 TiB  11
> >> >> > >>>> TiB 493 MiB  26 GiB 3.8 TiB 73.90 0.99 185     up
>  osd.7
> >> >> > >>>> >> >> >>> >> >> >>>>>  74   hdd  14.65039  1.00000  15 TiB
> 11 TiB  11
> >> >> > >>>> TiB 1.1 GiB  27 GiB 3.5 TiB 76.27 1.02 208     up
>  osd.74
> >> >> > >>>> >> >> >>> >> >> >>>>>  -4        58.49872        -  58 TiB
> 43 TiB  37
> >> >> > >>>> TiB  33 GiB  86 GiB  15 TiB 74.05 0.99   -            host
> s3db3
> >> >> > >>>> >> >> >>> >> >> >>>>>   2   hdd  14.65039  1.00000  15 TiB
> 11 TiB  11
> >> >> > >>>> TiB 850 MiB  26 GiB 4.0 TiB 72.78 0.98 203     up
>  osd.2
> >> >> > >>>> >> >> >>> >> >> >>>>>   9   hdd  14.65039  1.00000  15 TiB
> 11 TiB  11
> >> >> > >>>> TiB 8.3 GiB  33 GiB 3.6 TiB 75.62 1.01 189     up
>  osd.9
> >> >> > >>>> >> >> >>> >> >> >>>>>  10   hdd  14.65039  1.00000  15 TiB
> 11 TiB  11
> >> >> > >>>> TiB 663 MiB  28 GiB 3.5 TiB 76.34 1.02 211     up
>  osd.10
> >> >> > >>>> >> >> >>> >> >> >>>>>  12   hdd   3.63689  1.00000 3.6 TiB
> 2.4 TiB 1.2
> >> >> > >>>> TiB 633 MiB     0 B 1.2 TiB 66.22 0.89  44     up
>  osd.12
> >> >> > >>>> >> >> >>> >> >> >>>>>  13   hdd   3.63689  1.00000 3.6 TiB
> 2.9 TiB 720
> >> >> > >>>> GiB 2.3 GiB     0 B 720 GiB 80.66 1.08  66     up
>  osd.13
> >> >> > >>>> >> >> >>> >> >> >>>>>  14   hdd   3.63689  1.00000 3.6 TiB
> 3.1 TiB 552
> >> >> > >>>> GiB  18 GiB     0 B 552 GiB 85.18 1.14  60     up
>  osd.14
> >> >> > >>>> >> >> >>> >> >> >>>>>  15   hdd   3.63689  1.00000 3.6 TiB
> 2.0 TiB 1.7
> >> >> > >>>> TiB 2.1 GiB     0 B 1.7 TiB 53.72 0.72  44     up
>  osd.15
> >> >> > >>>> >> >> >>> >> >> >>>>>  -5        58.49872        -  58 TiB
> 45 TiB  37
> >> >> > >>>> TiB 7.2 GiB  99 GiB  14 TiB 76.37 1.02   -            host
> s3db4
> >> >> > >>>> >> >> >>> >> >> >>>>>  11   hdd  14.65039  1.00000  15 TiB
> 12 TiB  12
> >> >> > >>>> TiB 897 MiB  28 GiB 2.8 TiB 81.15 1.09 205     up
>  osd.11
> >> >> > >>>> >> >> >>> >> >> >>>>>  17   hdd  14.65039  1.00000  15 TiB
> 11 TiB  11
> >> >> > >>>> TiB 1.2 GiB  27 GiB 3.6 TiB 75.38 1.01 211     up
>  osd.17
> >> >> > >>>> >> >> >>> >> >> >>>>>  18   hdd  14.65039  1.00000  15 TiB
> 11 TiB  11
> >> >> > >>>> TiB 965 MiB  44 GiB 4.0 TiB 72.86 0.98 188     up
>  osd.18
> >> >> > >>>> >> >> >>> >> >> >>>>>  20   hdd   3.63689  1.00000 3.6 TiB
> 2.9 TiB 796
> >> >> > >>>> GiB 529 MiB     0 B 796 GiB 78.63 1.05  66     up
>  osd.20
> >> >> > >>>> >> >> >>> >> >> >>>>>  21   hdd   3.63689  1.00000 3.6 TiB
> 2.6 TiB 1.1
> >> >> > >>>> TiB 2.1 GiB     0 B 1.1 TiB 70.32 0.94  47     up
>  osd.21
> >> >> > >>>> >> >> >>> >> >> >>>>>  22   hdd   3.63689  1.00000 3.6 TiB
> 2.9 TiB 802
> >> >> > >>>> GiB 882 MiB     0 B 802 GiB 78.47 1.05  58     up
>  osd.22
> >> >> > >>>> >> >> >>> >> >> >>>>>  24   hdd   3.63689  1.00000 3.6 TiB
> 2.8 TiB 856
> >> >> > >>>> GiB 645 MiB     0 B 856 GiB 77.01 1.03  47     up
>  osd.24
> >> >> > >>>> >> >> >>> >> >> >>>>>  -6        58.89636        -  59 TiB
> 44 TiB  44
> >> >> > >>>> TiB 2.4 GiB 111 GiB  15 TiB 75.22 1.01   -            host
> s3db5
> >> >> > >>>> >> >> >>> >> >> >>>>>   0   hdd   3.73630  1.00000 3.7 TiB
> 2.4 TiB 2.3
> >> >> > >>>> TiB  70 MiB 6.6 GiB 1.3 TiB 65.00 0.87  48     up
>  osd.0
> >> >> > >>>> >> >> >>> >> >> >>>>>  25   hdd   3.73630  1.00000 3.7 TiB
> 2.4 TiB 2.3
> >> >> > >>>> TiB 5.3 MiB 6.6 GiB 1.4 TiB 63.86 0.86  41     up
>  osd.25
> >> >> > >>>> >> >> >>> >> >> >>>>>  26   hdd   3.73630  1.00000 3.7 TiB
> 2.9 TiB 2.8
> >> >> > >>>> TiB 181 MiB 7.6 GiB 862 GiB 77.47 1.04  48     up
>  osd.26
> >> >> > >>>> >> >> >>> >> >> >>>>>  27   hdd   3.73630  1.00000 3.7 TiB
> 2.3 TiB 2.2
> >> >> > >>>> TiB 7.0 MiB 6.1 GiB 1.5 TiB 61.00 0.82  48     up
>  osd.27
> >> >> > >>>> >> >> >>> >> >> >>>>>  28   hdd  14.65039  1.00000  15 TiB
> 12 TiB  12
> >> >> > >>>> TiB 937 MiB  30 GiB 2.8 TiB 81.19 1.09 203     up
>  osd.28
> >> >> > >>>> >> >> >>> >> >> >>>>>  29   hdd  14.65039  1.00000  15 TiB
> 11 TiB  11
> >> >> > >>>> TiB 536 MiB  26 GiB 3.8 TiB 73.95 0.99 200     up
>  osd.29
> >> >> > >>>> >> >> >>> >> >> >>>>>  30   hdd  14.65039  1.00000  15 TiB
> 12 TiB  11
> >> >> > >>>> TiB 744 MiB  28 GiB 3.1 TiB 79.07 1.06 207     up
>  osd.30
> >> >> > >>>> >> >> >>> >> >> >>>>>  -7        58.89636        -  59 TiB
> 44 TiB  44
> >> >> > >>>> TiB  14 GiB 122 GiB  14 TiB 75.41 1.01   -            host
> s3db6
> >> >> > >>>> >> >> >>> >> >> >>>>>  32   hdd   3.73630  1.00000 3.7 TiB
> 3.1 TiB 3.0
> >> >> > >>>> TiB  16 MiB 8.2 GiB 622 GiB 83.74 1.12  65     up
>  osd.32
> >> >> > >>>> >> >> >>> >> >> >>>>>  33   hdd   3.73630  0.79999 3.7 TiB
> 3.0 TiB 2.9
> >> >> > >>>> TiB  14 MiB 8.1 GiB 740 GiB 80.67 1.08  52     up
>  osd.33
> >> >> > >>>> >> >> >>> >> >> >>>>>  34   hdd   3.73630  0.79999 3.7 TiB
> 2.9 TiB 2.8
> >> >> > >>>> TiB 449 MiB 7.7 GiB 877 GiB 77.08 1.03  52     up
>  osd.34
> >> >> > >>>> >> >> >>> >> >> >>>>>  35   hdd   3.73630  0.79999 3.7 TiB
> 2.3 TiB 2.2
> >> >> > >>>> TiB 133 MiB 7.0 GiB 1.4 TiB 62.18 0.83  42     up
>  osd.35
> >> >> > >>>> >> >> >>> >> >> >>>>>  36   hdd  14.65039  1.00000  15 TiB
> 11 TiB  11
> >> >> > >>>> TiB 544 MiB  26 GiB 4.0 TiB 72.98 0.98 220     up
>  osd.36
> >> >> > >>>> >> >> >>> >> >> >>>>>  37   hdd  14.65039  1.00000  15 TiB
> 11 TiB  11
> >> >> > >>>> TiB  11 GiB  38 GiB 3.6 TiB 75.30 1.01 200     up
>  osd.37
> >> >> > >>>> >> >> >>> >> >> >>>>>  38   hdd  14.65039  1.00000  15 TiB
> 11 TiB  11
> >> >> > >>>> TiB 1.2 GiB  28 GiB 3.3 TiB 77.43 1.04 217     up
>  osd.38
> >> >> > >>>> >> >> >>> >> >> >>>>>  -8        58.89636        -  59 TiB
> 47 TiB  46
> >> >> > >>>> TiB 3.9 GiB 116 GiB  12 TiB 78.98 1.06   -            host
> s3db7
> >> >> > >>>> >> >> >>> >> >> >>>>>  39   hdd   3.73630  1.00000 3.7 TiB
> 3.2 TiB 3.2
> >> >> > >>>> TiB  19 MiB 8.5 GiB 499 GiB 86.96 1.17  43     up
>  osd.39
> >> >> > >>>> >> >> >>> >> >> >>>>>  40   hdd   3.73630  1.00000 3.7 TiB
> 2.6 TiB 2.5
> >> >> > >>>> TiB 144 MiB 7.0 GiB 1.2 TiB 68.33 0.92  39     up
>  osd.40
> >> >> > >>>> >> >> >>> >> >> >>>>>  41   hdd   3.73630  1.00000 3.7 TiB
> 3.0 TiB 2.9
> >> >> > >>>> TiB 218 MiB 7.9 GiB 732 GiB 80.86 1.08  64     up
>  osd.41
> >> >> > >>>> >> >> >>> >> >> >>>>>  42   hdd   3.73630  1.00000 3.7 TiB
> 2.5 TiB 2.4
> >> >> > >>>> TiB 594 MiB 7.0 GiB 1.2 TiB 67.97 0.91  50     up
>  osd.42
> >> >> > >>>> >> >> >>> >> >> >>>>>  43   hdd  14.65039  1.00000  15 TiB
> 12 TiB  12
> >> >> > >>>> TiB 564 MiB  28 GiB 2.9 TiB 80.32 1.08 213     up
>  osd.43
> >> >> > >>>> >> >> >>> >> >> >>>>>  44   hdd  14.65039  1.00000  15 TiB
> 12 TiB  11
> >> >> > >>>> TiB 1.3 GiB  28 GiB 3.1 TiB 78.59 1.05 198     up
>  osd.44
> >> >> > >>>> >> >> >>> >> >> >>>>>  45   hdd  14.65039  1.00000  15 TiB
> 12 TiB  12
> >> >> > >>>> TiB 1.2 GiB  30 GiB 2.8 TiB 81.05 1.09 214     up
>  osd.45
> >> >> > >>>> >> >> >>> >> >> >>>>>  -9        51.28331        -  51 TiB
> 41 TiB  41
> >> >> > >>>> TiB 4.9 GiB 108 GiB  10 TiB 79.75 1.07   -            host
> s3db8
> >> >> > >>>> >> >> >>> >> >> >>>>>   8   hdd   7.32619  1.00000 7.3 TiB
> 5.8 TiB 5.8
> >> >> > >>>> TiB 472 MiB  15 GiB 1.5 TiB 79.68 1.07  99     up
>  osd.8
> >> >> > >>>> >> >> >>> >> >> >>>>>  16   hdd   7.32619  1.00000 7.3 TiB
> 5.9 TiB 5.8
> >> >> > >>>> TiB 785 MiB  15 GiB 1.4 TiB 80.25 1.08  97     up
>  osd.16
> >> >> > >>>> >> >> >>> >> >> >>>>>  31   hdd   7.32619  1.00000 7.3 TiB
> 5.5 TiB 5.5
> >> >> > >>>> TiB 438 MiB  14 GiB 1.8 TiB 75.36 1.01  87     up
>  osd.31
> >> >> > >>>> >> >> >>> >> >> >>>>>  52   hdd   7.32619  1.00000 7.3 TiB
> 5.7 TiB 5.7
> >> >> > >>>> TiB 844 MiB  15 GiB 1.6 TiB 78.19 1.05 113     up
>  osd.52
> >> >> > >>>> >> >> >>> >> >> >>>>>  53   hdd   7.32619  1.00000 7.3 TiB
> 6.2 TiB 6.1
> >> >> > >>>> TiB 792 MiB  18 GiB 1.1 TiB 84.46 1.13 109     up
>  osd.53
> >> >> > >>>> >> >> >>> >> >> >>>>>  54   hdd   7.32619  1.00000 7.3 TiB
> 5.6 TiB 5.6
> >> >> > >>>> TiB 959 MiB  15 GiB 1.7 TiB 76.73 1.03 115     up
>  osd.54
> >> >> > >>>> >> >> >>> >> >> >>>>>  55   hdd   7.32619  1.00000 7.3 TiB
> 6.1 TiB 6.1
> >> >> > >>>> TiB 699 MiB  16 GiB 1.2 TiB 83.56 1.12 122     up
>  osd.55
> >> >> > >>>> >> >> >>> >> >> >>>>> -10        51.28331        -  51 TiB
> 39 TiB  39
> >> >> > >>>> TiB 4.7 GiB 100 GiB  12 TiB 76.05 1.02   -            host
> s3db9
> >> >> > >>>> >> >> >>> >> >> >>>>>  56   hdd   7.32619  1.00000 7.3 TiB
> 5.2 TiB 5.2
> >> >> > >>>> TiB 840 MiB  13 GiB 2.1 TiB 71.06 0.95 105     up
>  osd.56
> >> >> > >>>> >> >> >>> >> >> >>>>>  57   hdd   7.32619  1.00000 7.3 TiB
> 6.1 TiB 6.0
> >> >> > >>>> TiB 1.0 GiB  16 GiB 1.2 TiB 83.17 1.12 102     up
>  osd.57
> >> >> > >>>> >> >> >>> >> >> >>>>>  58   hdd   7.32619  1.00000 7.3 TiB
> 6.0 TiB 5.9
> >> >> > >>>> TiB  43 MiB  15 GiB 1.4 TiB 81.56 1.09 105     up
>  osd.58
> >> >> > >>>> >> >> >>> >> >> >>>>>  59   hdd   7.32619  1.00000 7.3 TiB
> 5.9 TiB 5.9
> >> >> > >>>> TiB 429 MiB  15 GiB 1.4 TiB 80.64 1.08  94     up
>  osd.59
> >> >> > >>>> >> >> >>> >> >> >>>>>  60   hdd   7.32619  1.00000 7.3 TiB
> 5.4 TiB 5.3
> >> >> > >>>> TiB 226 MiB  14 GiB 2.0 TiB 73.25 0.98 101     up
>  osd.60
> >> >> > >>>> >> >> >>> >> >> >>>>>  61   hdd   7.32619  1.00000 7.3 TiB
> 4.8 TiB 4.8
> >> >> > >>>> TiB 1.1 GiB  12 GiB 2.5 TiB 65.84 0.88 103     up
>  osd.61
> >> >> > >>>> >> >> >>> >> >> >>>>>  62   hdd   7.32619  1.00000 7.3 TiB
> 5.6 TiB 5.6
> >> >> > >>>> TiB 1.0 GiB  15 GiB 1.7 TiB 76.83 1.03 126     up
>  osd.62
> >> >> > >>>> >> >> >>> >> >> >>>>>                        TOTAL 674 TiB
> 501 TiB 473
> >> >> > >>>> TiB  96 GiB 1.2 TiB 173 TiB 74.57
> >> >> > >>>> >> >> >>> >> >> >>>>> MIN/MAX VAR: 0.17/1.20  STDDEV: 10.25
> >> >> > >>>> >> >> >>> >> >> >>>>>
> >> >> > >>>> >> >> >>> >> >> >>>>>
> >> >> > >>>> >> >> >>> >> >> >>>>>
> >> >> > >>>> >> >> >>> >> >> >>>>> Am Sa., 13. März 2021 um 15:57 Uhr
> schrieb Dan
> >> >> > >>>> van der Ster <dan@xxxxxxxxxxxxxx>:
> >> >> > >>>> >> >> >>> >> >> >>>>>>
> >> >> > >>>> >> >> >>> >> >> >>>>>> No, increasing num PGs won't help
> substantially.
> >> >> > >>>> >> >> >>> >> >> >>>>>>
> >> >> > >>>> >> >> >>> >> >> >>>>>> Can you share the entire output of
> ceph osd df
> >> >> > >>>> tree ?
> >> >> > >>>> >> >> >>> >> >> >>>>>>
> >> >> > >>>> >> >> >>> >> >> >>>>>> Did you already set
> >> >> > >>>> >> >> >>> >> >> >>>>>>
> >> >> > >>>> >> >> >>> >> >> >>>>>>   ceph config set mgr
> >> >> > >>>> mgr/balancer/upmap_max_deviation 1
> >> >> > >>>> >> >> >>> >> >> >>>>>>
> >> >> > >>>> >> >> >>> >> >> >>>>>>
> >> >> > >>>> >> >> >>> >> >> >>>>>> ??
> >> >> > >>>> >> >> >>> >> >> >>>>>> And I recommend debug_mgr 4/5 so you
> can see
> >> >> > >>>> some basic upmap balancer logging.
> >> >> > >>>> >> >> >>> >> >> >>>>>>
> >> >> > >>>> >> >> >>> >> >> >>>>>> .. Dan
> >> >> > >>>> >> >> >>> >> >> >>>>>>
> >> >> > >>>> >> >> >>> >> >> >>>>>>
> >> >> > >>>> >> >> >>> >> >> >>>>>>
> >> >> > >>>> >> >> >>> >> >> >>>>>>
> >> >> > >>>> >> >> >>> >> >> >>>>>>
> >> >> > >>>> >> >> >>> >> >> >>>>>>
> >> >> > >>>> >> >> >>> >> >> >>>>>> On Sat, Mar 13, 2021, 3:49 PM Boris
> Behrens <
> >> >> > >>>> bb@xxxxxxxxx> wrote:
> >> >> > >>>> >> >> >>> >> >> >>>>>>>
> >> >> > >>>> >> >> >>> >> >> >>>>>>> Hello people,
> >> >> > >>>> >> >> >>> >> >> >>>>>>>
> >> >> > >>>> >> >> >>> >> >> >>>>>>> I am still struggeling with the
> balancer
> >> >> > >>>> >> >> >>> >> >> >>>>>>> (
> >> >> > >>>> https://www.mail-archive.com/ceph-users@xxxxxxx/msg09124.html
> )
> >> >> > >>>> >> >> >>> >> >> >>>>>>> Now I've read some more and might
> think that I
> >> >> > >>>> do not have enough PGs.
> >> >> > >>>> >> >> >>> >> >> >>>>>>> Currently I have 84OSDs and 1024PGs
> for the
> >> >> > >>>> main pool (3008 total). I
> >> >> > >>>> >> >> >>> >> >> >>>>>>> have the autoscaler enabled, but I
> doesn't tell
> >> >> > >>>> me to increase the
> >> >> > >>>> >> >> >>> >> >> >>>>>>> PGs.
> >> >> > >>>> >> >> >>> >> >> >>>>>>>
> >> >> > >>>> >> >> >>> >> >> >>>>>>> What do you think?
> >> >> > >>>> >> >> >>> >> >> >>>>>>>
> >> >> > >>>> >> >> >>> >> >> >>>>>>> --
> >> >> > >>>> >> >> >>> >> >> >>>>>>> Die Selbsthilfegruppe
> "UTF-8-Probleme" trifft
> >> >> > >>>> sich diesmal abweichend
> >> >> > >>>> >> >> >>> >> >> >>>>>>> im groüen Saal.
> >> >> > >>>> >> >> >>> >> >> >>>>>>>
> _______________________________________________
> >> >> > >>>> >> >> >>> >> >> >>>>>>> ceph-users mailing list --
> ceph-users@xxxxxxx
> >> >> > >>>> >> >> >>> >> >> >>>>>>> To unsubscribe send an email to
> >> >> > >>>> ceph-users-leave@xxxxxxx
> >> >> > >>>> >> >> >>> >> >> >>>>>
> >> >> > >>>> >> >> >>> >> >> >>>>>
> >> >> > >>>> >> >> >>> >> >> >>>>>
> >> >> > >>>> >> >> >>> >> >> >>>>> --
> >> >> > >>>> >> >> >>> >> >> >>>>> Die Selbsthilfegruppe "UTF-8-Probleme"
> trifft
> >> >> > >>>> sich diesmal abweichend im groüen Saal.
> >> >> > >>>> >> >> >>> >> >> >>>
> >> >> > >>>> >> >> >>> >> >> >>>
> >> >> > >>>> >> >> >>> >> >> >>>
> >> >> > >>>> >> >> >>> >> >> >>> --
> >> >> > >>>> >> >> >>> >> >> >>> Die Selbsthilfegruppe "UTF-8-Probleme"
> trifft sich
> >> >> > >>>> diesmal abweichend im groüen Saal.
> >> >> > >>>> >> >> >>> >> >> >
> >> >> > >>>> >> >> >>> >> >> >
> >> >> > >>>> >> >> >>> >> >> >
> >> >> > >>>> >> >> >>> >> >> > --
> >> >> > >>>> >> >> >>> >> >> > Die Selbsthilfegruppe "UTF-8-Probleme"
> trifft sich
> >> >> > >>>> diesmal abweichend im groüen Saal.
> >> >> > >>>> >> >> >>> >> >
> >> >> > >>>> >> >> >>> >> >
> >> >> > >>>> >> >> >>> >> >
> >> >> > >>>> >> >> >>> >> > --
> >> >> > >>>> >> >> >>> >> > Die Selbsthilfegruppe "UTF-8-Probleme" trifft
> sich
> >> >> > >>>> diesmal abweichend im groüen Saal.
> >> >> > >>>> >> >> >>> >
> >> >> > >>>> >> >> >>> >
> >> >> > >>>> >> >> >>> >
> >> >> > >>>> >> >> >>> > --
> >> >> > >>>> >> >> >>> > Die Selbsthilfegruppe "UTF-8-Probleme" trifft
> sich diesmal
> >> >> > >>>> abweichend im groüen Saal.
> >> >> > >>>> >> >> >>
> >> >> > >>>> >> >> >>
> >> >> > >>>> >> >> >>
> >> >> > >>>> >> >> >> --
> >> >> > >>>> >> >> >> Die Selbsthilfegruppe "UTF-8-Probleme" trifft sich
> diesmal
> >> >> > >>>> abweichend im groüen Saal.
> >> >> > >>>> >> >> >
> >> >> > >>>> >> >> >
> >> >> > >>>> >> >> >
> >> >> > >>>> >> >> > --
> >> >> > >>>> >> >> > Die Selbsthilfegruppe "UTF-8-Probleme" trifft sich
> diesmal
> >> >> > >>>> abweichend im groüen Saal.
> >> >> > >>>> >> >
> >> >> > >>>> >> >
> >> >> > >>>> >> >
> >> >> > >>>> >> > --
> >> >> > >>>> >> > Die Selbsthilfegruppe "UTF-8-Probleme" trifft sich
> diesmal
> >> >> > >>>> abweichend im groüen Saal.
> >> >> > >>>> >
> >> >> > >>>> >
> >> >> > >>>> >
> >> >> > >>>> > --
> >> >> > >>>> > Die Selbsthilfegruppe "UTF-8-Probleme" trifft sich diesmal
> abweichend
> >> >> > >>>> im groüen Saal.
> >> >> > >>>>
> >> >> > >>>
> >> >> > >>>
> >> >> > >>> --
> >> >> > >>> Die Selbsthilfegruppe "UTF-8-Probleme" trifft sich diesmal
> abweichend im
> >> >> > >>> groüen Saal.
> >> >> > >>>
> >> >> > >>
> >> >> > >
> >> >> > > --
> >> >> > > Die Selbsthilfegruppe "UTF-8-Probleme" trifft sich diesmal
> abweichend im
> >> >> > > groüen Saal.
> >> >> > >
> >> >> >
> >> >> >
> >> >> > --
> >> >> > Die Selbsthilfegruppe "UTF-8-Probleme" trifft sich diesmal
> abweichend im
> >> >> > groüen Saal.
> >> >> > _______________________________________________
> >> >> > ceph-users mailing list -- ceph-users@xxxxxxx
> >> >> > To unsubscribe send an email to ceph-users-leave@xxxxxxx
> >> >
> >> >
> >> >
> >> > --
> >> > Die Selbsthilfegruppe "UTF-8-Probleme" trifft sich diesmal abweichend
> im groüen Saal.
> >
> >
> >
> > --
> > Die Selbsthilfegruppe "UTF-8-Probleme" trifft sich diesmal abweichend im
> groüen Saal.
>


-- 
Die Selbsthilfegruppe "UTF-8-Probleme" trifft sich diesmal abweichend im
groüen Saal.
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx





[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux