Re: Balancing PGs across OSDs

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hello Paul,

thanks for your analysis.

I want to share more statistics of my cluster to follow-up on your
response "You have way too few PGs in one of the roots".

Here are the pool details:
root@ld3955:~# ceph osd pool ls detail
pool 11 'hdb_backup' replicated size 3 min_size 2 crush_rule 1
object_hash rjenkins pg_num 8192 pgp_num 8192 autoscale_mode warn
last_change 294572 flags hashpspool,selfmanaged_snaps stripe_width 0
application rbd
        removed_snaps [1~3]
pool 59 'hdd' replicated size 2 min_size 2 crush_rule 3 object_hash
rjenkins pg_num 64 pgp_num 64 autoscale_mode warn last_change 267271
flags hashpspool,selfmanaged_snaps stripe_width 0 application rbd
        removed_snaps [1~3]
pool 60 'ssd' replicated size 2 min_size 2 crush_rule 4 object_hash
rjenkins pg_num 128 pgp_num 128 autoscale_mode warn last_change 299719
lfor 299717/299717/299717 flags hashpspool,selfmanaged_snaps
stripe_width 0 application rbd
        removed_snaps [1~3]
pool 61 'nvme' replicated size 2 min_size 2 crush_rule 2 object_hash
rjenkins pg_num 8 pgp_num 8 autoscale_mode warn last_change 267125 flags
hashpspool stripe_width 0 application rbd
pool 62 'cephfs_data' replicated size 3 min_size 2 crush_rule 3
object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode warn
last_change 300312 lfor 300310/300310/300310 flags hashpspool
stripe_width 0 application cephfs
pool 63 'cephfs_metadata' replicated size 3 min_size 2 crush_rule 3
object_hash rjenkins pg_num 8 pgp_num 8 autoscale_mode warn last_change
267069 flags hashpspool stripe_width 0 pg_autoscale_bias 4 pg_num_min 16
recovery_priority 5 application cephfs

Any pg_num / pgp_num is monitored by Ceph, means I get a warning in the
log / health status if a pool is undersized.
I didn't enable PG auto-scaler for any pool, though.
The calculation of PGs per Pool is done with pgcalc
<https://ceph.io/pgcalc/>.
Here's a screenshot <https://ibb.co/VjR6X3x> of this calculation.

My focus is on pool hdb_backup.
Based on these statistics
root@ld3955:~# ceph df detail
RAW STORAGE:
    CLASS     SIZE        AVAIL       USED        RAW USED     %RAW USED
    hdd       1.4 PiB     744 TiB     729 TiB      730 TiB         49.53
    nvme       23 TiB      23 TiB      43 GiB       51 GiB          0.22
    ssd        27 TiB      25 TiB     1.9 TiB      1.9 TiB          7.15
    TOTAL     1.5 PiB     792 TiB     731 TiB      732 TiB         48.02

POOLS:
    POOL                ID     STORED      OBJECTS     USED       
%USED     MAX AVAIL     QUOTA OBJECTS     QUOTA BYTES     DIRTY      
USED COMPR     UNDER COMPR
    hdb_backup          11     241 TiB      63.29M     241 TiB    
57.03        61 TiB     N/A               N/A              63.29M
0 B             0 B
    hdd                 59     553 GiB     142.16k     553 GiB     
0.50        54 TiB     N/A               N/A             142.16k
0 B             0 B
    ssd                 60     2.0 TiB     530.75k     2.0 TiB     
8.72        10 TiB     N/A               N/A             530.75k
0 B             0 B
    nvme                61         0 B           0         0 B        
0        11 TiB     N/A               N/A                   0
0 B             0 B
    cephfs_data         62     356 GiB     102.29k     356 GiB     
0.32        36 TiB     N/A               N/A             102.29k
0 B             0 B
    cephfs_metadata     63     117 MiB          52     117 MiB        
0        36 TiB     N/A               N/A                  52
0 B             0 B

there's only 57% used, but effectively I cannot store much more data
because some OSDs are filling up by +80%.

It is true that the disks that are used for this pool exclusively are
different in size, means
3x 48 disks à 7.2TB
4x 48 disks à 1.6TB
and the disk usage is for 7.2TB disk from 41% to 54% and for 1.6TB disks
from 52% to 81%.

If Ceph is not cabable to manage rebalancing automatically, how can I
proceed to rebalance the data manually?
OSD reweight is not an option in my opinion because it starts filling
OSDs that are not with lowest usage rate.
Can I move PGs to specific OSDs?


THX




Am 18.11.2019 um 20:18 schrieb Paul Emmerich:
> You have way too few PGs in one of the roots. Many OSDs have so few
> PGs that you should see a lot of health warnings because of it.
> The other root has a factor 5 difference in disk size which isn't ideal either.
>
>
> Paul
>

_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux