Re: Balancer=on with crush-compat mode

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



If I understand the balancer correct, it balances PGs not data.
This worked perfectly fine in your case.

I prefer a PG count of ~100 per OSD, you are at 30. Maybe it would
help to bump the PGs.

Kevin

Am Sa., 5. Jan. 2019 um 14:39 Uhr schrieb Marc Roos <M.Roos@xxxxxxxxxxxxxxxxx>:
>
>
> I have straw2, balancer=on, crush-compat and it gives worst spread over
> my ssd drives (4 only) being used by only 2 pools. One of these pools
> has pg 8. Should I increase this to 16 to create a better result, or
> will it never be any better.
>
> For now I like to stick to crush-compat, so I can use a default centos7
> kernel.
>
> Luminous 12.2.8, 3.10.0-862.14.4.el7.x86_64, CentOS Linux release
> 7.5.1804 (Core)
>
>
>
> [@c01 ~]# cat balancer-1-before.txt | egrep '^19|^20|^21|^30'
> 19   ssd 0.48000  1.00000  447GiB  164GiB  283GiB 36.79 0.93  31
> 20   ssd 0.48000  1.00000  447GiB  136GiB  311GiB 30.49 0.77  32
> 21   ssd 0.48000  1.00000  447GiB  215GiB  232GiB 48.02 1.22  30
> 30   ssd 0.48000  1.00000  447GiB  151GiB  296GiB 33.72 0.86  27
>
> [@c01 ~]# ceph osd df | egrep '^19|^20|^21|^30'
> 19   ssd 0.48000  1.00000  447GiB  157GiB  290GiB 35.18 0.87  30
> 20   ssd 0.48000  1.00000  447GiB  125GiB  322GiB 28.00 0.69  30
> 21   ssd 0.48000  1.00000  447GiB  245GiB  202GiB 54.71 1.35  30
> 30   ssd 0.48000  1.00000  447GiB  217GiB  230GiB 48.46 1.20  30
>
> [@c01 ~]# ceph osd pool ls detail | egrep 'fs_meta|rbd.ssd'
> pool 19 'fs_meta' replicated size 3 min_size 2 crush_rule 5 object_hash
> rjenkins pg_num 16 pgp_num 16 last_change 22425 lfor 0/9035 flags
> hashpspool stripe_width 0 application cephfs
> pool 54 'rbd.ssd' replicated size 3 min_size 2 crush_rule 5 object_hash
> rjenkins pg_num 8 pgp_num 8 last_change 24666 flags hashpspool
> stripe_width 0 application rbd
>
> [@c01 ~]# ceph df |egrep 'ssd|fs_meta'
>     fs_meta                           19      170MiB      0.07
> 240GiB     2451382
>     fs_data.ssd                       33          0B         0
> 240GiB           0
>     rbd.ssd                           54      266GiB     52.57
> 240GiB       75902
>     fs_data.ec21.ssd                  55          0B         0
> 480GiB           0
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@xxxxxxxxxxxxxx
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux