Re: Advice on enabling autoscaler

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Robert,


Am 07.02.22 um 13:15 schrieb Maarten van Ingen:
> As it's just a few pools affected, doing a manual increase would be and option for me as well, if recommended.
>
> As you can see one pool is basically lacking pg's while the others are mostly increasing due to the much higher target_bytes compared to the current usage.

Were the target sizes set by you based on your expectations? They seem
to differ quite a bit from the current usage.
>> They are set this way indeed.

And keep in mind that when PGs are increased that you also may need to
increase the number of OSDs as one OSD should carry a max of around 200
PGs. But I do not know if that is still the case with current Ceph versions.

>> We are currently on 32 PG's per OSD on average, if we apply what the autoscaler suggest, we should be att around 60. Thats still way lower than the 200 and even to the suggested 100. Not sure if the 100 on average still holds these days.


Regards
--
Robert Sander
Heinlein Consulting GmbH
Schwedter Str. 8/9b, 10119 Berlin

http://www.heinlein-support.de

Tel: 030 / 405051-43
Fax: 030 / 405051-19

Zwangsangaben lt. §35a GmbHG:
HRB 220009 B / Amtsgericht Berlin-Charlottenburg,
Geschäftsführer: Peer Heinlein -- Sitz: Berlin
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux