Can I get rid of PGs after trying to decrease the number on the pool again? Doing a backup and nuking the cluster seems a little too much work for me :) $ sudo ceph osd pool get cephfs_data pg_num pg_num: 128 $ sudo ceph osd pool set cephfs_data pg_num 16 $ sudo ceph osd pool get cephfs_data pg_num pg_num: 128 Am Fr., 2. Dez. 2022 um 10:22 Uhr schrieb Janne Johansson < icepic.dz@xxxxxxxxx>: > > my OSDs are running odroid-hc4's and they only have about 4GB of memory, > > and every 10 minutes a random OSD crashes due to out of memory. Sadly the > > whole machine gets unresponsive when the memory gets completely full, so > no > > ssh access or prometheus output in the meantime. > > > I've set the memory limit very low on all OSDs: > > > > for i in {0..17} ; do sudo ceph config set osd.$i osd_memory_target > > 939524096 ; done which is the absolute minimum, about 0.9GB. > > > > Why are the OSDs not respecting this limit? > > The memory limit you set with osd_memory_target is about the parts > that _can_ scale up and down memory usage, like read caches and so > forth, but it is not all of the needed RAM to run an OSD with many > PGs/Objects. If the box is too small, it is too small. > > -- > May the most significant bit of your life be positive. > _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx