I still think they are not part of the cluster somehow. "ceph osd status" shows most likely they are not used. When you add just 1 osd you should see something in your cluster capacity and some rebalancing. Status of ceph is HEALTH_OK? > Thanks for the prompt reply. > > Yes, it does. All of them are up, with the correct class that is used by > the crush algorithm. > > On Thu, Feb 13, 2025 at 7:47 AM Marc <Marc@xxxxxxxxxxxxxxxxx > <mailto:Marc@xxxxxxxxxxxxxxxxx> > wrote: > > > > Hello guys, > > Let's say I have a cluster with 4 nodes with 24 SSDs each, and a > single > > pool that consumes all OSDs of all nodes. After adding another > host, I > > noticed that no extra space was added. Can this be a result of > the > > number > > of PGs I am using? > > > > I mean, when adding more hosts/OSDs, should I always consider > increasing > > the number of PGs from a pool? > > > > ceph osd tree > > shows all up and with correct weight? > _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx