To expand on the comments below, "max avail" takes into account usage imbalance between OSDs. There's a pretty significant imbalance in this cluster and Ceph assumes that the imbalance will continue, and thus indicates that there's not much room left in the pool. Rebalancing that pool will make a big difference in terms of top-OSD fullness and the "max avail" metric. Josh On Mon, Nov 1, 2021 at 12:25 PM Alexander Closs <acloss@xxxxxxxxxxxxx> wrote: > > Max available = free space actually usable now based on OSD usage, not including already-used space. > > -Alex > MIT CSAIL > > On 11/1/21, 2:18 PM, "Szabo, Istvan (Agoda)" <Istvan.Szabo@xxxxxxxxx> wrote: > > It says max available: 115TB and current use is 104TB, what I don’t understand where the max available come from because on the pool no object and no size limit is set: > > quotas for pool 'sin.rgw.buckets.data': > max objects: N/A > max bytes : N/A > > Istvan Szabo > Senior Infrastructure Engineer > --------------------------------------------------- > Agoda Services Co., Ltd. > e: istvan.szabo@xxxxxxxxx<mailto:istvan.szabo@xxxxxxxxx> > --------------------------------------------------- > > On 2021. Nov 1., at 18:48, Etienne Menguy <etienne.menguy@xxxxxxxx> wrote: > > sin.rgw.buckets.data 24 128 104 TiB 104 TiB 0 B 1.30G 156 TiB 156 TiB 0 B 47.51 115 TiB N/A N/A 1.30G 0 B 0 B > _______________________________________________ > ceph-users mailing list -- ceph-users@xxxxxxx > To unsubscribe send an email to ceph-users-leave@xxxxxxx > > > _______________________________________________ > ceph-users mailing list -- ceph-users@xxxxxxx > To unsubscribe send an email to ceph-users-leave@xxxxxxx _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx