On Mon, Feb 13, 2023 at 4:16 AM Sake Paulusma <sake1989@xxxxxxxxxxx> wrote: > > Hello, > > I configured a stretched cluster on two datacenters. It's working fine, except this weekend the Raw Capicity exceeded 50% and the error POOL_TARGET_SIZE_BYTES_OVERCOMMITED showed up. > > The command "ceph df" is showing the correct cluster size, but "ceph osd pool autoscale-status" is showing half of the total Raw Capacity. > > What could be wrong? There's a bug with the statistics handling of pools in stretch mode, and others like them. :( https://tracker.ceph.com/issues/56650 -Greg > > > > ---------------------------------------------------------------------------- > [ceph: root@aqsoel11445 /]# ceph status > cluster: > id: adbe7bb6-5h6d-11ed-8511-004449ede0c > health: HEALTH_WARN > 1 MDSs report oversized cache > 1 subtrees have overcommitted pool target_size_bytes > > services: > mon: 5 daemons, quorum host1,host2,host3,host4,host5 (age 4w) > mgr: aqsoel11445.nqamuz(active, since 5w), standbys: host1.wujgas > mds: 2/2 daemons up, 2 standby > osd: 12 osds: 12 up (since 5w), 12 in (since 9w) > > data: > volumes: 2/2 healthy > pools: 5 pools, 193 pgs > objects: 17.31M objects, 1.2 TiB > usage: 5.0 TiB used, 3.8 TiB / 8.8 TiB avail > pgs: 192 active+clean > 1 active+clean+scrubbing > ---------------------------------------------------------------------------- > > ---------------------------------------------------------------------------- > [ceph: root@aqsoel11445 /]# ceph df > --- RAW STORAGE --- > CLASS SIZE AVAIL USED RAW USED %RAW USED > ssd 8.8 TiB 3.8 TiB 5.0 TiB 5.0 TiB 56.83 > TOTAL 8.8 TiB 3.8 TiB 5.0 TiB 5.0 TiB 56.83 > > --- POOLS --- > POOL ID PGS STORED OBJECTS USED %USED MAX AVAIL > .mgr 1 1 449 KiB 2 1.8 MiB 0 320 GiB > cephfs.application-tst.meta 2 16 540 MiB 18.79k 2.1 GiB 0.16 320 GiB > cephfs.application-tst.data 3 32 4.4 GiB 8.01k 17 GiB 1.33 320 GiB > cephfs.application-acc.meta 4 16 11 GiB 3.54M 45 GiB 3.37 320 GiB > cephfs.application-acc.data 5 128 1.2 TiB 13.74M 4.8 TiB 79.46 320 GiB > ---------------------------------------------------------------------------- > > ---------------------------------------------------------------------------- > [ceph: root@aqsoel11445 /]# ceph osd pool autoscale-status > POOL SIZE TARGET SIZE RATE RAW CAPACITY RATIO TARGET RATIO EFFECTIVE RATIO BIAS PG_NUM NEW PG_NUM AUTOSCALE BULK > .mgr 448.5k 4.0 4499G 0.0000 1.0 1 on False > cephfs.application-tst.meta 539.8M 4.0 4499G 0.0005 4.0 16 on False > cephfs.application-tst.data 4488M 51200M 4.0 4499G 0.0444 1.0 32 on False > cephfs.application-acc.meta 11430M 4.0 4499G 0.0099 4.0 16 on False > cephfs.application-acc.data 1244G 4.0 4499G 1.1062 1.0000 0.9556 1.0 128 on False > ---------------------------------------------------------------------------- > > _______________________________________________ > ceph-users mailing list -- ceph-users@xxxxxxx > To unsubscribe send an email to ceph-users-leave@xxxxxxx > _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx