On 9/25/19 3:22 PM, nalexandrov@xxxxxxxxxxxxxx wrote: > Hi everyone, > > We are running Nautilus 14.2.2 with 6 nodes and a total of 44 OSDs, all are 2TB spinning disks. > # ceph osd count-metadata osd_objectstore > "bluestore": 44 > # ceph osd pool get one size > size: 3 > # ceph df > RAW STORAGE: > CLASS SIZE AVAIL USED RAW USED %RAW USED > hdd 80 TiB 33 TiB 47 TiB 47 TiB 58.26 > TOTAL 80 TiB 33 TiB 47 TiB 47 TiB 58.26 > > POOLS: > POOL ID STORED OBJECTS USED %USED MAX AVAIL > one 2 15 TiB 4.06M 47 TiB 68.48 7.1 TiB > bench 5 250 MiB 67 250 MiB 0 21 TiB > > Why pool's stats are showing incorrect values for %USED and MAX AVAIL? > They should be much bigger. Keep in mind that the MAX AVAIL is the usable capacity which takes the replication factor (3?) into account and also the nearfull ratio (0.85) of the OSDs serving that pool. The fullest OSD of the pool will determine how much data can still be stored. Wido > The first 24 OSDs was created on jewell release and the osd_objectstore was 'filestore'. > While we were with mimic release, we added 20 more 'bluestore' OSDs. The first 24 was destroyed and recreated as 'bluestore'. > After the upgrade from mimic release, all the OSD's was updated with ceph-bluestore-tool repair. > The incorrect values appeared after the upgrade from 14.2.1 to 14.2.2. > Any help will be appreciated :) > > BR, > NAlexandrov > _______________________________________________ > ceph-users mailing list -- ceph-users@xxxxxxx > To unsubscribe send an email to ceph-users-leave@xxxxxxx > _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx