Umh....
Taking a look to your computation I think the ratio OSD/Overhead
it's really about 1.1Gb per OSD.
Because I have 9 NVMe OSD alive right now. So about 9.5Gb of
overhead.
So I guess this is just it's right behaviour.
Fine!
Il 02/03/2018 15:18, David Turner ha
scritto:
[1] Here is a ceph starts on a brand new cluster that has
never had any pools created or data or into it at all. 323GB
used out of 2.3PB. that's 0.01% overhead, but we're using 10TB
disks for this cluster, and the overhead is moreso per osd than
per TB. It is 1.1GB overhead per osd. 34 of the osds are pure
nvme and the other 255 have collocated DBs with their WAL on
flash.
The used space your string is most likely just osd overhead,
but you can double check if there are any orphaned rados objects
using up space with a `rados ls`. Another thing to note is that
deleting a pool in ceph is not instant. It goes into garbage
collection and is taken care of over time. Most likely you're
just looking at osd overhead, though.
[1]
$ ceph -s
cluster:
health: HEALTH_OK
services:
mon: 5 daemons, quorum mon1,mon2,mon4,mon3,mon5
mgr: mon1(active), standbys: mon3, mon2, mon5, mon4
osd: 289 osds: 289 up, 289 in
data:
pools: 0 pools, 0 pgs
objects: 0 objects, 0 bytes
usage: 323 GB used, 2324 TB / 2324 TB avail
|
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com