I'm trying out cephfs and I'm in the process of copying over some
real-world data to see what happens.
I have created a number of cephfs file systems, the only one I've
started working on is the one called jenkins specifically the one named
jenkins which lives in fs_jenkins_data and fs_jenkins_metadata.
According to ceph df I have about 1387 GB of data in all of the pools,
while the raw used space is 5918 GB, which gives a ratio of about 4.3, I
would have expected a ratio around 2 as the pool size has been set to 2.
Can anyone explain where half my space has been squandered?
> ceph df
GLOBAL:
SIZE AVAIL RAW USED %RAW USED
8382G 2463G 5918G 70.61
POOLS:
NAME ID USED %USED MAX
AVAIL OBJECTS
.rgw.root 1 1113 0 258G 4
default.rgw.control 2 0 0 258G 8
default.rgw.meta 3 0 0 258G 0
default.rgw.log 4 0 0 258G 207
fs_docker-nexus_data 5 66120M 11.09 258G 22655
fs_docker-nexus_metadata 6 39463k 0 258G 2376
fs_meta_data 7 330 0 258G 4
fs_meta_metadata 8 567k 0 258G 22
fs_jenkins_data 9 1321G 71.84 258G 28576278
fs_jenkins_metadata 10 52178k 0 258G 2285493
fs_nexus_data 11 0 0 258G 0
fs_nexus_metadata 12 4181 0 258G 21
--
Regards Flemming Frandsen - Stibo Systems - DK - STEP Release Manager
Please use release@xxxxxxxxx for all Release Management requests
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com