Each OSD lives on a separate HDD in bluestore with the journals
on 2GB partitions on a shared SSD.
On 16/02/18 21:08, Gregory Farnum
wrote:
What does the cluster deployment look like? Usually this happens
when you’re sharing disks with the OS, or have co-located file
journals or something.
I'm trying
out cephfs and I'm in the process of copying over some
real-world data to see what happens.
I have created a number of cephfs file systems, the only one
I've
started working on is the one called jenkins specifically the
one named
jenkins which lives in fs_jenkins_data and
fs_jenkins_metadata.
According to ceph df I have about 1387 GB of data in all of
the pools,
while the raw used space is 5918 GB, which gives a ratio of
about 4.3, I
would have expected a ratio around 2 as the pool size has been
set to 2.
Can anyone explain where half my space has been squandered?
> ceph df
GLOBAL:
SIZE AVAIL RAW USED %RAW USED
8382G 2463G 5918G 70.61
POOLS:
NAME ID USED %USED
MAX
AVAIL OBJECTS
.rgw.root 1 1113 0
258G 4
default.rgw.control 2 0 0
258G 8
default.rgw.meta 3 0 0
258G 0
default.rgw.log 4 0 0
258G 207
fs_docker-nexus_data 5 66120M 11.09
258G 22655
fs_docker-nexus_metadata 6 39463k 0
258G 2376
fs_meta_data 7 330 0
258G 4
fs_meta_metadata 8 567k 0
258G 22
fs_jenkins_data 9 1321G 71.84
258G 28576278
fs_jenkins_metadata 10 52178k 0
258G 2285493
fs_nexus_data 11 0 0
258G 0
fs_nexus_metadata 12 4181 0
258G 21
--
Regards Flemming Frandsen - Stibo Systems - DK - STEP
Release Manager
Please use release@xxxxxxxxx
for all Release Management requests
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
--
Regards Flemming Frandsen - Stibo Systems - DK - STEP Release Manager
Please use release@xxxxxxxxx for all Release Management requests
|
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com