On Wed, Nov 12, 2014 at 1:41 PM, houmles <houmles@xxxxxxxxx> wrote: > Hi, > > I have 2 hosts with 8 2TB drive in each. > I want to have 2 replicas between both hosts and then 2 replicas between osds on each host. That way even when I lost one host I still have 2 replicas. > > Currently I have this ruleset: > > rule repl { > ruleset 5 > type replicated > min_size 1 > max_size 10 > step take asterix > step choose firstn -2 type osd > step emit > step take obelix > step choose firstn 2 type osd > step emit > } > > Which works ok. I have 4 replicas as I want and PGs are distributed perfectly but when I run ceph df I have only 1/2 of my capacity which I should have. > In total it's 32TB, 16TB in each host. If there is a 2 replicas on each host it should report around 8TB, right? It's reporting only 4TB in pool which is 1/8 of total capacity. > Can anyone tell me what is wrong? What version are you running? Can you copy-paste the command and output, pointing out which bit you think is wrong? There are occasionally oddities in the source data that confuse things and I think there's new functionality to try and predict the "effective" size that might have an issue. -Greg _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com