Currently Firefly on Debian stable, all updated. I already tried it with Giant and it's same. But it's look like I solved it. I change crush tunables to optimal and now it shows the size right. And even when I switch back to default it shows it right. It's weird, but hopefully it's solved for now. On Tue, Nov 18, 2014 at 03:51:56PM -0800, Gregory Farnum wrote: > On Wed, Nov 12, 2014 at 1:41 PM, houmles <houmles@xxxxxxxxx> wrote: > > Hi, > > > > I have 2 hosts with 8 2TB drive in each. > > I want to have 2 replicas between both hosts and then 2 replicas between osds on each host. That way even when I lost one host I still have 2 replicas. > > > > Currently I have this ruleset: > > > > rule repl { > > ruleset 5 > > type replicated > > min_size 1 > > max_size 10 > > step take asterix > > step choose firstn -2 type osd > > step emit > > step take obelix > > step choose firstn 2 type osd > > step emit > > } > > > > Which works ok. I have 4 replicas as I want and PGs are distributed perfectly but when I run ceph df I have only 1/2 of my capacity which I should have. > > In total it's 32TB, 16TB in each host. If there is a 2 replicas on each host it should report around 8TB, right? It's reporting only 4TB in pool which is 1/8 of total capacity. > > Can anyone tell me what is wrong? > > What version are you running? Can you copy-paste the command and > output, pointing out which bit you think is wrong? There are > occasionally oddities in the source data that confuse things and I > think there's new functionality to try and predict the "effective" > size that might have an issue. > -Greg _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com