The default CRUSH placement rules want to put replica pgs on different hosts - so with pools of size 3 you need at least 3 hosts. You can get around this by editing your CRUSH rules to put replica pgs on different OSDs instead - but clearly redundancy is impacted in that case - best to have (number of hosts) >= (pool size). regards Mark On 12/05/14 20:30, Georg H?llrigl wrote: > Thank you soo much! That seems to work immidetately. > > ATM I still see 3 pgs in active+clean+scrubbing state - but that will > hopefully fix by time. > > So the way to go with firefly, is to either use at least 3 hosts for > OSDs - or reduce the number of replicas? > > Kind Regards, > Georg > > > On 09.05.2014 10:59, Martin B Nielsen wrote: >> Hi, >> >> I experienced exactly the same with 14.04 and the 0.79 release. >> >> It was a fresh clean install with default crushmap and ceph-deploy >> install as pr. the quick-start guide. >> >> Oddly enough changing replica size (incl min_size) from 3 - 2 (and 2->1) >> and back again it worked. >> >> I didn't have time to look into replicating the issue. >> >> Cheers, >> Martin >> >> >> On Thu, May 8, 2014 at 4:30 PM, Georg H?llrigl >> <georg.hoellrigl at xidras.com <mailto:georg.hoellrigl at xidras.com>> wrote: >> >> Hello, >> >> We've a fresh cluster setup - with Ubuntu 14.04 and ceph firefly. By >> now I've tried this multiple times - but the result keeps the same >> and shows me lots of troubles (the cluster is empty, no client has >> accessed it) >> >> #ceph -s >> cluster b04fc583-9e71-48b7-a741-__92f4dff4cfef >> health HEALTH_WARN 470 pgs stale; 470 pgs stuck stale; 18 pgs >> stuck unclean; 26 requests are blocked > 32 sec >> monmap e2: 3 mons at >> >> {ceph-m-01=10.0.0.100:6789/0,__ceph-m-02=10.0.1.101:6789/0,__ceph-m-03=10.0.1.102:6789/0 >> >> >> <http://10.0.0.100:6789/0,ceph-m-02=10.0.1.101:6789/0,ceph-m-03=10.0.1.102:6789/0>}, >> >> election epoch 8, quorum 0,1,2 ceph-m-01,ceph-m-02,ceph-m-03 >> osdmap e409: 9 osds: 9 up, 9 in >> pgmap v1231: 480 pgs, 9 pools, 822 bytes data, 43 objects >> 9373 MB used, 78317 GB / 78326 GB avail >> 451 stale+active+clean >> 1 stale+active+clean+scrubbing >> 10 active+clean >> 18 stale+active+remapped >> >> Anyone an idea what happens here? Should an empty cluster not show >> only active+clean pgs? >> >> >> Regards, >> Georg >> _________________________________________________ >> ceph-users mailing list >> ceph-users at lists.ceph.com <mailto:ceph-users at lists.ceph.com> >> http://lists.ceph.com/__listinfo.cgi/ceph-users-ceph.__com >> <http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com> > _______________________________________________ > ceph-users mailing list > ceph-users at lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com