Hello Vincenzo, Yes, those 6 OSDs are on different hosts. I've got 3 VMs each with 2 OSDs. So this should be enough for the requirement to have 3 replica's (even though I set it back to 2, as suggested in the howto's). I will try to have the replica's only over the OSDs en not over the hosts tomorrow. Met vriendelijke groet/With kind regards, Tijn Buijs Cloud.nl logo tijn at cloud.nl <mailto:tijn at cloud.nl> | T. 0800-CLOUDNL / +31 (0)162 820 000 | F. +31 (0)162 820 001 Cloud.nl B.V. | Minervum 7092D | 4817 ZK Breda | www.cloud.nl <http://www.cloud.nl> On 31/07/14 17:18, Vincenzo Pii wrote: > Are the 6 osds on different hosts? > > The default ruleset that ceph applies to pools states that object > replicas (3 per default) should be placed on OSDs of different hosts. > This cannot be satisfied if you don't have OSDs on separate hosts. > > I ran myself into this issue and wrote down the steps I needed to > solve it. > If this is your case, you can try to read it here: > http://blog.zhaw.ch/icclab/deploy-ceph-troubleshooting-part-23/ > (paragraph: "Check that replication requirements can be met"). > > Basically, you either specify a different crush ruleset or reduce the > size of the replicas for your pools. > > Hope this can help! > > Vincenzo. > > > 2014-07-31 16:36 GMT+02:00 Tijn Buijs <tijn at cloud.nl > <mailto:tijn at cloud.nl>>: > > Hello everybody, > > At cloud.nl <http://cloud.nl> we are going to use Ceph. So I find > it a good idea to get some handson experience with it, so I can > work with it :). So I'm installing a testcluster in a few > VirtualBox machines on my iMac, which runs OS X 10.9.4 offcourse. > I know I will get a lousy performance, but that's not the > objective here. The objective is to get some experience with Ceph, > to see how it works. > > But I hit an issue during the initial setup of the cluster. When > I'm done installing everything and following the howto's on > ceph.com <http://ceph.com> (the preflight > <http://ceph.com/docs/master/start/quick-start-preflight/> and the > Storage Cluster quick start > <http://ceph.com/docs/master/start/quick-ceph-deploy/>) I need to > run ceph health to see that everything is running perfectly. But > it doesn't run perfectly, I get the following output: > ceph at ceph-admin:~$ ceph health > HEALTH_WARN 192 pgs incomplete; 192 pgs stuck inactive; 192 pgs > stuck unclean > > And it stays at this information, it never ever changes. So > everything is really stuck. But I don't know what is stuck exactly > and how I can fix it. Some more info about my cluster: > ceph at ceph-admin:~$ ceph -s > cluster d31586a5-6dd6-454e-8835-0d6d9e204612 > health HEALTH_WARN 192 pgs incomplete; 192 pgs stuck > inactive; 192 pgs stuck unclean > monmap e3: 3 mons at > {ceph-mon1=10.28.28.18:6789/0,ceph-mon2=10.28.28.31:6789/0,ceph-mon3=10.28.28.50:6789/0 > <http://10.28.28.18:6789/0,ceph-mon2=10.28.28.31:6789/0,ceph-mon3=10.28.28.50:6789/0>}, > election epoch 4, quorum 0,1,2 ceph-mon1,ceph-mon2,ceph-mon3 > osdmap e25: 6 osds: 6 up, 6 in > pgmap v56: 192 pgs, 3 pools, 0 bytes data, 0 objects > 197 MB used, 30455 MB / 30653 MB avail > 192 creating+incomplete > > I'm running on Ubuntu 14.04.1 LTS Server. I did try to get it > running on CentOS 6.5 too (CentOS 6.5 is my actual distro of > choice, but Ceph has more affinity with Ubuntu, so I tried that > too), but I got exactly the same results. > > But because this is my first install of Ceph I don't know the > exact debug commands and stuff. I'm willing to get this working, > but I just don't know how :). Any help is appreciated :). > > Met vriendelijke groet/With kind regards, > > Tijn Buijs > > Cloud.nl logo > > tijn at cloud.nl <mailto:tijn at cloud.nl> | T. 0800-CLOUDNL / +31 > (0)162 820 000 <tel:%2B31%20%280%29162%20820%20000> | F. +31 > (0)162 820 001 <tel:%2B31%20%280%29162%20820%20001> > Cloud.nl B.V. | Minervum 7092D | 4817 ZK Breda | www.cloud.nl > <http://www.cloud.nl> > > > > > -- > Vincenzo Pii > Researcher, InIT Cloud Computing Lab > Zurich University of Applied Sciences (ZHAW) > http://www.cloudcomp.ch/ -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://lists.ceph.com/pipermail/ceph-users-ceph.com/attachments/20140731/8cba1dd8/attachment.htm> -------------- next part -------------- A non-text attachment was scrubbed... Name: logo_cloud_nl.jpg Type: image/jpeg Size: 14056 bytes Desc: not available URL: <http://lists.ceph.com/pipermail/ceph-users-ceph.com/attachments/20140731/8cba1dd8/attachment.jpg> -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: image/jpeg Size: 14056 bytes Desc: not available URL: <http://lists.ceph.com/pipermail/ceph-users-ceph.com/attachments/20140731/8cba1dd8/attachment.jpeg>