Hi, I just setup a test ceph installation on 3 node Centos 6.5 . two of the nodes are used for hosting osds and the third acts as mon . Please note I'm using LVM so had to set up the osd using the manual install guide. --snip-- ceph -s cluster 2929fa80-0841-4cb6-a133-90b2098fc802 health HEALTH_WARN 192 pgs stuck inactive; 192 pgs stuck unclean; noup,nodown,noout flag(s) set monmap e2: 3 mons at {ceph0= 10.0.12.220:6789/0,ceph1=10.0.12.221:6789/0,ceph2=10.0.12.222:6789/0}, election epoch 46, quorum 0,1,2 ceph0,ceph1,ceph2 osdmap e21: 2 osds: 0 up, 0 in flags noup,nodown,noout pgmap v22: 192 pgs, 3 pools, 0 bytes data, 0 objects 0 kB used, 0 kB / 0 kB avail 192 creating --snip-- osd tree --snip-- ceph osd tree # id weight type name up/down reweight -1 2 root default -3 1 host ceph1 0 1 osd.0 down 0 -2 1 host ceph2 1 1 osd.1 down 0 --snip-- --snip-- ceph daemon osd.0 status { "cluster_fsid": "99babb8f-c880-4b32-a227-94aa483d4871", "osd_fsid": "1ad28bde-c23c-44ba-a3b7-0aaaafd3372e", "whoami": 0, "state": "booting", "oldest_map": 1, "newest_map": 21, "num_pgs": 0} --snip-- --snip-- ceph daemon osd.1 status { "cluster_fsid": "99babb8f-c880-4b32-a227-94aa483d4871", "osd_fsid": "becc3252-6977-47d6-87af-7b1337e591d8", "whoami": 1, "state": "booting", "oldest_map": 1, "newest_map": 21, "num_pgs": 0} --snip-- # Cpus are idling # does anybody know what is wrong Thanks in advance -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://lists.ceph.com/pipermail/ceph-users-ceph.com/attachments/20140726/30172d39/attachment.htm>