Hello, I've already thought about that - but even after changing the replication level (size) I'm not getting a clean cluster (there are only the default pools ATM): root at ceph-m-02:~#ceph -s cluster b04fc583-9e71-48b7-a741-92f4dff4cfef health HEALTH_WARN 232 pgs stuck unclean; recovery 26/126 objects degraded (20.635%) monmap e2: 3 mons at {ceph-m-01=10.0.0.100:6789/0,ceph-m-02=10.0.1.101:6789/0,ceph-m-03=10.0.1.102:6789/0}, election epoch 8, quorum 0,1,2 ceph-m-01,ceph-m-02,ceph-m-03 osdmap e56: 9 osds: 9 up, 9 in pgmap v287: 232 pgs, 8 pools, 822 bytes data, 43 objects 9342 MB used, 78317 GB / 78326 GB avail 26/126 objects degraded (20.635%) 119 active 113 active+remapped root at ceph-m-02:~#ceph osd dump | grep size pool 0 'data' replicated size 2 min_size 2 crush_ruleset 0 object_hash rjenkins pg_num 64 pgp_num 64 last_change 48 owner 0 flags hashpspool crash_replay_interval 45 stripe_width 0 pool 1 'metadata' replicated size 2 min_size 2 crush_ruleset 0 object_hash rjenkins pg_num 64 pgp_num 64 last_change 49 owner 0 flags hashpspool stripe_width 0 pool 2 'rbd' replicated size 2 min_size 2 crush_ruleset 0 object_hash rjenkins pg_num 64 pgp_num 64 last_change 50 owner 0 flags hashpspool stripe_width 0 pool 3 '.rgw.root' replicated size 2 min_size 2 crush_ruleset 0 object_hash rjenkins pg_num 8 pgp_num 8 last_change 52 owner 0 flags hashpspool stripe_width 0 pool 4 '.rgw.control' replicated size 2 min_size 2 crush_ruleset 0 object_hash rjenkins pg_num 8 pgp_num 8 last_change 53 owner 0 flags hashpspool stripe_width 0 pool 5 '.rgw' replicated size 2 min_size 2 crush_ruleset 0 object_hash rjenkins pg_num 8 pgp_num 8 last_change 54 owner 18446744073709551615 flags hashpspool stripe_width 0 pool 6 '.rgw.gc' replicated size 2 min_size 2 crush_ruleset 0 object_hash rjenkins pg_num 8 pgp_num 8 last_change 55 owner 0 flags hashpspool stripe_width 0 pool 7 '.users.uid' replicated size 2 min_size 2 crush_ruleset 0 object_hash rjenkins pg_num 8 pgp_num 8 last_change 56 owner 18446744073709551615 flags hashpspool stripe_width 0 Kind Regards, Georg On 09.05.2014 08:29, Mark Kirkwood wrote: > So that's two hosts - if this is a new cluster chances are the pools > have replication size=3, and won't place replica pgs on the same host... > 'ceph osd dump' will let you know if this is the case. If it is ether > reduce size to 2, add another host or edit your crush rules to allow > replica pgs on the same host. > > Cheers > > Mark > > On 09/05/14 18:20, Georg H?llrigl wrote: >> #ceph osd tree >> # id weight type name up/down reweight >> -1 76.47 root default >> -2 32.72 host ceph-s-01 >> 0 7.27 osd.0 up 1 >> 1 7.27 osd.1 up 1 >> 2 9.09 osd.2 up 1 >> 3 9.09 osd.3 up 1 >> -3 43.75 host ceph-s-02 >> 4 10.91 osd.4 up 1 >> 5 0.11 osd.5 up 1 >> 6 10.91 osd.6 up 1 >> 7 10.91 osd.7 up 1 >> 8 10.91 osd.8 up 1 >> >> >> On 08.05.2014 19:11, Craig Lewis wrote: >>> What does `ceph osd tree` output? >>> >>> On 5/8/14 07:30 , Georg H?llrigl wrote: >>>> Hello, >>>> >>>> We've a fresh cluster setup - with Ubuntu 14.04 and ceph firefly. By >>>> now I've tried this multiple times - but the result keeps the same and >>>> shows me lots of troubles (the cluster is empty, no client has >>>> accessed it) >>>> >>>> #ceph -s >>>> cluster b04fc583-9e71-48b7-a741-92f4dff4cfef >>>> health HEALTH_WARN 470 pgs stale; 470 pgs stuck stale; 18 pgs >>>> stuck unclean; 26 requests are blocked > 32 sec >>>> monmap e2: 3 mons at >>>> {ceph-m-01=10.0.0.100:6789/0,ceph-m-02=10.0.1.101:6789/0,ceph-m-03=10.0.1.102:6789/0}, >>>> >>>> election epoch 8, quorum 0,1,2 ceph-m-01,ceph-m-02,ceph-m-03 >>>> osdmap e409: 9 osds: 9 up, 9 in >>>> pgmap v1231: 480 pgs, 9 pools, 822 bytes data, 43 objects >>>> 9373 MB used, 78317 GB / 78326 GB avail >>>> 451 stale+active+clean >>>> 1 stale+active+clean+scrubbing >>>> 10 active+clean >>>> 18 stale+active+remapped >>>> >> _______________________________________________ >> ceph-users mailing list >> ceph-users at lists.ceph.com >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com