#ceph osd tree # id weight type name up/down reweight -1 76.47 root default -2 32.72 host ceph-s-01 0 7.27 osd.0 up 1 1 7.27 osd.1 up 1 2 9.09 osd.2 up 1 3 9.09 osd.3 up 1 -3 43.75 host ceph-s-02 4 10.91 osd.4 up 1 5 0.11 osd.5 up 1 6 10.91 osd.6 up 1 7 10.91 osd.7 up 1 8 10.91 osd.8 up 1 On 08.05.2014 19:11, Craig Lewis wrote: > What does `ceph osd tree` output? > > On 5/8/14 07:30 , Georg H?llrigl wrote: >> Hello, >> >> We've a fresh cluster setup - with Ubuntu 14.04 and ceph firefly. By >> now I've tried this multiple times - but the result keeps the same and >> shows me lots of troubles (the cluster is empty, no client has >> accessed it) >> >> #ceph -s >> cluster b04fc583-9e71-48b7-a741-92f4dff4cfef >> health HEALTH_WARN 470 pgs stale; 470 pgs stuck stale; 18 pgs >> stuck unclean; 26 requests are blocked > 32 sec >> monmap e2: 3 mons at >> {ceph-m-01=10.0.0.100:6789/0,ceph-m-02=10.0.1.101:6789/0,ceph-m-03=10.0.1.102:6789/0}, >> election epoch 8, quorum 0,1,2 ceph-m-01,ceph-m-02,ceph-m-03 >> osdmap e409: 9 osds: 9 up, 9 in >> pgmap v1231: 480 pgs, 9 pools, 822 bytes data, 43 objects >> 9373 MB used, 78317 GB / 78326 GB avail >> 451 stale+active+clean >> 1 stale+active+clean+scrubbing >> 10 active+clean >> 18 stale+active+remapped >>