On Tue, Jul 1, 2014 at 11:45 AM, Gregory Farnum <greg at inktank.com> wrote: > On Tue, Jul 1, 2014 at 11:33 AM, Brian Lovett > <brian.lovett at prosperent.com> wrote: >> Brian Lovett <brian.lovett at ...> writes: >> >> >> I restarted all of the osd's and noticed that ceph shows 2 osd's up even if >> the servers are completely powered down: osdmap e95: 8 osds: 2 up, 8 in >> >> Why would that be? > > The OSDs report each other down much more quickly (~30s) than the > monitor timeout (~15 minutes). They'd get marked down eventually. > > On Tue, Jul 1, 2014 at 11:43 AM, Brian Lovett > <brian.lovett at prosperent.com> wrote: >> Gregory Farnum <greg at ...> writes: >> >>> >>> What's the output of "ceph osd map"? >>> >>> Your CRUSH map probably isn't trying to segregate properly, with 2 >>> hosts and 4 OSDs each. >>> Software Engineer #42 <at> http://inktank.com | http://ceph.com >>> >> Is this what you are looking for? >> >> ceph osd map rbd ceph >> osdmap e104 pool 'rbd' (2) object 'ceph' -> pg 2.3482c180 (2.0) -> up ([3,5], >> p3) acting ([3,5,0], p3) > > Whoops, I mean "ceph osd list", sorry! (That should output a textual > representation of how they're arranged in the CRUSH map.) ...and one more time, because apparently my brain's out to lunch today: ceph osd tree *sigh*