HEALTH_WARN active+degraded on fresh install CENTOS 6.5

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Tue, Jul 1, 2014 at 11:45 AM, Gregory Farnum <greg at inktank.com> wrote:
> On Tue, Jul 1, 2014 at 11:33 AM, Brian Lovett
> <brian.lovett at prosperent.com> wrote:
>> Brian Lovett <brian.lovett at ...> writes:
>>
>>
>> I restarted all of the osd's and noticed that ceph shows 2 osd's up even if
>> the servers are completely powered down:  osdmap e95: 8 osds: 2 up, 8 in
>>
>> Why would that be?
>
> The OSDs report each other down much more quickly (~30s) than the
> monitor timeout (~15 minutes). They'd get marked down eventually.
>
> On Tue, Jul 1, 2014 at 11:43 AM, Brian Lovett
> <brian.lovett at prosperent.com> wrote:
>> Gregory Farnum <greg at ...> writes:
>>
>>>
>>> What's the output of "ceph osd map"?
>>>
>>> Your CRUSH map probably isn't trying to segregate properly, with 2
>>> hosts and 4 OSDs each.
>>> Software Engineer #42  <at>  http://inktank.com | http://ceph.com
>>>
>>  Is this what you are looking for?
>>
>> ceph osd map rbd ceph
>> osdmap e104 pool 'rbd' (2) object 'ceph' -> pg 2.3482c180 (2.0) -> up ([3,5],
>> p3) acting ([3,5,0], p3)
>
> Whoops, I mean "ceph osd list", sorry! (That should output a textual
> representation of how they're arranged in the CRUSH map.)

...and one more time, because apparently my brain's out to lunch today:

ceph osd tree

*sigh*


[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux