Re: [ceph-users] cuttlefish countdown -- OSD doesn't get marked out

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi David,

did you test it with more than one rack as well? In my first problem I
used two racks, with a custom crushmap, so that the replicas are in the
two racks (replicationlevel = 2). Than I took one osd down, and expected
that the remaining osds in this rack would get the now missing replicas
from the osd of the other rack.
But nothing happened, the cluster stayed degraded.

-martin


On 26.04.2013 02:22, David Zafman wrote:
> 
> I filed tracker bug 4822 and have wip-4822 with a fix.  My manual testing shows that it works.  I'm building a teuthology test.
> 
> Given your osd tree has a single rack it should always mark OSDs down after 5 minutes by default.
> 
> David Zafman
> Senior Developer
> http://www.inktank.com
> 
> 
> 
> 
> On Apr 25, 2013, at 9:38 AM, Martin Mailand <martin@xxxxxxxxxxxx> wrote:
> 
>> Hi Sage,
>>
>> On 25.04.2013 18:17, Sage Weil wrote:
>>> What is the output from 'ceph osd tree' and the contents of your 
>>> [mon*] sections of ceph.conf?
>>>
>>> Thanks!
>>> sage
>>
>>
>> root@store1:~# ceph osd tree
>>
>> # id	weight	type name	up/down	reweight
>> -1	24	root default
>> -3	24		rack unknownrack
>> -2	4			host store1
>> 0	1				osd.0	up	1	
>> 1	1				osd.1	down	1	
>> 2	1				osd.2	up	1	
>> 3	1				osd.3	up	1	
>> -4	4			host store3
>> 10	1				osd.10	up	1	
>> 11	1				osd.11	up	1	
>> 8	1				osd.8	up	1	
>> 9	1				osd.9	up	1	
>> -5	4			host store4
>> 12	1				osd.12	up	1	
>> 13	1				osd.13	up	1	
>> 14	1				osd.14	up	1	
>> 15	1				osd.15	up	1	
>> -6	4			host store5
>> 16	1				osd.16	up	1	
>> 17	1				osd.17	up	1	
>> 18	1				osd.18	up	1	
>> 19	1				osd.19	up	1	
>> -7	4			host store6
>> 20	1				osd.20	up	1	
>> 21	1				osd.21	up	1	
>> 22	1				osd.22	up	1	
>> 23	1				osd.23	up	1	
>> -8	4			host store2
>> 4	1				osd.4	up	1	
>> 5	1				osd.5	up	1	
>> 6	1				osd.6	up	1	
>> 7	1				osd.7	up	1	
>>
>>
>>
>> [global]
>>        auth cluster requierd = none
>>        auth service required = none
>>        auth client required = none
>> #       log file = ""
>>        log_max_recent=100
>>        log_max_new=100
>>
>> [mon]
>>        mon data = /data/mon.$id
>> [mon.a]
>>        mon host = store1
>>        mon addr = 192.168.195.31:6789
>> [mon.b]
>>        mon host = store3
>>        mon addr = 192.168.195.33:6789
>> [mon.c]
>>        mon host = store5
>>        mon addr = 192.168.195.35:6789
>> _______________________________________________
>> ceph-users mailing list
>> ceph-users@xxxxxxxxxxxxxx
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> 
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html




[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]

  Powered by Linux