anti-cephalopod question

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Is the '15 minutes or so '  something that can be configured at run time?


On Mon, Jul 28, 2014 at 8:44 AM, Joao Eduardo Luis <joao.luis at inktank.com>
wrote:

> (CC'ing ceph-users)
>
> On 07/28/2014 12:34 PM, Marc wrote:
>
>> Hi,
>>
>>
>>  This said, if out of 3 monitors you have 2 monitors down, your cluster
>>> will cease functioning (no admin commands, no writes or reads served).
>>>
>>
>> This is not entirely true. (At least) RBDs will continue being fully
>> functional even if the mon quorum is lost. This only applies to RBDs
>> that are already mounted (qemu) at the time of quorum loss though.
>>
>> Meaning: (K)VMs running off of Ceph will remain fully functional even if
>> the mon quorum is lost (assuming you havent lost too many OSDs at the
>> same time).
>>
>
> True.  Clients will maintain the connections they have to OSDs for about
> 15 minutes or so, at which point timeouts will go off and all work will be
> halted.  New clients won't be able to do this though, as they have to grab
> maps from the monitors prior to connecting to OSDs, and the monitor will
> not serve those requests if quorum is not in place.
>
>   -Joao
>
>
>
>> On 28/07/2014 12:22, Joao Eduardo Luis wrote:
>>
>>> On 07/28/2014 08:49 AM, Christian Balzer wrote:
>>>
>>>>
>>>> Hello,
>>>>
>>>> On Sun, 27 Jul 2014 18:20:43 -0400 Robert Fantini wrote:
>>>>
>>>>  Hello Christian,
>>>>>
>>>>> Let me supply more info and answer some questions.
>>>>>
>>>>> * Our main concern is high availability, not speed.
>>>>> Our storage requirements are not huge.
>>>>> However we want good keyboard response 99.99% of the time.   We
>>>>> mostly do
>>>>> data entry and reporting.   20-25  users doing mostly order , invoice
>>>>> processing and email.
>>>>>
>>>>> * DRBD has been very reliable , but I am the SPOF .   Meaning that when
>>>>> split brain occurs [ every 18-24 months ] it is me or no one who knows
>>>>> what to do. Try to explain how to deal with split brain in advance....
>>>>> For the future ceph looks like it will be easier to maintain.
>>>>>
>>>>>  The DRBD people would of course tell you to configure things in a way
>>>> that
>>>> a split brain can't happen. ^o^
>>>>
>>>> Note that given the right circumstances (too many OSDs down, MONs down)
>>>> Ceph can wind up in a similar state.
>>>>
>>>
>>>
>>> I am not sure what you mean by ceph winding up in a similar state.  If
>>> you mean regarding 'split brain' in the usual sense of the term, it does
>>> not occur in Ceph.  If it does, you have surely found a bug and you
>>> should let us know with lots of CAPS.
>>>
>>> What you can incur though if you have too many monitors down is cluster
>>> downtime.  The monitors will ensure you need a strict majority of
>>> monitors up in order to operate the cluster, and will not serve requests
>>> if said majority is not in place.  The monitors will only serve requests
>>> when there's a formed 'quorum', and a quorum is only formed by (N/2)+1
>>> monitors, N being the total number of monitors in the cluster (via the
>>> monitor map -- monmap).
>>>
>>> This said, if out of 3 monitors you have 2 monitors down, your cluster
>>> will cease functioning (no admin commands, no writes or reads served).
>>> As there is no configuration in which you can have two strict
>>> majorities, thus no two partitions of the cluster are able to function
>>> at the same time, you do not incur in split brain.
>>>
>>> If you are a creative admin however, you may be able to enforce split
>>> brain by modifying monmaps.  In the end you'd obviously end up with two
>>> distinct monitor clusters, but if you so happened to not inform the
>>> clients about this there's a fair chance that it would cause havoc with
>>> unforeseen effects.  Then again, this would be the operator's fault, not
>>> Ceph itself -- especially because rewriting monitor maps is not trivial
>>> enough for someone to mistakenly do something like this.
>>>
>>>    -Joao
>>>
>>>
>>>
>>
>
> --
> Joao Eduardo Luis
> Software Engineer | http://inktank.com | http://ceph.com
> _______________________________________________
> ceph-users mailing list
> ceph-users at lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ceph.com/pipermail/ceph-users-ceph.com/attachments/20140728/cb6df31e/attachment.htm>


[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux