Jack,
to be correct all kudos shoud go to David Turner since this was his
comment.
I just wanted to thank you and him for taking the time to answer my
initial question.
Kind regards,
G.
Thanks Georgios for pointing out my mistake: I can still perform
reads
only because of IO cache on the host .. so the pool really block all
requests
On 02/24/2018 01:45 PM, Georgios Dimitrakakis wrote:
The pool will not actually go read only. All read and write
requests
will block until both osds are back up. If I were you, I would use
min_size=2 and change it to 1 temporarily if needed to do
maintenance
or troubleshooting where down time is not an option.
On Thu, Feb 22, 2018, 5:31 PM Georgios Dimitrakakis wrote:
All right! Thank you very much Jack!
The way I understand this is that its not necessarily a bad
thing. I
mean as long as it doesnt harm any data or
cannot cause any other issue.
Unfortunately my scenario consists of only two OSDs therefore
there is
a replication factor of 2 and min_size=1.
What I am trying to figure out is if its more dangerous to have
min_size=2 rather than 1 in the above scenario and if it gives me
any
benefits.
I am already aware of the *golden* rule about the minimum number
of
replicas (3) but the cluster will be reformed soon and until then
I
would like to know if its better to go with min_size=2 or not.
Regards,
G.
> If min_size == size, a single OSD failure will place your pool
read
> only
>
> On 02/22/2018 11:06 PM, Georgios Dimitrakakis wrote:
>> Dear all,
>>
>> I would like to know if there are additional risks when running
CEPH
>> with "Min Size" equal to "Replicated Size" for a given pool.
>>
>> What are the drawbacks and what could be go wrong in such a
>> scenario?
>>
>> Best regards,
>>
>> G.
>> _______________________________________________
>> ceph-users mailing list
>> ceph-users@xxxxxxxxxxxxxx [1]
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com [2]
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@xxxxxxxxxxxxxx [3]
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com [4]
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx [5]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com [6]
Links:
------
[1] mailto:ceph-users@xxxxxxxxxxxxxx
[2] http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
[3] mailto:ceph-users@xxxxxxxxxxxxxx
[4] http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
[5] mailto:ceph-users@xxxxxxxxxxxxxx
[6] http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
[7] mailto:giorgis@xxxxxxxxxxxx
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com