Re: Erasure coded PGs incomplete

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 




On 17/12/2014 19:46, Italo Santos wrote:> Understood.
> Thanks for your help, the cluster is healthy now :D
> 
> Also, using for example k=6,m=1 and failure domain by host I’ll be able lose all OSD on the same host, but if a lose 2 disks on different hosts I can lose data right? So, it is possible been a failure domain which allow me to lose an OSD or a host?

That's actually a good way to put it :-)

> 
> Regards.
> 
> *Italo Santos*
> http://italosantos.com.br/
> 
> On Wednesday, December 17, 2014 at 4:27 PM, Loic Dachary wrote:
> 
>>
>>
>> On 17/12/2014 19:22, Italo Santos wrote:
>>> Loic,
>>>
>>> So, if want have a failure domain by host, I’ll need set up a erasure profile which k+m = total number of hosts I have, right?
>>
>> Yes, k+m has to be <= number of hosts.
>>
>>>
>>> Regards.
>>>
>>> *Italo Santos*
>>> http://italosantos.com.br/
>>>
>>> On Wednesday, December 17, 2014 at 3:24 PM, Loic Dachary wrote:
>>>
>>>>
>>>>
>>>> On 17/12/2014 18:18, Italo Santos wrote:
>>>>> Hello,
>>>>>
>>>>> I’ve take a look to this documentation (which help a lot) and if I understand right, when I set a profile like:
>>>>>
>>>>> ===
>>>>> ceph osd erasure-code-profile set isilon k=8 m=2 ruleset-failure-domain=host
>>>>> ===
>>>>>
>>>>> And create a pool following the recommendations on doc, I’ll need (100*16)/2 = 800 PGs, I’ll need the sufficient number of hosts to support create total PGs?
>>>>
>>>> You will need k+m = 10 host per OSD. If you only have 10 hosts that should be ok and the 800 PGs will use these 10 OSD in various orders. It also means that you will end up having 800 PG per OSD which is a bit too mche. If you have 20 OSDs that will be better : each PG will get 10 OSD out of 20 and each OSD will have 400 PGs. Ideally you want the number of PG per OSD to be in the range (approximately) [20,300].
>>>>
>>>> Cheers
>>>>
>>>>>
>>>>> Regards.
>>>>>
>>>>> *Italo Santos*
>>>>> http://italosantos.com.br/
>>>>>
>>>>> On Wednesday, December 17, 2014 at 2:42 PM, Loic Dachary wrote:
>>>>>
>>>>>> Hi,
>>>>>>
>>>>>> Thanks for the update : good news are much appreciated :-) Would you have time to review the documentation at https://github.com/ceph/ceph/pull/3194/files ? It was partly motivated by the problem you had.
>>>>>>
>>>>>> Cheers
>>>>>>
>>>>>> On 17/12/2014 14:03, Italo Santos wrote:
>>>>>>> Hello Loic,
>>>>>>>
>>>>>>> Thanks for you help, I’ve take a look to my crush map and I replace "step chooseleaf indep 0 type osd” by "step choose indep 0 type osd” and all PGs was created successfully.
>>>>>>>
>>>>>>> At.
>>>>>>>
>>>>>>> *Italo Santos*
>>>>>>> http://italosantos.com.br/
>>>>>>>
>>>>>>> On Tuesday, December 16, 2014 at 8:39 PM, Loic Dachary wrote:
>>>>>>>
>>>>>>>> Hi,
>>>>>>>>
>>>>>>>> The 2147483647 means that CRUSH did not find enough OSD for a given PG. If you check the crush rule associated with the erasure coded pool, you will most probably find why.
>>>>>>>>
>>>>>>>> Cheers
>>>>>>>>
>>>>>>>> On 16/12/2014 23:32, Italo Santos wrote:
>>>>>>>>> Hello,
>>>>>>>>>
>>>>>>>>> I'm trying to create an erasure pool following http://docs.ceph.com/docs/master/rados/operations/erasure-code/, but when I try create a pool with a specifc erasure-code-profile ("myprofile") the PGs became on incomplete state.
>>>>>>>>>
>>>>>>>>> Anyone can help me?
>>>>>>>>>
>>>>>>>>> Below the profile I created:
>>>>>>>>> root@ceph0001:~# ceph osd erasure-code-profile get myprofile
>>>>>>>>> directory=/usr/lib/ceph/erasure-code
>>>>>>>>> k=6
>>>>>>>>> m=2
>>>>>>>>> plugin=jerasure
>>>>>>>>> technique=reed_sol_van
>>>>>>>>>
>>>>>>>>> The status of cluster:
>>>>>>>>> root@ceph0001:~# ceph health
>>>>>>>>> HEALTH_WARN 12 pgs incomplete; 12 pgs stuck inactive; 12 pgs stuck unclean
>>>>>>>>>
>>>>>>>>> health detail:
>>>>>>>>> root@ceph0001:~# ceph health detail
>>>>>>>>> HEALTH_WARN 12 pgs incomplete; 12 pgs stuck inactive; 12 pgs stuck unclean
>>>>>>>>> pg 2.9 is stuck inactive since forever, current state incomplete, last acting [4,10,15,2147483647,3,2147483647,2147483647,2147483647]
>>>>>>>>> pg 2.8 is stuck inactive since forever, current state incomplete, last acting [0,2147483647,4,2147483647,10,2147483647,15,2147483647]
>>>>>>>>> pg 2.b is stuck inactive since forever, current state incomplete, last acting [8,3,14,2147483647,5,2147483647,2147483647,2147483647]
>>>>>>>>> pg 2.a is stuck inactive since forever, current state incomplete, last acting [11,7,2,2147483647,2147483647,2147483647,15,2147483647]
>>>>>>>>> pg 2.5 is stuck inactive since forever, current state incomplete, last acting [12,8,5,1,2147483647,2147483647,2147483647,2147483647]
>>>>>>>>> pg 2.4 is stuck inactive since forever, current state incomplete, last acting [5,2147483647,13,1,2147483647,2147483647,8,2147483647]
>>>>>>>>> pg 2.7 is stuck inactive since forever, current state incomplete, last acting [12,2,10,7,2147483647,2147483647,2147483647,2147483647]
>>>>>>>>> pg 2.6 is stuck inactive since forever, current state incomplete, last acting [9,15,2147483647,4,2,2147483647,2147483647,2147483647]
>>>>>>>>> pg 2.1 is stuck inactive since forever, current state incomplete, last acting [2,4,2147483647,13,2147483647,10,2147483647,2147483647]
>>>>>>>>> pg 2.0 is stuck inactive since forever, current state incomplete, last acting [14,1,2147483647,4,10,2147483647,2147483647,2147483647]
>>>>>>>>> pg 2.3 is stuck inactive since forever, current state incomplete, last acting [14,11,6,2147483647,2147483647,2147483647,2,2147483647]
>>>>>>>>> pg 2.2 is stuck inactive since forever, current state incomplete, last acting [13,5,11,2147483647,2147483647,3,2147483647,2147483647]
>>>>>>>>> pg 2.9 is stuck unclean since forever, current state incomplete, last acting [4,10,15,2147483647,3,2147483647,2147483647,2147483647]
>>>>>>>>> pg 2.8 is stuck unclean since forever, current state incomplete, last acting [0,2147483647,4,2147483647,10,2147483647,15,2147483647]
>>>>>>>>> pg 2.b is stuck unclean since forever, current state incomplete, last acting [8,3,14,2147483647,5,2147483647,2147483647,2147483647]
>>>>>>>>> pg 2.a is stuck unclean since forever, current state incomplete, last acting [11,7,2,2147483647,2147483647,2147483647,15,2147483647]
>>>>>>>>> pg 2.5 is stuck unclean since forever, current state incomplete, last acting [12,8,5,1,2147483647,2147483647,2147483647,2147483647]
>>>>>>>>> pg 2.4 is stuck unclean since forever, current state incomplete, last acting [5,2147483647,13,1,2147483647,2147483647,8,2147483647]
>>>>>>>>> pg 2.7 is stuck unclean since forever, current state incomplete, last acting [12,2,10,7,2147483647,2147483647,2147483647,2147483647]
>>>>>>>>> pg 2.6 is stuck unclean since forever, current state incomplete, last acting [9,15,2147483647,4,2,2147483647,2147483647,2147483647]
>>>>>>>>> pg 2.1 is stuck unclean since forever, current state incomplete, last acting [2,4,2147483647,13,2147483647,10,2147483647,2147483647]
>>>>>>>>> pg 2.0 is stuck unclean since forever, current state incomplete, last acting [14,1,2147483647,4,10,2147483647,2147483647,2147483647]
>>>>>>>>> pg 2.3 is stuck unclean since forever, current state incomplete, last acting [14,11,6,2147483647,2147483647,2147483647,2,2147483647]
>>>>>>>>> pg 2.2 is stuck unclean since forever, current state incomplete, last acting [13,5,11,2147483647,2147483647,3,2147483647,2147483647]
>>>>>>>>> pg 2.9 is incomplete, acting [4,10,15,2147483647,3,2147483647,2147483647,2147483647] (reducing pool ecpool min_size from 6 may help; search ceph.com/docs <http://ceph.com/docs> for 'incomplete')
>>>>>>>>> pg 2.8 is incomplete, acting [0,2147483647,4,2147483647,10,2147483647,15,2147483647] (reducing pool ecpool min_size from 6 may help; search ceph.com/docs <http://ceph.com/docs> for 'incomplete')
>>>>>>>>> pg 2.b is incomplete, acting [8,3,14,2147483647,5,2147483647,2147483647,2147483647] (reducing pool ecpool min_size from 6 may help; search ceph.com/docs <http://ceph.com/docs> for 'incomplete')
>>>>>>>>> pg 2.a is incomplete, acting [11,7,2,2147483647,2147483647,2147483647,15,2147483647] (reducing pool ecpool min_size from 6 may help; search ceph.com/docs <http://ceph.com/docs> for 'incomplete')
>>>>>>>>> pg 2.5 is incomplete, acting [12,8,5,1,2147483647,2147483647,2147483647,2147483647] (reducing pool ecpool min_size from 6 may help; search ceph.com/docs <http://ceph.com/docs> for 'incomplete')
>>>>>>>>> pg 2.4 is incomplete, acting [5,2147483647,13,1,2147483647,2147483647,8,2147483647] (reducing pool ecpool min_size from 6 may help; search ceph.com/docs <http://ceph.com/docs> for 'incomplete')
>>>>>>>>> pg 2.7 is incomplete, acting [12,2,10,7,2147483647,2147483647,2147483647,2147483647] (reducing pool ecpool min_size from 6 may help; search ceph.com/docs <http://ceph.com/docs> for 'incomplete')
>>>>>>>>> pg 2.6 is incomplete, acting [9,15,2147483647,4,2,2147483647,2147483647,2147483647] (reducing pool ecpool min_size from 6 may help; search ceph.com/docs <http://ceph.com/docs> for 'incomplete')
>>>>>>>>> pg 2.1 is incomplete, acting [2,4,2147483647,13,2147483647,10,2147483647,2147483647] (reducing pool ecpool min_size from 6 may help; search ceph.com/docs <http://ceph.com/docs> for 'incomplete')
>>>>>>>>> pg 2.0 is incomplete, acting [14,1,2147483647,4,10,2147483647,2147483647,2147483647] (reducing pool ecpool min_size from 6 may help; search ceph.com/docs <http://ceph.com/docs> for 'incomplete')
>>>>>>>>> pg 2.3 is incomplete, acting [14,11,6,2147483647,2147483647,2147483647,2,2147483647] (reducing pool ecpool min_size from 6 may help; search ceph.com/docs <http://ceph.com/docs> for 'incomplete')
>>>>>>>>> pg 2.2 is incomplete, acting [13,5,11,2147483647,2147483647,3,2147483647,2147483647] (reducing pool ecpool min_size from 6 may help; search ceph.com/docs <http://ceph.com/docs> for 'incomplete')
>>>>>>>>>
>>>>>>>>> At.
>>>>>>>>>
>>>>>>>>> *Italo Santos*
>>>>>>>>> http://italosantos.com.br/
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> _______________________________________________
>>>>>>>>> ceph-users mailing list
>>>>>>>>> ceph-users@xxxxxxxxxxxxxx <mailto:ceph-users@xxxxxxxxxxxxxx>
>>>>>>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>>>>>>
>>>>>>>> -- 
>>>>>>>> Loïc Dachary, Artisan Logiciel Libre
>>>>>>
>>>>>> -- 
>>>>>> Loïc Dachary, Artisan Logiciel Libre
>>>>
>>>> -- 
>>>> Loïc Dachary, Artisan Logiciel Libre
>>
>> -- 
>> Loïc Dachary, Artisan Logiciel Libre
> 

-- 
Loïc Dachary, Artisan Logiciel Libre

Attachment: signature.asc
Description: OpenPGP digital signature

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux