On Thu, 29 Dec 2016, Łukasz Chrustek wrote: > Cześć, > > > >> >> > >> >> here i press ctrl+c > >> >> > >> >> 2016-12-28 17:38:10.055239 7f4576d7a700 0 monclient: hunting for new mon > >> >> 2016-12-28 17:38:10.055582 7f4574233700 0 -- 192.168.128.1:0/1201679761 >> 192.168.128.2:6789/0 pipe(0x7f456c023190 sd=3 :0 s=1 pgs=0 cs=0 l=1 c=0x7f456c024470).fault > >> >> 2016-12-28 17:38:30.550622 7f4574233700 0 -- 192.168.128.1:0/1201679761 >> 192.168.128.1:6789/0 pipe(0x7f45600008c0 sd=3 :0 s=1 pgs=0 cs=0 l=1 c=0x7f4560001df0).fault > >> >> 2016-12-28 17:38:54.551031 7f4574474700 0 -- 192.168.128.1:0/1201679761 >> 192.168.128.2:6789/0 pipe(0x7f45600046c0 sd=4 :0 s=1 pgs=0 cs=0 l=1 c=0x7f45600042b0).fault > >> > >> > What version is this? > >> > >> infernalis > >> > >> > Can you attach the crush map too? (ceph osd crush dump -f json-pretty) > >> > >> I can't - ceph-mons are crushing on diffrent ceph-mon hosts: > >> > >> ]# ceph osd crush dump -f json-pretty > > > Hmm, in that case, 'ceph osd getcrushmap -o cm' and post that somewhere? > > In attachment. The encoded tree bucket -11 had bad values. I don't really trust the tree bucket code in crush... it's not well tested (and is a poor balance computation and efficiency anyway). We should probably try to remove tree entirely. I've attached a fixed map that you can inject with ceph osd setcrushmap -i <filename> Bucket -11 is now empty; not sure what was supposed to be in it. I suggest switching all of your tree buckets over to straw2 as soon as possible. Note that this will result in some rebalancing. You could do it one bucket a time if that's concerning. sage
Attachment:
cmx2
Description: Binary data