Hi, Thank You for suggestion - I will read about it. Regards Luk > Have you thought through your use of '"min_size": 1' on many of these pools? > See recent discussions on this mailing list on the subject for background. > Just FYI. > On Fri, Dec 30, 2016 at 8:18 AM, Sage Weil <sage@xxxxxxxxxxxx> wrote: >> On Thu, 29 Dec 2016, Łukasz Chrustek wrote: >>> Hi, >>> >>> Thank You very much for analize and the file ! I had similar :) but >>> wasn't sure if it wan't distroy something in cluster. >>> >>> > The encoded tree bucket -11 had bad values. I don't really trust the tree >>> > bucket code in crush... it's not well tested (and is a poor balance >>> > computation and efficiency anyway). We should probably try to remove tree >>> > entirely. >>> >>> > I've attached a fixed map that you can inject with >>> >>> > ceph osd setcrushmap -i <filename> >>> >>> Now it works, and also ceph osd crush dump -f json-pretty runs OK. >> >> Great news! >> >>> > Bucket -11 is now empty; not sure what was supposed to be in it. >>> >>> this server will be reinstalled, there where three osds. >>> >>> > I suggest switching all of your tree buckets over to straw2 as soon as >>> > possible. Note that this will result in some rebalancing. You could do >>> > it one bucket a time if that's concerning. >>> >>> OK, changing alg to straw2 will rebalance ale PGs on all nodes ? >> >> For any bucket you change from tree -> straw2, you'll see PGs shuffle >> between the children of that bucket. So for hosts, you'll see data move >> between the disks. And fixing ssd-intel-s3700 will shuffle data between >> hosts. >> >> I'd also switch the straw buckets to straw2, although that will move a >> comparatively small amount of data. >> >> sage -- Pozdrowienia, Łukasz Chrustek -- To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html