On Mon 16-04-18 17:23:30, Sasha Levin wrote: > On Mon, Apr 16, 2018 at 07:06:04PM +0200, Pavel Machek wrote: > >On Mon 2018-04-16 16:37:56, Sasha Levin wrote: > >> On Mon, Apr 16, 2018 at 12:30:19PM -0400, Steven Rostedt wrote: > >> >On Mon, 16 Apr 2018 16:19:14 +0000 > >> >Sasha Levin <Alexander.Levin@xxxxxxxxxxxxx> wrote: > >> > > >> >> >Wait! What does that mean? What's the purpose of stable if it is as > >> >> >broken as mainline? > >> >> > >> >> This just means that if there is a fix that went in mainline, and the > >> >> fix is broken somehow, we'd rather take the broken fix than not. > >> >> > >> >> In this scenario, *something* will be broken, it's just a matter of > >> >> what. We'd rather have the same thing broken between mainline and > >> >> stable. > >> > > >> >Honestly, I think that removes all value of the stable series. I > >> >remember when the stable series were first created. People were saying > >> >that it wouldn't even get to more than 5 versions, because the bar for > >> >backporting was suppose to be very high. Today it's just a fork of the > >> >kernel at a given version. No more features, but we will be OK with > >> >regressions. I'm struggling to see what the benefit of it is suppose to > >> >be? > >> > >> It's not "OK with regressions". > >> > >> Let's look at a hypothetical example: You have a 4.15.1 kernel that has > >> a broken printf() behaviour so that when you: > >> > >> pr_err("%d", 5) > >> > >> Would print: > >> > >> "Microsoft Rulez" > >> > >> Bad, right? So you went ahead and fixed it, and now it prints "5" as you > >> might expect. But alas, with your patch, running: > >> > >> pr_err("%s", "hi!") > >> > >> Would show a cat picture for 5 seconds. > >> > >> Should we take your patch in -stable or not? If we don't, we're stuck > >> with the original issue while the mainline kernel will behave > >> differently, but if we do - we introduce a new regression. > > > >Of course not. > > > >- It must be obviously correct and tested. > > > >If it introduces new bug, it is not correct, and certainly not > >obviously correct. > > As you might have noticed, we don't strictly follow the rules. > > Take a look at the whole PTI story as an example. It's way more than 100 > lines, it's not obviously corrent, it fixed more than 1 thing, and so > on, and yet it went in -stable! > > Would you argue we shouldn't have backported PTI to -stable? So I agree with that being backported. But I think this nicely demostrates a point some people are trying to make in this thread. We do take fixes with high risk or regression if they fix serious enough issue. Also we do take fixes to non-serious stuff (such as addition of device ID) if the chances of regression are really low. So IMHO the metric for including the fix is not solely "how annoying to user this can be" but rather something like: score = (how annoying the bug is) * ((1 / (chance of regression due to including this)) - 1)^3 (constants are somewhat arbitrary subject to tuning ;). Now both 'annoying' and 'regression chance' parts are subjective and sometimes difficult to estimate so don't take the formula too seriously but it demonstrates the point. I think we all agree we want to fix annoying stuff and we don't want regressions. But you need to somehow weight this over your expected userbase - and this is where your argument "but someone might be annoyed by LEDs not working so let's include it" has problems - it should rather be "is the annoyance of non-working leds over expected user base high enough to risk a regression due to this patch for someone in the expected user base"? The answer to this second question is not clear at all to a casual reviewer and that's why we IMHO have CC stable tag as maintainer is supposed to have at least a bit better clue. Another point I wanted to make is that if chance a patch causes a regression is about 2% as you said somewhere else in a thread, then by adding 20 patches that "may fix a bug that is annoying for someone" you've just increased a chance there's a regression in the release by 34%. And this is not just a math game, this also roughly matches a real experience with maintaining our enterprise kernels. Do 20 "maybe" fixes outweight such regression chance? And I also note that for a regression to get reported so that it gets included into your 2% estimate of a patch regression rate, someone must be bothered enough by it to triage it and send an email somewhere so that already falls into a category of "serious" stuff to me. So these are the reasons why I think that merging tons of patches into stable isn't actually very good. Honza -- Jan Kara <jack@xxxxxxxx> SUSE Labs, CR