On 8/17/24 4:04 PM, Pavel Begunkov wrote: > On 8/17/24 22:09, Jens Axboe wrote: >> On 8/17/24 3:05 PM, Pavel Begunkov wrote: >>> On 8/17/24 21:20, Jens Axboe wrote: >>>> On 8/17/24 1:44 PM, Pavel Begunkov wrote: >>>>>> This patchset adds a IOURING_ENTER_NO_IOWAIT flag that can be set on >>>>>> enter. If set, then current->in_iowait is not set. By default this flag >>>>> >>>>> A worthwhile change but for _completely_ different reasons. So, first, >>>>> it's v3, not v2, considering the patchset from a couple month ago. And >>>>> since in essence nothing has changed, I can only repeat same points I >>>>> made back then. >>>>> >>>>> The description reads like the flag's purpose is to change accounting, >>>>> and I'm vividly oppose any user exposed (per ring) toggle doing that. >>>>> We don't want the overhead, it's a very confusing feature, and not even >>>> >>>> Come on, there's no overhead associated with this, in practice. >>> >>> Minor, right, ~ 1 "if" or ccmov, but that's for a feature that nobody >>> really cares about and arguably even harmful. That's assuming there >>> will be another flag for performance tuning in the future, at least >>> the description reads like that. >> >> You literally brought up the case where people care yourself in the >> reply, the one from Jamal. And the one I said is probably not the common > > Not really. Yes, by chance high iowait helped to narrow it to the > in_iowait io_uring patch, but that's rather a weird side effect, > not what iowait was "created for", i.e. system is idling because > storage waiting takes too long. It's definitely a smoking gun, if you're already looking for that. But the main question remains - was this an actual power issue or not? Because the answer to that is pretty key imho. > And that "use case" for iowait directly linked to cpufreq, so > if it still counts, then we shouldn't be separating stats from > cpufreq at all. This is what the cpufreq people want to do anyway, so it'll probably happen whether we like it or not. >> case, yet I think we should cater to it as it very well could be legit, >> just in the tiny minority of cases. > > I explained why it's a confusing feature. We can make up some niche > case (with enough of imagination we can justify basically anything), > but I explained why IMHO accounting flag (let's forget about > cpufreq) would have net negative effect. A sysctl knob would be > much more reasonable, but I don't think it's needed at all. The main thing for me is policy vs flexibility. The fact that boost and iowait accounting is currently tied together is pretty ugly and will hopefully go away with my patches. >>>> It's really simple for this stuff - the freq boost is useful (and >>>> needed) for some workloads, and the iowait accounting is never useful >>>> for anything but (currently) comes as an unfortunate side effect of the >>>> former. But even with those two separated, there are still going to be >>>> cases where you want to control when it happens. >>> >>> You can imagine such cases, but in reality I doubt it. If we >>> disable the stat part, nobody would notice as nobody cared for >>> last 3-4 years before in_iowait was added. >> >> That would be ideal. You're saying Jamal's complaint was purely iowait >> based? Because it looked like power concerns to me... If it's just >> iowait, then they just need to stop looking at that, that's pretty >> simple. > > Power consumption, and then, in search of what's wrong, it was > correlated to high iowait as well as difference in C state stats. But this means that it was indeed power consumption, and iowait was just the canary in the coal mine that lead them down the right path. And this in turn means that even with the split, we want to differentiate between short/busty sleeps and longer ones. >>>>> that helpful. iowait is monitored not by the app itself but by someone >>>>> else outside, likely by a different person, and even before trying to >>>>> make sense of numbers the monitor would need to learn first whether >>>>> _any_ program uses io_uring and what flags the application writer >>>>> decided to pass, even more fun when io_uring is used via a 3rd party >>>>> library. >>>>> >>>>> Exactly same patches could make sense if you flip the description >>>>> and say "in_iowait is good for perfomance in some cases but >>>>> degrades power consumption for others, so here is a way to tune >>>>> performance", (just take Jamal's numbers). And that would need to >>>>> clearly state (including man) that the iowait statistic is a side >>>>> effect of it, we don't give it a thought, and the time accounting >>>>> aspect may and hopefully will change in the future. >>>> >>>> Don't disagree with that, the boosting is the primary function here, >>>> iowait accounting is just an odd relic and side effect. >>>> >>>>> Jens, can you remind what happened with separating iowait stats >>>>> vs the optimisation? I believed you sent some patches >>>> >>>> Yep I still have them, and I'll dust them off and resend them. But it >>>> doesn't really change the need for this at all, other than perhaps >>>> wanting to rename the actual flag as it would not be about iowait at >>>> all, it'd just be about power consumption. >>>> >>>> Last cut was here: >>>> >>>> https://git.kernel.dk/cgit/linux/log/?h=iowait.2 >>>> >>>> just need simple rebasing. >>> >>> IMHO we should try to merge it first. And if there would be >>> some friction, we can get back and take this patch, with >>> additional note describing about iowait stat side effects. >> >> I did just rebase on top of sched/core and sent out a new version. I >> still think it makes sense regardless of this patch or not, only >> difference would be what you call the flag. > > There shouldn't be any difference in how it's called or how > it's recommended to be used, but maybe I don't understand > what you mean. As I see it, the difference in a blob of > text in the man mentioning the iowait stat side effect. See above. >> It's here: >> >> https://lore.kernel.org/lkml/20240817204639.132794-1-axboe@xxxxxxxxx/T/#mab3638659a54193927351ebdc5e345278637bc41 > > Perfect > > >>>>>> is not set to maintain existing behaviour i.e. in_iowait is always set. >>>>>> This is to prevent waiting for completions being accounted as CPU >>>>>> utilisation. >>>>> >>>>> For accounting, it's more reasonable to keep it disabled by >>>>> default, so we stop getting millions complaints per day about >>>>> high iowait. >>>> >>>> Agree, it should just go away. >>>> >>>>>> Not setting in_iowait does mean that we also lose cpufreq optimisations >>>>>> above because in_iowait semantics couples 1 and 2 together. Eventually >>>>>> we will untangle the two so the optimisations can be enabled >>>>>> independently of the accounting. >>>>>> >>>>>> IORING_FEAT_IOWAIT_TOGGLE is returned in io_uring_create() to indicate >>>>>> support. This will be used by liburing to check for this feature. >>>>>> Signed-off-by: David Wei <dw@xxxxxxxxxxx> >>>>>> --- >>>>>> v2: >>>>>> - squash patches into one >>>>>> - move no_iowait in struct io_wait_queue to the end >>>>>> - always set iowq.no_iowait >>>>>> >>>>>> --- >>>>>> include/uapi/linux/io_uring.h | 2 ++ >>>>>> io_uring/io_uring.c | 7 ++++--- >>>>>> io_uring/io_uring.h | 1 + >>>>>> 3 files changed, 7 insertions(+), 3 deletions(-) >>>>>> >>>>>> diff --git a/include/uapi/linux/io_uring.h b/include/uapi/linux/io_uring.h >>>>>> index 48c440edf674..3a94afa8665e 100644 >>>>>> --- a/include/uapi/linux/io_uring.h >>>>>> +++ b/include/uapi/linux/io_uring.h >>>>>> @@ -508,6 +508,7 @@ struct io_cqring_offsets { >>>>>> #define IORING_ENTER_EXT_ARG (1U << 3) >>>>>> #define IORING_ENTER_REGISTERED_RING (1U << 4) >>>>>> #define IORING_ENTER_ABS_TIMER (1U << 5) >>>>>> +#define IORING_ENTER_NO_IOWAIT (1U << 6) >>>>> >>>>> Just curious, why did we switch from a register opcode to an >>>>> ENTER flag? >>>> >>>> That was my suggestion, I think it's more flexible in that you can >>>> disable the boost if you know you're doing longer waits or slower IO, >>>> and enable it when frequencies go up. I don't particularly like the >>>> static register approach for this. >>> >>> Makes sense, it's easier for the app to count how many >>> request of what type is inflight. >> >> Exactly >> >>> The name might also be confusing. We need an explanation when >>> it could be useful, and name it accordingly. DEEP/SHALLOW_WAIT? >>> Do you remember how cpufreq accounts for it? >> >> I don't remember how it accounts for it, and was just pondering that >> with the above reply. Because if it just decays the sleep state, then >> you could just use it generically. If it stays high regardless of how >> long you wait, then it could be a power issue. Not on servers really >> (well a bit, depending on boosting), but more so on desktop apps. >> Laptops tend to be pretty power conservative! > > {SHORT,BRIEF}/LONG_WAIT maybe? I think that's a lot more descriptive. Ideally we'd want to tie this to wakeup latencies, eg we'd need to know about wakeup latencies. For example, if the user asks for a 100 usec wait, we'd want to influence what sleep state is picked in propagating that information. Things like the min-wait I posted would directly work for that, as it tells the story in two chapters on what waits we're expecting here. Currently there's no way to do that (hence iowait -> cpufreq boosting), but there clearly should be. Or even without min-wait, the timeout is clearly known here, combined with the expected/desired number of events the application is looking for. -- Jens Axboe