On Thu, May 10, 2018 at 09:41:32AM -0400, Johannes Weiner wrote: > So there is a reason I'm tracking productivity states per-cpu and not > globally. Consider the following example periods on two CPUs: > > CPU 0 > Task 1: | EXECUTING | memstalled | > Task 2: | runqueued | EXECUTING | > > CPU 1 > Task 3: | memstalled | EXECUTING | > > If we tracked only the global number of stalled tasks, similarly to > nr_uninterruptible, the number would be elevated throughout the whole > sampling period, giving a pressure value of 100% for "some stalled". > And, since there is always something executing, a "full stall" of 0%. But if you read the comment about SMP IO-wait; see commit: e33a9bba85a8 ("sched/core: move IO scheduling accounting from io_schedule_timeout() into scheduler") you'll see that per-cpu accounting has issues too. Also, note that in your example above you have 1 memstalled task (at any one time), but _2_ CPUs. So at most you should end up with a 50% value. There is no way 1 task could consume 2 CPUs worth of time. Furthermore, associating a blocked task to any particular CPU is fundamentally broken and I'll hard NAK anything that relies on it. > Now consider what happens when the Task 3 sequence is the other way > around: > > CPU 0 > Task 1: | EXECUTING | memstalled | > Task 2: | runqueued | EXECUTING | > > CPU 1 > Task 3: | EXECUTING | memstalled | > > Here the number of stalled tasks is elevated only during half of the > sampling period, this time giving a pressure reading of 50% for "some" > (and again 0% for "full"). That entirely depends on your averaging; an exponentially decaying average would not typically result in 50% for the above case. But I think we can agree that this results in one 0% and one 100% sample -- we have two stalled tasks and two CPUs. > That's a different measurement, but in terms of workload progress, the > sequences are functionally equivalent. In both scenarios the same > amount of productive CPU cycles is spent advancing tasks 1, 2 and 3, > and the same amount of potentially productive CPU time is lost due to > the contention of memory. We really ought to read the same pressure. And you do -- subject to the averaging used, as per the above. The first gives two 50% samples, the second gives 0%, 100%. > So what I'm doing is calculating the productivity loss on each CPU in > a sampling period as if they were independent time slices. It doesn't > matter how you slice and dice the sequences within each one - if used > CPU time and lost CPU time have the same proportion, we have the same > pressure. I'm still thinking you can do basically the same without the stong CPU relation. > To illustrate: > > CPU X > 1 2 3 4 > Task 1: | EXECUTING | memstalled | sleeping | sleeping | > Task 2: | runqueued | EXECUTING | sleeping | sleeping | > Task 3: | sleeping | sleeping | EXECUTING | memstalled | > > You can clearly see the 50% of walltime in which *somebody* isn't > advancing (2 and 4), and the 25% of walltime in which *no* tasks are > (3). Same amount of work, same memory stalls, same pressure numbers. > > Globalized state tracking would produce those numbers on the single > CPU (obviously), but once concurrency gets into the mix, it's > questionable what its results mean. It certainly isn't able to > reliably detect equivalent slowdowns of individual tasks ("some" is > all over the place), and in this example wasn't able to capture the > impact of contention on overall work completion ("full" is 0%). > > * CPU 0: some = 50%, full = 0% > CPU 1: some = 50%, full = 50% > avg: some = 50%, full = 25% I'm not entirely sure I get your point here; but note that a task doesn't sleep on a CPU. When it sleeps it is not strictly associated with a CPU, only when it runs does it have an association. What is the value of accounting a sleep state to a particular CPU if the task when wakes up on another? Where did the sleep take place? All we really can say is that a task slept, and if we can reduce the reason for its sleeping (IO, reclaim, whatever) then it could've ran sooner. And then you can make predictions based on the number of CPUs and global idle time, how much that could improve things.