On Mon, 22 Feb 2016 22:30:17 +0100 Peter Zijlstra <peterz@xxxxxxxxxxxxx> wrote: > On Mon, Feb 22, 2016 at 12:48:54PM -0500, Steven Rostedt wrote: > > > > As it stands, the existing tracepoint have already been an ABI > > > trainwreck, why would I want to add more? > > > > Yes, this may become a type of ABI, but even the sched switch > > tracepoints haven't been that bad. Has it really prevented us from > > changing anything? > > The whole wakeup thing where we _still_ have a dummy argument, and have > been lying about the value for a long time really stinks. Having a dummy argument is not that bad. Yes, it's inconvenient, and I'm not sure who even uses it (can we delete it without breaking anything?) But it doesn't prevent us from going forward with development. > > > But let me ask, what would you recommend to finding out if the kernel > > has really given your tasks the recommended runtime within a given > > period? We can't expect users of SCHED_DEADLINE to be modifying the > > kernel themselves. > > So why are these deadline specific tracepoint? Why not extend the ones > we already have? I'm not sure how to do that and be able to report when a period has elapsed, and when the next period is coming. > > Also, will these tracepoints still work if we implement SCHED_DEADLINE > using Least-Laxity-First or Pfair or some other exotic algorithm? Or > will be forever be bound to EDF just because tracepoint ABI shite? Can we come up with generic numbers? I mean, the user that asks for their task to have a specific runtime within a specific period/deadline, these seem to be generic already. I'll have to read up on those that you mention, but do that not have a "replenish" for when the period starts again? And then a yield, showing the task has given up its remaining time, or a block, where a task is scheduled out because it blocked on a lock? > > Worse, the proposed tracepoints are atrocious, look at crap like this: > > > + if (trace_sched_deadline_yield_enabled()) { > > + u64 delta_exec = rq_clock_task(rq) - p->se.exec_start; > > + /* Subtract the last run till now */ > > + if (likely((s64)delta_exec > 0)) > > + p->dl.runtime -= delta_exec; > > + trace_sched_deadline_yield(&p->dl); > > + } > > tracepoints should _NEVER_ change state, ever. Heh, it's not really changing state. The code directly after this is: p->dl.runtime = 0; Without updating dl.runtime, you the tracepoint would report inaccurate remaining time. Without that, you would get reports of yielding with full runtimes, making it look like you never ran at all. > > And there's the whole COND tracepoint muck, which also doesn't win any > prices. It keeps us from adding flags that may end up going away and us maintaining a dummy field forever ;-) > > So tell me why these specific tracepoints and why the existing ones > could not be extended to include this information. For example, why a > trace_sched_dealine_yield, and not a generic trace_sched_yield() that > works for all classes. But what about reporting actual runtime, and when the next period will come. That only matters for deadline. > > Tell me that the information presented does not pin the implementation. > > And clean up the crap. > > Then I might maybe consider this. > > But do not present me with a bunch of random arse, hacked together > tracepoints and tell me they might be useful, maybe. They ARE useful. These are the tracepoints I'm currently using to debug the deadline scheduler with. They have been indispensable for my current work. -- Steve -- To unsubscribe from this list: send the line "unsubscribe linux-rt-users" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html