Re: [PATCH v7 2/2] sched/fair: update scale invariance of PELT

From: Patrick Bellasi
Date: Thu Jan 24 2019 - 09:04:40 EST


On 24-Jan 10:07, Peter Zijlstra wrote:
>
> Sorry; trying to get back to this and re-reading the old conversations.
>
> On Thu, Nov 29, 2018 at 03:13:16PM +0000, Patrick Bellasi wrote:
> > On 29-Nov 13:53, Peter Zijlstra wrote:
> > > On Wed, Nov 28, 2018 at 11:53:36AM +0000, Patrick Bellasi wrote:
> > >
> > > > diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c
> > > > index ac855b2f4774..93e0cf5d8a76 100644
> > > > --- a/kernel/sched/fair.c
> > > > +++ b/kernel/sched/fair.c
> > > > @@ -3661,6 +3661,10 @@ util_est_dequeue(struct cfs_rq *cfs_rq, struct task_struct *p, bool task_sleep)
> > > > if (!task_sleep)
> > > > return;
> > > >
> > > > + /* Skip samples which do not represent an actual utilization */
> > > > + if (unlikely(task_util(p) > capacity_of(task_cpu(p))))
> > > > + return;
> > > > +
> > > > /*
> > > > * If the PELT values haven't changed since enqueue time,
> > > > * skip the util_est update.
> > >
> > > Would you not want something like:
> > >
> > > min(task_util(p), capacity_of(task_cpu(p)))
> > >
> > > And is this the only place where we need this?
> >
> > Mmm... even this could be an over-estimation:
> >
> > I've just posted an example in my last reply to Vincent, end of:
> >
> > Message-ID: <20181129150020.GF23094@e110439-lin>
> > https://lore.kernel.org/lkml/20181129150020.GF23094@e110439-lin/
>
> In particular this bit:
>
> | Seems we agree that, when there is no idle time:
> | - the two 15% tasks will be overestimated
> | - their utilization will reach 50% after a while
>
> Right?
>
> > > OTOH, if the task is always running, it will be always running
> > > irrespective of where it runs.
> >
> > That's not what I'm concerned about. I'm concerned about small tasks
> > which are running on limited capacity (e.g. due to thermal capping)
> > without idle time. In this case, the new "utilization" signal could
> > overestimate the real task needs.
> >
> > > Not storing these samples seems weird though; this is the exact
> > > condition you want to record -- the task is very active, if we skip
> > > these, we'll come back at a low frequency on the next wakeup.
> >
> > When there is not idle time, we don't know if the reported
> > utilization, above the cpu capacity, is due to the task being bigger...
> > or just the new utilization signal converging towards:
> >
> > 100% / RUNNABLE_TASKS_COUNT
>
> So if I'm not mistaken we then have 3 cases:
>
> 1) runnable == util <= capacity
>
> no contention, idle
>
> 2) runnable == util > capacity
>
> no contention, no idle
>
> 3) runnable > util
>
> contention, no idle
>
> For 1) we can use: 'util'
> For 2) we can use: 'capacity'
> For 3) we can use: 'util * capacity >> 10'
>
> (note that 2 is a special case of 3 when u=1)
>
> This should work right?

I think there is a case, similar to 2, in which the new 'util' could
potentially be used. That's the case for example of a 20% (estimated)
utilization task running alone on a 15% capacity CPU, for a single
activation. In that case such a task will complete and be dequeued
with:

runnable == util > capacity

The problem is that we need to be sure there was not contention... and
that seems to be difficult to detect.

> Now, instead of doing complicated things like that, you instead figure
> that when there's no idle there's also no dequeue happening and we can
> simply short-cut by skipping the entire thing, forgetting everything
> about 2,3.
>
> Did I get that right?

More or less... just saying that 1 is the only easy to detect scenario
in which we are granted the utilization represents an actual bandwidth
request and thus the only safe values to sample for estimated
utilization. For the other cases, since anyway:

util_est := max(max(ewma, last_util), util_avg)

util_est will just keep representing a safe and actually measured
lower-bound for the expected utilization of a task, without
side-affecting the EWMA which has a "slow" update dynamic.

--
#include <best/regards.h>

Patrick Bellasi