sched/fair: Fix cfs_rq_clock_pelt() for throttled cfs_rq
[ Upstream commit 64eaf50731ac0a8c76ce2fedd50ef6652aabc5ff ] Since commit2312729688
("sched/fair: Update scale invariance of PELT") change to use rq_clock_pelt() instead of rq_clock_task(), we should also use rq_clock_pelt() for throttled_clock_task_time and throttled_clock_task accounting to get correct cfs_rq_clock_pelt() of throttled cfs_rq. And rename throttled_clock_task(_time) to be clock_pelt rather than clock_task. Fixes:2312729688
("sched/fair: Update scale invariance of PELT") Signed-off-by: Chengming Zhou <zhouchengming@bytedance.com> Signed-off-by: Peter Zijlstra (Intel) <peterz@infradead.org> Reviewed-by: Ben Segall <bsegall@google.com> Reviewed-by: Vincent Guittot <vincent.guittot@linaro.org> Link: https://lore.kernel.org/r/20220408115309.81603-1-zhouchengming@bytedance.com Signed-off-by: Sasha Levin <sashal@kernel.org>
This commit is contained in:
parent
55fddbb1e2
commit
735b57a960
@ -4485,8 +4485,8 @@ static int tg_unthrottle_up(struct task_group *tg, void *data)
|
|||||||
|
|
||||||
cfs_rq->throttle_count--;
|
cfs_rq->throttle_count--;
|
||||||
if (!cfs_rq->throttle_count) {
|
if (!cfs_rq->throttle_count) {
|
||||||
cfs_rq->throttled_clock_task_time += rq_clock_task(rq) -
|
cfs_rq->throttled_clock_pelt_time += rq_clock_pelt(rq) -
|
||||||
cfs_rq->throttled_clock_task;
|
cfs_rq->throttled_clock_pelt;
|
||||||
|
|
||||||
/* Add cfs_rq with already running entity in the list */
|
/* Add cfs_rq with already running entity in the list */
|
||||||
if (cfs_rq->nr_running >= 1)
|
if (cfs_rq->nr_running >= 1)
|
||||||
@ -4503,7 +4503,7 @@ static int tg_throttle_down(struct task_group *tg, void *data)
|
|||||||
|
|
||||||
/* group is entering throttled state, stop time */
|
/* group is entering throttled state, stop time */
|
||||||
if (!cfs_rq->throttle_count) {
|
if (!cfs_rq->throttle_count) {
|
||||||
cfs_rq->throttled_clock_task = rq_clock_task(rq);
|
cfs_rq->throttled_clock_pelt = rq_clock_pelt(rq);
|
||||||
list_del_leaf_cfs_rq(cfs_rq);
|
list_del_leaf_cfs_rq(cfs_rq);
|
||||||
}
|
}
|
||||||
cfs_rq->throttle_count++;
|
cfs_rq->throttle_count++;
|
||||||
@ -4932,7 +4932,7 @@ static void sync_throttle(struct task_group *tg, int cpu)
|
|||||||
pcfs_rq = tg->parent->cfs_rq[cpu];
|
pcfs_rq = tg->parent->cfs_rq[cpu];
|
||||||
|
|
||||||
cfs_rq->throttle_count = pcfs_rq->throttle_count;
|
cfs_rq->throttle_count = pcfs_rq->throttle_count;
|
||||||
cfs_rq->throttled_clock_task = rq_clock_task(cpu_rq(cpu));
|
cfs_rq->throttled_clock_pelt = rq_clock_pelt(cpu_rq(cpu));
|
||||||
}
|
}
|
||||||
|
|
||||||
/* conditionally throttle active cfs_rq's from put_prev_entity() */
|
/* conditionally throttle active cfs_rq's from put_prev_entity() */
|
||||||
|
@ -127,9 +127,9 @@ static inline u64 rq_clock_pelt(struct rq *rq)
|
|||||||
static inline u64 cfs_rq_clock_pelt(struct cfs_rq *cfs_rq)
|
static inline u64 cfs_rq_clock_pelt(struct cfs_rq *cfs_rq)
|
||||||
{
|
{
|
||||||
if (unlikely(cfs_rq->throttle_count))
|
if (unlikely(cfs_rq->throttle_count))
|
||||||
return cfs_rq->throttled_clock_task - cfs_rq->throttled_clock_task_time;
|
return cfs_rq->throttled_clock_pelt - cfs_rq->throttled_clock_pelt_time;
|
||||||
|
|
||||||
return rq_clock_pelt(rq_of(cfs_rq)) - cfs_rq->throttled_clock_task_time;
|
return rq_clock_pelt(rq_of(cfs_rq)) - cfs_rq->throttled_clock_pelt_time;
|
||||||
}
|
}
|
||||||
#else
|
#else
|
||||||
static inline u64 cfs_rq_clock_pelt(struct cfs_rq *cfs_rq)
|
static inline u64 cfs_rq_clock_pelt(struct cfs_rq *cfs_rq)
|
||||||
|
@ -570,8 +570,8 @@ struct cfs_rq {
|
|||||||
s64 runtime_remaining;
|
s64 runtime_remaining;
|
||||||
|
|
||||||
u64 throttled_clock;
|
u64 throttled_clock;
|
||||||
u64 throttled_clock_task;
|
u64 throttled_clock_pelt;
|
||||||
u64 throttled_clock_task_time;
|
u64 throttled_clock_pelt_time;
|
||||||
int throttled;
|
int throttled;
|
||||||
int throttle_count;
|
int throttle_count;
|
||||||
struct list_head throttled_list;
|
struct list_head throttled_list;
|
||||||
|
Loading…
Reference in New Issue
Block a user