|
|
|
|
@@ -2200,7 +2200,7 @@ static void update_numa_stats(struct task_numa_env *env,
|
|
|
|
|
ns->load += cpu_load(rq);
|
|
|
|
|
ns->runnable += cpu_runnable(rq);
|
|
|
|
|
ns->util += cpu_util_cfs(cpu);
|
|
|
|
|
ns->nr_running += rq->cfs.h_nr_queued;
|
|
|
|
|
ns->nr_running += rq->cfs.h_nr_running;
|
|
|
|
|
ns->compute_capacity += capacity_of(cpu);
|
|
|
|
|
|
|
|
|
|
if (find_idle && idle_core < 0 && !rq->nr_running && idle_cpu(cpu)) {
|
|
|
|
|
@@ -5523,7 +5523,7 @@ enqueue_entity(struct cfs_rq *cfs_rq, struct sched_entity *se, int flags)
|
|
|
|
|
* When enqueuing a sched_entity, we must:
|
|
|
|
|
* - Update loads to have both entity and cfs_rq synced with now.
|
|
|
|
|
* - For group_entity, update its runnable_weight to reflect the new
|
|
|
|
|
* h_nr_queued of its group cfs_rq.
|
|
|
|
|
* h_nr_running of its group cfs_rq.
|
|
|
|
|
* - For group_entity, update its weight to reflect the new share of
|
|
|
|
|
* its group cfs_rq
|
|
|
|
|
* - Add its new weight to cfs_rq->load.weight
|
|
|
|
|
@@ -5680,7 +5680,7 @@ dequeue_entity(struct cfs_rq *cfs_rq, struct sched_entity *se, int flags)
|
|
|
|
|
* When dequeuing a sched_entity, we must:
|
|
|
|
|
* - Update loads to have both entity and cfs_rq synced with now.
|
|
|
|
|
* - For group_entity, update its runnable_weight to reflect the new
|
|
|
|
|
* h_nr_queued of its group cfs_rq.
|
|
|
|
|
* h_nr_running of its group cfs_rq.
|
|
|
|
|
* - Subtract its previous weight from cfs_rq->load.weight.
|
|
|
|
|
* - For group entity, update its weight to reflect the new share
|
|
|
|
|
* of its group cfs_rq.
|
|
|
|
|
@@ -6083,8 +6083,8 @@ static bool throttle_cfs_rq(struct cfs_rq *cfs_rq)
|
|
|
|
|
struct rq *rq = rq_of(cfs_rq);
|
|
|
|
|
struct cfs_bandwidth *cfs_b = tg_cfs_bandwidth(cfs_rq->tg);
|
|
|
|
|
struct sched_entity *se;
|
|
|
|
|
long queued_delta, idle_task_delta, delayed_delta, dequeue = 1;
|
|
|
|
|
long rq_h_nr_queued = rq->cfs.h_nr_queued;
|
|
|
|
|
long task_delta, idle_task_delta, delayed_delta, dequeue = 1;
|
|
|
|
|
long rq_h_nr_running = rq->cfs.h_nr_running;
|
|
|
|
|
|
|
|
|
|
raw_spin_lock(&cfs_b->lock);
|
|
|
|
|
/* This will start the period timer if necessary */
|
|
|
|
|
@@ -6114,7 +6114,7 @@ static bool throttle_cfs_rq(struct cfs_rq *cfs_rq)
|
|
|
|
|
walk_tg_tree_from(cfs_rq->tg, tg_throttle_down, tg_nop, (void *)rq);
|
|
|
|
|
rcu_read_unlock();
|
|
|
|
|
|
|
|
|
|
queued_delta = cfs_rq->h_nr_queued;
|
|
|
|
|
task_delta = cfs_rq->h_nr_running;
|
|
|
|
|
idle_task_delta = cfs_rq->idle_h_nr_running;
|
|
|
|
|
delayed_delta = cfs_rq->h_nr_delayed;
|
|
|
|
|
for_each_sched_entity(se) {
|
|
|
|
|
@@ -6136,9 +6136,9 @@ static bool throttle_cfs_rq(struct cfs_rq *cfs_rq)
|
|
|
|
|
dequeue_entity(qcfs_rq, se, flags);
|
|
|
|
|
|
|
|
|
|
if (cfs_rq_is_idle(group_cfs_rq(se)))
|
|
|
|
|
idle_task_delta = cfs_rq->h_nr_queued;
|
|
|
|
|
idle_task_delta = cfs_rq->h_nr_running;
|
|
|
|
|
|
|
|
|
|
qcfs_rq->h_nr_queued -= queued_delta;
|
|
|
|
|
qcfs_rq->h_nr_running -= task_delta;
|
|
|
|
|
qcfs_rq->idle_h_nr_running -= idle_task_delta;
|
|
|
|
|
qcfs_rq->h_nr_delayed -= delayed_delta;
|
|
|
|
|
|
|
|
|
|
@@ -6159,18 +6159,18 @@ static bool throttle_cfs_rq(struct cfs_rq *cfs_rq)
|
|
|
|
|
se_update_runnable(se);
|
|
|
|
|
|
|
|
|
|
if (cfs_rq_is_idle(group_cfs_rq(se)))
|
|
|
|
|
idle_task_delta = cfs_rq->h_nr_queued;
|
|
|
|
|
idle_task_delta = cfs_rq->h_nr_running;
|
|
|
|
|
|
|
|
|
|
qcfs_rq->h_nr_queued -= queued_delta;
|
|
|
|
|
qcfs_rq->h_nr_running -= task_delta;
|
|
|
|
|
qcfs_rq->idle_h_nr_running -= idle_task_delta;
|
|
|
|
|
qcfs_rq->h_nr_delayed -= delayed_delta;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* At this point se is NULL and we are at root level*/
|
|
|
|
|
sub_nr_running(rq, queued_delta);
|
|
|
|
|
sub_nr_running(rq, task_delta);
|
|
|
|
|
|
|
|
|
|
/* Stop the fair server if throttling resulted in no runnable tasks */
|
|
|
|
|
if (rq_h_nr_queued && !rq->cfs.h_nr_queued)
|
|
|
|
|
if (rq_h_nr_running && !rq->cfs.h_nr_running)
|
|
|
|
|
dl_server_stop(&rq->fair_server);
|
|
|
|
|
done:
|
|
|
|
|
/*
|
|
|
|
|
@@ -6189,8 +6189,8 @@ void unthrottle_cfs_rq(struct cfs_rq *cfs_rq)
|
|
|
|
|
struct rq *rq = rq_of(cfs_rq);
|
|
|
|
|
struct cfs_bandwidth *cfs_b = tg_cfs_bandwidth(cfs_rq->tg);
|
|
|
|
|
struct sched_entity *se;
|
|
|
|
|
long queued_delta, idle_task_delta, delayed_delta;
|
|
|
|
|
long rq_h_nr_queued = rq->cfs.h_nr_queued;
|
|
|
|
|
long task_delta, idle_task_delta, delayed_delta;
|
|
|
|
|
long rq_h_nr_running = rq->cfs.h_nr_running;
|
|
|
|
|
|
|
|
|
|
se = cfs_rq->tg->se[cpu_of(rq)];
|
|
|
|
|
|
|
|
|
|
@@ -6223,7 +6223,7 @@ void unthrottle_cfs_rq(struct cfs_rq *cfs_rq)
|
|
|
|
|
goto unthrottle_throttle;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
queued_delta = cfs_rq->h_nr_queued;
|
|
|
|
|
task_delta = cfs_rq->h_nr_running;
|
|
|
|
|
idle_task_delta = cfs_rq->idle_h_nr_running;
|
|
|
|
|
delayed_delta = cfs_rq->h_nr_delayed;
|
|
|
|
|
for_each_sched_entity(se) {
|
|
|
|
|
@@ -6239,9 +6239,9 @@ void unthrottle_cfs_rq(struct cfs_rq *cfs_rq)
|
|
|
|
|
enqueue_entity(qcfs_rq, se, ENQUEUE_WAKEUP);
|
|
|
|
|
|
|
|
|
|
if (cfs_rq_is_idle(group_cfs_rq(se)))
|
|
|
|
|
idle_task_delta = cfs_rq->h_nr_queued;
|
|
|
|
|
idle_task_delta = cfs_rq->h_nr_running;
|
|
|
|
|
|
|
|
|
|
qcfs_rq->h_nr_queued += queued_delta;
|
|
|
|
|
qcfs_rq->h_nr_running += task_delta;
|
|
|
|
|
qcfs_rq->idle_h_nr_running += idle_task_delta;
|
|
|
|
|
qcfs_rq->h_nr_delayed += delayed_delta;
|
|
|
|
|
|
|
|
|
|
@@ -6257,9 +6257,9 @@ void unthrottle_cfs_rq(struct cfs_rq *cfs_rq)
|
|
|
|
|
se_update_runnable(se);
|
|
|
|
|
|
|
|
|
|
if (cfs_rq_is_idle(group_cfs_rq(se)))
|
|
|
|
|
idle_task_delta = cfs_rq->h_nr_queued;
|
|
|
|
|
idle_task_delta = cfs_rq->h_nr_running;
|
|
|
|
|
|
|
|
|
|
qcfs_rq->h_nr_queued += queued_delta;
|
|
|
|
|
qcfs_rq->h_nr_running += task_delta;
|
|
|
|
|
qcfs_rq->idle_h_nr_running += idle_task_delta;
|
|
|
|
|
qcfs_rq->h_nr_delayed += delayed_delta;
|
|
|
|
|
|
|
|
|
|
@@ -6269,11 +6269,11 @@ void unthrottle_cfs_rq(struct cfs_rq *cfs_rq)
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Start the fair server if un-throttling resulted in new runnable tasks */
|
|
|
|
|
if (!rq_h_nr_queued && rq->cfs.h_nr_queued)
|
|
|
|
|
if (!rq_h_nr_running && rq->cfs.h_nr_running)
|
|
|
|
|
dl_server_start(&rq->fair_server);
|
|
|
|
|
|
|
|
|
|
/* At this point se is NULL and we are at root level*/
|
|
|
|
|
add_nr_running(rq, queued_delta);
|
|
|
|
|
add_nr_running(rq, task_delta);
|
|
|
|
|
|
|
|
|
|
unthrottle_throttle:
|
|
|
|
|
assert_list_leaf_cfs_rq(rq);
|
|
|
|
|
@@ -6988,7 +6988,7 @@ static void hrtick_start_fair(struct rq *rq, struct task_struct *p)
|
|
|
|
|
|
|
|
|
|
SCHED_WARN_ON(task_rq(p) != rq);
|
|
|
|
|
|
|
|
|
|
if (rq->cfs.h_nr_queued > 1) {
|
|
|
|
|
if (rq->cfs.h_nr_running > 1) {
|
|
|
|
|
u64 ran = se->sum_exec_runtime - se->prev_sum_exec_runtime;
|
|
|
|
|
u64 slice = se->slice;
|
|
|
|
|
s64 delta = slice - ran;
|
|
|
|
|
@@ -7136,7 +7136,7 @@ enqueue_task_fair(struct rq *rq, struct task_struct *p, int flags)
|
|
|
|
|
int idle_h_nr_running = task_has_idle_policy(p);
|
|
|
|
|
int h_nr_delayed = 0;
|
|
|
|
|
int task_new = !(flags & ENQUEUE_WAKEUP);
|
|
|
|
|
int rq_h_nr_queued = rq->cfs.h_nr_queued;
|
|
|
|
|
int rq_h_nr_running = rq->cfs.h_nr_running;
|
|
|
|
|
u64 slice = 0;
|
|
|
|
|
int should_iowait_boost;
|
|
|
|
|
|
|
|
|
|
@@ -7187,7 +7187,7 @@ enqueue_task_fair(struct rq *rq, struct task_struct *p, int flags)
|
|
|
|
|
enqueue_entity(cfs_rq, se, flags);
|
|
|
|
|
slice = cfs_rq_min_slice(cfs_rq);
|
|
|
|
|
|
|
|
|
|
cfs_rq->h_nr_queued++;
|
|
|
|
|
cfs_rq->h_nr_running++;
|
|
|
|
|
cfs_rq->idle_h_nr_running += idle_h_nr_running;
|
|
|
|
|
cfs_rq->h_nr_delayed += h_nr_delayed;
|
|
|
|
|
|
|
|
|
|
@@ -7214,7 +7214,7 @@ enqueue_task_fair(struct rq *rq, struct task_struct *p, int flags)
|
|
|
|
|
min_vruntime_cb_propagate(&se->run_node, NULL);
|
|
|
|
|
slice = cfs_rq_min_slice(cfs_rq);
|
|
|
|
|
|
|
|
|
|
cfs_rq->h_nr_queued++;
|
|
|
|
|
cfs_rq->h_nr_running++;
|
|
|
|
|
cfs_rq->idle_h_nr_running += idle_h_nr_running;
|
|
|
|
|
cfs_rq->h_nr_delayed += h_nr_delayed;
|
|
|
|
|
|
|
|
|
|
@@ -7226,7 +7226,7 @@ enqueue_task_fair(struct rq *rq, struct task_struct *p, int flags)
|
|
|
|
|
goto enqueue_throttle;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
if (!rq_h_nr_queued && rq->cfs.h_nr_queued) {
|
|
|
|
|
if (!rq_h_nr_running && rq->cfs.h_nr_running) {
|
|
|
|
|
/* Account for idle runtime */
|
|
|
|
|
if (!rq->nr_running)
|
|
|
|
|
dl_server_update_idle_time(rq, rq->curr);
|
|
|
|
|
@@ -7273,19 +7273,19 @@ static void set_next_buddy(struct sched_entity *se);
|
|
|
|
|
static int dequeue_entities(struct rq *rq, struct sched_entity *se, int flags)
|
|
|
|
|
{
|
|
|
|
|
bool was_sched_idle = sched_idle_rq(rq);
|
|
|
|
|
int rq_h_nr_queued = rq->cfs.h_nr_queued;
|
|
|
|
|
int rq_h_nr_running = rq->cfs.h_nr_running;
|
|
|
|
|
bool task_sleep = flags & DEQUEUE_SLEEP;
|
|
|
|
|
bool task_delayed = flags & DEQUEUE_DELAYED;
|
|
|
|
|
struct task_struct *p = NULL;
|
|
|
|
|
int idle_h_nr_running = 0;
|
|
|
|
|
int h_nr_queued = 0;
|
|
|
|
|
int h_nr_running = 0;
|
|
|
|
|
int h_nr_delayed = 0;
|
|
|
|
|
struct cfs_rq *cfs_rq;
|
|
|
|
|
u64 slice = 0;
|
|
|
|
|
|
|
|
|
|
if (entity_is_task(se)) {
|
|
|
|
|
p = task_of(se);
|
|
|
|
|
h_nr_queued = 1;
|
|
|
|
|
h_nr_running = 1;
|
|
|
|
|
idle_h_nr_running = task_has_idle_policy(p);
|
|
|
|
|
if (!task_sleep && !task_delayed)
|
|
|
|
|
h_nr_delayed = !!se->sched_delayed;
|
|
|
|
|
@@ -7302,12 +7302,12 @@ static int dequeue_entities(struct rq *rq, struct sched_entity *se, int flags)
|
|
|
|
|
break;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
cfs_rq->h_nr_queued -= h_nr_queued;
|
|
|
|
|
cfs_rq->h_nr_running -= h_nr_running;
|
|
|
|
|
cfs_rq->idle_h_nr_running -= idle_h_nr_running;
|
|
|
|
|
cfs_rq->h_nr_delayed -= h_nr_delayed;
|
|
|
|
|
|
|
|
|
|
if (cfs_rq_is_idle(cfs_rq))
|
|
|
|
|
idle_h_nr_running = h_nr_queued;
|
|
|
|
|
idle_h_nr_running = h_nr_running;
|
|
|
|
|
|
|
|
|
|
/* end evaluation on encountering a throttled cfs_rq */
|
|
|
|
|
if (cfs_rq_throttled(cfs_rq))
|
|
|
|
|
@@ -7344,21 +7344,21 @@ static int dequeue_entities(struct rq *rq, struct sched_entity *se, int flags)
|
|
|
|
|
min_vruntime_cb_propagate(&se->run_node, NULL);
|
|
|
|
|
slice = cfs_rq_min_slice(cfs_rq);
|
|
|
|
|
|
|
|
|
|
cfs_rq->h_nr_queued -= h_nr_queued;
|
|
|
|
|
cfs_rq->h_nr_running -= h_nr_running;
|
|
|
|
|
cfs_rq->idle_h_nr_running -= idle_h_nr_running;
|
|
|
|
|
cfs_rq->h_nr_delayed -= h_nr_delayed;
|
|
|
|
|
|
|
|
|
|
if (cfs_rq_is_idle(cfs_rq))
|
|
|
|
|
idle_h_nr_running = h_nr_queued;
|
|
|
|
|
idle_h_nr_running = h_nr_running;
|
|
|
|
|
|
|
|
|
|
/* end evaluation on encountering a throttled cfs_rq */
|
|
|
|
|
if (cfs_rq_throttled(cfs_rq))
|
|
|
|
|
return 0;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
sub_nr_running(rq, h_nr_queued);
|
|
|
|
|
sub_nr_running(rq, h_nr_running);
|
|
|
|
|
|
|
|
|
|
if (rq_h_nr_queued && !rq->cfs.h_nr_queued)
|
|
|
|
|
if (rq_h_nr_running && !rq->cfs.h_nr_running)
|
|
|
|
|
dl_server_stop(&rq->fair_server);
|
|
|
|
|
|
|
|
|
|
/* balance early to pull high priority tasks */
|
|
|
|
|
@@ -10577,7 +10577,7 @@ sched_reduced_capacity(struct rq *rq, struct sched_domain *sd)
|
|
|
|
|
* When there is more than 1 task, the group_overloaded case already
|
|
|
|
|
* takes care of cpu with reduced capacity
|
|
|
|
|
*/
|
|
|
|
|
if (rq->cfs.h_nr_queued != 1)
|
|
|
|
|
if (rq->cfs.h_nr_running != 1)
|
|
|
|
|
return false;
|
|
|
|
|
|
|
|
|
|
return check_cpu_capacity(rq, sd);
|
|
|
|
|
@@ -10612,7 +10612,7 @@ static inline void update_sg_lb_stats(struct lb_env *env,
|
|
|
|
|
sgs->group_load += load;
|
|
|
|
|
sgs->group_util += cpu_util_cfs(i);
|
|
|
|
|
sgs->group_runnable += cpu_runnable(rq);
|
|
|
|
|
sgs->sum_h_nr_running += rq->cfs.h_nr_queued;
|
|
|
|
|
sgs->sum_h_nr_running += rq->cfs.h_nr_running;
|
|
|
|
|
|
|
|
|
|
nr_running = rq->nr_running;
|
|
|
|
|
sgs->sum_nr_running += nr_running;
|
|
|
|
|
@@ -10930,7 +10930,7 @@ static inline void update_sg_wakeup_stats(struct sched_domain *sd,
|
|
|
|
|
sgs->group_util += cpu_util_without(i, p);
|
|
|
|
|
sgs->group_runnable += cpu_runnable_without(rq, p);
|
|
|
|
|
local = task_running_on_cpu(i, p);
|
|
|
|
|
sgs->sum_h_nr_running += rq->cfs.h_nr_queued - local;
|
|
|
|
|
sgs->sum_h_nr_running += rq->cfs.h_nr_running - local;
|
|
|
|
|
|
|
|
|
|
nr_running = rq->nr_running - local;
|
|
|
|
|
sgs->sum_nr_running += nr_running;
|
|
|
|
|
@@ -11726,7 +11726,7 @@ static struct rq *sched_balance_find_src_rq(struct lb_env *env,
|
|
|
|
|
if (rt > env->fbq_type)
|
|
|
|
|
continue;
|
|
|
|
|
|
|
|
|
|
nr_running = rq->cfs.h_nr_queued;
|
|
|
|
|
nr_running = rq->cfs.h_nr_running;
|
|
|
|
|
if (!nr_running)
|
|
|
|
|
continue;
|
|
|
|
|
|
|
|
|
|
@@ -11886,7 +11886,7 @@ static int need_active_balance(struct lb_env *env)
|
|
|
|
|
* available on dst_cpu.
|
|
|
|
|
*/
|
|
|
|
|
if (env->idle &&
|
|
|
|
|
(env->src_rq->cfs.h_nr_queued == 1)) {
|
|
|
|
|
(env->src_rq->cfs.h_nr_running == 1)) {
|
|
|
|
|
if ((check_cpu_capacity(env->src_rq, sd)) &&
|
|
|
|
|
(capacity_of(env->src_cpu)*sd->imbalance_pct < capacity_of(env->dst_cpu)*100))
|
|
|
|
|
return 1;
|
|
|
|
|
@@ -12645,7 +12645,7 @@ static void nohz_balancer_kick(struct rq *rq)
|
|
|
|
|
* If there's a runnable CFS task and the current CPU has reduced
|
|
|
|
|
* capacity, kick the ILB to see if there's a better CPU to run on:
|
|
|
|
|
*/
|
|
|
|
|
if (rq->cfs.h_nr_queued >= 1 && check_cpu_capacity(rq, sd)) {
|
|
|
|
|
if (rq->cfs.h_nr_running >= 1 && check_cpu_capacity(rq, sd)) {
|
|
|
|
|
flags = NOHZ_STATS_KICK | NOHZ_BALANCE_KICK;
|
|
|
|
|
goto unlock;
|
|
|
|
|
}
|
|
|
|
|
@@ -13148,11 +13148,11 @@ static int sched_balance_newidle(struct rq *this_rq, struct rq_flags *rf)
|
|
|
|
|
* have been enqueued in the meantime. Since we're not going idle,
|
|
|
|
|
* pretend we pulled a task.
|
|
|
|
|
*/
|
|
|
|
|
if (this_rq->cfs.h_nr_queued && !pulled_task)
|
|
|
|
|
if (this_rq->cfs.h_nr_running && !pulled_task)
|
|
|
|
|
pulled_task = 1;
|
|
|
|
|
|
|
|
|
|
/* Is there a task of a high priority class? */
|
|
|
|
|
if (this_rq->nr_running != this_rq->cfs.h_nr_queued)
|
|
|
|
|
if (this_rq->nr_running != this_rq->cfs.h_nr_running)
|
|
|
|
|
pulled_task = -1;
|
|
|
|
|
|
|
|
|
|
out:
|
|
|
|
|
@@ -13839,7 +13839,7 @@ int sched_group_set_idle(struct task_group *tg, long idle)
|
|
|
|
|
parent_cfs_rq->idle_nr_running--;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
idle_task_delta = grp_cfs_rq->h_nr_queued -
|
|
|
|
|
idle_task_delta = grp_cfs_rq->h_nr_running -
|
|
|
|
|
grp_cfs_rq->idle_h_nr_running;
|
|
|
|
|
if (!cfs_rq_is_idle(grp_cfs_rq))
|
|
|
|
|
idle_task_delta *= -1;
|
|
|
|
|
|