Lines Matching refs:this_rq

4715 static int newidle_balance(struct rq *this_rq, struct rq_flags *rf);
5869 if (rq == this_rq()) { in __unthrottle_cfs_rq_async()
7256 struct rq *this_rq = this_rq(); in select_idle_cpu() local
7276 if (unlikely(this_rq->wake_stamp < now)) { in select_idle_cpu()
7277 while (this_rq->wake_stamp < now && this_rq->wake_avg_idle) { in select_idle_cpu()
7278 this_rq->wake_stamp++; in select_idle_cpu()
7279 this_rq->wake_avg_idle >>= 1; in select_idle_cpu()
7283 avg_idle = this_rq->wake_avg_idle; in select_idle_cpu()
7331 this_rq->wake_avg_idle -= min(this_rq->wake_avg_idle, time); in select_idle_cpu()
7456 this_rq()->nr_running <= 1 && in select_idle_sibling()
7821 struct root_domain *rd = this_rq()->rd; in find_energy_efficient_cpu()
11143 static int load_balance(int this_cpu, struct rq *this_rq, in load_balance() argument
11156 .dst_rq = this_rq, in load_balance()
11865 SCHED_WARN_ON(rq != this_rq()); in nohz_balance_exit_idle()
11977 static void _nohz_idle_balance(struct rq *this_rq, unsigned int flags) in _nohz_idle_balance() argument
11984 int this_cpu = this_rq->cpu; in _nohz_idle_balance()
12080 static bool nohz_idle_balance(struct rq *this_rq, enum cpu_idle_type idle) in nohz_idle_balance() argument
12082 unsigned int flags = this_rq->nohz_idle_balance; in nohz_idle_balance()
12087 this_rq->nohz_idle_balance = 0; in nohz_idle_balance()
12092 _nohz_idle_balance(this_rq, flags); in nohz_idle_balance()
12115 static void nohz_newidle_balance(struct rq *this_rq) in nohz_newidle_balance() argument
12117 int this_cpu = this_rq->cpu; in nohz_newidle_balance()
12127 if (this_rq->avg_idle < sysctl_sched_migration_cost) in nohz_newidle_balance()
12145 static inline bool nohz_idle_balance(struct rq *this_rq, enum cpu_idle_type idle) in nohz_idle_balance() argument
12150 static inline void nohz_newidle_balance(struct rq *this_rq) { } in nohz_newidle_balance() argument
12162 static int newidle_balance(struct rq *this_rq, struct rq_flags *rf) in newidle_balance() argument
12165 int this_cpu = this_rq->cpu; in newidle_balance()
12170 update_misfit_status(NULL, this_rq); in newidle_balance()
12176 if (this_rq->ttwu_pending) in newidle_balance()
12183 this_rq->idle_stamp = rq_clock(this_rq); in newidle_balance()
12197 rq_unpin_lock(this_rq, rf); in newidle_balance()
12200 sd = rcu_dereference_check_sched_domain(this_rq->sd); in newidle_balance()
12202 if (!READ_ONCE(this_rq->rd->overload) || in newidle_balance()
12203 (sd && this_rq->avg_idle < sd->max_newidle_lb_cost)) { in newidle_balance()
12213 raw_spin_rq_unlock(this_rq); in newidle_balance()
12225 if (this_rq->avg_idle < curr_cost + sd->max_newidle_lb_cost) in newidle_balance()
12230 pulled_task = load_balance(this_cpu, this_rq, in newidle_balance()
12246 if (pulled_task || this_rq->nr_running > 0 || in newidle_balance()
12247 this_rq->ttwu_pending) in newidle_balance()
12252 raw_spin_rq_lock(this_rq); in newidle_balance()
12254 if (curr_cost > this_rq->max_idle_balance_cost) in newidle_balance()
12255 this_rq->max_idle_balance_cost = curr_cost; in newidle_balance()
12262 if (this_rq->cfs.h_nr_running && !pulled_task) in newidle_balance()
12266 if (this_rq->nr_running != this_rq->cfs.h_nr_running) in newidle_balance()
12271 if (time_after(this_rq->next_balance, next_balance)) in newidle_balance()
12272 this_rq->next_balance = next_balance; in newidle_balance()
12275 this_rq->idle_stamp = 0; in newidle_balance()
12277 nohz_newidle_balance(this_rq); in newidle_balance()
12279 rq_repin_lock(this_rq, rf); in newidle_balance()
12290 struct rq *this_rq = this_rq(); in run_rebalance_domains() local
12291 enum cpu_idle_type idle = this_rq->idle_balance ? in run_rebalance_domains()
12302 if (nohz_idle_balance(this_rq, idle)) in run_rebalance_domains()
12306 update_blocked_averages(this_rq->cpu); in run_rebalance_domains()
12307 rebalance_domains(this_rq, idle); in run_rebalance_domains()
12508 struct rq *rq = this_rq(); in task_fork_fair()