Lines Matching defs:cfs_rq

348 static inline bool list_add_leaf_cfs_rq(struct cfs_rq *cfs_rq)  in list_add_leaf_cfs_rq()
416 static inline void list_del_leaf_cfs_rq(struct cfs_rq *cfs_rq) in list_del_leaf_cfs_rq()
442 #define for_each_leaf_cfs_rq_safe(rq, cfs_rq, pos) \ argument
498 static int cfs_rq_is_idle(struct cfs_rq *cfs_rq) in cfs_rq_is_idle()
515 static inline bool list_add_leaf_cfs_rq(struct cfs_rq *cfs_rq) in list_add_leaf_cfs_rq()
520 static inline void list_del_leaf_cfs_rq(struct cfs_rq *cfs_rq) in list_del_leaf_cfs_rq()
528 #define for_each_leaf_cfs_rq_safe(rq, cfs_rq, pos) \ argument
546 static int cfs_rq_is_idle(struct cfs_rq *cfs_rq) in cfs_rq_is_idle()
592 static void update_min_vruntime(struct cfs_rq *cfs_rq) in update_min_vruntime()
628 static void __enqueue_entity(struct cfs_rq *cfs_rq, struct sched_entity *se) in __enqueue_entity()
633 static void __dequeue_entity(struct cfs_rq *cfs_rq, struct sched_entity *se) in __dequeue_entity()
638 struct sched_entity *__pick_first_entity(struct cfs_rq *cfs_rq) in __pick_first_entity()
659 struct sched_entity *__pick_last_entity(struct cfs_rq *cfs_rq) in __pick_last_entity()
726 static u64 sched_slice(struct cfs_rq *cfs_rq, struct sched_entity *se) in sched_slice()
772 static u64 sched_vslice(struct cfs_rq *cfs_rq, struct sched_entity *se) in sched_vslice()
832 struct cfs_rq *cfs_rq = cfs_rq_of(se); in post_init_entity_util_avg() local
874 static void update_tg_load_avg(struct cfs_rq *cfs_rq) in update_tg_load_avg()
882 static void update_curr(struct cfs_rq *cfs_rq) in update_curr()
928 update_stats_wait_start_fair(struct cfs_rq *cfs_rq, struct sched_entity *se) in update_stats_wait_start_fair()
945 update_stats_wait_end_fair(struct cfs_rq *cfs_rq, struct sched_entity *se) in update_stats_wait_end_fair()
971 update_stats_enqueue_sleeper_fair(struct cfs_rq *cfs_rq, struct sched_entity *se) in update_stats_enqueue_sleeper_fair()
991 update_stats_enqueue_fair(struct cfs_rq *cfs_rq, struct sched_entity *se, int flags) in update_stats_enqueue_fair()
1008 update_stats_dequeue_fair(struct cfs_rq *cfs_rq, struct sched_entity *se, int flags) in update_stats_dequeue_fair()
1040 update_stats_curr_start(struct cfs_rq *cfs_rq, struct sched_entity *se) in update_stats_curr_start()
3203 account_entity_enqueue(struct cfs_rq *cfs_rq, struct sched_entity *se) in account_entity_enqueue()
3220 account_entity_dequeue(struct cfs_rq *cfs_rq, struct sched_entity *se) in account_entity_dequeue()
3284 enqueue_load_avg(struct cfs_rq *cfs_rq, struct sched_entity *se) in enqueue_load_avg()
3291 dequeue_load_avg(struct cfs_rq *cfs_rq, struct sched_entity *se) in dequeue_load_avg()
3301 enqueue_load_avg(struct cfs_rq *cfs_rq, struct sched_entity *se) { } in enqueue_load_avg()
3303 dequeue_load_avg(struct cfs_rq *cfs_rq, struct sched_entity *se) { } in dequeue_load_avg()
3306 static void reweight_entity(struct cfs_rq *cfs_rq, struct sched_entity *se, in reweight_entity()
3336 struct cfs_rq *cfs_rq = cfs_rq_of(se); in reweight_task() local
3421 static long calc_group_shares(struct cfs_rq *cfs_rq) in calc_group_shares()
3489 static inline void cfs_rq_util_change(struct cfs_rq *cfs_rq, int flags) in cfs_rq_util_change()
3536 static inline u64 cfs_rq_last_update_time(struct cfs_rq *cfs_rq) in cfs_rq_last_update_time()
3550 static inline bool child_cfs_rq_on_list(struct cfs_rq *cfs_rq) in child_cfs_rq_on_list()
3568 static inline bool cfs_rq_is_decayed(struct cfs_rq *cfs_rq) in cfs_rq_is_decayed()
3596 static inline void update_tg_load_avg(struct cfs_rq *cfs_rq) in update_tg_load_avg()
3711 update_tg_cfs_util(struct cfs_rq *cfs_rq, struct sched_entity *se, struct cfs_rq *gcfs_rq) in update_tg_cfs_util()
3743 update_tg_cfs_runnable(struct cfs_rq *cfs_rq, struct sched_entity *se, struct cfs_rq *gcfs_rq) in update_tg_cfs_runnable()
3773 update_tg_cfs_load(struct cfs_rq *cfs_rq, struct sched_entity *se, struct cfs_rq *gcfs_rq) in update_tg_cfs_load()
3840 static inline void add_tg_cfs_propagate(struct cfs_rq *cfs_rq, long runnable_sum) in add_tg_cfs_propagate()
3849 struct cfs_rq *cfs_rq, *gcfs_rq; in propagate_entity_load_avg() local
3906 static inline void update_tg_load_avg(struct cfs_rq *cfs_rq) {} in update_tg_load_avg()
3913 static inline void add_tg_cfs_propagate(struct cfs_rq *cfs_rq, long runnable_sum) {} in add_tg_cfs_propagate()
3921 struct cfs_rq *cfs_rq; in migrate_se_pelt_lag() local
4016 update_cfs_rq_load_avg(u64 now, struct cfs_rq *cfs_rq) in update_cfs_rq_load_avg()
4087 static void attach_entity_load_avg(struct cfs_rq *cfs_rq, struct sched_entity *se) in attach_entity_load_avg()
4142 static void detach_entity_load_avg(struct cfs_rq *cfs_rq, struct sched_entity *se) in detach_entity_load_avg()
4173 static inline void update_load_avg(struct cfs_rq *cfs_rq, struct sched_entity *se, int flags) in update_load_avg()
4221 struct cfs_rq *cfs_rq = cfs_rq_of(se); in sync_entity_load_avg() local
4234 struct cfs_rq *cfs_rq = cfs_rq_of(se); in remove_entity_load_avg() local
4253 static inline unsigned long cfs_rq_runnable_avg(struct cfs_rq *cfs_rq) in cfs_rq_runnable_avg()
4258 static inline unsigned long cfs_rq_load_avg(struct cfs_rq *cfs_rq) in cfs_rq_load_avg()
4298 static inline void util_est_enqueue(struct cfs_rq *cfs_rq, in util_est_enqueue()
4314 static inline void util_est_dequeue(struct cfs_rq *cfs_rq, in util_est_dequeue()
4345 static inline void util_est_update(struct cfs_rq *cfs_rq, in util_est_update()
4586 static inline bool cfs_rq_is_decayed(struct cfs_rq *cfs_rq) in cfs_rq_is_decayed()
4596 static inline void update_load_avg(struct cfs_rq *cfs_rq, struct sched_entity *se, int not_used1) in update_load_avg()
4604 attach_entity_load_avg(struct cfs_rq *cfs_rq, struct sched_entity *se) {} in attach_entity_load_avg()
4606 detach_entity_load_avg(struct cfs_rq *cfs_rq, struct sched_entity *se) {} in detach_entity_load_avg()
4614 util_est_enqueue(struct cfs_rq *cfs_rq, struct task_struct *p) {} in util_est_enqueue()
4617 util_est_dequeue(struct cfs_rq *cfs_rq, struct task_struct *p) {} in util_est_dequeue()
4620 util_est_update(struct cfs_rq *cfs_rq, struct task_struct *p, in util_est_update()
4626 static void check_spread(struct cfs_rq *cfs_rq, struct sched_entity *se) in check_spread()
4640 place_entity(struct cfs_rq *cfs_rq, struct sched_entity *se, int initial) in place_entity()
4711 enqueue_entity(struct cfs_rq *cfs_rq, struct sched_entity *se, int flags) in enqueue_entity()
4768 struct cfs_rq *cfs_rq = cfs_rq_of(se); in __clear_buddies_last() local
4779 struct cfs_rq *cfs_rq = cfs_rq_of(se); in __clear_buddies_next() local
4790 struct cfs_rq *cfs_rq = cfs_rq_of(se); in __clear_buddies_skip() local
4798 static void clear_buddies(struct cfs_rq *cfs_rq, struct sched_entity *se) in clear_buddies()
4813 dequeue_entity(struct cfs_rq *cfs_rq, struct sched_entity *se, int flags) in dequeue_entity()
4877 check_preempt_tick(struct cfs_rq *cfs_rq, struct sched_entity *curr) in check_preempt_tick()
4914 set_next_entity(struct cfs_rq *cfs_rq, struct sched_entity *se) in set_next_entity()
4962 pick_next_entity(struct cfs_rq *cfs_rq, struct sched_entity *curr) in pick_next_entity()
5012 static void put_prev_entity(struct cfs_rq *cfs_rq, struct sched_entity *prev) in put_prev_entity()
5037 entity_tick(struct cfs_rq *cfs_rq, struct sched_entity *curr, int queued) in entity_tick()
5151 struct cfs_rq *cfs_rq, u64 target_runtime) in __assign_cfs_rq_runtime()
5178 static int assign_cfs_rq_runtime(struct cfs_rq *cfs_rq) in assign_cfs_rq_runtime()
5190 static void __account_cfs_rq_runtime(struct cfs_rq *cfs_rq, u64 delta_exec) in __account_cfs_rq_runtime()
5209 void account_cfs_rq_runtime(struct cfs_rq *cfs_rq, u64 delta_exec) in account_cfs_rq_runtime()
5217 static inline int cfs_rq_throttled(struct cfs_rq *cfs_rq) in cfs_rq_throttled()
5223 static inline int throttled_hierarchy(struct cfs_rq *cfs_rq) in throttled_hierarchy()
5248 struct cfs_rq *cfs_rq = tg->cfs_rq[cpu_of(rq)]; in tg_unthrottle_up() local
5266 struct cfs_rq *cfs_rq = tg->cfs_rq[cpu_of(rq)]; in tg_throttle_down() local
5278 static bool throttle_cfs_rq(struct cfs_rq *cfs_rq) in throttle_cfs_rq()
5365 void unthrottle_cfs_rq(struct cfs_rq *cfs_rq) in unthrottle_cfs_rq()
5450 struct cfs_rq *cfs_rq; in distribute_cfs_runtime() local
5601 static void __return_cfs_rq_runtime(struct cfs_rq *cfs_rq) in __return_cfs_rq_runtime()
5624 static __always_inline void return_cfs_rq_runtime(struct cfs_rq *cfs_rq) in return_cfs_rq_runtime()
5669 static void check_enqueue_throttle(struct cfs_rq *cfs_rq) in check_enqueue_throttle()
5690 struct cfs_rq *pcfs_rq, *cfs_rq; in sync_throttle() local
5706 static bool check_cfs_rq_runtime(struct cfs_rq *cfs_rq) in check_cfs_rq_runtime()
5807 static void init_cfs_rq_runtime(struct cfs_rq *cfs_rq) in init_cfs_rq_runtime()
5852 struct cfs_rq *cfs_rq = tg->cfs_rq[cpu_of(rq)]; in update_runtime_enabled() local
5870 struct cfs_rq *cfs_rq = tg->cfs_rq[cpu_of(rq)]; in unthrottle_offline_cfs_rqs() local
5899 static void account_cfs_rq_runtime(struct cfs_rq *cfs_rq, u64 delta_exec) {} in account_cfs_rq_runtime()
5900 static bool check_cfs_rq_runtime(struct cfs_rq *cfs_rq) { return false; } in check_cfs_rq_runtime()
5901 static void check_enqueue_throttle(struct cfs_rq *cfs_rq) {} in check_enqueue_throttle()
5903 static __always_inline void return_cfs_rq_runtime(struct cfs_rq *cfs_rq) {} in return_cfs_rq_runtime()
5905 static inline int cfs_rq_throttled(struct cfs_rq *cfs_rq) in cfs_rq_throttled()
5910 static inline int throttled_hierarchy(struct cfs_rq *cfs_rq) in throttled_hierarchy()
5924 static void init_cfs_rq_runtime(struct cfs_rq *cfs_rq) {} in init_cfs_rq_runtime()
5945 struct cfs_rq *cfs_rq = cfs_rq_of(se); in hrtick_start_fair() local
6021 static bool sched_idle_cfs_rq(struct cfs_rq *cfs_rq) in sched_idle_cfs_rq()
6042 struct cfs_rq *cfs_rq; in enqueue_task_fair() local
6135 struct cfs_rq *cfs_rq; in dequeue_task_fair() local
6242 struct cfs_rq *cfs_rq; in cpu_load_without() local
6265 struct cfs_rq *cfs_rq; in cpu_runnable_without() local
6939 struct cfs_rq *cfs_rq = &cpu_rq(cpu)->cfs; in cpu_util_next() local
7428 struct cfs_rq *cfs_rq = cfs_rq_of(se); in migrate_task_rq_fair() local
7557 struct cfs_rq *cfs_rq = task_cfs_rq(curr); in check_preempt_wakeup() local
7654 struct cfs_rq *cfs_rq; in pick_task_fair() local
7686 struct cfs_rq *cfs_rq = &rq->cfs; in pick_next_task_fair() local
7838 struct cfs_rq *cfs_rq; in put_prev_task_fair() local
7854 struct cfs_rq *cfs_rq = task_cfs_rq(curr); in yield_task_fair() local
8530 static inline bool cfs_rq_has_blocked(struct cfs_rq *cfs_rq) in cfs_rq_has_blocked()
8571 static inline bool cfs_rq_has_blocked(struct cfs_rq *cfs_rq) { return false; } in cfs_rq_has_blocked()
8607 struct cfs_rq *cfs_rq, *pos; in __update_blocked_fair() local
8653 static void update_cfs_rq_h_load(struct cfs_rq *cfs_rq) in update_cfs_rq_h_load()
8688 struct cfs_rq *cfs_rq = task_cfs_rq(p); in task_h_load() local
8697 struct cfs_rq *cfs_rq = &rq->cfs; in __update_blocked_fair() local
11658 struct cfs_rq *cfs_rq = cfs_rq_of(se); in se_fi_update() local
11740 struct cfs_rq *cfs_rq; in task_tick_fair() local
11764 struct cfs_rq *cfs_rq; in task_fork_fair() local
11853 struct cfs_rq *cfs_rq = cfs_rq_of(se); in propagate_entity_cfs_rq() local
11882 struct cfs_rq *cfs_rq = cfs_rq_of(se); in detach_entity_cfs_rq() local
11904 struct cfs_rq *cfs_rq = cfs_rq_of(se); in attach_entity_cfs_rq() local
11916 struct cfs_rq *cfs_rq = cfs_rq_of(se); in detach_task_cfs_rq() local
11933 struct cfs_rq *cfs_rq = cfs_rq_of(se); in attach_task_cfs_rq() local
11983 struct cfs_rq *cfs_rq = cfs_rq_of(se); in set_next_task_fair() local
11991 void init_cfs_rq(struct cfs_rq *cfs_rq) in init_cfs_rq()
12038 struct cfs_rq *cfs_rq; in alloc_fair_sched_group() local
12121 void init_tg_cfs_entry(struct task_group *tg, struct cfs_rq *cfs_rq, in init_tg_cfs_entry()
12251 struct cfs_rq *cfs_rq = cfs_rq_of(se); in sched_group_set_idle() local
12360 struct cfs_rq *cfs_rq, *pos; in print_cfs_stats() local