Searched refs:rq_clock_pelt (Results 1 – 4 of 4) sorted by relevance
/linux-6.6.21/kernel/sched/ |
D | pelt.h | 64 static inline u64 rq_clock_pelt(struct rq *rq) in rq_clock_pelt() function 80 u64_u32_store(rq->clock_pelt_idle, rq_clock_pelt(rq)); in _update_idle_rq_clock_pelt() 174 return rq_clock_pelt(rq_of(cfs_rq)) - cfs_rq->throttled_clock_pelt_time; in cfs_rq_clock_pelt() 180 return rq_clock_pelt(rq_of(cfs_rq)); in cfs_rq_clock_pelt() 221 static inline u64 rq_clock_pelt(struct rq *rq) in rq_clock_pelt() function
|
D | deadline.c | 1991 update_dl_rq_load_avg(rq_clock_pelt(rq), rq, 0); in set_next_task_dl() 2043 update_dl_rq_load_avg(rq_clock_pelt(rq), rq, 1); in put_prev_task_dl() 2060 update_dl_rq_load_avg(rq_clock_pelt(rq), rq, 1); in task_tick_dl() 2661 update_dl_rq_load_avg(rq_clock_pelt(rq), rq, 0); in switched_to_dl()
|
D | rt.c | 1768 update_rt_rq_load_avg(rq_clock_pelt(rq), rq, 0); in set_next_task_rt() 1838 update_rt_rq_load_avg(rq_clock_pelt(rq), rq, 1); in put_prev_task_rt() 2551 update_rt_rq_load_avg(rq_clock_pelt(rq), rq, 0); in switched_to_rt() 2652 update_rt_rq_load_avg(rq_clock_pelt(rq), rq, 1); in task_tick_rt()
|
D | fair.c | 5608 cfs_rq->throttled_clock_pelt_time += rq_clock_pelt(rq) - in tg_unthrottle_up() 5637 cfs_rq->throttled_clock_pelt = rq_clock_pelt(rq); in tg_throttle_down() 6182 cfs_rq->throttled_clock_pelt = rq_clock_pelt(cpu_rq(cpu)); in sync_throttle() 9154 u64 now = rq_clock_pelt(rq); in __update_blocked_others()
|