Searched refs:task_rq (Results 1 – 8 of 8) sorted by relevance
2096 struct rq *rq = task_rq(p); in direct_dispatch()2741 struct rq *task_rq = task_rq(p); in consume_dispatch_q() local2743 if (rq == task_rq) { in consume_dispatch_q()2785 struct rq *src_rq = task_rq(p); in dispatch_to_local_dsq()3704 struct rq *rq = task_rq(p); in scx_ops_enable_task()3730 struct rq *rq = task_rq(p); in scx_ops_disable_task()3746 lockdep_assert_rq_held(task_rq(p)); in scx_ops_exit_task()3870 lockdep_assert_rq_held(task_rq(p)); in reweight_task_scx()3904 lockdep_assert_rq_held(task_rq(p)); in scx_check_setscheduler()6283 src_rq = task_rq(p); in scx_dsq_move()[all …]
682 rq = task_rq(p); in __task_rq_lock()706 rq = task_rq(p); in task_rq_lock()2261 rq = task_rq(p); in wait_task_inactive()2583 if (task_rq(p) == rq) { in migration_cpu_stop()2662 if (task_rq(p) != rq) in push_cpu_stop()2679 if (task_rq(p) == rq) { in push_cpu_stop()2719 struct rq *rq = task_rq(p); in __do_set_cpus_allowed()3359 src_rq = task_rq(p); in __migrate_swap_task()6068 return (task_rq(t)->idle == t); in is_task_rq_idle()10719 struct rq *rq = task_rq(p); in sched_deq_and_put_task()[all …]
1146 if (!cpumask_subset(task_rq(p)->rd->span, mask)) in dl_task_check_affinity()1437 p_rq = task_rq(p); in yield_to()1446 if (task_rq(p) != p_rq) in yield_to()
142 if (task_on_cpu(task_rq(p), p)) in psi_enqueue()
77 rq = task_rq(dl_task_of(dl_se)); in rq_of_dl_se()375 dl_rq_change_utilization(task_rq(p), &p->dl, new_bw); in dl_change_utilization()2287 rq = task_rq(p); in migrate_task_rq_dl()2553 if (!cpudl_find(&task_rq(task)->rd->cpudl, task, later_mask)) in find_later_rq()2649 if (unlikely(task_rq(task) != rq || in find_lock_later_rq()2895 rq = task_rq(p); in set_cpus_allowed_dl()
298 return task_rq(p); in rq_of_rt_se()1816 ret = cpupri_find_fitness(&task_rq(task)->rd->cpupri, in find_lowest_rq()1821 ret = cpupri_find(&task_rq(task)->rd->cpupri, in find_lowest_rq()1922 if (unlikely(task_rq(task) != rq || in find_lock_lowest_rq()
1346 #define task_rq(p) cpu_rq(task_cpu(p)) macro1593 return &task_rq(p)->cfs; in task_cfs_rq()1599 struct rq *rq = task_rq(p); in cfs_rq_of()
1494 (lockdep_is_held(__rq_lockp(task_rq(p))) && !READ_ONCE(p->on_cpu))); in deref_task_numa_group()6790 WARN_ON_ONCE(task_rq(p) != rq); in hrtick_start_fair()9680 WARN_ON_ONCE(task_rq(p) != rq); in attach_task()13035 struct rq *rq = task_rq(a); in cfs_prio_less()13042 WARN_ON_ONCE(task_rq(b)->core != rq->core); in cfs_prio_less()13065 cfs_rqa = &task_rq(a)->cfs; in cfs_prio_less()13066 cfs_rqb = &task_rq(b)->cfs; in cfs_prio_less()13117 check_update_overutilized_status(task_rq(curr)); in task_tick_fair()