Searched refs:src_rq (Results 1 – 6 of 6) sorted by relevance
| /linux-6.15/kernel/sched/ |
| H A D | ext.c | 2594 raw_spin_rq_lock(src_rq); in unlink_dsq_and_lock_src_rq() 2648 if (src_rq != dst_rq && in move_task_between_dsqs() 2651 dst_rq = src_rq; in move_task_between_dsqs() 2655 dst_rq = src_rq; in move_task_between_dsqs() 2664 if (src_rq == dst_rq) { in move_task_between_dsqs() 2803 if (src_rq != dst_rq && in dispatch_to_local_dsq() 2829 locked_rq = src_rq; in dispatch_to_local_dsq() 2841 if (src_rq == dst_rq) { in dispatch_to_local_dsq() 6283 src_rq = task_rq(p); in scx_dsq_move() 6290 if (this_rq != src_rq) { in scx_dsq_move() [all …]
|
| H A D | rt.c | 2274 struct rq *src_rq; in pull_rt_task() local 2302 src_rq = cpu_rq(cpu); in pull_rt_task() 2311 if (src_rq->rt.highest_prio.next >= in pull_rt_task() 2321 double_lock_balance(this_rq, src_rq); in pull_rt_task() 2334 WARN_ON(p == src_rq->curr); in pull_rt_task() 2345 if (p->prio < src_rq->donor->prio) in pull_rt_task() 2349 push_task = get_push_task(src_rq); in pull_rt_task() 2351 move_queued_task_locked(src_rq, this_rq, p); in pull_rt_task() 2362 double_unlock_balance(this_rq, src_rq); in pull_rt_task() 2367 stop_one_cpu_nowait(src_rq->cpu, push_cpu_stop, in pull_rt_task() [all …]
|
| H A D | deadline.c | 2786 struct rq *src_rq; in pull_dl_task() local 2802 src_rq = cpu_rq(cpu); in pull_dl_task() 2810 src_rq->dl.earliest_dl.next)) in pull_dl_task() 2815 double_lock_balance(this_rq, src_rq); in pull_dl_task() 2821 if (src_rq->dl.dl_nr_running <= 1) in pull_dl_task() 2833 WARN_ON(p == src_rq->curr); in pull_dl_task() 2841 src_rq->donor->dl.deadline)) in pull_dl_task() 2845 push_task = get_push_task(src_rq); in pull_dl_task() 2847 move_queued_task_locked(src_rq, this_rq, p); in pull_dl_task() 2855 double_unlock_balance(this_rq, src_rq); in pull_dl_task() [all …]
|
| H A D | core.c | 3356 struct rq *src_rq, *dst_rq; in __migrate_swap_task() local 3359 src_rq = task_rq(p); in __migrate_swap_task() 3362 rq_pin_lock(src_rq, &srf); in __migrate_swap_task() 3369 rq_unpin_lock(src_rq, &srf); in __migrate_swap_task() 3389 struct rq *src_rq, *dst_rq; in migrate_swap_stop() local 3394 src_rq = cpu_rq(arg->src_cpu); in migrate_swap_stop() 3398 guard(double_rq_lock)(src_rq, dst_rq); in migrate_swap_stop() 10339 src_task = rcu_dereference(src_rq->curr); in __sched_mm_cid_migrate_from_fetch_cid() 10388 src_task = rcu_dereference(src_rq->curr); in __sched_mm_cid_migrate_from_try_steal_cid() 10419 struct rq *src_rq; in sched_mm_cid_migrate_to() local [all …]
|
| H A D | fair.c | 9226 struct rq *src_rq; member 9257 lockdep_assert_rq_held(env->src_rq); in task_hot() 9320 if (env->src_rq->nr_running > env->src_rq->nr_preferred_running) in migrate_degrades_locality() 9386 lockdep_assert_rq_held(env->src_rq); in can_migrate_task() 9453 if (task_on_cpu(env->src_rq, p)) { in can_migrate_task() 9489 lockdep_assert_rq_held(env->src_rq); in detach_task() 9511 lockdep_assert_rq_held(env->src_rq); in detach_one_task() 9545 lockdep_assert_rq_held(env->src_rq); in detach_tasks() 9551 if (env->src_rq->nr_running <= 1) { in detach_tasks() 11785 env.src_rq = busiest; in sched_balance_rq() [all …]
|
| H A D | sched.h | 3868 void move_queued_task_locked(struct rq *src_rq, struct rq *dst_rq, struct task_struct *task) in move_queued_task_locked() argument 3870 lockdep_assert_rq_held(src_rq); in move_queued_task_locked() 3873 deactivate_task(src_rq, task, 0); in move_queued_task_locked()
|