| /linux-6.15/kernel/sched/ |
| H A D | ext_idle.c | 452 numa_cpus = numa_span(prev_cpu); in scx_select_cpu_dfl() 455 llc_cpus = llc_span(prev_cpu); in scx_select_cpu_dfl() 471 cpu = prev_cpu; in scx_select_cpu_dfl() 508 cpu = prev_cpu; in scx_select_cpu_dfl() 556 cpu = prev_cpu; in scx_select_cpu_dfl() 822 if (!ops_cpu_valid(prev_cpu, NULL)) in scx_bpf_select_cpu_dfl() 823 goto prev_cpu; in scx_bpf_select_cpu_dfl() 826 goto prev_cpu; in scx_bpf_select_cpu_dfl() 829 goto prev_cpu; in scx_bpf_select_cpu_dfl() 839 prev_cpu: in scx_bpf_select_cpu_dfl() [all …]
|
| H A D | ext_idle.h | 30 s32 scx_select_cpu_dfl(struct task_struct *p, s32 prev_cpu, u64 wake_flags, u64 flags);
|
| H A D | fair.c | 7358 return available_idle_cpu(prev_cpu) ? prev_cpu : this_cpu; in wake_affine_idle() 7363 if (available_idle_cpu(prev_cpu)) in wake_affine_idle() 7364 return prev_cpu; in wake_affine_idle() 7425 return prev_cpu; in wake_affine() 7501 return prev_cpu; in sched_balance_find_dst_cpu() 8407 target = prev_cpu; in find_energy_efficient_cpu() 8478 if (cpu == prev_cpu) { in find_energy_efficient_cpu() 8505 prev_cpu); in find_energy_efficient_cpu() 8580 int new_cpu = prev_cpu; in select_task_rq_fair() 8600 new_cpu = prev_cpu; in select_task_rq_fair() [all …]
|
| /linux-6.15/arch/x86/include/asm/trace/ |
| H A D | irq_vectors.h | 156 unsigned int prev_cpu), 158 TP_ARGS(irq, vector, cpu, prev_vector, prev_cpu), 165 __field( unsigned int, prev_cpu ) 173 __entry->prev_cpu = prev_cpu; 179 __entry->prev_vector, __entry->prev_cpu) 186 unsigned int prev_cpu), \ 187 TP_ARGS(irq, vector, cpu, prev_vector, prev_cpu), NULL, NULL); \
|
| /linux-6.15/tools/testing/selftests/sched_ext/ |
| H A D | select_cpu_dispatch.bpf.c | 16 s32 prev_cpu, u64 wake_flags) in BPF_STRUCT_OPS() argument 19 s32 cpu = prev_cpu; in BPF_STRUCT_OPS() 29 cpu = prev_cpu; in BPF_STRUCT_OPS()
|
| H A D | enq_select_cpu_fails.bpf.c | 13 s32 scx_bpf_select_cpu_dfl(struct task_struct *p, s32 prev_cpu, u64 wake_flags, 17 s32 prev_cpu, u64 wake_flags) in BPF_STRUCT_OPS() argument 19 return prev_cpu; in BPF_STRUCT_OPS()
|
| H A D | select_cpu_dfl_nodispatch.bpf.c | 31 s32 scx_bpf_select_cpu_dfl(struct task_struct *p, s32 prev_cpu, u64 wake_flags, 35 s32 prev_cpu, u64 wake_flags) in BPF_STRUCT_OPS() argument 46 cpu = scx_bpf_select_cpu_dfl(p, prev_cpu, wake_flags, in BPF_STRUCT_OPS()
|
| H A D | select_cpu_dispatch_bad_dsq.bpf.c | 18 s32 prev_cpu, u64 wake_flags) in BPF_STRUCT_OPS() argument 23 return prev_cpu; in BPF_STRUCT_OPS()
|
| H A D | select_cpu_dispatch_dbl_dsp.bpf.c | 18 s32 prev_cpu, u64 wake_flags) in BPF_STRUCT_OPS() argument 24 return prev_cpu; in BPF_STRUCT_OPS()
|
| H A D | ddsp_bogus_dsq_fail.bpf.c | 14 s32 prev_cpu, u64 wake_flags) in BPF_STRUCT_OPS() argument 28 return prev_cpu; in BPF_STRUCT_OPS()
|
| H A D | ddsp_vtimelocal_fail.bpf.c | 14 s32 prev_cpu, u64 wake_flags) in BPF_STRUCT_OPS() argument 25 return prev_cpu; in BPF_STRUCT_OPS()
|
| H A D | dsp_local_on.bpf.c | 20 s32 prev_cpu, u64 wake_flags) in BPF_STRUCT_OPS() argument 22 return prev_cpu; in BPF_STRUCT_OPS()
|
| H A D | exit.bpf.c | 21 s32 prev_cpu, u64 wake_flags) in BPF_STRUCT_OPS() argument 28 return scx_bpf_select_cpu_dfl(p, prev_cpu, wake_flags, &found); in BPF_STRUCT_OPS()
|
| H A D | select_cpu_vtime.bpf.c | 39 s32 prev_cpu, u64 wake_flags) in BPF_STRUCT_OPS() argument 47 cpu = prev_cpu; in BPF_STRUCT_OPS()
|
| H A D | maximal.bpf.c | 17 s32 BPF_STRUCT_OPS(maximal_select_cpu, struct task_struct *p, s32 prev_cpu, in BPF_STRUCT_OPS() argument 20 return prev_cpu; in BPF_STRUCT_OPS()
|
| H A D | numa.bpf.c | 35 struct task_struct *p, s32 prev_cpu, u64 wake_flags) in BPF_STRUCT_OPS() argument
|
| /linux-6.15/drivers/irqchip/ |
| H A D | irq-loongarch-avec.c | 54 unsigned int prev_cpu; member 92 if (cpu_online(adata->prev_cpu)) { in avecintc_sync() 93 plist = per_cpu_ptr(&pending_list, adata->prev_cpu); in avecintc_sync() 96 mp_ops.send_ipi_single(adata->prev_cpu, ACTION_CLEAR_VECTOR); in avecintc_sync() 175 cpu = adata->prev_cpu; in complete_irq_moving() 201 adata->prev_cpu = adata->cpu; in complete_irq_moving() 264 adata->prev_cpu = adata->cpu = cpu; in avecintc_alloc_vector() 308 per_cpu(irq_map, adata->prev_cpu)[adata->prev_vec] = NULL; in avecintc_free_vector() 309 irq_matrix_free(loongarch_avec.vector_matrix, adata->prev_cpu, adata->prev_vec, false); in avecintc_free_vector()
|
| /linux-6.15/arch/powerpc/lib/ |
| H A D | qspinlock.c | 260 static struct qnode *get_tail_qnode(struct qspinlock *lock, int prev_cpu) in get_tail_qnode() argument 262 struct qnodes *qnodesp = per_cpu_ptr(&qnodes, prev_cpu); in get_tail_qnode() 376 static __always_inline bool yield_to_prev(struct qspinlock *lock, struct qnode *node, int prev_cpu,… in yield_to_prev() argument 391 if (node->sleepy || vcpu_is_preempted(prev_cpu)) { in yield_to_prev() 418 yield_count = yield_count_of(prev_cpu); in yield_to_prev() 430 yield_to_preempted(prev_cpu, yield_count); in yield_to_prev() 575 int prev_cpu = decode_tail_cpu(old); in queued_spin_lock_mcs_queue() local 576 struct qnode *prev = get_tail_qnode(lock, prev_cpu); in queued_spin_lock_mcs_queue() 586 if (yield_to_prev(lock, node, prev_cpu, paravirt)) in queued_spin_lock_mcs_queue()
|
| /linux-6.15/tools/sched_ext/ |
| H A D | scx_simple.bpf.c | 55 s32 BPF_STRUCT_OPS(simple_select_cpu, struct task_struct *p, s32 prev_cpu, u64 wake_flags) in BPF_STRUCT_OPS() argument 60 cpu = scx_bpf_select_cpu_dfl(p, prev_cpu, wake_flags, &is_idle); in BPF_STRUCT_OPS()
|
| H A D | scx_qmap.bpf.c | 134 static s32 pick_direct_dispatch_cpu(struct task_struct *p, s32 prev_cpu) in pick_direct_dispatch_cpu() argument 139 scx_bpf_test_and_clear_cpu_idle(prev_cpu)) in pick_direct_dispatch_cpu() 140 return prev_cpu; in pick_direct_dispatch_cpu() 161 s32 prev_cpu, u64 wake_flags) in BPF_STRUCT_OPS() argument 169 cpu = pick_direct_dispatch_cpu(p, prev_cpu); in BPF_STRUCT_OPS() 175 return prev_cpu; in BPF_STRUCT_OPS()
|
| /linux-6.15/Documentation/translations/zh_CN/scheduler/ |
| H A D | sched-energy.rst | 118 依然放在之前活动的prev_cpu是否可以节省能量。 132 prev_cpu = 0(上一次运行在CPU0)。 197 **情况3. P依旧留在prev_cpu/CPU0**::
|
| /linux-6.15/arch/sparc/kernel/ |
| H A D | cpumap.c | 193 int n, id, cpu, prev_cpu, last_cpu, level; in build_cpuinfo_tree() local 204 prev_cpu = cpu = cpumask_first(cpu_online_mask); in build_cpuinfo_tree() 268 (cpu == last_cpu) ? cpu : prev_cpu; in build_cpuinfo_tree() 290 prev_cpu = cpu; in build_cpuinfo_tree()
|
| /linux-6.15/arch/riscv/include/asm/ |
| H A D | switch_to.h | 99 bool thread_migrated = smp_processor_id() != task->thread.prev_cpu; in switch_to_should_flush_icache() 108 #define __set_prev_cpu(thread) ((thread).prev_cpu = smp_processor_id())
|
| H A D | processor.h | 115 unsigned int prev_cpu; member
|
| /linux-6.15/arch/x86/kernel/apic/ |
| H A D | vector.c | 31 unsigned int prev_cpu; member 174 apicd->prev_cpu = apicd->cpu; in apic_update_vector() 357 apicd->prev_cpu); in clear_irq_vector() 368 per_cpu(vector_irq, apicd->prev_cpu)[vector] = VECTOR_SHUTDOWN; in clear_irq_vector() 369 irq_matrix_free(vector_matrix, apicd->prev_cpu, vector, managed); in clear_irq_vector() 648 seq_printf(m, "%*sPrevious target: %5u\n", ind, "", apicd.prev_cpu); in x86_vector_debug_show() 894 unsigned int cpu = apicd->prev_cpu; in free_moved_vector() 934 if (!vector || (apicd->cpu != cpu && apicd->prev_cpu != cpu)) in apic_force_complete_move() 1088 unsigned int cpu = apicd->prev_cpu; in __vector_schedule_cleanup()
|