Home
last modified time | relevance | path

Searched refs:cpus_ptr (Results 1 – 25 of 30) sorted by relevance

12

/linux-6.15/tools/testing/selftests/sched_ext/
H A Dnuma.bpf.c45 cpu = __COMPAT_scx_bpf_pick_idle_cpu_node(p->cpus_ptr, node, in BPF_STRUCT_OPS()
48 cpu = __COMPAT_scx_bpf_pick_any_cpu_node(p->cpus_ptr, node, in BPF_STRUCT_OPS()
H A Dselect_cpu_dispatch.bpf.c24 cpu = scx_bpf_pick_idle_cpu(p->cpus_ptr, 0); in BPF_STRUCT_OPS()
H A Dddsp_bogus_dsq_fail.bpf.c16 s32 cpu = scx_bpf_pick_idle_cpu(p->cpus_ptr, 0); in BPF_STRUCT_OPS()
H A Dddsp_vtimelocal_fail.bpf.c16 s32 cpu = scx_bpf_pick_idle_cpu(p->cpus_ptr, 0); in BPF_STRUCT_OPS()
H A Dselect_cpu_vtime.bpf.c43 cpu = scx_bpf_pick_idle_cpu(p->cpus_ptr, 0); in BPF_STRUCT_OPS()
/linux-6.15/samples/trace_events/
H A Dtrace-events-sample.c41 current->cpus_ptr, fmt, &va); in do_simple_thread_func()
53 trace_foo_rel_loc("Hello __rel_loc", cnt, bitmask, current->cpus_ptr); in do_simple_thread_func()
/linux-6.15/tools/testing/selftests/bpf/progs/
H A Dcpumask_failure.c69 cpumask = bpf_cpumask_acquire((struct bpf_cpumask *)task->cpus_ptr); in BPF_PROG()
80 bpf_cpumask_set_cpu(0, (struct bpf_cpumask *)task->cpus_ptr); in BPF_PROG()
H A Dnested_trust_success.c24 bpf_cpumask_test_cpu(0, task->cpus_ptr); in BPF_PROG()
/linux-6.15/tools/sched_ext/
H A Dscx_qmap.bpf.c142 cpu = scx_bpf_pick_idle_cpu(p->cpus_ptr, 0); in pick_direct_dispatch_cpu()
251 cpu = scx_bpf_pick_idle_cpu(p->cpus_ptr, 0); in BPF_STRUCT_OPS()
338 if (bpf_cpumask_test_cpu(this_cpu, p->cpus_ptr)) in dispatch_highpri()
341 cpu = scx_bpf_pick_any_cpu(p->cpus_ptr, 0); in dispatch_highpri()
H A Dscx_central.bpf.c154 if (!bpf_cpumask_test_cpu(cpu, p->cpus_ptr)) { in dispatch_to_cpu()
/linux-6.15/kernel/sched/
H A Dext_idle.c493 if (cpumask_test_cpu(cpu, p->cpus_ptr)) in scx_select_cpu_dfl()
538 cpu = scx_pick_idle_cpu(p->cpus_ptr, node, flags | SCX_PICK_IDLE_CORE); in scx_select_cpu_dfl()
586 cpu = scx_pick_idle_cpu(p->cpus_ptr, node, flags); in scx_select_cpu_dfl()
H A Dfair.c2151 !cpumask_test_cpu(cpu, env->p->cpus_ptr)) in update_numa_stats()
2183 !cpumask_test_cpu(cpu, env->p->cpus_ptr)) { in task_numa_assign()
2296 if (!cpumask_test_cpu(env->src_cpu, cur->cpus_ptr)) in task_numa_compare()
2495 if (!cpumask_test_cpu(cpu, env->p->cpus_ptr)) in task_numa_find_cpu()
7700 cpumask_and(cpus, sched_domain_span(sd), p->cpus_ptr); in select_idle_cpu()
7893 cpumask_test_cpu(recent_used_cpu, p->cpus_ptr) && in select_idle_sibling()
8444 if (!cpumask_test_cpu(cpu, p->cpus_ptr)) in find_energy_efficient_cpu()
8593 cpumask_test_cpu(cpu, p->cpus_ptr)) in select_task_rq_fair()
8709 if (!cpumask_intersects(p->cpus_ptr, cpumask)) in set_task_max_allowed_capacity()
9418 if (!cpumask_test_cpu(env->dst_cpu, p->cpus_ptr)) { in can_migrate_task()
[all …]
H A Dcore.c2361 if (p->cpus_ptr != &p->cpus_mask) in migrate_disable_switch()
2418 if (p->cpus_ptr != &p->cpus_mask) in migrate_enable()
2624 if (cpumask_test_cpu(task_cpu(p), p->cpus_ptr)) { in migration_cpu_stop()
2702 p->cpus_ptr = ctx->new_mask; in set_cpus_allowed_common()
3406 if (!cpumask_test_cpu(arg->dst_cpu, arg->src_task->cpus_ptr)) in migrate_swap_stop()
3444 if (!cpumask_test_cpu(arg.dst_cpu, arg.src_task->cpus_ptr)) in migrate_swap()
3447 if (!cpumask_test_cpu(arg.src_cpu, arg.dst_task->cpus_ptr)) in migrate_swap()
3527 for_each_cpu(dest_cpu, p->cpus_ptr) { in select_fallback_rq()
3586 cpu = cpumask_any(p->cpus_ptr); in select_task_rq()
3940 if (!cpumask_test_cpu(cpu, p->cpus_ptr)) in ttwu_queue_cond()
[all …]
H A Dsyscalls.c654 if (!cpumask_subset(span, p->cpus_ptr) || in __sched_setscheduler()
H A Dext.c2478 WARN_ON_ONCE(!cpumask_test_cpu(cpu_of(dst_rq), p->cpus_ptr)); in move_remote_task_to_local_dsq()
3467 p, (struct cpumask *)p->cpus_ptr); in set_cpus_allowed_scx()
3891 p, (struct cpumask *)p->cpus_ptr); in switching_to_scx()
5000 dump_line(s, " cpus=%*pb", cpumask_pr_args(p->cpus_ptr)); in scx_dump_task()
H A Dsched.h1443 for_each_cpu_and(cpu, sched_group_span(group), p->cpus_ptr) { in sched_group_cookie_match()
2570 if (!cpumask_test_cpu(cpu, p->cpus_ptr)) in task_allowed_on_cpu()
/linux-6.15/Documentation/bpf/
H A Dcpumasks.rst105 if (!bpf_cpumask_full(task->cpus_ptr))
108 bpf_cpumask_copy(cpumask, task->cpus_ptr);
243 /* struct cpumask * pointers such as task->cpus_ptr can also be queried. */
244 if (bpf_cpumask_test_cpu(0, task->cpus_ptr))
/linux-6.15/arch/mips/kernel/
H A Dmips-mt-fpaff.c181 cpumask_or(&allowed, &p->thread.user_cpus_allowed, p->cpus_ptr); in mipsmt_sys_sched_getaffinity()
/linux-6.15/init/
H A Dinit_task.c79 .cpus_ptr = &init_task.cpus_mask,
/linux-6.15/kernel/trace/
H A Dtrace_hwlat.c323 if (!cpumask_equal(current_mask, current->cpus_ptr)) in move_to_next_cpu()
/linux-6.15/arch/x86/kernel/cpu/resctrl/
H A Dpseudo_lock.c1528 if (!cpumask_subset(current->cpus_ptr, &plr->d->hdr.cpu_mask)) { in pseudo_lock_dev_mmap()
/linux-6.15/arch/powerpc/platforms/cell/spufs/
H A Dsched.c131 cpumask_copy(&ctx->cpus_allowed, current->cpus_ptr); in __spu_update_sched_info()
/linux-6.15/drivers/infiniband/hw/hfi1/
H A Daffinity.c1000 *proc_mask = current->cpus_ptr; in hfi1_get_proc_affinity()
/linux-6.15/include/linux/
H A Dsched.h913 const cpumask_t *cpus_ptr; member
/linux-6.15/kernel/
H A Dfork.c1169 if (orig->cpus_ptr == &orig->cpus_mask) in dup_task_struct()
1170 tsk->cpus_ptr = &tsk->cpus_mask; in dup_task_struct()

12