| /linux-6.15/tools/sched_ext/include/scx/ |
| H A D | compat.bpf.h | 42 #define scx_bpf_dsq_insert(p, dsq_id, slice, enq_flags) \ argument 44 scx_bpf_dsq_insert((p), (dsq_id), (slice), (enq_flags)) : \ 45 scx_bpf_dispatch___compat((p), (dsq_id), (slice), (enq_flags))) 47 #define scx_bpf_dsq_insert_vtime(p, dsq_id, slice, vtime, enq_flags) \ argument 71 #define __COMPAT_scx_bpf_dsq_move(it__iter, p, dsq_id, enq_flags) \ argument 73 scx_bpf_dsq_move((it__iter), (p), (dsq_id), (enq_flags)) : \ 80 scx_bpf_dsq_move_vtime((it__iter), (p), (dsq_id), (enq_flags)) : \ 85 #define scx_bpf_dispatch(p, dsq_id, slice, enq_flags) \ argument 88 #define scx_bpf_dispatch_vtime(p, dsq_id, slice, vtime, enq_flags) \ argument 136 static inline bool __COMPAT_is_enq_cpu_selected(u64 enq_flags) in __COMPAT_is_enq_cpu_selected() argument [all …]
|
| H A D | common.bpf.h | 51 void scx_bpf_dsq_insert(struct task_struct *p, u64 dsq_id, u64 slice, u64 enq_flags) __ksym __weak; 52 …insert_vtime(struct task_struct *p, u64 dsq_id, u64 slice, u64 vtime, u64 enq_flags) __ksym __weak; 58 …struct bpf_iter_scx_dsq *it__iter, struct task_struct *p, u64 dsq_id, u64 enq_flags) __ksym __weak; 59 …struct bpf_iter_scx_dsq *it__iter, struct task_struct *p, u64 dsq_id, u64 enq_flags) __ksym __weak;
|
| /linux-6.15/kernel/sched/ |
| H A D | ext.c | 1000 u64 enq_flags; member 1853 u64 enq_flags) in dispatch_enqueue() argument 2066 u64 enq_flags) in mark_direct_dispatch() argument 2955 ent->enq_flags); in flush_dispatch_buf() 6125 u64 enq_flags) in scx_dsq_insert_commit() argument 6145 .enq_flags = enq_flags, in scx_dsq_insert_commit() 6186 u64 enq_flags) in scx_bpf_dsq_insert() argument 6201 u64 enq_flags) in scx_bpf_dispatch() argument 6515 u64 enq_flags) in scx_bpf_dsq_move() argument 6549 u64 enq_flags) in scx_bpf_dsq_move_vtime() argument [all …]
|
| /linux-6.15/tools/sched_ext/ |
| H A D | scx_simple.bpf.c | 69 void BPF_STRUCT_OPS(simple_enqueue, struct task_struct *p, u64 enq_flags) in BPF_STRUCT_OPS() argument 74 scx_bpf_dsq_insert(p, SHARED_DSQ, SCX_SLICE_DFL, enq_flags); in BPF_STRUCT_OPS() 86 enq_flags); in BPF_STRUCT_OPS()
|
| H A D | scx_qmap.bpf.c | 194 void BPF_STRUCT_OPS(qmap_enqueue, struct task_struct *p, u64 enq_flags) in BPF_STRUCT_OPS() argument 229 scx_bpf_dsq_insert(p, SCX_DSQ_LOCAL, slice_ns, enq_flags); in BPF_STRUCT_OPS() 234 if (!__COMPAT_is_enq_cpu_selected(enq_flags) && in BPF_STRUCT_OPS() 237 scx_bpf_dsq_insert(p, SCX_DSQ_LOCAL_ON | cpu, slice_ns, enq_flags); in BPF_STRUCT_OPS() 247 if (enq_flags & SCX_ENQ_REENQ) { in BPF_STRUCT_OPS() 250 scx_bpf_dsq_insert(p, SHARED_DSQ, 0, enq_flags); in BPF_STRUCT_OPS() 265 scx_bpf_dsq_insert(p, SHARED_DSQ, slice_ns, enq_flags); in BPF_STRUCT_OPS()
|
| H A D | scx_central.bpf.c | 102 void BPF_STRUCT_OPS(central_enqueue, struct task_struct *p, u64 enq_flags) in BPF_STRUCT_OPS() argument 117 enq_flags | SCX_ENQ_PREEMPT); in BPF_STRUCT_OPS() 123 scx_bpf_dsq_insert(p, FALLBACK_DSQ_ID, SCX_SLICE_INF, enq_flags); in BPF_STRUCT_OPS()
|
| H A D | scx_flatcg.bpf.c | 345 void BPF_STRUCT_OPS(fcg_enqueue, struct task_struct *p, u64 enq_flags) in BPF_STRUCT_OPS() argument 376 enq_flags); in BPF_STRUCT_OPS() 380 enq_flags); in BPF_STRUCT_OPS() 391 scx_bpf_dsq_insert(p, cgrp->kn->id, SCX_SLICE_DFL, enq_flags); in BPF_STRUCT_OPS() 403 tvtime, enq_flags); in BPF_STRUCT_OPS() 507 void BPF_STRUCT_OPS(fcg_runnable, struct task_struct *p, u64 enq_flags) in BPF_STRUCT_OPS() argument
|
| /linux-6.15/tools/testing/selftests/sched_ext/ |
| H A D | select_cpu_dfl.bpf.c | 23 u64 enq_flags) in BPF_STRUCT_OPS() argument 33 scx_bpf_dsq_insert(p, SCX_DSQ_GLOBAL, SCX_SLICE_DFL, enq_flags); in BPF_STRUCT_OPS()
|
| H A D | enq_select_cpu_fails.bpf.c | 23 u64 enq_flags) in BPF_STRUCT_OPS() argument 34 scx_bpf_dsq_insert(p, SCX_DSQ_GLOBAL, SCX_SLICE_DFL, enq_flags); in BPF_STRUCT_OPS()
|
| H A D | maximal.bpf.c | 23 void BPF_STRUCT_OPS(maximal_enqueue, struct task_struct *p, u64 enq_flags) in BPF_STRUCT_OPS() argument 25 scx_bpf_dsq_insert(p, DSQ_ID, SCX_SLICE_DFL, enq_flags); in BPF_STRUCT_OPS() 36 void BPF_STRUCT_OPS(maximal_runnable, struct task_struct *p, u64 enq_flags) in BPF_STRUCT_OPS() argument
|
| H A D | exit.bpf.c | 31 void BPF_STRUCT_OPS(exit_enqueue, struct task_struct *p, u64 enq_flags) in BPF_STRUCT_OPS() argument 36 scx_bpf_dsq_insert(p, DSQ_ID, SCX_SLICE_DFL, enq_flags); in BPF_STRUCT_OPS()
|
| H A D | select_cpu_dfl_nodispatch.bpf.c | 53 u64 enq_flags) in BPF_STRUCT_OPS() argument 70 scx_bpf_dsq_insert(p, dsq_id, SCX_SLICE_DFL, enq_flags); in BPF_STRUCT_OPS()
|
| H A D | numa.bpf.c | 60 void BPF_STRUCT_OPS(numa_enqueue, struct task_struct *p, u64 enq_flags) in BPF_STRUCT_OPS() argument 64 scx_bpf_dsq_insert(p, node, SCX_SLICE_DFL, enq_flags); in BPF_STRUCT_OPS()
|
| H A D | dsp_local_on.bpf.c | 26 u64 enq_flags) in BPF_STRUCT_OPS() argument
|
| /linux-6.15/Documentation/scheduler/ |
| H A D | sched-ext.rst | 166 void BPF_STRUCT_OPS(simple_enqueue, struct task_struct *p, u64 enq_flags) 168 scx_bpf_dsq_insert(p, SCX_DSQ_GLOBAL, SCX_SLICE_DFL, enq_flags);
|