Home
last modified time | relevance | path

Searched refs:exec (Results 1 – 25 of 259) sorted by relevance

1234567891011

/linux-6.15/drivers/gpu/drm/
H A Ddrm_exec.c65 exec->prelocked = NULL; in drm_exec_unlock_all()
84 exec->flags = flags; in drm_exec_init()
88 exec->max_objects = exec->objects ? nr : 0; in drm_exec_init()
89 exec->num_objects = 0; in drm_exec_init()
91 exec->prelocked = NULL; in drm_exec_init()
105 kvfree(exec->objects); in drm_exec_fini()
135 exec->num_objects = 0; in drm_exec_cleanup()
144 if (unlikely(exec->num_objects == exec->max_objects)) { in drm_exec_obj_locked()
152 exec->objects = tmp; in drm_exec_obj_locked()
156 exec->objects[exec->num_objects++] = obj; in drm_exec_obj_locked()
[all …]
H A Ddrm_gpuvm.c1112 struct drm_exec *exec, in drm_gpuvm_prepare_vm() argument
1121 struct drm_exec *exec, in __drm_gpuvm_prepare_objects() argument
1142 struct drm_exec *exec, in drm_gpuvm_prepare_objects_locked() argument
1186 struct drm_exec *exec, in drm_gpuvm_prepare_objects() argument
1249 struct drm_exec *exec = &vm_exec->exec; in drm_gpuvm_exec_lock() local
1255 drm_exec_until_all_locked(exec) { in drm_gpuvm_exec_lock()
1277 drm_exec_fini(exec); in drm_gpuvm_exec_lock()
1341 struct drm_exec *exec = &vm_exec->exec; in drm_gpuvm_exec_lock_range() local
1346 drm_exec_until_all_locked(exec) { in drm_gpuvm_exec_lock_range()
1357 drm_exec_fini(exec); in drm_gpuvm_exec_lock_range()
[all …]
/linux-6.15/drivers/gpu/drm/tests/
H A Ddrm_exec_test.c47 struct drm_exec exec; in sanitycheck() local
50 drm_exec_fini(&exec); in sanitycheck()
58 struct drm_exec exec; in test_lock() local
71 drm_exec_fini(&exec); in test_lock()
78 struct drm_exec exec; in test_lock_unlock() local
98 drm_exec_fini(&exec); in test_lock_unlock()
105 struct drm_exec exec; in test_duplicates() local
125 drm_exec_fini(&exec); in test_duplicates()
132 struct drm_exec exec; in test_prepare() local
145 drm_exec_fini(&exec); in test_prepare()
[all …]
/linux-6.15/drivers/gpu/drm/vc4/
H A Dvc4_gem.c172 if (!exec[0] && !exec[1]) { in vc4_save_hang_state()
483 if (!exec) in vc4_submit_next_bin_job()
497 if (exec->ct0ca != exec->ct0ea) { in vc4_submit_next_bin_job()
500 submit_cl(dev, 0, exec->ct0ca, exec->ct0ea); in vc4_submit_next_bin_job()
523 if (!exec) in vc4_submit_next_render_job()
537 trace_vc4_submit_cl(dev, true, exec->seqno, exec->ct1ca, exec->ct1ea); in vc4_submit_next_render_job()
538 submit_cl(dev, 1, exec->ct1ca, exec->ct1ea); in vc4_submit_next_render_job()
696 exec->bo_count, &exec->bo); in vc4_cl_lookup_bos()
723 for (i = 0; i < exec->bo_count && exec->bo[i]; i++) in vc4_cl_lookup_bos()
811 exec->ct0ca = exec->exec_bo->dma_addr + bin_offset; in vc4_get_bcl()
[all …]
H A Dvc4_validate.c135 return vc4_use_bo(exec, exec->bo_index[gem_handles_packet_index]); in vc4_use_handle()
224 if (!validate_bin_pos(exec, untrusted, exec->args->bin_cl_size - 1)) { in validate_flush()
253 if (!validate_bin_pos(exec, untrusted, exec->args->bin_cl_size - 2)) { in validate_increment_semaphore()
278 shader_state = &exec->shader_state[exec->shader_state_count - 1]; in validate_indexed_prim_list()
312 shader_state = &exec->shader_state[exec->shader_state_count - 1]; in validate_gl_array_primitive()
371 tile_count = exec->bin_tiles_x * exec->bin_tiles_y; in validate_tile_binning_config()
377 exec->bin_tiles_x, exec->bin_tiles_y); in validate_tile_binning_config()
433 memcpy(exec->bo_index, untrusted, sizeof(exec->bo_index)); in validate_gem_handles()
546 exec->ct0ea = exec->ct0ca + dst_offset; in vc4_validate_bin_cl()
880 ((uint32_t *)exec->uniforms_v)[o] = exec->uniforms_p; in validate_gl_shader_rec()
[all …]
H A Dvc4_irq.c67 struct vc4_exec_info *exec; in vc4_overflow_mem_work() local
92 exec = vc4_first_bin_job(vc4); in vc4_overflow_mem_work()
93 if (!exec) in vc4_overflow_mem_work()
95 if (exec) { in vc4_overflow_mem_work()
122 if (!exec) in vc4_irq_finish_bin_job()
144 if (!exec) in vc4_cancel_bin_job()
148 if (exec->perfmon) in vc4_cancel_bin_job()
162 if (!exec) in vc4_irq_finish_render_job()
192 if (exec->fence) { in vc4_irq_finish_render_job()
194 dma_fence_put(exec->fence); in vc4_irq_finish_render_job()
[all …]
H A Dvc4_render_cl.c186 (y * exec->bin_tiles_x + x) * 32)); in emit_tile()
335 &exec->unref_list); in vc4_create_rcl_bo()
377 exec->ct1ca = setup->rcl->dma_addr; in vc4_create_rcl_bo()
421 *obj = vc4_use_bo(exec, surf->hindex); in vc4_rcl_msaa_surface_setup()
425 exec->rcl_write_bo[exec->rcl_write_bo_count++] = *obj; in vc4_rcl_msaa_surface_setup()
462 exec->rcl_write_bo[exec->rcl_write_bo_count++] = *obj; in vc4_rcl_surface_setup()
465 if (surf == &exec->args->zs_write) { in vc4_rcl_surface_setup()
526 exec->args->width, exec->args->height, cpp)) { in vc4_rcl_surface_setup()
566 exec->rcl_write_bo[exec->rcl_write_bo_count++] = *obj; in vc4_rcl_render_config_surface_setup()
587 exec->args->width, exec->args->height, cpp)) { in vc4_rcl_render_config_surface_setup()
[all …]
/linux-6.15/include/drm/
H A Ddrm_exec.h65 return index < exec->num_objects ? exec->objects[index] : NULL; in drm_exec_obj()
91 for ((index) = (exec)->num_objects - 1; \
92 ((obj) = drm_exec_obj(exec, index)); --(index))
105 #define drm_exec_until_all_locked(exec) \ argument
110 drm_exec_cleanup(exec); \
120 #define drm_exec_retry_on_contention(exec) \ argument
122 if (unlikely(drm_exec_is_contended(exec))) \
135 return !!exec->contended; in drm_exec_is_contended()
139 void drm_exec_fini(struct drm_exec *exec);
140 bool drm_exec_cleanup(struct drm_exec *exec);
[all …]
H A Ddrm_gpuvm.h516 struct drm_exec exec; member
553 struct drm_exec *exec,
557 struct drm_exec *exec,
561 struct drm_exec *exec,
586 drm_exec_fini(&vm_exec->exec); in drm_gpuvm_exec_unlock()
589 int drm_gpuvm_validate(struct drm_gpuvm *gpuvm, struct drm_exec *exec);
591 struct drm_exec *exec,
611 drm_gpuvm_resv_add_fence(vm_exec->vm, &vm_exec->exec, fence, in drm_gpuvm_exec_resv_add_fence()
626 return drm_gpuvm_validate(vm_exec->vm, &vm_exec->exec); in drm_gpuvm_exec_validate()
1160 struct drm_exec *exec);
/linux-6.15/include/uapi/linux/
H A Da.out.h44 #define N_MAGIC(exec) ((exec).a_info & 0xffff) argument
46 #define N_MACHTYPE(exec) ((enum machine_type)(((exec).a_info >> 16) & 0xff)) argument
47 #define N_FLAGS(exec) (((exec).a_info >> 24) & 0xff) argument
48 #define N_SET_INFO(exec, magic, type, flags) \ argument
49 ((exec).a_info = ((magic) & 0xffff) \
52 #define N_SET_MAGIC(exec, magic) \ argument
53 ((exec).a_info = (((exec).a_info & 0xffff0000) | ((magic) & 0xffff)))
55 #define N_SET_MACHTYPE(exec, machtype) \ argument
56 ((exec).a_info = \
59 #define N_SET_FLAGS(exec, flags) \ argument
[all …]
/linux-6.15/drivers/gpu/drm/amd/amdgpu/
H A Damdgpu_csa.c69 struct drm_exec exec; in amdgpu_map_static_csa() local
73 drm_exec_until_all_locked(&exec) { in amdgpu_map_static_csa()
74 r = amdgpu_vm_lock_pd(vm, &exec, 0); in amdgpu_map_static_csa()
77 drm_exec_retry_on_contention(&exec); in amdgpu_map_static_csa()
101 drm_exec_fini(&exec); in amdgpu_map_static_csa()
109 struct drm_exec exec; in amdgpu_unmap_static_csa() local
112 drm_exec_init(&exec, 0, 0); in amdgpu_unmap_static_csa()
113 drm_exec_until_all_locked(&exec) { in amdgpu_unmap_static_csa()
114 r = amdgpu_vm_lock_pd(vm, &exec, 0); in amdgpu_unmap_static_csa()
117 drm_exec_retry_on_contention(&exec); in amdgpu_unmap_static_csa()
[all …]
H A Damdgpu_seq64.c67 struct drm_exec exec; in amdgpu_seq64_map() local
76 drm_exec_until_all_locked(&exec) { in amdgpu_seq64_map()
77 r = amdgpu_vm_lock_pd(vm, &exec, 0); in amdgpu_seq64_map()
79 r = drm_exec_lock_obj(&exec, &bo->tbo.base); in amdgpu_seq64_map()
80 drm_exec_retry_on_contention(&exec); in amdgpu_seq64_map()
108 drm_exec_fini(&exec); in amdgpu_seq64_map()
124 struct drm_exec exec; in amdgpu_seq64_unmap() local
137 drm_exec_until_all_locked(&exec) { in amdgpu_seq64_unmap()
138 r = amdgpu_vm_lock_pd(vm, &exec, 0); in amdgpu_seq64_unmap()
141 drm_exec_retry_on_contention(&exec); in amdgpu_seq64_unmap()
[all …]
H A Damdgpu_gem.c234 struct drm_exec exec; in amdgpu_gem_object_close() local
238 drm_exec_until_all_locked(&exec) { in amdgpu_gem_object_close()
240 drm_exec_retry_on_contention(&exec); in amdgpu_gem_object_close()
244 r = amdgpu_vm_lock_pd(vm, &exec, 0); in amdgpu_gem_object_close()
245 drm_exec_retry_on_contention(&exec); in amdgpu_gem_object_close()
272 drm_exec_fini(&exec); in amdgpu_gem_object_close()
716 struct drm_exec exec; in amdgpu_gem_va_ioctl() local
779 drm_exec_until_all_locked(&exec) { in amdgpu_gem_va_ioctl()
781 r = drm_exec_lock_obj(&exec, gobj); in amdgpu_gem_va_ioctl()
788 drm_exec_retry_on_contention(&exec); in amdgpu_gem_va_ioctl()
[all …]
/linux-6.15/tools/testing/selftests/exec/
H A Dcheck-exec.c269 const int exec = variant->mount_exec && variant->file_exec; in TEST_F() local
271 test_exec_path(_metadata, reg_file_path, exec ? 0 : EACCES); in TEST_F()
274 test_exec_path(_metadata, reg_file_path, exec ? 0 : EACCES); in TEST_F()
279 const int exec = variant->mount_exec && variant->file_exec; in TEST_F() local
292 const int exec = variant->file_exec; in TEST_F() local
294 test_exec_fd(_metadata, self->memfd, exec ? 0 : EACCES); in TEST_F()
297 test_exec_fd(_metadata, self->memfd, exec ? 0 : EACCES); in TEST_F()
302 const int exec = variant->file_exec; in TEST_F() local
306 test_exec_fd(_metadata, self->memfd, exec ? 0 : EACCES); in TEST_F()
309 test_exec_fd(_metadata, self->memfd, exec ? 0 : EACCES); in TEST_F()
[all …]
H A Dnull-argv.c10 #define FORK(exec) \ argument
15 exec; /* Some kind of exec */ \
16 perror("# " #exec); \
19 check_result(pid, #exec); \
/linux-6.15/drivers/gpu/drm/nouveau/nvkm/subdev/timer/
H A Dbase.c73 LIST_HEAD(exec); in nvkm_timer_alarm_trigger()
90 list_add(&alarm->exec, &exec); in nvkm_timer_alarm_trigger()
99 list_for_each_entry_safe(alarm, atemp, &exec, exec) { in nvkm_timer_alarm_trigger()
100 list_del(&alarm->exec); in nvkm_timer_alarm_trigger()
/linux-6.15/tools/perf/util/
H A Dcomm.h15 bool exec; member
23 struct comm *comm__new(const char *str, u64 timestamp, bool exec);
26 bool exec);
H A Dcomm.c190 struct comm *comm__new(const char *str, u64 timestamp, bool exec) in comm__new() argument
198 comm->exec = exec; in comm__new()
209 int comm__override(struct comm *comm, const char *str, u64 timestamp, bool exec) in comm__override() argument
220 if (exec) in comm__override()
221 comm->exec = true; in comm__override()
/linux-6.15/drivers/sbus/char/
H A Doradax.c861 ctx->result.exec.status = DAX_SUBMIT_ERR_THR_INIT; in dax_ccb_exec()
893 if (ctx->result.exec.status != DAX_SUBMIT_OK) in dax_ccb_exec()
897 &ctx->result.exec.status_data); in dax_ccb_exec()
898 if (ctx->result.exec.status != DAX_SUBMIT_OK) in dax_ccb_exec()
906 &accepted_len, &ctx->result.exec.status_data); in dax_ccb_exec()
919 ctx->result.exec.status = DAX_SUBMIT_OK; in dax_ccb_exec()
935 ctx->result.exec.status = DAX_SUBMIT_ERR_NOMAP; in dax_ccb_exec()
954 ctx->result.exec.status = DAX_SUBMIT_ERR_NOACCESS; in dax_ccb_exec()
963 ctx->result.exec.status = DAX_SUBMIT_ERR_UNAVAIL; in dax_ccb_exec()
966 ctx->result.exec.status = DAX_SUBMIT_ERR_INTERNAL; in dax_ccb_exec()
[all …]
/linux-6.15/tools/testing/selftests/bpf/progs/
H A Dtest_overhead.c13 int BPF_KPROBE(prog1, struct task_struct *tsk, const char *buf, bool exec) in BPF_KPROBE() argument
31 int BPF_PROG(prog4, struct task_struct *tsk, const char *buf, bool exec) in BPF_PROG() argument
37 int BPF_PROG(prog5, struct task_struct *tsk, const char *buf, bool exec) in BPF_PROG() argument
/linux-6.15/arch/sparc/mm/
H A Dtlb.c73 bool exec, unsigned int hugepage_shift) in tlb_batch_add_one() argument
79 if (exec) in tlb_batch_add_one()
162 bool exec = pte_exec(*pte); in tlb_batch_pmd_scan() local
164 tlb_batch_add_one(mm, vaddr, exec, PAGE_SHIFT); in tlb_batch_pmd_scan()
214 bool exec = pte_exec(orig_pte); in __set_pmd_acct() local
216 tlb_batch_add_one(mm, addr, exec, REAL_HPAGE_SHIFT); in __set_pmd_acct()
217 tlb_batch_add_one(mm, addr + REAL_HPAGE_SIZE, exec, in __set_pmd_acct()
/linux-6.15/drivers/gpu/drm/xe/
H A Dxe_gt_pagefault.c105 err = xe_vm_lock_vma(exec, vma); in xe_pf_begin()
134 struct drm_exec exec; in handle_vma_pagefault() local
161 drm_exec_init(&exec, 0, 0); in handle_vma_pagefault()
162 drm_exec_until_all_locked(&exec) { in handle_vma_pagefault()
164 drm_exec_retry_on_contention(&exec); in handle_vma_pagefault()
186 drm_exec_fini(&exec); in handle_vma_pagefault()
569 struct drm_exec exec; in handle_acc() local
598 drm_exec_init(&exec, 0, 0); in handle_acc()
599 drm_exec_until_all_locked(&exec) { in handle_acc()
601 drm_exec_retry_on_contention(&exec); in handle_acc()
[all …]
H A Dxe_exec.c102 return xe_vm_validate_rebind(vm, &vm_exec->exec, 1); in xe_exec_fn()
116 struct drm_exec *exec = &vm_exec.exec; in xe_exec_ioctl() local
243 drm_exec_init(exec, vm_exec.flags, 0); in xe_exec_ioctl()
247 if (xe_vm_validate_should_retry(exec, err, &end)) in xe_exec_ioctl()
312 drm_gpuvm_resv_add_fence(&vm->gpuvm, exec, &job->drm.s_fence->finished, in xe_exec_ioctl()
344 drm_exec_fini(exec); in xe_exec_ioctl()
/linux-6.15/drivers/gpu/drm/imagination/
H A Dpvr_job.c531 prepare_fw_obj_resv(struct drm_exec *exec, struct pvr_fw_object *fw_obj) in prepare_fw_obj_resv() argument
533 return drm_exec_prepare_obj(exec, gem_from_pvr_gem(fw_obj->gem), 1); in prepare_fw_obj_resv()
537 jobs_lock_all_objs(struct drm_exec *exec, struct pvr_job_data *job_data, in jobs_lock_all_objs() argument
546 int err = drm_exec_lock_obj(exec, in jobs_lock_all_objs()
553 err = prepare_fw_obj_resv(exec, in jobs_lock_all_objs()
567 drm_exec_until_all_locked(exec) { in prepare_job_resvs_for_each()
568 int err = jobs_lock_all_objs(exec, job_data, job_count); in prepare_job_resvs_for_each()
570 drm_exec_retry_on_contention(exec); in prepare_job_resvs_for_each()
711 struct drm_exec exec; in pvr_submit_jobs() local
752 err = prepare_job_resvs_for_each(&exec, job_data, args->jobs.count); in pvr_submit_jobs()
[all …]
/linux-6.15/drivers/gpu/drm/panthor/
H A Dpanthor_mmu.h50 int panthor_vm_prepare_mapped_bos_resvs(struct drm_exec *exec,
91 int panthor_vm_bind_job_prepare_resvs(struct drm_exec *exec,
93 void panthor_vm_bind_job_update_resvs(struct drm_exec *exec, struct drm_sched_job *job);
95 void panthor_vm_update_resvs(struct panthor_vm *vm, struct drm_exec *exec,

1234567891011