Home
last modified time | relevance | path

Searched refs:current_thread (Results 1 – 25 of 274) sorted by relevance

1234567891011

/xnu-11215/osfmk/kern/
H A Dkpc_thread.c107 if (!current_thread()->kpc_buf) { in kpc_set_thread_counting()
108 current_thread()->kperf_ast |= T_KPC_ALLOC; in kpc_set_thread_counting()
109 act_set_kperf(current_thread()); in kpc_set_thread_counting()
141 if (!current_thread()->kpc_buf) { in kpc_update_thread_counters()
142 current_thread()->kperf_ast |= T_KPC_ALLOC; in kpc_update_thread_counters()
143 act_set_kperf(current_thread()); in kpc_update_thread_counters()
156 thread_t thread = current_thread(); in kpc_get_curthread_counters()
172 kpc_update_thread_counters( current_thread()); in kpc_get_curthread_counters()
H A Dkern_apfs_reflock.c161 ctid_t my_ctid = thread_get_ctid(current_thread()); in kern_apfs_reflock_try_get_ref()
193 … panic("kern_apfs_reflock: too many waiters for %p thread %p", reflock, current_thread()); in kern_apfs_reflock_try_get_ref()
205 … panic("kern_apfs_reflock: too many refs for %p thread %p", reflock, current_thread()); in kern_apfs_reflock_try_get_ref()
229 ctid_t my_ctid = thread_get_ctid(current_thread()); in kern_apfs_reflock_try_put_ref()
246 … panic("kern_apfs_reflock: over releasing reflock %p thread %p", reflock, current_thread()); in kern_apfs_reflock_try_put_ref()
268 … panic("kern_apfs_reflock: too many waiters for %p thread %p", reflock, current_thread()); in kern_apfs_reflock_try_put_ref()
308 ctid_t my_ctid = thread_get_ctid(current_thread()); in kern_apfs_reflock_try_lock()
323 …nic("Trying to lock a reflock owned by the same thread %p, reflock %p", current_thread(), reflock); in kern_apfs_reflock_try_lock()
334 … panic("kern_apfs_reflock: too many waiters for %p thread %p", reflock, current_thread()); in kern_apfs_reflock_try_lock()
366 ctid_t my_ctid = thread_get_ctid(current_thread()); in kern_apfs_reflock_wait_for_unlock()
[all …]
H A Dsyscall_subr.c110 result = SCHED(thread_should_yield)(myprocessor, current_thread()); in swtch_continue()
127 if (!SCHED(thread_should_yield)(myprocessor, current_thread())) { in swtch()
143 thread_depress_abort(current_thread()); in swtch_pri_continue()
164 if (!SCHED(thread_should_yield)(myprocessor, current_thread())) { in swtch_pri()
179 thread_t self = current_thread(); in thread_switch_continue()
202 thread_t self = current_thread(); in thread_switch()
352 thread_t self = current_thread(); in thread_handoff_internal()
433 thread_t self = current_thread(); in thread_depress_abstime()
555 assert(self == current_thread()); in thread_poll_yield()
599 thread_t self = current_thread(); in thread_yield_internal()
[all …]
H A Dthread_test_context.c93 if (current_thread()->th_test_ctx != NULL) { in thread_test_context_tests()
105 if (current_thread()->th_test_ctx != NULL) { in thread_test_context_tests()
112 if (current_thread()->th_test_ctx != &ctx) { in thread_test_context_tests()
124 if (current_thread()->th_test_ctx != NULL) { in thread_test_context_tests()
199 if (current_thread()->th_test_ctx != NULL) { in thread_test_context_tests()
H A Dthread_test_context.h103 return current_thread()->th_test_ctx; in thread_get_test_context()
152 thread_t thread = current_thread(); in thread_set_test_context()
195 thread_test_context_t *thread_ctx = current_thread()->th_test_ctx; in thread_cleanup_test_context()
199 current_thread()->th_test_ctx = NULL; in thread_cleanup_test_context()
H A Dthread_act.c639 if (thread != current_thread()) { in thread_get_state_internal()
834 from_thread = current_thread(); in thread_convert_thread_state()
836 to_thread = current_thread(); in thread_convert_thread_state()
1123 if (thread == current_thread()) { in thread_set_apc_ast_locked()
1307 if (thread == current_thread()) { in act_set_state()
1321 if (thread == current_thread()) { in act_set_state_from_user()
1343 if (thread == current_thread()) { in act_get_state()
1357 if (thread == current_thread()) { in act_get_state_to_user()
1371 if (thread == current_thread()) { in act_set_ast()
1405 if (thread == current_thread()) { in act_set_ast_async()
[all …]
H A Dlock_ticket.c141 thread_t self = current_thread(); in __lck_ticket_owned_panic()
150 thread_t self = current_thread(); in __lck_ticket_not_owned_panic()
209 lock_disable_preemption_for_thread(current_thread()); in hw_lck_ticket_init_locked()
638 lock_disable_preemption_for_thread(current_thread()); in hw_lck_ticket_lock()
678 lock_disable_preemption_for_thread(current_thread()); in hw_lck_ticket_lock_to()
702 thread_t cthread = current_thread(); in lck_ticket_lock()
711 thread_t cthread = current_thread(); in lck_ticket_lock_nopreempt()
726 lock_disable_preemption_for_thread(current_thread()); in hw_lck_ticket_lock_try_internal()
761 thread_t cthread = current_thread(); in lck_ticket_lock_try_internal()
850 lock_disable_preemption_for_thread(current_thread()); in hw_lck_ticket_reserve()
[all …]
H A Dlocks.c695 thread_t thread = current_thread();
709 thread_t thread = current_thread();
726 thread_t thread = current_thread();
742 thread_t thread = current_thread();
768 thread_t thread = current_thread();
781 thread_t thread = current_thread();
792 current_thread(), lock); in __hw_lock_unlock_unowned_panic()
837 thread = current_thread(); in hw_lock_assert()
2065 if (holder == current_thread()) { in gate_try_close()
2259 if (holder != current_thread()) { in gate_handoff()
[all …]
H A Dthread.c426 thread_t thread = current_thread(); in thread_corpse_continue()
2388 thread_t thread = current_thread(); in thread_priority_floor_start()
2417 thread_t thread = current_thread(); in thread_priority_floor_end()
2444 assert(thread == current_thread()); in thread_guard_violation()
2820 thread_t thread = current_thread(); in thread_cpulimit_remaining()
2844 thread_t thread = current_thread(); in thread_cpulimit_interval_has_expired()
2861 thread_t thread = current_thread(); in thread_cpulimit_restart()
3120 if (thread == current_thread()) { in thread_shared_rsrc_policy_set()
3150 if (thread == current_thread()) { in thread_shared_rsrc_policy_clear()
4149 thread_t self = current_thread(); in thread_associate_txm_thread_stack()
[all …]
H A Dlock_rw.c1134 thread_t thread = current_thread(); in lck_rw_lock_exclusive_check_contended()
1170 thread_t thread = current_thread();
1407 thread_t thread = current_thread();
1555 thread_t thread = current_thread(); in lck_rw_lock_shared_to_exclusive_failure()
1711 thread_t thread = current_thread(); in lck_rw_lock_shared_to_exclusive()
1812 thread_t thread = current_thread(); in lck_rw_lock_exclusive_to_shared_gen()
1917 thread_t thread = current_thread(); in lck_rw_try_lock_shared_internal_inline()
2032 thread_t thread = current_thread(); in lck_rw_try_lock_exclusive_internal_inline()
2181 thread = current_thread(); in lck_rw_done_gen()
2219 thread_t thread = current_thread(); in lck_rw_done()
[all …]
H A Dexclaves.c792 thread_t thread = current_thread(); in exclaves_allocate_ipc_buffer()
820 thread_t thread = current_thread(); in exclaves_free_ipc_buffer()
850 assert(thread == current_thread()); in exclaves_thread_terminate()
874 thread_t thread = current_thread(); in exclaves_get_ipc_buffer()
928 thread_t thread = current_thread(); in exclaves_boot_early()
1556 thread_t thread = current_thread(); in handle_response_spawned()
1606 thread_t thread = current_thread(); in handle_response_wait()
2051 thread_t thread = current_thread(); in exclaves_scheduler_request_update_timer()
2109 thread_t thread = current_thread(); in exclaves_scheduler_request_boot()
2201 thread_t thread = current_thread(); in exclaves_scheduler_request_resume()
[all …]
H A Dexclaves_xnuproxy.c76 thread_t thread = current_thread(); in exclaves_bootstrap_context_acquire()
96 thread_t thread = current_thread(); in exclaves_bootstrap_context_release()
308 thread_exclaves_state_flags_t state = current_thread()->th_exclaves_state; in exclaves_xnuproxy_ctx_alloc()
309 current_thread()->th_exclaves_state |= TH_EXCLAVES_SPAWN_EXPECTED; in exclaves_xnuproxy_ctx_alloc()
318 current_thread()->th_exclaves_state = state; in exclaves_xnuproxy_ctx_alloc()
474 thread_t thread = current_thread(); in exclaves_xnuproxy_endpoint_call()
/xnu-11215/osfmk/tests/
H A Dkernel_tests.c944 thread_t thread = current_thread(); in thread_lock_unlock_kernel_primitive()
1487 uint my_pri = current_thread()->sched_pri; in thread_inheritor_like_mutex()
1538 uint my_pri = current_thread()->sched_pri; in thread_just_inheritor_do_work()
1578 uint my_pri = current_thread()->sched_pri; in thread_steal_work()
1639 uint my_pri = current_thread()->sched_pri; in thread_no_inheritor_work()
1666 uint my_pri = current_thread()->sched_pri; in thread_mtx_work()
1766 uint my_pri = current_thread()->sched_pri; in thread_rw_work()
2582 uint my_pri = current_thread()->sched_pri; in thread_gate_aggressive()
2923 thread_t self = current_thread(); in thread_sleep_gate_chain_work()
3019 thread_t self = current_thread(); in thread_gate_chain_work()
[all …]
/xnu-11215/osfmk/i386/
H A Dlocks_i386_opt.c151 ordered_store_mtx_owner(lock, current_thread()->ctid); in lck_mtx_lock()
200 ordered_store_mtx_owner(lock, current_thread()->ctid); in lck_mtx_try_lock()
261 ordered_store_mtx_owner(lock, current_thread()->ctid); in lck_mtx_lock_spin_always()
341 ordered_store_mtx_owner(lock, current_thread()->ctid); in lck_mtx_try_lock_spin_always()
429 if (__improbable(lock->lck_mtx_owner != current_thread()->ctid)) { in lck_mtx_unlock()
H A Dlocks_i386.c422 thread = current_thread(); in lck_spin_assert()
486 lowner, current_thread(), lsti->owner_cpu, in usimple_lock_acquire_timeout_panic()
859 l->debug.lock_thread = (void *)current_thread(); in usld_lock_post()
964 l->debug.lock_thread = (void *) current_thread(); in usld_lock_try_post()
1115 thread = current_thread(); in lck_mtx_unlock_slow()
1201 assert(current_thread()->turnstile != NULL); in lck_mtx_unlock_wakeup_tail()
1245 assert(current_thread()->turnstile != NULL); in lck_mtx_lock_acquire_inline()
1428 thread = current_thread(); in lck_mtx_lock_contended()
1462 assert(current_thread()->turnstile != NULL); in lck_mtx_lock_contended()
1861 thread = current_thread(); in lck_mtx_assert()
[all …]
H A Dbsd_i386_native.c133 machine_thread_set_tsd_base(current_thread(), self); in thread_fast_set_cthread_self()
149 machine_thread_set_tsd_base(current_thread(), self); in thread_fast_set_cthread_self64()
199 pcb = THREAD_TO_PCB(current_thread()); in thread_set_user_ldt()
/xnu-11215/osfmk/arm/
H A Dpreemption_disable.c84 if (current_thread()->machine.CpuDatap->cpu_pending_ast & AST_URGENT) { in kernel_preempt_check()
123 thread_t thread = current_thread(); in _disable_preemption()
154 thread_t thread = current_thread(); in _disable_preemption_without_measurements()
192 thread_t thread = current_thread(); in _enable_preemption()
229 return get_preemption_level_for_thread(current_thread()); in get_preemption_level()
348 thread_t thread = current_thread(); in _prepare_preemption_disable_measurement()
449 _enable_preemption_write_count(current_thread(), 0); in _collect_preemption_disable_measurement()
H A Dmachdep_call.c65 return machine_thread_set_tsd_base(current_thread(), self); in thread_set_cthread_self()
74 assert( self == current_thread()->machine.cthread_self); in thread_get_cthread_self()
/xnu-11215/osfmk/arm64/
H A Dcopyio.c291 vm_map_t map = current_thread()->map; in copyin()
325 vm_map_t map = current_thread()->map; in copyin_atomic32()
343 vm_map_t map = current_thread()->map; in copyin_atomic32_wait_if_equals()
361 vm_map_t map = current_thread()->map; in copyin_atomic64()
379 vm_map_t map = current_thread()->map; in copyout_atomic32()
397 vm_map_t map = current_thread()->map; in copyout_atomic64()
415 vm_map_t map = current_thread()->map; in copyinstr()
444 vm_map_t map = current_thread()->map; in copyout()
473 vm_map_t map = current_thread()->map; in copyoutstr_prevalidate()
H A Dbsd_arm64.c150 if (thread == current_thread()) { in thread_setsinglestep()
196 struct uthread *ut = get_bsdthread_info(current_thread()); in mach_syscall()
209 proc_pid(current_proc()), thread_tid(current_thread())); in mach_syscall()
286 proc_pid(current_proc()), thread_tid(current_thread())); in mach_syscall()
298 prior __assert_only = thread_get_kernel_state(current_thread())->allocation_name; in mach_syscall()
/xnu-11215/osfmk/vm/
H A Dvm_dyld_pager.c305 …ktriage_record(thread_tid(current_thread()), KDBG_TRIAGE_EVENTID(KDBG_TRIAGE_SUBSYS_DYLD_PAGER, KD… in fixupPage64()
321 …ktriage_record(thread_tid(current_thread()), KDBG_TRIAGE_EVENTID(KDBG_TRIAGE_SUBSYS_DYLD_PAGER, KD… in fixupPage64()
339 …ktriage_record(thread_tid(current_thread()), KDBG_TRIAGE_EVENTID(KDBG_TRIAGE_SUBSYS_DYLD_PAGER, KD… in fixupPage64()
385 …ktriage_record(thread_tid(current_thread()), KDBG_TRIAGE_EVENTID(KDBG_TRIAGE_SUBSYS_DYLD_PAGER, KD… in fixupChain32()
579 …ktriage_record(thread_tid(current_thread()), KDBG_TRIAGE_EVENTID(KDBG_TRIAGE_SUBSYS_DYLD_PAGER, KD… in fixupPageAuth64()
602 …ktriage_record(thread_tid(current_thread()), KDBG_TRIAGE_EVENTID(KDBG_TRIAGE_SUBSYS_DYLD_PAGER, KD… in fixupPageAuth64()
803 …ktriage_record(thread_tid(current_thread()), KDBG_TRIAGE_EVENTID(KDBG_TRIAGE_SUBSYS_DYLD_PAGER, KD… in fixup_page()
837 …ktriage_record(thread_tid(current_thread()), KDBG_TRIAGE_EVENTID(KDBG_TRIAGE_SUBSYS_DYLD_PAGER, KD… in fixup_page()
913 …ktriage_record(thread_tid(current_thread()), KDBG_TRIAGE_EVENTID(KDBG_TRIAGE_SUBSYS_DYLD_PAGER, KD… in dyld_pager_data_request()
979 …ktriage_record(thread_tid(current_thread()), KDBG_TRIAGE_EVENTID(KDBG_TRIAGE_SUBSYS_DYLD_PAGER, KD… in dyld_pager_data_request()
[all …]
/xnu-11215/bsd/kern/
H A Dtty_tty.c99 context.vc_thread = current_thread(); in cttyopen()
170 context.vc_thread = current_thread(); in cttyread()
190 context.vc_thread = current_thread(); in cttywrite()
225 context.vc_thread = current_thread(); in cttyioctl()
241 context.vc_thread = current_thread(); in cttyselect()
H A Dsys_work_interval.c81 kret = kern_work_interval_create(current_thread(), &create_args); in work_interval_ctl()
105 kern_work_interval_destroy(current_thread(), create_args.wica_id); in work_interval_ctl()
210 kret = kern_work_interval_destroy(current_thread(), uap->work_interval_id); in work_interval_ctl()
245 kret = kern_work_interval_notify(current_thread(), &kwi_args); in work_interval_ctl()
260 kret = kern_work_interval_join(current_thread(), in work_interval_ctl()
/xnu-11215/bsd/dev/dtrace/
H A Dprofile_prvd.c156 x86_saved_state_t *kern_regs = find_kern_regs(current_thread()); in profile_fire()
162 pal_register_cache_state(current_thread(), VALID); in profile_fire()
164 x86_saved_state_t *tagged_regs = (x86_saved_state_t *)find_user_regs(current_thread()); in profile_fire()
182 arm_saved_state_t *arm_kern_regs = (arm_saved_state_t *) find_kern_regs(current_thread()); in profile_fire()
194 arm_saved_state_t *arm_user_regs = (arm_saved_state_t *)find_user_regs(current_thread()); in profile_fire()
215 x86_saved_state_t *kern_regs = find_kern_regs(current_thread()); in profile_tick()
221 pal_register_cache_state(current_thread(), VALID); in profile_tick()
223 x86_saved_state_t *tagged_regs = (x86_saved_state_t *)find_user_regs(current_thread()); in profile_tick()
241 arm_saved_state_t *arm_kern_regs = (arm_saved_state_t *) find_kern_regs(current_thread()); in profile_tick()
250 arm_saved_state_t *arm_user_regs = (arm_saved_state_t *)find_user_regs(current_thread()); in profile_tick()
/xnu-11215/bsd/vm/
H A Dvm_unix.c936 (void *)VM_KERNEL_ADDRPERM(current_thread()), in shared_region_check_np()
1014 (void *)VM_KERNEL_ADDRPERM(current_thread()), in shared_region_check_np()
1038 (void *)VM_KERNEL_ADDRPERM(current_thread()), in shared_region_copyin()
1089 (void *)VM_KERNEL_ADDRPERM(current_thread()), in shared_region_map_and_slide_setup()
1127 (void *)VM_KERNEL_ADDRPERM(current_thread()), in shared_region_map_and_slide_setup()
1204 (void *)VM_KERNEL_ADDRPERM(current_thread()), in shared_region_map_and_slide_setup()
1620 (void *)VM_KERNEL_ADDRPERM(current_thread()), in _shared_region_map_and_slide()
1863 (void *)VM_KERNEL_ADDRPERM(current_thread()), in shared_region_map_and_slide_2_np()
1877 (void *)VM_KERNEL_ADDRPERM(current_thread()), in shared_region_map_and_slide_2_np()
1888 (void *)VM_KERNEL_ADDRPERM(current_thread()), in shared_region_map_and_slide_2_np()
[all …]

1234567891011