| /xnu-11215/osfmk/kern/ |
| H A D | kpc_thread.c | 107 if (!current_thread()->kpc_buf) { in kpc_set_thread_counting() 108 current_thread()->kperf_ast |= T_KPC_ALLOC; in kpc_set_thread_counting() 109 act_set_kperf(current_thread()); in kpc_set_thread_counting() 141 if (!current_thread()->kpc_buf) { in kpc_update_thread_counters() 142 current_thread()->kperf_ast |= T_KPC_ALLOC; in kpc_update_thread_counters() 143 act_set_kperf(current_thread()); in kpc_update_thread_counters() 156 thread_t thread = current_thread(); in kpc_get_curthread_counters() 172 kpc_update_thread_counters( current_thread()); in kpc_get_curthread_counters()
|
| H A D | kern_apfs_reflock.c | 161 ctid_t my_ctid = thread_get_ctid(current_thread()); in kern_apfs_reflock_try_get_ref() 193 … panic("kern_apfs_reflock: too many waiters for %p thread %p", reflock, current_thread()); in kern_apfs_reflock_try_get_ref() 205 … panic("kern_apfs_reflock: too many refs for %p thread %p", reflock, current_thread()); in kern_apfs_reflock_try_get_ref() 229 ctid_t my_ctid = thread_get_ctid(current_thread()); in kern_apfs_reflock_try_put_ref() 246 … panic("kern_apfs_reflock: over releasing reflock %p thread %p", reflock, current_thread()); in kern_apfs_reflock_try_put_ref() 268 … panic("kern_apfs_reflock: too many waiters for %p thread %p", reflock, current_thread()); in kern_apfs_reflock_try_put_ref() 308 ctid_t my_ctid = thread_get_ctid(current_thread()); in kern_apfs_reflock_try_lock() 323 …nic("Trying to lock a reflock owned by the same thread %p, reflock %p", current_thread(), reflock); in kern_apfs_reflock_try_lock() 334 … panic("kern_apfs_reflock: too many waiters for %p thread %p", reflock, current_thread()); in kern_apfs_reflock_try_lock() 366 ctid_t my_ctid = thread_get_ctid(current_thread()); in kern_apfs_reflock_wait_for_unlock() [all …]
|
| H A D | syscall_subr.c | 110 result = SCHED(thread_should_yield)(myprocessor, current_thread()); in swtch_continue() 127 if (!SCHED(thread_should_yield)(myprocessor, current_thread())) { in swtch() 143 thread_depress_abort(current_thread()); in swtch_pri_continue() 164 if (!SCHED(thread_should_yield)(myprocessor, current_thread())) { in swtch_pri() 179 thread_t self = current_thread(); in thread_switch_continue() 202 thread_t self = current_thread(); in thread_switch() 352 thread_t self = current_thread(); in thread_handoff_internal() 433 thread_t self = current_thread(); in thread_depress_abstime() 555 assert(self == current_thread()); in thread_poll_yield() 599 thread_t self = current_thread(); in thread_yield_internal() [all …]
|
| H A D | thread_test_context.c | 93 if (current_thread()->th_test_ctx != NULL) { in thread_test_context_tests() 105 if (current_thread()->th_test_ctx != NULL) { in thread_test_context_tests() 112 if (current_thread()->th_test_ctx != &ctx) { in thread_test_context_tests() 124 if (current_thread()->th_test_ctx != NULL) { in thread_test_context_tests() 199 if (current_thread()->th_test_ctx != NULL) { in thread_test_context_tests()
|
| H A D | thread_test_context.h | 103 return current_thread()->th_test_ctx; in thread_get_test_context() 152 thread_t thread = current_thread(); in thread_set_test_context() 195 thread_test_context_t *thread_ctx = current_thread()->th_test_ctx; in thread_cleanup_test_context() 199 current_thread()->th_test_ctx = NULL; in thread_cleanup_test_context()
|
| H A D | thread_act.c | 639 if (thread != current_thread()) { in thread_get_state_internal() 834 from_thread = current_thread(); in thread_convert_thread_state() 836 to_thread = current_thread(); in thread_convert_thread_state() 1123 if (thread == current_thread()) { in thread_set_apc_ast_locked() 1307 if (thread == current_thread()) { in act_set_state() 1321 if (thread == current_thread()) { in act_set_state_from_user() 1343 if (thread == current_thread()) { in act_get_state() 1357 if (thread == current_thread()) { in act_get_state_to_user() 1371 if (thread == current_thread()) { in act_set_ast() 1405 if (thread == current_thread()) { in act_set_ast_async() [all …]
|
| H A D | lock_ticket.c | 141 thread_t self = current_thread(); in __lck_ticket_owned_panic() 150 thread_t self = current_thread(); in __lck_ticket_not_owned_panic() 209 lock_disable_preemption_for_thread(current_thread()); in hw_lck_ticket_init_locked() 638 lock_disable_preemption_for_thread(current_thread()); in hw_lck_ticket_lock() 678 lock_disable_preemption_for_thread(current_thread()); in hw_lck_ticket_lock_to() 702 thread_t cthread = current_thread(); in lck_ticket_lock() 711 thread_t cthread = current_thread(); in lck_ticket_lock_nopreempt() 726 lock_disable_preemption_for_thread(current_thread()); in hw_lck_ticket_lock_try_internal() 761 thread_t cthread = current_thread(); in lck_ticket_lock_try_internal() 850 lock_disable_preemption_for_thread(current_thread()); in hw_lck_ticket_reserve() [all …]
|
| H A D | locks.c | 695 thread_t thread = current_thread(); 709 thread_t thread = current_thread(); 726 thread_t thread = current_thread(); 742 thread_t thread = current_thread(); 768 thread_t thread = current_thread(); 781 thread_t thread = current_thread(); 792 current_thread(), lock); in __hw_lock_unlock_unowned_panic() 837 thread = current_thread(); in hw_lock_assert() 2065 if (holder == current_thread()) { in gate_try_close() 2259 if (holder != current_thread()) { in gate_handoff() [all …]
|
| H A D | thread.c | 426 thread_t thread = current_thread(); in thread_corpse_continue() 2388 thread_t thread = current_thread(); in thread_priority_floor_start() 2417 thread_t thread = current_thread(); in thread_priority_floor_end() 2444 assert(thread == current_thread()); in thread_guard_violation() 2820 thread_t thread = current_thread(); in thread_cpulimit_remaining() 2844 thread_t thread = current_thread(); in thread_cpulimit_interval_has_expired() 2861 thread_t thread = current_thread(); in thread_cpulimit_restart() 3120 if (thread == current_thread()) { in thread_shared_rsrc_policy_set() 3150 if (thread == current_thread()) { in thread_shared_rsrc_policy_clear() 4149 thread_t self = current_thread(); in thread_associate_txm_thread_stack() [all …]
|
| H A D | lock_rw.c | 1134 thread_t thread = current_thread(); in lck_rw_lock_exclusive_check_contended() 1170 thread_t thread = current_thread(); 1407 thread_t thread = current_thread(); 1555 thread_t thread = current_thread(); in lck_rw_lock_shared_to_exclusive_failure() 1711 thread_t thread = current_thread(); in lck_rw_lock_shared_to_exclusive() 1812 thread_t thread = current_thread(); in lck_rw_lock_exclusive_to_shared_gen() 1917 thread_t thread = current_thread(); in lck_rw_try_lock_shared_internal_inline() 2032 thread_t thread = current_thread(); in lck_rw_try_lock_exclusive_internal_inline() 2181 thread = current_thread(); in lck_rw_done_gen() 2219 thread_t thread = current_thread(); in lck_rw_done() [all …]
|
| H A D | exclaves.c | 792 thread_t thread = current_thread(); in exclaves_allocate_ipc_buffer() 820 thread_t thread = current_thread(); in exclaves_free_ipc_buffer() 850 assert(thread == current_thread()); in exclaves_thread_terminate() 874 thread_t thread = current_thread(); in exclaves_get_ipc_buffer() 928 thread_t thread = current_thread(); in exclaves_boot_early() 1556 thread_t thread = current_thread(); in handle_response_spawned() 1606 thread_t thread = current_thread(); in handle_response_wait() 2051 thread_t thread = current_thread(); in exclaves_scheduler_request_update_timer() 2109 thread_t thread = current_thread(); in exclaves_scheduler_request_boot() 2201 thread_t thread = current_thread(); in exclaves_scheduler_request_resume() [all …]
|
| H A D | exclaves_xnuproxy.c | 76 thread_t thread = current_thread(); in exclaves_bootstrap_context_acquire() 96 thread_t thread = current_thread(); in exclaves_bootstrap_context_release() 308 thread_exclaves_state_flags_t state = current_thread()->th_exclaves_state; in exclaves_xnuproxy_ctx_alloc() 309 current_thread()->th_exclaves_state |= TH_EXCLAVES_SPAWN_EXPECTED; in exclaves_xnuproxy_ctx_alloc() 318 current_thread()->th_exclaves_state = state; in exclaves_xnuproxy_ctx_alloc() 474 thread_t thread = current_thread(); in exclaves_xnuproxy_endpoint_call()
|
| /xnu-11215/osfmk/tests/ |
| H A D | kernel_tests.c | 944 thread_t thread = current_thread(); in thread_lock_unlock_kernel_primitive() 1487 uint my_pri = current_thread()->sched_pri; in thread_inheritor_like_mutex() 1538 uint my_pri = current_thread()->sched_pri; in thread_just_inheritor_do_work() 1578 uint my_pri = current_thread()->sched_pri; in thread_steal_work() 1639 uint my_pri = current_thread()->sched_pri; in thread_no_inheritor_work() 1666 uint my_pri = current_thread()->sched_pri; in thread_mtx_work() 1766 uint my_pri = current_thread()->sched_pri; in thread_rw_work() 2582 uint my_pri = current_thread()->sched_pri; in thread_gate_aggressive() 2923 thread_t self = current_thread(); in thread_sleep_gate_chain_work() 3019 thread_t self = current_thread(); in thread_gate_chain_work() [all …]
|
| /xnu-11215/osfmk/i386/ |
| H A D | locks_i386_opt.c | 151 ordered_store_mtx_owner(lock, current_thread()->ctid); in lck_mtx_lock() 200 ordered_store_mtx_owner(lock, current_thread()->ctid); in lck_mtx_try_lock() 261 ordered_store_mtx_owner(lock, current_thread()->ctid); in lck_mtx_lock_spin_always() 341 ordered_store_mtx_owner(lock, current_thread()->ctid); in lck_mtx_try_lock_spin_always() 429 if (__improbable(lock->lck_mtx_owner != current_thread()->ctid)) { in lck_mtx_unlock()
|
| H A D | locks_i386.c | 422 thread = current_thread(); in lck_spin_assert() 486 lowner, current_thread(), lsti->owner_cpu, in usimple_lock_acquire_timeout_panic() 859 l->debug.lock_thread = (void *)current_thread(); in usld_lock_post() 964 l->debug.lock_thread = (void *) current_thread(); in usld_lock_try_post() 1115 thread = current_thread(); in lck_mtx_unlock_slow() 1201 assert(current_thread()->turnstile != NULL); in lck_mtx_unlock_wakeup_tail() 1245 assert(current_thread()->turnstile != NULL); in lck_mtx_lock_acquire_inline() 1428 thread = current_thread(); in lck_mtx_lock_contended() 1462 assert(current_thread()->turnstile != NULL); in lck_mtx_lock_contended() 1861 thread = current_thread(); in lck_mtx_assert() [all …]
|
| H A D | bsd_i386_native.c | 133 machine_thread_set_tsd_base(current_thread(), self); in thread_fast_set_cthread_self() 149 machine_thread_set_tsd_base(current_thread(), self); in thread_fast_set_cthread_self64() 199 pcb = THREAD_TO_PCB(current_thread()); in thread_set_user_ldt()
|
| /xnu-11215/osfmk/arm/ |
| H A D | preemption_disable.c | 84 if (current_thread()->machine.CpuDatap->cpu_pending_ast & AST_URGENT) { in kernel_preempt_check() 123 thread_t thread = current_thread(); in _disable_preemption() 154 thread_t thread = current_thread(); in _disable_preemption_without_measurements() 192 thread_t thread = current_thread(); in _enable_preemption() 229 return get_preemption_level_for_thread(current_thread()); in get_preemption_level() 348 thread_t thread = current_thread(); in _prepare_preemption_disable_measurement() 449 _enable_preemption_write_count(current_thread(), 0); in _collect_preemption_disable_measurement()
|
| H A D | machdep_call.c | 65 return machine_thread_set_tsd_base(current_thread(), self); in thread_set_cthread_self() 74 assert( self == current_thread()->machine.cthread_self); in thread_get_cthread_self()
|
| /xnu-11215/osfmk/arm64/ |
| H A D | copyio.c | 291 vm_map_t map = current_thread()->map; in copyin() 325 vm_map_t map = current_thread()->map; in copyin_atomic32() 343 vm_map_t map = current_thread()->map; in copyin_atomic32_wait_if_equals() 361 vm_map_t map = current_thread()->map; in copyin_atomic64() 379 vm_map_t map = current_thread()->map; in copyout_atomic32() 397 vm_map_t map = current_thread()->map; in copyout_atomic64() 415 vm_map_t map = current_thread()->map; in copyinstr() 444 vm_map_t map = current_thread()->map; in copyout() 473 vm_map_t map = current_thread()->map; in copyoutstr_prevalidate()
|
| H A D | bsd_arm64.c | 150 if (thread == current_thread()) { in thread_setsinglestep() 196 struct uthread *ut = get_bsdthread_info(current_thread()); in mach_syscall() 209 proc_pid(current_proc()), thread_tid(current_thread())); in mach_syscall() 286 proc_pid(current_proc()), thread_tid(current_thread())); in mach_syscall() 298 prior __assert_only = thread_get_kernel_state(current_thread())->allocation_name; in mach_syscall()
|
| /xnu-11215/osfmk/vm/ |
| H A D | vm_dyld_pager.c | 305 …ktriage_record(thread_tid(current_thread()), KDBG_TRIAGE_EVENTID(KDBG_TRIAGE_SUBSYS_DYLD_PAGER, KD… in fixupPage64() 321 …ktriage_record(thread_tid(current_thread()), KDBG_TRIAGE_EVENTID(KDBG_TRIAGE_SUBSYS_DYLD_PAGER, KD… in fixupPage64() 339 …ktriage_record(thread_tid(current_thread()), KDBG_TRIAGE_EVENTID(KDBG_TRIAGE_SUBSYS_DYLD_PAGER, KD… in fixupPage64() 385 …ktriage_record(thread_tid(current_thread()), KDBG_TRIAGE_EVENTID(KDBG_TRIAGE_SUBSYS_DYLD_PAGER, KD… in fixupChain32() 579 …ktriage_record(thread_tid(current_thread()), KDBG_TRIAGE_EVENTID(KDBG_TRIAGE_SUBSYS_DYLD_PAGER, KD… in fixupPageAuth64() 602 …ktriage_record(thread_tid(current_thread()), KDBG_TRIAGE_EVENTID(KDBG_TRIAGE_SUBSYS_DYLD_PAGER, KD… in fixupPageAuth64() 803 …ktriage_record(thread_tid(current_thread()), KDBG_TRIAGE_EVENTID(KDBG_TRIAGE_SUBSYS_DYLD_PAGER, KD… in fixup_page() 837 …ktriage_record(thread_tid(current_thread()), KDBG_TRIAGE_EVENTID(KDBG_TRIAGE_SUBSYS_DYLD_PAGER, KD… in fixup_page() 913 …ktriage_record(thread_tid(current_thread()), KDBG_TRIAGE_EVENTID(KDBG_TRIAGE_SUBSYS_DYLD_PAGER, KD… in dyld_pager_data_request() 979 …ktriage_record(thread_tid(current_thread()), KDBG_TRIAGE_EVENTID(KDBG_TRIAGE_SUBSYS_DYLD_PAGER, KD… in dyld_pager_data_request() [all …]
|
| /xnu-11215/bsd/kern/ |
| H A D | tty_tty.c | 99 context.vc_thread = current_thread(); in cttyopen() 170 context.vc_thread = current_thread(); in cttyread() 190 context.vc_thread = current_thread(); in cttywrite() 225 context.vc_thread = current_thread(); in cttyioctl() 241 context.vc_thread = current_thread(); in cttyselect()
|
| H A D | sys_work_interval.c | 81 kret = kern_work_interval_create(current_thread(), &create_args); in work_interval_ctl() 105 kern_work_interval_destroy(current_thread(), create_args.wica_id); in work_interval_ctl() 210 kret = kern_work_interval_destroy(current_thread(), uap->work_interval_id); in work_interval_ctl() 245 kret = kern_work_interval_notify(current_thread(), &kwi_args); in work_interval_ctl() 260 kret = kern_work_interval_join(current_thread(), in work_interval_ctl()
|
| /xnu-11215/bsd/dev/dtrace/ |
| H A D | profile_prvd.c | 156 x86_saved_state_t *kern_regs = find_kern_regs(current_thread()); in profile_fire() 162 pal_register_cache_state(current_thread(), VALID); in profile_fire() 164 x86_saved_state_t *tagged_regs = (x86_saved_state_t *)find_user_regs(current_thread()); in profile_fire() 182 arm_saved_state_t *arm_kern_regs = (arm_saved_state_t *) find_kern_regs(current_thread()); in profile_fire() 194 arm_saved_state_t *arm_user_regs = (arm_saved_state_t *)find_user_regs(current_thread()); in profile_fire() 215 x86_saved_state_t *kern_regs = find_kern_regs(current_thread()); in profile_tick() 221 pal_register_cache_state(current_thread(), VALID); in profile_tick() 223 x86_saved_state_t *tagged_regs = (x86_saved_state_t *)find_user_regs(current_thread()); in profile_tick() 241 arm_saved_state_t *arm_kern_regs = (arm_saved_state_t *) find_kern_regs(current_thread()); in profile_tick() 250 arm_saved_state_t *arm_user_regs = (arm_saved_state_t *)find_user_regs(current_thread()); in profile_tick()
|
| /xnu-11215/bsd/vm/ |
| H A D | vm_unix.c | 936 (void *)VM_KERNEL_ADDRPERM(current_thread()), in shared_region_check_np() 1014 (void *)VM_KERNEL_ADDRPERM(current_thread()), in shared_region_check_np() 1038 (void *)VM_KERNEL_ADDRPERM(current_thread()), in shared_region_copyin() 1089 (void *)VM_KERNEL_ADDRPERM(current_thread()), in shared_region_map_and_slide_setup() 1127 (void *)VM_KERNEL_ADDRPERM(current_thread()), in shared_region_map_and_slide_setup() 1204 (void *)VM_KERNEL_ADDRPERM(current_thread()), in shared_region_map_and_slide_setup() 1620 (void *)VM_KERNEL_ADDRPERM(current_thread()), in _shared_region_map_and_slide() 1863 (void *)VM_KERNEL_ADDRPERM(current_thread()), in shared_region_map_and_slide_2_np() 1877 (void *)VM_KERNEL_ADDRPERM(current_thread()), in shared_region_map_and_slide_2_np() 1888 (void *)VM_KERNEL_ADDRPERM(current_thread()), in shared_region_map_and_slide_2_np() [all …]
|