| /xnu-11215/osfmk/kern/ |
| H A D | kpc_common.c | 83 kpc_register_cpu(struct cpu_data *cpu_data) in kpc_register_cpu() argument 85 assert(cpu_data); in kpc_register_cpu() 123 kpc_unregister_cpu(struct cpu_data *cpu_data) in kpc_unregister_cpu() argument 125 assert(cpu_data); in kpc_unregister_cpu() 126 if (cpu_data->cpu_kpc_buf[0] != NULL) { in kpc_unregister_cpu() 128 cpu_data->cpu_kpc_buf[0] = NULL; in kpc_unregister_cpu() 130 if (cpu_data->cpu_kpc_buf[1] != NULL) { in kpc_unregister_cpu() 132 cpu_data->cpu_kpc_buf[1] = NULL; in kpc_unregister_cpu() 134 if (cpu_data->cpu_kpc_shadow != NULL) { in kpc_unregister_cpu() 136 cpu_data->cpu_kpc_shadow = NULL; in kpc_unregister_cpu() [all …]
|
| H A D | kpc.h | 83 struct cpu_data; 84 extern void kpc_register_cpu(struct cpu_data *cpu_data); 85 extern void kpc_unregister_cpu(struct cpu_data *cpu_data);
|
| H A D | exclaves.c | 1205 percpu_foreach(cpu_data, cpu_data) { in exclaves_init_multicore() 1206 if (cpu_data->cpu_phys_id != dt_phys_id) { in exclaves_init_multicore() 1209 req = PERCPU_GET_RELATIVE(exclaves_request, cpu_data, cpu_data); in exclaves_init_multicore() 1212 res = PERCPU_GET_RELATIVE(exclaves_response, cpu_data, cpu_data); in exclaves_init_multicore()
|
| H A D | sched_clutch.h | 204 } cpu_data; member
|
| H A D | Makefile | 66 cpu_data.h \
|
| H A D | sched_clutch.c | 1390 …os_atomic_store(&clutch_bucket_group->scbg_cpu_data.cpu_data.scbcd_cpu_blocked, (clutch_cpu_data_t… in sched_clutch_bucket_group_init() 1669 clutch_cpu_data_t cpu_used = scb_cpu_data.cpu_data.scbcd_cpu_used; in sched_clutch_interactivity_from_cpu_data() 1670 clutch_cpu_data_t cpu_blocked = scb_cpu_data.cpu_data.scbcd_cpu_blocked; in sched_clutch_interactivity_from_cpu_data() 1887 …os_atomic_add(&(clutch_bucket_group->scbg_cpu_data.cpu_data.scbcd_cpu_used), (clutch_cpu_data_t)de… in sched_clutch_bucket_group_cpu_usage_update() 1931 clutch_cpu_data_t cpu_used = old_cpu_data.cpu_data.scbcd_cpu_used; in sched_clutch_bucket_group_cpu_adjust() 1932 clutch_cpu_data_t cpu_blocked = old_cpu_data.cpu_data.scbcd_cpu_blocked; in sched_clutch_bucket_group_cpu_adjust() 1945 new_cpu_data.cpu_data.scbcd_cpu_used = cpu_used; in sched_clutch_bucket_group_cpu_adjust() 1946 new_cpu_data.cpu_data.scbcd_cpu_blocked = cpu_blocked; in sched_clutch_bucket_group_cpu_adjust() 2598 …os_atomic_add(&(clutch_bucket_group->scbg_cpu_data.cpu_data.scbcd_cpu_blocked), (clutch_cpu_data_t… in sched_clutch_bucket_group_run_count_inc()
|
| /xnu-11215/osfmk/i386/ |
| H A D | pmCPU.h | 133 uint64_t pmCPUGetDeadline(struct cpu_data *cpu); 134 uint64_t pmCPUSetDeadline(struct cpu_data *cpu, uint64_t deadline); 135 void pmCPUDeadline(struct cpu_data *cpu); 136 boolean_t pmCPUExitIdle(struct cpu_data *cpu); 137 void pmCPUMarkRunning(struct cpu_data *cpu);
|
| H A D | Makefile | 16 cpu_data.h \ 61 INSTALL_KF_MD_LCL_LIST = $(filter-out bit_routines.h cpu_data.h pal_i386.h, $(EXPORT_ONLY_FILES))
|
| H A D | cpu_topology.h | 70 struct cpu_data; 132 struct cpu_data *cpu; /* cpu_data structure */
|
| H A D | misc_protos.h | 38 struct cpu_data;
|
| H A D | cpu_data.h | 169 typedef struct cpu_data { struct 172 struct cpu_data *cpu_this; /* pointer to myself */ argument
|
| /xnu-11215/osfmk/arm/ |
| H A D | cpu_data_internal.h | 74 struct cpu_data *cpu_data_vaddr; /* Cpu data virtual address */ 133 typedef struct cpu_data { struct 295 PERCPU_DECL(cpu_data_t, cpu_data); argument 296 #define BootCpuData __PERCPU_NAME(cpu_data)
|
| H A D | cpu_common.c | 63 cpu_data_t PERCPU_DATA(cpu_data); 702 cpu_data_t *cpu_data = cpu_datap(cpu); in cpu_to_processor() local 703 if (cpu_data != NULL) { in cpu_to_processor() 704 return PERCPU_GET_RELATIVE(processor, cpu_data, cpu_data); in cpu_to_processor() 716 return PERCPU_GET_RELATIVE(cpu_data, processor, processor); in processor_to_cpu_datap() 756 cpu_data_ptr = PERCPU_GET_MASTER(cpu_data); in cpu_data_alloc() 761 cpu_data_ptr = PERCPU_GET_WITH_BASE(base, cpu_data); in cpu_data_alloc() 827 return (char *)cpu_datap(cpu) - __PERCPU_ADDR(cpu_data); in other_percpu_base()
|
| H A D | arm_timer.c | 108 processor = PERCPU_GET_RELATIVE(processor, cpu_data, cpu_data_ptr); in timer_intr() 171 PERCPU_GET_RELATIVE(processor, cpu_data, cpu_data_ptr)); in timer_resync_deadlines()
|
| H A D | thread.h | 160 struct cpu_data * CpuDatap; /* current per cpu data */
|
| H A D | arm_init.c | 783 processor_t processor = PERCPU_GET_RELATIVE(processor, cpu_data, cpu_data_ptr); in arm_init_cpu()
|
| /xnu-11215/tools/lldbmacros/kmemory/ |
| H A D | kmem.py | 264 cpu_data = self.arm64_CpuDataEntries.chkGetChildAtIndex(cpu) 267 return cpu_data.xGetIntegerByName('cpu_data_vaddr') - boot_vaddr 300 cpu_data = self.intel_cpu_data.chkGetChildAtIndex(cpu) 301 return cpu_data.xGetIntegerByName('cpu_pcpu_base')
|
| /xnu-11215/osfmk/arm64/ |
| H A D | cpu.c | 188 cpu_data_t *cpu_data; in arm64_ipi_test_callback() local 190 cpu_data = getCpuDatap(); in arm64_ipi_test_callback() 192 *ipi_test_data = cpu_data->cpu_number; in arm64_ipi_test_callback() 199 cpu_data_t *cpu_data; in arm64_immediate_ipi_test_callback() local 201 cpu_data = getCpuDatap(); in arm64_immediate_ipi_test_callback() 203 *ipi_test_data = cpu_data->cpu_number + MAX_CPUS; in arm64_immediate_ipi_test_callback() 879 processor_t processor = PERCPU_GET_RELATIVE(processor, cpu_data, cpu_data_ptr); in cpu_start() 908 ml_make_pcpu_base_and_cpu_number((char *)cpu_data_ptr - __PERCPU_ADDR(cpu_data), in cpu_start()
|
| /xnu-11215/tools/lldbmacros/ |
| H A D | misc.py | 162 cpu_data = kern.globals.cpu_data_ptr[cpu_id] 163 return cpu_data 240 cpu_data = GetCpuDataForCpuID(cpu) 241 rt_timer = cpu_data.rtclock_timer 298 cpu_data = GetCpuDataForCpuID(cpu_id) 300 print (format_string.format(cpu_id, processor, cpu_data))
|
| H A D | scheduler.py | 373 cpu_data = kern.globals.cpu_data_ptr[current_processor.cpu_id] 374 if (cpu_data != 0) : 375 cpu_debugger_time = max(cpu_data.debugger_entry_time, cpu_data.debugger_ipi_time)
|
| H A D | process.py | 1889 cpu_data = kern.globals.cpu_data_ptr[processor.cpu_id] 1890 if (cpu_data != 0) : 1891 ast = cpu_data.cpu_pending_ast 1892 preemption_disable = cpu_data.cpu_preemption_level
|
| /xnu-11215/osfmk/arm/pmap/ |
| H A D | pmap_data.c | 3512 pmap_cpu_data_t *cpu_data = pmap_get_cpu_data(); in validate_pmap_mutable_internal() 3521 os_atomic_load(&cpu_data->inflight_pmap, relaxed); in validate_pmap_mutable_internal() 3531 os_atomic_store(&cpu_data->inflight_pmap, pmap, release); in validate_pmap_mutable_internal() 4151 pmap_cpu_data_array[cpu_num].cpu_data.cpu_number = PMAP_INVALID_CPU_NUM; in pmap_cpu_data_array_init() 4152 pmap_cpu_data_array[cpu_num].cpu_data.ppl_state = PPL_STATE_KERNEL; in pmap_cpu_data_array_init() 4153 pmap_cpu_data_array[cpu_num].cpu_data.ppl_stack = (void*)(stack_va + PPL_STACK_SIZE); in pmap_cpu_data_array_init() 4250 pmap_cpu_data_array[cpu_num].cpu_data.scratch_page = (void*)phystokv(avail_start); in pmap_cpu_data_array_init() 4294 return &pmap_cpu_data_array[cpu].cpu_data; in pmap_get_remote_cpu_data() 4296 cpu_data_t *cpu_data = cpu_datap((int)cpu); in pmap_get_remote_cpu_data() 4297 if (cpu_data == NULL) { in pmap_get_remote_cpu_data() [all …]
|
| /xnu-11215/osfmk/kdp/ml/arm/ |
| H A D | kdp_vm.c | 310 struct cpu_data *cpudatap = cpuentryp->cpu_data_vaddr; in kern_collectth_state() 332 processor_t processor = PERCPU_GET_RELATIVE(processor, cpu_data, cpudatap); in kern_collectth_state()
|
| /xnu-11215/osfmk/arm64/sptm/pmap/ |
| H A D | pmap_data.c | 2147 cpu_data_t *cpu_data = cpu_datap((int)cpu); in pmap_get_remote_cpu_data() local 2148 if (cpu_data == NULL) { in pmap_get_remote_cpu_data() 2151 return &cpu_data->cpu_pmap_cpu_data; in pmap_get_remote_cpu_data()
|
| H A D | pmap.h | 667 pmap_cpu_data_t cpu_data; member
|