| /linux-6.15/drivers/gpu/drm/amd/amdkfd/ |
| H A D | kfd_flat_memory.c | 324 pdd->lds_limit = MAKE_LDS_APP_LIMIT(pdd->lds_base); in kfd_init_apertures_vi() 331 pdd->gpuvm_limit = in kfd_init_apertures_vi() 341 pdd->scratch_limit = MAKE_SCRATCH_APP_LIMIT(pdd->scratch_base); in kfd_init_apertures_vi() 347 pdd->lds_limit = MAKE_LDS_APP_LIMIT(pdd->lds_base); in kfd_init_apertures_v9() 350 pdd->gpuvm_limit = in kfd_init_apertures_v9() 354 pdd->scratch_limit = MAKE_SCRATCH_APP_LIMIT(pdd->scratch_base); in kfd_init_apertures_v9() 360 pdd->qpd.cwsr_base = AMDGPU_VA_RESERVED_TRAP_START(pdd->dev->adev); in kfd_init_apertures_v9() 382 if (!pdd) { in kfd_init_apertures() 393 pdd->lds_base = pdd->lds_limit = 0; in kfd_init_apertures() 394 pdd->gpuvm_base = pdd->gpuvm_limit = 0; in kfd_init_apertures() [all …]
|
| H A D | kfd_debug.c | 212 if (!pdd) in kfd_set_dbg_ev_from_interrupt() 273 if (!pdd) in kfd_dbg_send_exception_to_runtime() 445 pdd->watch_points[watch_id] = pdd->dev->kfd2kgd->clear_address_watch( in kfd_dbg_trap_clear_dev_address_watch() 483 pdd->dev->adev, in kfd_dbg_trap_set_dev_address_watch() 620 pdd->dev->adev, in kfd_dbg_trap_deactivate() 626 release_debug_trap_vmid(pdd->dev->dqm, &pdd->qpd)) in kfd_dbg_trap_deactivate() 705 r = reserve_debug_trap_vmid(pdd->dev->dqm, &pdd->qpd); in kfd_dbg_trap_activate() 726 pdd->dev->kfd2kgd->enable_debug_trap(pdd->dev->adev, true, in kfd_dbg_trap_activate() 729 pdd->spi_dbg_override = pdd->dev->kfd2kgd->enable_debug_trap( in kfd_dbg_trap_activate() 785 if (pdd->qpd.num_gws && (!kfd_dbg_has_gws_support(pdd->dev) || in kfd_dbg_trap_enable() [all …]
|
| H A D | kfd_process.c | 116 pdd = workarea->pdd; in kfd_sdma_activity_worker() 117 if (!pdd) in kfd_sdma_activity_worker() 345 sdma_activity_work_handler.pdd = pdd; in kfd_procfs_show() 1028 if (!pdd) in kfd_process_kunmap_signal_bo() 1065 if (pdd->qpd.cwsr_kaddr && !pdd->qpd.cwsr_base) in kfd_process_destroy_pdds() 1071 kfd_free_process_doorbells(pdd->dev->kfd, pdd); in kfd_process_destroy_pdds() 1617 pdd = kzalloc(sizeof(*pdd), GFP_KERNEL); in kfd_create_process_device_data() 1618 if (!pdd) in kfd_create_process_device_data() 1836 if (pdd) in kfd_lookup_process_by_pasid() 1846 if (pdd) in kfd_lookup_process_by_pasid() [all …]
|
| H A D | kfd_chardev.c | 77 if (pdd) in kfd_lock_pdd_by_id() 78 return pdd; in kfd_lock_pdd_by_id() 330 if (!pdd) { in kfd_ioctl_create_queue() 585 if (!pdd) { in kfd_ioctl_set_memory_policy() 591 pdd = kfd_bind_process_to_device(pdd->dev, p); in kfd_ioctl_set_memory_policy() 635 pdd = kfd_bind_process_to_device(pdd->dev, p); in kfd_ioctl_set_trap_handler() 684 if (pdd) in kfd_ioctl_get_clock_counters() 944 if (!pdd) in kfd_ioctl_get_tile_config() 1044 if (!pdd) in kfd_ioctl_get_available_memory() 1576 pdd = kfd_bind_process_to_device(pdd->dev, p); in kfd_ioctl_import_dmabuf() [all …]
|
| H A D | kfd_process_queue_manager.c | 125 if (!pdd) { in pqm_set_gws() 196 if (!pdd) { in pqm_clean_queue_resource() 208 pdd->qpd.num_gws = 0; in pqm_clean_queue_resource() 225 if (pdd) { in pqm_uninit() 229 WARN_ON(!pdd); in pqm_uninit() 340 if (!pdd) { in pqm_create_queue() 446 kq, &pdd->qpd); in pqm_create_queue() 528 if (!pdd) { in pqm_destroy_queue() 549 pdd->pasid, in pqm_destroy_queue() 594 if (!pdd) in pqm_update_queue_properties() [all …]
|
| H A D | kfd_doorbell.c | 110 struct kfd_process_device *pdd; in kfd_doorbell_mmap() local 119 pdd = kfd_get_process_device_data(dev, process); in kfd_doorbell_mmap() 120 if (!pdd) in kfd_doorbell_mmap() 124 address = kfd_get_process_doorbells(pdd); in kfd_doorbell_mmap() 237 struct amdgpu_device *adev = pdd->dev->adev; in kfd_get_process_doorbells() 240 if (!pdd->qpd.proc_doorbells) { in kfd_get_process_doorbells() 241 if (kfd_alloc_process_doorbells(pdd->dev->kfd, pdd)) in kfd_get_process_doorbells() 247 pdd->qpd.proc_doorbells, in kfd_get_process_doorbells() 256 struct qcm_process_device *qpd = &pdd->qpd; in kfd_alloc_process_doorbells() 266 r = init_doorbell_bitmap(&pdd->qpd, kfd); in kfd_alloc_process_doorbells() [all …]
|
| H A D | kfd_queue.c | 92 struct kfd_process *p = pdd->process; in kfd_queue_buffer_svm_get() 153 struct kfd_process *p = pdd->process; in kfd_queue_buffer_svm_put() 241 topo_dev = kfd_topology_device_by_id(pdd->dev->id); in kfd_queue_acquire_buffers() 254 vm = drm_priv_to_vm(pdd->drm_priv); in kfd_queue_acquire_buffers() 309 * NUM_XCC(pdd->dev->xcc_mask); in kfd_queue_acquire_buffers() 336 kfd_queue_unref_bo_vas(pdd, properties); in kfd_queue_acquire_buffers() 337 kfd_queue_release_buffers(pdd, properties); in kfd_queue_acquire_buffers() 352 topo_dev = kfd_topology_device_by_id(pdd->dev->id); in kfd_queue_release_buffers() 356 * NUM_XCC(pdd->dev->xcc_mask); in kfd_queue_release_buffers() 374 int kfd_queue_unref_bo_vas(struct kfd_process_device *pdd, in kfd_queue_unref_bo_vas() argument [all …]
|
| H A D | kfd_device_queue_manager_v9.c | 50 static uint32_t compute_sh_mem_bases_64bit(struct kfd_process_device *pdd) in compute_sh_mem_bases_64bit() argument 52 uint32_t shared_base = pdd->lds_base >> 48; in compute_sh_mem_bases_64bit() 53 uint32_t private_base = pdd->scratch_base >> 48; in compute_sh_mem_bases_64bit() 94 struct kfd_process_device *pdd = qpd_to_pdd(qpd); in update_qpd_v9() local 96 pdd = qpd_to_pdd(qpd); in update_qpd_v9() 116 if (!pdd->process->xnack_enabled) in update_qpd_v9() 122 qpd->sh_mem_bases = compute_sh_mem_bases_64bit(pdd); in update_qpd_v9()
|
| H A D | kfd_device_queue_manager.c | 809 if (!pdd) in dbgdev_wave_reset_wavefronts() 954 if (!pdd) { in update_queue() 1206 if (!pdd->drm_priv) in evict_process_queues_cpsch() 1351 if (!pdd->drm_priv) in restore_process_queues_cpsch() 2221 if (!pdd) { in reset_hung_queues() 2473 if (WARN_ON(!pdd)) in wait_on_destroy_queue() 3049 if (!pdd) in kfd_dqm_suspend_bad_queue_mes() 3054 if (pdd) { in kfd_dqm_suspend_bad_queue_mes() 3055 qpd = &pdd->qpd; in kfd_dqm_suspend_bad_queue_mes() 3134 p = pdd->process; in kfd_evict_process_device() [all …]
|
| H A D | cik_event_interrupt.c | 110 struct kfd_process_device *pdd = NULL; in cik_event_interrupt_wq() local 115 p = kfd_lookup_process_by_pasid(pasid, &pdd); in cik_event_interrupt_wq() 116 if (!pdd) in cik_event_interrupt_wq() 119 kfd_evict_process_device(pdd); in cik_event_interrupt_wq() 129 kfd_signal_vm_fault_event(pdd, &info, NULL); in cik_event_interrupt_wq() 131 kfd_signal_vm_fault_event(pdd, &info, NULL); in cik_event_interrupt_wq()
|
| H A D | kfd_svm.c | 221 if (!pdd) { in svm_range_dma_map() 269 if (!pdd) { in svm_range_dma_unmap() 422 if (pdd) in svm_range_bo_release() 650 if (pdd) in svm_range_vram_node_new() 686 if (!pdd) { in svm_range_get_node_by_id() 1485 pdd = kfd_bind_process_to_device(pdd->dev, p); in svm_range_map_to_gpus() 1587 if (!pdd) in kfd_svm_page_owner() 2303 if (!pdd) in svm_range_drain_retry_fault() 2493 if (!pdd) in svm_range_unmap_from_cpu() 2936 if (pdd) in svm_range_count_fault() [all …]
|
| H A D | kfd_priv.h | 1042 struct kfd_process_device **pdd); 1068 int kfd_process_device_init_vm(struct kfd_process_device *pdd, 1110 struct kfd_process_device *pdd, 1114 struct kfd_process_device *pdd); 1116 struct kfd_process_device *pdd); 1316 int kfd_queue_unref_bo_vas(struct kfd_process_device *pdd, 1337 int kfd_evict_process_device(struct kfd_process_device *pdd); 1510 void kfd_signal_vm_fault_event(struct kfd_process_device *pdd, 1518 static inline void kfd_flush_tlb(struct kfd_process_device *pdd, in kfd_flush_tlb() argument 1521 struct amdgpu_device *adev = pdd->dev->adev; in kfd_flush_tlb() [all …]
|
| H A D | kfd_packet_manager_v9.c | 38 struct kfd_process_device *pdd = in pm_map_process_v9() local 50 packet->bitfields2.pasid = pdd->pasid; in pm_map_process_v9() 58 if (kfd->dqm->trap_debug_vmid && pdd->process->debug_trap_enabled && in pm_map_process_v9() 59 pdd->process->runtime_info.runtime_state == DEBUG_RUNTIME_STATE_ENABLED) { in pm_map_process_v9() 96 struct kfd_process_device *pdd = in pm_map_process_aldebaran() local 109 packet->bitfields2.pasid = pdd->pasid; in pm_map_process_aldebaran() 116 packet->spi_gdbg_per_vmid_cntl = pdd->spi_dbg_override | in pm_map_process_aldebaran() 117 pdd->spi_dbg_launch_mode; in pm_map_process_aldebaran() 119 if (pdd->process->debug_trap_enabled) { in pm_map_process_aldebaran() 121 packet->tcp_watch_cntl[i] = pdd->watch_points[i]; in pm_map_process_aldebaran() [all …]
|
| H A D | kfd_device_queue_manager_v11.c | 50 static uint32_t compute_sh_mem_bases_64bit(struct kfd_process_device *pdd) in compute_sh_mem_bases_64bit() argument 52 uint32_t shared_base = pdd->lds_base >> 48; in compute_sh_mem_bases_64bit() 53 uint32_t private_base = pdd->scratch_base >> 48; in compute_sh_mem_bases_64bit()
|
| H A D | kfd_device_queue_manager_v10.c | 51 static uint32_t compute_sh_mem_bases_64bit(struct kfd_process_device *pdd) in compute_sh_mem_bases_64bit() argument 53 uint32_t shared_base = pdd->lds_base >> 48; in compute_sh_mem_bases_64bit() 54 uint32_t private_base = pdd->scratch_base >> 48; in compute_sh_mem_bases_64bit()
|
| H A D | kfd_device_queue_manager_v12.c | 50 static uint32_t compute_sh_mem_bases_64bit(struct kfd_process_device *pdd) in compute_sh_mem_bases_64bit() argument 52 uint32_t shared_base = pdd->lds_base >> 48; in compute_sh_mem_bases_64bit() 53 uint32_t private_base = pdd->scratch_base >> 48; in compute_sh_mem_bases_64bit()
|
| H A D | kfd_events.c | 353 struct kfd_process_device *pdd; in kfd_kmap_event_page() local 364 if (!pdd) { in kfd_kmap_event_page() 368 kfd = pdd->dev; in kfd_kmap_event_page() 370 pdd = kfd_bind_process_to_device(kfd, p); in kfd_kmap_event_page() 371 if (IS_ERR(pdd)) in kfd_kmap_event_page() 372 return PTR_ERR(pdd); in kfd_kmap_event_page() 374 mem = kfd_process_device_translate_handle(pdd, in kfd_kmap_event_page() 1186 struct kfd_process *p = pdd->process; in kfd_signal_vm_fault_event() 1193 pdd->dev->id); in kfd_signal_vm_fault_event() 1263 if (unlikely(!pdd)) { in kfd_signal_reset_event() [all …]
|
| H A D | kfd_debug.h | 58 int kfd_dbg_trap_clear_dev_address_watch(struct kfd_process_device *pdd, 60 int kfd_dbg_trap_set_dev_address_watch(struct kfd_process_device *pdd, 131 int kfd_dbg_set_mes_debug_mode(struct kfd_process_device *pdd, bool sq_trap_en);
|
| H A D | kfd_device_queue_manager.h | 331 static inline unsigned int get_sh_mem_bases_32(struct kfd_process_device *pdd) in get_sh_mem_bases_32() argument 333 return (pdd->lds_base >> 16) & 0xFF; in get_sh_mem_bases_32() 337 get_sh_mem_bases_nybble_64(struct kfd_process_device *pdd) in get_sh_mem_bases_nybble_64() argument 339 return (pdd->lds_base >> 60) & 0x0E; in get_sh_mem_bases_nybble_64()
|
| H A D | kfd_migrate.c | 394 struct kfd_process_device *pdd; in svm_migrate_vma_to_vram() local 463 pdd = svm_range_get_pdd_by_node(prange, node); in svm_migrate_vma_to_vram() 464 if (pdd) in svm_migrate_vma_to_vram() 465 WRITE_ONCE(pdd->page_in, pdd->page_in + mpages); in svm_migrate_vma_to_vram() 695 struct kfd_process_device *pdd; in svm_migrate_vma_to_ram() local 769 pdd = svm_range_get_pdd_by_node(prange, node); in svm_migrate_vma_to_ram() 770 if (pdd) in svm_migrate_vma_to_ram() 771 WRITE_ONCE(pdd->page_out, pdd->page_out + mpages); in svm_migrate_vma_to_ram()
|
| H A D | kfd_smi_events.c | 329 struct kfd_process_device *pdd = p->pdds[i]; in kfd_smi_event_queue_restore_rescheduled() local 331 kfd_smi_event_add(p->lead_thread->pid, pdd->dev, in kfd_smi_event_queue_restore_rescheduled() 334 p->lead_thread->pid, pdd->dev->id, 'R')); in kfd_smi_event_queue_restore_rescheduled()
|
| /linux-6.15/drivers/pmdomain/ |
| H A D | governor.c | 129 struct pm_domain_data *pdd; in update_domain_next_wakeup() local 143 list_for_each_entry(pdd, &genpd->dev_list, list_node) { in update_domain_next_wakeup() 144 next_wakeup = to_gpd_data(pdd)->td->next_wakeup; in update_domain_next_wakeup() 181 struct pm_domain_data *pdd; in __default_power_down_ok() local 217 list_for_each_entry(pdd, &genpd->dev_list, list_node) { in __default_power_down_ok() 226 td = to_gpd_data(pdd)->td; in __default_power_down_ok()
|
| H A D | core.c | 317 struct pm_domain_data *pdd; in _genpd_reeval_performance_state() local 330 pd_data = to_gpd_data(pdd); in _genpd_reeval_performance_state() 859 struct pm_domain_data *pdd; in genpd_power_off() local 904 if (to_gpd_data(pdd)->rpm_always_on) in genpd_power_off() 1020 struct pm_domain_data *pdd; in genpd_dev_pm_qos_notifier() local 1025 pdd = dev->power.subsys_data ? in genpd_dev_pm_qos_notifier() 1027 if (pdd) { in genpd_dev_pm_qos_notifier() 1028 td = to_gpd_data(pdd)->td; in genpd_dev_pm_qos_notifier() 1887 struct pm_domain_data *pdd; in genpd_remove_device() local 1893 gpd_data = to_gpd_data(pdd); in genpd_remove_device() [all …]
|
| /linux-6.15/drivers/input/touchscreen/ |
| H A D | wm9705.c | 77 static int pdd = 8; variable 78 module_param(pdd, int, 0); 79 MODULE_PARM_DESC(pdd, "Set pen detect comparator threshold"); 168 dig2 |= (pdd & 0x000f); in wm9705_phy_init() 169 dev_dbg(wm->dev, "setting pdd to Vmid/%d", 1 - (pdd & 0x000f)); in wm9705_phy_init()
|
| /linux-6.15/drivers/pmdomain/xilinx/ |
| H A D | zynqmp-pm-domains.c | 108 struct pm_domain_data *pdd, *tmp; in zynqmp_gpd_power_off() local 119 list_for_each_entry_safe(pdd, tmp, &domain->dev_list, list_node) { in zynqmp_gpd_power_off() 121 may_wakeup = zynqmp_gpd_is_active_wakeup_path(pdd->dev, NULL); in zynqmp_gpd_power_off() 123 dev_dbg(pdd->dev, "device is in wakeup path in %s\n", in zynqmp_gpd_power_off()
|