Home
last modified time | relevance | path

Searched refs:kfd_node (Results 1 – 25 of 40) sorted by relevance

12

/linux-6.15/drivers/gpu/drm/amd/amdkfd/
H A Dkfd_smi_events.h29 int kfd_smi_event_open(struct kfd_node *dev, uint32_t *fd);
30 void kfd_smi_event_update_vmfault(struct kfd_node *dev, uint16_t pasid);
31 void kfd_smi_event_update_thermal_throttling(struct kfd_node *dev,
33 void kfd_smi_event_update_gpu_reset(struct kfd_node *dev, bool post_reset,
35 void kfd_smi_event_page_fault_start(struct kfd_node *node, pid_t pid,
38 void kfd_smi_event_page_fault_end(struct kfd_node *node, pid_t pid,
40 void kfd_smi_event_migration_start(struct kfd_node *node, pid_t pid,
45 void kfd_smi_event_migration_end(struct kfd_node *node, pid_t pid,
49 void kfd_smi_event_queue_eviction(struct kfd_node *node, pid_t pid,
51 void kfd_smi_event_queue_restore(struct kfd_node *node, pid_t pid);
[all …]
H A Dkfd_priv.h213 struct kfd_node;
261 struct kfd_node { struct
607 struct kfd_node *device;
744 struct kfd_node *dev;
1321 struct kfd_node *dev);
1323 struct kfd_node *dev);
1325 struct kfd_node *dev);
1327 struct kfd_node *dev);
1329 struct kfd_node *dev);
1331 struct kfd_node *dev);
[all …]
H A Dkfd_interrupt.c53 int kfd_interrupt_init(struct kfd_node *node) in kfd_interrupt_init()
90 void kfd_interrupt_exit(struct kfd_node *node) in kfd_interrupt_exit()
108 bool enqueue_ih_ring_entry(struct kfd_node *node, const void *ih_ring_entry) in enqueue_ih_ring_entry()
123 static bool dequeue_ih_ring_entry(struct kfd_node *node, u32 **ih_ring_entry) in dequeue_ih_ring_entry()
138 struct kfd_node *dev = container_of(work, struct kfd_node, interrupt_work); in interrupt_wq()
157 bool interrupt_is_wanted(struct kfd_node *dev, in interrupt_is_wanted()
H A Dkfd_smi_events.c40 struct kfd_node *dev;
153 struct kfd_node *dev = client->dev; in kfd_smi_ev_release()
175 static void add_event_to_kfifo(pid_t pid, struct kfd_node *dev, in add_event_to_kfifo()
200 static void kfd_smi_event_add(pid_t pid, struct kfd_node *dev, in kfd_smi_event_add()
242 void kfd_smi_event_update_thermal_throttling(struct kfd_node *dev, in kfd_smi_event_update_thermal_throttling()
273 void kfd_smi_event_page_fault_end(struct kfd_node *node, pid_t pid, in kfd_smi_event_page_fault_end()
293 void kfd_smi_event_migration_end(struct kfd_node *node, pid_t pid, in kfd_smi_event_migration_end()
304 void kfd_smi_event_queue_eviction(struct kfd_node *node, pid_t pid, in kfd_smi_event_queue_eviction()
312 void kfd_smi_event_queue_restore(struct kfd_node *node, pid_t pid) in kfd_smi_event_queue_restore()
339 void kfd_smi_event_unmap_from_gpu(struct kfd_node *node, pid_t pid, in kfd_smi_event_unmap_from_gpu()
[all …]
H A Dkfd_debug.h35 bool kfd_set_dbg_ev_from_interrupt(struct kfd_node *dev,
42 struct kfd_process *process, struct kfd_node *dev,
77 static inline bool kfd_dbg_is_per_vmid_supported(struct kfd_node *dev) in kfd_dbg_is_per_vmid_supported()
100 static inline bool kfd_dbg_is_rlc_restore_supported(struct kfd_node *dev) in kfd_dbg_is_rlc_restore_supported()
106 static inline bool kfd_dbg_has_cwsr_workaround(struct kfd_node *dev) in kfd_dbg_has_cwsr_workaround()
112 static inline bool kfd_dbg_has_gws_support(struct kfd_node *dev) in kfd_dbg_has_gws_support()
133 static inline bool kfd_dbg_has_ttmps_always_setup(struct kfd_node *dev) in kfd_dbg_has_ttmps_always_setup()
H A Dkfd_mqd_manager.h71 struct kfd_mem_obj* (*allocate_mqd)(struct kfd_node *kfd,
127 struct kfd_node *dev;
156 struct kfd_mem_obj *allocate_hiq_mqd(struct kfd_node *dev,
159 struct kfd_mem_obj *allocate_sdma_mqd(struct kfd_node *dev,
195 void kfd_get_hiq_xcc_mqd(struct kfd_node *dev,
198 uint64_t kfd_hiq_mqd_stride(struct kfd_node *dev);
201 bool kfd_check_hiq_mqd_doorbell_id(struct kfd_node *node, uint32_t doorbell_id,
H A Dkfd_device.c65 static int kfd_resume(struct kfd_node *kfd);
660 struct kfd_node *knode; in kfd_cleanup_nodes()
722 struct kfd_node *node; in kgd2kfd_device_init()
968 struct kfd_node *node; in kgd2kfd_pre_reset()
996 struct kfd_node *node; in kgd2kfd_post_reset()
1029 struct kfd_node *node; in kgd2kfd_suspend()
1075 static int kfd_resume(struct kfd_node *node) in kfd_resume()
1094 struct kfd_node *node; in kgd2kfd_interrupt()
1473 struct kfd_node *node; in kgd2kfd_start_sched()
1495 struct kfd_node *node; in kgd2kfd_stop_sched()
[all …]
H A Dkfd_topology.h93 struct kfd_node *gpu;
111 struct kfd_node *gpu;
132 struct kfd_node *gpu;
154 struct kfd_node *gpu;
H A Dkfd_svm.h51 struct kfd_node *node;
171 struct kfd_node *svm_range_get_node_by_id(struct svm_range *prange,
173 int svm_range_vram_node_new(struct kfd_node *node, struct svm_range *prange,
198 svm_range_get_pdd_by_node(struct svm_range *prange, struct kfd_node *node);
H A Dkfd_packet_manager.c52 struct kfd_node *node = pm->dqm->dev; in pm_calc_rlib_size()
101 struct kfd_node *node = pm->dqm->dev; in pm_allocate_runlist_ib()
137 struct kfd_node *node = pm->dqm->dev; in pm_create_runlist_ib()
297 struct kfd_node *node = pm->dqm->dev; in pm_send_set_resources()
370 struct kfd_node *node = pm->dqm->dev; in pm_send_query_status()
412 struct kfd_node *node = pm->dqm->dev; in pm_config_dequeue_wait_counts()
462 struct kfd_node *node = pm->dqm->dev; in pm_send_unmap_queue()
521 struct kfd_node *node = pm->dqm->dev; in pm_debugfs_hang_hws()
H A Dkfd_mqd_manager.c49 struct kfd_mem_obj *allocate_hiq_mqd(struct kfd_node *dev, struct queue_properties *q) in allocate_hiq_mqd()
64 struct kfd_mem_obj *allocate_sdma_mqd(struct kfd_node *dev, in allocate_sdma_mqd()
271 uint64_t kfd_hiq_mqd_stride(struct kfd_node *dev) in kfd_hiq_mqd_stride()
276 void kfd_get_hiq_xcc_mqd(struct kfd_node *dev, struct kfd_mem_obj *mqd_mem_obj, in kfd_get_hiq_xcc_mqd()
296 bool kfd_check_hiq_mqd_doorbell_id(struct kfd_node *node, uint32_t doorbell_id, in kfd_check_hiq_mqd_doorbell_id()
H A Dkfd_crat.h298 struct kfd_node;
311 int kfd_get_gpu_cache_info(struct kfd_node *kdev, struct kfd_gpu_cache_info **pcache_info);
317 int flags, struct kfd_node *kdev,
H A Dcik_event_interrupt.c29 static bool cik_event_interrupt_isr(struct kfd_node *dev, in cik_event_interrupt_isr()
88 static void cik_event_interrupt_wq(struct kfd_node *dev, in cik_event_interrupt_wq()
H A Dkfd_process.c271 struct kfd_node *dev = NULL; in kfd_get_cu_occupancy()
721 struct kfd_node *dev = pdd->dev; in kfd_process_free_gpuvm()
743 struct kfd_node *kdev = pdd->dev; in kfd_process_alloc_gpuvm()
1018 struct kfd_node *kdev; in kfd_process_kunmap_signal_bo()
1339 struct kfd_node *dev = p->pdds[i]->dev; in kfd_process_init_cwsr_apu()
1376 struct kfd_node *dev = pdd->dev; in kfd_process_device_init_cwsr_dgpu()
1412 struct kfd_node *dev = pdd->dev; in kfd_process_device_destroy_cwsr_dgpu()
1459 struct kfd_node *dev = p->pdds[i]->dev; in kfd_process_xnack_mode()
1671 struct kfd_node *dev; in kfd_process_device_init_vm()
1738 struct kfd_process_device *kfd_bind_process_to_device(struct kfd_node *dev, in kfd_bind_process_to_device()
[all …]
H A Dkfd_int_process_v9.c144 static void event_interrupt_poison_consumption_v9(struct kfd_node *dev, in event_interrupt_poison_consumption_v9()
259 static bool event_interrupt_isr_v9(struct kfd_node *dev, in event_interrupt_isr_v9()
356 static void event_interrupt_wq_v9(struct kfd_node *dev, in event_interrupt_wq_v9()
495 static bool event_interrupt_isr_v9_4_3(struct kfd_node *node, in event_interrupt_isr_v9_4_3()
H A Dkfd_int_process_v11.c191 static void event_interrupt_poison_consumption_v11(struct kfd_node *dev, in event_interrupt_poison_consumption_v11()
233 static bool event_interrupt_isr_v11(struct kfd_node *dev, in event_interrupt_isr_v11()
282 static void event_interrupt_wq_v11(struct kfd_node *dev, in event_interrupt_wq_v11()
H A Dkfd_kernel_queue.h56 struct kfd_node *dev;
H A Dkfd_kernel_queue.c42 static bool kq_initialize(struct kernel_queue *kq, struct kfd_node *dev, in kq_initialize()
338 struct kernel_queue *kernel_queue_init(struct kfd_node *dev, in kernel_queue_init()
363 static __attribute__((unused)) void test_kq(struct kfd_node *dev) in test_kq()
H A Dkfd_device_queue_manager.h219 struct kfd_node *dev);
247 struct kfd_node *dev;
H A Dkfd_int_process_v10.c132 static bool event_interrupt_isr_v10(struct kfd_node *dev, in event_interrupt_isr_v10()
194 static void event_interrupt_wq_v10(struct kfd_node *dev, in event_interrupt_wq_v10()
H A Dkfd_debugfs.c46 struct kfd_node *dev; in kfd_debugfs_hang_hws_write()
H A Dkfd_process_queue_manager.c85 struct kfd_node *dev = pdd->dev; in kfd_process_dequeue_from_device()
107 struct kfd_node *dev = NULL; in pqm_set_gws()
190 struct kfd_node *dev; in pqm_clean_queue_resource()
245 struct kfd_node *dev, struct queue **q, in init_user_queue()
311 struct kfd_node *dev, in pqm_create_queue()
506 struct kfd_node *dev; in pqm_destroy_queue()
H A Dkfd_packet_manager_v9.c37 struct kfd_node *kfd = pm->dqm->dev; in pm_map_process_v9()
95 struct kfd_node *knode = pm->dqm->dev; in pm_map_process_aldebaran()
153 struct kfd_node *kfd = pm->dqm->dev; in pm_runlist_v9()
H A Dkfd_crat.c1579 int kfd_get_gpu_cache_info(struct kfd_node *kdev, struct kfd_gpu_cache_info **pcache_info) in kfd_get_gpu_cache_info()
1982 struct kfd_node *kdev, uint8_t type, uint64_t size, in kfd_fill_gpu_memory_affinity()
2011 static void kfd_find_numa_node_in_srat(struct kfd_node *kdev) in kfd_find_numa_node_in_srat()
2109 struct kfd_node *kdev, in kfd_fill_gpu_direct_io_link_to_cpu()
2181 struct kfd_node *kdev, in kfd_fill_gpu_xgmi_link_to_gpu()
2182 struct kfd_node *peer_kdev, in kfd_fill_gpu_xgmi_link_to_gpu()
2233 size_t *size, struct kfd_node *kdev, in kfd_create_vcrat_image_gpu()
2403 int flags, struct kfd_node *kdev, in kfd_create_crat_image_virtual()
H A Dkfd_migrate.c277 svm_migrate_copy_to_vram(struct kfd_node *node, struct svm_range *prange, in svm_migrate_copy_to_vram()
387 svm_migrate_vma_to_vram(struct kfd_node *node, struct svm_range *prange, in svm_migrate_vma_to_vram()
494 struct kfd_node *node; in svm_migrate_ram_to_vram()
685 svm_migrate_vma_to_ram(struct kfd_node *node, struct svm_range *prange, in svm_migrate_vma_to_ram()
795 struct kfd_node *node; in svm_migrate_vram_to_ram()

12