| /linux-6.15/drivers/gpu/drm/amd/pm/ |
| H A D | amdgpu_dpm.c | 37 ((adev)->powerplay.pp_funcs->enable_bapm((adev)->powerplay.pp_handle, (e))) 39 #define amdgpu_dpm_is_legacy_dpm(adev) ((adev)->powerplay.pp_handle == (adev)) argument 49 mutex_lock(&adev->pm.mutex); in amdgpu_dpm_get_sclk() 65 mutex_lock(&adev->pm.mutex); in amdgpu_dpm_get_mclk() 90 mutex_lock(&adev->pm.mutex); in amdgpu_dpm_set_powergating_by_smu() 232 if (adev->in_s3) in amdgpu_dpm_is_baco_supported() 496 amdgpu_dpm_enable_bapm(adev, adev->pm.ac_power); in amdgpu_pm_acpi_event_handler() 664 (is_support_sw_smu(adev) && (adev->flags & AMD_IS_APU))) in amdgpu_pm_load_smu_firmware() 738 if (adev->cper.enabled) in amdgpu_dpm_send_rma_reason() 1334 adev->pm.dpm.current_ps = adev->pm.dpm.boot_ps; in amdgpu_dpm_set_sclk_od() [all …]
|
| /linux-6.15/drivers/gpu/drm/amd/amdgpu/ |
| H A D | amdgpu_device.c | 740 ret = adev->pcie_rreg(adev, reg * 4); in amdgpu_device_rreg() 808 ret = adev->pcie_rreg(adev, reg * 4); in amdgpu_device_xcc_rreg() 867 adev->pcie_wreg(adev, reg * 4, v); in amdgpu_device_wreg() 896 adev->pcie_wreg(adev, reg * 4, v); in amdgpu_mm_wreg_mmio_rlc() 939 adev->pcie_wreg(adev, reg * 4, v); in amdgpu_device_xcc_wreg() 2416 adev->ip_blocks[adev->num_ip_blocks].adev = adev; in amdgpu_device_ip_block_add() 3287 adev = gpu_ins->adev; in amdgpu_device_enable_mgpu_fan_boost() 4108 adev->sdma_timeout = adev->video_timeout = adev->gfx_timeout; in amdgpu_device_get_job_timeout_settings() 4155 adev->sdma_timeout = adev->video_timeout = adev->gfx_timeout; in amdgpu_device_get_job_timeout_settings() 4354 adev->gfx.enforce_isolation[i].adev = adev; in amdgpu_device_init() [all …]
|
| H A D | gmc_v10_0.c | 428 struct amdgpu_device *adev = ring->adev; in gmc_v10_0_emit_pasid_mapping() local 690 adev->vm_manager.vram_base_offset = adev->gfxhub.funcs->get_mc_fb_offset(adev); in gmc_v10_0_vram_gtt_location() 725 adev->gmc.aper_base = adev->gfxhub.funcs->get_mc_fb_offset(adev); in gmc_v10_0_mc_init() 780 adev->gfxhub.funcs->init(adev); in gmc_v10_0_sw_init() 782 adev->mmhub.funcs->init(adev); in gmc_v10_0_sw_init() 965 r = adev->mmhub.funcs->gart_enable(adev); in gmc_v10_0_gart_enable() 969 adev->hdp.funcs->init_registers(adev); in gmc_v10_0_gart_enable() 1006 adev->gfxhub.funcs->utcl2_harvest(adev); in gmc_v10_0_hw_init() 1019 adev->umc.funcs->init_registers(adev); in gmc_v10_0_hw_init() 1034 adev->gfxhub.funcs->gart_disable(adev); in gmc_v10_0_gart_disable() [all …]
|
| H A D | gmc_v11_0.c | 76 if (!adev->in_s0ix && (adev->in_runpm || adev->in_suspend || in gmc_v11_0_vm_fault_interrupt_state() 393 struct amdgpu_device *adev = ring->adev; in gmc_v11_0_emit_pasid_mapping() local 611 struct amdgpu_device *adev = ip_block->adev; in gmc_v11_0_early_init() local 632 struct amdgpu_device *adev = ip_block->adev; in gmc_v11_0_late_init() local 665 adev->vm_manager.vram_base_offset = adev->mmhub.funcs->get_mc_fb_offset(adev); in gmc_v11_0_vram_gtt_location() 696 adev->gmc.aper_base = adev->mmhub.funcs->get_mc_fb_offset(adev); in gmc_v11_0_mc_init() 742 adev->mmhub.funcs->init(adev); in gmc_v11_0_sw_init() 744 adev->gfxhub.funcs->init(adev); in gmc_v11_0_sw_init() 909 r = adev->mmhub.funcs->gart_enable(adev); in gmc_v11_0_gart_enable() 944 adev->umc.funcs->init_registers(adev); in gmc_v11_0_hw_init() [all …]
|
| H A D | soc24.c | 93 return adev->nbio.funcs->get_memsize(adev); in soc24_get_config_memsize() 249 adev->nbio.funcs->program_aspm(adev); in soc24_program_aspm() 367 struct amdgpu_device *adev = ip_block->adev; in soc24_common_early_init() local 369 adev->nbio.funcs->set_reg_remap(adev); in soc24_common_early_init() 444 struct amdgpu_device *adev = ip_block->adev; in soc24_common_late_init() local 469 struct amdgpu_device *adev = ip_block->adev; in soc24_common_sw_init() local 479 struct amdgpu_device *adev = ip_block->adev; in soc24_common_hw_init() local 484 adev->nbio.funcs->init_registers(adev); in soc24_common_hw_init() 493 adev->df.funcs->hw_init(adev); in soc24_common_hw_init() 503 struct amdgpu_device *adev = ip_block->adev; in soc24_common_hw_fini() local [all …]
|
| H A D | gmc_v9_0.c | 991 struct amdgpu_device *adev = ring->adev; in gmc_v9_0_emit_flush_gpu_tlb() local 1039 struct amdgpu_device *adev = ring->adev; in gmc_v9_0_emit_pasid_mapping() local 1653 adev->smuio.funcs->get_pkg_type(adev); in gmc_v9_0_early_init() 1738 adev->vm_manager.vram_base_offset = adev->gfxhub.funcs->get_mc_fb_offset(adev); in gmc_v9_0_vram_gtt_location() 2086 adev->gfxhub.funcs->init(adev); in gmc_v9_0_sw_init() 2088 adev->mmhub.funcs->init(adev); in gmc_v9_0_sw_init() 2383 r = adev->mmhub.funcs->gart_enable(adev); in gmc_v9_0_gart_enable() 2435 adev->hdp.funcs->init_registers(adev); in gmc_v9_0_hw_init() 2457 adev->umc.funcs->init_registers(adev); in gmc_v9_0_hw_init() 2479 adev->gfxhub.funcs->gart_disable(adev); in gmc_v9_0_gart_disable() [all …]
|
| H A D | gmc_v12_0.c | 413 struct amdgpu_device *adev = ring->adev; in gmc_v12_0_emit_pasid_mapping() local 607 struct amdgpu_device *adev = ip_block->adev; in gmc_v12_0_early_init() local 627 struct amdgpu_device *adev = ip_block->adev; in gmc_v12_0_late_init() local 658 adev->vm_manager.vram_base_offset = adev->mmhub.funcs->get_mc_fb_offset(adev); in gmc_v12_0_vram_gtt_location() 690 adev->gmc.aper_base = adev->mmhub.funcs->get_mc_fb_offset(adev); in gmc_v12_0_mc_init() 735 struct amdgpu_device *adev = ip_block->adev; in gmc_v12_0_sw_init() local 737 adev->mmhub.funcs->init(adev); in gmc_v12_0_sw_init() 739 adev->gfxhub.funcs->init(adev); in gmc_v12_0_sw_init() 879 r = adev->mmhub.funcs->gart_enable(adev); in gmc_v12_0_gart_enable() 912 adev->umc.funcs->init_registers(adev); in gmc_v12_0_hw_init() [all …]
|
| H A D | amdgpu_virt.c | 68 adev->cg_flags = 0; in amdgpu_virt_init_setting() 69 adev->pg_flags = 0; in amdgpu_virt_init_setting() 198 if (!amdgpu_sriov_vf(adev) || adev->virt.mm_table.gpu_addr) in amdgpu_virt_alloc_mm_table() 226 if (!amdgpu_sriov_vf(adev) || !adev->virt.mm_table.gpu_addr) in amdgpu_virt_free_mm_table() 474 dev_err(adev->dev, in amdgpu_virt_read_pf2vf_data() 490 dev_err(adev->dev, in amdgpu_virt_read_pf2vf_data() 524 adev->unique_id = in amdgpu_virt_read_pf2vf_data() 677 if (adev->mman.fw_vram_usage_va && adev->mman.drv_vram_usage_va) { in amdgpu_virt_init_data_exchange() 702 if (adev->mman.fw_vram_usage_va || adev->mman.drv_vram_usage_va) { in amdgpu_virt_exchange_data() 1062 dev_err(adev->dev, in amdgpu_virt_rlcg_reg_rw() [all …]
|
| H A D | soc15.c | 342 return adev->nbio.funcs->get_memsize(adev); in soc15_get_config_memsize() 699 adev->nbio.funcs->program_aspm(adev); in soc15_program_aspm() 962 struct amdgpu_device *adev = ip_block->adev; in soc15_common_early_init() local 964 adev->nbio.funcs->set_reg_remap(adev); in soc15_common_early_init() 1232 struct amdgpu_device *adev = ip_block->adev; in soc15_common_late_init() local 1254 adev->df.funcs->sw_init(adev); in soc15_common_sw_init() 1265 adev->df.funcs->sw_fini(adev); in soc15_common_sw_fini() 1290 adev->nbio.funcs->init_registers(adev); in soc15_common_hw_init() 1427 adev->hdp.funcs->update_clock_gating(adev, in soc15_common_set_clockgating_state() 1445 adev->hdp.funcs->update_clock_gating(adev, in soc15_common_set_clockgating_state() [all …]
|
| H A D | soc21.c | 223 return adev->nbio.funcs->get_memsize(adev); in soc21_get_config_memsize() 441 adev->nbio.funcs->program_aspm(adev); in soc21_program_aspm() 556 struct amdgpu_device *adev = ip_block->adev; in soc21_common_early_init() local 558 adev->nbio.funcs->set_reg_remap(adev); in soc21_common_early_init() 649 adev->external_rev_id = adev->rev_id + 0x1; in soc21_common_early_init() 822 struct amdgpu_device *adev = ip_block->adev; in soc21_common_late_init() local 860 struct amdgpu_device *adev = ip_block->adev; in soc21_common_sw_init() local 870 struct amdgpu_device *adev = ip_block->adev; in soc21_common_hw_init() local 875 adev->nbio.funcs->init_registers(adev); in soc21_common_hw_init() 890 struct amdgpu_device *adev = ip_block->adev; in soc21_common_hw_fini() local [all …]
|
| H A D | amdgpu_bios.c | 89 kfree(adev->bios); in amdgpu_bios_release() 90 adev->bios = NULL; in amdgpu_bios_release() 115 adev->bios = NULL; in amdgpu_read_bios_from_vram() 122 if (!adev->bios) { in amdgpu_read_bios_from_vram() 171 if (!adev->asic_funcs || !adev->asic_funcs->read_bios_from_rom) in amdgpu_read_bios_from_rom() 196 amdgpu_asic_read_bios_from_rom(adev, adev->bios, len); in amdgpu_read_bios_from_rom() 218 if (!adev->bios) in amdgpu_read_platform_bios() 525 adev->is_atom_fw = adev->asic_type >= CHIP_VEGA10; in amdgpu_get_bios() 556 adev->smuio.funcs->get_rom_index_offset(adev); in amdgpu_soc15_read_bios_from_rom() 558 adev->smuio.funcs->get_rom_data_offset(adev); in amdgpu_soc15_read_bios_from_rom() [all …]
|
| H A D | amdgpu_rlc.c | 44 if (!adev->gfx.rlc.funcs->is_rlc_enabled(adev)) in amdgpu_gfx_rlc_enter_safe_mode() 47 if (adev->cg_flags & in amdgpu_gfx_rlc_enter_safe_mode() 50 adev->gfx.rlc.funcs->set_safe_mode(adev, xcc_id); in amdgpu_gfx_rlc_enter_safe_mode() 69 if (!adev->gfx.rlc.funcs->is_rlc_enabled(adev)) in amdgpu_gfx_rlc_exit_safe_mode() 72 if (adev->cg_flags & in amdgpu_gfx_rlc_exit_safe_mode() 75 adev->gfx.rlc.funcs->unset_safe_mode(adev, xcc_id); in amdgpu_gfx_rlc_exit_safe_mode() 105 amdgpu_gfx_rlc_fini(adev); in amdgpu_gfx_rlc_init_sr() 134 adev->gfx.rlc.clear_state_size = dws = adev->gfx.rlc.funcs->get_csb_size(adev); in amdgpu_gfx_rlc_init_csb() 162 r = amdgpu_bo_create_reserved(adev, adev->gfx.rlc.cp_table_size, in amdgpu_gfx_rlc_init_cpt() 197 max_me = adev->gfx.rlc.funcs->get_cp_table_num(adev); in amdgpu_gfx_rlc_setup_cp_table() [all …]
|
| H A D | nv.c | 308 return adev->nbio.funcs->get_memsize(adev); in nv_get_config_memsize() 520 adev->nbio.funcs->program_aspm(adev); in nv_program_aspm() 613 adev->nbio.funcs->enable_aspm(adev, !enter); in nv_update_umd_stable_pstate() 640 struct amdgpu_device *adev = ip_block->adev; in nv_common_early_init() local 642 adev->nbio.funcs->set_reg_remap(adev); in nv_common_early_init() 687 adev->external_rev_id = adev->rev_id + 0x1; in nv_common_early_init() 708 adev->external_rev_id = adev->rev_id + 20; in nv_common_early_init() 737 adev->external_rev_id = adev->rev_id + 0xa; in nv_common_early_init() 950 struct amdgpu_device *adev = ip_block->adev; in nv_common_late_init() local 1000 adev->nbio.funcs->init_registers(adev); in nv_common_hw_init() [all …]
|
| H A D | amdgpu_acp.c | 103 struct amdgpu_device *adev = ip_block->adev; in acp_sw_init() local 105 adev->acp.parent = adev->dev; in acp_sw_init() 117 struct amdgpu_device *adev = ip_block->adev; in acp_sw_fini() local 136 adev = apd->adev; in acp_poweroff() 153 adev = apd->adev; in acp_poweron() 233 struct amdgpu_device *adev = ip_block->adev; in acp_hw_init() local 256 adev->acp.acp_genpd->adev = adev; in acp_hw_init() 507 struct amdgpu_device *adev = ip_block->adev; in acp_hw_fini() local 564 struct amdgpu_device *adev = ip_block->adev; in acp_suspend() local 574 struct amdgpu_device *adev = ip_block->adev; in acp_resume() local [all …]
|
| H A D | amdgpu_virt.h | 305 #define amdgpu_sriov_vf(adev) \ argument 308 #define amdgpu_sriov_bios(adev) \ argument 315 (amdgpu_sriov_vf((adev)) && !amdgpu_sriov_runtime((adev))) 318 (amdgpu_sriov_vf((adev)) && \ 322 (amdgpu_sriov_vf((adev)) && \ 326 (amdgpu_sriov_vf((adev)) && \ 330 (amdgpu_sriov_vf((adev)) && \ 334 (amdgpu_sriov_reg_indirect_mmhub(adev) || amdgpu_sriov_reg_indirect_gc(adev)) 368 (amdgpu_sriov_vf(adev) && !amdgpu_sriov_is_pp_one_vf(adev)) 370 ((!amdgpu_in_reset(adev)) && adev->virt.tdr_debug) [all …]
|
| H A D | amdgpu_gfx.c | 1438 struct amdgpu_device *adev = ring->adev; in amdgpu_gfx_run_cleaner_shader_job() local 1552 if (adev->in_suspend && !adev->in_runpm) in amdgpu_gfx_set_run_cleaner_shader() 1692 if (!adev) in amdgpu_gfx_get_gfx_reset_mask() 1705 if (!adev) in amdgpu_gfx_get_compute_reset_mask() 1977 struct amdgpu_device *adev = isolation_work->adev; in amdgpu_gfx_enforce_isolation_handler() local 2080 struct amdgpu_device *adev = ring->adev; in amdgpu_gfx_enforce_isolation_ring_begin_use() local 2120 struct amdgpu_device *adev = ring->adev; in amdgpu_gfx_enforce_isolation_ring_end_use() local 2181 struct amdgpu_device *adev = ring->adev; in amdgpu_gfx_profile_ring_begin_use() local 2231 if (!adev) in amdgpu_debugfs_gfx_sched_mask_set() 2257 if (!adev) in amdgpu_debugfs_gfx_sched_mask_get() [all …]
|
| H A D | amdgpu_jpeg.c | 133 struct amdgpu_device *adev = ring->adev; in amdgpu_jpeg_ring_begin_use() local 152 struct amdgpu_device *adev = ring->adev; in amdgpu_jpeg_dec_ring_test_ring() local 189 struct amdgpu_device *adev = ring->adev; in amdgpu_jpeg_dec_set_reg() local 228 struct amdgpu_device *adev = ring->adev; in amdgpu_jpeg_dec_ring_test_ib() local 299 r = amdgpu_irq_get(adev, &adev->jpeg.inst[i].ras_poison_irq, 0); in amdgpu_jpeg_ras_late_init() 361 if (!adev) in amdgpu_debugfs_jpeg_sched_mask_set() 389 if (!adev) in amdgpu_debugfs_jpeg_sched_mask_get() 430 if (!adev) in amdgpu_get_jpeg_reset_mask() 463 adev->jpeg.ip_dump = kcalloc(adev->jpeg.num_jpeg_inst * count, in amdgpu_jpeg_reg_dump_init() 484 struct amdgpu_device *adev = ip_block->adev; in amdgpu_jpeg_dump_ip_state() local [all …]
|
| H A D | amdgpu_discovery.c | 463 adev, adev->mman.discovery_bin); in amdgpu_discovery_init() 470 dev_err(adev->dev, in amdgpu_discovery_init() 623 dev_err(adev->dev, in amdgpu_discovery_validate_ip() 766 adev->umc.active_mask = ((1 << adev->umc.node_inst_num) - 1) & in amdgpu_discovery_read_from_harvest_table() 1014 struct amdgpu_device *adev = ip_top->adev; in ip_disc_release() local 1218 adev->ip_top = kzalloc(sizeof(*adev->ip_top), GFP_KERNEL); in amdgpu_discovery_sysfs_init() 1219 if (!adev->ip_top) in amdgpu_discovery_sysfs_init() 1222 adev->ip_top->adev = adev; in amdgpu_discovery_sysfs_init() 1389 adev->vcn.inst[adev->vcn.num_vcn_inst].vcn_config = in amdgpu_discovery_reg_base_init() 2363 if (!(adev->asic_type == CHIP_VEGA20 && amdgpu_sriov_vf(adev))) in amdgpu_discovery_set_mm_ip_blocks() [all …]
|
| H A D | amdgpu_amdkfd.c | 77 adev->kfd.dev = kgd2kfd_probe(adev, vf); in amdgpu_amdkfd_device_probe() 152 if (!adev->kfd.init_complete || adev->kfd.client.dev) in amdgpu_amdkfd_drm_client_create() 155 ret = drm_client_init(&adev->ddev, &adev->kfd.client, "kfd", in amdgpu_amdkfd_drm_client_create() 226 adev->kfd.init_complete = kgd2kfd_device_init(adev->kfd.dev, in amdgpu_amdkfd_device_init() 247 if (adev->kfd.dev) in amdgpu_amdkfd_interrupt() 253 if (adev->kfd.dev) in amdgpu_amdkfd_suspend() 261 if (adev->kfd.dev) in amdgpu_amdkfd_resume() 272 if (adev->kfd.dev) in amdgpu_amdkfd_pre_reset() 282 if (adev->kfd.dev) in amdgpu_amdkfd_post_reset() 456 if (adev->gmc.real_vram_size == adev->gmc.visible_vram_size) in amdgpu_amdkfd_get_local_mem_info() [all …]
|
| H A D | mxgpu_nv.c | 73 xgpu_nv_mailbox_send_ack(adev); in xgpu_nv_mailbox_rcv_msg() 155 r = xgpu_nv_poll_ack(adev); in xgpu_nv_mailbox_trans_msg() 372 &adev->virt.flr_work), in xgpu_nv_mailbox_rcv_irq() 412 r = amdgpu_irq_add_id(adev, SOC15_IH_CLIENTID_BIF, 135, &adev->virt.rcv_irq); in xgpu_nv_mailbox_add_irq_id() 416 r = amdgpu_irq_add_id(adev, SOC15_IH_CLIENTID_BIF, 138, &adev->virt.ack_irq); in xgpu_nv_mailbox_add_irq_id() 418 amdgpu_irq_put(adev, &adev->virt.rcv_irq, 0); in xgpu_nv_mailbox_add_irq_id() 429 r = amdgpu_irq_get(adev, &adev->virt.rcv_irq, 0); in xgpu_nv_mailbox_get_irq() 432 r = amdgpu_irq_get(adev, &adev->virt.ack_irq, 0); in xgpu_nv_mailbox_get_irq() 434 amdgpu_irq_put(adev, &adev->virt.rcv_irq, 0); in xgpu_nv_mailbox_get_irq() 445 amdgpu_irq_put(adev, &adev->virt.ack_irq, 0); in xgpu_nv_mailbox_put_irq() [all …]
|
| H A D | vi.c | 1460 struct amdgpu_device *adev = ip_block->adev; in vi_common_early_init() local 1480 adev->rev_id = vi_get_rev_id(adev); in vi_common_early_init() 1507 adev->external_rev_id = adev->rev_id + 0x3c; in vi_common_early_init() 1524 adev->external_rev_id = adev->rev_id + 0x14; in vi_common_early_init() 1644 adev->external_rev_id = adev->rev_id + 0x1; in vi_common_early_init() 1684 struct amdgpu_device *adev = ip_block->adev; in vi_common_late_init() local 1694 struct amdgpu_device *adev = ip_block->adev; in vi_common_sw_init() local 1704 struct amdgpu_device *adev = ip_block->adev; in vi_common_hw_init() local 1718 struct amdgpu_device *adev = ip_block->adev; in vi_common_hw_fini() local 1951 struct amdgpu_device *adev = ip_block->adev; in vi_common_set_clockgating_state() local [all …]
|
| H A D | amdgpu_umsch_mm.c | 53 struct amdgpu_device *adev = ring->adev; in amdgpu_umsch_mm_query_fence() local 69 struct amdgpu_device *adev = ring->adev; in umsch_mm_ring_set_wptr() local 80 struct amdgpu_device *adev = ring->adev; in umsch_mm_ring_get_rptr() local 88 struct amdgpu_device *adev = ring->adev; in umsch_mm_ring_get_wptr() local 122 struct amdgpu_device *adev = umsch->ring.adev; in amdgpu_umsch_mm_init_microcode() local 180 struct amdgpu_device *adev = umsch->ring.adev; in amdgpu_umsch_mm_allocate_ucode_buffer() local 316 struct amdgpu_device *adev = ip_block->adev; in umsch_mm_early_init() local 335 struct amdgpu_device *adev = ip_block->adev; in umsch_mm_late_init() local 337 if (amdgpu_in_reset(adev) || adev->in_s0ix || adev->in_suspend) in umsch_mm_late_init() 345 struct amdgpu_device *adev = ip_block->adev; in umsch_mm_sw_init() local [all …]
|
| H A D | amdgpu_irq.c | 170 ret = amdgpu_ih_process(adev, &adev->irq.ih); in amdgpu_irq_handler() 191 amdgpu_ih_process(adev, &adev->irq.ih1); in amdgpu_irq_handle_ih1() 206 amdgpu_ih_process(adev, &adev->irq.ih2); in amdgpu_irq_handle_ih2() 221 amdgpu_ih_process(adev, &adev->irq.ih_soft); in amdgpu_irq_handle_ih_soft() 315 adev->irq.irq = irq; in amdgpu_irq_init() 332 free_irq(adev->irq.irq, adev_to_drm(adev)); in amdgpu_irq_fini_hw() 338 amdgpu_ih_ring_fini(adev, &adev->irq.ih_soft); in amdgpu_irq_fini_hw() 339 amdgpu_ih_ring_fini(adev, &adev->irq.ih); in amdgpu_irq_fini_hw() 340 amdgpu_ih_ring_fini(adev, &adev->irq.ih1); in amdgpu_irq_fini_hw() 341 amdgpu_ih_ring_fini(adev, &adev->irq.ih2); in amdgpu_irq_fini_hw() [all …]
|
| H A D | amdgpu_umc.c | 109 if (adev->umc.ras && adev->umc.ras->ras_block.hw_ops && in amdgpu_umc_handle_bad_pages() 113 if (adev->umc.ras && adev->umc.ras->ras_block.hw_ops && in amdgpu_umc_handle_bad_pages() 259 if (adev->virt.ops && adev->virt.ops->ras_poison_handler) in amdgpu_umc_pasid_poison_handler() 260 adev->virt.ops->ras_poison_handler(adev, block); in amdgpu_umc_pasid_poison_handler() 326 r = amdgpu_irq_get(adev, &adev->gmc.ecc_irq, 0); in amdgpu_umc_ras_late_init() 334 adev->umc.ras->err_cnt_init(adev); in amdgpu_umc_ras_late_init() 407 adev->umc.active_mask, adev->umc.umc_inst_num); in amdgpu_umc_loop_all_aid() 409 adev->umc.node_inst_num * adev->umc.umc_inst_num) { in amdgpu_umc_loop_all_aid() 467 return adev->umc.ras->update_ecc_status(adev, in amdgpu_umc_update_ecc_status() 501 if (adev->umc.ras && adev->umc.ras->convert_ras_err_addr) in amdgpu_umc_pages_in_a_row() [all …]
|
| /linux-6.15/sound/soc/intel/avs/ |
| H A D | loader.c | 171 ret, reg, snd_hdac_adsp_readl(adev, AVS_FW_REG_ERROR(adev))); in avs_cldma_load_basefw() 185 ret, reg, snd_hdac_adsp_readl(adev, AVS_FW_REG_ERROR(adev))); in avs_cldma_load_basefw() 350 snd_hdac_adsp_readl(adev, AVS_FW_REG_STATUS(adev)), in avs_imr_load_basefw() 351 snd_hdac_adsp_readl(adev, AVS_FW_REG_ERROR(adev))); in avs_imr_load_basefw() 592 snd_hdac_adsp_readl(adev, AVS_FW_REG_STATUS(adev)), in avs_dsp_load_basefw() 593 snd_hdac_adsp_readl(adev, AVS_FW_REG_ERROR(adev))); in avs_dsp_load_basefw() 688 ret = avs_ipc_get_hw_config(adev, &adev->hw_cfg); in avs_dsp_alloc_resources() 692 ret = avs_ipc_get_fw_config(adev, &adev->fw_cfg); in avs_dsp_alloc_resources() 696 adev->core_refs = devm_kcalloc(adev->dev, adev->hw_cfg.dsp_cores, in avs_dsp_alloc_resources() 698 adev->lib_names = devm_kcalloc(adev->dev, adev->fw_cfg.max_libs_count, in avs_dsp_alloc_resources() [all …]
|