| /linux-6.15/drivers/gpu/drm/amd/amdgpu/ |
| H A D | vcn_v5_0_1.c | 378 bool indirect) in vcn_v5_0_1_mc_resume_dpg_mode() argument 390 if (!indirect) { in vcn_v5_0_1_mc_resume_dpg_mode() 394 inst_idx].tmr_mc_addr_lo), 0, indirect); in vcn_v5_0_1_mc_resume_dpg_mode() 398 inst_idx].tmr_mc_addr_hi), 0, indirect); in vcn_v5_0_1_mc_resume_dpg_mode() 423 if (!indirect) in vcn_v5_0_1_mc_resume_dpg_mode() 431 if (!indirect) { in vcn_v5_0_1_mc_resume_dpg_mode() 455 AMDGPU_VCN_STACK_SIZE), 0, indirect); in vcn_v5_0_1_mc_resume_dpg_mode() 459 AMDGPU_VCN_STACK_SIZE), 0, indirect); in vcn_v5_0_1_mc_resume_dpg_mode() 560 bool indirect) in vcn_v5_0_1_start_dpg_mode() argument 582 if (indirect) { in vcn_v5_0_1_start_dpg_mode() [all …]
|
| H A D | vcn_v4_0_5.c | 460 bool indirect) in vcn_v4_0_5_mc_resume_dpg_mode() argument 472 if (!indirect) { in vcn_v4_0_5_mc_resume_dpg_mode() 476 0, indirect); in vcn_v4_0_5_mc_resume_dpg_mode() 480 0, indirect); in vcn_v4_0_5_mc_resume_dpg_mode() 505 if (!indirect) in vcn_v4_0_5_mc_resume_dpg_mode() 513 if (!indirect) { in vcn_v4_0_5_mc_resume_dpg_mode() 538 0, indirect); in vcn_v4_0_5_mc_resume_dpg_mode() 542 0, indirect); in vcn_v4_0_5_mc_resume_dpg_mode() 910 bool indirect) in vcn_v4_0_5_start_dpg_mode() argument 927 if (indirect) in vcn_v4_0_5_start_dpg_mode() [all …]
|
| H A D | vcn_v5_0_0.c | 438 bool indirect) in vcn_v5_0_0_mc_resume_dpg_mode() argument 450 if (!indirect) { in vcn_v5_0_0_mc_resume_dpg_mode() 481 if (!indirect) in vcn_v5_0_0_mc_resume_dpg_mode() 489 if (!indirect) { in vcn_v5_0_0_mc_resume_dpg_mode() 537 adev->gfx.config.gb_addr_config, 0, indirect); in vcn_v5_0_0_mc_resume_dpg_mode() 681 uint8_t indirect) 708 bool indirect) in vcn_v5_0_0_start_dpg_mode() argument 726 if (indirect) in vcn_v5_0_0_start_dpg_mode() 751 vcn_v5_0_0_mc_resume_dpg_mode(vinst, indirect); in vcn_v5_0_0_start_dpg_mode() 766 UVD_MASTINT_EN__VCPU_EN_MASK, 0, indirect); in vcn_v5_0_0_start_dpg_mode() [all …]
|
| H A D | vcn_v4_0_3.c | 518 bool indirect) in vcn_v4_0_3_mc_resume_dpg_mode() argument 530 if (!indirect) { in vcn_v4_0_3_mc_resume_dpg_mode() 563 if (!indirect) in vcn_v4_0_3_mc_resume_dpg_mode() 571 if (!indirect) { in vcn_v4_0_3_mc_resume_dpg_mode() 730 uint8_t indirect) in vcn_v4_0_3_disable_clock_gating_dpg_mode() argument 833 bool indirect) in vcn_v4_0_3_start_dpg_mode() argument 853 if (indirect) { in vcn_v4_0_3_start_dpg_mode() 933 if (indirect) in vcn_v4_0_3_start_dpg_mode() 2062 tmp, 0, indirect); in vcn_v4_0_3_enable_ras() 2067 tmp, 0, indirect); in vcn_v4_0_3_enable_ras() [all …]
|
| H A D | vcn_v2_5.c | 664 bool indirect) in vcn_v2_5_mc_resume_dpg_mode() argument 673 if (!indirect) { in vcn_v2_5_mc_resume_dpg_mode() 704 if (!indirect) in vcn_v2_5_mc_resume_dpg_mode() 712 if (!indirect) { in vcn_v2_5_mc_resume_dpg_mode() 879 uint8_t sram_sel, uint8_t indirect) in vcn_v2_5_clock_gating_dpg_mode() argument 990 bool indirect) in vcn_v2_6_enable_ras() argument 1005 tmp, 0, indirect); in vcn_v2_6_enable_ras() 1010 tmp, 0, indirect); in vcn_v2_6_enable_ras() 1015 tmp, 0, indirect); in vcn_v2_6_enable_ras() 1035 if (indirect) in vcn_v2_5_start_dpg_mode() [all …]
|
| H A D | vcn_v4_0.c | 521 bool indirect) in vcn_v4_0_mc_resume_dpg_mode() argument 532 if (!indirect) { in vcn_v4_0_mc_resume_dpg_mode() 563 if (!indirect) in vcn_v4_0_mc_resume_dpg_mode() 571 if (!indirect) { in vcn_v4_0_mc_resume_dpg_mode() 867 uint8_t indirect) in vcn_v4_0_disable_clock_gating_dpg_mode() argument 976 bool indirect) in vcn_v4_0_enable_ras() argument 991 tmp, 0, indirect); in vcn_v4_0_enable_ras() 996 tmp, 0, indirect); in vcn_v4_0_enable_ras() 1024 if (indirect) in vcn_v4_0_start_dpg_mode() 1088 vcn_v4_0_enable_ras(vinst, indirect); in vcn_v4_0_start_dpg_mode() [all …]
|
| H A D | jpeg_v5_0_0.c | 304 int inst_idx, uint8_t indirect) in jpeg_engine_5_0_0_dpg_clock_gating_mode() argument 317 if (indirect) { in jpeg_engine_5_0_0_dpg_clock_gating_mode() 318 ADD_SOC24_JPEG_TO_DPG_SRAM(inst_idx, vcnipJPEG_CGC_CTRL, data, indirect); in jpeg_engine_5_0_0_dpg_clock_gating_mode() 353 if (indirect) in jpeg_v5_0_0_start_dpg_mode() 357 jpeg_engine_5_0_0_dpg_clock_gating_mode(adev, inst_idx, indirect); in jpeg_v5_0_0_start_dpg_mode() 360 if (indirect) in jpeg_v5_0_0_start_dpg_mode() 362 adev->gfx.config.gb_addr_config, indirect); in jpeg_v5_0_0_start_dpg_mode() 368 if (indirect) in jpeg_v5_0_0_start_dpg_mode() 370 JPEG_SYS_INT_EN__DJRBC0_MASK, indirect); in jpeg_v5_0_0_start_dpg_mode() 375 if (indirect) { in jpeg_v5_0_0_start_dpg_mode() [all …]
|
| H A D | vcn_v2_0.c | 434 bool indirect) in vcn_v2_0_mc_resume_dpg_mode() argument 442 if (!indirect) { in vcn_v2_0_mc_resume_dpg_mode() 473 if (!indirect) in vcn_v2_0_mc_resume_dpg_mode() 481 if (!indirect) { in vcn_v2_0_mc_resume_dpg_mode() 645 uint8_t sram_sel, uint8_t indirect) in vcn_v2_0_clock_gating_dpg_mode() argument 863 if (indirect) in vcn_v2_0_start_dpg_mode() 878 UVD, 0, mmUVD_MASTINT_EN), 0, 0, indirect); in vcn_v2_0_start_dpg_mode() 890 UVD, 0, mmUVD_LMI_CTRL), tmp, 0, indirect); in vcn_v2_0_start_dpg_mode() 925 UVD, 0, mmUVD_SOFT_RESET), 0, 0, indirect); in vcn_v2_0_start_dpg_mode() 935 UVD_MASTINT_EN__VCPU_EN_MASK, 0, indirect); in vcn_v2_0_start_dpg_mode() [all …]
|
| H A D | vcn_v3_0.c | 570 bool indirect) in vcn_v3_0_mc_resume_dpg_mode() argument 579 if (!indirect) { in vcn_v3_0_mc_resume_dpg_mode() 607 AMDGPU_UVD_FIRMWARE_OFFSET >> 3, 0, indirect); in vcn_v3_0_mc_resume_dpg_mode() 610 if (!indirect) in vcn_v3_0_mc_resume_dpg_mode() 618 if (!indirect) { in vcn_v3_0_mc_resume_dpg_mode() 907 uint8_t indirect) in vcn_v3_0_clock_gating_dpg_mode() argument 1042 if (indirect) in vcn_v3_0_start_dpg_mode() 1095 vcn_v3_0_mc_resume_dpg_mode(vinst, indirect); in vcn_v3_0_start_dpg_mode() 1104 VCN, inst_idx, mmUVD_LMI_CTRL2), 0, 0, indirect); in vcn_v3_0_start_dpg_mode() 1118 UVD_MASTINT_EN__VCPU_EN_MASK, 0, indirect); in vcn_v3_0_start_dpg_mode() [all …]
|
| H A D | amdgpu_jpeg.h | 36 #define WREG32_SOC15_JPEG_DPG_MODE(inst_idx, offset, value, indirect) \ argument 38 if (!indirect) { \ 46 indirect << UVD_DPG_LMA_CTL__SRAM_SEL__SHIFT)); \ 64 #define WREG32_SOC24_JPEG_DPG_MODE(inst_idx, offset, value, indirect) \ argument 75 indirect << UVD_DPG_LMA_CTL__SRAM_SEL__SHIFT)); \ 89 #define ADD_SOC24_JPEG_TO_DPG_SRAM(inst_idx, offset, value, indirect) \ argument
|
| H A D | jpeg_v4_0_5.c | 355 int inst_idx, uint8_t indirect) in jpeg_engine_4_0_5_dpg_clock_gating_mode() argument 366 WREG32_SOC15_JPEG_DPG_MODE(inst_idx, regJPEG_CGC_CTRL_INTERNAL_OFFSET, data, indirect); in jpeg_engine_4_0_5_dpg_clock_gating_mode() 370 data, indirect); in jpeg_engine_4_0_5_dpg_clock_gating_mode() 420 static void jpeg_v4_0_5_start_dpg_mode(struct amdgpu_device *adev, int inst_idx, bool indirect) in jpeg_v4_0_5_start_dpg_mode() argument 443 if (indirect) in jpeg_v4_0_5_start_dpg_mode() 447 jpeg_engine_4_0_5_dpg_clock_gating_mode(adev, inst_idx, indirect); in jpeg_v4_0_5_start_dpg_mode() 451 adev->gfx.config.gb_addr_config, indirect); in jpeg_v4_0_5_start_dpg_mode() 454 JPEG_SYS_INT_EN__DJRBC_MASK, indirect); in jpeg_v4_0_5_start_dpg_mode() 457 WREG32_SOC15_JPEG_DPG_MODE(inst_idx, regUVD_NO_OP_INTERNAL_OFFSET, 0, indirect); in jpeg_v4_0_5_start_dpg_mode() 459 if (indirect) in jpeg_v4_0_5_start_dpg_mode()
|
| /linux-6.15/arch/x86/kernel/ |
| H A D | ksysfs.c | 95 struct setup_indirect *indirect; in get_setup_data_size() local 116 if (indirect->type != SETUP_INDIRECT) in get_setup_data_size() 117 *size = indirect->len; in get_setup_data_size() 138 struct setup_indirect *indirect; in type_show() local 162 indirect = (struct setup_indirect *)data->data; in type_show() 164 ret = sprintf(buf, "0x%x\n", indirect->type); in type_show() 179 struct setup_indirect *indirect; in setup_data_data_read() local 203 indirect = (struct setup_indirect *)data->data; in setup_data_data_read() 205 if (indirect->type != SETUP_INDIRECT) { in setup_data_data_read() 206 paddr = indirect->addr; in setup_data_data_read() [all …]
|
| H A D | kdebugfs.c | 91 struct setup_indirect *indirect; in create_setup_data_nodes() local 129 indirect = (struct setup_indirect *)data->data; in create_setup_data_nodes() 131 if (indirect->type != SETUP_INDIRECT) { in create_setup_data_nodes() 132 node->paddr = indirect->addr; in create_setup_data_nodes() 133 node->type = indirect->type; in create_setup_data_nodes() 134 node->len = indirect->len; in create_setup_data_nodes()
|
| /linux-6.15/Documentation/filesystems/ext4/ |
| H A D | blockmap.rst | 16 | 13 | Double-indirect block: (file blocks ``$block_size``/4 + 12 to (``$block_siz… 21 | | | 0 to (``$block_size`` / 4) | Map to (``$block_size`` / 4) indirect bl… 30 | 14 | Triple-indirect block: (file blocks (``$block_size`` / 4) ^ 2 + (``$block_s… 35 … | 0 to (``$block_size`` / 4) | Map to (``$block_size`` / 4) double indirect blocks (1024 if 4… 40 … | | 0 to (``$block_size`` / 4) | Map to (``$block_size`` / 4) indirect blocks (1024 if 4…
|
| /linux-6.15/drivers/net/ethernet/intel/idpf/ |
| H A D | idpf_controlq.c | 80 desc->params.indirect.addr_high = in idpf_ctlq_init_rxq_bufs() 82 desc->params.indirect.addr_low = in idpf_ctlq_init_rxq_bufs() 84 desc->params.indirect.param0 = 0; in idpf_ctlq_init_rxq_bufs() 85 desc->params.indirect.sw_cookie = 0; in idpf_ctlq_init_rxq_bufs() 86 desc->params.indirect.v_flags = 0; in idpf_ctlq_init_rxq_bufs() 307 desc->params.indirect.addr_high = in idpf_ctlq_send() 309 desc->params.indirect.addr_low = in idpf_ctlq_send() 312 memcpy(&desc->params, msg->ctx.indirect.context, in idpf_ctlq_send() 500 desc->params.indirect.addr_high = in idpf_ctlq_post_rx_buffs() 502 desc->params.indirect.addr_low = in idpf_ctlq_post_rx_buffs() [all …]
|
| /linux-6.15/arch/arm64/kvm/hyp/ |
| H A D | hyp-entry.S | 216 .macro hyp_ventry indirect, spectrev2 226 .if \indirect != 0 249 .macro generate_vectors indirect, spectrev2 252 hyp_ventry \indirect, \spectrev2 259 generate_vectors indirect = 0, spectrev2 = 1 // HYP_VECTOR_SPECTRE_DIRECT 260 generate_vectors indirect = 1, spectrev2 = 0 // HYP_VECTOR_INDIRECT 261 generate_vectors indirect = 1, spectrev2 = 1 // HYP_VECTOR_SPECTRE_INDIRECT
|
| /linux-6.15/drivers/block/xen-blkback/ |
| H A D | blkback.c | 1105 dst->u.indirect.indirect_op = src->u.indirect.indirect_op; in blkif_get_x86_32_req() 1106 dst->u.indirect.nr_segments = in blkif_get_x86_32_req() 1108 dst->u.indirect.handle = src->u.indirect.handle; in blkif_get_x86_32_req() 1109 dst->u.indirect.id = src->u.indirect.id; in blkif_get_x86_32_req() 1110 dst->u.indirect.sector_number = src->u.indirect.sector_number; in blkif_get_x86_32_req() 1114 dst->u.indirect.indirect_grefs[i] = in blkif_get_x86_32_req() 1158 dst->u.indirect.indirect_op = src->u.indirect.indirect_op; in blkif_get_x86_64_req() 1159 dst->u.indirect.nr_segments = in blkif_get_x86_64_req() 1161 dst->u.indirect.handle = src->u.indirect.handle; in blkif_get_x86_64_req() 1162 dst->u.indirect.id = src->u.indirect.id; in blkif_get_x86_64_req() [all …]
|
| /linux-6.15/tools/testing/selftests/bpf/progs/ |
| H A D | map_ptr_kern.c | 50 static inline int check_bpf_map_ptr(struct bpf_map *indirect, in check_bpf_map_ptr() argument 53 VERIFY(indirect->map_type == direct->map_type); in check_bpf_map_ptr() 54 VERIFY(indirect->key_size == direct->key_size); in check_bpf_map_ptr() 55 VERIFY(indirect->value_size == direct->value_size); in check_bpf_map_ptr() 56 VERIFY(indirect->max_entries == direct->max_entries); in check_bpf_map_ptr() 57 VERIFY(indirect->id == direct->id); in check_bpf_map_ptr() 65 VERIFY(check_bpf_map_ptr(indirect, direct)); in check() 66 VERIFY(check_bpf_map_fields(indirect, key_size, value_size, in check() 71 static inline int check_default(struct bpf_map *indirect, in check_default() argument 74 VERIFY(check(indirect, direct, sizeof(__u32), sizeof(__u32), in check_default() [all …]
|
| /linux-6.15/Documentation/admin-guide/hw-vuln/ |
| H A D | spectre.rst | 62 execution of indirect branches to leak privileged memory. 93 execution of indirect branches :ref:`[3] <spec_ref3>`. The indirect 104 buffer of a CPU used for predicting indirect branch addresses. Such 105 poisoning could be done by indirect branching into existing code, 106 with the address offset of the indirect branch under the attacker's 253 indirect branch speculation. This comes with a performance cost 291 guests from affecting indirect branching in the host kernel. 536 can be compiled with return trampolines for indirect branches. 542 can disable indirect branch speculation via prctl() (See 548 Restricting indirect branch speculation on a user program will [all …]
|
| H A D | indirect-target-selection.rst | 8 of indirect branches and RETs located in the lower half of a cacheline. 20 - **Indirect Branch Prediction Barrier (IBPB)**: After an IBPB, indirect 57 As only the indirect branches and RETs that have their last byte of instruction 59 the mitigation is to not allow indirect branches in the lower half. 63 added ITS-safe thunks. These safe thunks consists of indirect branch in the 66 indirect branch. 75 Note, for simplicity, indirect branches in eBPF programs are always replaced 82 thunks. But, RETs significantly outnumber indirect branches, and any benefit 88 Retpoline sequence also mitigates ITS-unsafe indirect branches. For this 119 indirect branches. [all …]
|
| /linux-6.15/drivers/net/can/sja1000/ |
| H A D | sja1000_isa.c | 37 static int indirect[MAXDEV] = {[0 ... (MAXDEV - 1)] = -1}; variable 46 module_param_hw_array(indirect, int, ioport, NULL, 0444); 47 MODULE_PARM_DESC(indirect, "Indirect access via address and data port"); 139 if (indirect[idx] > 0 || in sja1000_isa_probe() 140 (indirect[idx] == -1 && indirect[0] > 0)) in sja1000_isa_probe()
|
| /linux-6.15/drivers/net/can/cc770/ |
| H A D | cc770_isa.c | 75 static int indirect[MAXDEV] = {[0 ... (MAXDEV - 1)] = -1}; variable 83 module_param_hw_array(indirect, int, ioport, NULL, 0444); 84 MODULE_PARM_DESC(indirect, "Indirect access via address and data port"); 184 if (indirect[idx] > 0 || in cc770_isa_probe() 185 (indirect[idx] == -1 && indirect[0] > 0)) in cc770_isa_probe()
|
| /linux-6.15/arch/m68k/math-emu/ |
| H A D | fp_decode.h | 196 | test if %pc is the base register for the indirect addr mode 220 | addressing mode: address register indirect 244 | addressing mode: address register indirect with postincrement 263 | addressing mode: address register indirect with predecrement 289 | addressing mode: address register/programm counter indirect 331 | all other indirect addressing modes will finally end up here 345 | addressing mode: address register/programm counter indirect 355 3: | addressing mode: address register/programm counter memory indirect
|
| H A D | fp_move.S | 135 | addressing mode: address register indirect 140 | addressing mode: address register indirect with postincrement 145 | addressing mode: address register indirect with predecrement 150 | addressing mode: address register indirect with 16bit displacement
|
| /linux-6.15/Documentation/virt/ |
| H A D | paravirt_ops.rst | 23 - simple indirect call 25 known that the overhead of indirect call isn't very important. 27 - indirect call which allows optimization with binary patch
|