Home
last modified time | relevance | path

Searched refs:guest_context (Results 1 – 18 of 18) sorted by relevance

/linux-6.15/arch/riscv/kvm/
H A Dvcpu_exit.c70 old_hstatus = csr_swap(CSR_HSTATUS, vcpu->arch.guest_context.hstatus); in kvm_riscv_vcpu_unpriv_read()
142 if (vcpu->arch.guest_context.sstatus & SR_SPP) in kvm_riscv_vcpu_trap_redirect()
162 vcpu->arch.guest_context.sepc = csr_read(CSR_VSTVEC); in kvm_riscv_vcpu_trap_redirect()
165 vcpu->arch.guest_context.sstatus |= SR_SPP; in kvm_riscv_vcpu_trap_redirect()
172 if (vcpu->arch.guest_context.hstatus & HSTATUS_SPV) { in vcpu_redirect()
225 if (vcpu->arch.guest_context.hstatus & HSTATUS_SPV) in kvm_riscv_vcpu_exit()
231 if (vcpu->arch.guest_context.hstatus & HSTATUS_SPV) in kvm_riscv_vcpu_exit()
235 if (vcpu->arch.guest_context.hstatus & HSTATUS_SPV) in kvm_riscv_vcpu_exit()
250 vcpu->arch.guest_context.sepc, in kvm_riscv_vcpu_exit()
251 vcpu->arch.guest_context.sstatus, in kvm_riscv_vcpu_exit()
[all …]
H A Dvcpu_insn.c162 utrap.sepc = vcpu->arch.guest_context.sepc; in truly_illegal_insn()
178 utrap.sepc = vcpu->arch.guest_context.sepc; in truly_virtual_insn()
264 SET_RD(insn, &vcpu->arch.guest_context, in kvm_riscv_vcpu_csr_return()
268 vcpu->arch.guest_context.sepc += INSN_LEN(insn); in kvm_riscv_vcpu_csr_return()
418 vcpu->arch.guest_context.sepc += INSN_LEN(insn); in system_opcode_insn()
447 ct = &vcpu->arch.guest_context; in kvm_riscv_vcpu_virtual_insn()
641 data = GET_RS2(insn, &vcpu->arch.guest_context); in kvm_riscv_vcpu_mmio_store()
755 SET_RD(insn, &vcpu->arch.guest_context, in kvm_riscv_vcpu_mmio_return()
760 SET_RD(insn, &vcpu->arch.guest_context, in kvm_riscv_vcpu_mmio_return()
765 SET_RD(insn, &vcpu->arch.guest_context, in kvm_riscv_vcpu_mmio_return()
[all …]
H A Dvcpu_sbi_replace.c20 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_sbi_ext_time_handler()
51 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_sbi_ext_ipi_handler()
95 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_sbi_ext_rfence_handler()
150 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_sbi_ext_srst_handler()
192 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_sbi_ext_dbcn_handler()
H A Dvcpu_sbi_hsm.c19 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_sbi_hsm_vcpu_start()
76 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_sbi_hsm_vcpu_get_status()
95 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_sbi_ext_hsm_handler()
H A Dvcpu_fp.c19 struct kvm_cpu_context *cntx = &vcpu->arch.guest_context; in kvm_riscv_vcpu_fp_reset()
81 struct kvm_cpu_context *cntx = &vcpu->arch.guest_context; in kvm_riscv_vcpu_get_reg_fp()
126 struct kvm_cpu_context *cntx = &vcpu->arch.guest_context; in kvm_riscv_vcpu_set_reg_fp()
H A Dvcpu_vector.c22 struct kvm_cpu_context *cntx = &vcpu->arch.guest_context; in kvm_riscv_vcpu_vector_reset()
100 struct kvm_cpu_context *cntx = &vcpu->arch.guest_context; in kvm_riscv_vcpu_vreg_addr()
181 struct kvm_cpu_context *cntx = &vcpu->arch.guest_context; in kvm_riscv_vcpu_set_reg_vector()
H A Dvcpu.c58 struct kvm_cpu_context *cntx = &vcpu->arch.guest_context; in kvm_riscv_reset_vcpu()
233 return (vcpu->arch.guest_context.sstatus & SR_SPP) ? true : false; in kvm_arch_vcpu_in_kernel()
239 return vcpu->arch.guest_context.sepc; in kvm_arch_vcpu_get_ip()
632 kvm_riscv_vcpu_guest_fp_restore(&vcpu->arch.guest_context, in kvm_arch_vcpu_load()
635 kvm_riscv_vcpu_guest_vector_restore(&vcpu->arch.guest_context, in kvm_arch_vcpu_load()
654 kvm_riscv_vcpu_guest_fp_save(&vcpu->arch.guest_context, in kvm_arch_vcpu_put()
659 kvm_riscv_vcpu_guest_vector_save(&vcpu->arch.guest_context, in kvm_arch_vcpu_put()
782 struct kvm_cpu_context *gcntx = &vcpu->arch.guest_context; in kvm_riscv_vcpu_enter_exit()
H A Dvcpu_sbi.c121 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_riscv_vcpu_sbi_forward()
161 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_riscv_vcpu_sbi_return()
173 vcpu->arch.guest_context.sepc += 4; in kvm_riscv_vcpu_sbi_return()
427 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_riscv_vcpu_sbi_ecall()
H A Dvcpu_sbi_sta.c83 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_sbi_sta_steal_time_set_shmem()
131 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_sbi_ext_sta_handler()
H A Dtrace.h25 __entry->pc = vcpu->arch.guest_context.sepc;
H A Dvcpu_sbi_system.c15 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_sbi_ext_susp_handler()
H A Dvcpu_sbi_base.c19 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_sbi_ext_base_handler()
H A Dvcpu_sbi_pmu.c20 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_sbi_ext_pmu_handler()
H A Dvcpu_sbi_v01.c24 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_sbi_ext_v01_handler()
H A Dvcpu_onereg.c388 struct kvm_cpu_context *cntx = &vcpu->arch.guest_context; in kvm_riscv_vcpu_get_reg_core()
421 struct kvm_cpu_context *cntx = &vcpu->arch.guest_context; in kvm_riscv_vcpu_set_reg_core()
942 const struct kvm_cpu_context *cntx = &vcpu->arch.guest_context; in num_fp_f_regs()
971 const struct kvm_cpu_context *cntx = &vcpu->arch.guest_context; in num_fp_d_regs()
1115 const struct kvm_cpu_context *cntx = &vcpu->arch.guest_context; in copy_vector_reg_indices()
H A Daia_imsic.c828 vcpu->arch.guest_context.hstatus &= ~HSTATUS_VGEIN; in kvm_riscv_vcpu_aia_imsic_update()
830 vcpu->arch.guest_context.hstatus |= in kvm_riscv_vcpu_aia_imsic_update()
/linux-6.15/arch/riscv/kernel/
H A Dasm-offsets.c135 OFFSET(KVM_ARCH_GUEST_RA, kvm_vcpu_arch, guest_context.ra); in asm_offsets()
136 OFFSET(KVM_ARCH_GUEST_SP, kvm_vcpu_arch, guest_context.sp); in asm_offsets()
137 OFFSET(KVM_ARCH_GUEST_GP, kvm_vcpu_arch, guest_context.gp); in asm_offsets()
138 OFFSET(KVM_ARCH_GUEST_TP, kvm_vcpu_arch, guest_context.tp); in asm_offsets()
139 OFFSET(KVM_ARCH_GUEST_T0, kvm_vcpu_arch, guest_context.t0); in asm_offsets()
140 OFFSET(KVM_ARCH_GUEST_T1, kvm_vcpu_arch, guest_context.t1); in asm_offsets()
141 OFFSET(KVM_ARCH_GUEST_T2, kvm_vcpu_arch, guest_context.t2); in asm_offsets()
142 OFFSET(KVM_ARCH_GUEST_S0, kvm_vcpu_arch, guest_context.s0); in asm_offsets()
143 OFFSET(KVM_ARCH_GUEST_S1, kvm_vcpu_arch, guest_context.s1); in asm_offsets()
144 OFFSET(KVM_ARCH_GUEST_A0, kvm_vcpu_arch, guest_context.a0); in asm_offsets()
[all …]
/linux-6.15/arch/riscv/include/asm/
H A Dkvm_host.h222 struct kvm_cpu_context guest_context; member