Lines Matching refs:guest_context

133 	utrap.sepc = vcpu->arch.guest_context.sepc;  in truly_illegal_insn()
148 vcpu->arch.guest_context.sepc += INSN_LEN(insn); in system_opcode_insn()
164 ct = &vcpu->arch.guest_context; in virtual_inst_fault()
193 struct kvm_cpu_context *ct = &vcpu->arch.guest_context; in emulate_load()
306 struct kvm_cpu_context *ct = &vcpu->arch.guest_context; in emulate_store()
332 data = GET_RS2(insn, &vcpu->arch.guest_context); in emulate_store()
348 data64 = GET_RS2S(insn, &vcpu->arch.guest_context); in emulate_store()
352 data64 = GET_RS2C(insn, &vcpu->arch.guest_context); in emulate_store()
356 data32 = GET_RS2S(insn, &vcpu->arch.guest_context); in emulate_store()
360 data32 = GET_RS2C(insn, &vcpu->arch.guest_context); in emulate_store()
489 old_hstatus = csr_swap(CSR_HSTATUS, vcpu->arch.guest_context.hstatus); in kvm_riscv_vcpu_unpriv_read()
581 if (vcpu->arch.guest_context.sstatus & SR_SPP) in kvm_riscv_vcpu_trap_redirect()
601 vcpu->arch.guest_context.sepc = csr_read(CSR_VSTVEC); in kvm_riscv_vcpu_trap_redirect()
635 SET_RD(insn, &vcpu->arch.guest_context, in kvm_riscv_vcpu_mmio_return()
640 SET_RD(insn, &vcpu->arch.guest_context, in kvm_riscv_vcpu_mmio_return()
645 SET_RD(insn, &vcpu->arch.guest_context, in kvm_riscv_vcpu_mmio_return()
650 SET_RD(insn, &vcpu->arch.guest_context, in kvm_riscv_vcpu_mmio_return()
659 vcpu->arch.guest_context.sepc += vcpu->arch.mmio_decode.insn_len; in kvm_riscv_vcpu_mmio_return()
682 if (vcpu->arch.guest_context.hstatus & HSTATUS_SPV) in kvm_riscv_vcpu_exit()
688 if (vcpu->arch.guest_context.hstatus & HSTATUS_SPV) in kvm_riscv_vcpu_exit()
692 if (vcpu->arch.guest_context.hstatus & HSTATUS_SPV) in kvm_riscv_vcpu_exit()
703 vcpu->arch.guest_context.sepc, in kvm_riscv_vcpu_exit()
704 vcpu->arch.guest_context.sstatus, in kvm_riscv_vcpu_exit()
705 vcpu->arch.guest_context.hstatus); in kvm_riscv_vcpu_exit()