Lines Matching refs:esr
31 static void kvm_handle_guest_serror(struct kvm_vcpu *vcpu, u64 esr) in kvm_handle_guest_serror() argument
33 if (!arm64_is_ras_serror(esr) || arm64_is_fatal_ras_serror(NULL, esr)) in kvm_handle_guest_serror()
116 u64 esr = kvm_vcpu_get_esr(vcpu); in kvm_handle_wfx() local
118 if (esr & ESR_ELx_WFx_ISS_WFE) { in kvm_handle_wfx()
126 if (esr & ESR_ELx_WFx_ISS_WFxT) { in kvm_handle_wfx()
127 if (esr & ESR_ELx_WFx_ISS_RV) { in kvm_handle_wfx()
137 esr &= ~ESR_ELx_WFx_ISS_WFxT; in kvm_handle_wfx()
141 if (esr & ESR_ELx_WFx_ISS_WFE) { in kvm_handle_wfx()
144 if (esr & ESR_ELx_WFx_ISS_WFxT) in kvm_handle_wfx()
169 u64 esr = kvm_vcpu_get_esr(vcpu); in kvm_handle_guest_debug() local
172 run->debug.arch.hsr = lower_32_bits(esr); in kvm_handle_guest_debug()
173 run->debug.arch.hsr_high = upper_32_bits(esr); in kvm_handle_guest_debug()
176 switch (ESR_ELx_EC(esr)) { in kvm_handle_guest_debug()
190 u64 esr = kvm_vcpu_get_esr(vcpu); in kvm_handle_unknown_ec() local
193 esr, esr_get_class_string(esr)); in kvm_handle_unknown_ec()
285 u64 esr = kvm_vcpu_get_esr(vcpu); in kvm_get_exit_handler() local
286 u8 esr_ec = ESR_ELx_EC(esr); in kvm_get_exit_handler()
386 void __noreturn __cold nvhe_hyp_panic_handler(u64 esr, u64 spsr, in nvhe_hyp_panic_handler() argument
397 } else if (ESR_ELx_EC(esr) == ESR_ELx_EC_BRK64 && in nvhe_hyp_panic_handler()
398 (esr & ESR_ELx_BRK64_ISS_COMMENT_MASK) == BUG_BRK_IMM) { in nvhe_hyp_panic_handler()
433 spsr, elr_virt, esr, far, hpfar, par, vcpu); in nvhe_hyp_panic_handler()