Searched refs:MSR_SF (Results 1 – 15 of 15) sorted by relevance
70 #define MSR_SF __MASK(MSR_SF_LG) /* Enable 64 bit mode */ macro75 #define MSR_SF 0 macro131 #define MSR_64BIT MSR_SF137 #define MSR_IDLE (MSR_ME | MSR_SF | MSR_HV)140 #define MSR_IDLE (MSR_ME | MSR_SF | MSR_HV | MSR_LE)
1033 msr_64bit = MSR_SF; in kvmppc_get_ea_indexed()
497 MSR_ILE|MSR_HV|MSR_SF)) == (MSR_DR|MSR_SF)) { in pSeries_system_reset_exception()
300 LOAD_REG_IMMEDIATE(r12, MSR_SF | MSR_LE)
1435 {MSR_SF, "SF"},
238 smsr &= MSR_FE0 | MSR_FE1 | MSR_SF | MSR_SE | MSR_BE | MSR_LE | in kvmppc_recalc_shadow_msr()241 smsr &= MSR_FE0 | MSR_FE1 | MSR_SF | MSR_SE | MSR_BE | MSR_LE; in kvmppc_recalc_shadow_msr()541 !(old_msr & MSR_PR) && !(old_msr & MSR_SF) && (msr & MSR_SF)) { in kvmppc_set_msr_pr()677 if (!(kvmppc_get_msr(vcpu) & MSR_SF)) in kvmppc_visible_gpa()1773 vcpu->arch.intr_msr = MSR_SF; in kvmppc_core_vcpu_create_pr()
122 if (!(kvmppc_get_msr(vcpu) & MSR_SF)) { in kvmppc_critical_section()434 if (!(kvmppc_get_msr(vcpu) & MSR_SF)) in kvmppc_gpa_to_pfn()
460 if (!(kvmppc_get_msr(vcpu) & MSR_SF)) in kvmppc_core_emulate_op_pr()
426 if (!(vcpu->arch.shared->msr & MSR_SF)) { in kvmppc_booke_irqprio_deliver()
172 if (!(kvmppc_get_msr(vcpu) & MSR_SF)) { in kvmppc_kvm_pv()
2834 vcpu->arch.intr_msr = MSR_SF | MSR_ME; in kvmppc_core_vcpu_create_hv()
528 if (mfmsr() & MSR_SF) in calculate_cfg_state()
198 #define CXL_PSL_SR_An_SF MSR_SF /* 64bit */
541 if (mfmsr() & MSR_SF) in attach_afu_directed()
609 sr |= (mfmsr() & MSR_SF) | CXL_PSL_SR_An_HV; in cxl_calculate_sr()