Searched refs:guest_msr (Results 1 – 5 of 5) sorted by relevance
/linux-5.19.10/arch/powerpc/kvm/ |
D | book3s_emulate.c | 124 unsigned long guest_msr = kvmppc_get_msr(vcpu); in kvmppc_emulate_treclaim() local 130 (((guest_msr & MSR_TS_MASK) >> (MSR_TS_S_LG - 1)) in kvmppc_emulate_treclaim() 160 guest_msr &= ~(MSR_TS_MASK); in kvmppc_emulate_treclaim() 161 kvmppc_set_msr(vcpu, guest_msr); in kvmppc_emulate_treclaim() 170 unsigned long guest_msr = kvmppc_get_msr(vcpu); in kvmppc_emulate_trchkpt() local 185 guest_msr &= ~(MSR_TS_MASK); in kvmppc_emulate_trchkpt() 186 guest_msr |= MSR_TS_S; in kvmppc_emulate_trchkpt() 187 kvmppc_set_msr(vcpu, guest_msr); in kvmppc_emulate_trchkpt() 199 unsigned long guest_msr = kvmppc_get_msr(vcpu); in kvmppc_emulate_tabort() local 209 (((guest_msr & MSR_TS_MASK) >> (MSR_TS_S_LG - 1)) in kvmppc_emulate_tabort() [all …]
|
D | book3s_hv_p9_entry.c | 315 unsigned long guest_msr = vcpu->arch.shregs.msr; in load_vcpu_state() local 316 if (MSR_TM_ACTIVE(guest_msr)) { in load_vcpu_state() 317 kvmppc_restore_tm_hv(vcpu, guest_msr, true); in load_vcpu_state() 350 unsigned long guest_msr = vcpu->arch.shregs.msr; in store_vcpu_state() local 351 if (MSR_TM_ACTIVE(guest_msr)) { in store_vcpu_state() 352 kvmppc_save_tm_hv(vcpu, guest_msr, true); in store_vcpu_state()
|
D | book3s_pr.c | 233 ulong guest_msr = kvmppc_get_msr(vcpu); in kvmppc_recalc_shadow_msr() local 234 ulong smsr = guest_msr; in kvmppc_recalc_shadow_msr() 246 smsr |= (guest_msr & vcpu->arch.guest_owned_ext); in kvmppc_recalc_shadow_msr() 257 if (!(guest_msr & MSR_PR)) in kvmppc_recalc_shadow_msr()
|
/linux-5.19.10/arch/powerpc/include/asm/ |
D | asm-prototypes.h | 60 void _kvmppc_restore_tm_pr(struct kvm_vcpu *vcpu, u64 guest_msr); 61 void _kvmppc_save_tm_pr(struct kvm_vcpu *vcpu, u64 guest_msr);
|
/linux-5.19.10/tools/testing/selftests/kvm/x86_64/ |
D | hyperv_features.c | 103 static void guest_msr(struct msr_data *msr) in guest_msr() function 192 vm = vm_create_default(VCPU_ID, 0, guest_msr); in guest_test_msrs_access()
|