/linux-6.1.9/arch/x86/include/asm/ |
D | virtext.h | 128 rdmsrl(MSR_EFER, efer); in cpu_svm_disable() 129 wrmsrl(MSR_EFER, efer & ~EFER_SVME); in cpu_svm_disable()
|
D | msr-index.h | 15 #define MSR_EFER 0xc0000080 /* extended feature register */ macro
|
/linux-6.1.9/tools/testing/selftests/kvm/lib/x86_64/ |
D | svm.c | 75 efer = rdmsr(MSR_EFER); in generic_svm_setup() 76 wrmsr(MSR_EFER, efer | EFER_SVME); in generic_svm_setup() 90 save->efer = rdmsr(MSR_EFER); in generic_svm_setup()
|
D | vmx.c | 286 vmwrite(HOST_IA32_EFER, rdmsr(MSR_EFER)); in init_vmcs_host_state()
|
/linux-6.1.9/arch/x86/kernel/acpi/ |
D | sleep.c | 83 if (!rdmsr_safe(MSR_EFER, in x86_acpi_suspend_lowlevel() 86 !wrmsr_safe(MSR_EFER, in x86_acpi_suspend_lowlevel()
|
/linux-6.1.9/arch/x86/boot/compressed/ |
D | efi_thunk_64.S | 131 movl $MSR_EFER, %ecx 159 movl $MSR_EFER, %ecx
|
D | head_64.S | 241 movl $MSR_EFER, %ecx 649 movl $MSR_EFER, %ecx
|
/linux-6.1.9/arch/x86/realmode/rm/ |
D | reboot.S | 39 movl $MSR_EFER, %ecx
|
D | wakeup_asm.S | 123 movl $MSR_EFER, %ecx
|
D | trampoline_64.S | 146 movl $MSR_EFER, %ecx
|
/linux-6.1.9/arch/x86/power/ |
D | cpu.c | 116 rdmsrl(MSR_EFER, ctxt->efer); in __save_processor_state() 209 wrmsrl(MSR_EFER, ctxt->efer); in __restore_processor_state()
|
/linux-6.1.9/arch/x86/platform/pvh/ |
D | head.S | 80 mov $MSR_EFER, %ecx
|
/linux-6.1.9/arch/x86/realmode/ |
D | init.c | 148 rdmsrl(MSR_EFER, efer); in setup_real_mode()
|
/linux-6.1.9/arch/x86/kernel/ |
D | head_32.S | 242 movl $MSR_EFER, %ecx
|
D | head_64.S | 295 movl $MSR_EFER, %ecx
|
/linux-6.1.9/arch/x86/kvm/vmx/ |
D | vmx.c | 891 case MSR_EFER: in clear_atomic_switch_msr() 944 case MSR_EFER: in add_atomic_switch_msr() 1034 add_atomic_switch_msr(vmx, MSR_EFER, in update_transition_efer() 1037 clear_atomic_switch_msr(vmx, MSR_EFER); in update_transition_efer() 1041 i = kvm_find_user_return_msr(MSR_EFER); in update_transition_efer() 1045 clear_atomic_switch_msr(vmx, MSR_EFER); in update_transition_efer() 1774 vmx_setup_uret_msr(vmx, MSR_EFER, update_transition_efer(vmx)); in vmx_setup_uret_msrs() 1883 case MSR_EFER: in vmx_get_msr() 2054 case MSR_EFER: in vmx_set_msr() 2988 if (!vmx_find_uret_msr(vmx, MSR_EFER)) in vmx_set_efer() [all …]
|
D | nested.c | 4563 if (vmx->msr_autoload.guest.val[i].index == MSR_EFER) in nested_vmx_get_vmcs01_guest_efer() 4567 efer_msr = vmx_find_uret_msr(vmx, MSR_EFER); in nested_vmx_get_vmcs01_guest_efer()
|
/linux-6.1.9/tools/arch/x86/include/asm/ |
D | msr-index.h | 15 #define MSR_EFER 0xc0000080 /* extended feature register */ macro
|
/linux-6.1.9/arch/x86/kvm/ |
D | emulate.c | 817 ctxt->ops->get_msr(ctxt, MSR_EFER, &efer); in emulator_recalc_and_set_mode() 1543 ctxt->ops->get_msr(ctxt, MSR_EFER, &efer); in get_descriptor_ptr() 1709 ctxt->ops->get_msr(ctxt, MSR_EFER, &efer); in __load_segment_descriptor() 2543 if (ctxt->ops->set_msr(ctxt, MSR_EFER, val & ~EFER_LMA)) in rsm_load_state_64() 2636 ctxt->ops->set_msr(ctxt, MSR_EFER, efer); in em_rsm() 2758 ops->get_msr(ctxt, MSR_EFER, &efer); in em_syscall() 2809 ops->get_msr(ctxt, MSR_EFER, &efer); in em_sysenter() 4259 ctxt->ops->get_msr(ctxt, MSR_EFER, &efer); in check_svme()
|
D | x86.c | 3583 case MSR_EFER: in kvm_set_msr_common() 4083 case MSR_EFER: in kvm_get_msr_common() 12274 rdmsrl_safe(MSR_EFER, &host_efer); in kvm_arch_hardware_setup()
|
/linux-6.1.9/arch/x86/kvm/svm/ |
D | svm.c | 101 { .index = MSR_EFER, .always = false }, 570 rdmsrl(MSR_EFER, efer); in svm_hardware_enable() 587 wrmsrl(MSR_EFER, efer | EFER_SVME); in svm_hardware_enable() 2695 msr_info.index = MSR_EFER; in efer_trap()
|
D | sev.c | 2977 set_msr_interception(vcpu, svm->msrpm, MSR_EFER, 1, 1); in sev_es_init_vmcb()
|