Searched refs:hv_vcpu (Results 1 – 6 of 6) sorted by relevance
69 struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(vcpu); in to_hv_synic() local71 return &hv_vcpu->synic; in to_hv_synic()76 struct kvm_vcpu_hv *hv_vcpu = container_of(synic, struct kvm_vcpu_hv, synic); in hv_synic_to_vcpu() local78 return hv_vcpu->vcpu; in hv_synic_to_vcpu()88 struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(vcpu); in kvm_hv_get_vpindex() local90 return hv_vcpu ? hv_vcpu->vp_index : vcpu->vcpu_idx; in kvm_hv_get_vpindex()121 struct kvm_vcpu_hv *hv_vcpu; in hv_stimer_to_vcpu() local123 hv_vcpu = container_of(stimer - stimer->index, struct kvm_vcpu_hv, in hv_stimer_to_vcpu()125 return hv_vcpu->vcpu; in hv_stimer_to_vcpu()130 struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(vcpu); in kvm_hv_has_stimer_pending() local[all …]
223 struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(vcpu); in kvm_hv_notify_acked_sint() local230 for (idx = 0; idx < ARRAY_SIZE(hv_vcpu->stimer); idx++) { in kvm_hv_notify_acked_sint()231 stimer = &hv_vcpu->stimer[idx]; in kvm_hv_notify_acked_sint()248 struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(vcpu); in synic_exit() local250 hv_vcpu->exit.type = KVM_EXIT_HYPERV_SYNIC; in synic_exit()251 hv_vcpu->exit.u.synic.msr = msr; in synic_exit()252 hv_vcpu->exit.u.synic.control = synic->control; in synic_exit()253 hv_vcpu->exit.u.synic.evt_page = synic->evt_page; in synic_exit()254 hv_vcpu->exit.u.synic.msg_page = synic->msg_page; in synic_exit()330 struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(vcpu); in kvm_hv_is_syndbg_enabled() local[all …]
10635 struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(vcpu); in vcpu_enter_guest() local10638 vcpu->run->hyperv = hv_vcpu->exit; in vcpu_enter_guest()
18 struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(vcpu); in nested_svm_hv_update_vm_vp_ids() local20 if (!hv_vcpu) in nested_svm_hv_update_vm_vp_ids()23 hv_vcpu->nested.pa_page_gpa = hve->partition_assist_page; in nested_svm_hv_update_vm_vp_ids()24 hv_vcpu->nested.vm_id = hve->hv_vm_id; in nested_svm_hv_update_vm_vp_ids()25 hv_vcpu->nested.vp_id = hve->hv_vp_id; in nested_svm_hv_update_vm_vp_ids()32 struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(vcpu); in nested_svm_l2_tlb_flush_enabled() local34 if (!hv_vcpu) in nested_svm_l2_tlb_flush_enabled()40 return hv_vcpu->vp_assist_page.nested_control.features.directhypercall; in nested_svm_l2_tlb_flush_enabled()
431 struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(vcpu); in nested_get_evmptr() local436 if (unlikely(!hv_vcpu->vp_assist_page.enlighten_vmentry)) in nested_get_evmptr()439 return hv_vcpu->vp_assist_page.current_nested_vmcs; in nested_get_evmptr()509 struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(vcpu); in evmcs_has_perf_global_ctrl() local518 if (WARN_ON_ONCE(!hv_vcpu)) in evmcs_has_perf_global_ctrl()521 return hv_vcpu->cpuid_cache.nested_ebx & HV_X64_NESTED_EVMCS1_PERF_GLOBAL_CTRL; in evmcs_has_perf_global_ctrl()660 struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(vcpu); in nested_evmcs_l2_tlb_flush_enabled() local664 if (!hv_vcpu || !evmcs) in nested_evmcs_l2_tlb_flush_enabled()670 return hv_vcpu->vp_assist_page.nested_control.features.directhypercall; in nested_evmcs_l2_tlb_flush_enabled()
229 struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(vcpu); in nested_release_evmcs() local239 if (hv_vcpu) { in nested_release_evmcs()240 hv_vcpu->nested.pa_page_gpa = INVALID_GPA; in nested_release_evmcs()241 hv_vcpu->nested.vm_id = 0; in nested_release_evmcs()242 hv_vcpu->nested.vp_id = 0; in nested_release_evmcs()1583 struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(&vmx->vcpu); in copy_enlightened_to_vmcs12() local1591 hv_vcpu->nested.pa_page_gpa = evmcs->partition_assist_page; in copy_enlightened_to_vmcs12()1592 hv_vcpu->nested.vm_id = evmcs->hv_vm_id; in copy_enlightened_to_vmcs12()1593 hv_vcpu->nested.vp_id = evmcs->hv_vp_id; in copy_enlightened_to_vmcs12()