/linux-3.4.99/include/trace/events/ |
D | kvm.h | 221 TP_PROTO(u64 gva, u64 gfn), 223 TP_ARGS(gva, gfn), 226 __field(__u64, gva) 231 __entry->gva = gva; 235 TP_printk("gva = %#llx, gfn = %#llx", __entry->gva, __entry->gfn) 240 TP_PROTO(u64 gva, u64 gfn), 242 TP_ARGS(gva, gfn) 247 TP_PROTO(u64 gva, u64 gfn), 249 TP_ARGS(gva, gfn) 254 TP_PROTO(u64 token, u64 gva), [all …]
|
/linux-3.4.99/arch/x86/kvm/ |
D | x86.h | 76 gva_t gva, gfn_t gfn, unsigned access) in vcpu_cache_mmio_info() argument 78 vcpu->arch.mmio_gva = gva & PAGE_MASK; in vcpu_cache_mmio_info() 87 static inline void vcpu_clear_mmio_info(struct kvm_vcpu *vcpu, gva_t gva) in vcpu_clear_mmio_info() argument 89 if (gva != (~0ul) && vcpu->arch.mmio_gva != (gva & PAGE_MASK)) in vcpu_clear_mmio_info() 95 static inline bool vcpu_match_mmio_gva(struct kvm_vcpu *vcpu, unsigned long gva) in vcpu_match_mmio_gva() argument 97 if (vcpu->arch.mmio_gva && vcpu->arch.mmio_gva == (gva & PAGE_MASK)) in vcpu_match_mmio_gva()
|
D | trace.h | 801 TP_PROTO(gva_t gva, gpa_t gpa, bool write, bool gpa_match), 802 TP_ARGS(gva, gpa, write, gpa_match), 805 __field(gva_t, gva) 812 __entry->gva = gva; 818 TP_printk("gva %#lx gpa %#llx %s %s", __entry->gva, __entry->gpa,
|
D | mmu.c | 1372 static void nonpaging_invlpg(struct kvm_vcpu *vcpu, gva_t gva) in nonpaging_invlpg() argument 2564 static bool handle_abnormal_pfn(struct kvm_vcpu *vcpu, gva_t gva, gfn_t gfn, in handle_abnormal_pfn() argument 2576 vcpu_cache_mmio_info(vcpu, gva, gfn, access); in handle_abnormal_pfn() 2584 gva_t gva, pfn_t *pfn, bool write, bool *writable); 2978 static int nonpaging_page_fault(struct kvm_vcpu *vcpu, gva_t gva, in nonpaging_page_fault() argument 2984 pgprintk("%s: gva %lx error %x\n", __func__, gva, error_code); in nonpaging_page_fault() 2987 return handle_mmio_page_fault(vcpu, gva, error_code, true); in nonpaging_page_fault() 2996 gfn = gva >> PAGE_SHIFT; in nonpaging_page_fault() 2998 return nonpaging_map(vcpu, gva & PAGE_MASK, in nonpaging_page_fault() 3002 static int kvm_arch_setup_async_pf(struct kvm_vcpu *vcpu, gva_t gva, gfn_t gfn) in kvm_arch_setup_async_pf() argument [all …]
|
D | x86.c | 3532 gpa_t kvm_mmu_gva_to_gpa_read(struct kvm_vcpu *vcpu, gva_t gva, in kvm_mmu_gva_to_gpa_read() argument 3536 return vcpu->arch.walk_mmu->gva_to_gpa(vcpu, gva, access, exception); in kvm_mmu_gva_to_gpa_read() 3539 gpa_t kvm_mmu_gva_to_gpa_fetch(struct kvm_vcpu *vcpu, gva_t gva, in kvm_mmu_gva_to_gpa_fetch() argument 3544 return vcpu->arch.walk_mmu->gva_to_gpa(vcpu, gva, access, exception); in kvm_mmu_gva_to_gpa_fetch() 3547 gpa_t kvm_mmu_gva_to_gpa_write(struct kvm_vcpu *vcpu, gva_t gva, in kvm_mmu_gva_to_gpa_write() argument 3552 return vcpu->arch.walk_mmu->gva_to_gpa(vcpu, gva, access, exception); in kvm_mmu_gva_to_gpa_write() 3556 gpa_t kvm_mmu_gva_to_gpa_system(struct kvm_vcpu *vcpu, gva_t gva, in kvm_mmu_gva_to_gpa_system() argument 3559 return vcpu->arch.walk_mmu->gva_to_gpa(vcpu, gva, 0, exception); in kvm_mmu_gva_to_gpa_system() 3659 static int vcpu_mmio_gva_to_gpa(struct kvm_vcpu *vcpu, unsigned long gva, in vcpu_mmio_gva_to_gpa() argument 3665 if (vcpu_match_mmio_gva(vcpu, gva) && in vcpu_mmio_gva_to_gpa() [all …]
|
D | paging_tmpl.h | 669 static void FNAME(invlpg)(struct kvm_vcpu *vcpu, gva_t gva) in FNAME() 676 vcpu_clear_mmio_info(vcpu, gva); in FNAME() 685 for_each_shadow_entry(vcpu, gva, iterator) { in FNAME()
|
D | vmx.c | 897 static inline void __invvpid(int ext, u16 vpid, gva_t gva) in __invvpid() argument 902 u64 gva; in __invvpid() member 903 } operand = { vpid, 0, gva }; in __invvpid() 5214 gva_t gva; in handle_vmclear() local 5224 vmcs_read32(VMX_INSTRUCTION_INFO), &gva)) in handle_vmclear() 5227 if (kvm_read_guest_virt(&vcpu->arch.emulate_ctxt, gva, &vmptr, in handle_vmclear() 5361 gva_t gva = 0; in handle_vmread() local 5385 vmx_instruction_info, &gva)) in handle_vmread() 5388 kvm_write_guest_virt_system(&vcpu->arch.emulate_ctxt, gva, in handle_vmread() 5401 gva_t gva; in handle_vmwrite() local [all …]
|
/linux-3.4.99/virt/kvm/ |
D | async_pf.c | 65 gva_t gva = apf->gva; in async_pf_execute() local 86 trace_kvm_async_pf_completed(addr, page, gva); in async_pf_execute() 147 int kvm_setup_async_pf(struct kvm_vcpu *vcpu, gva_t gva, gfn_t gfn, in kvm_setup_async_pf() argument 168 work->gva = gva; in kvm_setup_async_pf()
|
/linux-3.4.99/arch/x86/include/asm/ |
D | kvm_host.h | 261 int (*page_fault)(struct kvm_vcpu *vcpu, gva_t gva, u32 err, 266 gpa_t (*gva_to_gpa)(struct kvm_vcpu *vcpu, gva_t gva, u32 access, 271 void (*invlpg)(struct kvm_vcpu *vcpu, gva_t gva); 813 int kvm_mmu_unprotect_page_virt(struct kvm_vcpu *vcpu, gva_t gva); 819 gpa_t kvm_mmu_gva_to_gpa_read(struct kvm_vcpu *vcpu, gva_t gva, 821 gpa_t kvm_mmu_gva_to_gpa_fetch(struct kvm_vcpu *vcpu, gva_t gva, 823 gpa_t kvm_mmu_gva_to_gpa_write(struct kvm_vcpu *vcpu, gva_t gva, 825 gpa_t kvm_mmu_gva_to_gpa_system(struct kvm_vcpu *vcpu, gva_t gva, 830 int kvm_mmu_page_fault(struct kvm_vcpu *vcpu, gva_t gva, u32 error_code, 832 void kvm_mmu_invlpg(struct kvm_vcpu *vcpu, gva_t gva);
|
/linux-3.4.99/Documentation/virtual/kvm/ |
D | mmu.txt | 36 gva guest virtual address 65 guest physical addresses, to host physical addresses (gva->gpa->hpa) 122 paging: gva->gpa->hpa 123 paging, tdp: (gva->)gpa->hpa 269 (gva->gpa or ngpa->gpa)
|
/linux-3.4.99/include/linux/ |
D | kvm_host.h | 99 gva_t gva; member 108 int kvm_setup_async_pf(struct kvm_vcpu *vcpu, gva_t gva, gfn_t gfn,
|