Home
last modified time | relevance | path

Searched refs:gva (Results 1 – 11 of 11) sorted by relevance

/linux-3.4.99/include/trace/events/
Dkvm.h221 TP_PROTO(u64 gva, u64 gfn),
223 TP_ARGS(gva, gfn),
226 __field(__u64, gva)
231 __entry->gva = gva;
235 TP_printk("gva = %#llx, gfn = %#llx", __entry->gva, __entry->gfn)
240 TP_PROTO(u64 gva, u64 gfn),
242 TP_ARGS(gva, gfn)
247 TP_PROTO(u64 gva, u64 gfn),
249 TP_ARGS(gva, gfn)
254 TP_PROTO(u64 token, u64 gva),
[all …]
/linux-3.4.99/arch/x86/kvm/
Dx86.h76 gva_t gva, gfn_t gfn, unsigned access) in vcpu_cache_mmio_info() argument
78 vcpu->arch.mmio_gva = gva & PAGE_MASK; in vcpu_cache_mmio_info()
87 static inline void vcpu_clear_mmio_info(struct kvm_vcpu *vcpu, gva_t gva) in vcpu_clear_mmio_info() argument
89 if (gva != (~0ul) && vcpu->arch.mmio_gva != (gva & PAGE_MASK)) in vcpu_clear_mmio_info()
95 static inline bool vcpu_match_mmio_gva(struct kvm_vcpu *vcpu, unsigned long gva) in vcpu_match_mmio_gva() argument
97 if (vcpu->arch.mmio_gva && vcpu->arch.mmio_gva == (gva & PAGE_MASK)) in vcpu_match_mmio_gva()
Dtrace.h801 TP_PROTO(gva_t gva, gpa_t gpa, bool write, bool gpa_match),
802 TP_ARGS(gva, gpa, write, gpa_match),
805 __field(gva_t, gva)
812 __entry->gva = gva;
818 TP_printk("gva %#lx gpa %#llx %s %s", __entry->gva, __entry->gpa,
Dmmu.c1372 static void nonpaging_invlpg(struct kvm_vcpu *vcpu, gva_t gva) in nonpaging_invlpg() argument
2564 static bool handle_abnormal_pfn(struct kvm_vcpu *vcpu, gva_t gva, gfn_t gfn, in handle_abnormal_pfn() argument
2576 vcpu_cache_mmio_info(vcpu, gva, gfn, access); in handle_abnormal_pfn()
2584 gva_t gva, pfn_t *pfn, bool write, bool *writable);
2978 static int nonpaging_page_fault(struct kvm_vcpu *vcpu, gva_t gva, in nonpaging_page_fault() argument
2984 pgprintk("%s: gva %lx error %x\n", __func__, gva, error_code); in nonpaging_page_fault()
2987 return handle_mmio_page_fault(vcpu, gva, error_code, true); in nonpaging_page_fault()
2996 gfn = gva >> PAGE_SHIFT; in nonpaging_page_fault()
2998 return nonpaging_map(vcpu, gva & PAGE_MASK, in nonpaging_page_fault()
3002 static int kvm_arch_setup_async_pf(struct kvm_vcpu *vcpu, gva_t gva, gfn_t gfn) in kvm_arch_setup_async_pf() argument
[all …]
Dx86.c3532 gpa_t kvm_mmu_gva_to_gpa_read(struct kvm_vcpu *vcpu, gva_t gva, in kvm_mmu_gva_to_gpa_read() argument
3536 return vcpu->arch.walk_mmu->gva_to_gpa(vcpu, gva, access, exception); in kvm_mmu_gva_to_gpa_read()
3539 gpa_t kvm_mmu_gva_to_gpa_fetch(struct kvm_vcpu *vcpu, gva_t gva, in kvm_mmu_gva_to_gpa_fetch() argument
3544 return vcpu->arch.walk_mmu->gva_to_gpa(vcpu, gva, access, exception); in kvm_mmu_gva_to_gpa_fetch()
3547 gpa_t kvm_mmu_gva_to_gpa_write(struct kvm_vcpu *vcpu, gva_t gva, in kvm_mmu_gva_to_gpa_write() argument
3552 return vcpu->arch.walk_mmu->gva_to_gpa(vcpu, gva, access, exception); in kvm_mmu_gva_to_gpa_write()
3556 gpa_t kvm_mmu_gva_to_gpa_system(struct kvm_vcpu *vcpu, gva_t gva, in kvm_mmu_gva_to_gpa_system() argument
3559 return vcpu->arch.walk_mmu->gva_to_gpa(vcpu, gva, 0, exception); in kvm_mmu_gva_to_gpa_system()
3659 static int vcpu_mmio_gva_to_gpa(struct kvm_vcpu *vcpu, unsigned long gva, in vcpu_mmio_gva_to_gpa() argument
3665 if (vcpu_match_mmio_gva(vcpu, gva) && in vcpu_mmio_gva_to_gpa()
[all …]
Dpaging_tmpl.h669 static void FNAME(invlpg)(struct kvm_vcpu *vcpu, gva_t gva) in FNAME()
676 vcpu_clear_mmio_info(vcpu, gva); in FNAME()
685 for_each_shadow_entry(vcpu, gva, iterator) { in FNAME()
Dvmx.c897 static inline void __invvpid(int ext, u16 vpid, gva_t gva) in __invvpid() argument
902 u64 gva; in __invvpid() member
903 } operand = { vpid, 0, gva }; in __invvpid()
5214 gva_t gva; in handle_vmclear() local
5224 vmcs_read32(VMX_INSTRUCTION_INFO), &gva)) in handle_vmclear()
5227 if (kvm_read_guest_virt(&vcpu->arch.emulate_ctxt, gva, &vmptr, in handle_vmclear()
5361 gva_t gva = 0; in handle_vmread() local
5385 vmx_instruction_info, &gva)) in handle_vmread()
5388 kvm_write_guest_virt_system(&vcpu->arch.emulate_ctxt, gva, in handle_vmread()
5401 gva_t gva; in handle_vmwrite() local
[all …]
/linux-3.4.99/virt/kvm/
Dasync_pf.c65 gva_t gva = apf->gva; in async_pf_execute() local
86 trace_kvm_async_pf_completed(addr, page, gva); in async_pf_execute()
147 int kvm_setup_async_pf(struct kvm_vcpu *vcpu, gva_t gva, gfn_t gfn, in kvm_setup_async_pf() argument
168 work->gva = gva; in kvm_setup_async_pf()
/linux-3.4.99/arch/x86/include/asm/
Dkvm_host.h261 int (*page_fault)(struct kvm_vcpu *vcpu, gva_t gva, u32 err,
266 gpa_t (*gva_to_gpa)(struct kvm_vcpu *vcpu, gva_t gva, u32 access,
271 void (*invlpg)(struct kvm_vcpu *vcpu, gva_t gva);
813 int kvm_mmu_unprotect_page_virt(struct kvm_vcpu *vcpu, gva_t gva);
819 gpa_t kvm_mmu_gva_to_gpa_read(struct kvm_vcpu *vcpu, gva_t gva,
821 gpa_t kvm_mmu_gva_to_gpa_fetch(struct kvm_vcpu *vcpu, gva_t gva,
823 gpa_t kvm_mmu_gva_to_gpa_write(struct kvm_vcpu *vcpu, gva_t gva,
825 gpa_t kvm_mmu_gva_to_gpa_system(struct kvm_vcpu *vcpu, gva_t gva,
830 int kvm_mmu_page_fault(struct kvm_vcpu *vcpu, gva_t gva, u32 error_code,
832 void kvm_mmu_invlpg(struct kvm_vcpu *vcpu, gva_t gva);
/linux-3.4.99/Documentation/virtual/kvm/
Dmmu.txt36 gva guest virtual address
65 guest physical addresses, to host physical addresses (gva->gpa->hpa)
122 paging: gva->gpa->hpa
123 paging, tdp: (gva->)gpa->hpa
269 (gva->gpa or ngpa->gpa)
/linux-3.4.99/include/linux/
Dkvm_host.h99 gva_t gva; member
108 int kvm_setup_async_pf(struct kvm_vcpu *vcpu, gva_t gva, gfn_t gfn,