Searched refs:kvm_mmu_page (Results 1 – 12 of 12) sorted by relevance
/linux-5.19.10/arch/x86/kvm/mmu/ |
D | tdp_mmu.c | 68 static void tdp_mmu_free_sp(struct kvm_mmu_page *sp) in tdp_mmu_free_sp() 84 struct kvm_mmu_page *sp = container_of(head, struct kvm_mmu_page, in tdp_mmu_free_sp_rcu_callback() 90 static void tdp_mmu_zap_root(struct kvm *kvm, struct kvm_mmu_page *root, 95 struct kvm_mmu_page *root = container_of(work, struct kvm_mmu_page, in tdp_mmu_zap_root_work() 123 static void tdp_mmu_schedule_zap_root(struct kvm *kvm, struct kvm_mmu_page *root) in tdp_mmu_schedule_zap_root() 130 static inline bool kvm_tdp_root_mark_invalid(struct kvm_mmu_page *page) in kvm_tdp_root_mark_invalid() 140 void kvm_tdp_mmu_put_root(struct kvm *kvm, struct kvm_mmu_page *root, in kvm_tdp_mmu_put_root() 204 static struct kvm_mmu_page *tdp_mmu_next_root(struct kvm *kvm, in tdp_mmu_next_root() 205 struct kvm_mmu_page *prev_root, in tdp_mmu_next_root() 208 struct kvm_mmu_page *next_root; in tdp_mmu_next_root() [all …]
|
D | tdp_mmu.h | 10 __must_check static inline bool kvm_tdp_mmu_get_root(struct kvm_mmu_page *root) in kvm_tdp_mmu_get_root() 15 void kvm_tdp_mmu_put_root(struct kvm *kvm, struct kvm_mmu_page *root, 20 bool kvm_tdp_mmu_zap_sp(struct kvm *kvm, struct kvm_mmu_page *sp); 71 static inline bool is_tdp_mmu_page(struct kvm_mmu_page *sp) { return sp->tdp_mmu_page; } in is_tdp_mmu_page() 75 struct kvm_mmu_page *sp; in is_tdp_mmu() 92 static inline bool is_tdp_mmu_page(struct kvm_mmu_page *sp) { return false; } in is_tdp_mmu_page()
|
D | mmu_internal.h | 35 struct kvm_mmu_page { struct 96 static inline struct kvm_mmu_page *to_shadow_page(hpa_t shadow_page) in to_shadow_page() argument 100 return (struct kvm_mmu_page *)page_private(page); in to_shadow_page() 103 static inline struct kvm_mmu_page *sptep_to_sp(u64 *sptep) in sptep_to_sp() 113 static inline int kvm_mmu_page_as_id(struct kvm_mmu_page *sp) in kvm_mmu_page_as_id() 118 static inline bool kvm_mmu_page_ad_need_write_protect(struct kvm_mmu_page *sp) in kvm_mmu_page_ad_need_write_protect() 290 void account_huge_nx_page(struct kvm *kvm, struct kvm_mmu_page *sp); 291 void unaccount_huge_nx_page(struct kvm *kvm, struct kvm_mmu_page *sp);
|
D | mmu.c | 367 struct kvm_mmu_page *sp = sptep_to_sp(sptep); in count_spte_clear() 451 struct kvm_mmu_page *sp = sptep_to_sp(sptep); in __get_spte_lockless() 712 static gfn_t kvm_mmu_page_get_gfn(struct kvm_mmu_page *sp, int index) in kvm_mmu_page_get_gfn() 723 static void kvm_mmu_page_set_gfn(struct kvm_mmu_page *sp, int index, gfn_t gfn) in kvm_mmu_page_set_gfn() 778 static void account_shadowed(struct kvm *kvm, struct kvm_mmu_page *sp) in account_shadowed() 797 void account_huge_nx_page(struct kvm *kvm, struct kvm_mmu_page *sp) in account_huge_nx_page() 808 static void unaccount_shadowed(struct kvm *kvm, struct kvm_mmu_page *sp) in unaccount_shadowed() 825 void unaccount_huge_nx_page(struct kvm *kvm, struct kvm_mmu_page *sp) in unaccount_huge_nx_page() 1028 struct kvm_mmu_page *sp; in rmap_remove() 1147 struct kvm_mmu_page *sp = sptep_to_sp(sptep); in drop_large_spte() [all …]
|
D | mmutrace.h | 159 TP_PROTO(struct kvm_mmu_page *sp, bool created), 178 TP_PROTO(struct kvm_mmu_page *sp), 193 TP_PROTO(struct kvm_mmu_page *sp), 199 TP_PROTO(struct kvm_mmu_page *sp), 205 TP_PROTO(struct kvm_mmu_page *sp),
|
D | paging_tmpl.h | 148 struct kvm_mmu_page *sp, u64 *spte, in FNAME() 515 FNAME(prefetch_gpte)(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp, in FNAME() 572 struct kvm_mmu_page *sp; in FNAME() 615 struct kvm_mmu_page *sp = NULL; in FNAME() 884 static gpa_t FNAME(get_level1_sp_gpa)(struct kvm_mmu_page *sp) in FNAME() 899 struct kvm_mmu_page *sp; in FNAME() 990 static int FNAME(sync_page)(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp) in FNAME()
|
D | spte.h | 227 static inline bool sp_ad_disabled(struct kvm_mmu_page *sp) in sp_ad_disabled() 427 bool make_spte(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp,
|
D | tdp_iter.h | 113 void tdp_iter_start(struct tdp_iter *iter, struct kvm_mmu_page *root,
|
D | tdp_iter.c | 43 void tdp_iter_start(struct tdp_iter *iter, struct kvm_mmu_page *root, in tdp_iter_start()
|
D | spte.c | 136 bool make_spte(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp, in make_spte()
|
/linux-5.19.10/arch/x86/include/asm/ |
D | kvm_host.h | 419 struct kvm_mmu_page; 437 struct kvm_mmu_page *sp);
|
/linux-5.19.10/Documentation/virt/kvm/x86/ |
D | mmu.rst | 121 The principal data structure is the shadow page, 'struct kvm_mmu_page'. A 198 shadow page; it is also used to go back from a struct kvm_mmu_page
|