Lines Matching refs:gfn_t

285 			    gfn_t gfn, union kvm_mmu_page_role role)  in tdp_mmu_init_sp()
342 static void handle_changed_spte(struct kvm *kvm, int as_id, gfn_t gfn,
357 static void handle_changed_spte_dirty_log(struct kvm *kvm, int as_id, gfn_t gfn, in handle_changed_spte_dirty_log()
432 gfn_t base_gfn = sp->gfn; in handle_removed_pt()
441 gfn_t gfn = base_gfn + i * KVM_PAGES_PER_HPAGE(level); in handle_removed_pt()
526 static void __handle_changed_spte(struct kvm *kvm, int as_id, gfn_t gfn, in __handle_changed_spte()
614 static void handle_changed_spte(struct kvm *kvm, int as_id, gfn_t gfn, in handle_changed_spte()
727 u64 old_spte, u64 new_spte, gfn_t gfn, int level, in __tdp_mmu_set_spte()
843 static inline gfn_t tdp_mmu_max_gfn_exclusive(void) in tdp_mmu_max_gfn_exclusive()
859 gfn_t end = tdp_mmu_max_gfn_exclusive(); in __tdp_mmu_zap_root()
860 gfn_t start = 0; in __tdp_mmu_zap_root()
945 gfn_t start, gfn_t end, bool can_yield, bool flush) in tdp_mmu_zap_leafs()
984 bool kvm_tdp_mmu_zap_leafs(struct kvm *kvm, int as_id, gfn_t start, gfn_t end, in kvm_tdp_mmu_zap_leafs()
1365 gfn_t start, gfn_t end, int min_level) in wrprot_gfn_range()
1517 gfn_t start, gfn_t end, in tdp_mmu_split_huge_pages_root()
1584 gfn_t start, gfn_t end, in kvm_tdp_mmu_try_split_huge_pages()
1609 gfn_t start, gfn_t end) in clear_dirty_gfn_range()
1677 gfn_t gfn, unsigned long mask, bool wrprot) in clear_dirty_pt_masked()
1722 gfn_t gfn, unsigned long mask, in kvm_tdp_mmu_clear_dirty_pt_masked()
1736 gfn_t start = slot->base_gfn; in zap_collapsible_spte_range()
1737 gfn_t end = start + slot->npages; in zap_collapsible_spte_range()
1804 gfn_t gfn, int min_level) in write_protect_gfn()
1840 struct kvm_memory_slot *slot, gfn_t gfn, in kvm_tdp_mmu_write_protect_gfn()
1864 gfn_t gfn = addr >> PAGE_SHIFT; in kvm_tdp_mmu_get_walk()
1893 gfn_t gfn = addr >> PAGE_SHIFT; in kvm_tdp_mmu_fast_pf_get_last_sptep()