/linux-6.1.9/arch/x86/kvm/mmu/ |
D | tdp_mmu.c | 343 u64 old_spte, u64 new_spte, int level, 346 static void handle_changed_spte_acc_track(u64 old_spte, u64 new_spte, int level) in handle_changed_spte_acc_track() argument 348 if (!is_shadow_present_pte(old_spte) || !is_last_spte(old_spte, level)) in handle_changed_spte_acc_track() 351 if (is_accessed_spte(old_spte) && in handle_changed_spte_acc_track() 353 spte_to_pfn(old_spte) != spte_to_pfn(new_spte))) in handle_changed_spte_acc_track() 354 kvm_set_pfn_accessed(spte_to_pfn(old_spte)); in handle_changed_spte_acc_track() 358 u64 old_spte, u64 new_spte, int level) in handle_changed_spte_dirty_log() argument 366 pfn_changed = spte_to_pfn(old_spte) != spte_to_pfn(new_spte); in handle_changed_spte_dirty_log() 368 if ((!is_writable_pte(old_spte) || pfn_changed) && in handle_changed_spte_dirty_log() 442 u64 old_spte; in handle_removed_pt() local [all …]
|
D | tdp_iter.h | 32 static inline u64 kvm_tdp_mmu_write_spte(tdp_ptep_t sptep, u64 old_spte, in kvm_tdp_mmu_write_spte() argument 47 if (is_shadow_present_pte(old_spte) && is_last_spte(old_spte, level) && in kvm_tdp_mmu_write_spte() 48 spte_has_volatile_bits(old_spte)) in kvm_tdp_mmu_write_spte() 52 return old_spte; in kvm_tdp_mmu_write_spte() 85 u64 old_spte; member
|
D | mmutrace.h | 257 u64 *sptep, u64 old_spte, int ret), 258 TP_ARGS(vcpu, fault, sptep, old_spte, ret), 265 __field(u64, old_spte) 275 __entry->old_spte = old_spte; 284 __entry->old_spte, __entry->new_spte, 393 TP_PROTO(int as_id, gfn_t gfn, int level, u64 old_spte, u64 new_spte), 394 TP_ARGS(as_id, gfn, level, old_spte, new_spte), 398 __field(u64, old_spte) 408 __entry->old_spte = old_spte; 416 __entry->old_spte, __entry->new_spte
|
D | tdp_iter.c | 15 iter->old_spte = kvm_tdp_mmu_read_spte(iter->sptep); in tdp_iter_refresh_sptep() 92 iter->old_spte = kvm_tdp_mmu_read_spte(iter->sptep); in try_step_down() 94 child_pt = spte_to_child_pt(iter->old_spte, iter->level); in try_step_down() 126 iter->old_spte = kvm_tdp_mmu_read_spte(iter->sptep); in try_step_side()
|
D | spte.c | 140 u64 old_spte, bool prefetch, bool can_unsync, in make_spte() argument 202 if (is_writable_pte(old_spte)) in make_spte() 315 u64 kvm_mmu_changed_pte_notifier_make_spte(u64 old_spte, kvm_pfn_t new_pfn) in kvm_mmu_changed_pte_notifier_make_spte() argument 319 new_spte = old_spte & ~SPTE_BASE_ADDR_MASK; in kvm_mmu_changed_pte_notifier_make_spte()
|
D | mmu.c | 462 u64 old_spte = *sptep; in mmu_spte_update_no_track() local 467 if (!is_shadow_present_pte(old_spte)) { in mmu_spte_update_no_track() 469 return old_spte; in mmu_spte_update_no_track() 472 if (!spte_has_volatile_bits(old_spte)) in mmu_spte_update_no_track() 475 old_spte = __update_clear_spte_slow(sptep, new_spte); in mmu_spte_update_no_track() 477 WARN_ON(spte_to_pfn(old_spte) != spte_to_pfn(new_spte)); in mmu_spte_update_no_track() 479 return old_spte; in mmu_spte_update_no_track() 494 u64 old_spte = mmu_spte_update_no_track(sptep, new_spte); in mmu_spte_update() local 496 if (!is_shadow_present_pte(old_spte)) in mmu_spte_update() 504 if (is_mmu_writable_spte(old_spte) && in mmu_spte_update() [all …]
|
D | spte.h | 447 u64 old_spte, bool prefetch, bool can_unsync, 469 u64 kvm_mmu_changed_pte_notifier_make_spte(u64 old_spte, kvm_pfn_t new_pfn);
|
D | paging_tmpl.h | 904 u64 old_spte; in FNAME() local 927 old_spte = *sptep; in FNAME() 928 if (is_last_spte(old_spte, level)) { in FNAME() 939 if (is_shadow_present_pte(old_spte)) in FNAME()
|
/linux-6.1.9/Documentation/virt/kvm/ |
D | locking.rst | 93 | old_spte = *spte; | | 108 | if (cmpxchg(spte, old_spte, old_spte+W) | 147 | old_spte = *spte; | | 151 | if (old_spte.Accessed == 1 && | | 152 | old_spte.W == 0) | | 166 | old_spte = xchg(spte, 0ull) | | 167 | if (old_spte.Accessed == 1) | | 169 | if (old_spte.Dirty == 1) | |
|