/linux-6.1.9/arch/x86/include/asm/ |
D | kfence.h | 33 if (level != PG_LEVEL_4K) in arch_kfence_init_pool() 46 if (WARN_ON(!pte || level != PG_LEVEL_4K)) in kfence_protect_page()
|
D | pgtable_types.h | 523 PG_LEVEL_4K, enumerator
|
D | kvm_host.h | 136 #define KVM_NR_PAGE_SIZES (KVM_MAX_HUGEPAGE_LEVEL - PG_LEVEL_4K + 1)
|
/linux-6.1.9/arch/x86/kvm/mmu/ |
D | page_track.c | 92 index = gfn_to_index(gfn, slot->base_gfn, PG_LEVEL_4K); in update_gfn_track() 135 if (kvm_mmu_slot_gfn_write_protect(kvm, slot, gfn, PG_LEVEL_4K)) in kvm_slot_page_track_add_page() 193 index = gfn_to_index(gfn, slot->base_gfn, PG_LEVEL_4K); in kvm_slot_page_track_is_active()
|
D | spte.c | 164 if (level > PG_LEVEL_4K && (pte_access & ACC_EXEC_MASK) && in make_spte() 177 if (level > PG_LEVEL_4K) in make_spte() 233 WARN_ON(level > PG_LEVEL_4K); in make_spte() 284 if (role.level == PG_LEVEL_4K) { in make_huge_page_split_spte()
|
D | paging_tmpl.h | 74 #define gpte_to_gfn(pte) gpte_to_gfn_lvl((pte), PG_LEVEL_4K) 163 if (FNAME(is_rsvd_bits_set)(vcpu->arch.mmu, gpte, PG_LEVEL_4K)) in FNAME() 295 gpte |= level - PG_LEVEL_4K - 1; in FNAME() 429 if (walker->level > PG_LEVEL_4K && is_cpuid_PSE36()) in FNAME() 560 if (level == PG_LEVEL_4K) { in FNAME() 585 if (sp->role.level > PG_LEVEL_4K) in FNAME() 837 fault->max_level = PG_LEVEL_4K; in FNAME() 892 WARN_ON(sp->role.level != PG_LEVEL_4K); in FNAME()
|
D | mmu.c | 795 if (sp->role.level > PG_LEVEL_4K) in account_shadowed() 801 if (kvm_mmu_slot_gfn_write_protect(kvm, slot, gfn, PG_LEVEL_4K)) in account_shadowed() 826 if (sp->role.level > PG_LEVEL_4K) in unaccount_shadowed() 1022 return &slot->arch.rmap[level - PG_LEVEL_4K][idx]; in gfn_to_rmap() 1146 WARN_ON(sp->role.level == PG_LEVEL_4K); in drop_large_spte() 1265 PG_LEVEL_4K, slot); in kvm_mmu_write_protect_pt_masked() 1298 PG_LEVEL_4K, slot); in kvm_mmu_clear_dirty_pt_masked() 1334 kvm_mmu_try_split_huge_pages(kvm, slot, start, end, PG_LEVEL_4K); in kvm_arch_mmu_enable_log_dirty_pt_masked() 1384 return kvm_mmu_slot_gfn_write_protect(vcpu->kvm, slot, gfn, PG_LEVEL_4K); in kvm_vcpu_write_protect_gfn() 1520 for_each_slot_rmap_range(range->slot, PG_LEVEL_4K, KVM_MAX_HUGEPAGE_LEVEL, in kvm_handle_gfn_range() [all …]
|
D | mmu_internal.h | 276 .req_level = PG_LEVEL_4K, in kvm_mmu_do_page_fault() 277 .goal_level = PG_LEVEL_4K, in kvm_mmu_do_page_fault()
|
D | tdp_iter.h | 109 for_each_tdp_pte_min_level(iter, root, PG_LEVEL_4K, start, end)
|
D | tdp_mmu.c | 363 if (level > PG_LEVEL_4K) in handle_changed_spte_dirty_log() 537 WARN_ON(level < PG_LEVEL_4K); in __handle_changed_spte() 955 for_each_tdp_pte_min_level(iter, root, PG_LEVEL_4K, start, end) { in tdp_mmu_zap_leafs() 1321 if (iter->level != PG_LEVEL_4K || in set_spte_gfn() 1689 if (iter.level > PG_LEVEL_4K || in clear_dirty_pt_masked()
|
D | spte.h | 290 return (level == PG_LEVEL_4K) || is_large_pte(pte); in is_last_spte()
|
/linux-6.1.9/arch/x86/xen/ |
D | p2m.c | 250 BUG_ON(!ptep || level != PG_LEVEL_4K); in xen_build_mfn_list_list() 448 BUG_ON(!ptep || level != PG_LEVEL_4K); in get_phys_to_machine() 542 BUG_ON(!ptep || level != PG_LEVEL_4K); in xen_alloc_p2m_entry() 667 BUG_ON(!ptep || level != PG_LEVEL_4K); in __set_phys_to_machine()
|
/linux-6.1.9/arch/x86/mm/ |
D | mem_encrypt_amd.c | 258 case PG_LEVEL_4K: in pg_level_to_pfn() 423 if (level == PG_LEVEL_4K) { in early_set_memory_enc_dec()
|
D | pti.c | 283 if (!pte || WARN_ON(level != PG_LEVEL_4K) || pte_none(*pte)) in pti_setup_vsyscall()
|
D | kmmio.c | 168 case PG_LEVEL_4K: in clear_page_presence()
|
D | init_32.c | 376 update_page_count(PG_LEVEL_4K, pages_4k); in kernel_physical_mapping_init()
|
D | init_64.c | 497 update_page_count(PG_LEVEL_4K, pages); in phys_pte_init() 1099 update_page_count(PG_LEVEL_4K, -pages); in remove_pte_table()
|
/linux-6.1.9/tools/testing/selftests/kvm/lib/x86_64/ |
D | vmx.c | 449 for (int level = PG_LEVEL_512G; level >= PG_LEVEL_4K; level--) { in __nested_pg_map() 473 __nested_pg_map(vmx, vm, nested_paddr, paddr, PG_LEVEL_4K); in nested_pg_map() 513 __nested_map(vmx, vm, nested_paddr, paddr, size, PG_LEVEL_4K); in nested_map()
|
D | processor.c | 214 pte = virt_get_pte(vm, PTE_GET_PFN(*pde), vaddr, PG_LEVEL_4K); in __virt_pg_map() 222 __virt_pg_map(vm, vaddr, paddr, PG_LEVEL_4K); in virt_arch_pg_map()
|
/linux-6.1.9/tools/testing/selftests/kvm/x86_64/ |
D | nx_huge_pages_test.c | 150 virt_map_level(vm, HPAGE_GVA, HPAGE_GPA, nr_bytes, PG_LEVEL_4K); in run_test()
|
/linux-6.1.9/arch/x86/coco/tdx/ |
D | tdx.c | 684 case PG_LEVEL_4K: in try_accept_one() 752 if (!try_accept_one(&start, len, PG_LEVEL_4K)) in tdx_enc_status_changed()
|
/linux-6.1.9/arch/x86/kvm/vmx/ |
D | capabilities.h | 330 return PG_LEVEL_4K; in ept_caps_to_lpage_level()
|
/linux-6.1.9/arch/x86/mm/pat/ |
D | cpa-test.c | 206 if (level != PG_LEVEL_4K) { in pageattr_test()
|
D | set_memory.c | 111 direct_pages_count[PG_LEVEL_4K] << 2); in arch_report_meminfo() 496 if (lookup_address(start, &level) && (level != PG_LEVEL_4K)) in protect_kernel_text_ro() 666 *level = PG_LEVEL_4K; in lookup_address_in_pgd() 1572 if (level == PG_LEVEL_4K) { in __change_page_attr()
|
/linux-6.1.9/tools/testing/selftests/kvm/include/x86_64/ |
D | processor.h | 858 PG_LEVEL_4K, enumerator 868 #define PG_SIZE_4K PG_LEVEL_SIZE(PG_LEVEL_4K)
|