/linux-5.19.10/arch/mips/kvm/ |
D | mmu.c | 205 int i_min = pud_index(start_gpa); in kvm_mips_flush_gpa_pud() 206 int i_max = pud_index(end_gpa); in kvm_mips_flush_gpa_pud() 332 int i_min = pud_index(start); \ 333 int i_max = pud_index(end); \
|
/linux-5.19.10/arch/riscv/mm/ |
D | init.c | 543 uintptr_t pud_index = pud_index(va); in create_pud_mapping() local 546 if (pud_val(pudp[pud_index]) == 0) in create_pud_mapping() 547 pudp[pud_index] = pfn_pud(PFN_DOWN(pa), prot); in create_pud_mapping() 551 if (pud_val(pudp[pud_index]) == 0) { in create_pud_mapping() 553 pudp[pud_index] = pfn_pud(PFN_DOWN(next_phys), PAGE_TABLE); in create_pud_mapping() 557 next_phys = PFN_PHYS(_pud_pfn(pudp[pud_index])); in create_pud_mapping()
|
D | kasan_init.c | 125 pudp = base_pud + pud_index(vaddr); in kasan_populate_pud() 346 pudp = base_pud + pud_index(vaddr); in kasan_shallow_populate_pud()
|
/linux-5.19.10/arch/riscv/include/asm/ |
D | pgtable-64.h | 298 #define pud_index(addr) (((addr) >> PUD_SHIFT) & (PTRS_PER_PUD - 1)) macro 304 return p4d_pgtable(*p4d) + pud_index(address); in pud_offset()
|
/linux-5.19.10/include/linux/ |
D | pgtable.h | 76 #ifndef pud_index 77 static inline unsigned long pud_index(unsigned long address) in pud_index() function 81 #define pud_index pud_index macro 119 return p4d_pgtable(*p4d) + pud_index(address); in pud_offset()
|
/linux-5.19.10/arch/x86/power/ |
D | hibernate_64.c | 70 set_pud(pud + pud_index(restore_jump_address), in set_up_temporary_text_mapping()
|
/linux-5.19.10/tools/testing/selftests/kvm/lib/aarch64/ |
D | processor.c | 33 static uint64_t pud_index(struct kvm_vm *vm, vm_vaddr_t gva) in pud_index() function 115 ptep = addr_gpa2hva(vm, pte_addr(vm, *ptep)) + pud_index(vm, vaddr) * 8; in _virt_pg_map() 155 ptep = addr_gpa2hva(vm, pte_addr(vm, *ptep)) + pud_index(vm, gva) * 8; in addr_gva2gpa()
|
/linux-5.19.10/arch/x86/mm/ |
D | kaslr.c | 164 pud_tramp = pud_page_tramp + pud_index(paddr); in init_trampoline_kaslr()
|
D | ident_map.c | 27 pud_t *pud = pud_page + pud_index(addr); in ident_pud_init()
|
D | init_64.c | 314 pud_t *pud = pud_page + pud_index(vaddr); in set_pte_vaddr_pud() 593 int i = pud_index(vaddr); in phys_pud_init() 601 pud = pud_page + pud_index(vaddr); in phys_pud_init() 1158 pud = pud_start + pud_index(addr); in remove_pud_table()
|
/linux-5.19.10/arch/x86/kernel/ |
D | head_64.S | 34 #define pud_index(x) (((x) >> PUD_SHIFT) & (PTRS_PER_PUD-1)) macro 39 L3_START_KERNEL = pud_index(__START_KERNEL_map)
|
D | espfix_64.c | 164 pud_p = &espfix_pud_page[pud_index(addr)]; in init_espfix_ap()
|
D | head64.c | 381 pud_p += pud_index(address); in __early_make_pgtable()
|
/linux-5.19.10/arch/csky/mm/ |
D | init.c | 166 j = pud_index(vaddr); in fixrange_init()
|
/linux-5.19.10/arch/x86/platform/efi/ |
D | efi_64.c | 141 num_entries = pud_index(EFI_VA_END); in efi_sync_low_kernel_mappings() 147 num_entries = PTRS_PER_PUD - pud_index(EFI_VA_START); in efi_sync_low_kernel_mappings()
|
/linux-5.19.10/arch/sh/mm/ |
D | init.c | 185 j = pud_index(vaddr); in page_table_range_init()
|
/linux-5.19.10/arch/s390/include/asm/ |
D | pgtable.h | 1298 #define pud_index(address) (((address) >> PUD_SHIFT) & (PTRS_PER_PUD-1)) macro 1375 return (pud_t *) p4d_deref(p4d) + pud_index(address); in pud_offset_lockless() 1578 r3o = __pa(pudp) - pud_index(addr) * sizeof(pud_t); in __pudp_idte()
|
/linux-5.19.10/arch/mips/mm/ |
D | init.c | 241 j = pud_index(vaddr); in fixrange_init()
|
/linux-5.19.10/arch/x86/xen/ |
D | mmu_pv.c | 558 nr = last ? pud_index(limit) + 1 : PTRS_PER_PUD; in xen_pud_walk() 1733 l2 = m2v(l3[pud_index(__START_KERNEL_map)].pud); in xen_setup_kernel_pagetable() 1852 pud = native_make_pud(xen_read_phys_ulong(pa + pud_index(vaddr) * in xen_early_virt_to_phys()
|
/linux-5.19.10/arch/powerpc/mm/book3s64/ |
D | radix_pgtable.c | 783 pud = pud_start + pud_index(addr); in remove_pud_table()
|
/linux-5.19.10/arch/arm64/include/asm/ |
D | pgtable.h | 765 #define pud_offset_phys(dir, addr) (p4d_page_paddr(READ_ONCE(*(dir))) + pud_index(addr) * sizeof(pu…
|