/linux-6.1.9/arch/ia64/include/asm/ |
D | pgalloc.h | 51 pmd_populate(struct mm_struct *mm, pmd_t * pmd_entry, pgtable_t pte) in pmd_populate() argument 53 pmd_val(*pmd_entry) = page_to_phys(pte); in pmd_populate() 57 pmd_populate_kernel(struct mm_struct *mm, pmd_t * pmd_entry, pte_t * pte) in pmd_populate_kernel() argument 59 pmd_val(*pmd_entry) = __pa(pte); in pmd_populate_kernel()
|
/linux-6.1.9/arch/x86/kernel/ |
D | head64.c | 187 pmdval_t *pmd, pmd_entry; in __startup_64() local 269 pmd_entry = __PAGE_KERNEL_LARGE_EXEC & ~_PAGE_GLOBAL; in __startup_64() 272 pmd_entry &= *mask_ptr; in __startup_64() 273 pmd_entry += sme_get_me_mask(); in __startup_64() 274 pmd_entry += physaddr; in __startup_64() 279 pmd[idx % PTRS_PER_PMD] = pmd_entry + i * PMD_SIZE; in __startup_64()
|
/linux-6.1.9/mm/ |
D | pagewalk.c | 127 if (ops->pmd_entry) in walk_pmd_range() 128 err = ops->pmd_entry(pmd, addr, next, walk); in walk_pmd_range() 194 !(ops->pmd_entry || ops->pte_entry)) in walk_pud_range() 239 else if (ops->pud_entry || ops->pmd_entry || ops->pte_entry) in walk_p4d_range() 276 else if (ops->p4d_entry || ops->pud_entry || ops->pmd_entry || ops->pte_entry) in walk_pgd_range()
|
D | mapping_dirty_helpers.c | 249 .pmd_entry = wp_clean_pmd_entry, 258 .pmd_entry = wp_clean_pmd_entry,
|
D | ptdump.c | 146 .pmd_entry = ptdump_pmd_entry,
|
D | mincore.c | 177 .pmd_entry = mincore_pte_range,
|
D | madvise.c | 234 .pmd_entry = swapin_walk_pmd_entry, 501 .pmd_entry = madvise_cold_or_pageout_pte_range, 733 .pmd_entry = madvise_free_pte_range,
|
D | hmm.c | 547 .pmd_entry = hmm_vma_walk_pmd,
|
D | migrate_device.c | 287 .pmd_entry = migrate_vma_collect_pmd,
|
D | mlock.c | 367 .pmd_entry = mlock_pte_range, in mlock_vma_pages_range()
|
D | memory-failure.c | 723 .pmd_entry = hwpoison_pte_range,
|
D | mempolicy.c | 707 .pmd_entry = queue_pages_pte_range,
|
D | memcontrol.c | 5945 .pmd_entry = mem_cgroup_count_precharge_pte_range, 6221 .pmd_entry = mem_cgroup_move_charge_pte_range,
|
/linux-6.1.9/include/linux/ |
D | pagewalk.h | 44 int (*pmd_entry)(pmd_t *pmd, unsigned long addr, member
|
/linux-6.1.9/arch/riscv/mm/ |
D | pageattr.c | 102 .pmd_entry = pageattr_pmd_entry,
|
/linux-6.1.9/mm/damon/ |
D | vaddr.c | 385 .pmd_entry = damon_mkold_pmd_entry, 522 .pmd_entry = damon_young_pmd_entry,
|
/linux-6.1.9/fs/ |
D | dax.c | 1175 pmd_t pmd_entry; in dax_pmd_load_hole() local 1203 pmd_entry = mk_pmd(zero_page, vmf->vma->vm_page_prot); in dax_pmd_load_hole() 1204 pmd_entry = pmd_mkhuge(pmd_entry); in dax_pmd_load_hole() 1205 set_pmd_at(vmf->vma->vm_mm, pmd_addr, vmf->pmd, pmd_entry); in dax_pmd_load_hole()
|
/linux-6.1.9/arch/powerpc/mm/book3s64/ |
D | subpage_prot.c | 145 .pmd_entry = subpage_walk_pmd_entry,
|
/linux-6.1.9/fs/proc/ |
D | task_mmu.c | 754 .pmd_entry = smaps_pte_range, 759 .pmd_entry = smaps_pte_range, 1242 .pmd_entry = clear_refs_pte_range, 1617 .pmd_entry = pagemap_pmd_range, 1925 .pmd_entry = gather_pte_stats,
|
/linux-6.1.9/arch/s390/mm/ |
D | gmap.c | 2512 .pmd_entry = thp_split_walk_pmd_entry, 2557 .pmd_entry = __zap_zero_pages, 2654 .pmd_entry = __s390_enable_skey_pmd,
|