/linux-6.6.21/mm/ |
D | pgtable-generic.c | 117 flush_pmd_tlb_range(vma, address, address + HPAGE_PMD_SIZE); in pmdp_set_access_flags() 131 flush_pmd_tlb_range(vma, address, address + HPAGE_PMD_SIZE); in pmdp_clear_flush_young() 145 flush_pmd_tlb_range(vma, address, address + HPAGE_PMD_SIZE); in pmdp_huge_clear_flush() 202 flush_pmd_tlb_range(vma, address, address + HPAGE_PMD_SIZE); in pmdp_invalidate() 230 flush_tlb_range(vma, address, address + HPAGE_PMD_SIZE); in pmdp_collapse_flush()
|
D | huge_memory.c | 113 !transhuge_vma_suitable(vma, (vma->vm_end - HPAGE_PMD_SIZE))) in hugepage_vma_check() 389 return sysfs_emit(buf, "%lu\n", HPAGE_PMD_SIZE); in hpage_pmd_size_show() 1589 tlb_change_page_size(tlb, HPAGE_PMD_SIZE); in madvise_free_huge_pmd() 1620 if (next - addr != HPAGE_PMD_SIZE) { in madvise_free_huge_pmd() 1665 tlb_change_page_size(tlb, HPAGE_PMD_SIZE); in zap_huge_pmd() 1717 tlb_remove_page_size(tlb, page, HPAGE_PMD_SIZE); in zap_huge_pmd() 1816 tlb_change_page_size(tlb, HPAGE_PMD_SIZE); in change_huge_pmd() 1929 tlb_flush_pmd_range(tlb, addr, HPAGE_PMD_SIZE); in change_huge_pmd() 2081 VM_BUG_ON_VMA(vma->vm_end < haddr + HPAGE_PMD_SIZE, vma); in __split_huge_pmd_locked() 2274 (address & HPAGE_PMD_MASK) + HPAGE_PMD_SIZE); in __split_huge_pmd() [all …]
|
D | khugepaged.c | 1152 address + HPAGE_PMD_SIZE); in collapse_huge_page() 1487 !range_in_vma(vma, haddr, haddr + HPAGE_PMD_SIZE)) in collapse_pte_mapped_thp() 1572 haddr, haddr + HPAGE_PMD_SIZE); in collapse_pte_mapped_thp() 1703 vma->vm_end < addr + HPAGE_PMD_SIZE) in retract_page_tables() 1724 addr, addr + HPAGE_PMD_SIZE); in retract_page_tables() 2377 hstart = round_up(vma->vm_start, HPAGE_PMD_SIZE); in khugepaged_scan_mm_slot() 2378 hend = round_down(vma->vm_end, HPAGE_PMD_SIZE); in khugepaged_scan_mm_slot() 2393 khugepaged_scan.address + HPAGE_PMD_SIZE > in khugepaged_scan_mm_slot() 2424 khugepaged_scan.address += HPAGE_PMD_SIZE; in khugepaged_scan_mm_slot() 2723 for (addr = hstart; addr < hend; addr += HPAGE_PMD_SIZE) { in madvise_collapse()
|
D | shmem.c | 641 round_up(inode->i_size, HPAGE_PMD_SIZE)) { in shmem_unused_huge_shrink() 1159 stat->blksize = HPAGE_PMD_SIZE; in shmem_getattr() 2282 if (len < HPAGE_PMD_SIZE) in shmem_get_unmapped_area() 2314 offset = (pgoff << PAGE_SHIFT) & (HPAGE_PMD_SIZE-1); in shmem_get_unmapped_area() 2315 if (offset && offset + len < 2 * HPAGE_PMD_SIZE) in shmem_get_unmapped_area() 2317 if ((addr & (HPAGE_PMD_SIZE-1)) == offset) in shmem_get_unmapped_area() 2320 inflated_len = len + HPAGE_PMD_SIZE - PAGE_SIZE; in shmem_get_unmapped_area() 2332 inflated_offset = inflated_addr & (HPAGE_PMD_SIZE-1); in shmem_get_unmapped_area() 2335 inflated_addr += HPAGE_PMD_SIZE; in shmem_get_unmapped_area()
|
D | madvise.c | 368 tlb_change_page_size(tlb, HPAGE_PMD_SIZE); in madvise_cold_or_pageout_pte_range() 392 if (next - addr != HPAGE_PMD_SIZE) { in madvise_cold_or_pageout_pte_range()
|
D | mprotect.c | 391 if ((next - addr != HPAGE_PMD_SIZE) || in change_pmd_range()
|
D | mremap.c | 554 if (extent == HPAGE_PMD_SIZE && in move_page_tables()
|
D | memory.c | 1156 VM_BUG_ON_VMA(next-addr != HPAGE_PMD_SIZE, src_vma); in copy_pmd_range() 1551 if (next - addr != HPAGE_PMD_SIZE) in zap_pmd_range() 1560 next - addr == HPAGE_PMD_SIZE && pmd_none(*pmd)) { in zap_pmd_range()
|
D | rmap.c | 976 address + HPAGE_PMD_SIZE); in page_vma_mkclean_one()
|
/linux-6.6.21/include/linux/ |
D | huge_mm.h | 72 #define HPAGE_PMD_SIZE ((1UL) << HPAGE_PMD_SHIFT) macro 73 #define HPAGE_PMD_MASK (~(HPAGE_PMD_SIZE - 1)) 114 if (haddr < vma->vm_start || haddr + HPAGE_PMD_SIZE > vma->vm_end) in transhuge_vma_suitable() 259 #define HPAGE_PMD_SIZE ({ BUILD_BUG(); 0; }) macro
|
/linux-6.6.21/arch/powerpc/mm/book3s64/ |
D | pgtable.c | 174 flush_pmd_tlb_range(vma, address, address + HPAGE_PMD_SIZE); in pmdp_invalidate() 192 flush_pmd_tlb_range(vma, addr, addr + HPAGE_PMD_SIZE); in pmdp_huge_get_and_clear_full()
|
D | radix_tlb.c | 1331 end = addr + HPAGE_PMD_SIZE; in radix__flush_tlb_collapsed_pmd()
|
D | hash_utils.c | 1974 max_hpte_count = HPAGE_PMD_SIZE >> shift; in flush_hash_hugepage()
|
/linux-6.6.21/arch/arc/mm/ |
D | tlb.c | 656 mmu->s_pg_sz_m != TO_MB(HPAGE_PMD_SIZE)) in arc_mmu_init() 658 (unsigned long)TO_MB(HPAGE_PMD_SIZE)); in arc_mmu_init()
|
/linux-6.6.21/arch/sparc/mm/ |
D | tlb.c | 254 flush_tlb_range(vma, address, address + HPAGE_PMD_SIZE); in pmdp_invalidate()
|
/linux-6.6.21/include/asm-generic/ |
D | tlb.h | 635 tlb_flush_pmd_range(tlb, address, HPAGE_PMD_SIZE); \
|
/linux-6.6.21/drivers/nvdimm/ |
D | pfn_devs.c | 102 alignments[1] = HPAGE_PMD_SIZE; in nd_pfn_supported_alignments() 117 return HPAGE_PMD_SIZE; in nd_pfn_default_alignment()
|
/linux-6.6.21/fs/proc/ |
D | task_mmu.c | 591 mss->anonymous_thp += HPAGE_PMD_SIZE; in smaps_pmd_entry() 593 mss->shmem_thp += HPAGE_PMD_SIZE; in smaps_pmd_entry() 597 mss->file_thp += HPAGE_PMD_SIZE; in smaps_pmd_entry() 1883 HPAGE_PMD_SIZE/PAGE_SIZE); in gather_pte_stats()
|
/linux-6.6.21/mm/damon/ |
D | vaddr.c | 469 *priv->folio_sz = HPAGE_PMD_SIZE; in damon_young_pmd_entry()
|
/linux-6.6.21/arch/x86/mm/ |
D | pgtable.c | 623 flush_tlb_range(vma, address, address + HPAGE_PMD_SIZE); in pmdp_clear_flush_young()
|