/linux-6.1.9/tools/testing/selftests/vm/ |
D | hugetlb-madvise.c | 36 unsigned long huge_page_size; variable 87 *((unsigned long *)(addr + (i * huge_page_size))) = i; in write_fault_pages() 96 dummy += *((unsigned long *)(addr + (i * huge_page_size))); in read_fault_pages() 111 huge_page_size = default_huge_page_size(); in main() 112 if (!huge_page_size) { in main() 117 if (!huge_page_size) { in main() 140 addr = mmap(NULL, (NR_HUGE_PAGES + 2) * huge_page_size, in main() 148 if (munmap(addr, huge_page_size) || in main() 149 munmap(addr + (NR_HUGE_PAGES + 1) * huge_page_size, in main() 150 huge_page_size)) { in main() [all …]
|
/linux-6.1.9/fs/hugetlbfs/ |
D | inode.c | 256 addr = ALIGN(addr, huge_page_size(h)); in generic_hugetlb_get_unmapped_area() 306 nr = huge_page_size(h); in hugetlbfs_read_iter() 392 huge_page_size(hstate_vma(vma))); in hugetlb_vma_maps_page() 708 end = huge_page_size(h); in hugetlbfs_zero_partial_page() 721 loff_t hpage_size = huge_page_size(h); in hugetlbfs_punch_hole() 778 loff_t hpage_size = huge_page_size(h); in hugetlbfs_fallocate() 1130 unsigned long hpage_size = huge_page_size(sbinfo->hstate); in hugetlbfs_show_options() 1169 buf->f_bsize = huge_page_size(h); in hugetlbfs_statfs() 1485 sb->s_blocksize = huge_page_size(ctx->hstate); in hugetlbfs_fill_super() 1645 huge_page_size(h) / SZ_1K); in mount_one_hugetlbfs()
|
/linux-6.1.9/mm/ |
D | hugetlb.c | 1842 if (huge_page_size(h) == size) in size_to_hstate() 3150 m = memblock_alloc_try_nid_raw(huge_page_size(h), huge_page_size(h), in __alloc_bootmem_huge_page() 3159 huge_page_size(h), huge_page_size(h), in __alloc_bootmem_huge_page() 3235 string_get_size(huge_page_size(h), 1, STRING_UNITS_2, buf, 32); in hugetlb_hstate_alloc_pages_onenode() 3297 string_get_size(huge_page_size(h), 1, STRING_UNITS_2, buf, 32); in hugetlb_hstate_alloc_pages() 3343 string_get_size(huge_page_size(h), 1, STRING_UNITS_2, buf, 32); in report_hugepages() 4226 string_get_size(huge_page_size(&default_hstate), in hugetlb_init() 4285 h->mask = ~(huge_page_size(h) - 1); in hugetlb_add_hstate() 4292 huge_page_size(h)/SZ_1K); in hugetlb_add_hstate() 4639 total += huge_page_size(h) * count; in hugetlb_report_meminfo() [all …]
|
D | pagewalk.c | 289 unsigned long boundary = (addr & huge_page_mask(h)) + huge_page_size(h); in hugetlb_entry_end() 300 unsigned long sz = huge_page_size(h); in walk_hugetlb_range()
|
D | page_vma_mapped.c | 167 unsigned long size = huge_page_size(hstate); in page_vma_mapped_walk()
|
D | mremap.c | 954 old_len = ALIGN(old_len, huge_page_size(h)); in SYSCALL_DEFINE5() 955 new_len = ALIGN(new_len, huge_page_size(h)); in SYSCALL_DEFINE5()
|
D | hugetlb_cgroup.c | 720 mem_fmt(buf, sizeof(buf), huge_page_size(h)); in __hugetlb_cgroup_file_dfl_init() 791 mem_fmt(buf, sizeof(buf), huge_page_size(h)); in __hugetlb_cgroup_file_legacy_init()
|
D | madvise.c | 822 *end = ALIGN_DOWN(*end, huge_page_size(hstate_vma(vma))); in madvise_dontneed_free_valid_vma()
|
D | mmap.c | 1432 len = ALIGN(len, huge_page_size(hstate_file(file))); in ksys_mmap_pgoff() 1444 len = ALIGN(len, huge_page_size(hs)); in ksys_mmap_pgoff()
|
/linux-6.1.9/include/linux/ |
D | hugetlb.h | 763 static inline unsigned long huge_page_size(const struct hstate *h) in huge_page_size() function 799 return huge_page_size(h) / 512; in blocks_per_huge_page() 927 if (huge_page_size(h) == PMD_SIZE) in huge_pte_lockptr() 929 VM_BUG_ON(huge_page_size(h) == PAGE_SIZE); in huge_pte_lockptr() 1048 static inline unsigned long huge_page_size(struct hstate *h)
|
/linux-6.1.9/arch/parisc/mm/ |
D | hugetlbpage.c | 40 addr = ALIGN(addr, huge_page_size(h)); in hugetlb_get_unmapped_area()
|
/linux-6.1.9/arch/arm64/mm/ |
D | hugetlbpage.c | 78 size_t pagesize = huge_page_size(h); in arch_hugetlb_migration_supported() 376 unsigned long hp_size = huge_page_size(h); in hugetlb_mask_last_page()
|
/linux-6.1.9/drivers/misc/habanalabs/common/ |
D | memory.c | 846 huge_page_size = ctx->hdev->asic_prop.pmmu_huge.page_size; in init_phys_pg_pack_from_userptr() local 847 u32 pgs_in_huge_page = huge_page_size >> __ffs(page_size); in init_phys_pg_pack_from_userptr() 879 (dma_addr & (huge_page_size - 1))) in init_phys_pg_pack_from_userptr() 884 page_size = huge_page_size; in init_phys_pg_pack_from_userptr() 1101 huge_page_size = hdev->asic_prop.pmmu_huge.page_size; in map_device_va() local 1130 if (addr & (huge_page_size - 1)) in map_device_va() 1133 va_block_align = huge_page_size; in map_device_va() 1141 va_block_align = huge_page_size; in map_device_va()
|
/linux-6.1.9/mm/damon/ |
D | vaddr.c | 350 addr + huge_page_size(hstate_vma(vma)))) in damon_hugetlb_mkold() 507 *priv->page_sz = huge_page_size(h); in damon_young_hugetlb_entry()
|
/linux-6.1.9/include/asm-generic/ |
D | tlb.h | 575 unsigned long _sz = huge_page_size(h); \
|
/linux-6.1.9/arch/s390/mm/ |
D | hugetlbpage.c | 318 addr = ALIGN(addr, huge_page_size(h)); in hugetlb_get_unmapped_area()
|
/linux-6.1.9/drivers/dma-buf/ |
D | udmabuf.c | 233 maxsubpgs = huge_page_size(hpstate) >> PAGE_SHIFT; in udmabuf_create()
|
/linux-6.1.9/arch/sparc/mm/ |
D | hugetlbpage.c | 120 addr = ALIGN(addr, huge_page_size(h)); in hugetlb_get_unmapped_area()
|
D | init_64.c | 419 hugepage_size = huge_page_size(hstate_vma(vma)); in update_mmu_cache()
|
/linux-6.1.9/fs/proc/ |
D | task_mmu.c | 743 mss->shared_hugetlb += huge_page_size(hstate_vma(vma)); in smaps_hugetlb_range() 745 mss->private_hugetlb += huge_page_size(hstate_vma(vma)); in smaps_hugetlb_range()
|
/linux-6.1.9/ipc/ |
D | shm.c | 752 hugesize = ALIGN(size, huge_page_size(hs)); in newseg()
|