Home
last modified time | relevance | path

Searched refs:hstate_vma (Results 1 – 18 of 18) sorted by relevance

/linux-6.1.9/mm/damon/
Dvaddr.c350 addr + huge_page_size(hstate_vma(vma)))) in damon_hugetlb_mkold()
365 struct hstate *h = hstate_vma(walk->vma); in damon_mkold_hugetlb_entry()
492 struct hstate *h = hstate_vma(walk->vma); in damon_young_hugetlb_entry()
/linux-6.1.9/mm/
Dhugetlb.c972 return vma_hugecache_offset(hstate_vma(vma), vma, address); in linear_hugepage_index()
3015 struct hstate *h = hstate_vma(vma); in alloc_huge_page()
4797 struct hstate *h = hstate_vma(vma); in hugetlb_vm_op_close()
4828 if (addr & ~(huge_page_mask(hstate_vma(vma)))) in hugetlb_vm_op_split()
4854 return huge_page_size(hstate_vma(vma)); in hugetlb_vm_op_pagesize()
4888 unsigned int shift = huge_page_shift(hstate_vma(vma)); in make_huge_pte()
4946 hugetlb_count_add(pages_per_huge_page(hstate_vma(vma)), vma->vm_mm); in hugetlb_install_page()
4959 struct hstate *h = hstate_vma(src_vma); in copy_hugetlb_page_range()
5137 struct hstate *h = hstate_vma(vma); in move_huge_pte()
5165 struct hstate *h = hstate_vma(vma); in move_hugetlb_page_tables()
[all …]
Dhmm.c486 ptl = huge_pte_lock(hstate_vma(vma), walk->mm, pte); in hmm_vma_walk_hugetlb_entry()
492 hmm_pfn_flags_order(huge_page_order(hstate_vma(vma))); in hmm_vma_walk_hugetlb_entry()
Dpage_vma_mapped.c166 struct hstate *hstate = hstate_vma(vma); in page_vma_mapped_walk()
Dmadvise.c813 if (start & ~huge_page_mask(hstate_vma(vma))) in madvise_dontneed_free_valid_vma()
822 *end = ALIGN_DOWN(*end, huge_page_size(hstate_vma(vma))); in madvise_dontneed_free_valid_vma()
Dpagewalk.c297 struct hstate *h = hstate_vma(vma); in walk_hugetlb_range()
Dmempolicy.c573 ptl = huge_pte_lock(hstate_vma(walk->vma), walk->mm, pte); in queue_pages_hugetlb()
1737 !hugepage_migration_supported(hstate_vma(vma))) in vma_migratable()
2029 huge_page_shift(hstate_vma(vma))); in huge_node()
Dmigrate.c238 unsigned int shift = huge_page_shift(hstate_vma(vma)); in remove_migration_pte()
338 spinlock_t *ptl = huge_pte_lockptr(hstate_vma(vma), vma->vm_mm, pte); in migration_entry_wait_huge()
Dmremap.c952 struct hstate *h __maybe_unused = hstate_vma(vma); in SYSCALL_DEFINE5()
Dmemory-failure.c713 struct hstate *h = hstate_vma(walk->vma); in hwpoison_hugetlb_range()
/linux-6.1.9/arch/x86/include/asm/
Dtlbflush.h229 ? huge_page_shift(hstate_vma(vma)) \
/linux-6.1.9/arch/riscv/kvm/
Dmmu.c643 vma_pageshift = huge_page_shift(hstate_vma(vma)); in kvm_riscv_gstage_map()
651 gfn = (gpa & huge_page_mask(hstate_vma(vma))) >> PAGE_SHIFT; in kvm_riscv_gstage_map()
/linux-6.1.9/include/linux/
Dhugetlb.h758 static inline struct hstate *hstate_vma(struct vm_area_struct *vma) in hstate_vma() function
1033 static inline struct hstate *hstate_vma(struct vm_area_struct *vma)
/linux-6.1.9/arch/powerpc/mm/
Dpgtable.c260 struct hstate *h = hstate_vma(vma); in huge_ptep_set_access_flags()
/linux-6.1.9/fs/proc/
Dtask_mmu.c743 mss->shared_hugetlb += huge_page_size(hstate_vma(vma)); in smaps_hugetlb_range()
745 mss->private_hugetlb += huge_page_size(hstate_vma(vma)); in smaps_hugetlb_range()
/linux-6.1.9/arch/arm64/kvm/
Dmmu.c1059 return huge_page_shift(hstate_vma(vma)); in get_vma_page_shift()
/linux-6.1.9/fs/hugetlbfs/
Dinode.c392 huge_page_size(hstate_vma(vma))); in hugetlb_vma_maps_page()
/linux-6.1.9/arch/sparc/mm/
Dinit_64.c419 hugepage_size = huge_page_size(hstate_vma(vma)); in update_mmu_cache()