Home
last modified time | relevance | path

Searched refs:huge_page_mask (Results 1 – 16 of 16) sorted by relevance

/linux-5.19.10/arch/x86/mm/
Dhugetlbpage.c96 info.align_mask = PAGE_MASK & ~huge_page_mask(h); in hugetlb_get_unmapped_area_bottomup()
120 info.align_mask = PAGE_MASK & ~huge_page_mask(h); in hugetlb_get_unmapped_area_topdown()
149 if (len & ~huge_page_mask(h)) in hugetlb_get_unmapped_area()
163 addr &= huge_page_mask(h); in hugetlb_get_unmapped_area()
/linux-5.19.10/arch/mips/include/asm/
Dhugetlb.h22 if (len & ~huge_page_mask(h)) in prepare_hugepage_range()
24 if (addr & ~huge_page_mask(h)) in prepare_hugepage_range()
/linux-5.19.10/arch/loongarch/include/asm/
Dhugetlb.h21 if (len & ~huge_page_mask(h)) in prepare_hugepage_range()
23 if (addr & ~huge_page_mask(h)) in prepare_hugepage_range()
/linux-5.19.10/include/asm-generic/
Dhugetlb.h120 if (len & ~huge_page_mask(h)) in prepare_hugepage_range()
122 if (addr & ~huge_page_mask(h)) in prepare_hugepage_range()
/linux-5.19.10/arch/s390/include/asm/
Dhugetlb.h33 if (len & ~huge_page_mask(h)) in prepare_hugepage_range()
35 if (addr & ~huge_page_mask(h)) in prepare_hugepage_range()
/linux-5.19.10/fs/hugetlbfs/
Dinode.c166 if (vma->vm_pgoff & (~huge_page_mask(h) >> PAGE_SHIFT)) in hugetlbfs_file_mmap()
209 info.align_mask = PAGE_MASK & ~huge_page_mask(h); in hugetlb_get_unmapped_area_bottomup()
225 info.align_mask = PAGE_MASK & ~huge_page_mask(h); in hugetlb_get_unmapped_area_topdown()
256 if (len & ~huge_page_mask(h)) in generic_hugetlb_get_unmapped_area()
338 unsigned long offset = iocb->ki_pos & ~huge_page_mask(h); in hugetlbfs_read_iter()
356 nr = ((isize - 1) & ~huge_page_mask(h)) + 1; in hugetlbfs_read_iter()
387 offset &= ~huge_page_mask(h); in hugetlbfs_read_iter()
591 BUG_ON(offset & ~huge_page_mask(h)); in hugetlb_vmtruncate()
615 start = start & ~huge_page_mask(h); in hugetlbfs_zero_partial_page()
616 end = end & ~huge_page_mask(h); in hugetlbfs_zero_partial_page()
[all …]
/linux-5.19.10/arch/s390/mm/
Dhugetlbpage.c271 info.align_mask = PAGE_MASK & ~huge_page_mask(h); in hugetlb_get_unmapped_area_bottomup()
288 info.align_mask = PAGE_MASK & ~huge_page_mask(h); in hugetlb_get_unmapped_area_topdown()
316 if (len & ~huge_page_mask(h)) in hugetlb_get_unmapped_area()
/linux-5.19.10/arch/sparc/mm/
Dhugetlbpage.c43 info.align_mask = PAGE_MASK & ~huge_page_mask(h); in hugetlb_get_unmapped_area_bottomup()
75 info.align_mask = PAGE_MASK & ~huge_page_mask(h); in hugetlb_get_unmapped_area_topdown()
108 if (len & ~huge_page_mask(h)) in hugetlb_get_unmapped_area()
/linux-5.19.10/arch/parisc/mm/
Dhugetlbpage.c30 if (len & ~huge_page_mask(h)) in hugetlb_get_unmapped_area()
/linux-5.19.10/mm/
Dpagewalk.c289 unsigned long boundary = (addr & huge_page_mask(h)) + huge_page_size(h); in hugetlb_entry_end()
299 unsigned long hmask = huge_page_mask(h); in walk_hugetlb_range()
Dmremap.c955 if (addr & ~huge_page_mask(h)) in SYSCALL_DEFINE5()
957 if (new_addr & ~huge_page_mask(h)) in SYSCALL_DEFINE5()
Dhugetlb.c4617 if (addr & ~(huge_page_mask(hstate_vma(vma)))) in hugetlb_vm_op_split()
5009 BUG_ON(start & ~huge_page_mask(h)); in __unmap_hugepage_range()
5010 BUG_ON(end & ~huge_page_mask(h)); in __unmap_hugepage_range()
5175 address = address & huge_page_mask(h); in unmap_ref_private()
5229 unsigned long haddr = address & huge_page_mask(h); in hugetlb_wp()
5509 unsigned long haddr = address & huge_page_mask(h); in hugetlb_no_page()
5714 unsigned long haddr = address & huge_page_mask(h); in hugetlb_fault()
6154 pte = huge_pte_offset(mm, vaddr & huge_page_mask(h), in follow_hugetlb_page()
6234 pfn_offset = (vaddr & ~huge_page_mask(h)) >> PAGE_SHIFT; in follow_hugetlb_page()
Dmadvise.c808 if (start & ~huge_page_mask(hstate_vma(vma))) in madvise_dontneed_free_valid_vma()
/linux-5.19.10/include/linux/
Dhugetlb.h727 static inline unsigned long huge_page_mask(struct hstate *h) in huge_page_mask() function
1003 static inline unsigned long huge_page_mask(struct hstate *h)
/linux-5.19.10/drivers/dma-buf/
Dudmabuf.c229 ~huge_page_mask(hpstate)) >> PAGE_SHIFT; in udmabuf_create()
/linux-5.19.10/arch/riscv/kvm/
Dmmu.c642 gfn = (gpa & huge_page_mask(hstate_vma(vma))) >> PAGE_SHIFT; in kvm_riscv_gstage_map()