/linux-6.1.9/mm/damon/ |
D | vaddr-test.h | 17 static void __link_vmas(struct maple_tree *mt, struct vm_area_struct *vmas, in __link_vmas() argument 28 vma_mas_store(&vmas[i], &mas); in __link_vmas() 64 struct vm_area_struct vmas[] = { in damon_test_three_regions_in_vmas() local 74 __link_vmas(&mm.mm_mt, vmas, ARRAY_SIZE(vmas)); in damon_test_three_regions_in_vmas()
|
/linux-6.1.9/mm/ |
D | gup.c | 1165 struct vm_area_struct **vmas, int *locked) in __get_user_pages() argument 1205 i = follow_hugetlb_page(mm, vma, pages, vmas, in __get_user_pages() 1272 if (vmas) { in __get_user_pages() 1273 vmas[i] = vma; in __get_user_pages() 1405 struct vm_area_struct **vmas, in __get_user_pages_locked() argument 1414 BUG_ON(vmas); in __get_user_pages_locked() 1438 vmas, locked); in __get_user_pages_locked() 1722 struct vm_area_struct **vmas, int *locked, in __get_user_pages_locked() argument 1752 if (vmas) in __get_user_pages_locked() 1753 vmas[i] = vma; in __get_user_pages_locked() [all …]
|
D | hugetlb.c | 6359 struct vm_area_struct **vmas) in record_subpages_vmas() argument 6366 if (vmas) in record_subpages_vmas() 6367 vmas[nr] = vma; in record_subpages_vmas() 6391 struct page **pages, struct vm_area_struct **vmas, in follow_hugetlb_page() argument 6514 if (!pages && !vmas && !pfn_offset && in follow_hugetlb_page() 6528 if (pages || vmas) in follow_hugetlb_page() 6532 vmas ? vmas + i : NULL); in follow_hugetlb_page()
|
/linux-6.1.9/drivers/gpu/drm/msm/ |
D | msm_gem.c | 318 list_add_tail(&vma->list, &msm_obj->vmas); in add_vma() 331 list_for_each_entry(vma, &msm_obj->vmas, list) { in lookup_vma() 362 list_for_each_entry(vma, &msm_obj->vmas, list) { in put_iova_spaces() 380 list_for_each_entry_safe(vma, tmp, &msm_obj->vmas, list) { in put_iova_vmas() 916 if (!list_empty(&msm_obj->vmas)) { in msm_gem_describe() 920 list_for_each_entry(vma, &msm_obj->vmas, list) { in msm_gem_describe() 1091 INIT_LIST_HEAD(&msm_obj->vmas); in msm_gem_new_impl()
|
D | msm_gem.h | 111 struct list_head vmas; /* list of msm_gem_vma */ member
|
/linux-6.1.9/io_uring/ |
D | rsrc.c | 1128 struct vm_area_struct **vmas = NULL; in io_pin_pages() local 1140 vmas = kvmalloc_array(nr_pages, sizeof(struct vm_area_struct *), in io_pin_pages() 1142 if (!vmas) in io_pin_pages() 1148 pages, vmas); in io_pin_pages() 1152 struct vm_area_struct *vma = vmas[i]; in io_pin_pages() 1178 kvfree(vmas); in io_pin_pages()
|
/linux-6.1.9/drivers/video/fbdev/vermilion/ |
D | vermilion.h | 210 atomic_t vmas; member
|
/linux-6.1.9/Documentation/locking/ |
D | robust-futexes.rst | 58 FUTEX_RECOVER. At do_exit() time, all vmas are searched to see whether 69 microsecond on Linux, but with thousands (or tens of thousands) of vmas
|
/linux-6.1.9/include/linux/ |
D | mm.h | 1967 struct vm_area_struct **vmas, int *locked); 1971 struct vm_area_struct **vmas, int *locked); 1974 struct vm_area_struct **vmas); 1977 struct vm_area_struct **vmas);
|
D | hugetlb.h | 277 struct vm_area_struct **vmas, unsigned long *position, in follow_hugetlb_page() argument
|
/linux-6.1.9/Documentation/admin-guide/mm/ |
D | userfaultfd.rst | 38 operations never involve heavyweight structures like vmas (in fact the 42 Terabytes. Too many vmas would be needed for that.
|
/linux-6.1.9/Documentation/gpu/rfc/ |
D | i915_vm_bind.rst | 122 invalidated vmas (due to eviction and userptr invalidation) etc.
|