/linux-5.19.10/drivers/gpu/drm/i915/selftests/ |
D | i915_gem_gtt.c | 241 u64 hole_start, u64 hole_end, in lowlevel_hole() argument 255 for (size = 12; (hole_end - hole_start) >> size; size++) { in lowlevel_hole() 262 hole_size = (hole_end - hole_start) >> aligned_size; in lowlevel_hole() 268 __func__, hole_start, hole_end, size, hole_size); in lowlevel_hole() 284 GEM_BUG_ON(hole_start + count * BIT_ULL(aligned_size) > hole_end); in lowlevel_hole() 315 hole_end = hole_start; /* quit */ in lowlevel_hole() 403 u64 hole_start, u64 hole_end, in fill_hole() argument 406 const u64 hole_size = hole_end - hole_start; in fill_hole() 432 { "top-down", hole_end, -1, }, in fill_hole() 484 if (offset + aligned_size > hole_end) in fill_hole() [all …]
|
/linux-5.19.10/drivers/gpu/drm/ |
D | drm_mm.c | 453 u64 hole_start, hole_end; in drm_mm_reserve_node() local 467 adj_end = hole_end = hole_start + hole->hole_size; in drm_mm_reserve_node() 485 if (end < hole_end) in drm_mm_reserve_node() 544 u64 hole_end = hole_start + hole->hole_size; in drm_mm_insert_node_in_range() local 551 if (mode == DRM_MM_INSERT_HIGH && hole_end <= range_start) in drm_mm_insert_node_in_range() 555 col_end = hole_end; in drm_mm_insert_node_in_range() 603 if (adj_start + size < hole_end) in drm_mm_insert_node_in_range() 785 u64 hole_start, hole_end; in drm_mm_scan_add_block() local 805 hole_end = __drm_mm_hole_node_end(hole); in drm_mm_scan_add_block() 808 col_end = hole_end; in drm_mm_scan_add_block() [all …]
|
/linux-5.19.10/arch/sparc/kernel/ |
D | adi_64.c | 163 unsigned long end_addr, hole_start, hole_end; in alloc_tag_store() local 168 hole_end = ULONG_MAX; in alloc_tag_store() 195 (tag_desc->start < hole_end)) in alloc_tag_store() 196 hole_end = tag_desc->start; in alloc_tag_store() 251 if (hole_end < end_addr) { in alloc_tag_store() 258 end_addr = hole_end - 1; in alloc_tag_store()
|
/linux-5.19.10/kernel/ |
D | kexec_core.c | 441 unsigned long hole_start, hole_end, size; in kimage_alloc_crash_control_pages() local 447 hole_end = hole_start + size - 1; in kimage_alloc_crash_control_pages() 448 while (hole_end <= crashk_res.end) { in kimage_alloc_crash_control_pages() 453 if (hole_end > KEXEC_CRASH_CONTROL_MEMORY_LIMIT) in kimage_alloc_crash_control_pages() 461 if ((hole_end >= mstart) && (hole_start <= mend)) { in kimage_alloc_crash_control_pages() 464 hole_end = hole_start + size - 1; in kimage_alloc_crash_control_pages() 471 image->control_page = hole_end; in kimage_alloc_crash_control_pages()
|
/linux-5.19.10/include/drm/ |
D | drm_mm.h | 394 #define drm_mm_for_each_hole(pos, mm, hole_start, hole_end) \ argument 399 hole_end = hole_start + pos->hole_size, \
|
/linux-5.19.10/drivers/gpu/drm/i915/gt/ |
D | intel_ggtt.c | 298 unsigned long hole_start, hole_end; in init_ggtt() local 361 drm_mm_for_each_hole(entry, &ggtt->vm.mm, hole_start, hole_end) { in init_ggtt() 364 hole_start, hole_end); in init_ggtt() 366 hole_end - hole_start); in init_ggtt()
|
/linux-5.19.10/fs/hugetlbfs/ |
D | inode.c | 632 loff_t hole_start, hole_end; in hugetlbfs_punch_hole() local 638 hole_end = round_down(offset + len, hpage_size); in hugetlbfs_punch_hole() 656 if (hole_end > hole_start) { in hugetlbfs_punch_hole() 660 hole_end >> PAGE_SHIFT, 0); in hugetlbfs_punch_hole() 664 if ((offset + len) > hole_end && (offset + len) > hole_start) in hugetlbfs_punch_hole() 666 hole_end, offset + len); in hugetlbfs_punch_hole() 671 if (hole_end > hole_start) in hugetlbfs_punch_hole() 672 remove_inode_hugepages(inode, hole_start, hole_end); in hugetlbfs_punch_hole()
|
/linux-5.19.10/drivers/gpu/drm/selftests/ |
D | test-drm_mm.c | 57 u64 hole_start, __always_unused hole_end; in assert_no_holes() local 61 drm_mm_for_each_hole(hole, mm, hole_start, hole_end) in assert_no_holes() 81 u64 hole_start, hole_end; in assert_one_hole() local 89 drm_mm_for_each_hole(hole, mm, hole_start, hole_end) { in assert_one_hole() 90 if (start != hole_start || end != hole_end) { in assert_one_hole() 93 hole_start, hole_end, in assert_one_hole() 1269 u64 hole_start, hole_end; in show_holes() local 1272 drm_mm_for_each_hole(hole, mm, hole_start, hole_end) { in show_holes() 1288 hole_start, hole_end, hole_end - hole_start, in show_holes()
|
/linux-5.19.10/fs/btrfs/ |
D | zoned.h | 48 u64 hole_end, u64 num_bytes); 145 u64 hole_start, u64 hole_end, in btrfs_find_allocatable_zones() argument
|
D | zoned.c | 1020 u64 hole_end, u64 num_bytes) in btrfs_find_allocatable_zones() argument 1033 while (pos < hole_end) { in btrfs_find_allocatable_zones() 1038 return hole_end; in btrfs_find_allocatable_zones()
|
D | volumes.c | 1475 u64 hole_end = *hole_start + *hole_size; in dev_extent_hole_check() local 1483 if (hole_end >= *hole_start) in dev_extent_hole_check() 1484 *hole_size = hole_end - *hole_start; in dev_extent_hole_check()
|
D | inode.c | 7137 const u64 hole_end = extent_map_end(hole_em); in btrfs_get_extent_fiemap() local 7153 if (hole_end <= start || hole_em->start > end) { in btrfs_get_extent_fiemap() 7158 hole_len = hole_end - hole_start; in btrfs_get_extent_fiemap()
|
/linux-5.19.10/fs/proc/ |
D | task_mmu.c | 1366 unsigned long hole_end; in pagemap_pte_hole() local 1369 hole_end = min(end, vma->vm_start); in pagemap_pte_hole() 1371 hole_end = end; in pagemap_pte_hole() 1373 for (; addr < hole_end; addr += PAGE_SIZE) { in pagemap_pte_hole()
|