Searched refs:zone_end_pfn (Results 1 – 13 of 13) sorted by relevance
/linux-6.6.21/arch/x86/mm/ |
D | highmem_32.c | 19 unsigned long zone_start_pfn, zone_end_pfn; in set_highmem_pages_init() local 25 zone_end_pfn = zone_start_pfn + zone->spanned_pages; in set_highmem_pages_init() 29 zone->name, nid, zone_start_pfn, zone_end_pfn); in set_highmem_pages_init() 32 zone_end_pfn); in set_highmem_pages_init()
|
/linux-6.6.21/mm/ |
D | mm_init.c | 849 unsigned long start_pfn, unsigned long zone_end_pfn, in memmap_init_range() argument 885 if (defer_init(nid, pfn, zone_end_pfn)) { in memmap_init_range() 915 unsigned long zone_end_pfn = zone_start_pfn + zone->spanned_pages; in memmap_init_zone_range() local 918 start_pfn = clamp(start_pfn, zone_start_pfn, zone_end_pfn); in memmap_init_zone_range() 919 end_pfn = clamp(end_pfn, zone_start_pfn, zone_end_pfn); in memmap_init_zone_range() 925 zone_end_pfn, MEMINIT_EARLY, NULL, MIGRATE_MOVABLE); in memmap_init_zone_range() 1120 unsigned long *zone_end_pfn) in adjust_zone_range_for_zone_movable() argument 1127 *zone_end_pfn = min(node_end_pfn, in adjust_zone_range_for_zone_movable() 1133 *zone_end_pfn > zone_movable_pfn[nid]) { in adjust_zone_range_for_zone_movable() 1134 *zone_end_pfn = zone_movable_pfn[nid]; in adjust_zone_range_for_zone_movable() [all …]
|
D | shuffle.c | 84 unsigned long end_pfn = zone_end_pfn(z); in __shuffle_zone()
|
D | memory_hotplug.c | 473 zone_end_pfn(zone)); in shrink_zone_span() 475 zone->spanned_pages = zone_end_pfn(zone) - pfn; in shrink_zone_span() 481 } else if (zone_end_pfn(zone) == end_pfn) { in shrink_zone_span() 506 unsigned long end_pfn = zone_end_pfn(zone); in update_pgdat_span() 707 unsigned long old_end_pfn = zone_end_pfn(zone); in resize_zone_range()
|
D | compaction.c | 229 pageblock_start_pfn(zone_end_pfn(zone) - 1); in reset_cached_positions() 344 block_pfn = min(block_pfn, zone_end_pfn(zone) - 1); in __reset_isolation_pfn() 379 unsigned long free_pfn = zone_end_pfn(zone) - 1; in __reset_isolation_suitable() 1437 end_pfn = min(pageblock_end_pfn(pfn), zone_end_pfn(cc->zone)); in fast_isolate_around() 1613 zone_end_pfn(cc->zone)), in fast_isolate_freepages() 1668 zone_end_pfn(zone)); in isolate_freepages() 2385 unsigned long end_pfn = zone_end_pfn(cc->zone); in compact_zone()
|
D | page_owner.c | 282 unsigned long end_pfn = zone_end_pfn(zone); in pagetypeinfo_showmixedcount_print() 624 unsigned long end_pfn = zone_end_pfn(zone); in init_pages_in_zone()
|
D | memblock.c | 1339 if (zone_end_pfn(zone) <= spfn) { in __next_mem_pfn_range_in_zone() 1347 *out_epfn = min(zone_end_pfn(zone), epfn); in __next_mem_pfn_range_in_zone()
|
D | kmemleak.c | 1581 unsigned long end_pfn = zone_end_pfn(zone); in kmemleak_scan()
|
D | vmstat.c | 1556 unsigned long end_pfn = zone_end_pfn(zone); in pagetypeinfo_showblockcount_print()
|
D | huge_memory.c | 2928 max_zone_pfn = zone_end_pfn(zone); in split_huge_pages_all()
|
/linux-6.6.21/include/linux/ |
D | mmzone.h | 1021 static inline unsigned long zone_end_pfn(const struct zone *zone) in zone_end_pfn() function 1028 return zone->zone_start_pfn <= pfn && pfn < zone_end_pfn(zone); in zone_spans_pfn() 1161 if (start_pfn >= zone_end_pfn(zone) || in zone_intersects()
|
/linux-6.6.21/arch/arm64/kernel/ |
D | hibernate.c | 266 max_zone_pfn = zone_end_pfn(zone); in swsusp_mte_save_tags()
|
/linux-6.6.21/kernel/power/ |
D | snapshot.c | 637 zone_end = zone_end_pfn(zone); in create_mem_extents() 1258 max_zone_pfn = zone_end_pfn(zone); in mark_free_pages() 1359 max_zone_pfn = zone_end_pfn(zone); in count_highmem_pages() 1426 max_zone_pfn = zone_end_pfn(zone); in count_data_pages() 1539 max_zone_pfn = zone_end_pfn(zone); in copy_data_pages()
|