/linux-6.1.9/include/linux/ |
D | memory_hotplug.h | 29 extern pg_data_t *arch_alloc_nodedata(int nid); 30 extern void arch_refresh_nodedata(int nid, pg_data_t *pgdat); 47 extern pg_data_t *node_data[]; 48 static inline void arch_refresh_nodedata(int nid, pg_data_t *pgdat) in arch_refresh_nodedata() 56 static inline pg_data_t *generic_alloc_nodedata(int nid) in generic_alloc_nodedata() 61 static inline void arch_refresh_nodedata(int nid, pg_data_t *pgdat) in arch_refresh_nodedata() 211 static inline void pgdat_kswapd_lock(pg_data_t *pgdat) in pgdat_kswapd_lock() 216 static inline void pgdat_kswapd_unlock(pg_data_t *pgdat) in pgdat_kswapd_unlock() 221 static inline void pgdat_kswapd_lock_init(pg_data_t *pgdat) in pgdat_kswapd_lock_init() 263 static inline void pgdat_kswapd_lock(pg_data_t *pgdat) {} in pgdat_kswapd_lock() [all …]
|
D | compaction.h | 98 extern void reset_isolation_suitable(pg_data_t *pgdat); 182 extern void wakeup_kcompactd(pg_data_t *pgdat, int order, int highest_zoneidx); 185 static inline void reset_isolation_suitable(pg_data_t *pgdat) in reset_isolation_suitable() 222 static inline void wakeup_kcompactd(pg_data_t *pgdat, in wakeup_kcompactd()
|
D | memory-tiers.h | 42 void node_get_allowed_targets(pg_data_t *pgdat, nodemask_t *targets); 50 static inline void node_get_allowed_targets(pg_data_t *pgdat, nodemask_t *targets) in node_get_allowed_targets() 92 static inline void node_get_allowed_targets(pg_data_t *pgdat, nodemask_t *targets) in node_get_allowed_targets()
|
D | page_owner.h | 19 pg_data_t *pgdat, struct zone *zone);
|
/linux-6.1.9/mm/ |
D | vmstat.c | 317 void set_pgdat_percpu_threshold(pg_data_t *pgdat, in set_pgdat_percpu_threshold() 1407 pg_data_t *pgdat; in frag_start() 1420 pg_data_t *pgdat = (pg_data_t *)arg; in frag_next() 1434 static void walk_zones_in_node(struct seq_file *m, pg_data_t *pgdat, in walk_zones_in_node() argument 1436 void (*print)(struct seq_file *m, pg_data_t *, struct zone *)) in walk_zones_in_node() 1456 static void frag_show_print(struct seq_file *m, pg_data_t *pgdat, in frag_show_print() 1476 pg_data_t *pgdat = (pg_data_t *)arg; in frag_show() 1482 pg_data_t *pgdat, struct zone *zone) in pagetypeinfo_showfree_print() 1527 pg_data_t *pgdat = (pg_data_t *)arg; in pagetypeinfo_showfree() 1539 pg_data_t *pgdat, struct zone *zone) in pagetypeinfo_showblockcount_print() [all …]
|
D | shuffle.h | 11 extern void __shuffle_free_memory(pg_data_t *pgdat); 13 static inline void __meminit shuffle_free_memory(pg_data_t *pgdat) in shuffle_free_memory() 40 static inline void shuffle_free_memory(pg_data_t *pgdat) in shuffle_free_memory()
|
D | compaction.c | 377 void reset_isolation_suitable(pg_data_t *pgdat) in reset_isolation_suitable() 748 static bool too_many_isolated(pg_data_t *pgdat) in too_many_isolated() 789 pg_data_t *pgdat = cc->zone->zone_pgdat; in isolate_migratepages_block() 1976 static bool kswapd_is_running(pg_data_t *pgdat) in kswapd_is_running() 2021 static unsigned int fragmentation_score_node(pg_data_t *pgdat) in fragmentation_score_node() 2036 static unsigned int fragmentation_score_wmark(pg_data_t *pgdat, bool low) in fragmentation_score_wmark() 2049 static bool should_proactive_compact_node(pg_data_t *pgdat) in should_proactive_compact_node() 2088 pg_data_t *pgdat; in __compact_finished() 2646 static void proactive_compact_node(pg_data_t *pgdat) in proactive_compact_node() 2676 pg_data_t *pgdat = NODE_DATA(nid); in compact_node() [all …]
|
D | memory-tiers.c | 227 pg_data_t *pgdat; in __node_get_memory_tier() 245 pg_data_t *pgdat; in node_is_toptier() 267 void node_get_allowed_targets(pg_data_t *pgdat, nodemask_t *targets) in node_get_allowed_targets() 480 pg_data_t *pgdat = NODE_DATA(node); in set_node_memory_tier() 507 pg_data_t *pgdat; in clear_node_memory_tier()
|
D | internal.h | 62 void __acct_reclaim_writeback(pg_data_t *pgdat, struct folio *folio, 66 pg_data_t *pgdat = folio_pgdat(folio); in acct_reclaim_writeback() 73 static inline void wake_throttle_isolated(pg_data_t *pgdat) in wake_throttle_isolated() 187 extern void reclaim_throttle(pg_data_t *pgdat, enum vmscan_throttle_state reason);
|
D | page_owner.c | 274 pg_data_t *pgdat, struct zone *zone) in pagetypeinfo_showmixedcount_print() 619 static void init_pages_in_zone(pg_data_t *pgdat, struct zone *zone) in init_pages_in_zone() 688 static void init_zones_in_node(pg_data_t *pgdat) in init_zones_in_node() 703 pg_data_t *pgdat; in init_early_allocated_pages()
|
D | vmscan.c | 1082 static bool skip_throttle_noprogress(pg_data_t *pgdat) in skip_throttle_noprogress() 1115 void reclaim_throttle(pg_data_t *pgdat, enum vmscan_throttle_state reason) in reclaim_throttle() 1189 void __acct_reclaim_writeback(pg_data_t *pgdat, struct folio *folio, in __acct_reclaim_writeback() 2785 static void prepare_scan_count(pg_data_t *pgdat, struct scan_control *sc) in prepare_scan_count() 6066 static void shrink_node_memcgs(pg_data_t *pgdat, struct scan_control *sc) in shrink_node_memcgs() 6124 static void shrink_node(pg_data_t *pgdat, struct scan_control *sc) in shrink_node() 6262 static void consider_reclaim_throttle(pg_data_t *pgdat, struct scan_control *sc) in consider_reclaim_throttle() 6307 pg_data_t *last_pgdat = NULL; in shrink_zones() 6308 pg_data_t *first_pgdat = NULL; in shrink_zones() 6392 static void snapshot_refaults(struct mem_cgroup *target_memcg, pg_data_t *pgdat) in snapshot_refaults() [all …]
|
D | mmzone.c | 32 pg_data_t *pgdat = zone->zone_pgdat; in next_zone()
|
D | shuffle.c | 153 void __meminit __shuffle_free_memory(pg_data_t *pgdat) in __shuffle_free_memory()
|
D | workingset.c | 187 static void *pack_shadow(int memcgid, pg_data_t *pgdat, unsigned long eviction, in pack_shadow() 198 static void unpack_shadow(void *shadow, int *memcgidp, pg_data_t **pgdat, in unpack_shadow()
|
D | mm_init.c | 31 pg_data_t *pgdat = NODE_DATA(nid); in mminit_verify_zonelist()
|
/linux-6.1.9/arch/ia64/mm/ |
D | discontig.c | 48 pg_data_t *pgdat_list[MAX_NUMNODES]; 120 pernodesize += L1_CACHE_ALIGN(sizeof(pg_data_t)); in compute_pernodesize() 122 pernodesize += L1_CACHE_ALIGN(sizeof(pg_data_t)); in compute_pernodesize() 274 pernode += L1_CACHE_ALIGN(sizeof(pg_data_t)); in fill_pernode() 278 pernode += L1_CACHE_ALIGN(sizeof(pg_data_t)); in fill_pernode() 372 pg_data_t **dst; in scatter_node_data() 611 pg_data_t * __init arch_alloc_nodedata(int nid) in arch_alloc_nodedata() 618 void arch_refresh_nodedata(int update_node, pg_data_t *update_pgdat) in arch_refresh_nodedata()
|
/linux-6.1.9/arch/mips/loongson64/ |
D | numa.c | 91 const size_t nd_size = roundup(sizeof(pg_data_t), SMP_CACHE_BYTES); in node_mem_init() 200 pg_data_t * __init arch_alloc_nodedata(int nid) in arch_alloc_nodedata() 202 return memblock_alloc(sizeof(pg_data_t), SMP_CACHE_BYTES); in arch_alloc_nodedata() 205 void arch_refresh_nodedata(int nid, pg_data_t *pgdat) in arch_refresh_nodedata()
|
/linux-6.1.9/arch/s390/kernel/ |
D | numa.c | 28 NODE_DATA(nid) = memblock_alloc(sizeof(pg_data_t), 8); in numa_setup() 31 __func__, sizeof(pg_data_t), 8); in numa_setup()
|
/linux-6.1.9/arch/mips/sgi-ip27/ |
D | ip27-memory.c | 426 pg_data_t * __init arch_alloc_nodedata(int nid) in arch_alloc_nodedata() 428 return memblock_alloc(sizeof(pg_data_t), SMP_CACHE_BYTES); in arch_alloc_nodedata() 431 void arch_refresh_nodedata(int nid, pg_data_t *pgdat) in arch_refresh_nodedata()
|
/linux-6.1.9/tools/testing/memblock/linux/ |
D | mmzone.h | 35 } pg_data_t; typedef
|
/linux-6.1.9/lib/ |
D | show_mem.c | 13 pg_data_t *pgdat; in __show_mem()
|
/linux-6.1.9/arch/ia64/include/asm/ |
D | numa.h | 27 extern pg_data_t *pgdat_list[MAX_NUMNODES];
|
/linux-6.1.9/drivers/base/ |
D | arch_numa.c | 221 const size_t nd_size = roundup(sizeof(pg_data_t), SMP_CACHE_BYTES); in setup_node_data() 244 memset(NODE_DATA(nid), 0, sizeof(pg_data_t)); in setup_node_data()
|
/linux-6.1.9/arch/loongarch/kernel/ |
D | numa.c | 229 size_t nd_sz = roundup(sizeof(pg_data_t), PAGE_SIZE); in alloc_node_data() 240 memset(nd, 0, sizeof(pg_data_t)); in alloc_node_data()
|
/linux-6.1.9/arch/x86/kernel/ |
D | setup_percpu.c | 70 pg_data_t *last = NULL; in pcpu_need_numa()
|