/linux-6.6.21/mm/ |
D | show_mem.c | 173 int zone_idx; in node_has_managed_zones() local 174 for (zone_idx = 0; zone_idx <= max_zone_idx; zone_idx++) in node_has_managed_zones() 175 if (zone_managed_pages(pgdat->node_zones + zone_idx)) in node_has_managed_zones() 197 if (zone_idx(zone) > max_zone_idx) in show_free_areas() 298 if (zone_idx(zone) > max_zone_idx) in show_free_areas() 361 if (zone_idx(zone) > max_zone_idx) in show_free_areas()
|
D | mm_init.c | 916 int nid = zone_to_nid(zone), zone_id = zone_idx(zone); in memmap_init_zone_range() 972 unsigned long zone_idx, int nid, in __init_zone_device_page() argument 976 __init_single_page(page, pfn, zone_idx, nid); in __init_zone_device_page() 1038 unsigned long zone_idx, int nid, in memmap_init_compound() argument 1049 __init_zone_device_page(page, pfn, zone_idx, nid, pgmap); in memmap_init_compound() 1072 unsigned long zone_idx = zone_idx(zone); in memmap_init_zone_device() local 1076 if (WARN_ON_ONCE(!pgmap || zone_idx != ZONE_DEVICE)) in memmap_init_zone_device() 1092 __init_zone_device_page(page, pfn, zone_idx, nid, pgmap); in memmap_init_zone_device() 1097 memmap_init_compound(page, pfn, zone_idx, nid, pgmap, in memmap_init_zone_device() 1402 int zone_idx = zone_idx(zone) + 1; in init_currently_empty_zone() local [all …]
|
D | memory_hotplug.c | 691 if (zone_idx(zone) <= ZONE_NORMAL && !node_state(nid, N_NORMAL_MEMORY)) in node_states_check_changes_online() 782 memmap_init_range(nr_pages, nid, zone_idx(zone), start_pfn, 0, in move_pfn_range_to_zone() 796 if (zone_idx(zone) == ZONE_MOVABLE) { in auto_movable_stats_account_zone() 1071 const bool movable = zone_idx(zone) == ZONE_MOVABLE; in adjust_present_page_count() 1832 if (zone_idx(zone) <= ZONE_NORMAL && nr_pages >= present_pages) in node_states_check_changes_offline() 2277 if (page && zone_idx(page_zone(page)) == ZONE_MOVABLE) in try_offline_memory_block()
|
D | page_alloc.c | 2791 wakeup_kswapd(zone, 0, 0, zone_idx(zone)); in rmqueue() 3015 if (zone_idx(zone) != ZONE_NORMAL) in alloc_flags_nofragment() 4364 __count_zid_vm_events(PGALLOC, zone_idx(zone), nr_account); in __alloc_pages_bulk() 4802 zoneref->zone_idx = zone_idx(zone); in zoneref_set_zone() 4942 zonerefs->zone_idx = 0; in build_zonelists_in_node_order() 4957 zonerefs->zone_idx = 0; in build_thisnode_zonelists() 5057 zonerefs->zone_idx = 0; in build_zonelists() 5646 if (!is_highmem(zone) && zone_idx(zone) != ZONE_MOVABLE) in __setup_per_zone_wmarks() 5656 if (is_highmem(zone) || zone_idx(zone) == ZONE_MOVABLE) { in __setup_per_zone_wmarks()
|
D | page_isolation.c | 72 if (zone_idx(zone) == ZONE_MOVABLE) in has_unmovable_pages()
|
D | migrate.c | 2022 zidx = zone_idx(folio_zone(src)); in alloc_migration_target()
|
D | vmscan.c | 655 int zone_idx) in lruvec_lru_size() argument 660 for (zid = 0; zid <= zone_idx; zid++) { in lruvec_lru_size() 7036 if (zone_idx(zone) > ZONE_NORMAL) in throttle_direct_reclaim()
|
D | khugepaged.c | 2589 if (zone_idx(zone) > gfp_zone(GFP_USER)) in set_recommended_min_free_kbytes()
|
/linux-6.6.21/drivers/nvme/host/ |
D | zns.c | 181 int ret, zone_idx = 0; in nvme_ns_report_zones() local 200 while (zone_idx < nr_zones && sector < get_capacity(ns->disk)) { in nvme_ns_report_zones() 215 for (i = 0; i < nz && zone_idx < nr_zones; i++) { in nvme_ns_report_zones() 217 zone_idx, cb, data); in nvme_ns_report_zones() 220 zone_idx++; in nvme_ns_report_zones() 226 if (zone_idx > 0) in nvme_ns_report_zones() 227 ret = zone_idx; in nvme_ns_report_zones()
|
/linux-6.6.21/include/trace/events/ |
D | oom.h | 46 __field( int, zone_idx) 57 __entry->zone_idx = zoneref->zone_idx; 67 __entry->node, __print_symbolic(__entry->zone_idx, ZONE_TYPE),
|
D | compaction.h | 209 __entry->idx = zone_idx(zone); 256 __entry->idx = zone_idx(zone);
|
/linux-6.6.21/drivers/scsi/ |
D | sd_zbc.c | 265 int zone_idx = 0; in sd_zbc_report_zones() local 280 while (zone_idx < nr_zones && lba < sdkp->capacity) { in sd_zbc_report_zones() 290 for (i = 0; i < nr && zone_idx < nr_zones; i++) { in sd_zbc_report_zones() 294 if ((zone_idx == 0 && in sd_zbc_report_zones() 297 (zone_idx > 0 && start_lba != lba) || in sd_zbc_report_zones() 301 zone_idx, lba, start_lba, zone_length); in sd_zbc_report_zones() 315 ret = sd_zbc_parse_report(sdkp, buf + offset, zone_idx, in sd_zbc_report_zones() 320 zone_idx++; in sd_zbc_report_zones() 324 ret = zone_idx; in sd_zbc_report_zones()
|
/linux-6.6.21/drivers/md/ |
D | dm-zone.c | 42 nr_zones - args.zone_idx); in dm_blk_do_report_zones() 45 } while (args.zone_idx < nr_zones && in dm_blk_do_report_zones() 48 return args.zone_idx; in dm_blk_do_report_zones() 104 return args->orig_cb(zone, args->zone_idx++, args->orig_data); in dm_report_zones_cb()
|
/linux-6.6.21/include/linux/ |
D | mmzone.h | 1196 int zone_idx; /* zone_idx(zoneref->zone) */ member 1463 #define zone_idx(zone) ((zone) - (zone)->zone_pgdat->node_zones) macro 1468 return zone_idx(zone) == ZONE_DEVICE; in zone_is_zone_device() 1534 return is_highmem_idx(zone_idx(zone)); in is_highmem() 1600 return zoneref->zone_idx; in zonelist_zone_idx()
|
D | memcontrol.h | 915 enum lru_list lru, int zone_idx) in mem_cgroup_get_zone_lru_size() argument 920 return READ_ONCE(mz->lru_zone_size[zone_idx][lru]); in mem_cgroup_get_zone_lru_size() 1416 enum lru_list lru, int zone_idx) in mem_cgroup_get_zone_lru_size() argument
|
D | device-mapper.h | 514 unsigned int zone_idx; member
|
/linux-6.6.21/drivers/block/ |
D | virtio_blk.c | 674 unsigned int zone_idx = 0; in virtblk_report_zones() local 691 while (zone_idx < nr_zones && sector < get_capacity(vblk->disk)) { in virtblk_report_zones() 704 for (i = 0; i < nz && zone_idx < nr_zones; i++) { in virtblk_report_zones() 706 zone_idx, cb, data); in virtblk_report_zones() 713 zone_idx++; in virtblk_report_zones() 717 if (zone_idx > 0) in virtblk_report_zones() 718 ret = zone_idx; in virtblk_report_zones()
|
/linux-6.6.21/drivers/net/ethernet/sfc/ |
D | tc.h | 245 u8 zone_idx; /* for TABLE_FIELD_ID_DOMAIN */ member
|