/kernel/linux/linux-5.10/drivers/nvme/host/ |
H A D | zns.c | 174 int ret, zone_idx = 0; in nvme_ns_report_zones() local 190 while (zone_idx < nr_zones && sector < get_capacity(ns->disk)) { in nvme_ns_report_zones() 205 for (i = 0; i < nz && zone_idx < nr_zones; i++) { in nvme_ns_report_zones() 207 zone_idx, cb, data); in nvme_ns_report_zones() 210 zone_idx++; in nvme_ns_report_zones() 216 if (zone_idx > 0) in nvme_ns_report_zones() 217 ret = zone_idx; in nvme_ns_report_zones()
|
/kernel/linux/linux-6.6/drivers/nvme/host/ |
H A D | zns.c | 181 int ret, zone_idx = 0; in nvme_ns_report_zones() local 200 while (zone_idx < nr_zones && sector < get_capacity(ns->disk)) { in nvme_ns_report_zones() 215 for (i = 0; i < nz && zone_idx < nr_zones; i++) { in nvme_ns_report_zones() 217 zone_idx, cb, data); in nvme_ns_report_zones() 220 zone_idx++; in nvme_ns_report_zones() 226 if (zone_idx > 0) in nvme_ns_report_zones() 227 ret = zone_idx; in nvme_ns_report_zones()
|
/kernel/linux/linux-5.10/include/trace/events/ |
H A D | oom.h | 46 __field( int, zone_idx) 57 __entry->zone_idx = zoneref->zone_idx; 67 __entry->node, __print_symbolic(__entry->zone_idx, ZONE_TYPE),
|
H A D | compaction.h | 214 __entry->idx = zone_idx(zone); 261 __entry->idx = zone_idx(zone);
|
/kernel/linux/linux-6.6/include/trace/events/ |
H A D | oom.h | 46 __field( int, zone_idx) 57 __entry->zone_idx = zoneref->zone_idx; 67 __entry->node, __print_symbolic(__entry->zone_idx, ZONE_TYPE),
|
H A D | compaction.h | 209 __entry->idx = zone_idx(zone); 256 __entry->idx = zone_idx(zone);
|
/kernel/linux/linux-6.6/mm/ |
H A D | show_mem.c | 173 int zone_idx; in node_has_managed_zones() local 174 for (zone_idx = 0; zone_idx <= max_zone_idx; zone_idx++) in node_has_managed_zones() 175 if (zone_managed_pages(pgdat->node_zones + zone_idx)) in node_has_managed_zones() 197 if (zone_idx(zone) > max_zone_idx) in show_free_areas() 298 if (zone_idx(zone) > max_zone_idx) in show_free_areas() 361 if (zone_idx(zone) > max_zone_idx) in show_free_areas()
|
H A D | mm_init.c | 916 int nid = zone_to_nid(zone), zone_id = zone_idx(zone); in memmap_init_zone_range() 972 unsigned long zone_idx, int nid, in __init_zone_device_page() 976 __init_single_page(page, pfn, zone_idx, nid); in __init_zone_device_page() 1038 unsigned long zone_idx, int nid, in memmap_init_compound() 1049 __init_zone_device_page(page, pfn, zone_idx, nid, pgmap); in memmap_init_compound() 1072 unsigned long zone_idx = zone_idx(zone); in memmap_init_zone_device() local 1076 if (WARN_ON_ONCE(!pgmap || zone_idx != ZONE_DEVICE)) in memmap_init_zone_device() 1092 __init_zone_device_page(page, pfn, zone_idx, nid, pgmap); in memmap_init_zone_device() 1097 memmap_init_compound(page, pfn, zone_idx, ni in memmap_init_zone_device() 971 __init_zone_device_page(struct page *page, unsigned long pfn, unsigned long zone_idx, int nid, struct dev_pagemap *pgmap) __init_zone_device_page() argument 1036 memmap_init_compound(struct page *head, unsigned long head_pfn, unsigned long zone_idx, int nid, struct dev_pagemap *pgmap, unsigned long nr_pages) memmap_init_compound() argument 1408 int zone_idx = zone_idx(zone) + 1; init_currently_empty_zone() local [all...] |
H A D | memory_hotplug.c | 692 if (zone_idx(zone) <= ZONE_NORMAL && !node_state(nid, N_NORMAL_MEMORY)) in node_states_check_changes_online() 783 memmap_init_range(nr_pages, nid, zone_idx(zone), start_pfn, 0, in move_pfn_range_to_zone() 797 if (zone_idx(zone) == ZONE_MOVABLE) { in auto_movable_stats_account_zone() 1072 const bool movable = zone_idx(zone) == ZONE_MOVABLE; in adjust_present_page_count() 1836 if (zone_idx(zone) <= ZONE_NORMAL && nr_pages >= present_pages) in node_states_check_changes_offline() 2284 if (page && zone_idx(page_zone(page)) == ZONE_MOVABLE) in try_offline_memory_block()
|
H A D | page_alloc.c | 2825 wakeup_kswapd(zone, 0, 0, zone_idx(zone)); in rmqueue() 3049 if (zone_idx(zone) != ZONE_NORMAL) in alloc_flags_nofragment() 4403 __count_zid_vm_events(PGALLOC, zone_idx(zone), nr_account); in __alloc_pages_bulk() 4841 zoneref->zone_idx = zone_idx(zone); in zoneref_set_zone() 4981 zonerefs->zone_idx = 0; in build_zonelists_in_node_order() 4996 zonerefs->zone_idx = 0; in build_thisnode_zonelists() 5096 zonerefs->zone_idx = 0; in build_zonelists() 5685 if (!is_highmem(zone) && zone_idx(zone) != ZONE_MOVABLE) in __setup_per_zone_wmarks() 5695 if (is_highmem(zone) || zone_idx(zon in __setup_per_zone_wmarks() [all...] |
H A D | page_isolation.c | 72 if (zone_idx(zone) == ZONE_MOVABLE) in has_unmovable_pages()
|
/kernel/linux/linux-6.6/drivers/scsi/ |
H A D | sd_zbc.c | 265 int zone_idx = 0; in sd_zbc_report_zones() local 280 while (zone_idx < nr_zones && lba < sdkp->capacity) { in sd_zbc_report_zones() 290 for (i = 0; i < nr && zone_idx < nr_zones; i++) { in sd_zbc_report_zones() 294 if ((zone_idx == 0 && in sd_zbc_report_zones() 297 (zone_idx > 0 && start_lba != lba) || in sd_zbc_report_zones() 301 zone_idx, lba, start_lba, zone_length); in sd_zbc_report_zones() 315 ret = sd_zbc_parse_report(sdkp, buf + offset, zone_idx, in sd_zbc_report_zones() 320 zone_idx++; in sd_zbc_report_zones() 324 ret = zone_idx; in sd_zbc_report_zones()
|
/kernel/linux/linux-5.10/drivers/scsi/ |
H A D | sd_zbc.c | 201 int zone_idx = 0; in sd_zbc_report_zones() local 216 while (zone_idx < nr_zones && sector < capacity) { in sd_zbc_report_zones() 227 for (i = 0; i < nr && zone_idx < nr_zones; i++) { in sd_zbc_report_zones() 229 ret = sd_zbc_parse_report(sdkp, buf + offset, zone_idx, in sd_zbc_report_zones() 233 zone_idx++; in sd_zbc_report_zones() 239 ret = zone_idx; in sd_zbc_report_zones()
|
/kernel/linux/linux-6.6/drivers/md/ |
H A D | dm-zone.c | 42 nr_zones - args.zone_idx); in dm_blk_do_report_zones() 45 } while (args.zone_idx < nr_zones && in dm_blk_do_report_zones() 48 return args.zone_idx; in dm_blk_do_report_zones() 104 return args->orig_cb(zone, args->zone_idx++, args->orig_data); in dm_report_zones_cb()
|
/kernel/linux/linux-6.6/include/linux/ |
H A D | mmzone.h | 1224 int zone_idx; /* zone_idx(zoneref->zone) */ member 1500 extern unsigned long lruvec_lru_size(struct lruvec *lruvec, enum lru_list lru, int zone_idx); 1509 * zone_idx() returns 0 for the ZONE_DMA zone, 1 for the ZONE_NORMAL zone, etc. 1511 #define zone_idx(zone) ((zone) - (zone)->zone_pgdat->node_zones) macro 1516 return zone_idx(zone) == ZONE_DEVICE; in zone_is_zone_device() 1582 return is_highmem_idx(zone_idx(zone)); in is_highmem() 1648 return zoneref->zone_idx; in zonelist_zone_idx()
|
H A D | memcontrol.h | 944 enum lru_list lru, int zone_idx) in mem_cgroup_get_zone_lru_size() 949 return READ_ONCE(mz->lru_zone_size[zone_idx][lru]); in mem_cgroup_get_zone_lru_size() 1465 enum lru_list lru, int zone_idx) in mem_cgroup_get_zone_lru_size() 943 mem_cgroup_get_zone_lru_size(struct lruvec *lruvec, enum lru_list lru, int zone_idx) mem_cgroup_get_zone_lru_size() argument 1464 mem_cgroup_get_zone_lru_size(struct lruvec *lruvec, enum lru_list lru, int zone_idx) mem_cgroup_get_zone_lru_size() argument
|
H A D | device-mapper.h | 514 unsigned int zone_idx; member
|
/kernel/linux/linux-5.10/mm/ |
H A D | memory_hotplug.c | 496 if (zone_idx(zone) == ZONE_DEVICE) in remove_pfn_range_from_zone() 641 if (zone_idx(zone) <= ZONE_NORMAL && !node_state(nid, N_NORMAL_MEMORY)) in node_states_check_changes_online() 644 if (zone_idx(zone) <= ZONE_HIGHMEM && !node_state(nid, N_HIGH_MEMORY)) in node_states_check_changes_online() 718 memmap_init_zone(nr_pages, nid, zone_idx(zone), start_pfn, 0, in move_pfn_range_to_zone() 1408 if (zone_idx(zone) <= ZONE_NORMAL && nr_pages >= present_pages) in node_states_check_changes_offline() 1421 if (zone_idx(zone) <= ZONE_HIGHMEM && nr_pages >= present_pages) in node_states_check_changes_offline() 1817 if (page && zone_idx(page_zone(page)) == ZONE_MOVABLE) in try_offline_memory_block()
|
H A D | page_alloc.c | 1785 int zid = zone_idx(zone); in deferred_init_pages() 3516 wakeup_kswapd(zone, 0, 0, zone_idx(zone)); in rmqueue() 3793 if (zone_idx(zone) != ZONE_NORMAL) in alloc_flags_nofragment() 5777 zoneref->zone_idx = zone_idx(zone); in zoneref_set_zone() 5919 zonerefs->zone_idx = 0; in build_zonelists_in_node_order() 5934 zonerefs->zone_idx = 0; in build_thisnode_zonelists() 6032 zonerefs->zone_idx = 0; in build_zonelists() 6279 unsigned long zone_idx = zone_idx(zon in memmap_init_zone_device() local 6662 int zone_idx = zone_idx(zone) + 1; init_currently_empty_zone() local [all...] |
/kernel/linux/linux-5.10/include/linux/ |
H A D | mmzone.h | 695 int zone_idx; /* zone_idx(zoneref->zone) */ member 918 extern unsigned long lruvec_lru_size(struct lruvec *lruvec, enum lru_list lru, int zone_idx); 927 * zone_idx() returns 0 for the ZONE_DMA zone, 1 for the ZONE_NORMAL zone, etc. 929 #define zone_idx(zone) ((zone) - (zone)->zone_pgdat->node_zones) macro 1008 return is_highmem_idx(zone_idx(zone)); in is_highmem() 1087 return zoneref->zone_idx; in zonelist_zone_idx()
|
H A D | memcontrol.h | 654 enum lru_list lru, int zone_idx) in mem_cgroup_get_zone_lru_size() 659 return READ_ONCE(mz->lru_zone_size[zone_idx][lru]); in mem_cgroup_get_zone_lru_size() 1146 enum lru_list lru, int zone_idx) in mem_cgroup_get_zone_lru_size() 653 mem_cgroup_get_zone_lru_size(struct lruvec *lruvec, enum lru_list lru, int zone_idx) mem_cgroup_get_zone_lru_size() argument 1145 mem_cgroup_get_zone_lru_size(struct lruvec *lruvec, enum lru_list lru, int zone_idx) mem_cgroup_get_zone_lru_size() argument
|
H A D | device-mapper.h | 458 unsigned int zone_idx; member
|
/kernel/linux/linux-6.6/drivers/block/ |
H A D | virtio_blk.c | 674 unsigned int zone_idx = 0; in virtblk_report_zones() local 691 while (zone_idx < nr_zones && sector < get_capacity(vblk->disk)) { in virtblk_report_zones() 704 for (i = 0; i < nz && zone_idx < nr_zones; i++) { in virtblk_report_zones() 706 zone_idx, cb, data); in virtblk_report_zones() 713 zone_idx++; in virtblk_report_zones() 717 if (zone_idx > 0) in virtblk_report_zones() 718 ret = zone_idx; in virtblk_report_zones()
|
/kernel/linux/linux-6.6/drivers/net/ethernet/sfc/ |
H A D | tc.h | 245 u8 zone_idx; /* for TABLE_FIELD_ID_DOMAIN */ member
|
/kernel/linux/linux-5.10/drivers/md/ |
H A D | dm.c | 468 return args->orig_cb(zone, args->zone_idx++, args->orig_data); in dm_report_zones_cb() 504 nr_zones - args.zone_idx); in dm_blk_report_zones() 507 } while (args.zone_idx < nr_zones && in dm_blk_report_zones() 510 ret = args.zone_idx; in dm_blk_report_zones()
|