/kernel/linux/linux-5.10/include/linux/ |
H A D | nodemask.h | 390 enum node_states { enum 410 extern nodemask_t node_states[NR_NODE_STATES]; 413 static inline int node_state(int node, enum node_states state) in node_state() 415 return node_isset(node, node_states[state]); in node_state() 418 static inline void node_set_state(int node, enum node_states state) in node_set_state() 420 __node_set(node, &node_states[state]); in node_set_state() 423 static inline void node_clear_state(int node, enum node_states state) in node_clear_state() 425 __node_clear(node, &node_states[state]); in node_clear_state() 428 static inline int num_node_state(enum node_states state) in num_node_state() 430 return nodes_weight(node_states[stat in num_node_state() [all...] |
H A D | cpuset.h | 203 #define cpuset_current_mems_allowed (node_states[N_MEMORY])
|
/kernel/linux/linux-6.6/include/linux/ |
H A D | nodemask.h | 398 enum node_states { enum 418 extern nodemask_t node_states[NR_NODE_STATES]; 421 static inline int node_state(int node, enum node_states state) in node_state() 423 return node_isset(node, node_states[state]); in node_state() 426 static inline void node_set_state(int node, enum node_states state) in node_set_state() 428 __node_set(node, &node_states[state]); in node_set_state() 431 static inline void node_clear_state(int node, enum node_states state) in node_clear_state() 433 __node_clear(node, &node_states[state]); in node_clear_state() 436 static inline int num_node_state(enum node_states state) in num_node_state() 438 return nodes_weight(node_states[stat in num_node_state() [all...] |
H A D | cpuset.h | 215 #define cpuset_current_mems_allowed (node_states[N_MEMORY])
|
/kernel/linux/linux-6.6/mm/ |
H A D | memory-tiers.c | 392 nodes_andnot(tier_nodes, node_states[N_MEMORY], tier_nodes); in establish_demotion_targets() 423 nodes_and(tier_nodes, node_states[N_CPU], tier_nodes); in establish_demotion_targets() 440 lower_tier = node_states[N_MEMORY]; in establish_demotion_targets()
|
H A D | mempolicy.c | 243 cpuset_current_mems_allowed, node_states[N_MEMORY]); in mpol_set_nodemask() 1849 * policy->nodes is intersect with node_states[N_MEMORY]. in apply_policy_zone() 1853 if (!nodes_intersects(policy->nodes, node_states[N_HIGH_MEMORY])) in apply_policy_zone() 3023 if (!nodes_subset(nodes, node_states[N_MEMORY])) in mpol_parse_str() 3054 nodes = node_states[N_MEMORY]; in mpol_parse_str()
|
H A D | mm_init.c | 346 nodemask_t saved_node_state = node_states[N_MEMORY]; in find_zone_movable_pfns_for_nodes() 348 int usable_nodes = nodes_weight(node_states[N_MEMORY]); in find_zone_movable_pfns_for_nodes() 561 node_states[N_MEMORY] = saved_node_state; in find_zone_movable_pfns_for_nodes()
|
H A D | hugetlb.c | 2664 page = remove_pool_huge_page(h, &node_states[N_MEMORY], 1); in return_unused_surplus_pages() 3191 for_each_node_mask_to_alloc(h, nr_nodes, node, &node_states[N_MEMORY]) { in __alloc_bootmem_huge_page() 3260 &node_states[N_MEMORY], NULL); in hugetlb_hstate_alloc_pages_onenode() 3324 &node_states[N_MEMORY], in hugetlb_hstate_alloc_pages() 3763 n_mask = &node_states[N_MEMORY]; in __nr_hugepages_store_common() 3922 n_mask = &node_states[N_MEMORY]; in demote_store()
|
H A D | oom_kill.c | 286 !nodes_subset(node_states[N_MEMORY], *oc->nodemask)) { in constrained_alloc()
|
H A D | memory_hotplug.c | 681 /* check which state of node_states will be changed when online memory */ 1772 nodemask_t nmask = node_states[N_MEMORY]; in do_migrate_range() 1815 /* check which state of node_states will be changed when offline memory */ 1827 * Check whether node_states[N_NORMAL_MEMORY] will be changed. in node_states_check_changes_offline() 1832 * node_states[N_NORMAL_MEMORY]. in node_states_check_changes_offline()
|
/kernel/linux/linux-5.10/drivers/base/ |
H A D | node.c | 994 enum node_states state; 1003 nodemask_pr_args(&node_states[na->state])); in show_node_state()
|
/kernel/linux/linux-6.6/drivers/base/ |
H A D | node.c | 908 enum node_states state; 917 nodemask_pr_args(&node_states[na->state])); in show_node_state()
|
/kernel/linux/linux-5.10/mm/ |
H A D | mempolicy.c | 239 cpuset_current_mems_allowed, node_states[N_MEMORY]); in mpol_set_nodemask() 1855 * policy->v.nodes is intersect with node_states[N_MEMORY]. in apply_policy_zone() 1859 if (!nodes_intersects(policy->v.nodes, node_states[N_HIGH_MEMORY])) in apply_policy_zone() 2888 if (!nodes_subset(nodes, node_states[N_MEMORY])) in mpol_parse_str() 2919 nodes = node_states[N_MEMORY]; in mpol_parse_str()
|
H A D | oom_kill.c | 286 !nodes_subset(node_states[N_MEMORY], *oc->nodemask)) { in constrained_alloc()
|
H A D | page_alloc.c | 151 nodemask_t node_states[NR_NODE_STATES] __read_mostly = { 163 EXPORT_SYMBOL(node_states); variable 7372 nodemask_t saved_node_state = node_states[N_MEMORY]; in find_zone_movable_pfns_for_nodes() 7374 int usable_nodes = nodes_weight(node_states[N_MEMORY]); in find_zone_movable_pfns_for_nodes() 7577 node_states[N_MEMORY] = saved_node_state; in find_zone_movable_pfns_for_nodes()
|
H A D | hugetlb.c | 2155 if (!free_pool_huge_page(h, &node_states[N_MEMORY], 1)) in return_unused_surplus_pages() 2471 for_each_node_mask_to_alloc(h, nr_nodes, node, &node_states[N_MEMORY]) { in __alloc_bootmem_huge_page() 2559 &node_states[N_MEMORY], in hugetlb_hstate_alloc_pages() 2854 n_mask = &node_states[N_MEMORY]; in __nr_hugepages_store_common()
|
H A D | memory_hotplug.c | 629 /* check which state of node_states will be changed when online memory */ 1342 nodemask_t nmask = node_states[N_MEMORY]; in do_migrate_range() 1386 /* check which state of node_states will be changed when offline memory */ 1399 * Check whether node_states[N_NORMAL_MEMORY] will be changed. in node_states_check_changes_offline() 1404 * node_states[N_NORMAL_MEMORY]. in node_states_check_changes_offline() 1413 * node_states[N_HIGH_MEMORY] contains nodes which in node_states_check_changes_offline()
|
/kernel/linux/linux-5.10/kernel/cgroup/ |
H A D | cpuset.c | 435 * of node_states[N_MEMORY]. 441 while (!nodes_intersects(cs->effective_mems, node_states[N_MEMORY])) in guarantee_online_mems() 443 nodes_and(*pmask, cs->effective_mems, node_states[N_MEMORY]); in guarantee_online_mems() 3276 new_mems = node_states[N_MEMORY]; in cpuset_hotplug_workfn() 3375 * Keep top_cpuset.mems_allowed tracking node_states[N_MEMORY]. 3376 * Call this routine anytime after node_states[N_MEMORY] changes. 3406 top_cpuset.effective_mems = node_states[N_MEMORY]; in cpuset_init_smp() 3485 * subset of node_states[N_MEMORY], even if this means going outside the
|
/kernel/linux/linux-6.6/kernel/cgroup/ |
H A D | cpuset.c | 539 * of node_states[N_MEMORY]. 545 while (!nodes_intersects(cs->effective_mems, node_states[N_MEMORY])) in guarantee_online_mems() 547 nodes_and(*pmask, cs->effective_mems, node_states[N_MEMORY]); in guarantee_online_mems() 3796 new_mems = node_states[N_MEMORY]; in cpuset_hotplug_workfn() 3895 * Keep top_cpuset.mems_allowed tracking node_states[N_MEMORY]. 3896 * Call this routine anytime after node_states[N_MEMORY] changes. 3921 top_cpuset.effective_mems = node_states[N_MEMORY]; in cpuset_init_smp() 4031 * subset of node_states[N_MEMORY], even if this means going outside the
|
/kernel/linux/linux-5.10/kernel/ |
H A D | kthread.c | 659 set_mems_allowed(node_states[N_MEMORY]); in kthreadd()
|
/kernel/linux/linux-5.10/init/ |
H A D | main.c | 1501 set_mems_allowed(node_states[N_MEMORY]); in kernel_init_freeable()
|
/kernel/linux/linux-6.6/kernel/ |
H A D | kthread.c | 726 set_mems_allowed(node_states[N_MEMORY]); in kthreadd()
|
/kernel/linux/linux-6.6/init/ |
H A D | main.c | 1531 set_mems_allowed(node_states[N_MEMORY]); in kernel_init_freeable()
|
/kernel/linux/linux-5.10/fs/proc/ |
H A D | task_mmu.c | 1800 if (!node_isset(nid, node_states[N_MEMORY])) in can_gather_numa_stats() 1825 if (!node_isset(nid, node_states[N_MEMORY])) in can_gather_numa_stats_pmd()
|
/kernel/linux/linux-6.6/fs/proc/ |
H A D | task_mmu.c | 1848 if (!node_isset(nid, node_states[N_MEMORY])) in can_gather_numa_stats() 1873 if (!node_isset(nid, node_states[N_MEMORY])) in can_gather_numa_stats_pmd()
|