/linux-master/mm/ |
H A D | memory-tiers.c | 343 for_each_node_state(node, N_MEMORY) { 366 for_each_node_state(node, N_MEMORY) { 402 for_each_node_state(node, N_MEMORY) { 420 nodes_andnot(tier_nodes, node_states[N_MEMORY], tier_nodes); 468 lower_tier = node_states[N_MEMORY]; 514 if (!node_state(node, N_MEMORY)) 834 * Look at all the existing N_MEMORY nodes and add them to 838 for_each_node_state(node, N_MEMORY) {
|
H A D | mm_init.c | 295 * Populate N_MEMORY for calculating usable_nodes. 308 node_set_state(nid, N_MEMORY); 346 nodemask_t saved_node_state = node_states[N_MEMORY]; 348 int usable_nodes = nodes_weight(node_states[N_MEMORY]); 461 for_each_node_state(nid, N_MEMORY) { 561 node_states[N_MEMORY] = saved_node_state; 645 if (num_node_state(N_MEMORY) == 1) 1883 * N_MEMORY and not marked online so that no sysfs 1897 node_set_state(nid, N_MEMORY); 2378 /* There will be num_node_state(N_MEMORY) thread [all...] |
H A D | page_ext.c | 450 for_each_node_state(nid, N_MEMORY) {
|
H A D | hugetlb_cgroup.c | 471 for_each_node_state(nid, N_MEMORY) 476 for_each_node_state(nid, N_MEMORY) 494 for_each_node_state(nid, N_MEMORY) {
|
H A D | memory_hotplug.c | 690 if (!node_state(nid, N_MEMORY)) 702 node_set_state(node, N_MEMORY); 1840 nodemask_t nmask = node_states[N_MEMORY]; 1914 * it for N_MEMORY as well. 1928 node_clear_state(node, N_MEMORY);
|
H A D | hugetlb.c | 2752 folio = remove_pool_hugetlb_folio(h, &node_states[N_MEMORY], 1); 3302 for_each_node_mask_to_alloc(&h->next_nid_to_alloc, nr_nodes, node, &node_states[N_MEMORY]) { 3458 .size = num_node_state(N_MEMORY), 3461 .max_threads = num_node_state(N_MEMORY), 3482 &node_states[N_MEMORY], NULL); 3538 struct folio *folio = alloc_pool_huge_folio(h, &node_states[N_MEMORY], 3576 * job.max_threads is twice the num_node_state(N_MEMORY), 3594 job.max_threads = num_node_state(N_MEMORY) * 2; 3595 job.min_chunk = h->max_huge_pages / num_node_state(N_MEMORY) / 2; 4098 n_mask = &node_states[N_MEMORY]; [all...] |
H A D | mempolicy.c | 275 /* Check N_MEMORY */ 277 cpuset_current_mems_allowed, node_states[N_MEMORY]); 1866 * policy->nodes is intersect with node_states[N_MEMORY]. 3091 for_each_node_state(nid, N_MEMORY) { 3163 if (!nodes_subset(nodes, node_states[N_MEMORY])) 3195 nodes = node_states[N_MEMORY];
|
H A D | vmstat.c | 1636 if (!node_state(pgdat->node_id, N_MEMORY)) 2207 if (!node_state(pgdat->node_id, N_MEMORY))
|
H A D | oom_kill.c | 287 !nodes_subset(node_states[N_MEMORY], *oc->nodemask)) {
|
H A D | compaction.c | 3280 for_each_node_state(nid, N_MEMORY) { 3364 for_each_node_state(nid, N_MEMORY)
|
H A D | vmscan.c | 2789 for_each_node_state(nid, N_MEMORY) { 5185 for_each_node_state(nid, N_MEMORY) { 5396 if (nid < 0 || nid >= MAX_NUMNODES || !node_state(nid, N_MEMORY)) 7310 for_each_node_state(nid, N_MEMORY)
|
H A D | memcontrol.c | 4235 for_each_node_state(nid, N_MEMORY) 4247 for_each_node_state(nid, N_MEMORY) 5815 for_each_node_state(nid, N_MEMORY) { 6924 for_each_node_state(nid, N_MEMORY) {
|
H A D | page_alloc.c | 201 [N_MEMORY] = { { [0] = 1UL } }, 5040 if (!node_isset(node, *used_node_mask) && node_state(node, N_MEMORY)) { 5045 for_each_node_state(n, N_MEMORY) {
|
H A D | migrate.c | 2212 if (!node_state(node, N_MEMORY))
|
/linux-master/include/linux/ |
H A D | nodemask.h | 407 N_MEMORY, /* The node has memory(regular, high, movable) */ enumerator in enum:node_states 445 #define first_memory_node first_node(node_states[N_MEMORY]) 452 return next_node(nid, node_states[N_MEMORY]);
|
H A D | cpuset.h | 216 #define cpuset_current_mems_allowed (node_states[N_MEMORY])
|
/linux-master/drivers/base/ |
H A D | node.c | 931 [N_MEMORY] = _NODE_ATTR(has_memory, N_MEMORY), 944 &node_state_attr[N_MEMORY].attr.attr,
|
/linux-master/fs/proc/ |
H A D | kcore.c | 241 for_each_node_state(nid, N_MEMORY) {
|
H A D | task_mmu.c | 2591 if (!node_isset(nid, node_states[N_MEMORY])) 2616 if (!node_isset(nid, node_states[N_MEMORY])) 2765 for_each_node_state(nid, N_MEMORY)
|
/linux-master/kernel/ |
H A D | kthread.c | 744 set_mems_allowed(node_states[N_MEMORY]);
|
/linux-master/kernel/cgroup/ |
H A D | cpuset.c | 569 * of node_states[N_MEMORY]. 575 while (!nodes_intersects(cs->effective_mems, node_states[N_MEMORY])) 577 nodes_and(*pmask, cs->effective_mems, node_states[N_MEMORY]); 2889 * top_cpuset.mems_allowed tracks node_stats[N_MEMORY]; 4566 * synchronized to cpu_active_mask and N_MEMORY, which is necessary in 4592 new_mems = node_states[N_MEMORY]; 4635 /* synchronize mems_allowed to N_MEMORY */ 4691 * Keep top_cpuset.mems_allowed tracking node_states[N_MEMORY]. 4692 * Call this routine anytime after node_states[N_MEMORY] changes. 4717 top_cpuset.effective_mems = node_states[N_MEMORY]; [all...] |
/linux-master/init/ |
H A D | main.c | 1533 set_mems_allowed(node_states[N_MEMORY]);
|
/linux-master/arch/x86/mm/ |
H A D | init_64.c | 824 node_clear_state(0, N_MEMORY);
|
/linux-master/kernel/sched/ |
H A D | fair.c | 1842 if (!node_state(dst_nid, N_MEMORY))
|