Searched refs:N_MEMORY (Results 1 - 24 of 24) sorted by relevance

/linux-master/mm/
H A Dmemory-tiers.c343 for_each_node_state(node, N_MEMORY) {
366 for_each_node_state(node, N_MEMORY) {
402 for_each_node_state(node, N_MEMORY) {
420 nodes_andnot(tier_nodes, node_states[N_MEMORY], tier_nodes);
468 lower_tier = node_states[N_MEMORY];
514 if (!node_state(node, N_MEMORY))
834 * Look at all the existing N_MEMORY nodes and add them to
838 for_each_node_state(node, N_MEMORY) {
H A Dmm_init.c295 * Populate N_MEMORY for calculating usable_nodes.
308 node_set_state(nid, N_MEMORY);
346 nodemask_t saved_node_state = node_states[N_MEMORY];
348 int usable_nodes = nodes_weight(node_states[N_MEMORY]);
461 for_each_node_state(nid, N_MEMORY) {
561 node_states[N_MEMORY] = saved_node_state;
645 if (num_node_state(N_MEMORY) == 1)
1883 * N_MEMORY and not marked online so that no sysfs
1897 node_set_state(nid, N_MEMORY);
2378 /* There will be num_node_state(N_MEMORY) thread
[all...]
H A Dpage_ext.c450 for_each_node_state(nid, N_MEMORY) {
H A Dhugetlb_cgroup.c471 for_each_node_state(nid, N_MEMORY)
476 for_each_node_state(nid, N_MEMORY)
494 for_each_node_state(nid, N_MEMORY) {
H A Dmemory_hotplug.c690 if (!node_state(nid, N_MEMORY))
702 node_set_state(node, N_MEMORY);
1840 nodemask_t nmask = node_states[N_MEMORY];
1914 * it for N_MEMORY as well.
1928 node_clear_state(node, N_MEMORY);
H A Dhugetlb.c2752 folio = remove_pool_hugetlb_folio(h, &node_states[N_MEMORY], 1);
3302 for_each_node_mask_to_alloc(&h->next_nid_to_alloc, nr_nodes, node, &node_states[N_MEMORY]) {
3458 .size = num_node_state(N_MEMORY),
3461 .max_threads = num_node_state(N_MEMORY),
3482 &node_states[N_MEMORY], NULL);
3538 struct folio *folio = alloc_pool_huge_folio(h, &node_states[N_MEMORY],
3576 * job.max_threads is twice the num_node_state(N_MEMORY),
3594 job.max_threads = num_node_state(N_MEMORY) * 2;
3595 job.min_chunk = h->max_huge_pages / num_node_state(N_MEMORY) / 2;
4098 n_mask = &node_states[N_MEMORY];
[all...]
H A Dmempolicy.c275 /* Check N_MEMORY */
277 cpuset_current_mems_allowed, node_states[N_MEMORY]);
1866 * policy->nodes is intersect with node_states[N_MEMORY].
3091 for_each_node_state(nid, N_MEMORY) {
3163 if (!nodes_subset(nodes, node_states[N_MEMORY]))
3195 nodes = node_states[N_MEMORY];
H A Dvmstat.c1636 if (!node_state(pgdat->node_id, N_MEMORY))
2207 if (!node_state(pgdat->node_id, N_MEMORY))
H A Doom_kill.c287 !nodes_subset(node_states[N_MEMORY], *oc->nodemask)) {
H A Dcompaction.c3280 for_each_node_state(nid, N_MEMORY) {
3364 for_each_node_state(nid, N_MEMORY)
H A Dvmscan.c2789 for_each_node_state(nid, N_MEMORY) {
5185 for_each_node_state(nid, N_MEMORY) {
5396 if (nid < 0 || nid >= MAX_NUMNODES || !node_state(nid, N_MEMORY))
7310 for_each_node_state(nid, N_MEMORY)
H A Dmemcontrol.c4235 for_each_node_state(nid, N_MEMORY)
4247 for_each_node_state(nid, N_MEMORY)
5815 for_each_node_state(nid, N_MEMORY) {
6924 for_each_node_state(nid, N_MEMORY) {
H A Dpage_alloc.c201 [N_MEMORY] = { { [0] = 1UL } },
5040 if (!node_isset(node, *used_node_mask) && node_state(node, N_MEMORY)) {
5045 for_each_node_state(n, N_MEMORY) {
H A Dmigrate.c2212 if (!node_state(node, N_MEMORY))
/linux-master/include/linux/
H A Dnodemask.h407 N_MEMORY, /* The node has memory(regular, high, movable) */ enumerator in enum:node_states
445 #define first_memory_node first_node(node_states[N_MEMORY])
452 return next_node(nid, node_states[N_MEMORY]);
H A Dcpuset.h216 #define cpuset_current_mems_allowed (node_states[N_MEMORY])
/linux-master/drivers/base/
H A Dnode.c931 [N_MEMORY] = _NODE_ATTR(has_memory, N_MEMORY),
944 &node_state_attr[N_MEMORY].attr.attr,
/linux-master/fs/proc/
H A Dkcore.c241 for_each_node_state(nid, N_MEMORY) {
H A Dtask_mmu.c2591 if (!node_isset(nid, node_states[N_MEMORY]))
2616 if (!node_isset(nid, node_states[N_MEMORY]))
2765 for_each_node_state(nid, N_MEMORY)
/linux-master/kernel/
H A Dkthread.c744 set_mems_allowed(node_states[N_MEMORY]);
/linux-master/kernel/cgroup/
H A Dcpuset.c569 * of node_states[N_MEMORY].
575 while (!nodes_intersects(cs->effective_mems, node_states[N_MEMORY]))
577 nodes_and(*pmask, cs->effective_mems, node_states[N_MEMORY]);
2889 * top_cpuset.mems_allowed tracks node_stats[N_MEMORY];
4566 * synchronized to cpu_active_mask and N_MEMORY, which is necessary in
4592 new_mems = node_states[N_MEMORY];
4635 /* synchronize mems_allowed to N_MEMORY */
4691 * Keep top_cpuset.mems_allowed tracking node_states[N_MEMORY].
4692 * Call this routine anytime after node_states[N_MEMORY] changes.
4717 top_cpuset.effective_mems = node_states[N_MEMORY];
[all...]
/linux-master/init/
H A Dmain.c1533 set_mems_allowed(node_states[N_MEMORY]);
/linux-master/arch/x86/mm/
H A Dinit_64.c824 node_clear_state(0, N_MEMORY);
/linux-master/kernel/sched/
H A Dfair.c1842 if (!node_state(dst_nid, N_MEMORY))

Completed in 351 milliseconds