/linux-master/lib/ |
H A D | group_cpus.c | 50 masks = kcalloc(nr_node_ids, sizeof(cpumask_var_t), GFP_KERNEL); 54 for (node = 0; node < nr_node_ids; node++) { 72 for (node = 0; node < nr_node_ids; node++) 139 for (n = 0; n < nr_node_ids; n++) { 158 sort(node_groups, nr_node_ids, sizeof(node_groups[0]), 229 for (n = 0; n < nr_node_ids; n++) { 280 node_groups = kcalloc(nr_node_ids, 289 for (i = 0; i < nr_node_ids; i++) {
|
/linux-master/drivers/base/ |
H A D | arch_numa.c | 52 if (WARN_ON(node < 0 || node >= nr_node_ids)) 104 /* setup nr_node_ids if not done yet */ 105 if (nr_node_ids == MAX_NUMNODES) 109 for (node = 0; node < nr_node_ids; node++) { 115 pr_debug("Node to cpumask map for %u nodes\n", nr_node_ids); 278 size = nr_node_ids * nr_node_ids * sizeof(numa_distance[0]); 283 numa_distance_cnt = nr_node_ids;
|
/linux-master/scripts/gdb/linux/ |
H A D | slab.py | 183 nr_node_ids = int(gdb.parse_and_eval("nr_node_ids")) 185 nr_node_ids = 1 198 for i in range(0, nr_node_ids): 278 nr_node_ids = None 294 nr_node_ids = int(gdb.parse_and_eval("nr_node_ids")) 296 nr_node_ids = 1 302 for i in range(0, nr_node_ids):
|
/linux-master/mm/ |
H A D | shrinker_debug.c | 56 count_per_node = kcalloc(nr_node_ids, sizeof(unsigned long), GFP_KERNEL); 125 if (nid < 0 || nid >= nr_node_ids)
|
H A D | mempolicy.c | 1474 unsigned int nbytes = BITS_TO_LONGS(nr_node_ids) * sizeof(long); 1478 nbytes = BITS_TO_COMPAT_LONGS(nr_node_ids) * sizeof(compat_long_t); 1486 maxnode = nr_node_ids; 1740 if (nmask != NULL && maxnode < nr_node_ids) 2436 weights = kzalloc(nr_node_ids, GFP_KERNEL); 2443 memcpy(weights, table, nr_node_ids); 3360 new = kzalloc(nr_node_ids, GFP_KERNEL); 3368 memcpy(new, old, nr_node_ids); 3393 for (i = 0; i < nr_node_ids; i++) 3494 node_attrs = kcalloc(nr_node_ids, sizeo [all...] |
H A D | list_lru.c | 333 mlru = kmalloc(struct_size(mlru, node, nr_node_ids), gfp); 564 lru->node = kcalloc(nr_node_ids, sizeof(*lru->node), GFP_KERNEL);
|
H A D | shrinker.c | 720 size *= nr_node_ids;
|
H A D | ksm.c | 3520 buf = kcalloc(nr_node_ids + nr_node_ids, sizeof(*buf), 3527 root_unstable_tree = buf + nr_node_ids; 3534 ksm_nr_node_ids = knob ? 1 : nr_node_ids;
|
H A D | memory-tiers.c | 820 node_demotion = kcalloc(nr_node_ids, sizeof(struct demotion_nodes),
|
H A D | hugetlb_cgroup.c | 142 h_cgroup = kzalloc(struct_size(h_cgroup, nodeinfo, nr_node_ids),
|
H A D | slub.c | 446 for (__node = 0; __node < nr_node_ids; __node++) \ 5642 nr_node_ids * sizeof(struct kmem_cache_node *), 5661 nr_cpu_ids, nr_node_ids); 6002 nodes = kcalloc(nr_node_ids, sizeof(unsigned long), GFP_KERNEL); 6093 for (node = 0; node < nr_node_ids; node++) {
|
/linux-master/tools/workqueue/ |
H A D | wq_dump.py | 223 nr_node_ids = prog['nr_node_ids'].value_() variable 250 nna = wq.node_nr_active[nr_node_ids]
|
/linux-master/arch/x86/mm/ |
H A D | numa.c | 115 /* setup nr_node_ids if not done yet */ 116 if (nr_node_ids == MAX_NUMNODES) 120 for (node = 0; node < nr_node_ids; node++) 124 pr_debug("Node to cpumask map for %u nodes\n", nr_node_ids); 876 if ((unsigned)node >= nr_node_ids) { 878 "cpumask_of_node(%d): (unsigned)node >= nr_node_ids(%u)\n", 879 node, nr_node_ids);
|
/linux-master/arch/x86/kernel/ |
H A D | setup_percpu.c | 122 pr_info("NR_CPUS:%d nr_cpumask_bits:%d nr_cpu_ids:%u nr_node_ids:%u\n", 123 NR_CPUS, nr_cpumask_bits, nr_cpu_ids, nr_node_ids);
|
/linux-master/arch/loongarch/kernel/ |
H A D | numa.c | 80 if (nr_node_ids >= 8) 383 loongson_sysconf.nr_nodes = nr_node_ids;
|
/linux-master/include/linux/ |
H A D | nodemask.h | 455 extern unsigned int nr_node_ids; 497 #define nr_node_ids 1U macro
|
/linux-master/kernel/sched/ |
H A D | topology.c | 1739 for (i = 0; i < nr_node_ids; i++) { 1741 for (j = 0; j < nr_node_ids; j++) { 1918 masks[i] = kzalloc(nr_node_ids * sizeof(void *), GFP_KERNEL); 2049 for (j = 0; j < nr_node_ids; j++) { 2065 for (j = 0; j < nr_node_ids; j++) { 2195 if (node >= nr_node_ids || hops >= sched_domains_numa_levels)
|
/linux-master/drivers/hv/ |
H A D | hv.c | 112 hv_context.hv_numa_map = kcalloc(nr_node_ids, sizeof(struct cpumask),
|
H A D | channel_mgmt.c | 761 if (numa_node == nr_node_ids) {
|
/linux-master/arch/powerpc/mm/ |
H A D | numa.c | 79 /* setup nr_node_ids if not done yet */ 80 if (nr_node_ids == MAX_NUMNODES) 88 pr_debug("Node to cpumask map for %u nodes\n", nr_node_ids); 185 if (nid == 0xffff || nid >= nr_node_ids) 190 * Returns nid in the range [0..nr_node_ids], or -1 if no useful NUMA
|
/linux-master/drivers/infiniband/sw/siw/ |
H A D | siw_main.c | 125 int i, num_nodes = nr_node_ids;
|
/linux-master/arch/powerpc/platforms/pseries/ |
H A D | hotplug-cpu.c | 241 if (rc && nr_node_ids > 1) {
|
/linux-master/arch/powerpc/sysdev/xive/ |
H A D | common.c | 1144 ipi_domain = irq_domain_create_linear(fwnode, nr_node_ids, 1149 xive_ipis = kcalloc(nr_node_ids, sizeof(*xive_ipis), GFP_KERNEL | __GFP_NOFAIL);
|
/linux-master/kernel/ |
H A D | workqueue.c | 1553 * - node_nr_active[nr_node_ids] if @node is %NUMA_NO_NODE. 1564 node = nr_node_ids; 4686 kfree(nna_ar[nr_node_ids]); 4687 nna_ar[nr_node_ids] = NULL; 4715 /* [nr_node_ids] is used as the fallback */ 4720 nna_ar[nr_node_ids] = nna; 5538 wq_size = struct_size(wq, node_nr_active, nr_node_ids + 1);
|
/linux-master/drivers/net/ethernet/cavium/thunder/ |
H A D | nic_main.c | 1244 if (nr_node_ids > 1)
|