/linux-master/mm/ |
H A D | vmalloc.c | 2567 node = numa_node_id();
|
H A D | shmem.c | 1614 page = alloc_pages_mpol(gfp, HPAGE_PMD_ORDER, mpol, ilx, numa_node_id()); 1628 page = alloc_pages_mpol(gfp, 0, mpol, ilx, numa_node_id());
|
H A D | memory.c | 5047 if (page_nid == numa_node_id()) {
|
H A D | vmscan.c | 6523 struct zonelist *zonelist = node_zonelist(numa_node_id(), sc.gfp_mask); 7248 struct zonelist *zonelist = node_zonelist(numa_node_id(), sc.gfp_mask); 7474 if (node_state(pgdat->node_id, N_CPU) && pgdat->node_id != numa_node_id())
|
H A D | swapfile.c | 1077 node = numa_node_id();
|
H A D | swap_state.c | 471 mpol, ilx, numa_node_id());
|
H A D | page_alloc.c | 2707 if (zone_to_nid(z) != numa_node_id()) 4927 struct zonelist *zonelist = node_zonelist(numa_node_id(), GFP_KERNEL);
|
H A D | page-writeback.c | 2628 reclaim_throttle(NODE_DATA(numa_node_id()),
|
H A D | mprotect.c | 103 target_node = numa_node_id();
|
H A D | mempolicy.c | 207 node = numa_node_id(); 1217 int nid = numa_node_id(); 1990 return numa_node_id(); 2032 return numa_node_id(); 2108 nid = numa_node_id(); 2149 init_nodemask_of_node(mask, numa_node_id()); 2216 * @nid: Preferred node (usually numa_node_id() but @mpol may override it). 2304 pol, ilx, numa_node_id()); 2336 pol, NO_INTERLEAVE_INDEX, numa_node_id()); 2510 nr_allocated += __alloc_pages_bulk(gfp, numa_node_id(), NUL [all...] |
/linux-master/drivers/gpu/drm/amd/amdkfd/ |
H A D | kfd_topology.c | 2244 * of the given NUMA node (numa_node_id) 2247 int kfd_numa_node_to_apic_id(int numa_node_id) argument 2249 if (numa_node_id == -1) { 2253 return kfd_cpumask_to_apic_id(cpumask_of_node(numa_node_id));
|
H A D | kfd_device.c | 1028 if (cpu_to_node(new_cpu) == numa_node_id())
|
H A D | kfd_priv.h | 1141 int kfd_numa_node_to_apic_id(int numa_node_id);
|
H A D | kfd_crat.c | 1703 * @numa_node_id: CPU NUMA node id 1709 static int kfd_fill_cu_for_cpu(int numa_node_id, int *avail_size, argument 1726 cpumask = cpumask_of_node(numa_node_id); 1731 sub_type_hdr->processor_id_low = kfd_numa_node_to_apic_id(numa_node_id); 1742 * @numa_node_id: CPU NUMA node id 1748 static int kfd_fill_mem_info_for_cpu(int numa_node_id, int *avail_size, argument 1773 pgdat = NODE_DATA(numa_node_id); 1786 static int kfd_fill_iolink_info_for_cpu(int numa_node_id, int *avail_size, argument 1803 if (nid == numa_node_id) /* node itself */ 1818 sub_type_hdr->proximity_domain_from = numa_node_id; 1843 int numa_node_id; local [all...] |
/linux-master/drivers/nvme/host/ |
H A D | multipath.c | 342 int node = numa_node_id();
|
/linux-master/drivers/md/ |
H A D | dm.c | 734 td = kmalloc_node(sizeof(*td), GFP_KERNEL, md->numa_node_id); 2065 int r, numa_node_id = dm_get_numa_node(); local 2070 md = kvzalloc_node(sizeof(*md), GFP_KERNEL, numa_node_id); 2091 md->numa_node_id = numa_node_id; 2111 md->disk = blk_alloc_disk(NULL, md->numa_node_id);
|
H A D | dm-table.c | 1031 pools = kzalloc_node(sizeof(*pools), GFP_KERNEL, md->numa_node_id);
|
/linux-master/net/core/ |
H A D | dev.c | 2545 int i, j, tci, numa_node_id = -2; local 2644 if (numa_node_id == -2) 2645 numa_node_id = cpu_to_node(j); 2646 else if (numa_node_id != cpu_to_node(j)) 2647 numa_node_id = -1; 2690 (numa_node_id >= 0) ? 2691 numa_node_id : NUMA_NO_NODE);
|
/linux-master/drivers/gpu/drm/nouveau/ |
H A D | nouveau_dmem.c | 268 ptr = memremap_pages(&chunk->pagemap, numa_node_id());
|
/linux-master/kernel/bpf/ |
H A D | helpers.c | 165 return numa_node_id();
|
/linux-master/drivers/crypto/intel/qat/qat_common/ |
H A D | qat_comp_algs.c | 200 node = numa_node_id();
|
/linux-master/arch/sparc/mm/ |
H A D | tsb.c | 425 gfp_flags, numa_node_id());
|
/linux-master/arch/sparc/kernel/ |
H A D | irq_64.c | 246 irq = __irq_alloc_descs(-1, 1, 1, numa_node_id(), NULL, NULL);
|
/linux-master/include/linux/ |
H A D | gfp.h | 274 return alloc_pages_node(numa_node_id(), gfp_mask, order); 283 return __folio_alloc_node(gfp, order, numa_node_id());
|
/linux-master/arch/x86/include/asm/ |
H A D | topology.h | 85 static inline int numa_node_id(void) function 92 #define numa_node_id numa_node_id macro
|