/linux-master/include/linux/ |
H A D | mempolicy.h | 122 void mpol_shared_policy_init(struct shared_policy *sp, struct mempolicy *mpol); 124 struct vm_area_struct *vma, struct mempolicy *mpol); 143 struct mempolicy **mpol, nodemask_t **nodemask); 162 extern int mpol_parse_str(char *str, struct mempolicy **mpol); 209 struct mempolicy *mpol) 255 struct mempolicy **mpol, nodemask_t **nodemask) 257 *mpol = NULL; 278 static inline int mpol_parse_str(char *str, struct mempolicy **mpol) argument 208 mpol_shared_policy_init(struct shared_policy *sp, struct mempolicy *mpol) argument 253 huge_node(struct vm_area_struct *vma, unsigned long addr, gfp_t gfp_flags, struct mempolicy **mpol, nodemask_t **nodemask) argument
|
H A D | shmem_fs.h | 71 struct mempolicy *mpol; /* default memory policy for mappings */ member in struct:shmem_sb_info
|
H A D | gfp.h | 267 struct mempolicy *mpol, pgoff_t ilx, int nid); 277 struct mempolicy *mpol, pgoff_t ilx, int nid) 276 alloc_pages_mpol(gfp_t gfp, unsigned int order, struct mempolicy *mpol, pgoff_t ilx, int nid) argument
|
/linux-master/mm/ |
H A D | swap.h | 54 struct mempolicy *mpol, pgoff_t ilx, bool *new_page_allocated, 57 struct mempolicy *mpol, pgoff_t ilx); 85 gfp_t gfp_mask, struct mempolicy *mpol, pgoff_t ilx) 84 swap_cluster_readahead(swp_entry_t entry, gfp_t gfp_mask, struct mempolicy *mpol, pgoff_t ilx) argument
|
H A D | swap_state.c | 430 struct mempolicy *mpol, pgoff_t ilx, bool *new_page_allocated, 471 mpol, ilx, numa_node_id()); 557 struct mempolicy *mpol; local 561 mpol = get_vma_policy(vma, addr, 0, &ilx); 562 folio = __read_swap_cache_async(entry, gfp_mask, mpol, ilx, 564 mpol_cond_put(mpol); 636 * @mpol: NUMA memory allocation policy to be applied 651 struct mempolicy *mpol, pgoff_t ilx) 680 gfp_mask, mpol, ilx, &page_allocated, false); 697 folio = __read_swap_cache_async(entry, gfp_mask, mpol, il 429 __read_swap_cache_async(swp_entry_t entry, gfp_t gfp_mask, struct mempolicy *mpol, pgoff_t ilx, bool *new_page_allocated, bool skip_if_exists) argument 650 swap_cluster_readahead(swp_entry_t entry, gfp_t gfp_mask, struct mempolicy *mpol, pgoff_t ilx) argument 815 swap_vma_readahead(swp_entry_t targ_entry, gfp_t gfp_mask, struct mempolicy *mpol, pgoff_t targ_ilx, struct vm_fault *vmf) argument 897 struct mempolicy *mpol; local [all...] |
H A D | shmem.c | 116 struct mempolicy *mpol; member in struct:shmem_options 1532 static void shmem_show_mpol(struct seq_file *seq, struct mempolicy *mpol) argument 1536 if (!mpol || mpol->mode == MPOL_DEFAULT) 1539 mpol_to_str(buffer, sizeof(buffer), mpol); 1541 seq_printf(seq, ",mpol=%s", buffer); 1546 struct mempolicy *mpol = NULL; local 1547 if (sbinfo->mpol) { 1549 mpol = sbinfo->mpol; 1556 shmem_show_mpol(struct seq_file *seq, struct mempolicy *mpol) argument 1571 struct mempolicy *mpol; local 1609 struct mempolicy *mpol; local 1623 struct mempolicy *mpol; local 2356 shmem_set_policy(struct vm_area_struct *vma, struct mempolicy *mpol) argument 2382 struct mempolicy *mpol; local 4154 struct mempolicy *mpol = NULL; local 4248 struct mempolicy *mpol; local [all...] |
H A D | mempolicy.c | 343 /* Slow path of a mpol destructor. */ 971 * Take a refcount on the mpol, because we are about to 2090 * huge_node(@vma, @addr, @gfp_flags, @mpol) 2094 * @mpol: pointer to mempolicy pointer for reference counted mempolicy 2103 struct mempolicy **mpol, nodemask_t **nodemask) 2109 *mpol = get_vma_policy(vma, addr, hstate_vma(vma)->order, &ilx); 2110 *nodemask = policy_nodemask(gfp_flags, *mpol, ilx, &nid); 2216 * @nid: Preferred node (usually numa_node_id() but @mpol may override it). 2933 * @mpol: struct mempolicy to install 2935 * Install non-NULL @mpol i 2102 huge_node(struct vm_area_struct *vma, unsigned long addr, gfp_t gfp_flags, struct mempolicy **mpol, nodemask_t **nodemask) argument 2940 mpol_shared_policy_init(struct shared_policy *sp, struct mempolicy *mpol) argument 3146 mpol_parse_str(char *str, struct mempolicy **mpol) argument [all...] |
H A D | hugetlb.c | 1396 struct mempolicy *mpol; local 1414 nid = huge_node(vma, address, gfp_mask, &mpol, &nodemask); 1416 if (mpol_is_preferred_many(mpol)) { 1433 mpol_cond_put(mpol); 2600 struct mempolicy *mpol; local 2605 nid = huge_node(vma, addr, gfp_mask, &mpol, &nodemask); 2606 if (mpol_is_preferred_many(mpol)) { 2618 mpol_cond_put(mpol); 4900 struct mempolicy *mpol = get_task_policy(current); local 4906 if (mpol 6648 struct mempolicy *mpol; local [all...] |
H A D | zswap.c | 1131 struct mempolicy *mpol; local 1138 mpol = get_task_policy(current); 1139 folio = __read_swap_cache_async(swpentry, GFP_KERNEL, mpol,
|
/linux-master/ipc/ |
H A D | shm.c | 566 static int shm_set_policy(struct vm_area_struct *vma, struct mempolicy *mpol) argument 572 err = sfd->vm_ops->set_policy(vma, mpol); 580 struct mempolicy *mpol = vma->vm_policy; local 583 mpol = sfd->vm_ops->get_policy(vma, addr, ilx); 584 return mpol;
|