Lines Matching refs:mpol
430 struct mempolicy *mpol, pgoff_t ilx, bool *new_page_allocated,
471 mpol, ilx, numa_node_id());
557 struct mempolicy *mpol;
561 mpol = get_vma_policy(vma, addr, 0, &ilx);
562 folio = __read_swap_cache_async(entry, gfp_mask, mpol, ilx,
564 mpol_cond_put(mpol);
636 * @mpol: NUMA memory allocation policy to be applied
651 struct mempolicy *mpol, pgoff_t ilx)
680 gfp_mask, mpol, ilx, &page_allocated, false);
697 folio = __read_swap_cache_async(entry, gfp_mask, mpol, ilx,
803 * @mpol: NUMA memory allocation policy to be applied
816 struct mempolicy *mpol, pgoff_t targ_ilx, struct vm_fault *vmf)
853 folio = __read_swap_cache_async(entry, gfp_mask, mpol, ilx,
873 folio = __read_swap_cache_async(targ_entry, gfp_mask, mpol, targ_ilx,
897 struct mempolicy *mpol;
901 mpol = get_vma_policy(vmf->vma, vmf->address, 0, &ilx);
903 swap_vma_readahead(entry, gfp_mask, mpol, ilx, vmf) :
904 swap_cluster_readahead(entry, gfp_mask, mpol, ilx);
905 mpol_cond_put(mpol);