/linux-master/arch/powerpc/kernel/ |
H A D | vdso.c | 61 static int vdso_mremap(const struct vm_special_mapping *sm, struct vm_area_struct *new_vma, argument 64 unsigned long new_size = new_vma->vm_end - new_vma->vm_start; 69 current->mm->context.vdso = (void __user *)new_vma->vm_start; 74 static int vdso32_mremap(const struct vm_special_mapping *sm, struct vm_area_struct *new_vma) argument 76 return vdso_mremap(sm, new_vma, &vdso32_end - &vdso32_start); 79 static int vdso64_mremap(const struct vm_special_mapping *sm, struct vm_area_struct *new_vma) argument 81 return vdso_mremap(sm, new_vma, &vdso64_end - &vdso64_start);
|
/linux-master/mm/ |
H A D | mremap.c | 138 struct vm_area_struct *new_vma, pmd_t *new_pmd, 160 * - During mremap(), new_vma is often known to be placed after vma 208 pte = move_pte(pte, new_vma->vm_page_prot, old_addr, new_addr); 524 unsigned long *new_addr, struct vm_area_struct *new_vma, 537 !can_align_down(new_vma, *new_addr, mask, for_stack)) 545 unsigned long old_addr, struct vm_area_struct *new_vma, 560 return move_hugetlb_page_tables(vma, new_vma, old_addr, 568 try_realign_addr(&old_addr, vma, &new_addr, new_vma, PMD_MASK, 631 if (pte_alloc(new_vma->vm_mm, new_pmd)) 634 new_vma, new_pm 136 move_ptes(struct vm_area_struct *vma, pmd_t *old_pmd, unsigned long old_addr, unsigned long old_end, struct vm_area_struct *new_vma, pmd_t *new_pmd, unsigned long new_addr, bool need_rmap_locks) argument 523 try_realign_addr(unsigned long *old_addr, struct vm_area_struct *old_vma, unsigned long *new_addr, struct vm_area_struct *new_vma, unsigned long mask, bool for_stack) argument 544 move_page_tables(struct vm_area_struct *vma, unsigned long old_addr, struct vm_area_struct *new_vma, unsigned long new_addr, unsigned long len, bool need_rmap_locks, bool for_stack) argument 658 struct vm_area_struct *new_vma; local [all...] |
H A D | mmap.c | 3359 struct vm_area_struct *new_vma, *prev; local 3372 new_vma = find_vma_prev(mm, addr, &prev); 3373 if (new_vma && new_vma->vm_start < addr + len) 3376 new_vma = vma_merge_new_vma(&vmi, prev, vma, addr, addr + len, pgoff); 3377 if (new_vma) { 3379 * Source vma may have been merged into new_vma 3381 if (unlikely(vma_start >= new_vma->vm_start && 3382 vma_start < new_vma->vm_end)) { 3395 VM_BUG_ON_VMA(faulted_in_anon_vma, new_vma); 3489 special_mapping_mremap(struct vm_area_struct *new_vma) argument [all...] |
H A D | hugetlb.c | 1228 * It has already been transferred to new_vma. 1231 * which copies vma into new_vma and unmaps vma. After the copy 1232 * operation both new_vma and vma share a reference to the resv_map 1235 * the reservation still lives on in new_vma, so simply decrement the 5587 struct vm_area_struct *new_vma, 5632 dst_pte = huge_pte_alloc(mm, new_vma, new_addr, sz); 5586 move_hugetlb_page_tables(struct vm_area_struct *vma, struct vm_area_struct *new_vma, unsigned long old_addr, unsigned long new_addr, unsigned long len) argument
|
/linux-master/arch/x86/entry/vdso/ |
H A D | vma.c | 76 struct vm_area_struct *new_vma) 87 regs->ip = new_vma->vm_start + vdso_land; 93 struct vm_area_struct *new_vma) 97 vdso_fix_landing(image, new_vma); 98 current->mm->context.vdso = (void __user *)new_vma->vm_start; 75 vdso_fix_landing(const struct vdso_image *image, struct vm_area_struct *new_vma) argument 92 vdso_mremap(const struct vm_special_mapping *sm, struct vm_area_struct *new_vma) argument
|
/linux-master/arch/loongarch/kernel/ |
H A D | vdso.c | 41 static int vdso_mremap(const struct vm_special_mapping *sm, struct vm_area_struct *new_vma) argument 43 current->mm->context.vdso = (void *)(new_vma->vm_start);
|
/linux-master/arch/arm64/kernel/ |
H A D | vdso.c | 76 struct vm_area_struct *new_vma) 78 current->mm->context.vdso = (void *)new_vma->vm_start; 246 struct vm_area_struct *new_vma) 248 current->mm->context.sigpage = (void *)new_vma->vm_start; 75 vdso_mremap(const struct vm_special_mapping *sm, struct vm_area_struct *new_vma) argument 245 aarch32_sigpage_mremap(const struct vm_special_mapping *sm, struct vm_area_struct *new_vma) argument
|
/linux-master/arch/riscv/kernel/ |
H A D | vdso.c | 53 struct vm_area_struct *new_vma) 55 current->mm->context.vdso = (void *)new_vma->vm_start; 52 vdso_mremap(const struct vm_special_mapping *sm, struct vm_area_struct *new_vma) argument
|
/linux-master/arch/arm/kernel/ |
H A D | vdso.c | 47 struct vm_area_struct *new_vma) 49 current->mm->context.vdso = new_vma->vm_start; 46 vdso_mremap(const struct vm_special_mapping *sm, struct vm_area_struct *new_vma) argument
|
H A D | process.c | 383 struct vm_area_struct *new_vma) 385 current->mm->context.sigpage = new_vma->vm_start; 382 sigpage_mremap(const struct vm_special_mapping *sm, struct vm_area_struct *new_vma) argument
|
/linux-master/include/linux/ |
H A D | mm_inline.h | 387 struct vm_area_struct *new_vma) 392 new_vma->anon_name = anon_vma_name_reuse(anon_name); 418 struct vm_area_struct *new_vma) {} 386 dup_anon_vma_name(struct vm_area_struct *orig_vma, struct vm_area_struct *new_vma) argument 417 dup_anon_vma_name(struct vm_area_struct *orig_vma, struct vm_area_struct *new_vma) argument
|
H A D | hugetlb.h | 131 struct vm_area_struct *new_vma, 349 struct vm_area_struct *new_vma, 348 move_hugetlb_page_tables(struct vm_area_struct *vma, struct vm_area_struct *new_vma, unsigned long old_addr, unsigned long new_addr, unsigned long len) argument
|
H A D | mm_types.h | 1301 struct vm_area_struct *new_vma);
|
H A D | mm.h | 2530 unsigned long old_addr, struct vm_area_struct *new_vma,
|
/linux-master/drivers/gpu/drm/i915/ |
H A D | i915_gem.c | 951 new_vma: 979 goto new_vma;
|
/linux-master/drivers/gpu/drm/xe/ |
H A D | xe_vm.c | 2229 static struct xe_vma *new_vma(struct xe_vm *vm, struct drm_gpuva_op_map *op, 2415 vma = new_vma(vm, &op->base.map, op->map.pat_index, 2442 vma = new_vma(vm, op->base.remap.prev, 2479 vma = new_vma(vm, op->base.remap.next, 2226 static struct xe_vma *new_vma(struct xe_vm *vm, struct drm_gpuva_op_map *op, function
|