Lines Matching refs:vma
388 static unsigned long dev_pagemap_mapping_shift(struct vm_area_struct *vma,
399 VM_BUG_ON_VMA(address == -EFAULT, vma);
400 pgd = pgd_offset(vma->vm_mm, address);
444 struct vm_area_struct *vma, struct list_head *to_kill,
455 tk->addr = ksm_addr ? ksm_addr : page_address_in_vma(p, vma);
458 tk->addr = vma_pgoff_address(fsdax_pgoff, 1, vma);
459 tk->size_shift = dev_pagemap_mapping_shift(vma, tk->addr);
487 struct vm_area_struct *vma,
490 __add_to_kill(tsk, p, vma, to_kill, 0, FSDAX_INVALID_PGOFF);
507 struct vm_area_struct *vma, struct list_head *to_kill,
511 __add_to_kill(tsk, p, vma, to_kill, ksm_addr, FSDAX_INVALID_PGOFF);
613 struct vm_area_struct *vma;
632 vma = vmac->vma;
633 if (vma->vm_mm != t->mm)
635 if (!page_mapped_in_vma(page, vma))
637 add_to_kill_anon_file(t, page, vma, to_kill);
650 struct vm_area_struct *vma;
663 vma_interval_tree_foreach(vma, &mapping->i_mmap, pgoff,
666 * Send early kill signal to tasks where a vma covers
672 if (vma->vm_mm == t->mm)
673 add_to_kill_anon_file(t, page, vma, to_kill);
682 struct vm_area_struct *vma,
685 __add_to_kill(tsk, p, vma, to_kill, 0, pgoff);
695 struct vm_area_struct *vma;
712 vma_interval_tree_foreach(vma, &mapping->i_mmap, pgoff, pgoff) {
713 if (vma->vm_mm == t->mm)
714 add_to_kill_fsdax(t, page, vma, to_kill, pgoff);
805 ptl = pmd_trans_huge_lock(pmdp, walk->vma);
812 mapped_pte = ptep = pte_offset_map_lock(walk->vma->vm_mm, pmdp,
836 struct hstate *h = hstate_vma(walk->vma);