Lines Matching defs:old_addr
72 unsigned long old_addr, unsigned long old_end,
82 old_start = old_addr;
103 old_pte = pte_offset_map_lock(mm, old_pmd, old_addr, &old_ptl);
110 for (; old_addr < old_end; old_pte++, old_addr += PAGE_SIZE,
114 pte = ptep_clear_flush(vma, old_addr, old_pte);
115 pte = move_pte(pte, new_vma->vm_page_prot, old_addr, new_addr);
132 unsigned long old_addr, struct vm_area_struct *new_vma,
138 old_end = old_addr + len;
139 flush_cache_range(vma, old_addr, old_end);
141 for (; old_addr < old_end; old_addr += extent, new_addr += extent) {
143 next = (old_addr + PMD_SIZE) & PMD_MASK;
146 extent = next - old_addr;
147 old_pmd = get_old_pmd(vma->vm_mm, old_addr);
158 move_ptes(vma, old_pmd, old_addr, old_addr + extent,
162 return len + old_addr - old_end; /* how much done */
166 unsigned long old_addr, unsigned long old_len,
193 err = ksm_madvise(vma, old_addr, old_addr + old_len,
198 new_pgoff = vma->vm_pgoff + ((old_addr - vma->vm_start) >> PAGE_SHIFT);
203 moved_len = move_page_tables(vma, old_addr, new_vma, new_addr, old_len);
210 move_page_tables(new_vma, new_addr, vma, old_addr, moved_len);
213 old_addr = new_addr;
221 if (old_addr > vma->vm_start &&
222 old_addr + old_len < vma->vm_end)
239 if (do_munmap(mm, old_addr, old_len) < 0) {