Lines Matching defs:range

864 			 * the range of VM_LOCKED vma. As page reclaim
866 * the range of VM_LOCKED vma.
1006 struct mmu_notifier_range range;
1013 mmu_notifier_range_init(&range, MMU_NOTIFY_PROTECTION_PAGE, 0,
1015 mmu_notifier_invalidate_range_start(&range);
1059 mmu_notifier_invalidate_range_end(&range);
1109 * pfn_mkclean_range - Cleans the PTEs (including PMDs) mapped with range of
1332 * folio_add_anon_rmap_ptes - add PTE mappings to a page range of an anon folio
1340 * The page range of folio is defined by [first_page, first_page + nr_pages)
1356 * folio_add_anon_rmap_pmd - add a PMD mapping to a page range of an anon folio
1363 * The page range of folio is defined by [first_page, first_page + HPAGE_PMD_NR)
1456 * folio_add_file_rmap_ptes - add PTE mappings to a page range of a folio
1462 * The page range of the folio is defined by [page, page + nr_pages)
1473 * folio_add_file_rmap_pmd - add a PMD mapping to a page range of a folio
1478 * The page range of the folio is defined by [page, page + HPAGE_PMD_NR)
1582 * folio_remove_rmap_ptes - remove PTE mappings from a page range of a folio
1588 * The page range of the folio is defined by [page, page + nr_pages)
1599 * folio_remove_rmap_pmd - remove a PMD mapping from a page range of a folio
1604 * The page range of the folio is defined by [page, page + HPAGE_PMD_NR)
1629 struct mmu_notifier_range range;
1654 range.end = vma_address_end(&pvmw);
1655 mmu_notifier_range_init(&range, MMU_NOTIFY_CLEAR, 0, vma->vm_mm,
1656 address, range.end);
1662 adjust_range_if_pmd_sharing_possible(vma, &range.start,
1663 &range.end);
1668 mmu_notifier_invalidate_range_start(&range);
1706 * range.
1708 flush_cache_range(vma, range.start, range.end);
1730 range.start, range.end);
1923 mmu_notifier_invalidate_range_end(&range);
1978 struct mmu_notifier_range range;
2007 range.end = vma_address_end(&pvmw);
2008 mmu_notifier_range_init(&range, MMU_NOTIFY_CLEAR, 0, vma->vm_mm,
2009 address, range.end);
2015 adjust_range_if_pmd_sharing_possible(vma, &range.start,
2016 &range.end);
2021 mmu_notifier_invalidate_range_start(&range);
2074 * range.
2076 flush_cache_range(vma, range.start, range.end);
2098 range.start, range.end);
2285 mmu_notifier_invalidate_range_end(&range);
2353 struct mmu_notifier_range range;
2358 mmu_notifier_range_init_owner(&range, MMU_NOTIFY_EXCLUSIVE, 0,
2362 mmu_notifier_invalidate_range_start(&range);
2421 mmu_notifier_invalidate_range_end(&range);
2431 * @owner: passed to MMU_NOTIFY_EXCLUSIVE range notifier callbacks
2470 * make_device_exclusive_range() - Mark a range for exclusive use by a device
2475 * @owner: passed to MMU_NOTIFY_EXCLUSIVE range notifier to allow filtering
2477 * Returns: number of pages found in the range by GUP. A page is marked for
2480 * This function finds ptes mapping page(s) to the given address range, locks