Lines Matching refs:iova
194 * iopt_pages. If the iova is sub page-size then the area has an iova that
221 static void iommu_unmap_nofail(struct iommu_domain *domain, unsigned long iova,
226 ret = iommu_unmap(domain, iova, size);
381 unsigned long iova;
384 iova = iopt_area_index_to_iova(area, start_index);
393 phys = iommu_iova_to_phys(domain, iova) - page_offset;
396 iova += PAGE_SIZE - page_offset;
409 unsigned long iova;
412 iova = iopt_area_index_to_iova(area, start_index);
416 phys = iommu_iova_to_phys(domain, iova) - page_offset;
418 iova += PAGE_SIZE - page_offset;
448 unsigned long iova, phys_addr_t paddr,
451 unsigned long start_iova = iova;
455 WARN_ON(paddr % PAGE_SIZE || iova % PAGE_SIZE ||
459 rc = iommu_map(domain, iova, paddr, PAGE_SIZE, prot,
463 iova += PAGE_SIZE;
470 if (start_iova != iova)
471 iommu_unmap_nofail(domain, start_iova, iova - start_iova);
484 unsigned long iova;
490 next_iova = iova = start_iova =
498 domain, iova,
500 next_iova - iova, area->iommu_prot);
502 rc = iommu_map(domain, iova,
504 next_iova - iova, area->iommu_prot,
508 iova = next_iova;
514 if (start_iova != iova)
515 iommu_unmap_nofail(domain, start_iova, iova - start_iova);
1489 * Called during area destruction. This unmaps the iova's covered by all the