/linux-master/drivers/s390/cio/ |
H A D | vfio_ccw_cp.c | 94 int unpinned = 0, npage = 1; local 98 dma_addr_t *last = &first[npage]; 100 if (unpinned + npage < pa_nr && 101 *first + npage * PAGE_SIZE == *last && 103 npage++; 107 vfio_unpin_pages(vdev, *first, npage); 108 unpinned += npage; 109 npage = 1; 133 int pinned = 0, npage = 1; local 138 dma_addr_t *last = &first[npage]; [all...] |
/linux-master/drivers/vfio/ |
H A D | vfio.h | 217 int npage, int prot, 220 dma_addr_t user_iova, int npage); 248 dma_addr_t iova, int npage, 251 dma_addr_t iova, int npage); 293 dma_addr_t iova, int npage, 300 dma_addr_t iova, int npage) 292 vfio_device_container_pin_pages(struct vfio_device *device, dma_addr_t iova, int npage, int prot, struct page **pages) argument 299 vfio_device_container_unpin_pages(struct vfio_device *device, dma_addr_t iova, int npage) argument
|
H A D | container.c | 532 dma_addr_t iova, int npage, 539 if (npage > VFIO_PIN_PAGES_MAX_ENTRIES) 545 npage, prot, pages); 549 dma_addr_t iova, int npage) 553 if (WARN_ON(npage <= 0 || npage > VFIO_PIN_PAGES_MAX_ENTRIES)) 557 npage); 531 vfio_device_container_pin_pages(struct vfio_device *device, dma_addr_t iova, int npage, int prot, struct page **pages) argument 548 vfio_device_container_unpin_pages(struct vfio_device *device, dma_addr_t iova, int npage) argument
|
H A D | vfio_iommu_type1.c | 413 bool lock_cap, long npage) 420 ret = __account_locked_vm(mm, abs(npage), npage > 0, task, lock_cap); 425 static int vfio_lock_acct(struct vfio_dma *dma, long npage, bool async) argument 430 if (!npage) 437 ret = mm_lock_acct(dma->task, mm, dma->lock_cap, npage); 439 dma->locked_vm += npage; 602 long npage, unsigned long *pfn_base, 624 while (npage) { 627 long req_pages = min_t(long, npage, batc 412 mm_lock_acct(struct task_struct *task, struct mm_struct *mm, bool lock_cap, long npage) argument 601 vfio_pin_pages_remote(struct vfio_dma *dma, unsigned long vaddr, long npage, unsigned long *pfn_base, unsigned long limit, struct vfio_batch *batch) argument 712 vfio_unpin_pages_remote(struct vfio_dma *dma, dma_addr_t iova, unsigned long pfn, long npage, bool do_accounting) argument 784 vfio_iommu_type1_pin_pages(void *iommu_data, struct iommu_group *iommu_group, dma_addr_t user_iova, int npage, int prot, struct page **pages) argument 904 vfio_iommu_type1_unpin_pages(void *iommu_data, dma_addr_t user_iova, int npage) argument 1418 vfio_iommu_map(struct vfio_iommu *iommu, dma_addr_t iova, unsigned long pfn, long npage, int prot) argument 1452 long npage; local 1517 long npage = dma->locked_vm; local 1724 long npage; local [all...] |
H A D | vfio_main.c | 1536 * @npage [in] : count of pages to be pinned. This count should not 1546 int npage, int prot, struct page **pages) 1549 if (!pages || !npage || WARN_ON(!vfio_assert_device_open(device))) 1555 npage, prot, pages); 1569 npage * PAGE_SIZE, pages, 1573 return npage; 1583 * @npage [in] : count of pages to be unpinned. This count should not 1586 void vfio_unpin_pages(struct vfio_device *device, dma_addr_t iova, int npage) argument 1594 vfio_device_container_unpin_pages(device, iova, npage); 1602 npage * PAGE_SIZ 1545 vfio_pin_pages(struct vfio_device *device, dma_addr_t iova, int npage, int prot, struct page **pages) argument [all...] |
/linux-master/drivers/infiniband/hw/hns/ |
H A D | hns_roce_mr.c | 597 int count, npage; local 605 npage = 0; 606 while (offset < end && npage < max_count) { 613 for (i = 0; i < count && npage < max_count; i++) { 614 addr = pages[npage]; 617 npage++; 622 return npage; 748 int npage; local 759 npage = hns_roce_get_umem_bufs(hr_dev, pages, page_count, 762 npage 853 u32 npage; local 885 u32 npage; local [all...] |
/linux-master/drivers/infiniband/hw/mthca/ |
H A D | mthca_allocator.c | 157 int npage = (nent * sizeof (void *) + PAGE_SIZE - 1) / PAGE_SIZE; local 160 array->page_list = kmalloc_array(npage, sizeof(*array->page_list), 165 for (i = 0; i < npage; ++i) {
|
/linux-master/fs/f2fs/ |
H A D | node.c | 764 struct page *npage[4]; local 777 npage[0] = dn->inode_page; 779 if (!npage[0]) { 780 npage[0] = f2fs_get_node_page(sbi, nids[0]); 781 if (IS_ERR(npage[0])) 782 return PTR_ERR(npage[0]); 788 f2fs_put_page(npage[0], 1); 792 parent = npage[0]; 795 dn->inode_page = npage[0]; 810 npage[ 1217 struct page *npage; local [all...] |
H A D | inline.c | 275 int f2fs_recover_inline_data(struct inode *inode, struct page *npage) argument 290 if (IS_INODE(npage)) 291 ri = F2FS_INODE(npage); 302 src_addr = inline_data_addr(inode, npage);
|
H A D | f2fs.h | 985 struct page *ipage, struct page *npage, nid_t nid) 990 dn->node_page = npage; 4159 int f2fs_recover_inline_data(struct inode *inode, struct page *npage); 984 set_new_dnode(struct dnode_of_data *dn, struct inode *inode, struct page *ipage, struct page *npage, nid_t nid) argument
|
/linux-master/lib/ |
H A D | kfifo.c | 315 struct page *npage; local 319 npage = virt_to_page(buf); 320 if (page_to_phys(page) != page_to_phys(npage) - l) { 325 page = npage;
|
/linux-master/include/linux/ |
H A D | vfio.h | 322 int npage, int prot, struct page **pages); 323 void vfio_unpin_pages(struct vfio_device *device, dma_addr_t iova, int npage);
|
/linux-master/drivers/gpu/drm/i915/gvt/ |
H A D | kvmgt.c | 141 int npage; local 148 for (npage = 0; npage < total_pages; npage++) { 149 dma_addr_t cur_iova = (gfn + npage) << PAGE_SHIFT; 160 if (npage == 0) 162 else if (page_to_pfn(base_page) + npage != page_to_pfn(cur_page)) { 164 npage++; 172 if (npage) 173 gvt_unpin_guest_page(vgpu, gfn, npage * PAGE_SIZ [all...] |
/linux-master/drivers/iommu/iommufd/ |
H A D | pages.c | 671 unsigned int npage = 0; local 677 copy_data_page(pfn_to_page(batch->pfns[cur] + npage), data, 683 npage++; 684 if (npage == batch->npfns[cur]) { 685 npage = 0;
|