Searched refs:npage (Results 1 - 14 of 14) sorted by relevance

/linux-master/drivers/s390/cio/
H A Dvfio_ccw_cp.c94 int unpinned = 0, npage = 1; local
98 dma_addr_t *last = &first[npage];
100 if (unpinned + npage < pa_nr &&
101 *first + npage * PAGE_SIZE == *last &&
103 npage++;
107 vfio_unpin_pages(vdev, *first, npage);
108 unpinned += npage;
109 npage = 1;
133 int pinned = 0, npage = 1; local
138 dma_addr_t *last = &first[npage];
[all...]
/linux-master/drivers/vfio/
H A Dvfio.h217 int npage, int prot,
220 dma_addr_t user_iova, int npage);
248 dma_addr_t iova, int npage,
251 dma_addr_t iova, int npage);
293 dma_addr_t iova, int npage,
300 dma_addr_t iova, int npage)
292 vfio_device_container_pin_pages(struct vfio_device *device, dma_addr_t iova, int npage, int prot, struct page **pages) argument
299 vfio_device_container_unpin_pages(struct vfio_device *device, dma_addr_t iova, int npage) argument
H A Dcontainer.c532 dma_addr_t iova, int npage,
539 if (npage > VFIO_PIN_PAGES_MAX_ENTRIES)
545 npage, prot, pages);
549 dma_addr_t iova, int npage)
553 if (WARN_ON(npage <= 0 || npage > VFIO_PIN_PAGES_MAX_ENTRIES))
557 npage);
531 vfio_device_container_pin_pages(struct vfio_device *device, dma_addr_t iova, int npage, int prot, struct page **pages) argument
548 vfio_device_container_unpin_pages(struct vfio_device *device, dma_addr_t iova, int npage) argument
H A Dvfio_iommu_type1.c413 bool lock_cap, long npage)
420 ret = __account_locked_vm(mm, abs(npage), npage > 0, task, lock_cap);
425 static int vfio_lock_acct(struct vfio_dma *dma, long npage, bool async) argument
430 if (!npage)
437 ret = mm_lock_acct(dma->task, mm, dma->lock_cap, npage);
439 dma->locked_vm += npage;
602 long npage, unsigned long *pfn_base,
624 while (npage) {
627 long req_pages = min_t(long, npage, batc
412 mm_lock_acct(struct task_struct *task, struct mm_struct *mm, bool lock_cap, long npage) argument
601 vfio_pin_pages_remote(struct vfio_dma *dma, unsigned long vaddr, long npage, unsigned long *pfn_base, unsigned long limit, struct vfio_batch *batch) argument
712 vfio_unpin_pages_remote(struct vfio_dma *dma, dma_addr_t iova, unsigned long pfn, long npage, bool do_accounting) argument
784 vfio_iommu_type1_pin_pages(void *iommu_data, struct iommu_group *iommu_group, dma_addr_t user_iova, int npage, int prot, struct page **pages) argument
904 vfio_iommu_type1_unpin_pages(void *iommu_data, dma_addr_t user_iova, int npage) argument
1418 vfio_iommu_map(struct vfio_iommu *iommu, dma_addr_t iova, unsigned long pfn, long npage, int prot) argument
1452 long npage; local
1517 long npage = dma->locked_vm; local
1724 long npage; local
[all...]
H A Dvfio_main.c1536 * @npage [in] : count of pages to be pinned. This count should not
1546 int npage, int prot, struct page **pages)
1549 if (!pages || !npage || WARN_ON(!vfio_assert_device_open(device)))
1555 npage, prot, pages);
1569 npage * PAGE_SIZE, pages,
1573 return npage;
1583 * @npage [in] : count of pages to be unpinned. This count should not
1586 void vfio_unpin_pages(struct vfio_device *device, dma_addr_t iova, int npage) argument
1594 vfio_device_container_unpin_pages(device, iova, npage);
1602 npage * PAGE_SIZ
1545 vfio_pin_pages(struct vfio_device *device, dma_addr_t iova, int npage, int prot, struct page **pages) argument
[all...]
/linux-master/drivers/infiniband/hw/hns/
H A Dhns_roce_mr.c597 int count, npage; local
605 npage = 0;
606 while (offset < end && npage < max_count) {
613 for (i = 0; i < count && npage < max_count; i++) {
614 addr = pages[npage];
617 npage++;
622 return npage;
748 int npage; local
759 npage = hns_roce_get_umem_bufs(hr_dev, pages, page_count,
762 npage
853 u32 npage; local
885 u32 npage; local
[all...]
/linux-master/drivers/infiniband/hw/mthca/
H A Dmthca_allocator.c157 int npage = (nent * sizeof (void *) + PAGE_SIZE - 1) / PAGE_SIZE; local
160 array->page_list = kmalloc_array(npage, sizeof(*array->page_list),
165 for (i = 0; i < npage; ++i) {
/linux-master/fs/f2fs/
H A Dnode.c764 struct page *npage[4]; local
777 npage[0] = dn->inode_page;
779 if (!npage[0]) {
780 npage[0] = f2fs_get_node_page(sbi, nids[0]);
781 if (IS_ERR(npage[0]))
782 return PTR_ERR(npage[0]);
788 f2fs_put_page(npage[0], 1);
792 parent = npage[0];
795 dn->inode_page = npage[0];
810 npage[
1217 struct page *npage; local
[all...]
H A Dinline.c275 int f2fs_recover_inline_data(struct inode *inode, struct page *npage) argument
290 if (IS_INODE(npage))
291 ri = F2FS_INODE(npage);
302 src_addr = inline_data_addr(inode, npage);
H A Df2fs.h985 struct page *ipage, struct page *npage, nid_t nid)
990 dn->node_page = npage;
4159 int f2fs_recover_inline_data(struct inode *inode, struct page *npage);
984 set_new_dnode(struct dnode_of_data *dn, struct inode *inode, struct page *ipage, struct page *npage, nid_t nid) argument
/linux-master/lib/
H A Dkfifo.c315 struct page *npage; local
319 npage = virt_to_page(buf);
320 if (page_to_phys(page) != page_to_phys(npage) - l) {
325 page = npage;
/linux-master/include/linux/
H A Dvfio.h322 int npage, int prot, struct page **pages);
323 void vfio_unpin_pages(struct vfio_device *device, dma_addr_t iova, int npage);
/linux-master/drivers/gpu/drm/i915/gvt/
H A Dkvmgt.c141 int npage; local
148 for (npage = 0; npage < total_pages; npage++) {
149 dma_addr_t cur_iova = (gfn + npage) << PAGE_SHIFT;
160 if (npage == 0)
162 else if (page_to_pfn(base_page) + npage != page_to_pfn(cur_page)) {
164 npage++;
172 if (npage)
173 gvt_unpin_guest_page(vgpu, gfn, npage * PAGE_SIZ
[all...]
/linux-master/drivers/iommu/iommufd/
H A Dpages.c671 unsigned int npage = 0; local
677 copy_data_page(pfn_to_page(batch->pfns[cur] + npage), data,
683 npage++;
684 if (npage == batch->npfns[cur]) {
685 npage = 0;

Completed in 189 milliseconds