Lines Matching defs:wiov

295 	     struct vringh_kiov *wiov,
316 if (WARN_ON(!riov && !wiov))
321 if (wiov)
322 wiov->i = wiov->used = wiov->consumed = 0;
374 iov = wiov;
377 if (unlikely(wiov && wiov->used)) {
387 !wiov ? "writable" : "readable");
687 * @wiov: where to put the writable descriptors (or NULL)
698 * Note that you can reuse riov and wiov with subsequent calls. Content is
700 * When you don't have to use riov and wiov anymore, you should clean up them
705 struct vringh_iov *wiov,
743 (struct vringh_kiov *)wiov,
769 * @wiov: the wiov as passed to vringh_getdesc_user() (updated as we consume)
775 ssize_t vringh_iov_push_user(struct vringh_iov *wiov,
778 return vringh_iov_xfer(NULL, (struct vringh_kiov *)wiov,
960 * @wiov: where to put the writable descriptors (or NULL)
962 * @gfp: flags for allocating larger riov/wiov.
971 * Note that you can reuse riov and wiov with subsequent calls. Content is
973 * When you don't have to use riov and wiov anymore, you should clean up them
978 struct vringh_kiov *wiov,
993 err = __vringh_iov(vrh, *head, riov, wiov, no_range_check, NULL,
1018 * @wiov: the wiov as passed to vringh_getdesc_kern() (updated as we consume)
1024 ssize_t vringh_iov_push_kern(struct vringh_kiov *wiov,
1027 return vringh_iov_xfer(NULL, wiov, (void *)src, len, kern_xfer);
1464 * @wiov: where to put the writable descriptors (or NULL)
1466 * @gfp: flags for allocating larger riov/wiov.
1475 * Note that you can reuse riov and wiov with subsequent calls. Content is
1477 * When you don't have to use riov and wiov anymore, you should clean up them
1482 struct vringh_kiov *wiov,
1497 err = __vringh_iov(vrh, *head, riov, wiov, no_range_check, NULL,
1526 * @wiov: the wiov as passed to vringh_getdesc_iotlb() (updated as we consume)
1533 struct vringh_kiov *wiov,
1536 return vringh_iov_xfer(vrh, wiov, (void *)src, len, xfer_to_iotlb);