Lines Matching refs:riov

294 	     struct vringh_kiov *riov,
316 if (WARN_ON(!riov && !wiov))
319 if (riov)
320 riov->i = riov->used = riov->consumed = 0;
376 iov = riov;
686 * @riov: where to put the readable descriptors (or NULL)
698 * Note that you can reuse riov and wiov with subsequent calls. Content is
700 * When you don't have to use riov and wiov anymore, you should clean up them
704 struct vringh_iov *riov,
742 err = __vringh_iov(vrh, *head, (struct vringh_kiov *)riov,
754 * @riov: the riov as passed to vringh_getdesc_user() (updated as we consume)
760 ssize_t vringh_iov_pull_user(struct vringh_iov *riov, void *dst, size_t len)
762 return vringh_iov_xfer(NULL, (struct vringh_kiov *)riov,
959 * @riov: where to put the readable descriptors (or NULL)
962 * @gfp: flags for allocating larger riov/wiov.
971 * Note that you can reuse riov and wiov with subsequent calls. Content is
973 * When you don't have to use riov and wiov anymore, you should clean up them
977 struct vringh_kiov *riov,
993 err = __vringh_iov(vrh, *head, riov, wiov, no_range_check, NULL,
1004 * @riov: the riov as passed to vringh_getdesc_kern() (updated as we consume)
1010 ssize_t vringh_iov_pull_kern(struct vringh_kiov *riov, void *dst, size_t len)
1012 return vringh_iov_xfer(NULL, riov, dst, len, xfer_kern);
1463 * @riov: where to put the readable descriptors (or NULL)
1466 * @gfp: flags for allocating larger riov/wiov.
1475 * Note that you can reuse riov and wiov with subsequent calls. Content is
1477 * When you don't have to use riov and wiov anymore, you should clean up them
1481 struct vringh_kiov *riov,
1497 err = __vringh_iov(vrh, *head, riov, wiov, no_range_check, NULL,
1509 * @riov: the riov as passed to vringh_getdesc_iotlb() (updated as we consume)
1516 struct vringh_kiov *riov,
1519 return vringh_iov_xfer(vrh, riov, dst, len, xfer_from_iotlb);