Lines Matching refs:umem
139 struct ib_umem *umem = base->ubuffer.umem;
156 if (offset > umem->length ||
157 (send && offset + sizeof(struct mlx5_wqe_ctrl_seg) > umem->length))
161 ret = ib_umem_copy_from(buffer, umem, offset, first_copy_length);
177 ret = ib_umem_copy_from(buffer + first_copy_length, umem, wq->offset,
650 struct ib_umem **umem,
656 *umem = ib_umem_get(pd->uobject->context, addr, size, 0, 0);
657 if (IS_ERR(*umem)) {
659 return PTR_ERR(*umem);
662 mlx5_ib_cont_pages(*umem, addr, npages, page_shift, ncont, NULL);
676 ib_umem_release(*umem);
677 *umem = NULL;
688 if (rwq->umem)
689 ib_umem_release(rwq->umem);
707 rwq->umem = ib_umem_get(pd->uobject->context, ucmd->buf_addr,
709 if (IS_ERR(rwq->umem)) {
711 err = PTR_ERR(rwq->umem);
715 mlx5_ib_cont_pages(rwq->umem, ucmd->buf_addr, &npages, &page_shift,
743 ib_umem_release(rwq->umem);
824 &ubuffer->umem, &npages, &page_shift,
829 ubuffer->umem = NULL;
841 if (ubuffer->umem)
842 mlx5_ib_populate_pas(dev, ubuffer->umem, page_shift, pas, 0);
878 if (ubuffer->umem)
879 ib_umem_release(ubuffer->umem);
894 if (base->ubuffer.umem)
895 ib_umem_release(base->ubuffer.umem);
1080 &sq->ubuffer.umem, &npages, &page_shift,
1111 mlx5_ib_populate_pas(dev, sq->ubuffer.umem, page_shift, pas, 0);
1123 ib_umem_release(sq->ubuffer.umem);
1124 sq->ubuffer.umem = NULL;
1133 ib_umem_release(sq->ubuffer.umem);
4632 mlx5_ib_populate_pas(dev, rwq->umem, rwq->page_shift, rq_pas0, 0);