Lines Matching defs:xprt
72 static int fast_reg_xdr(struct svcxprt_rdma *xprt,
84 frmr = svc_rdma_get_frmr(xprt);
106 ib_dma_map_single(xprt->sc_cm_id->device,
109 if (ib_dma_mapping_error(xprt->sc_cm_id->device,
112 atomic_inc(&xprt->sc_dma_used);
131 ib_dma_map_single(xprt->sc_cm_id->device,
134 if (ib_dma_mapping_error(xprt->sc_cm_id->device,
138 atomic_inc(&xprt->sc_dma_used);
169 ib_dma_map_single(xprt->sc_cm_id->device, va, PAGE_SIZE,
171 if (ib_dma_mapping_error(xprt->sc_cm_id->device,
174 atomic_inc(&xprt->sc_dma_used);
180 if (svc_rdma_fastreg(xprt, frmr))
186 printk("svcrdma: Error fast registering memory for xprt %p\n", xprt);
188 svc_rdma_put_frmr(xprt, frmr);
192 static int map_xdr(struct svcxprt_rdma *xprt,
205 if (xprt->sc_frmr_pg_list_len)
206 return fast_reg_xdr(xprt, xdr, vec);
253 static int send_write(struct svcxprt_rdma *xprt, struct svc_rqst *rqstp,
273 ctxt = svc_rdma_get_context(xprt);
296 ib_dma_map_single(xprt->sc_cm_id->device,
300 if (ib_dma_mapping_error(xprt->sc_cm_id->device,
303 atomic_inc(&xprt->sc_dma_used);
304 sge[sge_no].lkey = xprt->sc_dma_lkey;
332 if (svc_rdma_send(xprt, &write_wr))
341 static int send_write_chunks(struct svcxprt_rdma *xprt,
366 max_write = xprt->sc_max_sge * PAGE_SIZE;
389 ret = send_write(xprt, rqstp,
412 static int send_reply_chunks(struct svcxprt_rdma *xprt,
438 max_write = xprt->sc_max_sge * PAGE_SIZE;
459 ret = send_write(xprt, rqstp,
633 struct svc_xprt *xprt = rqstp->rq_xprt;
635 container_of(xprt, struct svcxprt_rdma, sc_xprt);