/linux-master/drivers/nvme/target/ |
H A D | rdma.c | 21 #include <rdma/ib_verbs.h> 22 #include <rdma/rdma_cm.h> 23 #include <rdma/rw.h> 24 #include <rdma/ib_cm.h> 26 #include <linux/nvme-rdma.h> 1711 * @queue: nvmet rdma queue (cm id qp_context)
|
H A D | configfs.c | 37 { NVMF_TRTYPE_RDMA, "rdma" }, 347 port->disc_addr.tsas.rdma.qptype = NVMF_RDMA_QPTYPE_CONNECTED; 348 port->disc_addr.tsas.rdma.prtype = NVMF_RDMA_PRTYPE_NOT_SPECIFIED; 349 port->disc_addr.tsas.rdma.cms = NVMF_RDMA_CMS_RDMA_CM; 409 if (port->disc_addr.tsas.rdma.qptype == nvmet_addr_tsas_rdma[i].type)
|
/linux-master/net/smc/ |
H A D | smc_ib.c | 23 #include <rdma/ib_verbs.h> 24 #include <rdma/ib_cache.h>
|
/linux-master/lib/ |
H A D | dynamic_debug.c | 40 #include <rdma/ib_verbs.h>
|
/linux-master/net/sunrpc/xprtrdma/ |
H A D | svc_rdma_sendto.c | 105 #include <rdma/ib_verbs.h> 106 #include <rdma/rdma_cm.h> 117 svc_rdma_send_ctxt_alloc(struct svcxprt_rdma *rdma) argument 119 int node = ibdev_to_node(rdma->sc_cm_id->device); 125 ctxt = kzalloc_node(struct_size(ctxt, sc_sges, rdma->sc_max_send_sges), 129 buffer = kmalloc_node(rdma->sc_max_req_size, GFP_KERNEL, node); 132 addr = ib_dma_map_single(rdma->sc_pd->device, buffer, 133 rdma->sc_max_req_size, DMA_TO_DEVICE); 134 if (ib_dma_mapping_error(rdma->sc_pd->device, addr)) 137 svc_rdma_send_cid_init(rdma, 167 svc_rdma_send_ctxts_destroy(struct svcxprt_rdma *rdma) argument 190 svc_rdma_send_ctxt_get(struct svcxprt_rdma *rdma) argument 224 svc_rdma_send_ctxt_release(struct svcxprt_rdma *rdma, struct svc_rdma_send_ctxt *ctxt) argument 266 svc_rdma_send_ctxt_put(struct svcxprt_rdma *rdma, struct svc_rdma_send_ctxt *ctxt) argument 279 svc_rdma_wake_send_waiters(struct svcxprt_rdma *rdma, int avail) argument 297 struct svcxprt_rdma *rdma = cq->cq_context; local 338 svc_rdma_post_send(struct svcxprt_rdma *rdma, struct svc_rdma_send_ctxt *ctxt) argument 566 struct svcxprt_rdma *rdma = args->md_rdma; local 703 svc_rdma_pull_up_needed(const struct svcxprt_rdma *rdma, const struct svc_rdma_send_ctxt *sctxt, const struct svc_rdma_pcl *write_pcl, const struct xdr_buf *xdr) argument 784 svc_rdma_pull_up_reply_msg(const struct svcxprt_rdma *rdma, struct svc_rdma_send_ctxt *sctxt, const struct svc_rdma_pcl *write_pcl, const struct xdr_buf *xdr) argument 818 svc_rdma_map_reply_msg(struct svcxprt_rdma *rdma, struct svc_rdma_send_ctxt *sctxt, const struct svc_rdma_pcl *write_pcl, const struct svc_rdma_pcl *reply_pcl, const struct xdr_buf *xdr) argument 880 svc_rdma_send_reply_msg(struct svcxprt_rdma *rdma, struct svc_rdma_send_ctxt *sctxt, const struct svc_rdma_recv_ctxt *rctxt, struct svc_rqst *rqstp) argument 923 svc_rdma_send_error_msg(struct svcxprt_rdma *rdma, struct svc_rdma_send_ctxt *sctxt, struct svc_rdma_recv_ctxt *rctxt, int status) argument 992 struct svcxprt_rdma *rdma = local [all...] |
H A D | svc_rdma_rw.c | 8 #include <rdma/rw.h> 55 svc_rdma_get_rw_ctxt(struct svcxprt_rdma *rdma, unsigned int sges) argument 57 struct ib_device *dev = rdma->sc_cm_id->device; 62 spin_lock(&rdma->sc_rw_ctxt_lock); 63 node = llist_del_first(&rdma->sc_rw_ctxts); 64 spin_unlock(&rdma->sc_rw_ctxt_lock); 87 trace_svcrdma_rwctx_empty(rdma, sges); 98 static void svc_rdma_put_rw_ctxt(struct svcxprt_rdma *rdma, argument 101 __svc_rdma_put_rw_ctxt(ctxt, &rdma->sc_rw_ctxts); 106 * @rdma 109 svc_rdma_destroy_rw_ctxts(struct svcxprt_rdma *rdma) argument 131 svc_rdma_rw_ctx_init(struct svcxprt_rdma *rdma, struct svc_rdma_rw_ctxt *ctxt, u64 offset, u32 handle, enum dma_data_direction direction) argument 154 svc_rdma_cc_init(struct svcxprt_rdma *rdma, struct svc_rdma_chunk_ctxt *cc) argument 172 svc_rdma_cc_release(struct svcxprt_rdma *rdma, struct svc_rdma_chunk_ctxt *cc, enum dma_data_direction dir) argument 201 svc_rdma_write_info_alloc(struct svcxprt_rdma *rdma, const struct svc_rdma_chunk *chunk) argument 238 svc_rdma_reply_chunk_release(struct svcxprt_rdma *rdma, struct svc_rdma_send_ctxt *ctxt) argument 260 struct svcxprt_rdma *rdma = cq->cq_context; local 285 struct svcxprt_rdma *rdma = cq->cq_context; local 319 struct svcxprt_rdma *rdma = cq->cq_context; local 362 svc_rdma_post_chunk_ctxt(struct svcxprt_rdma *rdma, struct svc_rdma_chunk_ctxt *cc) argument 475 struct svcxprt_rdma *rdma = info->wi_rdma; local 604 svc_rdma_send_write_chunk(struct svcxprt_rdma *rdma, const struct svc_rdma_chunk *chunk, const struct xdr_buf *xdr) argument 646 svc_rdma_send_write_list(struct svcxprt_rdma *rdma, const struct svc_rdma_recv_ctxt *rctxt, const struct xdr_buf *xdr) argument 678 svc_rdma_prepare_reply_chunk(struct svcxprt_rdma *rdma, const struct svc_rdma_pcl *write_pcl, const struct svc_rdma_pcl *reply_pcl, struct svc_rdma_send_ctxt *sctxt, const struct xdr_buf *xdr) argument 735 struct svcxprt_rdma *rdma = svc_rdma_rqst_rdma(rqstp); local 1115 svc_rdma_process_read_list(struct svcxprt_rdma *rdma, struct svc_rqst *rqstp, struct svc_rdma_recv_ctxt *head) argument [all...] |
/linux-master/include/linux/sunrpc/ |
H A D | svc_rdma.h | 53 #include <rdma/ib_verbs.h> 54 #include <rdma/rdma_cm.h> 140 * @rdma: controlling transport 143 static inline void svc_rdma_recv_cid_init(struct svcxprt_rdma *rdma, argument 146 cid->ci_queue_id = rdma->sc_rq_cq->res.id; 147 cid->ci_completion_id = atomic_inc_return(&rdma->sc_completion_ids); 152 * @rdma: controlling transport 155 static inline void svc_rdma_send_cid_init(struct svcxprt_rdma *rdma, argument 158 cid->ci_queue_id = rdma->sc_sq_cq->res.id; 159 cid->ci_completion_id = atomic_inc_return(&rdma [all...] |
/linux-master/fs/smb/client/ |
H A D | smb2pdu.c | 4368 /* we can only offload on an rdma connection */ 4369 if (!server->rdma || !server->smbd_conn)
|
H A D | smb2ops.c | 462 if (server->rdma) { 510 if (server->rdma) { 536 * return 1 if a is rdma capable, or rss capable, or has higher link speed
|
H A D | fs_context.h | 251 bool rdma:1; member in struct:smb3_fs_context
|
H A D | connect.c | 1447 if (server->rdma) 1583 if (server->rdma != ctx->rdma) 1725 tcp_ses->rdma = ctx->rdma; 1782 if (tcp_ses->rdma) {
|
H A D | fs_context.c | 124 fsparam_flag("rdma", Opt_rdma), 705 if (ctx->rdma && ctx->vals->protocol_id < SMB30_PROT_ID) { 1626 ctx->rdma = true;
|
H A D | cifsglob.h | 753 bool rdma; member in struct:TCP_Server_Info
|
H A D | cifsfs.c | 457 if (server->rdma) 458 seq_puts(s, ",rdma");
|
H A D | cifs_debug.c | 221 seq_puts(m, "rdma "); 376 if (!server->rdma) 455 if (server->rdma)
|
H A D | sess.c | 247 /* do not mix rdma and non-rdma interfaces */ 248 if (iface->rdma_capable != ses->server->rdma) 419 /* do not mix rdma and non-rdma interfaces */ 420 if (iface->rdma_capable != server->rdma) 525 cifs_dbg(FYI, "adding channel to ses %p (speed:%zu bps rdma:%s ip:%pI4)\n", 529 cifs_dbg(FYI, "adding channel to ses %p (speed:%zu bps rdma:%s ip:%pI6)\n", 593 ctx->rdma = iface->rdma_capable;
|
/linux-master/drivers/infiniband/sw/rxe/ |
H A D | rxe.c | 7 #include <rdma/rdma_netlink.h>
|
/linux-master/drivers/infiniband/hw/mlx5/ |
H A D | mad.c | 34 #include <rdma/ib_mad.h> 35 #include <rdma/ib_smi.h> 36 #include <rdma/ib_pma.h>
|
/linux-master/drivers/infiniband/core/ |
H A D | cm.c | 26 #include <rdma/ib_cache.h> 27 #include <rdma/ib_cm.h> 28 #include <rdma/ib_sysfs.h>
|
/linux-master/sound/soc/codecs/ |
H A D | wm_adsp.c | 96 __be32 rdma[6]; member in struct:wm_adsp_system_config_xm_hdr
|
/linux-master/net/rds/ |
H A D | rdma.c | 44 * - an rdma is an mlock, apply rlimit? 616 struct rm_rdma_op *op = &rm->rdma; 626 || rm->rdma.op_active) 705 rdsdebug("RDS: rdma prepare nr_local %llu rva %llx rkey %x\n", 843 rm->rdma.op_rdma_mr = mr; 862 &rm->rdma.op_rdma_mr, rm->m_conn_path);
|
H A D | send.c | 280 (rm->rdma.op_active && 308 /* The transport either sends the whole rdma or none of it */ 309 if (rm->rdma.op_active && !cp->cp_xmit_rdma_sent) { 310 rm->m_final_op = &rm->rdma; 315 ret = conn->c_trans->xmit_rdma(conn, &rm->rdma); 352 ops_present = (rm->atomic.op_active || rm->rdma.op_active); 355 if (rm->rdma.op_active && !rm->rdma.op_silent) 508 ro = &rm->rdma; 579 ro = &rm->rdma; [all...] |
/linux-master/drivers/net/ethernet/intel/ice/ |
H A D | ice_common.c | 1219 uld_mask = ICE_RESET_DONE_MASK | (hw->func_caps.common_cap.rdma ? 2146 caps->rdma = (number == 1); 2147 ice_debug(hw, ICE_DBG_INIT, "%s: rdma = %d\n", prefix, caps->rdma); 2202 if (caps->rdma) { 2204 caps->rdma = 0;
|
/linux-master/include/rdma/ |
H A D | uverbs_ioctl.h | 9 #include <rdma/uverbs_types.h> 11 #include <rdma/rdma_user_ioctl.h> 12 #include <rdma/ib_user_ioctl_verbs.h> 13 #include <rdma/ib_user_ioctl_cmds.h> 458 * in a header under include/uapi/rdma. 466 * zero. The structure must be declared in a header under include/uapi/rdma.
|
H A D | ib_verbs.h | 39 #include <uapi/rdma/ib_user_verbs.h> 40 #include <rdma/rdma_counter.h> 41 #include <rdma/restrack.h> 42 #include <rdma/signature.h> 43 #include <uapi/rdma/rdma_user_ioctl.h> 44 #include <uapi/rdma/ib_user_ioctl_verbs.h> 1510 struct rdma_cgroup *cg; /* owner rdma cgroup */ 2207 /* rdma netdev type - specifies protocol type */ 2214 * struct rdma_netdev - rdma netdev 2384 * rdma netde [all...] |