Searched refs:cm_id (Results 1 - 25 of 54) sorted by last modified time

123

/linux-master/drivers/nvme/target/
H A Drdma.c88 struct rdma_cm_id *cm_id; member in struct:nvmet_rdma_queue
122 struct rdma_cm_id *cm_id; member in struct:nvmet_rdma_port
629 struct rdma_cm_id *cm_id = rsp->queue->cm_id; local
634 ret = rdma_rw_ctx_signature_init(&rsp->rw, cm_id->qp,
635 cm_id->port_num, req->sg, req->sg_cnt,
639 ret = rdma_rw_ctx_init(&rsp->rw, cm_id->qp, cm_id->port_num,
648 struct rdma_cm_id *cm_id = rsp->queue->cm_id; local
713 struct rdma_cm_id *cm_id = rsp->queue->cm_id; local
784 struct rdma_cm_id *cm_id = rsp->queue->cm_id; local
1183 nvmet_rdma_find_get_device(struct rdma_cm_id *cm_id) argument
1396 nvmet_rdma_cm_reject(struct rdma_cm_id *cm_id, enum nvme_rdma_cm_status status) argument
1412 nvmet_rdma_alloc_queue(struct nvmet_rdma_device *ndev, struct rdma_cm_id *cm_id, struct rdma_cm_event *event) argument
1534 nvmet_rdma_cm_accept(struct rdma_cm_id *cm_id, struct nvmet_rdma_queue *queue, struct rdma_conn_param *p) argument
1558 nvmet_rdma_queue_connect(struct rdma_cm_id *cm_id, struct rdma_cm_event *event) argument
1694 nvmet_rdma_queue_connect_fail(struct rdma_cm_id *cm_id, struct nvmet_rdma_queue *queue) argument
1723 nvmet_rdma_device_removal(struct rdma_cm_id *cm_id, struct nvmet_rdma_queue *queue) argument
1755 nvmet_rdma_cm_handler(struct rdma_cm_id *cm_id, struct rdma_cm_event *event) argument
1844 struct rdma_cm_id *cm_id = xchg(&port->cm_id, NULL); local
1860 struct rdma_cm_id *cm_id; local
1993 struct rdma_cm_id *cm_id = port->cm_id; local
[all...]
/linux-master/drivers/thunderbolt/
H A Dusb4.c2387 * @cm_id: CM ID to assign
2393 int usb4_dp_port_set_cm_id(struct tb_port *port, int cm_id) argument
2407 val |= cm_id << ADP_DP_CS_2_CM_ID_SHIFT;
H A Dtb.h1370 int usb4_dp_port_set_cm_id(struct tb_port *port, int cm_id);
/linux-master/drivers/infiniband/core/
H A Dcm.c991 * The cm_id could be destroyed by the user before we exit timewait.
992 * To protect against this, we search for the cm_id after exiting
1029 static noinline void cm_destroy_id_wait_timeout(struct ib_cm_id *cm_id, argument
1034 cm_id_priv = container_of(cm_id, struct cm_id_private, id);
1035 pr_err("%s: cm_id=%p timed out. state %d -> %d, refcnt=%d\n", __func__,
1036 cm_id, old_state, cm_id->state, refcount_read(&cm_id_priv->refcount));
1039 static void cm_destroy_id(struct ib_cm_id *cm_id, int err) argument
1046 cm_id_priv = container_of(cm_id, struct cm_id_private, id);
1048 old_state = cm_id
1169 ib_destroy_cm_id(struct ib_cm_id *cm_id) argument
1198 ib_cm_listen(struct ib_cm_id *cm_id, __be64 service_id) argument
1494 ib_send_cm_req(struct ib_cm_id *cm_id, struct ib_cm_req_param *param) argument
2275 ib_send_cm_rep(struct ib_cm_id *cm_id, struct ib_cm_rep_param *param) argument
2348 ib_send_cm_rtu(struct ib_cm_id *cm_id, const void *private_data, u8 private_data_len) argument
2699 ib_send_cm_dreq(struct ib_cm_id *cm_id, const void *private_data, u8 private_data_len) argument
2766 ib_send_cm_drep(struct ib_cm_id *cm_id, const void *private_data, u8 private_data_len) argument
2977 ib_send_cm_rej(struct ib_cm_id *cm_id, enum ib_cm_rej_reason reason, void *ari, u8 ari_length, const void *private_data, u8 private_data_len) argument
3088 ib_send_cm_mra(struct ib_cm_id *cm_id, u8 service_timeout, const void *private_data, u8 private_data_len) argument
3488 ib_send_cm_sidr_req(struct ib_cm_id *cm_id, struct ib_cm_sidr_req_param *param) argument
3704 ib_send_cm_sidr_rep(struct ib_cm_id *cm_id, struct ib_cm_sidr_rep_param *param) argument
3913 cm_establish(struct ib_cm_id *cm_id) argument
3977 cm_migrate(struct ib_cm_id *cm_id) argument
3997 ib_cm_notify(struct ib_cm_id *cm_id, enum ib_event_type event) argument
4260 ib_cm_init_qp_attr(struct ib_cm_id *cm_id, struct ib_qp_attr *qp_attr, int *qp_attr_mask) argument
[all...]
H A Ducma.c92 struct rdma_cm_id *cm_id; member in struct:ucma_context
159 * Same as ucm_get_ctx but requires that ->cm_id->device is valid, eg that the
168 if (!ctx->cm_id->device) {
186 rdma_destroy_id(ctx->cm_id);
188 /* Reading the cm_id without holding a positive ref is not allowed */
189 ctx->cm_id = NULL;
216 struct rdma_cm_id *cm_id)
219 ctx->cm_id = cm_id;
283 if (ctx->cm_id
215 ucma_set_ctx_cm_id(struct ucma_context *ctx, struct rdma_cm_id *cm_id) argument
295 ucma_connect_event_handler(struct rdma_cm_id *cm_id, struct rdma_cm_event *event) argument
332 ucma_event_handler(struct rdma_cm_id *cm_id, struct rdma_cm_event *event) argument
444 struct rdma_cm_id *cm_id; local
869 ucma_query_device_addr(struct rdma_cm_id *cm_id, struct rdma_ucm_query_addr_resp *resp) argument
[all...]
H A Dnldev.c408 [RDMA_RESTRACK_CM_ID] = "cm_id",
565 struct rdma_cm_id *cm_id = &id_priv->id; local
567 if (port && port != cm_id->port_num)
570 if (cm_id->port_num &&
571 nla_put_u32(msg, RDMA_NLDEV_ATTR_PORT_INDEX, cm_id->port_num))
577 if (nla_put_u8(msg, RDMA_NLDEV_ATTR_RES_TYPE, cm_id->qp_type))
581 if (nla_put_u32(msg, RDMA_NLDEV_ATTR_RES_PS, cm_id->ps))
587 if (cm_id->route.addr.src_addr.ss_family &&
589 sizeof(cm_id->route.addr.src_addr),
590 &cm_id
[all...]
H A Diwcm.c97 struct iwcm_id_private *cm_id; member in struct:iwcm_work
118 * elements. The design pre-allocates them based on the cm_id type:
128 * One exception is when creating the cm_id for incoming connection requests.
130 * 1) in the event upcall, cm_event_handler(), for a listening cm_id. If
135 * If work elements cannot be allocated for the new connect request cm_id,
154 list_add(&work->free_list, &work->cm_id->work_free_list);
178 work->cm_id = cm_id_priv;
208 * Release a reference on cm_id. If the last reference is being
209 * released, free the cm_id and return 1.
222 static void add_ref(struct iw_cm_id *cm_id) argument
229 rem_ref(struct iw_cm_id *cm_id) argument
305 iw_cm_disconnect(struct iw_cm_id *cm_id, int abrupt) argument
372 destroy_cm_id(struct iw_cm_id *cm_id) argument
451 iw_destroy_cm_id(struct iw_cm_id *cm_id) argument
506 iw_cm_map(struct iw_cm_id *cm_id, bool active) argument
560 iw_cm_listen(struct iw_cm_id *cm_id, int backlog) argument
602 iw_cm_reject(struct iw_cm_id *cm_id, const void *private_data, u8 private_data_len) argument
640 iw_cm_accept(struct iw_cm_id *cm_id, struct iw_cm_conn_param *iw_param) argument
696 iw_cm_connect(struct iw_cm_id *cm_id, struct iw_cm_conn_param *iw_param) argument
767 struct iw_cm_id *cm_id; local
1064 cm_event_handler(struct iw_cm_id *cm_id, struct iw_cm_event *iw_event) argument
1156 iw_cm_init_qp_attr(struct iw_cm_id *cm_id, struct ib_qp_attr *qp_attr, int *qp_attr_mask) argument
[all...]
/linux-master/drivers/gpu/drm/amd/display/dmub/inc/
H A Ddmub_cmd.h2118 uint8_t cm_id; /**< CM ID */ member in struct:dpia_bw_allocation_notify_data
/linux-master/include/rdma/
H A Dib_verbs.h2614 int (*iw_connect)(struct iw_cm_id *cm_id,
2616 int (*iw_accept)(struct iw_cm_id *cm_id,
2618 int (*iw_reject)(struct iw_cm_id *cm_id, const void *pdata,
2620 int (*iw_create_listen)(struct iw_cm_id *cm_id, int backlog);
2621 int (*iw_destroy_listen)(struct iw_cm_id *cm_id);
/linux-master/drivers/infiniband/ulp/srpt/
H A Dib_srpt.c230 rdma_notify(ch->rdma_cm.cm_id, event->event);
232 ib_cm_notify(ch->ib_cm.cm_id, event->event);
1166 ret = ib_cm_init_qp_attr(ch->ib_cm.cm_id, &qp_attr, &attr_mask);
1196 ret = ib_cm_init_qp_attr(ch->ib_cm.cm_id, &qp_attr, &attr_mask);
1839 ret = rdma_create_qp(ch->rdma_cm.cm_id, sdev->pd, qp_init);
1840 ch->qp = ch->rdma_cm.cm_id->qp;
1951 ret = rdma_disconnect(ch->rdma_cm.cm_id);
1953 ret = ib_send_cm_dreq(ch->ib_cm.cm_id, NULL, 0);
1955 ret = ib_send_cm_drep(ch->ib_cm.cm_id, NULL, 0);
2104 rdma_destroy_id(ch->rdma_cm.cm_id);
2532 srpt_ib_cm_req_recv(struct ib_cm_id *cm_id, const struct ib_cm_req_event_param *param, void *private_data) argument
2546 srpt_rdma_cm_req_recv(struct rdma_cm_id *cm_id, struct rdma_cm_event *event) argument
2647 srpt_cm_handler(struct ib_cm_id *cm_id, const struct ib_cm_event *event) argument
2699 srpt_rdma_cm_handler(struct rdma_cm_id *cm_id, struct rdma_cm_event *event) argument
[all...]
/linux-master/drivers/infiniband/hw/irdma/
H A Dverbs.c1604 !iwqp->cm_id) {
1737 if (iwqp->cm_id && close_timer_started == 1)
H A Dverbs.h179 struct iw_cm_id *cm_id; member in struct:irdma_qp
H A Dtrace_cm.h201 TP_PROTO(struct irdma_cm_node *cm_node, struct iw_cm_id *cm_id,
203 TP_ARGS(cm_node, cm_id, type, status, caller),
206 __field(struct iw_cm_id *, cm_id)
222 __entry->cm_id = cm_id;
238 TP_printk("iwdev=%p caller=%pS cm_id=%p node=%p refcnt=%d vlan_id=%d accel=%d state=%s event_type=%s status=%d loc: %s rem: %s",
241 __entry->cm_id,
257 TP_PROTO(struct iw_cm_id *cm_id, enum iw_cm_event_type type,
259 TP_ARGS(cm_id, type, status, caller),
260 TP_STRUCT__entry(__field(struct iw_cm_id *, cm_id)
[all...]
H A Dcm.h245 struct iw_cm_id *cm_id; member in struct:irdma_cm_listener
296 struct iw_cm_id *cm_id; member in struct:irdma_cm_node
330 struct iw_cm_id *cm_id; member in struct:irdma_cm_info
394 int irdma_accept(struct iw_cm_id *cm_id, struct iw_cm_conn_param *conn_param);
395 int irdma_reject(struct iw_cm_id *cm_id, const void *pdata, u8 pdata_len);
396 int irdma_connect(struct iw_cm_id *cm_id, struct iw_cm_conn_param *conn_param);
397 int irdma_create_listen(struct iw_cm_id *cm_id, int backlog);
398 int irdma_destroy_listen(struct iw_cm_id *cm_id);
H A Dcm.c133 * @cm_id: upper layers cm struct for the event
137 struct iw_cm_id *cm_id,
140 memcpy(&event->local_addr, &cm_id->m_local_addr,
142 memcpy(&event->remote_addr, &cm_id->m_remote_addr,
155 * @cm_id: upper layer's cm info struct
160 struct iw_cm_id *cm_id,
167 trace_irdma_send_cm_event(cm_node, cm_id, type, status,
171 "CM: cm_node %p cm_id=%p state=%d accel=%d event_type=%d status=%d\n",
172 cm_node, cm_id, cm_node->accelerated, cm_node->state, type,
187 irdma_get_cmevent_info(cm_node, cm_id,
136 irdma_get_cmevent_info(struct irdma_cm_node *cm_node, struct iw_cm_id *cm_id, struct iw_cm_event *event) argument
159 irdma_send_cm_event(struct irdma_cm_node *cm_node, struct iw_cm_id *cm_id, enum iw_cm_event_type type, int status) argument
664 struct iw_cm_id *cm_id; local
3486 struct iw_cm_id *cm_id; local
3635 irdma_accept(struct iw_cm_id *cm_id, struct iw_cm_conn_param *conn_param) argument
3789 irdma_reject(struct iw_cm_id *cm_id, const void *pdata, u8 pdata_len) argument
3816 irdma_connect(struct iw_cm_id *cm_id, struct iw_cm_conn_param *conn_param) argument
3964 irdma_create_listen(struct iw_cm_id *cm_id, int backlog) argument
4085 irdma_destroy_listen(struct iw_cm_id *cm_id) argument
4139 struct iw_cm_id *cm_id; local
4197 struct iw_cm_id *cm_id = cm_node->cm_id; local
[all...]
/linux-master/drivers/infiniband/hw/cxgb4/
H A Diw_cxgb4.h842 struct iw_cm_id *cm_id; member in struct:c4iw_ep_common
906 static inline struct c4iw_ep *to_ep(struct iw_cm_id *cm_id) argument
908 return cm_id->provider_data;
911 static inline struct c4iw_listen_ep *to_listen_ep(struct iw_cm_id *cm_id) argument
913 return cm_id->provider_data;
961 int c4iw_connect(struct iw_cm_id *cm_id, struct iw_cm_conn_param *conn_param);
962 int c4iw_create_listen(struct iw_cm_id *cm_id, int backlog);
963 int c4iw_destroy_listen(struct iw_cm_id *cm_id);
964 int c4iw_accept_cr(struct iw_cm_id *cm_id, struct iw_cm_conn_param *conn_param);
965 int c4iw_reject_cr(struct iw_cm_id *cm_id, cons
[all...]
/linux-master/drivers/nvme/host/
H A Drdma.c95 struct rdma_cm_id *cm_id; member in struct:nvme_rdma_queue
151 static int nvme_rdma_cm_handler(struct rdma_cm_id *cm_id,
279 ret = rdma_create_qp(queue->cm_id, dev->pd, &init_attr);
281 queue->qp = queue->cm_id->qp;
367 nvme_rdma_find_get_device(struct rdma_cm_id *cm_id) argument
373 if (ndev->dev->node_guid == cm_id->device->node_guid &&
382 ndev->dev = cm_id->device;
437 * The cm_id object might have been destroyed during RDMA connection
496 queue->device = nvme_rdma_find_get_device(queue->cm_id);
498 dev_err(queue->cm_id
1794 struct rdma_cm_id *cm_id = queue->cm_id; local
1888 nvme_rdma_cm_handler(struct rdma_cm_id *cm_id, struct rdma_cm_event *ev) argument
[all...]
/linux-master/fs/smb/server/
H A Dtransport_rdma.c84 struct rdma_cm_id *cm_id; member in struct:smb_direct_listener
103 struct rdma_cm_id *cm_id; member in struct:smb_direct_transport
269 ib_dma_unmap_single(t->cm_id->device, recvmsg->sge.addr,
295 ib_dma_unmap_single(t->cm_id->device, recvmsg->sge.addr,
338 rdma_disconnect(t->cm_id);
360 static struct smb_direct_transport *alloc_transport(struct rdma_cm_id *cm_id) argument
369 t->cm_id = cm_id;
370 cm_id->context = t;
452 if (t->cm_id)
1523 smb_direct_cm_handler(struct rdma_cm_id *cm_id, struct rdma_cm_event *event) argument
2075 smb_direct_listen_handler(struct rdma_cm_id *cm_id, struct rdma_cm_event *event) argument
2102 struct rdma_cm_id *cm_id; local
[all...]
/linux-master/drivers/infiniband/ulp/rtrs/
H A Drtrs.c218 rdma_notify(con->cm_id, IB_EVENT_COMM_EST);
235 struct rdma_cm_id *cm_id = con->cm_id; local
239 cq = ib_alloc_cq(cm_id->device, con, nr_cqe, cq_vector,
242 cq = ib_cq_pool_get(cm_id->device, nr_cqe, cq_vector, poll_ctx);
259 struct rdma_cm_id *cm_id = con->cm_id; local
274 ret = rdma_create_qp(cm_id, pd, &init_attr);
279 con->qp = cm_id->qp;
321 rdma_destroy_qp(con->cm_id);
[all...]
H A Drtrs-clt.c1666 clt_path->s.dev = rtrs_ib_dev_find_or_add(con->c.cm_id->device,
1760 rdma_disconnect(con->c.cm_id);
1767 rdma_destroy_id(con->c.cm_id);
1768 con->c.cm_id = NULL;
1783 err = rdma_resolve_route(con->c.cm_id, RTRS_CONNECT_TIMEOUT_MS);
1817 err = rdma_connect_locked(con->c.cm_id, &param);
1903 clt_path->hca_port = con->c.cm_id->port_num;
1906 clt_path->s.src_addr = con->c.cm_id->route.addr.src_addr;
1932 rej_msg = rdma_reject_msg(con->c.cm_id, status);
1933 msg = rdma_consumer_reject_data(con->c.cm_id, e
1976 rtrs_clt_rdma_cm_handler(struct rdma_cm_id *cm_id, struct rdma_cm_event *ev) argument
2056 struct rdma_cm_id *cm_id; local
[all...]
H A Drtrs-srv.c784 static int rtrs_rdma_do_reject(struct rdma_cm_id *cm_id, int errno);
1541 rdma_disconnect(con->c.cm_id);
1567 rdma_destroy_id(con->c.cm_id);
1583 struct rdma_cm_id *cm_id)
1607 err = rdma_accept(cm_id, &param);
1614 static int rtrs_rdma_do_reject(struct rdma_cm_id *cm_id, int errno) argument
1625 err = rdma_reject(cm_id, &msg, sizeof(msg), IB_CM_REJ_CONSUMER_DEFINED);
1647 struct rdma_cm_id *cm_id,
1665 con->c.cm_id = cm_id;
1582 rtrs_rdma_do_accept(struct rtrs_srv_path *srv_path, struct rdma_cm_id *cm_id) argument
1646 create_con(struct rtrs_srv_path *srv_path, struct rdma_cm_id *cm_id, unsigned int cid) argument
1731 __alloc_path(struct rtrs_srv_sess *srv, struct rdma_cm_id *cm_id, unsigned int con_num, unsigned int recon_cnt, const uuid_t *uuid) argument
1831 rtrs_rdma_connect(struct rdma_cm_id *cm_id, const struct rtrs_msg_conn_req *msg, size_t len) argument
1954 rtrs_srv_rdma_cm_handler(struct rdma_cm_id *cm_id, struct rdma_cm_event *ev) argument
2002 struct rdma_cm_id *cm_id; local
[all...]
/linux-master/drivers/infiniband/sw/siw/
H A Dsiw_cm.c318 id = cep->listen_cep->cm_id;
320 id = cep->cm_id;
363 if (!cep->cm_id)
366 cep->cm_id->rem_ref(cep->cm_id);
367 cep->cm_id = NULL;
410 if (cep->cm_id) {
1076 if (cep->cm_id)
1083 if (cep->cm_id) {
1135 if (cep->cm_id)
[all...]
H A Dsiw_cm.h38 struct iw_cm_id *cm_id; member in struct:siw_cep
/linux-master/drivers/infiniband/ulp/srp/
H A Dib_srp.h178 struct ib_cm_id *cm_id; member in struct:srp_rdma_ch::__anon1029::ib_cm
181 struct rdma_cm_id *cm_id; member in struct:srp_rdma_ch::__anon1029::rdma_cm
H A Dib_srp.c145 static int srp_ib_cm_handler(struct ib_cm_id *cm_id,
147 static int srp_rdma_cm_handler(struct rdma_cm_id *cm_id,
313 if (ch->ib_cm.cm_id)
314 ib_destroy_cm_id(ch->ib_cm.cm_id);
315 ch->ib_cm.cm_id = new_cm_id;
364 swap(ch->rdma_cm.cm_id, new_cm_id);
567 ret = rdma_create_qp(ch->rdma_cm.cm_id, dev->pd, init_attr);
568 qp = ch->rdma_cm.cm_id->qp;
617 rdma_destroy_qp(ch->rdma_cm.cm_id);
646 if (ch->rdma_cm.cm_id) {
2330 srp_cm_rep_handler(struct ib_cm_id *cm_id, const struct srp_login_rsp *lrsp, struct srp_rdma_ch *ch) argument
2422 srp_ib_cm_rej_handler(struct ib_cm_id *cm_id, const struct ib_cm_event *event, struct srp_rdma_ch *ch) argument
2508 srp_ib_cm_handler(struct ib_cm_id *cm_id, const struct ib_cm_event *event) argument
2620 srp_rdma_cm_handler(struct rdma_cm_id *cm_id, struct rdma_cm_event *event) argument
[all...]

Completed in 647 milliseconds

123