/linux-master/drivers/nvme/target/ |
H A D | rdma.c | 88 struct rdma_cm_id *cm_id; member in struct:nvmet_rdma_queue 122 struct rdma_cm_id *cm_id; member in struct:nvmet_rdma_port 629 struct rdma_cm_id *cm_id = rsp->queue->cm_id; local 634 ret = rdma_rw_ctx_signature_init(&rsp->rw, cm_id->qp, 635 cm_id->port_num, req->sg, req->sg_cnt, 639 ret = rdma_rw_ctx_init(&rsp->rw, cm_id->qp, cm_id->port_num, 648 struct rdma_cm_id *cm_id = rsp->queue->cm_id; local 713 struct rdma_cm_id *cm_id = rsp->queue->cm_id; local 784 struct rdma_cm_id *cm_id = rsp->queue->cm_id; local 1183 nvmet_rdma_find_get_device(struct rdma_cm_id *cm_id) argument 1396 nvmet_rdma_cm_reject(struct rdma_cm_id *cm_id, enum nvme_rdma_cm_status status) argument 1412 nvmet_rdma_alloc_queue(struct nvmet_rdma_device *ndev, struct rdma_cm_id *cm_id, struct rdma_cm_event *event) argument 1534 nvmet_rdma_cm_accept(struct rdma_cm_id *cm_id, struct nvmet_rdma_queue *queue, struct rdma_conn_param *p) argument 1558 nvmet_rdma_queue_connect(struct rdma_cm_id *cm_id, struct rdma_cm_event *event) argument 1694 nvmet_rdma_queue_connect_fail(struct rdma_cm_id *cm_id, struct nvmet_rdma_queue *queue) argument 1723 nvmet_rdma_device_removal(struct rdma_cm_id *cm_id, struct nvmet_rdma_queue *queue) argument 1755 nvmet_rdma_cm_handler(struct rdma_cm_id *cm_id, struct rdma_cm_event *event) argument 1844 struct rdma_cm_id *cm_id = xchg(&port->cm_id, NULL); local 1860 struct rdma_cm_id *cm_id; local 1993 struct rdma_cm_id *cm_id = port->cm_id; local [all...] |
/linux-master/drivers/thunderbolt/ |
H A D | usb4.c | 2387 * @cm_id: CM ID to assign 2393 int usb4_dp_port_set_cm_id(struct tb_port *port, int cm_id) argument 2407 val |= cm_id << ADP_DP_CS_2_CM_ID_SHIFT;
|
H A D | tb.h | 1370 int usb4_dp_port_set_cm_id(struct tb_port *port, int cm_id);
|
/linux-master/drivers/infiniband/core/ |
H A D | cm.c | 991 * The cm_id could be destroyed by the user before we exit timewait. 992 * To protect against this, we search for the cm_id after exiting 1029 static noinline void cm_destroy_id_wait_timeout(struct ib_cm_id *cm_id, argument 1034 cm_id_priv = container_of(cm_id, struct cm_id_private, id); 1035 pr_err("%s: cm_id=%p timed out. state %d -> %d, refcnt=%d\n", __func__, 1036 cm_id, old_state, cm_id->state, refcount_read(&cm_id_priv->refcount)); 1039 static void cm_destroy_id(struct ib_cm_id *cm_id, int err) argument 1046 cm_id_priv = container_of(cm_id, struct cm_id_private, id); 1048 old_state = cm_id 1169 ib_destroy_cm_id(struct ib_cm_id *cm_id) argument 1198 ib_cm_listen(struct ib_cm_id *cm_id, __be64 service_id) argument 1494 ib_send_cm_req(struct ib_cm_id *cm_id, struct ib_cm_req_param *param) argument 2275 ib_send_cm_rep(struct ib_cm_id *cm_id, struct ib_cm_rep_param *param) argument 2348 ib_send_cm_rtu(struct ib_cm_id *cm_id, const void *private_data, u8 private_data_len) argument 2699 ib_send_cm_dreq(struct ib_cm_id *cm_id, const void *private_data, u8 private_data_len) argument 2766 ib_send_cm_drep(struct ib_cm_id *cm_id, const void *private_data, u8 private_data_len) argument 2977 ib_send_cm_rej(struct ib_cm_id *cm_id, enum ib_cm_rej_reason reason, void *ari, u8 ari_length, const void *private_data, u8 private_data_len) argument 3088 ib_send_cm_mra(struct ib_cm_id *cm_id, u8 service_timeout, const void *private_data, u8 private_data_len) argument 3488 ib_send_cm_sidr_req(struct ib_cm_id *cm_id, struct ib_cm_sidr_req_param *param) argument 3704 ib_send_cm_sidr_rep(struct ib_cm_id *cm_id, struct ib_cm_sidr_rep_param *param) argument 3913 cm_establish(struct ib_cm_id *cm_id) argument 3977 cm_migrate(struct ib_cm_id *cm_id) argument 3997 ib_cm_notify(struct ib_cm_id *cm_id, enum ib_event_type event) argument 4260 ib_cm_init_qp_attr(struct ib_cm_id *cm_id, struct ib_qp_attr *qp_attr, int *qp_attr_mask) argument [all...] |
H A D | ucma.c | 92 struct rdma_cm_id *cm_id; member in struct:ucma_context 159 * Same as ucm_get_ctx but requires that ->cm_id->device is valid, eg that the 168 if (!ctx->cm_id->device) { 186 rdma_destroy_id(ctx->cm_id); 188 /* Reading the cm_id without holding a positive ref is not allowed */ 189 ctx->cm_id = NULL; 216 struct rdma_cm_id *cm_id) 219 ctx->cm_id = cm_id; 283 if (ctx->cm_id 215 ucma_set_ctx_cm_id(struct ucma_context *ctx, struct rdma_cm_id *cm_id) argument 295 ucma_connect_event_handler(struct rdma_cm_id *cm_id, struct rdma_cm_event *event) argument 332 ucma_event_handler(struct rdma_cm_id *cm_id, struct rdma_cm_event *event) argument 444 struct rdma_cm_id *cm_id; local 869 ucma_query_device_addr(struct rdma_cm_id *cm_id, struct rdma_ucm_query_addr_resp *resp) argument [all...] |
H A D | nldev.c | 408 [RDMA_RESTRACK_CM_ID] = "cm_id", 565 struct rdma_cm_id *cm_id = &id_priv->id; local 567 if (port && port != cm_id->port_num) 570 if (cm_id->port_num && 571 nla_put_u32(msg, RDMA_NLDEV_ATTR_PORT_INDEX, cm_id->port_num)) 577 if (nla_put_u8(msg, RDMA_NLDEV_ATTR_RES_TYPE, cm_id->qp_type)) 581 if (nla_put_u32(msg, RDMA_NLDEV_ATTR_RES_PS, cm_id->ps)) 587 if (cm_id->route.addr.src_addr.ss_family && 589 sizeof(cm_id->route.addr.src_addr), 590 &cm_id [all...] |
H A D | iwcm.c | 97 struct iwcm_id_private *cm_id; member in struct:iwcm_work 118 * elements. The design pre-allocates them based on the cm_id type: 128 * One exception is when creating the cm_id for incoming connection requests. 130 * 1) in the event upcall, cm_event_handler(), for a listening cm_id. If 135 * If work elements cannot be allocated for the new connect request cm_id, 154 list_add(&work->free_list, &work->cm_id->work_free_list); 178 work->cm_id = cm_id_priv; 208 * Release a reference on cm_id. If the last reference is being 209 * released, free the cm_id and return 1. 222 static void add_ref(struct iw_cm_id *cm_id) argument 229 rem_ref(struct iw_cm_id *cm_id) argument 305 iw_cm_disconnect(struct iw_cm_id *cm_id, int abrupt) argument 372 destroy_cm_id(struct iw_cm_id *cm_id) argument 451 iw_destroy_cm_id(struct iw_cm_id *cm_id) argument 506 iw_cm_map(struct iw_cm_id *cm_id, bool active) argument 560 iw_cm_listen(struct iw_cm_id *cm_id, int backlog) argument 602 iw_cm_reject(struct iw_cm_id *cm_id, const void *private_data, u8 private_data_len) argument 640 iw_cm_accept(struct iw_cm_id *cm_id, struct iw_cm_conn_param *iw_param) argument 696 iw_cm_connect(struct iw_cm_id *cm_id, struct iw_cm_conn_param *iw_param) argument 767 struct iw_cm_id *cm_id; local 1064 cm_event_handler(struct iw_cm_id *cm_id, struct iw_cm_event *iw_event) argument 1156 iw_cm_init_qp_attr(struct iw_cm_id *cm_id, struct ib_qp_attr *qp_attr, int *qp_attr_mask) argument [all...] |
/linux-master/drivers/gpu/drm/amd/display/dmub/inc/ |
H A D | dmub_cmd.h | 2118 uint8_t cm_id; /**< CM ID */ member in struct:dpia_bw_allocation_notify_data
|
/linux-master/include/rdma/ |
H A D | ib_verbs.h | 2614 int (*iw_connect)(struct iw_cm_id *cm_id, 2616 int (*iw_accept)(struct iw_cm_id *cm_id, 2618 int (*iw_reject)(struct iw_cm_id *cm_id, const void *pdata, 2620 int (*iw_create_listen)(struct iw_cm_id *cm_id, int backlog); 2621 int (*iw_destroy_listen)(struct iw_cm_id *cm_id);
|
/linux-master/drivers/infiniband/ulp/srpt/ |
H A D | ib_srpt.c | 230 rdma_notify(ch->rdma_cm.cm_id, event->event); 232 ib_cm_notify(ch->ib_cm.cm_id, event->event); 1166 ret = ib_cm_init_qp_attr(ch->ib_cm.cm_id, &qp_attr, &attr_mask); 1196 ret = ib_cm_init_qp_attr(ch->ib_cm.cm_id, &qp_attr, &attr_mask); 1839 ret = rdma_create_qp(ch->rdma_cm.cm_id, sdev->pd, qp_init); 1840 ch->qp = ch->rdma_cm.cm_id->qp; 1951 ret = rdma_disconnect(ch->rdma_cm.cm_id); 1953 ret = ib_send_cm_dreq(ch->ib_cm.cm_id, NULL, 0); 1955 ret = ib_send_cm_drep(ch->ib_cm.cm_id, NULL, 0); 2104 rdma_destroy_id(ch->rdma_cm.cm_id); 2532 srpt_ib_cm_req_recv(struct ib_cm_id *cm_id, const struct ib_cm_req_event_param *param, void *private_data) argument 2546 srpt_rdma_cm_req_recv(struct rdma_cm_id *cm_id, struct rdma_cm_event *event) argument 2647 srpt_cm_handler(struct ib_cm_id *cm_id, const struct ib_cm_event *event) argument 2699 srpt_rdma_cm_handler(struct rdma_cm_id *cm_id, struct rdma_cm_event *event) argument [all...] |
/linux-master/drivers/infiniband/hw/irdma/ |
H A D | verbs.c | 1604 !iwqp->cm_id) { 1737 if (iwqp->cm_id && close_timer_started == 1)
|
H A D | verbs.h | 179 struct iw_cm_id *cm_id; member in struct:irdma_qp
|
H A D | trace_cm.h | 201 TP_PROTO(struct irdma_cm_node *cm_node, struct iw_cm_id *cm_id, 203 TP_ARGS(cm_node, cm_id, type, status, caller), 206 __field(struct iw_cm_id *, cm_id) 222 __entry->cm_id = cm_id; 238 TP_printk("iwdev=%p caller=%pS cm_id=%p node=%p refcnt=%d vlan_id=%d accel=%d state=%s event_type=%s status=%d loc: %s rem: %s", 241 __entry->cm_id, 257 TP_PROTO(struct iw_cm_id *cm_id, enum iw_cm_event_type type, 259 TP_ARGS(cm_id, type, status, caller), 260 TP_STRUCT__entry(__field(struct iw_cm_id *, cm_id) [all...] |
H A D | cm.h | 245 struct iw_cm_id *cm_id; member in struct:irdma_cm_listener 296 struct iw_cm_id *cm_id; member in struct:irdma_cm_node 330 struct iw_cm_id *cm_id; member in struct:irdma_cm_info 394 int irdma_accept(struct iw_cm_id *cm_id, struct iw_cm_conn_param *conn_param); 395 int irdma_reject(struct iw_cm_id *cm_id, const void *pdata, u8 pdata_len); 396 int irdma_connect(struct iw_cm_id *cm_id, struct iw_cm_conn_param *conn_param); 397 int irdma_create_listen(struct iw_cm_id *cm_id, int backlog); 398 int irdma_destroy_listen(struct iw_cm_id *cm_id);
|
H A D | cm.c | 133 * @cm_id: upper layers cm struct for the event 137 struct iw_cm_id *cm_id, 140 memcpy(&event->local_addr, &cm_id->m_local_addr, 142 memcpy(&event->remote_addr, &cm_id->m_remote_addr, 155 * @cm_id: upper layer's cm info struct 160 struct iw_cm_id *cm_id, 167 trace_irdma_send_cm_event(cm_node, cm_id, type, status, 171 "CM: cm_node %p cm_id=%p state=%d accel=%d event_type=%d status=%d\n", 172 cm_node, cm_id, cm_node->accelerated, cm_node->state, type, 187 irdma_get_cmevent_info(cm_node, cm_id, 136 irdma_get_cmevent_info(struct irdma_cm_node *cm_node, struct iw_cm_id *cm_id, struct iw_cm_event *event) argument 159 irdma_send_cm_event(struct irdma_cm_node *cm_node, struct iw_cm_id *cm_id, enum iw_cm_event_type type, int status) argument 664 struct iw_cm_id *cm_id; local 3486 struct iw_cm_id *cm_id; local 3635 irdma_accept(struct iw_cm_id *cm_id, struct iw_cm_conn_param *conn_param) argument 3789 irdma_reject(struct iw_cm_id *cm_id, const void *pdata, u8 pdata_len) argument 3816 irdma_connect(struct iw_cm_id *cm_id, struct iw_cm_conn_param *conn_param) argument 3964 irdma_create_listen(struct iw_cm_id *cm_id, int backlog) argument 4085 irdma_destroy_listen(struct iw_cm_id *cm_id) argument 4139 struct iw_cm_id *cm_id; local 4197 struct iw_cm_id *cm_id = cm_node->cm_id; local [all...] |
/linux-master/drivers/infiniband/hw/cxgb4/ |
H A D | iw_cxgb4.h | 842 struct iw_cm_id *cm_id; member in struct:c4iw_ep_common 906 static inline struct c4iw_ep *to_ep(struct iw_cm_id *cm_id) argument 908 return cm_id->provider_data; 911 static inline struct c4iw_listen_ep *to_listen_ep(struct iw_cm_id *cm_id) argument 913 return cm_id->provider_data; 961 int c4iw_connect(struct iw_cm_id *cm_id, struct iw_cm_conn_param *conn_param); 962 int c4iw_create_listen(struct iw_cm_id *cm_id, int backlog); 963 int c4iw_destroy_listen(struct iw_cm_id *cm_id); 964 int c4iw_accept_cr(struct iw_cm_id *cm_id, struct iw_cm_conn_param *conn_param); 965 int c4iw_reject_cr(struct iw_cm_id *cm_id, cons [all...] |
/linux-master/drivers/nvme/host/ |
H A D | rdma.c | 95 struct rdma_cm_id *cm_id; member in struct:nvme_rdma_queue 151 static int nvme_rdma_cm_handler(struct rdma_cm_id *cm_id, 279 ret = rdma_create_qp(queue->cm_id, dev->pd, &init_attr); 281 queue->qp = queue->cm_id->qp; 367 nvme_rdma_find_get_device(struct rdma_cm_id *cm_id) argument 373 if (ndev->dev->node_guid == cm_id->device->node_guid && 382 ndev->dev = cm_id->device; 437 * The cm_id object might have been destroyed during RDMA connection 496 queue->device = nvme_rdma_find_get_device(queue->cm_id); 498 dev_err(queue->cm_id 1794 struct rdma_cm_id *cm_id = queue->cm_id; local 1888 nvme_rdma_cm_handler(struct rdma_cm_id *cm_id, struct rdma_cm_event *ev) argument [all...] |
/linux-master/fs/smb/server/ |
H A D | transport_rdma.c | 84 struct rdma_cm_id *cm_id; member in struct:smb_direct_listener 103 struct rdma_cm_id *cm_id; member in struct:smb_direct_transport 269 ib_dma_unmap_single(t->cm_id->device, recvmsg->sge.addr, 295 ib_dma_unmap_single(t->cm_id->device, recvmsg->sge.addr, 338 rdma_disconnect(t->cm_id); 360 static struct smb_direct_transport *alloc_transport(struct rdma_cm_id *cm_id) argument 369 t->cm_id = cm_id; 370 cm_id->context = t; 452 if (t->cm_id) 1523 smb_direct_cm_handler(struct rdma_cm_id *cm_id, struct rdma_cm_event *event) argument 2075 smb_direct_listen_handler(struct rdma_cm_id *cm_id, struct rdma_cm_event *event) argument 2102 struct rdma_cm_id *cm_id; local [all...] |
/linux-master/drivers/infiniband/ulp/rtrs/ |
H A D | rtrs.c | 218 rdma_notify(con->cm_id, IB_EVENT_COMM_EST); 235 struct rdma_cm_id *cm_id = con->cm_id; local 239 cq = ib_alloc_cq(cm_id->device, con, nr_cqe, cq_vector, 242 cq = ib_cq_pool_get(cm_id->device, nr_cqe, cq_vector, poll_ctx); 259 struct rdma_cm_id *cm_id = con->cm_id; local 274 ret = rdma_create_qp(cm_id, pd, &init_attr); 279 con->qp = cm_id->qp; 321 rdma_destroy_qp(con->cm_id); [all...] |
H A D | rtrs-clt.c | 1666 clt_path->s.dev = rtrs_ib_dev_find_or_add(con->c.cm_id->device, 1760 rdma_disconnect(con->c.cm_id); 1767 rdma_destroy_id(con->c.cm_id); 1768 con->c.cm_id = NULL; 1783 err = rdma_resolve_route(con->c.cm_id, RTRS_CONNECT_TIMEOUT_MS); 1817 err = rdma_connect_locked(con->c.cm_id, ¶m); 1903 clt_path->hca_port = con->c.cm_id->port_num; 1906 clt_path->s.src_addr = con->c.cm_id->route.addr.src_addr; 1932 rej_msg = rdma_reject_msg(con->c.cm_id, status); 1933 msg = rdma_consumer_reject_data(con->c.cm_id, e 1976 rtrs_clt_rdma_cm_handler(struct rdma_cm_id *cm_id, struct rdma_cm_event *ev) argument 2056 struct rdma_cm_id *cm_id; local [all...] |
H A D | rtrs-srv.c | 784 static int rtrs_rdma_do_reject(struct rdma_cm_id *cm_id, int errno); 1541 rdma_disconnect(con->c.cm_id); 1567 rdma_destroy_id(con->c.cm_id); 1583 struct rdma_cm_id *cm_id) 1607 err = rdma_accept(cm_id, ¶m); 1614 static int rtrs_rdma_do_reject(struct rdma_cm_id *cm_id, int errno) argument 1625 err = rdma_reject(cm_id, &msg, sizeof(msg), IB_CM_REJ_CONSUMER_DEFINED); 1647 struct rdma_cm_id *cm_id, 1665 con->c.cm_id = cm_id; 1582 rtrs_rdma_do_accept(struct rtrs_srv_path *srv_path, struct rdma_cm_id *cm_id) argument 1646 create_con(struct rtrs_srv_path *srv_path, struct rdma_cm_id *cm_id, unsigned int cid) argument 1731 __alloc_path(struct rtrs_srv_sess *srv, struct rdma_cm_id *cm_id, unsigned int con_num, unsigned int recon_cnt, const uuid_t *uuid) argument 1831 rtrs_rdma_connect(struct rdma_cm_id *cm_id, const struct rtrs_msg_conn_req *msg, size_t len) argument 1954 rtrs_srv_rdma_cm_handler(struct rdma_cm_id *cm_id, struct rdma_cm_event *ev) argument 2002 struct rdma_cm_id *cm_id; local [all...] |
/linux-master/drivers/infiniband/sw/siw/ |
H A D | siw_cm.c | 318 id = cep->listen_cep->cm_id; 320 id = cep->cm_id; 363 if (!cep->cm_id) 366 cep->cm_id->rem_ref(cep->cm_id); 367 cep->cm_id = NULL; 410 if (cep->cm_id) { 1076 if (cep->cm_id) 1083 if (cep->cm_id) { 1135 if (cep->cm_id) [all...] |
H A D | siw_cm.h | 38 struct iw_cm_id *cm_id; member in struct:siw_cep
|
/linux-master/drivers/infiniband/ulp/srp/ |
H A D | ib_srp.h | 178 struct ib_cm_id *cm_id; member in struct:srp_rdma_ch::__anon1029::ib_cm 181 struct rdma_cm_id *cm_id; member in struct:srp_rdma_ch::__anon1029::rdma_cm
|
H A D | ib_srp.c | 145 static int srp_ib_cm_handler(struct ib_cm_id *cm_id, 147 static int srp_rdma_cm_handler(struct rdma_cm_id *cm_id, 313 if (ch->ib_cm.cm_id) 314 ib_destroy_cm_id(ch->ib_cm.cm_id); 315 ch->ib_cm.cm_id = new_cm_id; 364 swap(ch->rdma_cm.cm_id, new_cm_id); 567 ret = rdma_create_qp(ch->rdma_cm.cm_id, dev->pd, init_attr); 568 qp = ch->rdma_cm.cm_id->qp; 617 rdma_destroy_qp(ch->rdma_cm.cm_id); 646 if (ch->rdma_cm.cm_id) { 2330 srp_cm_rep_handler(struct ib_cm_id *cm_id, const struct srp_login_rsp *lrsp, struct srp_rdma_ch *ch) argument 2422 srp_ib_cm_rej_handler(struct ib_cm_id *cm_id, const struct ib_cm_event *event, struct srp_rdma_ch *ch) argument 2508 srp_ib_cm_handler(struct ib_cm_id *cm_id, const struct ib_cm_event *event) argument 2620 srp_rdma_cm_handler(struct rdma_cm_id *cm_id, struct rdma_cm_event *event) argument [all...] |