/linux-master/drivers/infiniband/hw/qedr/ |
H A D | qedr_iw_cm.h | 34 int qedr_iw_connect(struct iw_cm_id *cm_id, 37 int qedr_iw_create_listen(struct iw_cm_id *cm_id, int backlog); 39 int qedr_iw_destroy_listen(struct iw_cm_id *cm_id); 41 int qedr_iw_accept(struct iw_cm_id *cm_id, struct iw_cm_conn_param *conn_param); 43 int qedr_iw_reject(struct iw_cm_id *cm_id, const void *pdata, u8 pdata_len);
|
H A D | qedr_iw_cm.c | 97 if (ep->cm_id) 98 ep->cm_id->rem_ref(ep->cm_id); 135 listener->cm_id->event_handler(listener->cm_id, &event); 166 if (ep->cm_id) 167 ep->cm_id->event_handler(ep->cm_id, &event); 175 if (ep->cm_id) 239 if (ep->cm_id) 531 qedr_iw_connect(struct iw_cm_id *cm_id, struct iw_cm_conn_param *conn_param) argument 657 qedr_iw_create_listen(struct iw_cm_id *cm_id, int backlog) argument 721 qedr_iw_destroy_listen(struct iw_cm_id *cm_id) argument 736 qedr_iw_accept(struct iw_cm_id *cm_id, struct iw_cm_conn_param *conn_param) argument 784 qedr_iw_reject(struct iw_cm_id *cm_id, const void *pdata, u8 pdata_len) argument [all...] |
/linux-master/include/rdma/ |
H A D | iw_cm.h | 39 * @cm_id: The IW CM identifier associated with the event. 42 typedef int (*iw_cm_handler)(struct iw_cm_id *cm_id, 50 * @cm_id: The IW CM identifier associated with the event. 53 typedef int (*iw_event_handler)(struct iw_cm_id *cm_id, 67 /* Used by provider to add and remove refs on IW cm_id */ 110 * @cm_id: The previously created IW CM identifier to destroy. 115 void iw_destroy_cm_id(struct iw_cm_id *cm_id); 121 * @cm_id: The IW CM identifier. 128 int iw_cm_listen(struct iw_cm_id *cm_id, int backlog); 133 * @cm_id [all...] |
H A D | ib_cm.h | 275 * @cm_id: Communication identifier associated with the reported event. 280 * new @cm_id. The new @cm_id is returned to the user through this callback. 281 * Clients are responsible for destroying the new @cm_id. For peer-to-peer 282 * IB_CM_REQ_RECEIVED and all other events, the returned @cm_id corresponds 287 * destroy the @cm_id after the callback completes. 289 typedef int (*ib_cm_handler)(struct ib_cm_id *cm_id, 306 * @device: Device associated with the cm_id. All related communication will 321 * @cm_id: Connection identifier to destroy. 325 void ib_destroy_cm_id(struct ib_cm_id *cm_id); [all...] |
/linux-master/net/rds/ |
H A D | rdma_transport.h | 20 int rds_rdma_cm_event_handler(struct rdma_cm_id *cm_id, 22 int rds6_rdma_cm_event_handler(struct rdma_cm_id *cm_id,
|
H A D | rdma_transport.c | 40 /* Global IPv4 and IPv6 RDS RDMA listener cm_id */ 49 static int rds_rdma_cm_event_handler_cmn(struct rdma_cm_id *cm_id, argument 54 struct rds_connection *conn = cm_id->context; 60 rdsdebug("conn %p id %p handling event %u (%s)\n", conn, cm_id, 63 if (cm_id->device->node_type == RDMA_NODE_IB_CA) 72 * right away. We return 0 so cm_id doesn't get 85 ret = trans->cm_handle_connect(cm_id, event, isv6); 90 rdma_set_service_type(cm_id, conn->c_tos); 91 rdma_set_min_rnr_timer(cm_id, IB_RNR_TIMER_000_32); 93 ret = rdma_resolve_route(cm_id, 183 rds_rdma_cm_event_handler(struct rdma_cm_id *cm_id, struct rdma_cm_event *event) argument 190 rds6_rdma_cm_event_handler(struct rdma_cm_id *cm_id, struct rdma_cm_event *event) argument 201 struct rdma_cm_id *cm_id; local [all...] |
H A D | ib.c | 410 struct rdma_cm_id *cm_id; local 422 cm_id = rdma_create_id(&init_net, rds_rdma_cm_event_handler, 424 if (IS_ERR(cm_id)) 425 return PTR_ERR(cm_id); 475 ret = rdma_bind_addr(cm_id, sa); 478 if (ret || !cm_id->device || 479 cm_id->device->node_type != RDMA_NODE_IB_CA) 484 cm_id->device ? cm_id->device->node_type : -1); 487 rdma_destroy_id(cm_id); [all...] |
/linux-master/drivers/infiniband/core/ |
H A D | iwcm.c | 97 struct iwcm_id_private *cm_id; member in struct:iwcm_work 118 * elements. The design pre-allocates them based on the cm_id type: 128 * One exception is when creating the cm_id for incoming connection requests. 130 * 1) in the event upcall, cm_event_handler(), for a listening cm_id. If 135 * If work elements cannot be allocated for the new connect request cm_id, 154 list_add(&work->free_list, &work->cm_id->work_free_list); 178 work->cm_id = cm_id_priv; 208 * Release a reference on cm_id. If the last reference is being 209 * released, free the cm_id and return 1. 222 static void add_ref(struct iw_cm_id *cm_id) argument 229 rem_ref(struct iw_cm_id *cm_id) argument 305 iw_cm_disconnect(struct iw_cm_id *cm_id, int abrupt) argument 372 destroy_cm_id(struct iw_cm_id *cm_id) argument 451 iw_destroy_cm_id(struct iw_cm_id *cm_id) argument 506 iw_cm_map(struct iw_cm_id *cm_id, bool active) argument 560 iw_cm_listen(struct iw_cm_id *cm_id, int backlog) argument 602 iw_cm_reject(struct iw_cm_id *cm_id, const void *private_data, u8 private_data_len) argument 640 iw_cm_accept(struct iw_cm_id *cm_id, struct iw_cm_conn_param *iw_param) argument 696 iw_cm_connect(struct iw_cm_id *cm_id, struct iw_cm_conn_param *iw_param) argument 767 struct iw_cm_id *cm_id; local 1064 cm_event_handler(struct iw_cm_id *cm_id, struct iw_cm_event *iw_event) argument 1156 iw_cm_init_qp_attr(struct iw_cm_id *cm_id, struct ib_qp_attr *qp_attr, int *qp_attr_mask) argument [all...] |
H A D | cm_trace.h | 139 const struct ib_cm_id *cm_id 142 TP_ARGS(cm_id), 145 __field(const void *, cm_id) /* for eBPF scripts */ 153 __entry->cm_id = cm_id; 154 __entry->local_id = be32_to_cpu(cm_id->local_id); 155 __entry->remote_id = be32_to_cpu(cm_id->remote_id); 156 __entry->state = cm_id->state; 157 __entry->lap_state = cm_id->lap_state; 171 const struct ib_cm_id *cm_id \ [all...] |
H A D | ucma.c | 92 struct rdma_cm_id *cm_id; member in struct:ucma_context 159 * Same as ucm_get_ctx but requires that ->cm_id->device is valid, eg that the 168 if (!ctx->cm_id->device) { 186 rdma_destroy_id(ctx->cm_id); 188 /* Reading the cm_id without holding a positive ref is not allowed */ 189 ctx->cm_id = NULL; 216 struct rdma_cm_id *cm_id) 219 ctx->cm_id = cm_id; 283 if (ctx->cm_id 215 ucma_set_ctx_cm_id(struct ucma_context *ctx, struct rdma_cm_id *cm_id) argument 295 ucma_connect_event_handler(struct rdma_cm_id *cm_id, struct rdma_cm_event *event) argument 332 ucma_event_handler(struct rdma_cm_id *cm_id, struct rdma_cm_event *event) argument 444 struct rdma_cm_id *cm_id; local 869 ucma_query_device_addr(struct rdma_cm_id *cm_id, struct rdma_ucm_query_addr_resp *resp) argument [all...] |
H A D | cma_trace.h | 29 __field(u32, cm_id) 36 __entry->cm_id = id_priv->res.id; 45 __entry->cm_id, __entry->srcaddr, __entry->dstaddr, __entry->tos 75 __field(u32, cm_id) 82 __entry->cm_id = id_priv->res.id; 91 __entry->cm_id, __entry->srcaddr, __entry->dstaddr, 104 __field(u32, cm_id) 112 __entry->cm_id = id_priv->res.id; 122 __entry->cm_id, __entry->srcaddr, __entry->dstaddr, __entry->tos, 182 __field(u32, cm_id) [all...] |
H A D | cma.c | 133 * rdma_iw_cm_id() - return the iw_cm_id pointer for this cm_id. 142 return id_priv->cm_id.iw; 748 * @id_priv: cm_id which should be bound to cma device 752 * It is applicable to active and passive side cm_id. 1273 if (!id_priv->cm_id.ib || (id_priv->id.qp_type == IB_QPT_UD)) 1276 ret = ib_cm_init_qp_attr(id_priv->cm_id.ib, qp_attr, 1282 if (!id_priv->cm_id.iw) { 1286 ret = iw_cm_init_qp_attr(id_priv->cm_id.iw, qp_attr, 1787 const struct ib_cm_id *cm_id, 1801 if (id_priv->id.device == cm_id 1785 cma_find_listener( const struct rdma_bind_list *bind_list, const struct ib_cm_id *cm_id, const struct ib_cm_event *ib_event, const struct cma_req_info *req, const struct net_device *net_dev) argument 1819 cma_ib_id_from_event(struct ib_cm_id *cm_id, const struct ib_cm_event *ib_event, struct cma_req_info *req, struct net_device **net_dev) argument 2157 cma_ib_handler(struct ib_cm_id *cm_id, const struct ib_cm_event *ib_event) argument 2379 cma_ib_req_handler(struct ib_cm_id *cm_id, const struct ib_cm_event *ib_event) argument 2467 rdma_read_gids(struct rdma_cm_id *cm_id, union ib_gid *sgid, union ib_gid *dgid) argument 2558 iw_conn_req_handler(struct iw_cm_id *cm_id, struct iw_cm_event *iw_event) argument 4189 cma_sidr_rep_handler(struct ib_cm_id *cm_id, const struct ib_cm_event *ib_event) argument 4393 struct iw_cm_id *cm_id; local [all...] |
H A D | cm.c | 991 * The cm_id could be destroyed by the user before we exit timewait. 992 * To protect against this, we search for the cm_id after exiting 1029 static noinline void cm_destroy_id_wait_timeout(struct ib_cm_id *cm_id, argument 1034 cm_id_priv = container_of(cm_id, struct cm_id_private, id); 1035 pr_err("%s: cm_id=%p timed out. state %d -> %d, refcnt=%d\n", __func__, 1036 cm_id, old_state, cm_id->state, refcount_read(&cm_id_priv->refcount)); 1039 static void cm_destroy_id(struct ib_cm_id *cm_id, int err) argument 1046 cm_id_priv = container_of(cm_id, struct cm_id_private, id); 1048 old_state = cm_id 1169 ib_destroy_cm_id(struct ib_cm_id *cm_id) argument 1198 ib_cm_listen(struct ib_cm_id *cm_id, __be64 service_id) argument 1494 ib_send_cm_req(struct ib_cm_id *cm_id, struct ib_cm_req_param *param) argument 2275 ib_send_cm_rep(struct ib_cm_id *cm_id, struct ib_cm_rep_param *param) argument 2348 ib_send_cm_rtu(struct ib_cm_id *cm_id, const void *private_data, u8 private_data_len) argument 2699 ib_send_cm_dreq(struct ib_cm_id *cm_id, const void *private_data, u8 private_data_len) argument 2766 ib_send_cm_drep(struct ib_cm_id *cm_id, const void *private_data, u8 private_data_len) argument 2977 ib_send_cm_rej(struct ib_cm_id *cm_id, enum ib_cm_rej_reason reason, void *ari, u8 ari_length, const void *private_data, u8 private_data_len) argument 3088 ib_send_cm_mra(struct ib_cm_id *cm_id, u8 service_timeout, const void *private_data, u8 private_data_len) argument 3488 ib_send_cm_sidr_req(struct ib_cm_id *cm_id, struct ib_cm_sidr_req_param *param) argument 3704 ib_send_cm_sidr_rep(struct ib_cm_id *cm_id, struct ib_cm_sidr_rep_param *param) argument 3913 cm_establish(struct ib_cm_id *cm_id) argument 3977 cm_migrate(struct ib_cm_id *cm_id) argument 3997 ib_cm_notify(struct ib_cm_id *cm_id, enum ib_event_type event) argument 4260 ib_cm_init_qp_attr(struct ib_cm_id *cm_id, struct ib_qp_attr *qp_attr, int *qp_attr_mask) argument [all...] |
/linux-master/net/9p/ |
H A D | trans_rdma.c | 49 * @cm_id: The RDMA CM ID 76 struct rdma_cm_id *cm_id; member in struct:p9_trans_rdma 280 rdma_disconnect(rdma->cm_id); 301 ib_dma_unmap_single(rdma->cm_id->device, c->busa, client->msize, 348 ib_dma_unmap_single(rdma->cm_id->device, 376 if (rdma->cm_id && !IS_ERR(rdma->cm_id)) 377 rdma_destroy_id(rdma->cm_id); 390 c->busa = ib_dma_map_single(rdma->cm_id->device, 393 if (ib_dma_mapping_error(rdma->cm_id [all...] |
/linux-master/drivers/nvme/target/ |
H A D | rdma.c | 88 struct rdma_cm_id *cm_id; member in struct:nvmet_rdma_queue 122 struct rdma_cm_id *cm_id; member in struct:nvmet_rdma_port 637 struct rdma_cm_id *cm_id = rsp->queue->cm_id; local 642 ret = rdma_rw_ctx_signature_init(&rsp->rw, cm_id->qp, 643 cm_id->port_num, req->sg, req->sg_cnt, 647 ret = rdma_rw_ctx_init(&rsp->rw, cm_id->qp, cm_id->port_num, 656 struct rdma_cm_id *cm_id = rsp->queue->cm_id; local 721 struct rdma_cm_id *cm_id = rsp->queue->cm_id; local 792 struct rdma_cm_id *cm_id = rsp->queue->cm_id; local 1191 nvmet_rdma_find_get_device(struct rdma_cm_id *cm_id) argument 1404 nvmet_rdma_cm_reject(struct rdma_cm_id *cm_id, enum nvme_rdma_cm_status status) argument 1420 nvmet_rdma_alloc_queue(struct nvmet_rdma_device *ndev, struct rdma_cm_id *cm_id, struct rdma_cm_event *event) argument 1542 nvmet_rdma_cm_accept(struct rdma_cm_id *cm_id, struct nvmet_rdma_queue *queue, struct rdma_conn_param *p) argument 1566 nvmet_rdma_queue_connect(struct rdma_cm_id *cm_id, struct rdma_cm_event *event) argument 1702 nvmet_rdma_queue_connect_fail(struct rdma_cm_id *cm_id, struct nvmet_rdma_queue *queue) argument 1731 nvmet_rdma_device_removal(struct rdma_cm_id *cm_id, struct nvmet_rdma_queue *queue) argument 1763 nvmet_rdma_cm_handler(struct rdma_cm_id *cm_id, struct rdma_cm_event *event) argument 1852 struct rdma_cm_id *cm_id = xchg(&port->cm_id, NULL); local 1868 struct rdma_cm_id *cm_id; local 2001 struct rdma_cm_id *cm_id = port->cm_id; local [all...] |
/linux-master/drivers/infiniband/hw/irdma/ |
H A D | cm.c | 133 * @cm_id: upper layers cm struct for the event 137 struct iw_cm_id *cm_id, 140 memcpy(&event->local_addr, &cm_id->m_local_addr, 142 memcpy(&event->remote_addr, &cm_id->m_remote_addr, 155 * @cm_id: upper layer's cm info struct 160 struct iw_cm_id *cm_id, 167 trace_irdma_send_cm_event(cm_node, cm_id, type, status, 171 "CM: cm_node %p cm_id=%p state=%d accel=%d event_type=%d status=%d\n", 172 cm_node, cm_id, cm_node->accelerated, cm_node->state, type, 187 irdma_get_cmevent_info(cm_node, cm_id, 136 irdma_get_cmevent_info(struct irdma_cm_node *cm_node, struct iw_cm_id *cm_id, struct iw_cm_event *event) argument 159 irdma_send_cm_event(struct irdma_cm_node *cm_node, struct iw_cm_id *cm_id, enum iw_cm_event_type type, int status) argument 664 struct iw_cm_id *cm_id; local 3486 struct iw_cm_id *cm_id; local 3635 irdma_accept(struct iw_cm_id *cm_id, struct iw_cm_conn_param *conn_param) argument 3789 irdma_reject(struct iw_cm_id *cm_id, const void *pdata, u8 pdata_len) argument 3816 irdma_connect(struct iw_cm_id *cm_id, struct iw_cm_conn_param *conn_param) argument 3964 irdma_create_listen(struct iw_cm_id *cm_id, int backlog) argument 4085 irdma_destroy_listen(struct iw_cm_id *cm_id) argument 4139 struct iw_cm_id *cm_id; local 4197 struct iw_cm_id *cm_id = cm_node->cm_id; local [all...] |
H A D | cm.h | 245 struct iw_cm_id *cm_id; member in struct:irdma_cm_listener 296 struct iw_cm_id *cm_id; member in struct:irdma_cm_node 330 struct iw_cm_id *cm_id; member in struct:irdma_cm_info 394 int irdma_accept(struct iw_cm_id *cm_id, struct iw_cm_conn_param *conn_param); 395 int irdma_reject(struct iw_cm_id *cm_id, const void *pdata, u8 pdata_len); 396 int irdma_connect(struct iw_cm_id *cm_id, struct iw_cm_conn_param *conn_param); 397 int irdma_create_listen(struct iw_cm_id *cm_id, int backlog); 398 int irdma_destroy_listen(struct iw_cm_id *cm_id);
|
H A D | trace_cm.h | 201 TP_PROTO(struct irdma_cm_node *cm_node, struct iw_cm_id *cm_id, 203 TP_ARGS(cm_node, cm_id, type, status, caller), 206 __field(struct iw_cm_id *, cm_id) 222 __entry->cm_id = cm_id; 238 TP_printk("iwdev=%p caller=%pS cm_id=%p node=%p refcnt=%d vlan_id=%d accel=%d state=%s event_type=%s status=%d loc: %s rem: %s", 241 __entry->cm_id, 257 TP_PROTO(struct iw_cm_id *cm_id, enum iw_cm_event_type type, 259 TP_ARGS(cm_id, type, status, caller), 260 TP_STRUCT__entry(__field(struct iw_cm_id *, cm_id) [all...] |
/linux-master/drivers/infiniband/hw/erdma/ |
H A D | erdma_cm.c | 238 struct iw_cm_id *cm_id; local 246 cm_id = cep->listen_cep->cm_id; 251 cm_id = cep->cm_id; 267 return cm_id->event_handler(cm_id, &event); 283 if (cep->cm_id) { 301 cep->cm_id->rem_ref(cep->cm_id); [all...] |
/linux-master/fs/smb/server/ |
H A D | transport_rdma.c | 84 struct rdma_cm_id *cm_id; member in struct:smb_direct_listener 103 struct rdma_cm_id *cm_id; member in struct:smb_direct_transport 269 ib_dma_unmap_single(t->cm_id->device, recvmsg->sge.addr, 295 ib_dma_unmap_single(t->cm_id->device, recvmsg->sge.addr, 338 rdma_disconnect(t->cm_id); 360 static struct smb_direct_transport *alloc_transport(struct rdma_cm_id *cm_id) argument 369 t->cm_id = cm_id; 370 cm_id->context = t; 452 if (t->cm_id) 1523 smb_direct_cm_handler(struct rdma_cm_id *cm_id, struct rdma_cm_event *event) argument 2075 smb_direct_listen_handler(struct rdma_cm_id *cm_id, struct rdma_cm_event *event) argument 2102 struct rdma_cm_id *cm_id; local [all...] |
/linux-master/drivers/infiniband/ulp/srpt/ |
H A D | ib_srpt.h | 255 * @ib_cm.cm_id: IB CM ID associated with the channel. 257 * @rdma_cm.cm_id: RDMA CM ID associated with the channel. 286 * against concurrent modification by the cm_id spinlock. 297 struct ib_cm_id *cm_id; member in struct:srpt_rdma_ch::__anon523::__anon524 300 struct rdma_cm_id *cm_id; member in struct:srpt_rdma_ch::__anon523::__anon525 442 * @cm_id: Connection identifier. 458 struct ib_cm_id *cm_id; member in struct:srpt_device
|
/linux-master/drivers/infiniband/ulp/ipoib/ |
H A D | ipoib_cm.c | 80 static int ipoib_cm_tx_handler(struct ib_cm_id *cm_id, 274 struct ib_cm_id *cm_id, struct ib_qp *qp, 282 ret = ib_cm_init_qp_attr(cm_id, &qp_attr, &qp_attr_mask); 293 ret = ib_cm_init_qp_attr(cm_id, &qp_attr, &qp_attr_mask); 314 ret = ib_cm_init_qp_attr(cm_id, &qp_attr, &qp_attr_mask); 347 static int ipoib_cm_nonsrq_init_rx(struct net_device *dev, struct ib_cm_id *cm_id, argument 375 ib_send_cm_rej(cm_id, IB_CM_REJ_NO_QP, NULL, 0, NULL, 0); 420 static int ipoib_cm_send_rep(struct net_device *dev, struct ib_cm_id *cm_id, argument 439 return ib_send_cm_rep(cm_id, &rep); 442 static int ipoib_cm_req_handler(struct ib_cm_id *cm_id, argument 273 ipoib_cm_modify_rx_qp(struct net_device *dev, struct ib_cm_id *cm_id, struct ib_qp *qp, unsigned int psn) argument 504 ipoib_cm_rx_handler(struct ib_cm_id *cm_id, const struct ib_cm_event *event) argument 985 ipoib_cm_rep_handler(struct ib_cm_id *cm_id, const struct ib_cm_event *event) argument 1119 ipoib_cm_modify_tx_init(struct net_device *dev, struct ib_cm_id *cm_id, struct ib_qp *qp) argument 1249 ipoib_cm_tx_handler(struct ib_cm_id *cm_id, const struct ib_cm_event *event) argument [all...] |
/linux-master/drivers/infiniband/ulp/rtrs/ |
H A D | rtrs.c | 218 rdma_notify(con->cm_id, IB_EVENT_COMM_EST); 235 struct rdma_cm_id *cm_id = con->cm_id; local 239 cq = ib_alloc_cq(cm_id->device, con, nr_cqe, cq_vector, 242 cq = ib_cq_pool_get(cm_id->device, nr_cqe, cq_vector, poll_ctx); 259 struct rdma_cm_id *cm_id = con->cm_id; local 274 ret = rdma_create_qp(cm_id, pd, &init_attr); 279 con->qp = cm_id->qp; 321 rdma_destroy_qp(con->cm_id); [all...] |
H A D | rtrs-srv.c | 784 static int rtrs_rdma_do_reject(struct rdma_cm_id *cm_id, int errno); 1541 rdma_disconnect(con->c.cm_id); 1567 rdma_destroy_id(con->c.cm_id); 1583 struct rdma_cm_id *cm_id) 1607 err = rdma_accept(cm_id, ¶m); 1614 static int rtrs_rdma_do_reject(struct rdma_cm_id *cm_id, int errno) argument 1625 err = rdma_reject(cm_id, &msg, sizeof(msg), IB_CM_REJ_CONSUMER_DEFINED); 1647 struct rdma_cm_id *cm_id, 1665 con->c.cm_id = cm_id; 1582 rtrs_rdma_do_accept(struct rtrs_srv_path *srv_path, struct rdma_cm_id *cm_id) argument 1646 create_con(struct rtrs_srv_path *srv_path, struct rdma_cm_id *cm_id, unsigned int cid) argument 1731 __alloc_path(struct rtrs_srv_sess *srv, struct rdma_cm_id *cm_id, unsigned int con_num, unsigned int recon_cnt, const uuid_t *uuid) argument 1831 rtrs_rdma_connect(struct rdma_cm_id *cm_id, const struct rtrs_msg_conn_req *msg, size_t len) argument 1954 rtrs_srv_rdma_cm_handler(struct rdma_cm_id *cm_id, struct rdma_cm_event *ev) argument 2002 struct rdma_cm_id *cm_id; local [all...] |
/linux-master/drivers/infiniband/hw/cxgb4/ |
H A D | cm.c | 152 epc->cm_id->rem_ref(epc->cm_id); 153 epc->cm_id = NULL; 160 epc->cm_id->add_ref(epc->cm_id); 1275 if (ep->com.cm_id) { 1276 pr_debug("close complete delivered ep %p cm_id %p tid %u\n", 1277 ep, ep->com.cm_id, ep->hwtid); 1278 ep->com.cm_id->event_handler(ep->com.cm_id, 3090 c4iw_reject_cr(struct iw_cm_id *cm_id, const void *pdata, u8 pdata_len) argument 3116 c4iw_accept_cr(struct iw_cm_id *cm_id, struct iw_cm_conn_param *conn_param) argument 3231 pick_local_ipaddrs(struct c4iw_dev *dev, struct iw_cm_id *cm_id) argument 3283 pick_local_ip6addrs(struct c4iw_dev *dev, struct iw_cm_id *cm_id) argument 3297 c4iw_connect(struct iw_cm_id *cm_id, struct iw_cm_conn_param *conn_param) argument 3536 c4iw_create_listen(struct iw_cm_id *cm_id, int backlog) argument 3600 c4iw_destroy_listen(struct iw_cm_id *cm_id) argument [all...] |