Searched refs:cm_id (Results 1 - 25 of 54) sorted by relevance

123

/linux-master/drivers/infiniband/hw/qedr/
H A Dqedr_iw_cm.h34 int qedr_iw_connect(struct iw_cm_id *cm_id,
37 int qedr_iw_create_listen(struct iw_cm_id *cm_id, int backlog);
39 int qedr_iw_destroy_listen(struct iw_cm_id *cm_id);
41 int qedr_iw_accept(struct iw_cm_id *cm_id, struct iw_cm_conn_param *conn_param);
43 int qedr_iw_reject(struct iw_cm_id *cm_id, const void *pdata, u8 pdata_len);
H A Dqedr_iw_cm.c97 if (ep->cm_id)
98 ep->cm_id->rem_ref(ep->cm_id);
135 listener->cm_id->event_handler(listener->cm_id, &event);
166 if (ep->cm_id)
167 ep->cm_id->event_handler(ep->cm_id, &event);
175 if (ep->cm_id)
239 if (ep->cm_id)
531 qedr_iw_connect(struct iw_cm_id *cm_id, struct iw_cm_conn_param *conn_param) argument
657 qedr_iw_create_listen(struct iw_cm_id *cm_id, int backlog) argument
721 qedr_iw_destroy_listen(struct iw_cm_id *cm_id) argument
736 qedr_iw_accept(struct iw_cm_id *cm_id, struct iw_cm_conn_param *conn_param) argument
784 qedr_iw_reject(struct iw_cm_id *cm_id, const void *pdata, u8 pdata_len) argument
[all...]
/linux-master/include/rdma/
H A Diw_cm.h39 * @cm_id: The IW CM identifier associated with the event.
42 typedef int (*iw_cm_handler)(struct iw_cm_id *cm_id,
50 * @cm_id: The IW CM identifier associated with the event.
53 typedef int (*iw_event_handler)(struct iw_cm_id *cm_id,
67 /* Used by provider to add and remove refs on IW cm_id */
110 * @cm_id: The previously created IW CM identifier to destroy.
115 void iw_destroy_cm_id(struct iw_cm_id *cm_id);
121 * @cm_id: The IW CM identifier.
128 int iw_cm_listen(struct iw_cm_id *cm_id, int backlog);
133 * @cm_id
[all...]
H A Dib_cm.h275 * @cm_id: Communication identifier associated with the reported event.
280 * new @cm_id. The new @cm_id is returned to the user through this callback.
281 * Clients are responsible for destroying the new @cm_id. For peer-to-peer
282 * IB_CM_REQ_RECEIVED and all other events, the returned @cm_id corresponds
287 * destroy the @cm_id after the callback completes.
289 typedef int (*ib_cm_handler)(struct ib_cm_id *cm_id,
306 * @device: Device associated with the cm_id. All related communication will
321 * @cm_id: Connection identifier to destroy.
325 void ib_destroy_cm_id(struct ib_cm_id *cm_id);
[all...]
/linux-master/net/rds/
H A Drdma_transport.h20 int rds_rdma_cm_event_handler(struct rdma_cm_id *cm_id,
22 int rds6_rdma_cm_event_handler(struct rdma_cm_id *cm_id,
H A Drdma_transport.c40 /* Global IPv4 and IPv6 RDS RDMA listener cm_id */
49 static int rds_rdma_cm_event_handler_cmn(struct rdma_cm_id *cm_id, argument
54 struct rds_connection *conn = cm_id->context;
60 rdsdebug("conn %p id %p handling event %u (%s)\n", conn, cm_id,
63 if (cm_id->device->node_type == RDMA_NODE_IB_CA)
72 * right away. We return 0 so cm_id doesn't get
85 ret = trans->cm_handle_connect(cm_id, event, isv6);
90 rdma_set_service_type(cm_id, conn->c_tos);
91 rdma_set_min_rnr_timer(cm_id, IB_RNR_TIMER_000_32);
93 ret = rdma_resolve_route(cm_id,
183 rds_rdma_cm_event_handler(struct rdma_cm_id *cm_id, struct rdma_cm_event *event) argument
190 rds6_rdma_cm_event_handler(struct rdma_cm_id *cm_id, struct rdma_cm_event *event) argument
201 struct rdma_cm_id *cm_id; local
[all...]
H A Dib.c410 struct rdma_cm_id *cm_id; local
422 cm_id = rdma_create_id(&init_net, rds_rdma_cm_event_handler,
424 if (IS_ERR(cm_id))
425 return PTR_ERR(cm_id);
475 ret = rdma_bind_addr(cm_id, sa);
478 if (ret || !cm_id->device ||
479 cm_id->device->node_type != RDMA_NODE_IB_CA)
484 cm_id->device ? cm_id->device->node_type : -1);
487 rdma_destroy_id(cm_id);
[all...]
/linux-master/drivers/infiniband/core/
H A Diwcm.c97 struct iwcm_id_private *cm_id; member in struct:iwcm_work
118 * elements. The design pre-allocates them based on the cm_id type:
128 * One exception is when creating the cm_id for incoming connection requests.
130 * 1) in the event upcall, cm_event_handler(), for a listening cm_id. If
135 * If work elements cannot be allocated for the new connect request cm_id,
154 list_add(&work->free_list, &work->cm_id->work_free_list);
178 work->cm_id = cm_id_priv;
208 * Release a reference on cm_id. If the last reference is being
209 * released, free the cm_id and return 1.
222 static void add_ref(struct iw_cm_id *cm_id) argument
229 rem_ref(struct iw_cm_id *cm_id) argument
305 iw_cm_disconnect(struct iw_cm_id *cm_id, int abrupt) argument
372 destroy_cm_id(struct iw_cm_id *cm_id) argument
451 iw_destroy_cm_id(struct iw_cm_id *cm_id) argument
506 iw_cm_map(struct iw_cm_id *cm_id, bool active) argument
560 iw_cm_listen(struct iw_cm_id *cm_id, int backlog) argument
602 iw_cm_reject(struct iw_cm_id *cm_id, const void *private_data, u8 private_data_len) argument
640 iw_cm_accept(struct iw_cm_id *cm_id, struct iw_cm_conn_param *iw_param) argument
696 iw_cm_connect(struct iw_cm_id *cm_id, struct iw_cm_conn_param *iw_param) argument
767 struct iw_cm_id *cm_id; local
1064 cm_event_handler(struct iw_cm_id *cm_id, struct iw_cm_event *iw_event) argument
1156 iw_cm_init_qp_attr(struct iw_cm_id *cm_id, struct ib_qp_attr *qp_attr, int *qp_attr_mask) argument
[all...]
H A Dcm_trace.h139 const struct ib_cm_id *cm_id
142 TP_ARGS(cm_id),
145 __field(const void *, cm_id) /* for eBPF scripts */
153 __entry->cm_id = cm_id;
154 __entry->local_id = be32_to_cpu(cm_id->local_id);
155 __entry->remote_id = be32_to_cpu(cm_id->remote_id);
156 __entry->state = cm_id->state;
157 __entry->lap_state = cm_id->lap_state;
171 const struct ib_cm_id *cm_id \
[all...]
H A Ducma.c92 struct rdma_cm_id *cm_id; member in struct:ucma_context
159 * Same as ucm_get_ctx but requires that ->cm_id->device is valid, eg that the
168 if (!ctx->cm_id->device) {
186 rdma_destroy_id(ctx->cm_id);
188 /* Reading the cm_id without holding a positive ref is not allowed */
189 ctx->cm_id = NULL;
216 struct rdma_cm_id *cm_id)
219 ctx->cm_id = cm_id;
283 if (ctx->cm_id
215 ucma_set_ctx_cm_id(struct ucma_context *ctx, struct rdma_cm_id *cm_id) argument
295 ucma_connect_event_handler(struct rdma_cm_id *cm_id, struct rdma_cm_event *event) argument
332 ucma_event_handler(struct rdma_cm_id *cm_id, struct rdma_cm_event *event) argument
444 struct rdma_cm_id *cm_id; local
869 ucma_query_device_addr(struct rdma_cm_id *cm_id, struct rdma_ucm_query_addr_resp *resp) argument
[all...]
H A Dcma_trace.h29 __field(u32, cm_id)
36 __entry->cm_id = id_priv->res.id;
45 __entry->cm_id, __entry->srcaddr, __entry->dstaddr, __entry->tos
75 __field(u32, cm_id)
82 __entry->cm_id = id_priv->res.id;
91 __entry->cm_id, __entry->srcaddr, __entry->dstaddr,
104 __field(u32, cm_id)
112 __entry->cm_id = id_priv->res.id;
122 __entry->cm_id, __entry->srcaddr, __entry->dstaddr, __entry->tos,
182 __field(u32, cm_id)
[all...]
H A Dcma.c133 * rdma_iw_cm_id() - return the iw_cm_id pointer for this cm_id.
142 return id_priv->cm_id.iw;
748 * @id_priv: cm_id which should be bound to cma device
752 * It is applicable to active and passive side cm_id.
1273 if (!id_priv->cm_id.ib || (id_priv->id.qp_type == IB_QPT_UD))
1276 ret = ib_cm_init_qp_attr(id_priv->cm_id.ib, qp_attr,
1282 if (!id_priv->cm_id.iw) {
1286 ret = iw_cm_init_qp_attr(id_priv->cm_id.iw, qp_attr,
1787 const struct ib_cm_id *cm_id,
1801 if (id_priv->id.device == cm_id
1785 cma_find_listener( const struct rdma_bind_list *bind_list, const struct ib_cm_id *cm_id, const struct ib_cm_event *ib_event, const struct cma_req_info *req, const struct net_device *net_dev) argument
1819 cma_ib_id_from_event(struct ib_cm_id *cm_id, const struct ib_cm_event *ib_event, struct cma_req_info *req, struct net_device **net_dev) argument
2157 cma_ib_handler(struct ib_cm_id *cm_id, const struct ib_cm_event *ib_event) argument
2379 cma_ib_req_handler(struct ib_cm_id *cm_id, const struct ib_cm_event *ib_event) argument
2467 rdma_read_gids(struct rdma_cm_id *cm_id, union ib_gid *sgid, union ib_gid *dgid) argument
2558 iw_conn_req_handler(struct iw_cm_id *cm_id, struct iw_cm_event *iw_event) argument
4189 cma_sidr_rep_handler(struct ib_cm_id *cm_id, const struct ib_cm_event *ib_event) argument
4393 struct iw_cm_id *cm_id; local
[all...]
H A Dcm.c991 * The cm_id could be destroyed by the user before we exit timewait.
992 * To protect against this, we search for the cm_id after exiting
1029 static noinline void cm_destroy_id_wait_timeout(struct ib_cm_id *cm_id, argument
1034 cm_id_priv = container_of(cm_id, struct cm_id_private, id);
1035 pr_err("%s: cm_id=%p timed out. state %d -> %d, refcnt=%d\n", __func__,
1036 cm_id, old_state, cm_id->state, refcount_read(&cm_id_priv->refcount));
1039 static void cm_destroy_id(struct ib_cm_id *cm_id, int err) argument
1046 cm_id_priv = container_of(cm_id, struct cm_id_private, id);
1048 old_state = cm_id
1169 ib_destroy_cm_id(struct ib_cm_id *cm_id) argument
1198 ib_cm_listen(struct ib_cm_id *cm_id, __be64 service_id) argument
1494 ib_send_cm_req(struct ib_cm_id *cm_id, struct ib_cm_req_param *param) argument
2275 ib_send_cm_rep(struct ib_cm_id *cm_id, struct ib_cm_rep_param *param) argument
2348 ib_send_cm_rtu(struct ib_cm_id *cm_id, const void *private_data, u8 private_data_len) argument
2699 ib_send_cm_dreq(struct ib_cm_id *cm_id, const void *private_data, u8 private_data_len) argument
2766 ib_send_cm_drep(struct ib_cm_id *cm_id, const void *private_data, u8 private_data_len) argument
2977 ib_send_cm_rej(struct ib_cm_id *cm_id, enum ib_cm_rej_reason reason, void *ari, u8 ari_length, const void *private_data, u8 private_data_len) argument
3088 ib_send_cm_mra(struct ib_cm_id *cm_id, u8 service_timeout, const void *private_data, u8 private_data_len) argument
3488 ib_send_cm_sidr_req(struct ib_cm_id *cm_id, struct ib_cm_sidr_req_param *param) argument
3704 ib_send_cm_sidr_rep(struct ib_cm_id *cm_id, struct ib_cm_sidr_rep_param *param) argument
3913 cm_establish(struct ib_cm_id *cm_id) argument
3977 cm_migrate(struct ib_cm_id *cm_id) argument
3997 ib_cm_notify(struct ib_cm_id *cm_id, enum ib_event_type event) argument
4260 ib_cm_init_qp_attr(struct ib_cm_id *cm_id, struct ib_qp_attr *qp_attr, int *qp_attr_mask) argument
[all...]
/linux-master/net/9p/
H A Dtrans_rdma.c49 * @cm_id: The RDMA CM ID
76 struct rdma_cm_id *cm_id; member in struct:p9_trans_rdma
280 rdma_disconnect(rdma->cm_id);
301 ib_dma_unmap_single(rdma->cm_id->device, c->busa, client->msize,
348 ib_dma_unmap_single(rdma->cm_id->device,
376 if (rdma->cm_id && !IS_ERR(rdma->cm_id))
377 rdma_destroy_id(rdma->cm_id);
390 c->busa = ib_dma_map_single(rdma->cm_id->device,
393 if (ib_dma_mapping_error(rdma->cm_id
[all...]
/linux-master/drivers/nvme/target/
H A Drdma.c88 struct rdma_cm_id *cm_id; member in struct:nvmet_rdma_queue
122 struct rdma_cm_id *cm_id; member in struct:nvmet_rdma_port
637 struct rdma_cm_id *cm_id = rsp->queue->cm_id; local
642 ret = rdma_rw_ctx_signature_init(&rsp->rw, cm_id->qp,
643 cm_id->port_num, req->sg, req->sg_cnt,
647 ret = rdma_rw_ctx_init(&rsp->rw, cm_id->qp, cm_id->port_num,
656 struct rdma_cm_id *cm_id = rsp->queue->cm_id; local
721 struct rdma_cm_id *cm_id = rsp->queue->cm_id; local
792 struct rdma_cm_id *cm_id = rsp->queue->cm_id; local
1191 nvmet_rdma_find_get_device(struct rdma_cm_id *cm_id) argument
1404 nvmet_rdma_cm_reject(struct rdma_cm_id *cm_id, enum nvme_rdma_cm_status status) argument
1420 nvmet_rdma_alloc_queue(struct nvmet_rdma_device *ndev, struct rdma_cm_id *cm_id, struct rdma_cm_event *event) argument
1542 nvmet_rdma_cm_accept(struct rdma_cm_id *cm_id, struct nvmet_rdma_queue *queue, struct rdma_conn_param *p) argument
1566 nvmet_rdma_queue_connect(struct rdma_cm_id *cm_id, struct rdma_cm_event *event) argument
1702 nvmet_rdma_queue_connect_fail(struct rdma_cm_id *cm_id, struct nvmet_rdma_queue *queue) argument
1731 nvmet_rdma_device_removal(struct rdma_cm_id *cm_id, struct nvmet_rdma_queue *queue) argument
1763 nvmet_rdma_cm_handler(struct rdma_cm_id *cm_id, struct rdma_cm_event *event) argument
1852 struct rdma_cm_id *cm_id = xchg(&port->cm_id, NULL); local
1868 struct rdma_cm_id *cm_id; local
2001 struct rdma_cm_id *cm_id = port->cm_id; local
[all...]
/linux-master/drivers/infiniband/hw/irdma/
H A Dcm.c133 * @cm_id: upper layers cm struct for the event
137 struct iw_cm_id *cm_id,
140 memcpy(&event->local_addr, &cm_id->m_local_addr,
142 memcpy(&event->remote_addr, &cm_id->m_remote_addr,
155 * @cm_id: upper layer's cm info struct
160 struct iw_cm_id *cm_id,
167 trace_irdma_send_cm_event(cm_node, cm_id, type, status,
171 "CM: cm_node %p cm_id=%p state=%d accel=%d event_type=%d status=%d\n",
172 cm_node, cm_id, cm_node->accelerated, cm_node->state, type,
187 irdma_get_cmevent_info(cm_node, cm_id,
136 irdma_get_cmevent_info(struct irdma_cm_node *cm_node, struct iw_cm_id *cm_id, struct iw_cm_event *event) argument
159 irdma_send_cm_event(struct irdma_cm_node *cm_node, struct iw_cm_id *cm_id, enum iw_cm_event_type type, int status) argument
664 struct iw_cm_id *cm_id; local
3486 struct iw_cm_id *cm_id; local
3635 irdma_accept(struct iw_cm_id *cm_id, struct iw_cm_conn_param *conn_param) argument
3789 irdma_reject(struct iw_cm_id *cm_id, const void *pdata, u8 pdata_len) argument
3816 irdma_connect(struct iw_cm_id *cm_id, struct iw_cm_conn_param *conn_param) argument
3964 irdma_create_listen(struct iw_cm_id *cm_id, int backlog) argument
4085 irdma_destroy_listen(struct iw_cm_id *cm_id) argument
4139 struct iw_cm_id *cm_id; local
4197 struct iw_cm_id *cm_id = cm_node->cm_id; local
[all...]
H A Dcm.h245 struct iw_cm_id *cm_id; member in struct:irdma_cm_listener
296 struct iw_cm_id *cm_id; member in struct:irdma_cm_node
330 struct iw_cm_id *cm_id; member in struct:irdma_cm_info
394 int irdma_accept(struct iw_cm_id *cm_id, struct iw_cm_conn_param *conn_param);
395 int irdma_reject(struct iw_cm_id *cm_id, const void *pdata, u8 pdata_len);
396 int irdma_connect(struct iw_cm_id *cm_id, struct iw_cm_conn_param *conn_param);
397 int irdma_create_listen(struct iw_cm_id *cm_id, int backlog);
398 int irdma_destroy_listen(struct iw_cm_id *cm_id);
H A Dtrace_cm.h201 TP_PROTO(struct irdma_cm_node *cm_node, struct iw_cm_id *cm_id,
203 TP_ARGS(cm_node, cm_id, type, status, caller),
206 __field(struct iw_cm_id *, cm_id)
222 __entry->cm_id = cm_id;
238 TP_printk("iwdev=%p caller=%pS cm_id=%p node=%p refcnt=%d vlan_id=%d accel=%d state=%s event_type=%s status=%d loc: %s rem: %s",
241 __entry->cm_id,
257 TP_PROTO(struct iw_cm_id *cm_id, enum iw_cm_event_type type,
259 TP_ARGS(cm_id, type, status, caller),
260 TP_STRUCT__entry(__field(struct iw_cm_id *, cm_id)
[all...]
/linux-master/drivers/infiniband/hw/erdma/
H A Derdma_cm.c238 struct iw_cm_id *cm_id; local
246 cm_id = cep->listen_cep->cm_id;
251 cm_id = cep->cm_id;
267 return cm_id->event_handler(cm_id, &event);
283 if (cep->cm_id) {
301 cep->cm_id->rem_ref(cep->cm_id);
[all...]
/linux-master/fs/smb/server/
H A Dtransport_rdma.c84 struct rdma_cm_id *cm_id; member in struct:smb_direct_listener
103 struct rdma_cm_id *cm_id; member in struct:smb_direct_transport
269 ib_dma_unmap_single(t->cm_id->device, recvmsg->sge.addr,
295 ib_dma_unmap_single(t->cm_id->device, recvmsg->sge.addr,
338 rdma_disconnect(t->cm_id);
360 static struct smb_direct_transport *alloc_transport(struct rdma_cm_id *cm_id) argument
369 t->cm_id = cm_id;
370 cm_id->context = t;
452 if (t->cm_id)
1523 smb_direct_cm_handler(struct rdma_cm_id *cm_id, struct rdma_cm_event *event) argument
2075 smb_direct_listen_handler(struct rdma_cm_id *cm_id, struct rdma_cm_event *event) argument
2102 struct rdma_cm_id *cm_id; local
[all...]
/linux-master/drivers/infiniband/ulp/srpt/
H A Dib_srpt.h255 * @ib_cm.cm_id: IB CM ID associated with the channel.
257 * @rdma_cm.cm_id: RDMA CM ID associated with the channel.
286 * against concurrent modification by the cm_id spinlock.
297 struct ib_cm_id *cm_id; member in struct:srpt_rdma_ch::__anon523::__anon524
300 struct rdma_cm_id *cm_id; member in struct:srpt_rdma_ch::__anon523::__anon525
442 * @cm_id: Connection identifier.
458 struct ib_cm_id *cm_id; member in struct:srpt_device
/linux-master/drivers/infiniband/ulp/ipoib/
H A Dipoib_cm.c80 static int ipoib_cm_tx_handler(struct ib_cm_id *cm_id,
274 struct ib_cm_id *cm_id, struct ib_qp *qp,
282 ret = ib_cm_init_qp_attr(cm_id, &qp_attr, &qp_attr_mask);
293 ret = ib_cm_init_qp_attr(cm_id, &qp_attr, &qp_attr_mask);
314 ret = ib_cm_init_qp_attr(cm_id, &qp_attr, &qp_attr_mask);
347 static int ipoib_cm_nonsrq_init_rx(struct net_device *dev, struct ib_cm_id *cm_id, argument
375 ib_send_cm_rej(cm_id, IB_CM_REJ_NO_QP, NULL, 0, NULL, 0);
420 static int ipoib_cm_send_rep(struct net_device *dev, struct ib_cm_id *cm_id, argument
439 return ib_send_cm_rep(cm_id, &rep);
442 static int ipoib_cm_req_handler(struct ib_cm_id *cm_id, argument
273 ipoib_cm_modify_rx_qp(struct net_device *dev, struct ib_cm_id *cm_id, struct ib_qp *qp, unsigned int psn) argument
504 ipoib_cm_rx_handler(struct ib_cm_id *cm_id, const struct ib_cm_event *event) argument
985 ipoib_cm_rep_handler(struct ib_cm_id *cm_id, const struct ib_cm_event *event) argument
1119 ipoib_cm_modify_tx_init(struct net_device *dev, struct ib_cm_id *cm_id, struct ib_qp *qp) argument
1249 ipoib_cm_tx_handler(struct ib_cm_id *cm_id, const struct ib_cm_event *event) argument
[all...]
/linux-master/drivers/infiniband/ulp/rtrs/
H A Drtrs.c218 rdma_notify(con->cm_id, IB_EVENT_COMM_EST);
235 struct rdma_cm_id *cm_id = con->cm_id; local
239 cq = ib_alloc_cq(cm_id->device, con, nr_cqe, cq_vector,
242 cq = ib_cq_pool_get(cm_id->device, nr_cqe, cq_vector, poll_ctx);
259 struct rdma_cm_id *cm_id = con->cm_id; local
274 ret = rdma_create_qp(cm_id, pd, &init_attr);
279 con->qp = cm_id->qp;
321 rdma_destroy_qp(con->cm_id);
[all...]
H A Drtrs-srv.c784 static int rtrs_rdma_do_reject(struct rdma_cm_id *cm_id, int errno);
1541 rdma_disconnect(con->c.cm_id);
1567 rdma_destroy_id(con->c.cm_id);
1583 struct rdma_cm_id *cm_id)
1607 err = rdma_accept(cm_id, &param);
1614 static int rtrs_rdma_do_reject(struct rdma_cm_id *cm_id, int errno) argument
1625 err = rdma_reject(cm_id, &msg, sizeof(msg), IB_CM_REJ_CONSUMER_DEFINED);
1647 struct rdma_cm_id *cm_id,
1665 con->c.cm_id = cm_id;
1582 rtrs_rdma_do_accept(struct rtrs_srv_path *srv_path, struct rdma_cm_id *cm_id) argument
1646 create_con(struct rtrs_srv_path *srv_path, struct rdma_cm_id *cm_id, unsigned int cid) argument
1731 __alloc_path(struct rtrs_srv_sess *srv, struct rdma_cm_id *cm_id, unsigned int con_num, unsigned int recon_cnt, const uuid_t *uuid) argument
1831 rtrs_rdma_connect(struct rdma_cm_id *cm_id, const struct rtrs_msg_conn_req *msg, size_t len) argument
1954 rtrs_srv_rdma_cm_handler(struct rdma_cm_id *cm_id, struct rdma_cm_event *ev) argument
2002 struct rdma_cm_id *cm_id; local
[all...]
/linux-master/drivers/infiniband/hw/cxgb4/
H A Dcm.c152 epc->cm_id->rem_ref(epc->cm_id);
153 epc->cm_id = NULL;
160 epc->cm_id->add_ref(epc->cm_id);
1275 if (ep->com.cm_id) {
1276 pr_debug("close complete delivered ep %p cm_id %p tid %u\n",
1277 ep, ep->com.cm_id, ep->hwtid);
1278 ep->com.cm_id->event_handler(ep->com.cm_id,
3090 c4iw_reject_cr(struct iw_cm_id *cm_id, const void *pdata, u8 pdata_len) argument
3116 c4iw_accept_cr(struct iw_cm_id *cm_id, struct iw_cm_conn_param *conn_param) argument
3231 pick_local_ipaddrs(struct c4iw_dev *dev, struct iw_cm_id *cm_id) argument
3283 pick_local_ip6addrs(struct c4iw_dev *dev, struct iw_cm_id *cm_id) argument
3297 c4iw_connect(struct iw_cm_id *cm_id, struct iw_cm_conn_param *conn_param) argument
3536 c4iw_create_listen(struct iw_cm_id *cm_id, int backlog) argument
3600 c4iw_destroy_listen(struct iw_cm_id *cm_id) argument
[all...]

Completed in 330 milliseconds

123