Searched refs:qhp (Results 1 - 14 of 14) sorted by relevance

/freebsd-11-stable/contrib/ofed/libcxgb4/
H A Dverbs.c308 struct c4iw_qp *qhp; local
314 qhp = calloc(1, sizeof *qhp);
315 if (!qhp)
318 ret = ibv_cmd_create_qp(pd, &qhp->ibv_qp, attr, &cmd,
330 qhp->wq.qid_mask = resp.qid_mask;
331 qhp->rhp = dev;
332 qhp->wq.sq.qid = resp.sqid;
333 qhp->wq.sq.size = resp.sq_size;
334 qhp
411 struct c4iw_qp *qhp; local
567 reset_qp(struct c4iw_qp *qhp) argument
584 struct c4iw_qp *qhp = to_c4iw_qp(ibqp); local
601 struct c4iw_qp *qhp = to_c4iw_qp(ibqp); local
636 struct c4iw_qp *qhp = to_c4iw_qp(ibqp); local
660 struct c4iw_qp *qhp = to_c4iw_qp(ibqp); local
674 struct c4iw_qp *qhp = to_c4iw_qp(ibqp); local
697 struct c4iw_qp *qhp = to_c4iw_qp(event->element.qp); local
[all...]
H A Dqp.c275 static int build_rdma_recv(struct c4iw_qp *qhp, union t4_recv_wr *wqe, argument
288 static void ring_kernel_db(struct c4iw_qp *qhp, u32 qid, u16 idx) argument
298 if (qid == qhp->wq.sq.qid) {
305 ret = ibv_cmd_modify_qp(&qhp->ibv_qp, &attr, mask, &cmd, sizeof cmd);
316 struct c4iw_qp *qhp; local
322 qhp = to_c4iw_qp(ibqp);
323 pthread_spin_lock(&qhp->lock);
324 if (t4_wq_in_error(&qhp->wq)) {
325 pthread_spin_unlock(&qhp->lock);
329 num_wrs = t4_sq_avail(&qhp
421 struct c4iw_qp *qhp; local
485 update_qp_state(struct c4iw_qp *qhp) argument
502 c4iw_flush_qp(struct c4iw_qp *qhp) argument
545 struct c4iw_qp *qhp = dev->qpid2ptr[i]; local
[all...]
H A Ddev.c274 static void dump_qp(struct c4iw_qp *qhp) argument
287 qhp,
288 qhp->wq.sq.qid,
289 qhp->wq.error,
290 qhp->wq.flushed,
291 qhp->wq.qid_mask,
292 qhp->wq.sq.qid,
293 qhp->wq.sq.queue,
294 qhp->wq.sq.sw_sq,
295 qhp
393 struct c4iw_qp *qhp = dev->qpid2ptr[i]; local
[all...]
H A Dcq.c95 void c4iw_flush_sq(struct c4iw_qp *qhp) argument
98 struct t4_wq *wq = &qhp->wq;
99 struct c4iw_cq *chp = to_c4iw_cq(qhp->ibv_qp.send_cq);
201 struct c4iw_qp *qhp; local
214 qhp = get_qhp(chp->rhp, CQE_QPID(hw_cqe));
219 if (qhp == NULL)
247 if (!qhp->wq.sq.oldest_read->signaled) {
248 advance_oldest_read(&qhp->wq);
256 create_read_req_cqe(&qhp->wq, hw_cqe, &read_cqe);
258 advance_oldest_read(&qhp
562 struct c4iw_qp *qhp = NULL; local
[all...]
H A Dlibcxgb4.h215 void c4iw_flush_qp(struct c4iw_qp *qhp);
231 void c4iw_flush_sq(struct c4iw_qp *qhp);
/freebsd-11-stable/sys/dev/cxgb/ulp/iw_cxgb/
H A Diw_cxgb_qp.c246 static int build_rdma_recv(struct iwch_qp *qhp, union t3_wr *wqe, argument
257 err = iwch_sgl2pbl_map(qhp->rhp, wr->sg_list, wr->num_sge, pbl_addr,
282 qhp->wq.rq[Q_PTR2IDX(qhp->wq.rq_wptr,
283 qhp->wq.rq_size_log2)].wr_id = wr->wr_id;
284 qhp->wq.rq[Q_PTR2IDX(qhp->wq.rq_wptr,
285 qhp->wq.rq_size_log2)].pbl_addr = 0;
290 static int build_zero_stag_recv(struct iwch_qp *qhp, union t3_wr *wqe, argument
304 pbl_addr = cxio_hal_pblpool_alloc(&qhp
360 struct iwch_qp *qhp; local
456 struct iwch_qp *qhp; local
519 struct iwch_qp *qhp; local
728 iwch_post_terminate(struct iwch_qp *qhp, struct respQ_msg_t *rsp_msg) argument
764 __flush_qp(struct iwch_qp *qhp, struct iwch_cq *rchp, struct iwch_cq *schp) argument
803 flush_qp(struct iwch_qp *qhp) argument
828 rqes_posted(struct iwch_qp *qhp) argument
839 rdma_init(struct iwch_dev *rhp, struct iwch_qp *qhp, enum iwch_qp_attr_mask mask, struct iwch_qp_attributes *attrs) argument
888 iwch_modify_qp(struct iwch_dev *rhp, struct iwch_qp *qhp, enum iwch_qp_attr_mask mask, struct iwch_qp_attributes *attrs, int internal) argument
[all...]
H A Diw_cxgb_ev.c80 post_qp_event(struct iwch_dev *rnicp, struct iwch_qp *qhp, struct iwch_cq *chp, argument
90 if (!qhp) {
98 if ((qhp->attr.state == IWCH_QP_STATE_ERROR) ||
99 (qhp->attr.state == IWCH_QP_STATE_TERMINATE)) {
102 qhp->attr.state, qhp->wq.qpid, CQE_STATUS(rsp_msg->cqe));
115 if (qhp->attr.state == IWCH_QP_STATE_RTS) {
117 iwch_modify_qp(qhp->rhp, qhp, IWCH_QP_ATTR_NEXT_STATE,
120 iwch_post_terminate(qhp, rsp_ms
141 struct iwch_qp *qhp; local
[all...]
H A Diw_cxgb_provider.c737 struct iwch_qp *qhp; local
741 qhp = to_iwch_qp(ib_qp);
742 rhp = qhp->rhp;
745 iwch_modify_qp(rhp, qhp, IWCH_QP_ATTR_NEXT_STATE, &attrs, 0);
746 mtx_lock(&qhp->lock);
747 if (qhp->ep)
748 msleep(qhp, &qhp->lock, 0, "iwch_destroy_qp1", 0);
749 mtx_unlock(&qhp->lock);
751 remove_handle(rhp, &rhp->qpidr, qhp
774 struct iwch_qp *qhp; local
922 struct iwch_qp *qhp; local
[all...]
H A Diw_cxgb_cq.c91 struct iwch_qp *qhp = NULL; local
104 qhp = get_qhp(rhp, CQE_QPID(*rd_cqe));
105 if (!qhp)
108 mtx_lock(&qhp->lock);
109 wq = &(qhp->wq);
126 wc->qp = &qhp->ibqp;
226 mtx_unlock(&qhp->lock);
H A Diw_cxgb_provider.h169 qp_quiesced(struct iwch_qp *qhp) argument
171 return qhp->flags & QP_QUIESCED;
258 struct iwch_qp *qhp,
337 int iwch_post_terminate(struct iwch_qp *qhp, struct respQ_msg_t *rsp_msg);
340 void stop_read_rep_timer(struct iwch_qp *qhp);
/freebsd-11-stable/sys/dev/cxgbe/iw_cxgbe/
H A Dqp.c95 static void set_state(struct c4iw_qp *qhp, enum c4iw_qp_state state) argument
98 spin_lock_irqsave(&qhp->lock, flag);
99 qhp->attr.state = state;
100 spin_unlock_irqrestore(&qhp->lock, flag);
545 static int build_rdma_recv(struct c4iw_qp *qhp, union t4_recv_wr *wqe, argument
550 ret = build_isgl((__be64 *)qhp->wq.rq.queue,
551 (__be64 *)&qhp->wq.rq.queue[qhp->wq.rq.size],
572 struct c4iw_qp *qhp; local
575 qhp
591 struct c4iw_qp *qhp; local
610 complete_sq_drain_wr(struct c4iw_qp *qhp, struct ib_send_wr *wr) argument
640 complete_rq_drain_wr(struct c4iw_qp *qhp, struct ib_recv_wr *wr) argument
775 struct c4iw_qp *qhp; local
912 struct c4iw_qp *qhp; local
1108 post_terminate(struct c4iw_qp *qhp, struct t4_cqe *err_cqe, gfp_t gfp) argument
1151 __flush_qp(struct c4iw_qp *qhp, struct c4iw_cq *rchp, struct c4iw_cq *schp) argument
1211 flush_qp(struct c4iw_qp *qhp) argument
1238 rdma_fini(struct c4iw_dev *rhp, struct c4iw_qp *qhp, struct c4iw_ep *ep) argument
1333 rdma_init(struct c4iw_dev *rhp, struct c4iw_qp *qhp) argument
1424 c4iw_modify_qp(struct c4iw_dev *rhp, struct c4iw_qp *qhp, enum c4iw_qp_attr_mask mask, struct c4iw_qp_attributes *attrs, int internal) argument
1677 struct c4iw_qp *qhp; local
1706 struct c4iw_qp *qhp; local
1910 struct c4iw_qp *qhp; local
1954 struct c4iw_qp *qhp = to_c4iw_qp(ibqp); local
[all...]
H A Dev.c43 struct c4iw_qp *qhp,
51 if ((qhp->attr.state == C4IW_QP_STATE_ERROR) ||
52 (qhp->attr.state == C4IW_QP_STATE_TERMINATE)) {
55 qhp->attr.state, qhp->wq.sq.qid, CQE_STATUS(err_cqe));
65 if (qhp->attr.state == C4IW_QP_STATE_RTS) {
67 c4iw_modify_qp(qhp->rhp, qhp, C4IW_QP_ATTR_NEXT_STATE,
76 event.element.qp = &qhp->ibqp;
77 if (qhp
42 post_qp_event(struct c4iw_dev *dev, struct c4iw_cq *chp, struct c4iw_qp *qhp, struct t4_cqe *err_cqe, enum ib_event_type ib_event) argument
88 struct c4iw_qp *qhp; local
[all...]
H A Dcq.c255 int c4iw_flush_sq(struct c4iw_qp *qhp) argument
258 struct t4_wq *wq = &qhp->wq;
259 struct c4iw_cq *chp = to_c4iw_cq(qhp->ibqp.send_cq);
362 struct c4iw_qp *qhp; local
376 qhp = get_qhp(chp->rhp, CQE_QPID(hw_cqe));
381 if (qhp == NULL)
404 if (!qhp->wq.sq.oldest_read->signaled) {
405 advance_oldest_read(&qhp->wq);
413 create_read_req_cqe(&qhp->wq, hw_cqe, &read_cqe);
415 advance_oldest_read(&qhp
715 struct c4iw_qp *qhp = NULL; local
[all...]
H A Diw_cxgbe.h578 struct c4iw_qp *qhp,
969 int c4iw_flush_sq(struct c4iw_qp *qhp);
971 u16 c4iw_rqes_posted(struct c4iw_qp *qhp);
972 int c4iw_post_terminate(struct c4iw_qp *qhp, struct t4_cqe *err_cqe);

Completed in 201 milliseconds