/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/drivers/infiniband/hw/cxgb3/ |
H A D | iwch_qp.c | 247 static int build_rdma_recv(struct iwch_qp *qhp, union t3_wr *wqe, argument 254 err = iwch_sgl2pbl_map(qhp->rhp, wr->sg_list, wr->num_sge, pbl_addr, 280 qhp->wq.rq[Q_PTR2IDX(qhp->wq.rq_wptr, 281 qhp->wq.rq_size_log2)].wr_id = wr->wr_id; 282 qhp->wq.rq[Q_PTR2IDX(qhp->wq.rq_wptr, 283 qhp->wq.rq_size_log2)].pbl_addr = 0; 287 static int build_zero_stag_recv(struct iwch_qp *qhp, union t3_wr *wqe, argument 301 pbl_addr = cxio_hal_pblpool_alloc(&qhp 357 struct iwch_qp *qhp; local 469 struct iwch_qp *qhp; local 534 struct iwch_qp *qhp; local 741 iwch_post_zb_read(struct iwch_qp *qhp) argument 773 iwch_post_terminate(struct iwch_qp *qhp, struct respQ_msg_t *rsp_msg) argument 805 __flush_qp(struct iwch_qp *qhp, unsigned long *flag) argument 848 flush_qp(struct iwch_qp *qhp, unsigned long *flag) argument 860 iwch_rqes_posted(struct iwch_qp *qhp) argument 872 rdma_init(struct iwch_dev *rhp, struct iwch_qp *qhp, enum iwch_qp_attr_mask mask, struct iwch_qp_attributes *attrs) argument 924 iwch_modify_qp(struct iwch_dev *rhp, struct iwch_qp *qhp, enum iwch_qp_attr_mask mask, struct iwch_qp_attributes *attrs, int internal) argument 1142 quiesce_qp(struct iwch_qp *qhp) argument 1151 resume_qp(struct iwch_qp *qhp) argument 1163 struct iwch_qp *qhp; local 1182 struct iwch_qp *qhp; local [all...] |
H A D | iwch_ev.c | 48 struct iwch_qp *qhp; local 51 qhp = get_qhp(rnicp, CQE_QPID(rsp_msg->cqe)); 53 if (!qhp) { 61 if ((qhp->attr.state == IWCH_QP_STATE_ERROR) || 62 (qhp->attr.state == IWCH_QP_STATE_TERMINATE)) { 65 qhp->attr.state, qhp->wq.qpid, CQE_STATUS(rsp_msg->cqe)); 76 atomic_inc(&qhp->refcnt); 84 event.element.qp = &qhp->ibqp; 86 if (qhp 106 struct iwch_qp *qhp; local [all...] |
H A D | iwch_cq.c | 47 struct iwch_qp *qhp = NULL; local 60 qhp = get_qhp(rhp, CQE_QPID(*rd_cqe)); 61 if (!qhp) 64 spin_lock(&qhp->lock); 65 wq = &(qhp->wq); 82 wc->qp = &qhp->ibqp; 191 spin_unlock(&qhp->lock);
|
H A D | iwch_provider.c | 855 struct iwch_qp *qhp; local 859 qhp = to_iwch_qp(ib_qp); 860 rhp = qhp->rhp; 863 iwch_modify_qp(rhp, qhp, IWCH_QP_ATTR_NEXT_STATE, &attrs, 0); 864 wait_event(qhp->wait, !qhp->ep); 866 remove_handle(rhp, &rhp->qpidr, qhp->wq.qpid); 868 atomic_dec(&qhp->refcnt); 869 wait_event(qhp->wait, !atomic_read(&qhp 887 struct iwch_qp *qhp; local 1039 struct iwch_qp *qhp; local [all...] |
H A D | iwch_provider.h | 171 static inline int qp_quiesced(struct iwch_qp *qhp) argument 173 return qhp->flags & QP_QUIESCED; 258 struct iwch_qp *qhp, 325 u16 iwch_rqes_posted(struct iwch_qp *qhp); 334 int iwch_post_terminate(struct iwch_qp *qhp, struct respQ_msg_t *rsp_msg); 335 int iwch_post_zb_read(struct iwch_qp *qhp); 340 void stop_read_rep_timer(struct iwch_qp *qhp);
|
H A D | iwch.c | 68 struct iwch_qp *qhp = p; local 70 cxio_disable_wq_db(&qhp->wq); 76 struct iwch_qp *qhp = p; local 79 ring_doorbell(qhp->rhp->rdev.ctrl_qp.doorbell, qhp->wq.qpid); 80 cxio_enable_wq_db(&qhp->wq);
|
/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/drivers/infiniband/hw/cxgb4/ |
H A D | ev.c | 39 struct c4iw_qp *qhp, 46 if ((qhp->attr.state == C4IW_QP_STATE_ERROR) || 47 (qhp->attr.state == C4IW_QP_STATE_TERMINATE)) { 50 qhp->attr.state, qhp->wq.sq.qid, CQE_STATUS(err_cqe)); 60 if (qhp->attr.state == C4IW_QP_STATE_RTS) { 62 c4iw_modify_qp(qhp->rhp, qhp, C4IW_QP_ATTR_NEXT_STATE, 71 event.element.qp = &qhp->ibqp; 72 if (qhp 38 post_qp_event(struct c4iw_dev *dev, struct c4iw_cq *chp, struct c4iw_qp *qhp, struct t4_cqe *err_cqe, enum ib_event_type ib_event) argument 81 struct c4iw_qp *qhp; local [all...] |
H A D | qp.c | 441 static int build_rdma_recv(struct c4iw_qp *qhp, union t4_recv_wr *wqe, argument 446 ret = build_isgl((__be64 *)qhp->wq.rq.queue, 447 (__be64 *)&qhp->wq.rq.queue[qhp->wq.rq.size], 535 struct c4iw_qp *qhp; local 542 qhp = to_c4iw_qp(ibqp); 543 spin_lock_irqsave(&qhp->lock, flag); 544 if (t4_wq_in_error(&qhp->wq)) { 545 spin_unlock_irqrestore(&qhp->lock, flag); 548 num_wrs = t4_sq_avail(&qhp 646 struct c4iw_qp *qhp; local 845 c4iw_post_zb_read(struct c4iw_qp *qhp) argument 875 post_terminate(struct c4iw_qp *qhp, struct t4_cqe *err_cqe, gfp_t gfp) argument 907 __flush_qp(struct c4iw_qp *qhp, struct c4iw_cq *rchp, struct c4iw_cq *schp, unsigned long *flag) argument 947 flush_qp(struct c4iw_qp *qhp, unsigned long *flag) argument 964 rdma_fini(struct c4iw_dev *rhp, struct c4iw_qp *qhp, struct c4iw_ep *ep) argument 1039 rdma_init(struct c4iw_dev *rhp, struct c4iw_qp *qhp) argument 1118 c4iw_modify_qp(struct c4iw_dev *rhp, struct c4iw_qp *qhp, enum c4iw_qp_attr_mask mask, struct c4iw_qp_attributes *attrs, int internal) argument 1346 struct c4iw_qp *qhp; local 1375 struct c4iw_qp *qhp; local 1547 struct c4iw_qp *qhp; local [all...] |
H A D | cq.c | 576 struct c4iw_qp *qhp = NULL; local 589 qhp = get_qhp(chp->rhp, CQE_QPID(rd_cqe)); 590 if (!qhp) 593 spin_lock(&qhp->lock); 594 wq = &(qhp->wq); 601 wc->qp = &qhp->ibqp; 713 spin_unlock(&qhp->lock);
|
H A D | iw_cxgb4.h | 414 struct c4iw_qp *qhp, 732 u16 c4iw_rqes_posted(struct c4iw_qp *qhp); 733 int c4iw_post_zb_read(struct c4iw_qp *qhp); 734 int c4iw_post_terminate(struct c4iw_qp *qhp, struct t4_cqe *err_cqe);
|