Lines Matching defs:rhp

106 	struct c4iw_dev *rhp = rdev_to_c4iw_dev(rdev);
111 dma_free_coherent(rhp->ibdev.dma_device,
114 dma_free_coherent(rhp->ibdev.dma_device,
130 struct c4iw_dev *rhp = rdev_to_c4iw_dev(rdev);
178 wq->sq.queue = dma_alloc_coherent(rhp->ibdev.dma_device, wq->sq.memsize,
188 wq->rq.queue = dma_alloc_coherent(rhp->ibdev.dma_device,
315 dma_free_coherent(rhp->ibdev.dma_device,
319 dma_free_coherent(rhp->ibdev.dma_device,
573 struct c4iw_dev *rhp;
577 rhp = qhp->rhp;
581 destroy_qp(&rhp->rdev, &qhp->wq,
582 ucontext ? &ucontext->uctx : &rhp->rdev.uctx);
595 queue_work(qhp->rhp->rdev.free_workq, &qhp->free_work);
688 PBL_OFF(&mhp->rhp->rdev, mhp->attr.pbl_addr)>>3));
784 rdev = &qhp->rhp->rdev;
834 c4iw_invalidate_mr(qhp->rhp,
873 c4iw_invalidate_mr(qhp->rhp, wr->ex.invalidate_rkey);
967 t4_ring_rq_db(&qhp->wq, idx, wqe, qhp->rhp->rdev.adap->iwt.wc_en);
1147 t4_wrq_tx(qhp->rhp->rdev.adap, wr);
1238 rdma_fini(struct c4iw_dev *rhp, struct c4iw_qp *qhp, struct c4iw_ep *ep)
1240 struct c4iw_rdev *rdev = &rhp->rdev;
1250 KASSERT(rhp == qhp->rhp && ep == qhp->ep, ("%s: EDOOFUS", __func__));
1333 static int rdma_init(struct c4iw_dev *rhp, struct c4iw_qp *qhp)
1339 struct c4iw_rdev *rdev = &qhp->rhp->rdev;
1353 ret = alloc_ird(rhp, qhp->attr.max_ird);
1410 free_ird(rhp, qhp->attr.max_ird);
1419 free_ird(rhp, qhp->attr.max_ird);
1424 int c4iw_modify_qp(struct c4iw_dev *rhp, struct c4iw_qp *qhp,
1464 if (attrs->max_ird > cur_max_read_depth(rhp)) {
1518 ret = rdma_init(rhp, qhp);
1543 ret = rdma_fini(rhp, qhp, ep);
1559 ret = rdma_fini(rhp, qhp, ep);
1676 struct c4iw_dev *rhp;
1682 rhp = qhp->rhp;
1686 c4iw_modify_qp(rhp, qhp, C4IW_QP_ATTR_NEXT_STATE, &attrs, 1);
1688 c4iw_modify_qp(rhp, qhp, C4IW_QP_ATTR_NEXT_STATE, &attrs, 0);
1691 remove_handle(rhp, &rhp->qpidr, qhp->wq.sq.qid);
1693 free_ird(rhp, qhp->attr.max_ird);
1705 struct c4iw_dev *rhp;
1723 rhp = php->rhp;
1724 schp = get_chp(rhp, ((struct c4iw_cq *)attrs->send_cq)->cq.cqid);
1725 rchp = get_chp(rhp, ((struct c4iw_cq *)attrs->recv_cq)->cq.cqid);
1732 if (attrs->cap.max_recv_wr > rhp->rdev.hw_queue.t4_max_rq_size)
1738 if (attrs->cap.max_send_wr > rhp->rdev.hw_queue.t4_max_sq_size)
1751 (sqsize + rhp->rdev.hw_queue.t4_eq_status_entries) *
1756 (rqsize + rhp->rdev.hw_queue.t4_eq_status_entries) *
1767 ret = create_qp(&rhp->rdev, &qhp->wq, &schp->cq, &rchp->cq,
1768 ucontext ? &ucontext->uctx : &rhp->rdev.uctx);
1776 qhp->rhp = rhp;
1799 ret = insert_handle(rhp, &rhp->qpidr, qhp, qhp->wq.sq.qid);
1825 uresp.qid_mask = rhp->rdev.qpmask;
1897 remove_handle(rhp, &rhp->qpidr, qhp->wq.sq.qid);
1899 destroy_qp(&rhp->rdev, &qhp->wq,
1900 ucontext ? &ucontext->uctx : &rhp->rdev.uctx);
1909 struct c4iw_dev *rhp;
1926 rhp = qhp->rhp;
1942 return c4iw_modify_qp(rhp, qhp, mask, &attrs, 0);