Lines Matching defs:send_cq
729 static void mlx4_ib_lock_cqs(struct mlx4_ib_cq *send_cq, struct mlx4_ib_cq *recv_cq)
731 if (send_cq == recv_cq)
732 spin_lock_irq(&send_cq->lock);
733 else if (send_cq->mcq.cqn < recv_cq->mcq.cqn) {
734 spin_lock_irq(&send_cq->lock);
738 spin_lock_nested(&send_cq->lock, SINGLE_DEPTH_NESTING);
742 static void mlx4_ib_unlock_cqs(struct mlx4_ib_cq *send_cq, struct mlx4_ib_cq *recv_cq)
744 if (send_cq == recv_cq)
745 spin_unlock_irq(&send_cq->lock);
746 else if (send_cq->mcq.cqn < recv_cq->mcq.cqn) {
748 spin_unlock_irq(&send_cq->lock);
750 spin_unlock(&send_cq->lock);
768 struct mlx4_ib_cq *send_cq, *recv_cq;
776 send_cq = to_mcq(qp->ibqp.send_cq);
779 mlx4_ib_lock_cqs(send_cq, recv_cq);
784 if (send_cq != recv_cq)
785 __mlx4_ib_cq_clean(send_cq, qp->mqp.qpn, NULL);
790 mlx4_ib_unlock_cqs(send_cq, recv_cq);
1235 context->cqn_send = cpu_to_be32(to_mcq(ibqp->send_cq)->mcq.cqn);
1358 if (ibqp->send_cq != ibqp->recv_cq)
1359 mlx4_ib_cq_clean(to_mcq(ibqp->send_cq), qp->mqp.qpn, NULL);
1961 if (mlx4_wq_overflow(&qp->sq, nreq, qp->ibqp.send_cq)) {
2514 ibqp->send_cq = ibqp->recv_cq = xrcd->cq;