Lines Matching defs:send_cq
719 qp_context->cqn_snd = cpu_to_be32(to_mcq(ibqp->send_cq)->cqn);
824 if (qp->ibqp.send_cq != qp->ibqp.recv_cq)
825 mthca_cq_clean(dev, to_mcq(qp->ibqp.send_cq), qp->qpn, NULL);
1171 struct mthca_cq *send_cq,
1292 struct mthca_cq *send_cq,
1319 err = mthca_alloc_qp_common(dev, pd, send_cq, recv_cq,
1334 static void mthca_lock_cqs(struct mthca_cq *send_cq, struct mthca_cq *recv_cq)
1336 if (send_cq == recv_cq)
1337 spin_lock_irq(&send_cq->lock);
1338 else if (send_cq->cqn < recv_cq->cqn) {
1339 spin_lock_irq(&send_cq->lock);
1343 spin_lock_nested(&send_cq->lock, SINGLE_DEPTH_NESTING);
1347 static void mthca_unlock_cqs(struct mthca_cq *send_cq, struct mthca_cq *recv_cq)
1349 if (send_cq == recv_cq)
1350 spin_unlock_irq(&send_cq->lock);
1351 else if (send_cq->cqn < recv_cq->cqn) {
1353 spin_unlock_irq(&send_cq->lock);
1355 spin_unlock(&send_cq->lock);
1362 struct mthca_cq *send_cq,
1398 err = mthca_alloc_qp_common(dev, pd, send_cq, recv_cq,
1412 mthca_lock_cqs(send_cq, recv_cq);
1418 mthca_unlock_cqs(send_cq, recv_cq);
1442 struct mthca_cq *send_cq;
1445 send_cq = to_mcq(qp->ibqp.send_cq);
1452 mthca_lock_cqs(send_cq, recv_cq);
1460 mthca_unlock_cqs(send_cq, recv_cq);
1476 if (send_cq != recv_cq)
1477 mthca_cq_clean(dev, send_cq, qp->qpn, NULL);
1604 if (mthca_wq_overflow(&qp->sq, nreq, qp->ibqp.send_cq)) {
1985 if (mthca_wq_overflow(&qp->sq, nreq, qp->ibqp.send_cq)) {