• Home
  • History
  • Annotate
  • Raw
  • Download
  • only in /freebsd-12-stable/sys/dev/mlx4/mlx4_ib/

Lines Matching refs:send_cq

54 static void mlx4_ib_lock_cqs(struct mlx4_ib_cq *send_cq,
56 static void mlx4_ib_unlock_cqs(struct mlx4_ib_cq *send_cq,
887 mlx4_ib_lock_cqs(to_mcq(init_attr->send_cq),
896 mcq = to_mcq(init_attr->send_cq);
900 mlx4_ib_unlock_cqs(to_mcq(init_attr->send_cq),
957 static void mlx4_ib_lock_cqs(struct mlx4_ib_cq *send_cq, struct mlx4_ib_cq *recv_cq)
958 __acquires(&send_cq->lock) __acquires(&recv_cq->lock)
960 if (send_cq == recv_cq) {
961 spin_lock(&send_cq->lock);
963 } else if (send_cq->mcq.cqn < recv_cq->mcq.cqn) {
964 spin_lock(&send_cq->lock);
968 spin_lock_nested(&send_cq->lock, SINGLE_DEPTH_NESTING);
972 static void mlx4_ib_unlock_cqs(struct mlx4_ib_cq *send_cq, struct mlx4_ib_cq *recv_cq)
973 __releases(&send_cq->lock) __releases(&recv_cq->lock)
975 if (send_cq == recv_cq) {
977 spin_unlock(&send_cq->lock);
978 } else if (send_cq->mcq.cqn < recv_cq->mcq.cqn) {
980 spin_unlock(&send_cq->lock);
982 spin_unlock(&send_cq->lock);
1006 struct mlx4_ib_cq **send_cq, struct mlx4_ib_cq **recv_cq)
1010 *send_cq = to_mcq(to_mxrcd(qp->ibqp.xrcd)->cq);
1011 *recv_cq = *send_cq;
1014 *send_cq = to_mcq(qp->ibqp.send_cq);
1015 *recv_cq = *send_cq;
1018 *send_cq = to_mcq(qp->ibqp.send_cq);
1027 struct mlx4_ib_cq *send_cq, *recv_cq;
1058 get_cqs(qp, &send_cq, &recv_cq);
1061 mlx4_ib_lock_cqs(send_cq, recv_cq);
1070 if (send_cq != recv_cq)
1071 __mlx4_ib_cq_clean(send_cq, qp->mqp.qpn, NULL);
1076 mlx4_ib_unlock_cqs(send_cq, recv_cq);
1176 init_attr->send_cq = to_mxrcd(init_attr->xrcd)->cq;
1181 init_attr->recv_cq = init_attr->send_cq;
1626 struct mlx4_ib_cq *send_cq, *recv_cq;
1837 get_cqs(qp, &send_cq, &recv_cq);
1839 context->cqn_send = cpu_to_be32(send_cq->mcq.cqn);
2049 if (send_cq != recv_cq)
2050 mlx4_ib_cq_clean(send_cq, qp->mqp.qpn, NULL);
2989 if (mlx4_wq_overflow(&qp->sq, nreq, qp->ibqp.send_cq)) {