Lines Matching refs:cqe64
80 struct mlx5_cqe64 *cqe64;
82 cqe64 = (cq->mcq.cqe_sz == 64) ? cqe : cqe + 64;
84 if (likely((cqe64->op_own) >> 4 != MLX5_CQE_INVALID) &&
85 !((cqe64->op_own & MLX5_CQE_OWNER_MASK) ^ !!(n & (cq->ibcq.cqe + 1)))) {
355 static void handle_atomic(struct mlx5_ib_qp *qp, struct mlx5_cqe64 *cqe64,
365 byte_count = be32_to_cpu(cqe64->byte_cnt);
380 static void handle_atomics(struct mlx5_ib_qp *qp, struct mlx5_cqe64 *cqe64,
387 handle_atomic(qp, cqe64, idx);
517 struct mlx5_cqe64 *cqe64;
535 cqe64 = (cq->mcq.cqe_sz == 64) ? cqe : cqe + 64;
544 opcode = cqe64->op_own >> 4;
557 qpn = ntohl(cqe64->sop_drop_qpn) & 0xffffff;
571 wqe_ctr = be16_to_cpu(cqe64->wqe_counter);
573 handle_good_req(wc, cqe64, wq, idx);
574 handle_atomics(*cur_qp, cqe64, wq->last_poll, idx);
583 handle_responder(wc, cqe64, *cur_qp);
590 err_cqe = (struct mlx5_err_cqe *)cqe64;
599 wqe_ctr = be16_to_cpu(cqe64->wqe_counter);
608 wqe_ctr = be16_to_cpu(cqe64->wqe_counter);
619 sig_err_cqe = (struct mlx5_sig_err_cqe *)cqe64;
836 struct mlx5_cqe64 *cqe64;
840 cqe64 = buf->cqe_size == 64 ? cqe : cqe + 64;
841 cqe64->op_own = MLX5_CQE_INVALID << 4;
1039 static int is_equal_rsn(struct mlx5_cqe64 *cqe64, u32 rsn)
1041 return rsn == (ntohl(cqe64->sop_drop_qpn) & 0xffffff);
1046 struct mlx5_cqe64 *cqe64, *dest64;
1070 cqe64 = (cq->mcq.cqe_sz == 64) ? cqe : cqe + 64;
1071 if (is_equal_rsn(cqe64, rsn)) {
1072 if (srq && (ntohl(cqe64->srqn) & 0xffffff))
1073 mlx5_ib_free_srq_wqe(srq, be16_to_cpu(cqe64->wqe_counter));