• Home
  • History
  • Annotate
  • Raw
  • Download
  • only in /asuswrt-rt-n18u-9.0.0.4.380.2695/release/src-rt-6.x.4708/linux/linux-2.6/drivers/infiniband/hw/cxgb4/

Lines Matching defs:cqe

193 	struct t4_cqe cqe;
197 memset(&cqe, 0, sizeof(cqe));
198 cqe.header = cpu_to_be32(V_CQE_STATUS(T4_ERR_SWFLUSH) |
203 cqe.bits_type_ts = cpu_to_be64(V_CQE_GENBIT((u64)cq->gen));
204 cq->sw_queue[cq->sw_pidx] = cqe;
226 struct t4_cqe cqe;
230 memset(&cqe, 0, sizeof(cqe));
231 cqe.header = cpu_to_be32(V_CQE_STATUS(T4_ERR_SWFLUSH) |
236 CQE_WRID_SQ_IDX(&cqe) = swcqe->idx;
237 cqe.bits_type_ts = cpu_to_be64(V_CQE_GENBIT((u64)cq->gen));
238 cq->sw_queue[cq->sw_pidx] = cqe;
265 struct t4_cqe *cqe = NULL, *swcqe;
269 ret = t4_next_hw_cqe(cq, &cqe);
274 *swcqe = *cqe;
278 ret = t4_next_hw_cqe(cq, &cqe);
282 static int cqe_completes_wr(struct t4_cqe *cqe, struct t4_wq *wq)
284 if (CQE_OPCODE(cqe) == FW_RI_TERMINATE)
287 if ((CQE_OPCODE(cqe) == FW_RI_RDMA_WRITE) && RQ_TYPE(cqe))
290 if ((CQE_OPCODE(cqe) == FW_RI_READ_RESP) && SQ_TYPE(cqe))
293 if (CQE_SEND_OPCODE(cqe) && RQ_TYPE(cqe) && t4_rq_empty(wq))
300 struct t4_cqe *cqe;
306 cqe = &cq->sw_queue[ptr];
307 if ((SQ_TYPE(cqe) || ((CQE_OPCODE(cqe) == FW_RI_READ_RESP) &&
309 (CQE_QPID(cqe) == wq->sq.qid))
319 struct t4_cqe *cqe;
326 cqe = &cq->sw_queue[ptr];
327 if (RQ_TYPE(cqe) && (CQE_OPCODE(cqe) != FW_RI_READ_RESP) &&
328 (CQE_QPID(cqe) == wq->rq.qid) && cqe_completes_wr(cqe, wq))
353 * Insert this completed cqe into the swcq.
355 PDBG("%s moving cqe into swcq sq idx %u cq idx %u\n",
357 swsqe->cqe.header |= htonl(V_CQE_SWCQE(1));
358 cq->sw_queue[cq->sw_pidx] = swsqe->cqe;
409 * cqe: copy of the polled CQE.
416 static int poll_cq(struct t4_wq *wq, struct t4_cq *cq, struct t4_cqe *cqe,
436 * skip cqe's not affiliated with a QP.
445 * 1) the cqe doesn't contain the sq_wptr from the wr.
521 swsqe->cqe = *hw_cqe;
528 *cqe = *hw_cqe;
554 PDBG("%s cq %p cqid 0x%x skip sw cqe cidx %u\n",
558 PDBG("%s cq %p cqid 0x%x skip hw cqe cidx %u\n",
569 * 0 cqe returned
577 struct t4_cqe cqe = {0, 0}, *rd_cqe;
596 ret = poll_cq(wq, &(chp->cq), &cqe, &cqe_flushed, &cookie, &credit);
602 wc->vendor_err = CQE_STATUS(&cqe);
606 "lo 0x%x cookie 0x%llx\n", __func__, CQE_QPID(&cqe),
607 CQE_TYPE(&cqe), CQE_OPCODE(&cqe), CQE_STATUS(&cqe), CQE_LEN(&cqe),
608 CQE_WRID_HI(&cqe), CQE_WRID_LOW(&cqe), (unsigned long long)cookie);
610 if (CQE_TYPE(&cqe) == 0) {
611 if (!CQE_STATUS(&cqe))
612 wc->byte_len = CQE_LEN(&cqe);
616 if (CQE_OPCODE(&cqe) == FW_RI_SEND_WITH_INV ||
617 CQE_OPCODE(&cqe) == FW_RI_SEND_WITH_SE_INV) {
618 wc->ex.invalidate_rkey = CQE_WRID_STAG(&cqe);
622 switch (CQE_OPCODE(&cqe)) {
628 wc->byte_len = CQE_LEN(&cqe);
652 CQE_OPCODE(&cqe), CQE_QPID(&cqe));
662 switch (CQE_STATUS(&cqe)) {
707 CQE_STATUS(&cqe), CQE_QPID(&cqe));
823 chp->ibcq.cqe = entries - 2;
882 int c4iw_resize_cq(struct ib_cq *cq, int cqe, struct ib_udata *udata)