Lines Matching refs:cqe

80 				   struct lpfc_queue *cq, struct lpfc_cqe *cqe);
92 struct lpfc_cqe *cqe);
592 struct lpfc_cqe *cqe = NULL;
611 cqe = lpfc_sli4_cq_get(cq);
612 while (cqe) {
613 __lpfc_sli4_consume_cqe(phba, cq, cqe);
615 cqe = lpfc_sli4_cq_get(cq);
685 struct lpfc_cqe *cqe;
690 cqe = lpfc_sli4_qe(q, q->host_index);
693 if (bf_get_le32(lpfc_cqe_valid, cqe) != q->qe_valid)
705 return cqe;
710 struct lpfc_cqe *cqe)
713 bf_set_le32(lpfc_cqe_valid, cqe, 0);
3235 size = bf_get(lpfc_rcqe_length, &nvmebuf->cq_event.cqe.rcqe_cmpl);
4535 switch (bf_get(lpfc_wcqe_c_code, &cq_event->cqe.wcqe_cmpl)) {
14140 lpfc_sli4_els_xri_aborted(phba, &cq_event->cqe.wcqe_axri);
14174 wcqe = &irspiocbq->cq_event.cqe.wcqe_cmpl;
14221 memcpy(&cq_event->cqe, entry, size);
14416 * @cqe: Pointer to mailbox completion queue entry.
14426 struct lpfc_cqe *cqe)
14434 lpfc_sli4_pcimem_bcopy(cqe, &mcqe, sizeof(struct lpfc_mcqe));
14491 memcpy(&irspiocbq->cq_event.cqe.wcqe_cmpl, wcqe, sizeof(*wcqe));
14638 memcpy(&dma_buf->cq_event.cqe.rcqe_cmpl, rcqe, sizeof(*rcqe));
14723 * @cqe: Pointer to a completion queue entry.
14732 struct lpfc_cqe *cqe)
14738 lpfc_sli4_pcimem_bcopy(cqe, &cqevt, sizeof(struct lpfc_cqe));
14833 * @handler: Routine to process each cqe
14854 struct lpfc_cqe *cqe;
14867 cqe = lpfc_sli4_cq_get(cq);
14868 while (cqe) {
14869 workposted |= handler(phba, cq, cqe);
14870 __lpfc_sli4_consume_cqe(phba, cq, cqe);
14886 cqe = lpfc_sli4_cq_get(cq);
15270 * @cqe: Pointer to fast-path completion queue entry.
15279 struct lpfc_cqe *cqe)
15285 lpfc_sli4_pcimem_bcopy(cqe, &wcqe, sizeof(struct lpfc_cqe));
19183 &seq_dmabuf->cq_event.cqe.rcqe_cmpl);
19216 &hbq_buf->cq_event.cqe.rcqe_cmpl);
19236 &hbq_buf->cq_event.cqe.rcqe_cmpl);
19328 frame_len = bf_get(lpfc_rcqe_length, &dmabuf->cq_event.cqe.rcqe_cmpl);
19457 &dmabuf->cq_event.cqe.rcqe_cmpl) == CQE_CODE_RECEIVE_V1))
19459 &dmabuf->cq_event.cqe.rcqe_cmpl);
19462 &dmabuf->cq_event.cqe.rcqe_cmpl);
19469 &dmabuf->cq_event.cqe.rcqe_cmpl));