Searched refs:cq (Results 101 - 125 of 350) sorted by relevance

1234567891011>>

/linux-master/drivers/net/ethernet/marvell/octeontx2/nic/
H A Dotx2_common.c574 int otx2_alloc_buffer(struct otx2_nic *pfvf, struct otx2_cq_queue *cq, argument
577 if (unlikely(__otx2_alloc_rbuf(pfvf, cq->rbpool, dma)))
863 aq->rq.cq = qidx;
899 aq->sq.cq = pfvf->hw.rx_queues + qidx;
990 struct otx2_cq_queue *cq; local
992 cq = &qset->cq[qidx];
993 cq->cq_idx = qidx;
996 cq->cq_type = CQ_RX;
997 cq
1077 struct otx2_cq_queue *cq; local
[all...]
H A Dotx2_txrx.h152 struct otx2_cq_queue *cq; member in struct:otx2_qset
173 int otx2_refill_pool_ptrs(void *dev, struct otx2_cq_queue *cq);
174 int cn10k_refill_pool_ptrs(void *dev, struct otx2_cq_queue *cq);
/linux-master/drivers/scsi/lpfc/
H A Dlpfc_debugfs.h462 struct lpfc_queue *wq, *cq, *eq; local
466 /* io wq and cq are 1:1, thus same indexes */
471 cq = phba->sli4_hba.hdwq[wqidx].io_cq;
475 cq = phba->sli4_hba.mbx_cq;
479 cq = phba->sli4_hba.els_cq;
483 cq = phba->sli4_hba.nvmels_cq;
490 if (cq->assoc_qid == eq->queue_id)
502 qtypestr, wqidx, wq->queue_id, wqidx, cq->queue_id,
507 qtypestr, wq->queue_id, cq->queue_id,
510 lpfc_debug_dump_q(cq);
[all...]
/linux-master/drivers/infiniband/core/
H A Dverbs.c1030 srq->ext.cq = srq_init_attr->ext.cq;
1031 atomic_inc(&srq->ext.cq->usecnt);
1050 atomic_dec(&srq->ext.cq->usecnt);
1094 atomic_dec(&srq->ext.cq->usecnt);
2128 struct ib_cq *cq; local
2131 cq = rdma_zalloc_drv_obj(device, ib_cq);
2132 if (!cq)
2135 cq->device = device;
2136 cq
2157 rdma_set_cq_moderation(struct ib_cq *cq, u16 cq_count, u16 cq_period) argument
2168 ib_destroy_cq_user(struct ib_cq *cq, struct ib_udata *udata) argument
2188 ib_resize_cq(struct ib_cq *cq, int cqe) argument
2559 struct ib_cq *cq = wq->cq; local
2806 ib_drain_qp_done(struct ib_cq *cq, struct ib_wc *wc) argument
2819 struct ib_cq *cq = qp->send_cq; local
2858 struct ib_cq *cq = qp->recv_cq; local
[all...]
/linux-master/drivers/infiniband/sw/siw/
H A Dsiw_verbs.c1101 struct siw_cq *cq = to_siw_cq(base_cq); local
1107 siw_dbg_cq(cq, "free CQ resources\n");
1109 siw_cq_flush(cq);
1112 rdma_user_mmap_entry_remove(cq->cq_entry);
1116 vfree(cq->queue);
1134 struct siw_cq *cq = to_siw_cq(base_cq); local
1151 cq->base_cq.cqe = size;
1152 cq->num_cqe = size;
1155 cq->queue = vmalloc_user(size * sizeof(struct siw_cqe) +
1158 cq
1229 struct siw_cq *cq = to_siw_cq(base_cq); local
1258 struct siw_cq *cq = to_siw_cq(base_cq); local
1831 siw_cq_event(struct siw_cq *cq, enum ib_event_type etype) argument
[all...]
/linux-master/drivers/vfio/pci/mlx5/
H A Dcmd.c1025 struct mlx5_vhca_cq *cq)
1027 mlx5_core_destroy_cq(mdev, &cq->mcq);
1028 mlx5_frag_buf_free(mdev, &cq->buf.frag_buf);
1029 mlx5_db_free(mdev, &cq->db);
1038 tracker.cq.mcq));
1086 tracker.cq.mcq);
1097 struct mlx5_vhca_cq *cq; local
1103 cq = &tracker->cq;
1105 err = mlx5_db_alloc_node(mdev, &cq
1024 mlx5vf_destroy_cq(struct mlx5_core_dev *mdev, struct mlx5_vhca_cq *cq) argument
1664 get_cqe(struct mlx5_vhca_cq *cq, int n) argument
1669 get_sw_cqe(struct mlx5_vhca_cq *cq, int n) argument
1685 mlx5vf_cq_poll_one(struct mlx5_vhca_cq *cq, struct mlx5_vhca_qp *qp, struct iova_bitmap *dirty, int *tracker_status) argument
1718 struct mlx5_vhca_cq *cq = &tracker->cq; local
[all...]
/linux-master/drivers/infiniband/hw/irdma/
H A Dpuda.c207 * irdma_puda_poll_info - poll cq for completion
208 * @cq: cq for poll
211 static int irdma_puda_poll_info(struct irdma_sc_cq *cq, argument
214 struct irdma_cq_uk *cq_uk = &cq->cq_uk;
227 cqe = IRDMA_GET_CURRENT_CQ_ELEM(&cq->cq_uk);
236 if (cq->dev->hw_attrs.uk_attrs.hw_rev >= IRDMA_GEN_2)
255 /* update cq tail in cq shadow memory also */
267 ibdev_dbg(to_ibdev(cq
323 irdma_puda_poll_cmpl(struct irdma_sc_dev *dev, struct irdma_sc_cq *cq, u32 *compl_err) argument
723 irdma_puda_cq_wqe(struct irdma_sc_dev *dev, struct irdma_sc_cq *cq) argument
778 struct irdma_sc_cq *cq = &rsrc->cq; local
[all...]
H A Dctrl.c2467 * @cq: cq struct
2469 static inline void irdma_sc_cq_ack(struct irdma_sc_cq *cq) argument
2471 writel(cq->cq_uk.cq_id, cq->cq_uk.cq_ack_db);
2476 * @cq: cq struct
2477 * @info: cq initialization info
2479 int irdma_sc_cq_init(struct irdma_sc_cq *cq, struct irdma_cq_init_info *info) argument
2487 cq
2516 irdma_sc_cq_create(struct irdma_sc_cq *cq, u64 scratch, bool check_overflow, bool post_sq) argument
2589 irdma_sc_cq_destroy(struct irdma_sc_cq *cq, u64 scratch, bool post_sq) argument
2639 irdma_sc_cq_resize(struct irdma_sc_cq *cq, struct irdma_modify_cq_info *info) argument
2655 irdma_sc_cq_modify(struct irdma_sc_cq *cq, struct irdma_modify_cq_info *info, u64 scratch, bool post_sq) argument
3021 irdma_sc_find_reg_cq(struct irdma_sc_ceq *ceq, struct irdma_sc_cq *cq) argument
3039 irdma_sc_add_cq_ctx(struct irdma_sc_ceq *ceq, struct irdma_sc_cq *cq) argument
3062 irdma_sc_remove_cq_ctx(struct irdma_sc_ceq *ceq, struct irdma_sc_cq *cq) argument
3807 struct irdma_sc_cq *cq = NULL; local
3857 irdma_sc_cleanup_ceqes(struct irdma_sc_cq *cq, struct irdma_sc_ceq *ceq) argument
4166 irdma_sc_ccq_init(struct irdma_sc_cq *cq, struct irdma_ccq_init_info *info) argument
[all...]
H A Dpuda.h100 struct irdma_sc_cq cq; member in struct:irdma_puda_rsrc
159 int irdma_puda_poll_cmpl(struct irdma_sc_dev *dev, struct irdma_sc_cq *cq,
172 int irdma_cqp_cq_create_cmd(struct irdma_sc_dev *dev, struct irdma_sc_cq *cq);
174 void irdma_cqp_cq_destroy_cmd(struct irdma_sc_dev *dev, struct irdma_sc_cq *cq);
/linux-master/net/9p/
H A Dtrans_rdma.c52 * @cq: Completion Queue pointer
79 struct ib_cq *cq; member in struct:p9_trans_rdma
290 recv_done(struct ib_cq *cq, struct ib_wc *wc) argument
292 struct p9_client *client = cq->cq_context;
341 send_done(struct ib_cq *cq, struct ib_wc *wc) argument
343 struct p9_client *client = cq->cq_context;
373 if (rdma->cq && !IS_ERR(rdma->cq))
374 ib_free_cq(rdma->cq);
697 rdma->cq
[all...]
/linux-master/drivers/net/ethernet/intel/ice/
H A Dice_vf_mbx.c242 struct ice_ctl_q_info *cq = &hw->mailboxq; local
290 snap_buf->head = ICE_RQ_DATA_MASK(cq->rq.next_to_clean +
292 snap_buf->tail = ICE_RQ_DATA_MASK(cq->rq.next_to_clean - 1);
/linux-master/drivers/net/ethernet/pensando/ionic/
H A Dionic_dev.h368 int ionic_cq_init(struct ionic_lif *lif, struct ionic_cq *cq,
371 void ionic_cq_map(struct ionic_cq *cq, void *base, dma_addr_t base_pa);
372 void ionic_cq_bind(struct ionic_cq *cq, struct ionic_queue *q);
373 typedef bool (*ionic_cq_cb)(struct ionic_cq *cq);
375 unsigned int ionic_cq_service(struct ionic_cq *cq, unsigned int work_to_do,
378 unsigned int ionic_tx_cq_service(struct ionic_cq *cq, unsigned int work_to_do);
/linux-master/drivers/net/ethernet/mellanox/mlx4/
H A Den_rx.c36 #include <linux/mlx4/cq.h>
713 int mlx4_en_process_rx_cq(struct net_device *dev, struct mlx4_en_cq *cq, int budget) argument
720 int cq_ring = cq->ring;
740 index = cq->mcq.cons_index & ring->size_mask;
741 cqe = mlx4_en_get_cqe(cq->buf, index, priv->cqe_size) + factor;
745 cq->mcq.cons_index & cq->size)) {
883 skb = napi_get_frags(&cq->napi);
952 napi_gro_frags(&cq->napi);
958 ++cq
987 struct mlx4_en_cq *cq = container_of(mcq, struct mlx4_en_cq, mcq); local
999 struct mlx4_en_cq *cq = container_of(napi, struct mlx4_en_cq, napi); local
[all...]
/linux-master/drivers/infiniband/hw/qedr/
H A Dmain.c440 struct qedr_cq *cq; local
454 cq = (struct qedr_cq *)(uintptr_t)HILO_U64(cq_handle->hi,
457 if (cq == NULL) {
466 if (cq->sig != QEDR_CQ_MAGIC_NUMBER) {
468 "Problem with cq signature, cq_handle->hi=%d ch_handle->lo=%d cq=%p\n",
469 cq_handle->hi, cq_handle->lo, cq);
473 cq->arm_flags = 0;
475 if (!cq->destroyed && cq
658 struct qedr_cq *cq; local
[all...]
/linux-master/drivers/infiniband/hw/ocrdma/
H A Docrdma_verbs.c925 static int ocrdma_copy_cq_uresp(struct ocrdma_dev *dev, struct ocrdma_cq *cq, argument
938 uresp.cq_id = cq->id;
939 uresp.page_size = PAGE_ALIGN(cq->len);
941 uresp.max_hw_cqe = cq->max_hw_cqe;
942 uresp.page_addr[0] = virt_to_phys(cq->va);
945 uresp.phase_change = cq->phase_change ? 1 : 0;
949 __func__, dev->id, cq->id);
960 cq->ucontext = uctx;
970 struct ocrdma_cq *cq = get_ocrdma_cq(ibcq); local
987 spin_lock_init(&cq
1017 struct ocrdma_cq *cq = get_ocrdma_cq(ibcq); local
1027 ocrdma_flush_cq(struct ocrdma_cq *cq) argument
1055 struct ocrdma_cq *cq = get_ocrdma_cq(ibcq); local
1588 ocrdma_discard_cqes(struct ocrdma_qp *qp, struct ocrdma_cq *cq) argument
2739 ocrdma_change_cq_phase(struct ocrdma_cq *cq, struct ocrdma_cqe *cqe, u16 cur_getp) argument
2751 ocrdma_poll_hwcq(struct ocrdma_cq *cq, int num_entries, struct ib_wc *ibwc) argument
2811 ocrdma_add_err_cqe(struct ocrdma_cq *cq, int num_entries, struct ocrdma_qp *qp, struct ib_wc *ibwc) argument
2840 struct ocrdma_cq *cq = get_ocrdma_cq(ibcq); local
2875 struct ocrdma_cq *cq = get_ocrdma_cq(ibcq); local
[all...]
/linux-master/drivers/scsi/elx/efct/
H A Defct_hw.h388 struct sli4_queue cq[EFCT_HW_MAX_NUM_CQ]; member in struct:efct_hw
505 struct hw_cq *cq; member in struct:hw_mq
519 struct hw_cq *cq; member in struct:hw_wq
551 struct hw_cq *cq; member in struct:hw_rq
658 struct hw_cq *cq, u8 *cqe);
682 void efct_hw_cq_process(struct efct_hw *hw, struct hw_cq *cq);
684 efct_hw_wq_process(struct efct_hw *hw, struct hw_cq *cq,
687 efct_hw_xabt_process(struct efct_hw *hw, struct hw_cq *cq,
742 struct hw_mq *efct_hw_new_mq(struct hw_cq *cq, u32 entry_count);
744 *efct_hw_new_wq(struct hw_cq *cq, u3
[all...]
/linux-master/drivers/infiniband/hw/bnxt_re/
H A Dqplib_fp.h100 struct bnxt_qplib_cq *cq; member in struct:bnxt_qplib_srq
484 struct bnxt_qplib_cq *cq);
509 struct bnxt_qplib_cq *cq; member in struct:bnxt_qplib_nq_work
551 int bnxt_qplib_create_cq(struct bnxt_qplib_res *res, struct bnxt_qplib_cq *cq);
552 int bnxt_qplib_resize_cq(struct bnxt_qplib_res *res, struct bnxt_qplib_cq *cq,
555 struct bnxt_qplib_cq *cq);
556 int bnxt_qplib_destroy_cq(struct bnxt_qplib_res *res, struct bnxt_qplib_cq *cq);
557 int bnxt_qplib_poll_cq(struct bnxt_qplib_cq *cq, struct bnxt_qplib_cqe *cqe,
559 bool bnxt_qplib_is_cq_empty(struct bnxt_qplib_cq *cq);
560 void bnxt_qplib_req_notify_cq(struct bnxt_qplib_cq *cq, u3
[all...]
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/en/
H A Dxdp.c281 struct mlx5e_cq *cq; member in struct:mlx5e_xsk_tx_complete
291 if (mlx5_is_real_time_rq(priv->cq->mdev) || mlx5_is_real_time_sq(priv->cq->mdev))
292 return mlx5_real_time_cyc2time(&priv->cq->mdev->clock, ts);
294 return mlx5_timecounter_cyc2time(&priv->cq->mdev->clock, ts);
663 struct mlx5e_cq *cq,
729 .cq = cq,
748 bool mlx5e_poll_xdpsq_cq(struct mlx5e_cq *cq) argument
759 sq = container_of(cq, struc
659 mlx5e_free_xdpsq_desc(struct mlx5e_xdpsq *sq, struct mlx5e_xdp_wqe_info *wi, u32 *xsk_frames, struct xdp_frame_bulk *bq, struct mlx5e_cq *cq, struct mlx5_cqe64 *cqe) argument
[all...]
/linux-master/drivers/net/ethernet/brocade/bna/
H A Dbna_tx_rx.c567 rxf->rit[offset] = rxp->cq.ccb->id;
1475 bna_ib_start(rx->bna, &rxp->cq.ib, is_regular);
1664 bfi_enet_datapath_q_init(&cfg_req->q_cfg[i].cq.q,
1665 &rxp->cq.qpt);
1668 rxp->cq.ib.ib_seg_host_addr.lsb;
1670 rxp->cq.ib.ib_seg_host_addr.msb;
1672 htons((u16)rxp->cq.ib.intr_vector);
1679 cfg_req->ib_cfg.msix = (rxp->cq.ib.intr_type == BNA_INTR_T_MSIX)
1683 htonl((u32)rxp->cq.ib.coalescing_timeo);
1685 htonl((u32)rxp->cq
[all...]
/linux-master/drivers/scsi/hisi_sas/
H A Dhisi_sas_main.c690 static void hisi_sas_sync_poll_cq(struct hisi_sas_cq *cq) argument
693 spin_lock(&cq->poll_lock);
694 spin_unlock(&cq->poll_lock);
697 static bool hisi_sas_queue_is_poll(struct hisi_sas_cq *cq) argument
699 struct hisi_hba *hisi_hba = cq->hisi_hba;
701 if (cq->id < hisi_hba->queue_count - hisi_hba->iopoll_q_cnt)
706 static void hisi_sas_sync_cq(struct hisi_sas_cq *cq) argument
708 if (hisi_sas_queue_is_poll(cq))
709 hisi_sas_sync_poll_cq(cq);
711 synchronize_irq(cq
719 struct hisi_sas_cq *cq = &hisi_hba->cq[i]; local
732 struct hisi_sas_cq *cq = &hisi_hba->cq[i]; local
747 struct hisi_sas_cq *cq = local
1150 struct hisi_sas_cq *cq = &hisi_hba->cq[i]; local
1632 struct hisi_sas_cq *cq; local
1702 struct hisi_sas_cq *cq = &hisi_hba->cq[slot->dlvry_queue]; local
1992 struct hisi_sas_cq *cq = local
2141 struct hisi_sas_cq *cq = &hisi_hba->cq[i]; local
2193 struct hisi_sas_cq *cq = &hisi_hba->cq[i]; local
[all...]
/linux-master/drivers/net/ethernet/ibm/ehea/
H A Dehea_hw.h232 static inline void ehea_update_feca(struct ehea_cq *cq, u32 nr_cqes) argument
234 struct h_epa epa = cq->epas.kernel;
239 static inline void ehea_reset_cq_n1(struct ehea_cq *cq) argument
241 struct h_epa epa = cq->epas.kernel;
/linux-master/drivers/infiniband/hw/usnic/
H A Dusnic_ib_verbs.h60 int usnic_ib_destroy_cq(struct ib_cq *cq, struct ib_udata *udata);
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/en/rep/
H A Dtc.h69 struct sk_buff *skb) { napi_gro_receive(rq->cq.napi, skb); }
/linux-master/drivers/scsi/snic/
H A Dsnic_res.c126 svnic_cq_free(&snic->cq[i]);
156 SNIC_INFO("wq %d cq %d intr %d\n", snic->wq_count,
175 &snic->cq[i],
187 &snic->cq[i],
219 svnic_cq_init(&snic->cq[i],
/linux-master/drivers/infiniband/sw/rxe/
H A Drxe_loc.h21 int rxe_cq_chk_attr(struct rxe_dev *rxe, struct rxe_cq *cq,
24 int rxe_cq_from_init(struct rxe_dev *rxe, struct rxe_cq *cq, int cqe,
28 int rxe_cq_resize_queue(struct rxe_cq *cq, int new_cqe,
32 int rxe_cq_post(struct rxe_cq *cq, struct rxe_cqe *cqe, int solicited);

Completed in 405 milliseconds

1234567891011>>