/linux-master/drivers/net/ethernet/marvell/octeontx2/nic/ |
H A D | otx2_common.c | 574 int otx2_alloc_buffer(struct otx2_nic *pfvf, struct otx2_cq_queue *cq, argument 577 if (unlikely(__otx2_alloc_rbuf(pfvf, cq->rbpool, dma))) 863 aq->rq.cq = qidx; 899 aq->sq.cq = pfvf->hw.rx_queues + qidx; 990 struct otx2_cq_queue *cq; local 992 cq = &qset->cq[qidx]; 993 cq->cq_idx = qidx; 996 cq->cq_type = CQ_RX; 997 cq 1077 struct otx2_cq_queue *cq; local [all...] |
H A D | otx2_txrx.h | 152 struct otx2_cq_queue *cq; member in struct:otx2_qset 173 int otx2_refill_pool_ptrs(void *dev, struct otx2_cq_queue *cq); 174 int cn10k_refill_pool_ptrs(void *dev, struct otx2_cq_queue *cq);
|
/linux-master/drivers/scsi/lpfc/ |
H A D | lpfc_debugfs.h | 462 struct lpfc_queue *wq, *cq, *eq; local 466 /* io wq and cq are 1:1, thus same indexes */ 471 cq = phba->sli4_hba.hdwq[wqidx].io_cq; 475 cq = phba->sli4_hba.mbx_cq; 479 cq = phba->sli4_hba.els_cq; 483 cq = phba->sli4_hba.nvmels_cq; 490 if (cq->assoc_qid == eq->queue_id) 502 qtypestr, wqidx, wq->queue_id, wqidx, cq->queue_id, 507 qtypestr, wq->queue_id, cq->queue_id, 510 lpfc_debug_dump_q(cq); [all...] |
/linux-master/drivers/infiniband/core/ |
H A D | verbs.c | 1030 srq->ext.cq = srq_init_attr->ext.cq; 1031 atomic_inc(&srq->ext.cq->usecnt); 1050 atomic_dec(&srq->ext.cq->usecnt); 1094 atomic_dec(&srq->ext.cq->usecnt); 2128 struct ib_cq *cq; local 2131 cq = rdma_zalloc_drv_obj(device, ib_cq); 2132 if (!cq) 2135 cq->device = device; 2136 cq 2157 rdma_set_cq_moderation(struct ib_cq *cq, u16 cq_count, u16 cq_period) argument 2168 ib_destroy_cq_user(struct ib_cq *cq, struct ib_udata *udata) argument 2188 ib_resize_cq(struct ib_cq *cq, int cqe) argument 2559 struct ib_cq *cq = wq->cq; local 2806 ib_drain_qp_done(struct ib_cq *cq, struct ib_wc *wc) argument 2819 struct ib_cq *cq = qp->send_cq; local 2858 struct ib_cq *cq = qp->recv_cq; local [all...] |
/linux-master/drivers/infiniband/sw/siw/ |
H A D | siw_verbs.c | 1101 struct siw_cq *cq = to_siw_cq(base_cq); local 1107 siw_dbg_cq(cq, "free CQ resources\n"); 1109 siw_cq_flush(cq); 1112 rdma_user_mmap_entry_remove(cq->cq_entry); 1116 vfree(cq->queue); 1134 struct siw_cq *cq = to_siw_cq(base_cq); local 1151 cq->base_cq.cqe = size; 1152 cq->num_cqe = size; 1155 cq->queue = vmalloc_user(size * sizeof(struct siw_cqe) + 1158 cq 1229 struct siw_cq *cq = to_siw_cq(base_cq); local 1258 struct siw_cq *cq = to_siw_cq(base_cq); local 1831 siw_cq_event(struct siw_cq *cq, enum ib_event_type etype) argument [all...] |
/linux-master/drivers/vfio/pci/mlx5/ |
H A D | cmd.c | 1025 struct mlx5_vhca_cq *cq) 1027 mlx5_core_destroy_cq(mdev, &cq->mcq); 1028 mlx5_frag_buf_free(mdev, &cq->buf.frag_buf); 1029 mlx5_db_free(mdev, &cq->db); 1038 tracker.cq.mcq)); 1086 tracker.cq.mcq); 1097 struct mlx5_vhca_cq *cq; local 1103 cq = &tracker->cq; 1105 err = mlx5_db_alloc_node(mdev, &cq 1024 mlx5vf_destroy_cq(struct mlx5_core_dev *mdev, struct mlx5_vhca_cq *cq) argument 1664 get_cqe(struct mlx5_vhca_cq *cq, int n) argument 1669 get_sw_cqe(struct mlx5_vhca_cq *cq, int n) argument 1685 mlx5vf_cq_poll_one(struct mlx5_vhca_cq *cq, struct mlx5_vhca_qp *qp, struct iova_bitmap *dirty, int *tracker_status) argument 1718 struct mlx5_vhca_cq *cq = &tracker->cq; local [all...] |
/linux-master/drivers/infiniband/hw/irdma/ |
H A D | puda.c | 207 * irdma_puda_poll_info - poll cq for completion 208 * @cq: cq for poll 211 static int irdma_puda_poll_info(struct irdma_sc_cq *cq, argument 214 struct irdma_cq_uk *cq_uk = &cq->cq_uk; 227 cqe = IRDMA_GET_CURRENT_CQ_ELEM(&cq->cq_uk); 236 if (cq->dev->hw_attrs.uk_attrs.hw_rev >= IRDMA_GEN_2) 255 /* update cq tail in cq shadow memory also */ 267 ibdev_dbg(to_ibdev(cq 323 irdma_puda_poll_cmpl(struct irdma_sc_dev *dev, struct irdma_sc_cq *cq, u32 *compl_err) argument 723 irdma_puda_cq_wqe(struct irdma_sc_dev *dev, struct irdma_sc_cq *cq) argument 778 struct irdma_sc_cq *cq = &rsrc->cq; local [all...] |
H A D | ctrl.c | 2467 * @cq: cq struct 2469 static inline void irdma_sc_cq_ack(struct irdma_sc_cq *cq) argument 2471 writel(cq->cq_uk.cq_id, cq->cq_uk.cq_ack_db); 2476 * @cq: cq struct 2477 * @info: cq initialization info 2479 int irdma_sc_cq_init(struct irdma_sc_cq *cq, struct irdma_cq_init_info *info) argument 2487 cq 2516 irdma_sc_cq_create(struct irdma_sc_cq *cq, u64 scratch, bool check_overflow, bool post_sq) argument 2589 irdma_sc_cq_destroy(struct irdma_sc_cq *cq, u64 scratch, bool post_sq) argument 2639 irdma_sc_cq_resize(struct irdma_sc_cq *cq, struct irdma_modify_cq_info *info) argument 2655 irdma_sc_cq_modify(struct irdma_sc_cq *cq, struct irdma_modify_cq_info *info, u64 scratch, bool post_sq) argument 3021 irdma_sc_find_reg_cq(struct irdma_sc_ceq *ceq, struct irdma_sc_cq *cq) argument 3039 irdma_sc_add_cq_ctx(struct irdma_sc_ceq *ceq, struct irdma_sc_cq *cq) argument 3062 irdma_sc_remove_cq_ctx(struct irdma_sc_ceq *ceq, struct irdma_sc_cq *cq) argument 3807 struct irdma_sc_cq *cq = NULL; local 3857 irdma_sc_cleanup_ceqes(struct irdma_sc_cq *cq, struct irdma_sc_ceq *ceq) argument 4166 irdma_sc_ccq_init(struct irdma_sc_cq *cq, struct irdma_ccq_init_info *info) argument [all...] |
H A D | puda.h | 100 struct irdma_sc_cq cq; member in struct:irdma_puda_rsrc 159 int irdma_puda_poll_cmpl(struct irdma_sc_dev *dev, struct irdma_sc_cq *cq, 172 int irdma_cqp_cq_create_cmd(struct irdma_sc_dev *dev, struct irdma_sc_cq *cq); 174 void irdma_cqp_cq_destroy_cmd(struct irdma_sc_dev *dev, struct irdma_sc_cq *cq);
|
/linux-master/net/9p/ |
H A D | trans_rdma.c | 52 * @cq: Completion Queue pointer 79 struct ib_cq *cq; member in struct:p9_trans_rdma 290 recv_done(struct ib_cq *cq, struct ib_wc *wc) argument 292 struct p9_client *client = cq->cq_context; 341 send_done(struct ib_cq *cq, struct ib_wc *wc) argument 343 struct p9_client *client = cq->cq_context; 373 if (rdma->cq && !IS_ERR(rdma->cq)) 374 ib_free_cq(rdma->cq); 697 rdma->cq [all...] |
/linux-master/drivers/net/ethernet/intel/ice/ |
H A D | ice_vf_mbx.c | 242 struct ice_ctl_q_info *cq = &hw->mailboxq; local 290 snap_buf->head = ICE_RQ_DATA_MASK(cq->rq.next_to_clean + 292 snap_buf->tail = ICE_RQ_DATA_MASK(cq->rq.next_to_clean - 1);
|
/linux-master/drivers/net/ethernet/pensando/ionic/ |
H A D | ionic_dev.h | 368 int ionic_cq_init(struct ionic_lif *lif, struct ionic_cq *cq, 371 void ionic_cq_map(struct ionic_cq *cq, void *base, dma_addr_t base_pa); 372 void ionic_cq_bind(struct ionic_cq *cq, struct ionic_queue *q); 373 typedef bool (*ionic_cq_cb)(struct ionic_cq *cq); 375 unsigned int ionic_cq_service(struct ionic_cq *cq, unsigned int work_to_do, 378 unsigned int ionic_tx_cq_service(struct ionic_cq *cq, unsigned int work_to_do);
|
/linux-master/drivers/net/ethernet/mellanox/mlx4/ |
H A D | en_rx.c | 36 #include <linux/mlx4/cq.h> 713 int mlx4_en_process_rx_cq(struct net_device *dev, struct mlx4_en_cq *cq, int budget) argument 720 int cq_ring = cq->ring; 740 index = cq->mcq.cons_index & ring->size_mask; 741 cqe = mlx4_en_get_cqe(cq->buf, index, priv->cqe_size) + factor; 745 cq->mcq.cons_index & cq->size)) { 883 skb = napi_get_frags(&cq->napi); 952 napi_gro_frags(&cq->napi); 958 ++cq 987 struct mlx4_en_cq *cq = container_of(mcq, struct mlx4_en_cq, mcq); local 999 struct mlx4_en_cq *cq = container_of(napi, struct mlx4_en_cq, napi); local [all...] |
/linux-master/drivers/infiniband/hw/qedr/ |
H A D | main.c | 440 struct qedr_cq *cq; local 454 cq = (struct qedr_cq *)(uintptr_t)HILO_U64(cq_handle->hi, 457 if (cq == NULL) { 466 if (cq->sig != QEDR_CQ_MAGIC_NUMBER) { 468 "Problem with cq signature, cq_handle->hi=%d ch_handle->lo=%d cq=%p\n", 469 cq_handle->hi, cq_handle->lo, cq); 473 cq->arm_flags = 0; 475 if (!cq->destroyed && cq 658 struct qedr_cq *cq; local [all...] |
/linux-master/drivers/infiniband/hw/ocrdma/ |
H A D | ocrdma_verbs.c | 925 static int ocrdma_copy_cq_uresp(struct ocrdma_dev *dev, struct ocrdma_cq *cq, argument 938 uresp.cq_id = cq->id; 939 uresp.page_size = PAGE_ALIGN(cq->len); 941 uresp.max_hw_cqe = cq->max_hw_cqe; 942 uresp.page_addr[0] = virt_to_phys(cq->va); 945 uresp.phase_change = cq->phase_change ? 1 : 0; 949 __func__, dev->id, cq->id); 960 cq->ucontext = uctx; 970 struct ocrdma_cq *cq = get_ocrdma_cq(ibcq); local 987 spin_lock_init(&cq 1017 struct ocrdma_cq *cq = get_ocrdma_cq(ibcq); local 1027 ocrdma_flush_cq(struct ocrdma_cq *cq) argument 1055 struct ocrdma_cq *cq = get_ocrdma_cq(ibcq); local 1588 ocrdma_discard_cqes(struct ocrdma_qp *qp, struct ocrdma_cq *cq) argument 2739 ocrdma_change_cq_phase(struct ocrdma_cq *cq, struct ocrdma_cqe *cqe, u16 cur_getp) argument 2751 ocrdma_poll_hwcq(struct ocrdma_cq *cq, int num_entries, struct ib_wc *ibwc) argument 2811 ocrdma_add_err_cqe(struct ocrdma_cq *cq, int num_entries, struct ocrdma_qp *qp, struct ib_wc *ibwc) argument 2840 struct ocrdma_cq *cq = get_ocrdma_cq(ibcq); local 2875 struct ocrdma_cq *cq = get_ocrdma_cq(ibcq); local [all...] |
/linux-master/drivers/scsi/elx/efct/ |
H A D | efct_hw.h | 388 struct sli4_queue cq[EFCT_HW_MAX_NUM_CQ]; member in struct:efct_hw 505 struct hw_cq *cq; member in struct:hw_mq 519 struct hw_cq *cq; member in struct:hw_wq 551 struct hw_cq *cq; member in struct:hw_rq 658 struct hw_cq *cq, u8 *cqe); 682 void efct_hw_cq_process(struct efct_hw *hw, struct hw_cq *cq); 684 efct_hw_wq_process(struct efct_hw *hw, struct hw_cq *cq, 687 efct_hw_xabt_process(struct efct_hw *hw, struct hw_cq *cq, 742 struct hw_mq *efct_hw_new_mq(struct hw_cq *cq, u32 entry_count); 744 *efct_hw_new_wq(struct hw_cq *cq, u3 [all...] |
/linux-master/drivers/infiniband/hw/bnxt_re/ |
H A D | qplib_fp.h | 100 struct bnxt_qplib_cq *cq; member in struct:bnxt_qplib_srq 484 struct bnxt_qplib_cq *cq); 509 struct bnxt_qplib_cq *cq; member in struct:bnxt_qplib_nq_work 551 int bnxt_qplib_create_cq(struct bnxt_qplib_res *res, struct bnxt_qplib_cq *cq); 552 int bnxt_qplib_resize_cq(struct bnxt_qplib_res *res, struct bnxt_qplib_cq *cq, 555 struct bnxt_qplib_cq *cq); 556 int bnxt_qplib_destroy_cq(struct bnxt_qplib_res *res, struct bnxt_qplib_cq *cq); 557 int bnxt_qplib_poll_cq(struct bnxt_qplib_cq *cq, struct bnxt_qplib_cqe *cqe, 559 bool bnxt_qplib_is_cq_empty(struct bnxt_qplib_cq *cq); 560 void bnxt_qplib_req_notify_cq(struct bnxt_qplib_cq *cq, u3 [all...] |
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/en/ |
H A D | xdp.c | 281 struct mlx5e_cq *cq; member in struct:mlx5e_xsk_tx_complete 291 if (mlx5_is_real_time_rq(priv->cq->mdev) || mlx5_is_real_time_sq(priv->cq->mdev)) 292 return mlx5_real_time_cyc2time(&priv->cq->mdev->clock, ts); 294 return mlx5_timecounter_cyc2time(&priv->cq->mdev->clock, ts); 663 struct mlx5e_cq *cq, 729 .cq = cq, 748 bool mlx5e_poll_xdpsq_cq(struct mlx5e_cq *cq) argument 759 sq = container_of(cq, struc 659 mlx5e_free_xdpsq_desc(struct mlx5e_xdpsq *sq, struct mlx5e_xdp_wqe_info *wi, u32 *xsk_frames, struct xdp_frame_bulk *bq, struct mlx5e_cq *cq, struct mlx5_cqe64 *cqe) argument [all...] |
/linux-master/drivers/net/ethernet/brocade/bna/ |
H A D | bna_tx_rx.c | 567 rxf->rit[offset] = rxp->cq.ccb->id; 1475 bna_ib_start(rx->bna, &rxp->cq.ib, is_regular); 1664 bfi_enet_datapath_q_init(&cfg_req->q_cfg[i].cq.q, 1665 &rxp->cq.qpt); 1668 rxp->cq.ib.ib_seg_host_addr.lsb; 1670 rxp->cq.ib.ib_seg_host_addr.msb; 1672 htons((u16)rxp->cq.ib.intr_vector); 1679 cfg_req->ib_cfg.msix = (rxp->cq.ib.intr_type == BNA_INTR_T_MSIX) 1683 htonl((u32)rxp->cq.ib.coalescing_timeo); 1685 htonl((u32)rxp->cq [all...] |
/linux-master/drivers/scsi/hisi_sas/ |
H A D | hisi_sas_main.c | 690 static void hisi_sas_sync_poll_cq(struct hisi_sas_cq *cq) argument 693 spin_lock(&cq->poll_lock); 694 spin_unlock(&cq->poll_lock); 697 static bool hisi_sas_queue_is_poll(struct hisi_sas_cq *cq) argument 699 struct hisi_hba *hisi_hba = cq->hisi_hba; 701 if (cq->id < hisi_hba->queue_count - hisi_hba->iopoll_q_cnt) 706 static void hisi_sas_sync_cq(struct hisi_sas_cq *cq) argument 708 if (hisi_sas_queue_is_poll(cq)) 709 hisi_sas_sync_poll_cq(cq); 711 synchronize_irq(cq 719 struct hisi_sas_cq *cq = &hisi_hba->cq[i]; local 732 struct hisi_sas_cq *cq = &hisi_hba->cq[i]; local 747 struct hisi_sas_cq *cq = local 1150 struct hisi_sas_cq *cq = &hisi_hba->cq[i]; local 1632 struct hisi_sas_cq *cq; local 1702 struct hisi_sas_cq *cq = &hisi_hba->cq[slot->dlvry_queue]; local 1992 struct hisi_sas_cq *cq = local 2141 struct hisi_sas_cq *cq = &hisi_hba->cq[i]; local 2193 struct hisi_sas_cq *cq = &hisi_hba->cq[i]; local [all...] |
/linux-master/drivers/net/ethernet/ibm/ehea/ |
H A D | ehea_hw.h | 232 static inline void ehea_update_feca(struct ehea_cq *cq, u32 nr_cqes) argument 234 struct h_epa epa = cq->epas.kernel; 239 static inline void ehea_reset_cq_n1(struct ehea_cq *cq) argument 241 struct h_epa epa = cq->epas.kernel;
|
/linux-master/drivers/infiniband/hw/usnic/ |
H A D | usnic_ib_verbs.h | 60 int usnic_ib_destroy_cq(struct ib_cq *cq, struct ib_udata *udata);
|
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/en/rep/ |
H A D | tc.h | 69 struct sk_buff *skb) { napi_gro_receive(rq->cq.napi, skb); }
|
/linux-master/drivers/scsi/snic/ |
H A D | snic_res.c | 126 svnic_cq_free(&snic->cq[i]); 156 SNIC_INFO("wq %d cq %d intr %d\n", snic->wq_count, 175 &snic->cq[i], 187 &snic->cq[i], 219 svnic_cq_init(&snic->cq[i],
|
/linux-master/drivers/infiniband/sw/rxe/ |
H A D | rxe_loc.h | 21 int rxe_cq_chk_attr(struct rxe_dev *rxe, struct rxe_cq *cq, 24 int rxe_cq_from_init(struct rxe_dev *rxe, struct rxe_cq *cq, int cqe, 28 int rxe_cq_resize_queue(struct rxe_cq *cq, int new_cqe, 32 int rxe_cq_post(struct rxe_cq *cq, struct rxe_cqe *cqe, int solicited);
|