/linux-master/drivers/net/ethernet/marvell/octeontx2/nic/ |
H A D | otx2_common.c | 1021 err = qmem_alloc(pfvf->dev, &cq->cqe, cq->cqe_cnt, cq->cqe_size); 1026 cq->cqe_base = cq->cqe->base; 1043 aq->cq.base = cq->cqe->iova;
|
/linux-master/drivers/net/ethernet/qlogic/qede/ |
H A D | qede_ethtool.c | 1552 union eth_rx_cqe *cqe; local 1579 cqe = (union eth_rx_cqe *)qed_chain_consume(&rxq->rx_comp_ring); 1584 fp_cqe = &cqe->fast_path_regular;
|
/linux-master/drivers/infiniband/sw/siw/ |
H A D | siw_verbs.c | 1135 int rv, size = attr->cqe; 1151 cq->base_cq.cqe = size;
|
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/en_accel/ |
H A D | macsec.c | 1680 struct mlx5_cqe64 *cqe) 1683 u32 macsec_meta_data = be32_to_cpu(cqe->ft_metadata); 1678 mlx5e_macsec_offload_handle_rx_skb(struct net_device *netdev, struct sk_buff *skb, struct mlx5_cqe64 *cqe) argument
|
/linux-master/drivers/infiniband/hw/mlx5/ |
H A D | qp.c | 2038 /* 0xffffff means we ask to work with cqe version 0 */ 2183 /* 0xffffff means we ask to work with cqe version 0 */ 2378 /* 0xffffff means we ask to work with cqe version 0 */ 2521 /* 0xffffff means we ask to work with cqe version 0 */ 5673 struct ib_cqe cqe; member in struct:mlx5_ib_drain_cqe 5679 struct mlx5_ib_drain_cqe *cqe = container_of(wc->wr_cqe, local 5681 cqe); 5683 complete(&cqe->done); 5745 { .wr_cqe = &sdrain.cqe, }, 5759 sdrain.cqe [all...] |
H A D | mlx5_ib.h | 732 struct ib_cqe cqe; member in struct:mlx5_ib_umr_context
|
/linux-master/drivers/net/ethernet/broadcom/bnx2x/ |
H A D | bnx2x_sp.h | 425 * @param cqe Completion element we are handling 435 * error in the cqe). 438 union event_ring_elem *cqe,
|
H A D | bnx2x.h | 815 #define BNX2X_RX_SUM_FIX(cqe) \ 816 BNX2X_PRS_FLAG_OVERETH_IPV4(cqe->fast_path_cqe.pars_flags.flags)
|
/linux-master/drivers/infiniband/hw/mlx4/ |
H A D | mlx4_ib.h | 113 int cqe; member in struct:mlx4_ib_cq_resize
|
/linux-master/include/rdma/ |
H A D | ib_verbs.h | 372 unsigned int cqe; member in struct:ib_cq_init_attr 1599 int cqe; member in struct:ib_cq 2469 int (*resize_cq)(struct ib_cq *cq, int cqe, struct ib_udata *udata); 3913 * @cqe: The minimum size of the CQ. 3917 int ib_resize_cq(struct ib_cq *cq, int cqe);
|
/linux-master/drivers/infiniband/hw/cxgb4/ |
H A D | iw_cxgb4.h | 1026 extern void c4iw_log_wr_stats(struct t4_wq *wq, struct t4_cqe *cqe);
|
/linux-master/drivers/crypto/hisilicon/ |
H A D | debugfs.c | 360 cqe_curr = qp->cqe + cqe_id;
|
/linux-master/drivers/ufs/core/ |
H A D | ufshcd.c | 798 * @cqe: pointer to the completion queue entry 805 struct cq_entry *cqe) 807 if (cqe) 808 return le32_to_cpu(cqe->status) & MASK_OCS; 5304 * @cqe: pointer to the completion queue entry 5310 struct cq_entry *cqe) 5328 ocs = ufshcd_get_tr_ocs(lrbp, cqe); 5485 * @cqe: pointer to the completion queue entry 5488 struct cq_entry *cqe) 5501 cmd->result = ufshcd_transfer_rsp_status(hba, lrbp, cqe); 804 ufshcd_get_tr_ocs(struct ufshcd_lrb *lrbp, struct cq_entry *cqe) argument 5309 ufshcd_transfer_rsp_status(struct ufs_hba *hba, struct ufshcd_lrb *lrbp, struct cq_entry *cqe) argument 5487 ufshcd_compl_one_cqe(struct ufs_hba *hba, int task_tag, struct cq_entry *cqe) argument [all...] |
/linux-master/drivers/infiniband/hw/irdma/ |
H A D | ctrl.c | 3349 __le64 *cqe; local 3357 cqe = IRDMA_GET_CURRENT_EXTENDED_CQ_ELEM(&ccq->cq_uk); 3359 cqe = IRDMA_GET_CURRENT_CQ_ELEM(&ccq->cq_uk); 3361 get_64bit_val(cqe, 24, &temp); 3369 get_64bit_val(cqe, 8, &qp_ctx); 3384 get_64bit_val(cqe, 16, &temp1);
|
H A D | verbs.c | 1819 * @lcqe_buf: the buffer where the last cqe is received 2010 ibcq->cqe = info.cq_size - 1; 2059 int entries = attr->cqe; 2083 iwcq->ibcq.cqe = info.cq_uk_init_info.cq_size; 3671 * irdma_process_cqe - process cqe info 3672 * @entry: processed cqe 3673 * @cq_poll_info: cqe info 3808 /* save the resized CQ buffer which received the last cqe */
|
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/ |
H A D | en_main.c | 1100 struct mlx5_cqe64 *cqe; local 1103 while ((cqe = mlx5_cqwq_get_cqe_enahnced_comp(cqwq))) 1106 while ((cqe = mlx5_cqwq_get_cqe(cqwq))) 2090 struct mlx5_cqe64 *cqe = mlx5_cqwq_get_wqe(&cq->wq, i); local 2092 cqe->op_own = 0xf1; 2093 cqe->validity_iteration_count = 0xff; 4553 netdev_warn(priv->netdev, "Disabling RX cqe compression\n"); 4556 netdev_err(priv->netdev, "Failed disabling cqe compression err=%d\n", err);
|
H A D | en_stats.c | 1197 ts_stats->pkts += stats->cqe; 2226 { MLX5E_DECLARE_PTP_CQ_STAT(struct mlx5e_ptp_cq_stats, cqe) },
|
H A D | en_tc.c | 5636 bool mlx5e_tc_update_skb(struct mlx5_cqe64 *cqe, struct sk_buff *skb, argument 5673 bool mlx5e_tc_update_skb_nic(struct mlx5_cqe64 *cqe, struct sk_buff *skb) argument 5681 reg_b = be32_to_cpu(cqe->ft_metadata); 5689 return mlx5e_tc_update_skb(cqe, skb, mapping_ctx, mapped_obj_id, ct_priv, zone_restore_id,
|
/linux-master/drivers/infiniband/ulp/rtrs/ |
H A D | rtrs-srv.c | 704 iu = container_of(wc->wr_cqe, struct rtrs_iu, cqe); 909 iu = container_of(wc->wr_cqe, struct rtrs_iu, cqe);
|
/linux-master/drivers/infiniband/hw/erdma/ |
H A D | erdma_verbs.c | 1635 unsigned int depth = attr->cqe; 1644 cq->ibcq.cqe = depth;
|
/linux-master/drivers/infiniband/hw/efa/ |
H A D | efa_verbs.c | 1098 int entries = attr->cqe; 1191 cq->ibcq.cqe = result.actual_depth;
|
/linux-master/drivers/gpu/drm/nouveau/nvkm/subdev/gsp/ |
H A D | r535.c | 158 struct r535_gsp_msg *cqe; local 197 cqe = (void *)((u8 *)gsp->shm.cmdq.ptr + 0x1000 + wptr * 0x1000); 199 memcpy(cqe, (u8 *)cmd + off, size);
|
/linux-master/drivers/net/ethernet/mellanox/mlx4/ |
H A D | en_netdev.c | 1679 struct mlx4_cqe *cqe = NULL; local 1681 cqe = mlx4_en_get_cqe(cq->buf, j, priv->cqe_size) + 1683 cqe->owner_sr_opcode = MLX4_CQE_OWNER_MASK;
|
/linux-master/drivers/net/ethernet/broadcom/ |
H A D | cnic.c | 2922 union eth_rx_cqe *cqe, *cqe_ring = (union eth_rx_cqe *) local 2938 cqe = &cqe_ring[sw_cons & BNX2X_MAX_RCQ_DESC_CNT]; 2939 cqe_fp_flags = cqe->fast_path_cqe.type_error_flags; 2941 cmd = le32_to_cpu(cqe->ramrod_cqe.conn_and_cmd_data);
|
/linux-master/drivers/scsi/lpfc/ |
H A D | lpfc_init.c | 938 switch (bf_get(lpfc_wcqe_c_code, &cq_event->cqe.wcqe_cmpl)) { 7324 switch (bf_get(lpfc_trailer_code, &cq_event->cqe.mcqe_cmpl)) { 7327 &cq_event->cqe.acqe_link); 7330 lpfc_sli4_async_fip_evt(phba, &cq_event->cqe.acqe_fip); 7334 &cq_event->cqe.acqe_dcbx); 7338 &cq_event->cqe.acqe_grp5); 7341 lpfc_sli4_async_fc_evt(phba, &cq_event->cqe.acqe_fc); 7344 lpfc_sli4_async_sli_evt(phba, &cq_event->cqe.acqe_sli); 7351 &cq_event->cqe.mcqe_cmpl));
|