Searched refs:cqe (Results 201 - 225 of 233) sorted by relevance

12345678910

/linux-master/drivers/net/ethernet/marvell/octeontx2/nic/
H A Dotx2_common.c1021 err = qmem_alloc(pfvf->dev, &cq->cqe, cq->cqe_cnt, cq->cqe_size);
1026 cq->cqe_base = cq->cqe->base;
1043 aq->cq.base = cq->cqe->iova;
/linux-master/drivers/net/ethernet/qlogic/qede/
H A Dqede_ethtool.c1552 union eth_rx_cqe *cqe; local
1579 cqe = (union eth_rx_cqe *)qed_chain_consume(&rxq->rx_comp_ring);
1584 fp_cqe = &cqe->fast_path_regular;
/linux-master/drivers/infiniband/sw/siw/
H A Dsiw_verbs.c1135 int rv, size = attr->cqe;
1151 cq->base_cq.cqe = size;
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/en_accel/
H A Dmacsec.c1680 struct mlx5_cqe64 *cqe)
1683 u32 macsec_meta_data = be32_to_cpu(cqe->ft_metadata);
1678 mlx5e_macsec_offload_handle_rx_skb(struct net_device *netdev, struct sk_buff *skb, struct mlx5_cqe64 *cqe) argument
/linux-master/drivers/infiniband/hw/mlx5/
H A Dqp.c2038 /* 0xffffff means we ask to work with cqe version 0 */
2183 /* 0xffffff means we ask to work with cqe version 0 */
2378 /* 0xffffff means we ask to work with cqe version 0 */
2521 /* 0xffffff means we ask to work with cqe version 0 */
5673 struct ib_cqe cqe; member in struct:mlx5_ib_drain_cqe
5679 struct mlx5_ib_drain_cqe *cqe = container_of(wc->wr_cqe, local
5681 cqe);
5683 complete(&cqe->done);
5745 { .wr_cqe = &sdrain.cqe, },
5759 sdrain.cqe
[all...]
H A Dmlx5_ib.h732 struct ib_cqe cqe; member in struct:mlx5_ib_umr_context
/linux-master/drivers/net/ethernet/broadcom/bnx2x/
H A Dbnx2x_sp.h425 * @param cqe Completion element we are handling
435 * error in the cqe).
438 union event_ring_elem *cqe,
H A Dbnx2x.h815 #define BNX2X_RX_SUM_FIX(cqe) \
816 BNX2X_PRS_FLAG_OVERETH_IPV4(cqe->fast_path_cqe.pars_flags.flags)
/linux-master/drivers/infiniband/hw/mlx4/
H A Dmlx4_ib.h113 int cqe; member in struct:mlx4_ib_cq_resize
/linux-master/include/rdma/
H A Dib_verbs.h372 unsigned int cqe; member in struct:ib_cq_init_attr
1599 int cqe; member in struct:ib_cq
2469 int (*resize_cq)(struct ib_cq *cq, int cqe, struct ib_udata *udata);
3913 * @cqe: The minimum size of the CQ.
3917 int ib_resize_cq(struct ib_cq *cq, int cqe);
/linux-master/drivers/infiniband/hw/cxgb4/
H A Diw_cxgb4.h1026 extern void c4iw_log_wr_stats(struct t4_wq *wq, struct t4_cqe *cqe);
/linux-master/drivers/crypto/hisilicon/
H A Ddebugfs.c360 cqe_curr = qp->cqe + cqe_id;
/linux-master/drivers/ufs/core/
H A Dufshcd.c798 * @cqe: pointer to the completion queue entry
805 struct cq_entry *cqe)
807 if (cqe)
808 return le32_to_cpu(cqe->status) & MASK_OCS;
5304 * @cqe: pointer to the completion queue entry
5310 struct cq_entry *cqe)
5328 ocs = ufshcd_get_tr_ocs(lrbp, cqe);
5485 * @cqe: pointer to the completion queue entry
5488 struct cq_entry *cqe)
5501 cmd->result = ufshcd_transfer_rsp_status(hba, lrbp, cqe);
804 ufshcd_get_tr_ocs(struct ufshcd_lrb *lrbp, struct cq_entry *cqe) argument
5309 ufshcd_transfer_rsp_status(struct ufs_hba *hba, struct ufshcd_lrb *lrbp, struct cq_entry *cqe) argument
5487 ufshcd_compl_one_cqe(struct ufs_hba *hba, int task_tag, struct cq_entry *cqe) argument
[all...]
/linux-master/drivers/infiniband/hw/irdma/
H A Dctrl.c3349 __le64 *cqe; local
3357 cqe = IRDMA_GET_CURRENT_EXTENDED_CQ_ELEM(&ccq->cq_uk);
3359 cqe = IRDMA_GET_CURRENT_CQ_ELEM(&ccq->cq_uk);
3361 get_64bit_val(cqe, 24, &temp);
3369 get_64bit_val(cqe, 8, &qp_ctx);
3384 get_64bit_val(cqe, 16, &temp1);
H A Dverbs.c1819 * @lcqe_buf: the buffer where the last cqe is received
2010 ibcq->cqe = info.cq_size - 1;
2059 int entries = attr->cqe;
2083 iwcq->ibcq.cqe = info.cq_uk_init_info.cq_size;
3671 * irdma_process_cqe - process cqe info
3672 * @entry: processed cqe
3673 * @cq_poll_info: cqe info
3808 /* save the resized CQ buffer which received the last cqe */
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/
H A Den_main.c1100 struct mlx5_cqe64 *cqe; local
1103 while ((cqe = mlx5_cqwq_get_cqe_enahnced_comp(cqwq)))
1106 while ((cqe = mlx5_cqwq_get_cqe(cqwq)))
2090 struct mlx5_cqe64 *cqe = mlx5_cqwq_get_wqe(&cq->wq, i); local
2092 cqe->op_own = 0xf1;
2093 cqe->validity_iteration_count = 0xff;
4553 netdev_warn(priv->netdev, "Disabling RX cqe compression\n");
4556 netdev_err(priv->netdev, "Failed disabling cqe compression err=%d\n", err);
H A Den_stats.c1197 ts_stats->pkts += stats->cqe;
2226 { MLX5E_DECLARE_PTP_CQ_STAT(struct mlx5e_ptp_cq_stats, cqe) },
H A Den_tc.c5636 bool mlx5e_tc_update_skb(struct mlx5_cqe64 *cqe, struct sk_buff *skb, argument
5673 bool mlx5e_tc_update_skb_nic(struct mlx5_cqe64 *cqe, struct sk_buff *skb) argument
5681 reg_b = be32_to_cpu(cqe->ft_metadata);
5689 return mlx5e_tc_update_skb(cqe, skb, mapping_ctx, mapped_obj_id, ct_priv, zone_restore_id,
/linux-master/drivers/infiniband/ulp/rtrs/
H A Drtrs-srv.c704 iu = container_of(wc->wr_cqe, struct rtrs_iu, cqe);
909 iu = container_of(wc->wr_cqe, struct rtrs_iu, cqe);
/linux-master/drivers/infiniband/hw/erdma/
H A Derdma_verbs.c1635 unsigned int depth = attr->cqe;
1644 cq->ibcq.cqe = depth;
/linux-master/drivers/infiniband/hw/efa/
H A Defa_verbs.c1098 int entries = attr->cqe;
1191 cq->ibcq.cqe = result.actual_depth;
/linux-master/drivers/gpu/drm/nouveau/nvkm/subdev/gsp/
H A Dr535.c158 struct r535_gsp_msg *cqe; local
197 cqe = (void *)((u8 *)gsp->shm.cmdq.ptr + 0x1000 + wptr * 0x1000);
199 memcpy(cqe, (u8 *)cmd + off, size);
/linux-master/drivers/net/ethernet/mellanox/mlx4/
H A Den_netdev.c1679 struct mlx4_cqe *cqe = NULL; local
1681 cqe = mlx4_en_get_cqe(cq->buf, j, priv->cqe_size) +
1683 cqe->owner_sr_opcode = MLX4_CQE_OWNER_MASK;
/linux-master/drivers/net/ethernet/broadcom/
H A Dcnic.c2922 union eth_rx_cqe *cqe, *cqe_ring = (union eth_rx_cqe *) local
2938 cqe = &cqe_ring[sw_cons & BNX2X_MAX_RCQ_DESC_CNT];
2939 cqe_fp_flags = cqe->fast_path_cqe.type_error_flags;
2941 cmd = le32_to_cpu(cqe->ramrod_cqe.conn_and_cmd_data);
/linux-master/drivers/scsi/lpfc/
H A Dlpfc_init.c938 switch (bf_get(lpfc_wcqe_c_code, &cq_event->cqe.wcqe_cmpl)) {
7324 switch (bf_get(lpfc_trailer_code, &cq_event->cqe.mcqe_cmpl)) {
7327 &cq_event->cqe.acqe_link);
7330 lpfc_sli4_async_fip_evt(phba, &cq_event->cqe.acqe_fip);
7334 &cq_event->cqe.acqe_dcbx);
7338 &cq_event->cqe.acqe_grp5);
7341 lpfc_sli4_async_fc_evt(phba, &cq_event->cqe.acqe_fc);
7344 lpfc_sli4_async_sli_evt(phba, &cq_event->cqe.acqe_sli);
7351 &cq_event->cqe.mcqe_cmpl));

Completed in 579 milliseconds

12345678910