Searched refs:cq_size (Results 1 - 25 of 44) sorted by relevance

12

/linux-master/include/uapi/rdma/
H A Dirdma-abi.h94 __u32 cq_size; member in struct:irdma_create_cq_resp
/linux-master/drivers/net/ethernet/mellanox/mlxbf_gige/
H A Dmlxbf_gige_rx.c84 size_t wq_size, cq_size; local
118 cq_size = MLXBF_GIGE_RX_CQE_SZ * priv->rx_q_entries;
119 priv->rx_cqe_base = dma_alloc_coherent(priv->dev, cq_size,
/linux-master/drivers/infiniband/ulp/iser/
H A Diser_verbs.c235 unsigned int max_send_wr, cq_size; local
250 cq_size = max_send_wr + ISER_QP_MAX_RECV_DTOS;
251 ib_conn->cq = ib_cq_pool_get(ib_dev, cq_size, -1, IB_POLL_SOFTIRQ);
256 ib_conn->cq_size = cq_size;
285 ib_cq_pool_put(ib_conn->cq, ib_conn->cq_size);
382 ib_cq_pool_put(ib_conn->cq, ib_conn->cq_size);
H A Discsi_iser.h361 * @cq_size: The number of max outstanding completions
371 u32 cq_size; member in struct:ib_conn
/linux-master/drivers/infiniband/ulp/isert/
H A Dib_isert.h184 u32 cq_size; member in struct:isert_conn
H A Dib_isert.c106 u32 cq_size = ISERT_QP_MAX_REQ_DTOS + ISERT_QP_MAX_RECV_DTOS + 2; local
112 isert_conn->cq = ib_cq_pool_get(ib_dev, cq_size, -1, IB_POLL_WORKQUEUE);
118 isert_conn->cq_size = cq_size;
140 ib_cq_pool_put(isert_conn->cq, isert_conn->cq_size);
412 ib_cq_pool_put(isert_conn->cq, isert_conn->cq_size);
/linux-master/drivers/infiniband/ulp/srpt/
H A Dib_srpt.h259 * @cq_size: Number of CQEs in @cq.
304 u32 cq_size; member in struct:srpt_rdma_ch
/linux-master/drivers/net/ethernet/amd/pds_core/
H A Ddebugfs.c122 debugfs_create_x32("cq_size", 0400, qcq_dentry, &qcq->cq_size);
H A Dcore.c153 dma_free_coherent(dev, qcq->cq_size,
237 qcq->q_size + qcq->cq_size,
270 qcq->cq_size = PDS_PAGE_SIZE + (num_descs * cq_desc_size);
271 qcq->cq_base = dma_alloc_coherent(dev, qcq->cq_size,
H A Dcore.h122 u32 cq_size; member in struct:pdsc_qcq
/linux-master/drivers/infiniband/hw/irdma/
H A Duser.h357 u32 cq_size; member in struct:irdma_cq_uk
393 u32 cq_size; member in struct:irdma_cq_uk_init_info
H A Dpuda.h115 u32 cq_size; member in struct:irdma_puda_rsrc
H A Dverbs.h119 u16 cq_size; member in struct:irdma_cq
H A Duk.c891 * @cq_size: number of cqes
893 void irdma_uk_cq_resize(struct irdma_cq_uk *cq, void *cq_base, int cq_size) argument
896 cq->cq_size = cq_size;
897 IRDMA_RING_INIT(cq->cq_ring, cq->cq_size);
1477 cq->cq_size = info->cq_size;
1482 IRDMA_RING_INIT(cq->cq_ring, cq->cq_size);
H A Dpuda.c736 set_64bit_val(wqe, 0, cq->cq_uk.cq_size);
786 cqsize = rsrc->cq_size * (sizeof(struct irdma_cqe));
798 info.shadow_read_threshold = rsrc->cq_size >> 2;
803 init_info->cq_size = rsrc->cq_size;
1062 rsrc->cq_size = info->rq_size + info->sq_size;
1065 rsrc->cq_size += info->rq_size;
H A Dverbs.c1921 info.cq_size = max(entries, 4);
1923 if (info.cq_size == iwcq->sc_cq.cq_uk.cq_size - 1)
1960 rsize = info.cq_size * sizeof(struct irdma_cqe);
2010 ibcq->cqe = info.cq_size - 1;
2045 #define IRDMA_CREATE_CQ_MIN_RESP_LEN offsetofend(struct irdma_create_cq_resp, cq_size)
2081 ukinfo->cq_size = max(entries, 4);
2083 iwcq->ibcq.cqe = info.cq_uk_init_info.cq_size;
2160 ukinfo->cq_size = entries;
2162 rsize = info.cq_uk_init_info.cq_size * sizeo
[all...]
/linux-master/drivers/net/ethernet/microsoft/mana/
H A Dhw_channel.c360 u32 eq_size, cq_size; local
367 cq_size = roundup_pow_of_two(GDMA_CQE_SIZE * q_depth);
368 if (cq_size < MINIMUM_SUPPORTED_PAGE_SIZE)
369 cq_size = MINIMUM_SUPPORTED_PAGE_SIZE;
382 err = mana_hwc_create_gdma_cq(hwc, cq_size, hwc_cq, mana_hwc_comp_event,
H A Dmana_en.c1140 req.cq_size = cq_spec->queue_size;
1879 u32 cq_size; local
1896 cq_size = MAX_SEND_BUFFERS_PER_QUEUE * COMP_ENTRY_SIZE;
1897 cq_size = PAGE_ALIGN(cq_size);
1930 spec.queue_size = cq_size;
2076 struct mana_rxq *rxq, u32 *rxq_size, u32 *cq_size)
2087 *cq_size = 0;
2109 *cq_size += COMP_ENTRY_SIZE;
2164 u32 cq_size, rq_siz local
2075 mana_alloc_rx_wqe(struct mana_port_context *apc, struct mana_rxq *rxq, u32 *rxq_size, u32 *cq_size) argument
[all...]
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/fpga/
H A Dconn.c411 static int mlx5_fpga_conn_create_cq(struct mlx5_fpga_conn *conn, int cq_size) argument
424 cq_size = roundup_pow_of_two(cq_size);
425 MLX5_SET(cqc, temp_cqc, log_cq_size, ilog2(cq_size));
455 MLX5_SET(cqc, cqc, log_cq_size, ilog2(cq_size));
/linux-master/drivers/net/ethernet/pensando/ionic/
H A Dionic_debugfs.c133 debugfs_create_x32("cq_size", 0400, qcq_dentry, &qcq->cq_size);
H A Dionic_lif.h74 u32 cq_size; member in struct:ionic_qcq
H A Dionic_lif.c422 dma_free_coherent(dev, qcq->cq_size, qcq->cq_base, qcq->cq_base_pa);
651 new->cq_size = PAGE_SIZE + (num_descs * cq_desc_size);
652 new->cq_base = dma_alloc_coherent(dev, new->cq_size, &new->cq_base_pa,
685 dma_free_coherent(dev, new->cq_size, new->cq_base, new->cq_base_pa);
776 memset(qcq->cq_base, 0, qcq->cq_size);
2889 swap(a->cq_size, b->cq_size);
/linux-master/drivers/dma/
H A Dhisi_dma.c586 size_t cq_size = sizeof(struct hisi_dma_cqe) * hdma_dev->chan_depth; local
598 chan->cq = dmam_alloc_coherent(dev, cq_size, &chan->cq_dma,
/linux-master/drivers/nvme/host/
H A Drdma.c99 int cq_size; member in struct:nvme_rdma_queue
418 ib_cq_pool_put(queue->ib_cq, queue->cq_size);
475 queue->ib_cq = ib_alloc_cq(ibdev, queue, queue->cq_size,
478 queue->ib_cq = ib_cq_pool_get(ibdev, queue->cq_size,
505 queue->cq_size = cq_factor * queue->queue_size + 1;
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/steering/
H A Ddr_send.c1223 int cq_size; local
1231 cq_size = QUEUE_SIZE + 1;
1232 dmn->send_ring->cq = dr_create_cq(dmn->mdev, dmn->uar, cq_size);

Completed in 305 milliseconds

12