/linux-master/include/uapi/rdma/ |
H A D | irdma-abi.h | 94 __u32 cq_size; member in struct:irdma_create_cq_resp
|
/linux-master/drivers/net/ethernet/mellanox/mlxbf_gige/ |
H A D | mlxbf_gige_rx.c | 84 size_t wq_size, cq_size; local 118 cq_size = MLXBF_GIGE_RX_CQE_SZ * priv->rx_q_entries; 119 priv->rx_cqe_base = dma_alloc_coherent(priv->dev, cq_size,
|
/linux-master/drivers/infiniband/ulp/iser/ |
H A D | iser_verbs.c | 235 unsigned int max_send_wr, cq_size; local 250 cq_size = max_send_wr + ISER_QP_MAX_RECV_DTOS; 251 ib_conn->cq = ib_cq_pool_get(ib_dev, cq_size, -1, IB_POLL_SOFTIRQ); 256 ib_conn->cq_size = cq_size; 285 ib_cq_pool_put(ib_conn->cq, ib_conn->cq_size); 382 ib_cq_pool_put(ib_conn->cq, ib_conn->cq_size);
|
H A D | iscsi_iser.h | 361 * @cq_size: The number of max outstanding completions 371 u32 cq_size; member in struct:ib_conn
|
/linux-master/drivers/infiniband/ulp/isert/ |
H A D | ib_isert.h | 184 u32 cq_size; member in struct:isert_conn
|
H A D | ib_isert.c | 106 u32 cq_size = ISERT_QP_MAX_REQ_DTOS + ISERT_QP_MAX_RECV_DTOS + 2; local 112 isert_conn->cq = ib_cq_pool_get(ib_dev, cq_size, -1, IB_POLL_WORKQUEUE); 118 isert_conn->cq_size = cq_size; 140 ib_cq_pool_put(isert_conn->cq, isert_conn->cq_size); 412 ib_cq_pool_put(isert_conn->cq, isert_conn->cq_size);
|
/linux-master/drivers/infiniband/ulp/srpt/ |
H A D | ib_srpt.h | 259 * @cq_size: Number of CQEs in @cq. 304 u32 cq_size; member in struct:srpt_rdma_ch
|
/linux-master/drivers/net/ethernet/amd/pds_core/ |
H A D | debugfs.c | 122 debugfs_create_x32("cq_size", 0400, qcq_dentry, &qcq->cq_size);
|
H A D | core.c | 153 dma_free_coherent(dev, qcq->cq_size, 237 qcq->q_size + qcq->cq_size, 270 qcq->cq_size = PDS_PAGE_SIZE + (num_descs * cq_desc_size); 271 qcq->cq_base = dma_alloc_coherent(dev, qcq->cq_size,
|
H A D | core.h | 122 u32 cq_size; member in struct:pdsc_qcq
|
/linux-master/drivers/infiniband/hw/irdma/ |
H A D | user.h | 357 u32 cq_size; member in struct:irdma_cq_uk 393 u32 cq_size; member in struct:irdma_cq_uk_init_info
|
H A D | puda.h | 115 u32 cq_size; member in struct:irdma_puda_rsrc
|
H A D | verbs.h | 119 u16 cq_size; member in struct:irdma_cq
|
H A D | uk.c | 891 * @cq_size: number of cqes 893 void irdma_uk_cq_resize(struct irdma_cq_uk *cq, void *cq_base, int cq_size) argument 896 cq->cq_size = cq_size; 897 IRDMA_RING_INIT(cq->cq_ring, cq->cq_size); 1477 cq->cq_size = info->cq_size; 1482 IRDMA_RING_INIT(cq->cq_ring, cq->cq_size);
|
H A D | puda.c | 736 set_64bit_val(wqe, 0, cq->cq_uk.cq_size); 786 cqsize = rsrc->cq_size * (sizeof(struct irdma_cqe)); 798 info.shadow_read_threshold = rsrc->cq_size >> 2; 803 init_info->cq_size = rsrc->cq_size; 1062 rsrc->cq_size = info->rq_size + info->sq_size; 1065 rsrc->cq_size += info->rq_size;
|
H A D | verbs.c | 1921 info.cq_size = max(entries, 4); 1923 if (info.cq_size == iwcq->sc_cq.cq_uk.cq_size - 1) 1960 rsize = info.cq_size * sizeof(struct irdma_cqe); 2010 ibcq->cqe = info.cq_size - 1; 2045 #define IRDMA_CREATE_CQ_MIN_RESP_LEN offsetofend(struct irdma_create_cq_resp, cq_size) 2081 ukinfo->cq_size = max(entries, 4); 2083 iwcq->ibcq.cqe = info.cq_uk_init_info.cq_size; 2160 ukinfo->cq_size = entries; 2162 rsize = info.cq_uk_init_info.cq_size * sizeo [all...] |
/linux-master/drivers/net/ethernet/microsoft/mana/ |
H A D | hw_channel.c | 360 u32 eq_size, cq_size; local 367 cq_size = roundup_pow_of_two(GDMA_CQE_SIZE * q_depth); 368 if (cq_size < MINIMUM_SUPPORTED_PAGE_SIZE) 369 cq_size = MINIMUM_SUPPORTED_PAGE_SIZE; 382 err = mana_hwc_create_gdma_cq(hwc, cq_size, hwc_cq, mana_hwc_comp_event,
|
H A D | mana_en.c | 1140 req.cq_size = cq_spec->queue_size; 1879 u32 cq_size; local 1896 cq_size = MAX_SEND_BUFFERS_PER_QUEUE * COMP_ENTRY_SIZE; 1897 cq_size = PAGE_ALIGN(cq_size); 1930 spec.queue_size = cq_size; 2076 struct mana_rxq *rxq, u32 *rxq_size, u32 *cq_size) 2087 *cq_size = 0; 2109 *cq_size += COMP_ENTRY_SIZE; 2164 u32 cq_size, rq_siz local 2075 mana_alloc_rx_wqe(struct mana_port_context *apc, struct mana_rxq *rxq, u32 *rxq_size, u32 *cq_size) argument [all...] |
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/fpga/ |
H A D | conn.c | 411 static int mlx5_fpga_conn_create_cq(struct mlx5_fpga_conn *conn, int cq_size) argument 424 cq_size = roundup_pow_of_two(cq_size); 425 MLX5_SET(cqc, temp_cqc, log_cq_size, ilog2(cq_size)); 455 MLX5_SET(cqc, cqc, log_cq_size, ilog2(cq_size));
|
/linux-master/drivers/net/ethernet/pensando/ionic/ |
H A D | ionic_debugfs.c | 133 debugfs_create_x32("cq_size", 0400, qcq_dentry, &qcq->cq_size);
|
H A D | ionic_lif.h | 74 u32 cq_size; member in struct:ionic_qcq
|
H A D | ionic_lif.c | 422 dma_free_coherent(dev, qcq->cq_size, qcq->cq_base, qcq->cq_base_pa); 651 new->cq_size = PAGE_SIZE + (num_descs * cq_desc_size); 652 new->cq_base = dma_alloc_coherent(dev, new->cq_size, &new->cq_base_pa, 685 dma_free_coherent(dev, new->cq_size, new->cq_base, new->cq_base_pa); 776 memset(qcq->cq_base, 0, qcq->cq_size); 2889 swap(a->cq_size, b->cq_size);
|
/linux-master/drivers/dma/ |
H A D | hisi_dma.c | 586 size_t cq_size = sizeof(struct hisi_dma_cqe) * hdma_dev->chan_depth; local 598 chan->cq = dmam_alloc_coherent(dev, cq_size, &chan->cq_dma,
|
/linux-master/drivers/nvme/host/ |
H A D | rdma.c | 99 int cq_size; member in struct:nvme_rdma_queue 418 ib_cq_pool_put(queue->ib_cq, queue->cq_size); 475 queue->ib_cq = ib_alloc_cq(ibdev, queue, queue->cq_size, 478 queue->ib_cq = ib_cq_pool_get(ibdev, queue->cq_size, 505 queue->cq_size = cq_factor * queue->queue_size + 1;
|
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/steering/ |
H A D | dr_send.c | 1223 int cq_size; local 1231 cq_size = QUEUE_SIZE + 1; 1232 dmn->send_ring->cq = dr_create_cq(dmn->mdev, dmn->uar, cq_size);
|