Searched refs:rq_size (Results 1 - 25 of 34) sorted by relevance

12

/linux-master/include/uapi/rdma/
H A Dcxgb4-abi.h83 __u32 rq_size; member in struct:c4iw_create_qp_resp
/linux-master/drivers/infiniband/hw/irdma/
H A Dpuda.c127 rsrc->rx_wqe_idx = ((rsrc->rx_wqe_idx + 1) % rsrc->rq_size);
649 u32 sq_size, rq_size; local
653 rq_size = rsrc->rq_size * IRDMA_QP_WQE_MIN_SIZE;
654 rsrc->qpmem.size = ALIGN((sq_size + rq_size + (IRDMA_SHADOW_AREA_SIZE << 3) + IRDMA_QP_CTX_SIZE),
665 qp->hw_rq_size = irdma_get_encoded_wqe_size(rsrc->rq_size, IRDMA_QUEUE_TYPE_SQ_RQ);
675 ukqp->shadow_area = ukqp->rq_base[rsrc->rq_size].elem;
677 qp->shadow_area_pa = qp->rq_pa + rq_size;
685 ukqp->rq_size = rsrc->rq_size;
[all...]
H A Duser.h328 u32 rq_size; member in struct:irdma_qp_uk
374 u32 rq_size; member in struct:irdma_qp_uk_init_info
410 int irdma_get_rqdepth(struct irdma_uk_attrs *uk_attrs, u32 rq_size, u8 shift,
H A Dpuda.h91 u32 rq_size; member in struct:irdma_puda_rsrc_info
92 u32 tx_buf_cnt; /* total bufs allocated will be rq_size + tx_buf_cnt */
114 u32 rq_size; member in struct:irdma_puda_rsrc
H A Duk.c1284 * @rq_size: RQ size
1288 int irdma_get_rqdepth(struct irdma_uk_attrs *uk_attrs, u32 rq_size, u8 shift, argument
1293 *rqdepth = irdma_qp_round_up((rq_size << shift) + IRDMA_RQ_RSVD);
1404 status = irdma_get_rqdepth(ukinfo->uk_attrs, ukinfo->rq_size,
1454 qp->rq_size = info->rq_size;
1458 IRDMA_RING_INIT(qp->rq_ring, qp->rq_size);
/linux-master/drivers/net/ethernet/qlogic/netxen/
H A Dnetxen_nic_ctx.c270 size_t rq_size, rsp_size; local
280 rq_size =
285 addr = dma_alloc_coherent(&adapter->pdev->dev, rq_size,
350 cmd.req.arg3 = rq_size;
394 dma_free_coherent(&adapter->pdev->dev, rq_size, prq, hostrq_phys_addr);
424 size_t rq_size, rsp_size; local
433 rq_size = SIZEOF_HOSTRQ_TX(nx_hostrq_tx_ctx_t);
434 rq_addr = dma_alloc_coherent(&adapter->pdev->dev, rq_size,
476 cmd.req.arg3 = rq_size;
500 dma_free_coherent(&adapter->pdev->dev, rq_size, rq_add
[all...]
/linux-master/drivers/scsi/bnx2i/
H A Dbnx2i_init.c65 unsigned int rq_size = BNX2I_RQ_WQES_DEFAULT; variable
66 module_param(rq_size, int, 0664);
67 MODULE_PARM_DESC(rq_size, "Configure RQ size");
H A Dbnx2i.h796 extern unsigned int rq_size;
H A Dbnx2i_iscsi.c855 hba->max_rqes = rq_size;
856 hba->max_cqes = hba->max_sqes + rq_size;
/linux-master/drivers/scsi/
H A Dstex.c347 u16 rq_size; member in struct:st_hba
361 u16 rq_size; member in struct:st_card_info
412 struct req_msg *req = hba->dma_mem + hba->req_head * hba->rq_size;
423 hba->req_head * hba->rq_size + sizeof(struct st_msg_header));
549 addr = hba->dma_handle + hba->req_head * hba->rq_size;
1063 h->req_sz = cpu_to_le16(hba->rq_size);
1075 status_phys = hba->dma_handle + (hba->rq_count+1) * hba->rq_size;
1154 h->req_sz = cpu_to_le16(hba->rq_size);
1545 .rq_size = 1048,
1558 .rq_size
[all...]
/linux-master/drivers/infiniband/ulp/srpt/
H A Dib_srpt.h265 * @rq_size: IB receive queue size.
309 int rq_size; member in struct:srpt_rdma_ch
H A Dib_srpt.c1797 WARN_ON(ch->rq_size < 1);
1805 ch->cq = ib_cq_pool_get(sdev->device, ch->rq_size + sq_size, -1,
1810 ch->rq_size + sq_size, ret);
1813 ch->cq_size = ch->rq_size + sq_size;
1836 qp_init->cap.max_recv_wr = ch->rq_size;
1874 for (i = 0; i < ch->rq_size; i++)
2119 ch->sport->sdev, ch->rq_size,
2125 sdev, ch->rq_size,
2238 * ch->rq_size should be at least as large as the initiator queue
2242 ch->rq_size
[all...]
/linux-master/drivers/net/ethernet/qlogic/qlcnic/
H A Dqlcnic_ctx.c250 size_t rq_size, rsp_size; local
260 rq_size = SIZEOF_HOSTRQ_RX(struct qlcnic_hostrq_rx_ctx, nrds_rings,
265 addr = dma_alloc_coherent(&adapter->pdev->dev, rq_size,
343 cmd.req.arg[3] = rq_size;
387 dma_free_coherent(&adapter->pdev->dev, rq_size, prq, hostrq_phys_addr);
426 size_t rq_size, rsp_size; local
435 rq_size = SIZEOF_HOSTRQ_TX(struct qlcnic_hostrq_tx_ctx);
436 rq_addr = dma_alloc_coherent(&adapter->pdev->dev, rq_size,
489 cmd.req.arg[3] = rq_size;
518 dma_free_coherent(&adapter->pdev->dev, rq_size, rq_add
[all...]
/linux-master/drivers/infiniband/hw/erdma/
H A Derdma_verbs.c55 ilog2(qp->attrs.rq_size)) |
82 (qp->attrs.rq_size << RQE_SHIFT);
493 WARPPED_BUFSIZE(qp->attrs.rq_size << RQE_SHIFT),
515 kqp->rwr_tbl = vmalloc_array(qp->attrs.rq_size, sizeof(u64));
525 size = (qp->attrs.rq_size << RQE_SHIFT) + ERDMA_EXTRA_BUFFER_SIZE;
532 kqp->rq_db_info = kqp->rq_buf + (qp->attrs.rq_size << RQE_SHIFT);
874 qp->attrs.rq_size * RQE_SIZE))
887 qp->attrs.rq_size << RQE_SHIFT, 0, va + rq_offset,
957 qp->attrs.rq_size = roundup_pow_of_two(attrs->cap.max_recv_wr);
973 uresp.num_rqe = qp->attrs.rq_size;
[all...]
H A Derdma_cq.c155 depth = qp->attrs.rq_size;
H A Derdma_qp.c545 qp->attrs.rq_size, RQE_SHIFT);
563 qp->kern_qp.rwr_tbl[qp->kern_qp.rq_pi & (qp->attrs.rq_size - 1)] =
H A Derdma_verbs.h205 u32 rq_size; member in struct:erdma_qp_attrs
/linux-master/drivers/infiniband/hw/efa/
H A Defa_verbs.c468 qp->rq_cpu_addr, qp->rq_size,
471 qp->rq_size, DMA_TO_DEVICE);
536 if (qp->rq_size) {
552 address, qp->rq_size,
558 resp->rq_mmap_size = qp->rq_size;
710 qp->rq_size = PAGE_ALIGN(create_qp_params.rq_ring_size_in_bytes);
711 if (qp->rq_size) {
713 qp->rq_size, DMA_TO_DEVICE);
721 qp->rq_cpu_addr, qp->rq_size, &qp->rq_dma_addr);
769 if (qp->rq_size)
[all...]
H A Defa.h116 size_t rq_size; member in struct:efa_qp
/linux-master/drivers/infiniband/sw/siw/
H A Dsiw_verbs.c407 qp->attrs.rq_size = 0;
421 qp->attrs.rq_size = num_rqe;
520 qp_attr->cap.max_recv_wr = qp->attrs.rq_size;
1003 if (qp->srq || qp->attrs.rq_size == 0) {
1063 u32 idx = qp->rq_put % qp->attrs.rq_size;
H A Dsiw.h261 u32 rq_size; member in struct:siw_qp_attrs
H A Dsiw_qp.c1299 while (qp->attrs.rq_size) {
1301 &qp->recvq[qp->rq_get % qp->attrs.rq_size];
/linux-master/drivers/vfio/pci/mlx5/
H A Dcmd.c1438 u64 rq_size)
1440 unsigned int npages = DIV_ROUND_UP_ULL(rq_size, PAGE_SIZE);
1517 u64 rq_size = SZ_2M; local
1544 max_recv_wr = DIV_ROUND_UP_ULL(rq_size, max_msg_size);
1568 rq_size);
1436 mlx5vf_alloc_qp_recv_resources(struct mlx5_core_dev *mdev, struct mlx5_vhca_qp *qp, u32 pdn, u64 rq_size) argument
/linux-master/block/
H A Dblk-mq.c3388 size_t rq_size, left; local
3396 * rq_size is the size of the request plus driver payload, rounded
3399 rq_size = round_up(sizeof(struct request) + set->cmd_size,
3401 left = rq_size * depth;
3420 if (order_to_size(this_order) < rq_size)
3436 entries_per_page = order_to_size(this_order) / rq_size;
3438 left -= to_do * rq_size;
3448 p += rq_size;
/linux-master/drivers/net/ethernet/microsoft/mana/
H A Dmana_en.c2164 u32 cq_size, rq_size; local
2190 err = mana_alloc_rx_wqe(apc, rxq, &rq_size, &cq_size);
2194 rq_size = PAGE_ALIGN(rq_size);
2201 spec.queue_size = rq_size;

Completed in 428 milliseconds

12