/linux-master/include/uapi/rdma/ |
H A D | cxgb4-abi.h | 83 __u32 rq_size; member in struct:c4iw_create_qp_resp
|
/linux-master/drivers/infiniband/hw/irdma/ |
H A D | puda.c | 127 rsrc->rx_wqe_idx = ((rsrc->rx_wqe_idx + 1) % rsrc->rq_size); 649 u32 sq_size, rq_size; local 653 rq_size = rsrc->rq_size * IRDMA_QP_WQE_MIN_SIZE; 654 rsrc->qpmem.size = ALIGN((sq_size + rq_size + (IRDMA_SHADOW_AREA_SIZE << 3) + IRDMA_QP_CTX_SIZE), 665 qp->hw_rq_size = irdma_get_encoded_wqe_size(rsrc->rq_size, IRDMA_QUEUE_TYPE_SQ_RQ); 675 ukqp->shadow_area = ukqp->rq_base[rsrc->rq_size].elem; 677 qp->shadow_area_pa = qp->rq_pa + rq_size; 685 ukqp->rq_size = rsrc->rq_size; [all...] |
H A D | user.h | 328 u32 rq_size; member in struct:irdma_qp_uk 374 u32 rq_size; member in struct:irdma_qp_uk_init_info 410 int irdma_get_rqdepth(struct irdma_uk_attrs *uk_attrs, u32 rq_size, u8 shift,
|
H A D | puda.h | 91 u32 rq_size; member in struct:irdma_puda_rsrc_info 92 u32 tx_buf_cnt; /* total bufs allocated will be rq_size + tx_buf_cnt */ 114 u32 rq_size; member in struct:irdma_puda_rsrc
|
H A D | uk.c | 1284 * @rq_size: RQ size 1288 int irdma_get_rqdepth(struct irdma_uk_attrs *uk_attrs, u32 rq_size, u8 shift, argument 1293 *rqdepth = irdma_qp_round_up((rq_size << shift) + IRDMA_RQ_RSVD); 1404 status = irdma_get_rqdepth(ukinfo->uk_attrs, ukinfo->rq_size, 1454 qp->rq_size = info->rq_size; 1458 IRDMA_RING_INIT(qp->rq_ring, qp->rq_size);
|
/linux-master/drivers/net/ethernet/qlogic/netxen/ |
H A D | netxen_nic_ctx.c | 270 size_t rq_size, rsp_size; local 280 rq_size = 285 addr = dma_alloc_coherent(&adapter->pdev->dev, rq_size, 350 cmd.req.arg3 = rq_size; 394 dma_free_coherent(&adapter->pdev->dev, rq_size, prq, hostrq_phys_addr); 424 size_t rq_size, rsp_size; local 433 rq_size = SIZEOF_HOSTRQ_TX(nx_hostrq_tx_ctx_t); 434 rq_addr = dma_alloc_coherent(&adapter->pdev->dev, rq_size, 476 cmd.req.arg3 = rq_size; 500 dma_free_coherent(&adapter->pdev->dev, rq_size, rq_add [all...] |
/linux-master/drivers/scsi/bnx2i/ |
H A D | bnx2i_init.c | 65 unsigned int rq_size = BNX2I_RQ_WQES_DEFAULT; variable 66 module_param(rq_size, int, 0664); 67 MODULE_PARM_DESC(rq_size, "Configure RQ size");
|
H A D | bnx2i.h | 796 extern unsigned int rq_size;
|
H A D | bnx2i_iscsi.c | 855 hba->max_rqes = rq_size; 856 hba->max_cqes = hba->max_sqes + rq_size;
|
/linux-master/drivers/scsi/ |
H A D | stex.c | 347 u16 rq_size; member in struct:st_hba 361 u16 rq_size; member in struct:st_card_info 412 struct req_msg *req = hba->dma_mem + hba->req_head * hba->rq_size; 423 hba->req_head * hba->rq_size + sizeof(struct st_msg_header)); 549 addr = hba->dma_handle + hba->req_head * hba->rq_size; 1063 h->req_sz = cpu_to_le16(hba->rq_size); 1075 status_phys = hba->dma_handle + (hba->rq_count+1) * hba->rq_size; 1154 h->req_sz = cpu_to_le16(hba->rq_size); 1545 .rq_size = 1048, 1558 .rq_size [all...] |
/linux-master/drivers/infiniband/ulp/srpt/ |
H A D | ib_srpt.h | 265 * @rq_size: IB receive queue size. 309 int rq_size; member in struct:srpt_rdma_ch
|
H A D | ib_srpt.c | 1797 WARN_ON(ch->rq_size < 1); 1805 ch->cq = ib_cq_pool_get(sdev->device, ch->rq_size + sq_size, -1, 1810 ch->rq_size + sq_size, ret); 1813 ch->cq_size = ch->rq_size + sq_size; 1836 qp_init->cap.max_recv_wr = ch->rq_size; 1874 for (i = 0; i < ch->rq_size; i++) 2119 ch->sport->sdev, ch->rq_size, 2125 sdev, ch->rq_size, 2238 * ch->rq_size should be at least as large as the initiator queue 2242 ch->rq_size [all...] |
/linux-master/drivers/net/ethernet/qlogic/qlcnic/ |
H A D | qlcnic_ctx.c | 250 size_t rq_size, rsp_size; local 260 rq_size = SIZEOF_HOSTRQ_RX(struct qlcnic_hostrq_rx_ctx, nrds_rings, 265 addr = dma_alloc_coherent(&adapter->pdev->dev, rq_size, 343 cmd.req.arg[3] = rq_size; 387 dma_free_coherent(&adapter->pdev->dev, rq_size, prq, hostrq_phys_addr); 426 size_t rq_size, rsp_size; local 435 rq_size = SIZEOF_HOSTRQ_TX(struct qlcnic_hostrq_tx_ctx); 436 rq_addr = dma_alloc_coherent(&adapter->pdev->dev, rq_size, 489 cmd.req.arg[3] = rq_size; 518 dma_free_coherent(&adapter->pdev->dev, rq_size, rq_add [all...] |
/linux-master/drivers/infiniband/hw/erdma/ |
H A D | erdma_verbs.c | 55 ilog2(qp->attrs.rq_size)) | 82 (qp->attrs.rq_size << RQE_SHIFT); 493 WARPPED_BUFSIZE(qp->attrs.rq_size << RQE_SHIFT), 515 kqp->rwr_tbl = vmalloc_array(qp->attrs.rq_size, sizeof(u64)); 525 size = (qp->attrs.rq_size << RQE_SHIFT) + ERDMA_EXTRA_BUFFER_SIZE; 532 kqp->rq_db_info = kqp->rq_buf + (qp->attrs.rq_size << RQE_SHIFT); 874 qp->attrs.rq_size * RQE_SIZE)) 887 qp->attrs.rq_size << RQE_SHIFT, 0, va + rq_offset, 957 qp->attrs.rq_size = roundup_pow_of_two(attrs->cap.max_recv_wr); 973 uresp.num_rqe = qp->attrs.rq_size; [all...] |
H A D | erdma_cq.c | 155 depth = qp->attrs.rq_size;
|
H A D | erdma_qp.c | 545 qp->attrs.rq_size, RQE_SHIFT); 563 qp->kern_qp.rwr_tbl[qp->kern_qp.rq_pi & (qp->attrs.rq_size - 1)] =
|
H A D | erdma_verbs.h | 205 u32 rq_size; member in struct:erdma_qp_attrs
|
/linux-master/drivers/infiniband/hw/efa/ |
H A D | efa_verbs.c | 468 qp->rq_cpu_addr, qp->rq_size, 471 qp->rq_size, DMA_TO_DEVICE); 536 if (qp->rq_size) { 552 address, qp->rq_size, 558 resp->rq_mmap_size = qp->rq_size; 710 qp->rq_size = PAGE_ALIGN(create_qp_params.rq_ring_size_in_bytes); 711 if (qp->rq_size) { 713 qp->rq_size, DMA_TO_DEVICE); 721 qp->rq_cpu_addr, qp->rq_size, &qp->rq_dma_addr); 769 if (qp->rq_size) [all...] |
H A D | efa.h | 116 size_t rq_size; member in struct:efa_qp
|
/linux-master/drivers/infiniband/sw/siw/ |
H A D | siw_verbs.c | 407 qp->attrs.rq_size = 0; 421 qp->attrs.rq_size = num_rqe; 520 qp_attr->cap.max_recv_wr = qp->attrs.rq_size; 1003 if (qp->srq || qp->attrs.rq_size == 0) { 1063 u32 idx = qp->rq_put % qp->attrs.rq_size;
|
H A D | siw.h | 261 u32 rq_size; member in struct:siw_qp_attrs
|
H A D | siw_qp.c | 1299 while (qp->attrs.rq_size) { 1301 &qp->recvq[qp->rq_get % qp->attrs.rq_size];
|
/linux-master/drivers/vfio/pci/mlx5/ |
H A D | cmd.c | 1438 u64 rq_size) 1440 unsigned int npages = DIV_ROUND_UP_ULL(rq_size, PAGE_SIZE); 1517 u64 rq_size = SZ_2M; local 1544 max_recv_wr = DIV_ROUND_UP_ULL(rq_size, max_msg_size); 1568 rq_size); 1436 mlx5vf_alloc_qp_recv_resources(struct mlx5_core_dev *mdev, struct mlx5_vhca_qp *qp, u32 pdn, u64 rq_size) argument
|
/linux-master/block/ |
H A D | blk-mq.c | 3388 size_t rq_size, left; local 3396 * rq_size is the size of the request plus driver payload, rounded 3399 rq_size = round_up(sizeof(struct request) + set->cmd_size, 3401 left = rq_size * depth; 3420 if (order_to_size(this_order) < rq_size) 3436 entries_per_page = order_to_size(this_order) / rq_size; 3438 left -= to_do * rq_size; 3448 p += rq_size;
|
/linux-master/drivers/net/ethernet/microsoft/mana/ |
H A D | mana_en.c | 2164 u32 cq_size, rq_size; local 2190 err = mana_alloc_rx_wqe(apc, rxq, &rq_size, &cq_size); 2194 rq_size = PAGE_ALIGN(rq_size); 2201 spec.queue_size = rq_size;
|