/linux-master/drivers/infiniband/hw/bnxt_re/ |
H A D | ib_verbs.c | 1060 if (!qp->qplib_qp.srq) { 1213 if (init_attr->srq) { 1214 struct bnxt_re_srq *srq; local 1216 srq = container_of(init_attr->srq, struct bnxt_re_srq, ib_srq); 1217 qplqp->srq = &srq->qplib_srq; 1680 struct bnxt_re_srq *srq = container_of(ib_srq, struct bnxt_re_srq, local 1682 struct bnxt_re_dev *rdev = srq->rdev; 1683 struct bnxt_qplib_srq *qplib_srq = &srq 1696 bnxt_re_init_user_srq(struct bnxt_re_dev *rdev, struct bnxt_re_pd *pd, struct bnxt_re_srq *srq, struct ib_udata *udata) argument 1736 struct bnxt_re_srq *srq; local 1822 struct bnxt_re_srq *srq = container_of(ib_srq, struct bnxt_re_srq, local 1855 struct bnxt_re_srq *srq = container_of(ib_srq, struct bnxt_re_srq, local 1878 struct bnxt_re_srq *srq = container_of(ib_srq, struct bnxt_re_srq, local [all...] |
H A D | main.c | 978 struct bnxt_re_srq *srq = container_of(qp->qplib_qp.srq, struct bnxt_re_srq, local 1072 if (srq) 1091 if (srq) 1095 if (event.event == IB_EVENT_SRQ_ERR && srq->ib_srq.event_handler) { 1096 (*srq->ib_srq.event_handler)(&event, 1097 srq->ib_srq.srq_context); 1190 struct bnxt_re_srq *srq = container_of(handle, struct bnxt_re_srq, local 1194 ib_event.device = &srq->rdev->ibdev; 1195 ib_event.element.srq [all...] |
/linux-master/drivers/infiniband/sw/rdmavt/ |
H A D | qp.c | 1053 if (!init_attr->srq) { 1083 if (init_attr->srq) { 1084 struct rvt_srq *srq = ibsrq_to_rvtsrq(init_attr->srq); local 1086 if (srq->rq.max_sge > 1) 1088 (srq->rq.max_sge - 1); 1127 if (init_attr->srq) { 1735 attr->cap.max_recv_wr = qp->ibqp.srq ? 0 : qp->r_rq.size - 1; 1760 init_attr->srq = qp->ibqp.srq; 2207 struct rvt_srq *srq = ibsrq_to_rvtsrq(ibsrq); local 2349 struct rvt_srq *srq; local [all...] |
/linux-master/drivers/infiniband/hw/mthca/ |
H A D | mthca_provider.c | 398 struct mthca_srq *srq = to_msrq(ibsrq); local 415 srq->mr.ibmr.lkey = ucmd.lkey; 416 srq->db_index = ucmd.db_index; 420 &init_attr->attr, srq, udata); 429 if (context && ib_copy_to_udata(udata, &srq->srqn, sizeof(__u32))) { 430 mthca_free_srq(to_mdev(ibsrq->device), srq); 437 static int mthca_destroy_srq(struct ib_srq *srq, struct ib_udata *udata) argument 446 mthca_unmap_user_db(to_mdev(srq->device), &context->uar, 447 context->db_tab, to_msrq(srq)->db_index); 450 mthca_free_srq(to_mdev(srq [all...] |
/linux-master/drivers/net/ethernet/mellanox/mlx4/ |
H A D | resource_tracker.c | 113 struct res_srq *srq; member in struct:res_qp 455 dev->quotas.srq = dev->caps.num_srqs - dev->caps.reserved_srqs; 466 dev->quotas.srq = 1712 enum res_srq_states state, struct res_srq **srq) 1738 if (srq) 1739 *srq = r; 2724 int srq = (be32_to_cpu(qpc->srqn) >> 24) & 1; local 2736 rq_size = (srq|rss|xrc) ? 0 : (1 << (log_rq_size + log_rq_stride + 4)); 2971 struct res_srq *srq; local 3011 err = get_res(dev, slave, srqn, RES_SRQ, &srq); 1711 srq_res_start_move_to(struct mlx4_dev *dev, int slave, int index, enum res_srq_states state, struct res_srq **srq) argument 3619 struct res_srq *srq = NULL; local 3663 struct res_srq *srq = NULL; local 3692 struct res_srq *srq; local 3715 struct res_srq *srq; local 4739 struct res_srq *srq; local [all...] |
/linux-master/drivers/infiniband/core/ |
H A D | nldev.c | 411 [RDMA_RESTRACK_SRQ] = "srq", 773 static int fill_res_srq_qps(struct sk_buff *msg, struct ib_srq *srq) argument 786 rt = &srq->device->res[RDMA_RESTRACK_QP]; 793 if (!qp->srq || (qp->srq->res.id != srq->res.id)) { 833 struct ib_srq *srq = container_of(res, struct ib_srq, res); local 834 struct ib_device *dev = srq->device; 836 if (nla_put_u32(msg, RDMA_NLDEV_ATTR_RES_SRQN, srq->res.id)) 839 if (nla_put_u8(msg, RDMA_NLDEV_ATTR_RES_TYPE, srq 869 struct ib_srq *srq = container_of(res, struct ib_srq, res); local [all...] |
/linux-master/drivers/infiniband/hw/ocrdma/ |
H A D | ocrdma_hw.h | 140 void ocrdma_mbx_destroy_srq(struct ocrdma_dev *dev, struct ocrdma_srq *srq);
|
/linux-master/drivers/infiniband/ulp/srpt/ |
H A D | ib_srpt.h | 441 * @srq: Per-HCA SRQ (shared receive queue). 457 struct ib_srq *srq; member in struct:srpt_device
|
H A D | ib_srpt.c | 846 return ib_post_srq_recv(sdev->srq, &wr, NULL); 1834 qp_init->srq = sdev->srq; 2438 rep_param->ib_cm.srq = 1; 3051 if (!sdev->srq) 3054 ib_destroy_srq(sdev->srq); 3059 sdev->srq = NULL; 3072 struct ib_srq *srq; local 3075 WARN_ON_ONCE(sdev->srq); 3076 srq [all...] |
/linux-master/include/rdma/ |
H A D | rdma_cm.h | 77 u8 srq; member in struct:rdma_conn_param
|
H A D | ib_verbs.h | 759 struct ib_srq *srq; member in union:ib_event::__anon163 1179 struct ib_srq *srq; member in struct:ib_qp_init_attr 1790 struct ib_srq *srq; member in struct:ib_qp 2342 int (*post_srq_recv)(struct ib_srq *srq, 2450 int (*create_srq)(struct ib_srq *srq, 2453 int (*modify_srq)(struct ib_srq *srq, struct ib_srq_attr *srq_attr, 2456 int (*query_srq)(struct ib_srq *srq, struct ib_srq_attr *srq_attr); 2457 int (*destroy_srq)(struct ib_srq *srq, struct ib_udata *udata); 3657 * @srq: The SRQ to modify. 3667 int ib_modify_srq(struct ib_srq *srq, 3693 ib_destroy_srq(struct ib_srq *srq) argument 3707 ib_post_srq_recv(struct ib_srq *srq, const struct ib_recv_wr *recv_wr, const struct ib_recv_wr **bad_recv_wr) argument [all...] |
/linux-master/drivers/infiniband/sw/siw/ |
H A D | siw_verbs.h | 87 void siw_srq_event(struct siw_srq *srq, enum ib_event_type type);
|
/linux-master/drivers/infiniband/hw/vmw_pvrdma/ |
H A D | pvrdma_qp.c | 205 bool is_srq = !!init_attr->srq; 278 qp->srq = NULL; 281 qp->srq = to_vsrq(init_attr->srq); 362 cmd->srq_handle = to_vsrq(init_attr->srq)->srq_handle; 893 if (qp->srq) { 1036 init_attr->srq = qp->ibqp.srq;
|
/linux-master/drivers/infiniband/hw/qedr/ |
H A D | main.c | 654 struct qedr_srq *srq; local 759 srq = xa_load(&dev->srqs, srq_id); 760 if (srq) { 761 ibsrq = &srq->ibsrq; 764 event.element.srq = ibsrq; 774 DP_NOTICE(dev, "SRQ event %d on handle %p\n", e_code, srq);
|
H A D | qedr.h | 370 /* lock to protect srq recv post */ 406 struct qedr_srq *srq; member in struct:qedr_qp 616 return qp->srq;
|
/linux-master/drivers/infiniband/hw/mlx4/ |
H A D | mlx4_ib.h | 774 void __mlx4_ib_cq_clean(struct mlx4_ib_cq *cq, u32 qpn, struct mlx4_ib_srq *srq); 775 void mlx4_ib_cq_clean(struct mlx4_ib_cq *cq, u32 qpn, struct mlx4_ib_srq *srq); 787 int mlx4_ib_create_srq(struct ib_srq *srq, struct ib_srq_init_attr *init_attr, 791 int mlx4_ib_query_srq(struct ib_srq *srq, struct ib_srq_attr *srq_attr); 792 int mlx4_ib_destroy_srq(struct ib_srq *srq, struct ib_udata *udata); 793 void mlx4_ib_free_srq_wqe(struct mlx4_ib_srq *srq, int wqe_index);
|
/linux-master/drivers/infiniband/hw/hfi1/ |
H A D | qp.c | 606 struct rvt_srq *srq = qp->ibqp.srq ? local 607 ibsrq_to_rvtsrq(qp->ibqp.srq) : NULL; 663 srq ? "SRQ" : "RQ", 664 srq ? srq->rq.size : qp->r_rq.size
|
/linux-master/drivers/infiniband/hw/mlx5/ |
H A D | odp.c | 1114 struct mlx5_ib_srq *srq, 1118 int wqe_size = 1 << srq->msrq.wqe_shift; 1158 struct mlx5_core_srq *srq; local 1162 srq = mlx5_cmd_get_srq(dev, wq_num); 1163 if (srq) 1164 common = &srq->common; 1242 struct mlx5_ib_srq *srq = res_to_srq(res); local 1244 ret = mlx5_ib_read_wqe_srq(srq, wqe_index, wqe, PAGE_SIZE, 1249 dev, srq, &wqe, &wqe_end, bytes_copied); 1113 mlx5_ib_mr_responder_pfault_handler_srq(struct mlx5_ib_dev *dev, struct mlx5_ib_srq *srq, void **wqe, void **wqe_end, int wqe_length) argument
|
H A D | mlx5_ib.h | 27 #include "srq.h" 1278 void __mlx5_ib_cq_clean(struct mlx5_ib_cq *cq, u32 qpn, struct mlx5_ib_srq *srq); 1279 void mlx5_ib_cq_clean(struct mlx5_ib_cq *cq, u32 qpn, struct mlx5_ib_srq *srq); 1280 void mlx5_ib_free_srq_wqe(struct mlx5_ib_srq *srq, int wqe_index); 1288 int mlx5_ib_create_srq(struct ib_srq *srq, struct ib_srq_init_attr *init_attr, 1293 int mlx5_ib_destroy_srq(struct ib_srq *srq, struct ib_udata *udata); 1311 int mlx5_ib_read_wqe_srq(struct mlx5_ib_srq *srq, int wqe_index, void *buffer,
|
H A D | qp.c | 285 static int mlx5_ib_read_user_wqe_srq(struct mlx5_ib_srq *srq, int wqe_index, argument 288 struct ib_umem *umem = srq->umem; 293 srq->msrq.max, srq->msrq.wqe_shift, 302 int mlx5_ib_read_wqe_srq(struct mlx5_ib_srq *srq, int wqe_index, void *buffer, argument 305 struct ib_umem *umem = srq->umem; 306 size_t wqe_size = 1 << srq->msrq.wqe_shift; 314 return mlx5_ib_read_user_wqe_srq(srq, wqe_index, buffer, buflen, bc); 680 attr->qp_type == IB_QPT_XRC_TGT || attr->srq || 1221 if (attr->srq || (q [all...] |
/linux-master/drivers/infiniband/hw/cxgb4/ |
H A D | iw_cxgb4.h | 135 struct c4iw_stat srq; member in struct:c4iw_stats 489 struct c4iw_srq *srq; member in struct:c4iw_qp 511 spinlock_t lock; /* protects srq */ 987 int c4iw_create_srq(struct ib_srq *srq, struct ib_srq_init_attr *attrs, 1032 void c4iw_dispatch_srq_limit_reached_event(struct c4iw_srq *srq); 1033 void c4iw_copy_wr_to_srq(struct t4_srq *srq, union t4_recv_wr *wqe, u8 len16);
|
H A D | device.c | 271 qp->wq.sq.qid, qp->srq ? "srq" : "rq", 272 qp->srq ? qp->srq->idx : qp->wq.rq.qid, 823 pr_debug("dev %s stag start 0x%0x size 0x%0x num stags %d pbl start 0x%0x size 0x%0x rq start 0x%0x size 0x%0x qp qid start %u size %u cq qid start %u size %u srq size %u\n", 833 rdev->lldi.vr->srq.size); 846 rdev->stats.srqt.total = rdev->lldi.vr->srq.size; 851 T4_MAX_NUM_PD, rdev->lldi.vr->srq.size);
|
/linux-master/drivers/nvme/target/ |
H A D | rdma.c | 127 struct ib_srq *srq; member in struct:nvmet_rdma_srq 504 ret = ib_post_srq_recv(cmd->nsrq->srq, &cmd->wr, NULL); 1060 ib_destroy_srq(nsrq->srq); 1084 struct ib_srq *srq; local 1095 srq = ib_create_srq(ndev->pd, &srq_attr); 1096 if (IS_ERR(srq)) { 1097 ret = PTR_ERR(srq); 1107 nsrq->srq = srq; 1122 ib_destroy_srq(srq); [all...] |
/linux-master/include/linux/mlx4/ |
H A D | device.h | 833 int srq; member in struct:mlx4_quotas 928 } __packed srq; member in union:mlx4_eqe::__anon2624 1159 struct mlx4_mtt *mtt, u64 db_rec, struct mlx4_srq *srq); 1160 void mlx4_srq_free(struct mlx4_dev *dev, struct mlx4_srq *srq); 1161 int mlx4_srq_arm(struct mlx4_dev *dev, struct mlx4_srq *srq, int limit_watermark); 1162 int mlx4_srq_query(struct mlx4_dev *dev, struct mlx4_srq *srq, int *limit_watermark);
|
/linux-master/include/uapi/rdma/ |
H A D | rdma_user_cm.h | 195 __u8 srq; member in struct:rdma_ucm_conn_param
|