Searched refs:srq (Results 51 - 75 of 106) sorted by relevance

12345

/linux-master/drivers/infiniband/hw/bnxt_re/
H A Dib_verbs.c1060 if (!qp->qplib_qp.srq) {
1213 if (init_attr->srq) {
1214 struct bnxt_re_srq *srq; local
1216 srq = container_of(init_attr->srq, struct bnxt_re_srq, ib_srq);
1217 qplqp->srq = &srq->qplib_srq;
1680 struct bnxt_re_srq *srq = container_of(ib_srq, struct bnxt_re_srq, local
1682 struct bnxt_re_dev *rdev = srq->rdev;
1683 struct bnxt_qplib_srq *qplib_srq = &srq
1696 bnxt_re_init_user_srq(struct bnxt_re_dev *rdev, struct bnxt_re_pd *pd, struct bnxt_re_srq *srq, struct ib_udata *udata) argument
1736 struct bnxt_re_srq *srq; local
1822 struct bnxt_re_srq *srq = container_of(ib_srq, struct bnxt_re_srq, local
1855 struct bnxt_re_srq *srq = container_of(ib_srq, struct bnxt_re_srq, local
1878 struct bnxt_re_srq *srq = container_of(ib_srq, struct bnxt_re_srq, local
[all...]
H A Dmain.c978 struct bnxt_re_srq *srq = container_of(qp->qplib_qp.srq, struct bnxt_re_srq, local
1072 if (srq)
1091 if (srq)
1095 if (event.event == IB_EVENT_SRQ_ERR && srq->ib_srq.event_handler) {
1096 (*srq->ib_srq.event_handler)(&event,
1097 srq->ib_srq.srq_context);
1190 struct bnxt_re_srq *srq = container_of(handle, struct bnxt_re_srq, local
1194 ib_event.device = &srq->rdev->ibdev;
1195 ib_event.element.srq
[all...]
/linux-master/drivers/infiniband/sw/rdmavt/
H A Dqp.c1053 if (!init_attr->srq) {
1083 if (init_attr->srq) {
1084 struct rvt_srq *srq = ibsrq_to_rvtsrq(init_attr->srq); local
1086 if (srq->rq.max_sge > 1)
1088 (srq->rq.max_sge - 1);
1127 if (init_attr->srq) {
1735 attr->cap.max_recv_wr = qp->ibqp.srq ? 0 : qp->r_rq.size - 1;
1760 init_attr->srq = qp->ibqp.srq;
2207 struct rvt_srq *srq = ibsrq_to_rvtsrq(ibsrq); local
2349 struct rvt_srq *srq; local
[all...]
/linux-master/drivers/infiniband/hw/mthca/
H A Dmthca_provider.c398 struct mthca_srq *srq = to_msrq(ibsrq); local
415 srq->mr.ibmr.lkey = ucmd.lkey;
416 srq->db_index = ucmd.db_index;
420 &init_attr->attr, srq, udata);
429 if (context && ib_copy_to_udata(udata, &srq->srqn, sizeof(__u32))) {
430 mthca_free_srq(to_mdev(ibsrq->device), srq);
437 static int mthca_destroy_srq(struct ib_srq *srq, struct ib_udata *udata) argument
446 mthca_unmap_user_db(to_mdev(srq->device), &context->uar,
447 context->db_tab, to_msrq(srq)->db_index);
450 mthca_free_srq(to_mdev(srq
[all...]
/linux-master/drivers/net/ethernet/mellanox/mlx4/
H A Dresource_tracker.c113 struct res_srq *srq; member in struct:res_qp
455 dev->quotas.srq = dev->caps.num_srqs - dev->caps.reserved_srqs;
466 dev->quotas.srq =
1712 enum res_srq_states state, struct res_srq **srq)
1738 if (srq)
1739 *srq = r;
2724 int srq = (be32_to_cpu(qpc->srqn) >> 24) & 1; local
2736 rq_size = (srq|rss|xrc) ? 0 : (1 << (log_rq_size + log_rq_stride + 4));
2971 struct res_srq *srq; local
3011 err = get_res(dev, slave, srqn, RES_SRQ, &srq);
1711 srq_res_start_move_to(struct mlx4_dev *dev, int slave, int index, enum res_srq_states state, struct res_srq **srq) argument
3619 struct res_srq *srq = NULL; local
3663 struct res_srq *srq = NULL; local
3692 struct res_srq *srq; local
3715 struct res_srq *srq; local
4739 struct res_srq *srq; local
[all...]
/linux-master/drivers/infiniband/core/
H A Dnldev.c411 [RDMA_RESTRACK_SRQ] = "srq",
773 static int fill_res_srq_qps(struct sk_buff *msg, struct ib_srq *srq) argument
786 rt = &srq->device->res[RDMA_RESTRACK_QP];
793 if (!qp->srq || (qp->srq->res.id != srq->res.id)) {
833 struct ib_srq *srq = container_of(res, struct ib_srq, res); local
834 struct ib_device *dev = srq->device;
836 if (nla_put_u32(msg, RDMA_NLDEV_ATTR_RES_SRQN, srq->res.id))
839 if (nla_put_u8(msg, RDMA_NLDEV_ATTR_RES_TYPE, srq
869 struct ib_srq *srq = container_of(res, struct ib_srq, res); local
[all...]
/linux-master/drivers/infiniband/hw/ocrdma/
H A Docrdma_hw.h140 void ocrdma_mbx_destroy_srq(struct ocrdma_dev *dev, struct ocrdma_srq *srq);
/linux-master/drivers/infiniband/ulp/srpt/
H A Dib_srpt.h441 * @srq: Per-HCA SRQ (shared receive queue).
457 struct ib_srq *srq; member in struct:srpt_device
H A Dib_srpt.c846 return ib_post_srq_recv(sdev->srq, &wr, NULL);
1834 qp_init->srq = sdev->srq;
2438 rep_param->ib_cm.srq = 1;
3051 if (!sdev->srq)
3054 ib_destroy_srq(sdev->srq);
3059 sdev->srq = NULL;
3072 struct ib_srq *srq; local
3075 WARN_ON_ONCE(sdev->srq);
3076 srq
[all...]
/linux-master/include/rdma/
H A Drdma_cm.h77 u8 srq; member in struct:rdma_conn_param
H A Dib_verbs.h759 struct ib_srq *srq; member in union:ib_event::__anon163
1179 struct ib_srq *srq; member in struct:ib_qp_init_attr
1790 struct ib_srq *srq; member in struct:ib_qp
2342 int (*post_srq_recv)(struct ib_srq *srq,
2450 int (*create_srq)(struct ib_srq *srq,
2453 int (*modify_srq)(struct ib_srq *srq, struct ib_srq_attr *srq_attr,
2456 int (*query_srq)(struct ib_srq *srq, struct ib_srq_attr *srq_attr);
2457 int (*destroy_srq)(struct ib_srq *srq, struct ib_udata *udata);
3657 * @srq: The SRQ to modify.
3667 int ib_modify_srq(struct ib_srq *srq,
3693 ib_destroy_srq(struct ib_srq *srq) argument
3707 ib_post_srq_recv(struct ib_srq *srq, const struct ib_recv_wr *recv_wr, const struct ib_recv_wr **bad_recv_wr) argument
[all...]
/linux-master/drivers/infiniband/sw/siw/
H A Dsiw_verbs.h87 void siw_srq_event(struct siw_srq *srq, enum ib_event_type type);
/linux-master/drivers/infiniband/hw/vmw_pvrdma/
H A Dpvrdma_qp.c205 bool is_srq = !!init_attr->srq;
278 qp->srq = NULL;
281 qp->srq = to_vsrq(init_attr->srq);
362 cmd->srq_handle = to_vsrq(init_attr->srq)->srq_handle;
893 if (qp->srq) {
1036 init_attr->srq = qp->ibqp.srq;
/linux-master/drivers/infiniband/hw/qedr/
H A Dmain.c654 struct qedr_srq *srq; local
759 srq = xa_load(&dev->srqs, srq_id);
760 if (srq) {
761 ibsrq = &srq->ibsrq;
764 event.element.srq = ibsrq;
774 DP_NOTICE(dev, "SRQ event %d on handle %p\n", e_code, srq);
H A Dqedr.h370 /* lock to protect srq recv post */
406 struct qedr_srq *srq; member in struct:qedr_qp
616 return qp->srq;
/linux-master/drivers/infiniband/hw/mlx4/
H A Dmlx4_ib.h774 void __mlx4_ib_cq_clean(struct mlx4_ib_cq *cq, u32 qpn, struct mlx4_ib_srq *srq);
775 void mlx4_ib_cq_clean(struct mlx4_ib_cq *cq, u32 qpn, struct mlx4_ib_srq *srq);
787 int mlx4_ib_create_srq(struct ib_srq *srq, struct ib_srq_init_attr *init_attr,
791 int mlx4_ib_query_srq(struct ib_srq *srq, struct ib_srq_attr *srq_attr);
792 int mlx4_ib_destroy_srq(struct ib_srq *srq, struct ib_udata *udata);
793 void mlx4_ib_free_srq_wqe(struct mlx4_ib_srq *srq, int wqe_index);
/linux-master/drivers/infiniband/hw/hfi1/
H A Dqp.c606 struct rvt_srq *srq = qp->ibqp.srq ? local
607 ibsrq_to_rvtsrq(qp->ibqp.srq) : NULL;
663 srq ? "SRQ" : "RQ",
664 srq ? srq->rq.size : qp->r_rq.size
/linux-master/drivers/infiniband/hw/mlx5/
H A Dodp.c1114 struct mlx5_ib_srq *srq,
1118 int wqe_size = 1 << srq->msrq.wqe_shift;
1158 struct mlx5_core_srq *srq; local
1162 srq = mlx5_cmd_get_srq(dev, wq_num);
1163 if (srq)
1164 common = &srq->common;
1242 struct mlx5_ib_srq *srq = res_to_srq(res); local
1244 ret = mlx5_ib_read_wqe_srq(srq, wqe_index, wqe, PAGE_SIZE,
1249 dev, srq, &wqe, &wqe_end, bytes_copied);
1113 mlx5_ib_mr_responder_pfault_handler_srq(struct mlx5_ib_dev *dev, struct mlx5_ib_srq *srq, void **wqe, void **wqe_end, int wqe_length) argument
H A Dmlx5_ib.h27 #include "srq.h"
1278 void __mlx5_ib_cq_clean(struct mlx5_ib_cq *cq, u32 qpn, struct mlx5_ib_srq *srq);
1279 void mlx5_ib_cq_clean(struct mlx5_ib_cq *cq, u32 qpn, struct mlx5_ib_srq *srq);
1280 void mlx5_ib_free_srq_wqe(struct mlx5_ib_srq *srq, int wqe_index);
1288 int mlx5_ib_create_srq(struct ib_srq *srq, struct ib_srq_init_attr *init_attr,
1293 int mlx5_ib_destroy_srq(struct ib_srq *srq, struct ib_udata *udata);
1311 int mlx5_ib_read_wqe_srq(struct mlx5_ib_srq *srq, int wqe_index, void *buffer,
H A Dqp.c285 static int mlx5_ib_read_user_wqe_srq(struct mlx5_ib_srq *srq, int wqe_index, argument
288 struct ib_umem *umem = srq->umem;
293 srq->msrq.max, srq->msrq.wqe_shift,
302 int mlx5_ib_read_wqe_srq(struct mlx5_ib_srq *srq, int wqe_index, void *buffer, argument
305 struct ib_umem *umem = srq->umem;
306 size_t wqe_size = 1 << srq->msrq.wqe_shift;
314 return mlx5_ib_read_user_wqe_srq(srq, wqe_index, buffer, buflen, bc);
680 attr->qp_type == IB_QPT_XRC_TGT || attr->srq ||
1221 if (attr->srq || (q
[all...]
/linux-master/drivers/infiniband/hw/cxgb4/
H A Diw_cxgb4.h135 struct c4iw_stat srq; member in struct:c4iw_stats
489 struct c4iw_srq *srq; member in struct:c4iw_qp
511 spinlock_t lock; /* protects srq */
987 int c4iw_create_srq(struct ib_srq *srq, struct ib_srq_init_attr *attrs,
1032 void c4iw_dispatch_srq_limit_reached_event(struct c4iw_srq *srq);
1033 void c4iw_copy_wr_to_srq(struct t4_srq *srq, union t4_recv_wr *wqe, u8 len16);
H A Ddevice.c271 qp->wq.sq.qid, qp->srq ? "srq" : "rq",
272 qp->srq ? qp->srq->idx : qp->wq.rq.qid,
823 pr_debug("dev %s stag start 0x%0x size 0x%0x num stags %d pbl start 0x%0x size 0x%0x rq start 0x%0x size 0x%0x qp qid start %u size %u cq qid start %u size %u srq size %u\n",
833 rdev->lldi.vr->srq.size);
846 rdev->stats.srqt.total = rdev->lldi.vr->srq.size;
851 T4_MAX_NUM_PD, rdev->lldi.vr->srq.size);
/linux-master/drivers/nvme/target/
H A Drdma.c127 struct ib_srq *srq; member in struct:nvmet_rdma_srq
504 ret = ib_post_srq_recv(cmd->nsrq->srq, &cmd->wr, NULL);
1060 ib_destroy_srq(nsrq->srq);
1084 struct ib_srq *srq; local
1095 srq = ib_create_srq(ndev->pd, &srq_attr);
1096 if (IS_ERR(srq)) {
1097 ret = PTR_ERR(srq);
1107 nsrq->srq = srq;
1122 ib_destroy_srq(srq);
[all...]
/linux-master/include/linux/mlx4/
H A Ddevice.h833 int srq; member in struct:mlx4_quotas
928 } __packed srq; member in union:mlx4_eqe::__anon2624
1159 struct mlx4_mtt *mtt, u64 db_rec, struct mlx4_srq *srq);
1160 void mlx4_srq_free(struct mlx4_dev *dev, struct mlx4_srq *srq);
1161 int mlx4_srq_arm(struct mlx4_dev *dev, struct mlx4_srq *srq, int limit_watermark);
1162 int mlx4_srq_query(struct mlx4_dev *dev, struct mlx4_srq *srq, int *limit_watermark);
/linux-master/include/uapi/rdma/
H A Drdma_user_cm.h195 __u8 srq; member in struct:rdma_ucm_conn_param

Completed in 296 milliseconds

12345