/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/drivers/infiniband/core/ |
H A D | mad.c | 2713 qp_init_attr.recv_cq = qp_info->port_priv->cq;
|
H A D | uverbs_cmd.c | 1060 attr.recv_cq = rcq; 1084 qp->recv_cq = attr.recv_cq; 1092 atomic_inc(&attr.recv_cq->usecnt);
|
H A D | verbs.c | 294 qp->recv_cq = qp_init_attr->recv_cq; 302 atomic_inc(&qp_init_attr->recv_cq->usecnt); 591 rcq = qp->recv_cq;
|
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/drivers/infiniband/hw/amso1100/ |
H A D | c2_qp.c | 420 struct c2_cq *recv_cq = to_c2cq(qp_attrs->recv_cq); local 460 wr.rq_cq_handle = recv_cq->adapter_handle; 570 static inline void c2_lock_cqs(struct c2_cq *send_cq, struct c2_cq *recv_cq) argument 572 if (send_cq == recv_cq) 574 else if (send_cq > recv_cq) { 576 spin_lock_nested(&recv_cq->lock, SINGLE_DEPTH_NESTING); 578 spin_lock_irq(&recv_cq->lock); 583 static inline void c2_unlock_cqs(struct c2_cq *send_cq, struct c2_cq *recv_cq) argument 585 if (send_cq == recv_cq) 599 struct c2_cq *recv_cq; local [all...] |
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/drivers/infiniband/hw/cxgb3/ |
H A D | iwch_provider.c | 781 rchp = get_chp(rhp, ((struct iwch_cq *) attrs->recv_cq)->cq.cqid); 826 qhp->attr.rcq = ((struct iwch_cq *) attrs->recv_cq)->cq.cqid;
|
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/drivers/infiniband/hw/ehca/ |
H A D | ehca_classes.h | 134 struct ehca_cq *recv_cq; member in struct:ehca_qp
|
H A D | ehca_main.c | 395 qp_init_attr.recv_cq = ibcq;
|
H A D | ehca_qp.c | 463 my_qp->recv_cq = 464 container_of(init_attr->recv_cq, struct ehca_cq, ib_cq); 597 my_qp->ib_qp.recv_cq = init_attr->recv_cq;
|
H A D | hcp_if.c | 326 qp->recv_cq->ipz_cq_handle.handle,
|
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/drivers/infiniband/hw/ipath/ |
H A D | ipath_qp.c | 450 ipath_cq_enter(to_icq(qp->ibqp.recv_cq), &wc, 1); 644 init_attr->recv_cq = qp->ibqp.recv_cq;
|
H A D | ipath_rc.c | 1666 ipath_cq_enter(to_icq(qp->ibqp.recv_cq), &wc,
|
H A D | ipath_ruc.c | 141 ipath_cq_enter(to_icq(qp->ibqp.recv_cq), &wc, 1); 432 ipath_cq_enter(to_icq(qp->ibqp.recv_cq), &wc,
|
H A D | ipath_uc.c | 420 ipath_cq_enter(to_icq(qp->ibqp.recv_cq), &wc,
|
H A D | ipath_ud.c | 78 ipath_cq_enter(to_icq(qp->ibqp.recv_cq), &wc, 1); 267 ipath_cq_enter(to_icq(qp->ibqp.recv_cq), wc, 667 ipath_cq_enter(to_icq(qp->ibqp.recv_cq), &wc,
|
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/drivers/infiniband/hw/mlx4/ |
H A D | qp.c | 456 static void mlx4_ib_lock_cqs(struct mlx4_ib_cq *send_cq, struct mlx4_ib_cq *recv_cq) argument 458 if (send_cq == recv_cq) 460 else if (send_cq->mcq.cqn < recv_cq->mcq.cqn) { 462 spin_lock_nested(&recv_cq->lock, SINGLE_DEPTH_NESTING); 464 spin_lock_irq(&recv_cq->lock); 469 static void mlx4_ib_unlock_cqs(struct mlx4_ib_cq *send_cq, struct mlx4_ib_cq *recv_cq) argument 471 if (send_cq == recv_cq) 473 else if (send_cq->mcq.cqn < recv_cq->mcq.cqn) { 474 spin_unlock(&recv_cq->lock); 478 spin_unlock_irq(&recv_cq 485 struct mlx4_ib_cq *send_cq, *recv_cq; local [all...] |
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/drivers/infiniband/hw/mthca/ |
H A D | mthca_dev.h | 548 struct mthca_cq *recv_cq, 556 struct mthca_cq *recv_cq,
|
H A D | mthca_provider.c | 589 to_mcq(init_attr->recv_cq), 624 to_mcq(init_attr->recv_cq),
|
H A D | mthca_qp.c | 758 qp_context->cqn_rcv = cpu_to_be32(to_mcq(ibqp->recv_cq)->cqn); 822 mthca_cq_clean(dev, to_mcq(qp->ibqp.recv_cq), qp->qpn, 824 if (qp->ibqp.send_cq != qp->ibqp.recv_cq) 1172 struct mthca_cq *recv_cq, 1293 struct mthca_cq *recv_cq, 1319 err = mthca_alloc_qp_common(dev, pd, send_cq, recv_cq, 1334 static void mthca_lock_cqs(struct mthca_cq *send_cq, struct mthca_cq *recv_cq) argument 1336 if (send_cq == recv_cq) 1338 else if (send_cq->cqn < recv_cq->cqn) { 1340 spin_lock_nested(&recv_cq 1169 mthca_alloc_qp_common(struct mthca_dev *dev, struct mthca_pd *pd, struct mthca_cq *send_cq, struct mthca_cq *recv_cq, enum ib_sig_type send_policy, struct mthca_qp *qp) argument 1290 mthca_alloc_qp(struct mthca_dev *dev, struct mthca_pd *pd, struct mthca_cq *send_cq, struct mthca_cq *recv_cq, enum ib_qp_type type, enum ib_sig_type send_policy, struct ib_qp_cap *cap, struct mthca_qp *qp) argument 1347 mthca_unlock_cqs(struct mthca_cq *send_cq, struct mthca_cq *recv_cq) argument 1360 mthca_alloc_sqp(struct mthca_dev *dev, struct mthca_pd *pd, struct mthca_cq *send_cq, struct mthca_cq *recv_cq, enum ib_sig_type send_policy, struct ib_qp_cap *cap, int qpn, int port, struct mthca_sqp *sqp) argument 1443 struct mthca_cq *recv_cq; local [all...] |
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/drivers/infiniband/ulp/ipoib/ |
H A D | ipoib_cm.c | 203 .recv_cq = priv->cq, 805 attr.recv_cq = priv->cq;
|
H A D | ipoib_verbs.c | 188 init_attr.recv_cq = priv->cq,
|
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/drivers/infiniband/ulp/iser/ |
H A D | iser_verbs.c | 177 init_attr.recv_cq = device->cq;
|
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/drivers/infiniband/ulp/srp/ |
H A D | ib_srp.c | 216 init_attr->recv_cq = target->cq;
|
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/drivers/net/ehea/ |
H A D | ehea.h | 349 struct ehea_cq *recv_cq; member in struct:ehea_port_res
|
H A D | ehea_main.c | 558 ehea_reset_cq_ep(pr->recv_cq); 560 ehea_reset_cq_n1(pr->recv_cq); 1120 pr->recv_cq = ehea_create_cq(adapter, pr_cfg->max_entries_rcq, 1123 if (!pr->recv_cq) { 1139 pr->recv_cq->attr.act_nr_of_cqes); 1164 init_attr->recv_cq_handle = pr->recv_cq->fw_handle; 1219 ehea_destroy_cq(pr->recv_cq); 1235 ehea_destroy_cq(pr->recv_cq);
|
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/include/rdma/ |
H A D | ib_verbs.h | 491 struct ib_cq *recv_cq; member in struct:ib_qp_init_attr 776 struct ib_cq *recv_cq; member in struct:ib_qp
|