/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/drivers/infiniband/core/ |
H A D | mad.c | 2712 qp_init_attr.send_cq = qp_info->port_priv->cq;
|
H A D | uverbs_cmd.c | 1059 attr.send_cq = scq; 1083 qp->send_cq = attr.send_cq; 1091 atomic_inc(&attr.send_cq->usecnt);
|
H A D | verbs.c | 293 qp->send_cq = qp_init_attr->send_cq; 301 atomic_inc(&qp_init_attr->send_cq->usecnt); 590 scq = qp->send_cq;
|
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/drivers/infiniband/hw/amso1100/ |
H A D | c2_qp.c | 419 struct c2_cq *send_cq = to_c2cq(qp_attrs->send_cq); local 459 wr.sq_cq_handle = send_cq->adapter_handle; 570 static inline void c2_lock_cqs(struct c2_cq *send_cq, struct c2_cq *recv_cq) argument 572 if (send_cq == recv_cq) 573 spin_lock_irq(&send_cq->lock); 574 else if (send_cq > recv_cq) { 575 spin_lock_irq(&send_cq->lock); 579 spin_lock_nested(&send_cq->lock, SINGLE_DEPTH_NESTING); 583 static inline void c2_unlock_cqs(struct c2_cq *send_cq, struc argument 598 struct c2_cq *send_cq; local [all...] |
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/drivers/infiniband/hw/cxgb3/ |
H A D | iwch_provider.c | 780 schp = get_chp(rhp, ((struct iwch_cq *) attrs->send_cq)->cq.cqid); 825 qhp->attr.scq = ((struct iwch_cq *) attrs->send_cq)->cq.cqid;
|
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/drivers/infiniband/hw/ehca/ |
H A D | ehca_classes.h | 133 struct ehca_cq *send_cq; member in struct:ehca_qp
|
H A D | ehca_main.c | 394 qp_init_attr.send_cq = ibcq;
|
H A D | ehca_qp.c | 465 my_qp->send_cq = 466 container_of(init_attr->send_cq, struct ehca_cq, ib_cq); 598 my_qp->ib_qp.send_cq = init_attr->send_cq; 624 if (init_attr->send_cq) { 625 struct ehca_cq *cq = container_of(init_attr->send_cq, 629 ehca_err(pd->device, "Couldn't assign qp to send_cq ret=%x", 633 my_qp->send_cq = cq; 1399 if (my_qp->send_cq) { 1400 ret = ehca_cq_unassign_qp(my_qp->send_cq, [all...] |
H A D | hcp_if.c | 325 qp->send_cq->ipz_cq_handle.handle,
|
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/drivers/infiniband/hw/ipath/ |
H A D | ipath_qp.c | 413 ipath_cq_enter(to_icq(qp->ibqp.send_cq), &wc, 1); 424 ipath_cq_enter(to_icq(qp->ibqp.send_cq), &wc, 1); 643 init_attr->send_cq = qp->ibqp.send_cq; 1007 ipath_cq_enter(to_icq(qp->ibqp.send_cq), wc, 1); 1017 ipath_cq_enter(to_icq(qp->ibqp.send_cq), wc, 1);
|
H A D | ipath_rc.c | 905 ipath_cq_enter(to_icq(qp->ibqp.send_cq), &wc, 0);
|
H A D | ipath_ruc.c | 452 ipath_cq_enter(to_icq(sqp->ibqp.send_cq), &wc, 0);
|
H A D | ipath_uc.c | 59 ipath_cq_enter(to_icq(qp->ibqp.send_cq), wc, 0);
|
H A D | ipath_ud.c | 482 ipath_cq_enter(to_icq(qp->ibqp.send_cq), &wc, 0);
|
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/drivers/infiniband/hw/mlx4/ |
H A D | qp.c | 456 static void mlx4_ib_lock_cqs(struct mlx4_ib_cq *send_cq, struct mlx4_ib_cq *recv_cq) argument 458 if (send_cq == recv_cq) 459 spin_lock_irq(&send_cq->lock); 460 else if (send_cq->mcq.cqn < recv_cq->mcq.cqn) { 461 spin_lock_irq(&send_cq->lock); 465 spin_lock_nested(&send_cq->lock, SINGLE_DEPTH_NESTING); 469 static void mlx4_ib_unlock_cqs(struct mlx4_ib_cq *send_cq, struct mlx4_ib_cq *recv_cq) argument 471 if (send_cq == recv_cq) 472 spin_unlock_irq(&send_cq->lock); 473 else if (send_cq 485 struct mlx4_ib_cq *send_cq, *recv_cq; local [all...] |
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/drivers/infiniband/hw/mthca/ |
H A D | mthca_dev.h | 547 struct mthca_cq *send_cq, 555 struct mthca_cq *send_cq,
|
H A D | mthca_provider.c | 588 to_mcq(init_attr->send_cq), 623 to_mcq(init_attr->send_cq),
|
H A D | mthca_qp.c | 719 qp_context->cqn_snd = cpu_to_be32(to_mcq(ibqp->send_cq)->cqn); 824 if (qp->ibqp.send_cq != qp->ibqp.recv_cq) 825 mthca_cq_clean(dev, to_mcq(qp->ibqp.send_cq), qp->qpn, NULL); 1171 struct mthca_cq *send_cq, 1292 struct mthca_cq *send_cq, 1319 err = mthca_alloc_qp_common(dev, pd, send_cq, recv_cq, 1334 static void mthca_lock_cqs(struct mthca_cq *send_cq, struct mthca_cq *recv_cq) argument 1336 if (send_cq == recv_cq) 1337 spin_lock_irq(&send_cq->lock); 1338 else if (send_cq 1169 mthca_alloc_qp_common(struct mthca_dev *dev, struct mthca_pd *pd, struct mthca_cq *send_cq, struct mthca_cq *recv_cq, enum ib_sig_type send_policy, struct mthca_qp *qp) argument 1290 mthca_alloc_qp(struct mthca_dev *dev, struct mthca_pd *pd, struct mthca_cq *send_cq, struct mthca_cq *recv_cq, enum ib_qp_type type, enum ib_sig_type send_policy, struct ib_qp_cap *cap, struct mthca_qp *qp) argument 1347 mthca_unlock_cqs(struct mthca_cq *send_cq, struct mthca_cq *recv_cq) argument 1360 mthca_alloc_sqp(struct mthca_dev *dev, struct mthca_pd *pd, struct mthca_cq *send_cq, struct mthca_cq *recv_cq, enum ib_sig_type send_policy, struct ib_qp_cap *cap, int qpn, int port, struct mthca_sqp *sqp) argument 1442 struct mthca_cq *send_cq; local [all...] |
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/drivers/infiniband/ulp/ipoib/ |
H A D | ipoib_cm.c | 202 .send_cq = priv->cq, /* For drain WR */ 811 attr.send_cq = cq;
|
H A D | ipoib_verbs.c | 187 init_attr.send_cq = priv->cq;
|
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/drivers/infiniband/ulp/iser/ |
H A D | iser_verbs.c | 176 init_attr.send_cq = device->cq;
|
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/drivers/infiniband/ulp/srp/ |
H A D | ib_srp.c | 215 init_attr->send_cq = target->cq;
|
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/drivers/net/ehea/ |
H A D | ehea.h | 348 struct ehea_cq *send_cq; member in struct:ehea_port_res
|
H A D | ehea_main.c | 485 struct ehea_cq *send_cq = pr->send_cq; local 493 cqe = ehea_poll_cq(send_cq); 495 ehea_inc_cq(send_cq); 523 cqe = ehea_poll_cq(send_cq); 526 ehea_update_feca(send_cq, cqe_counter); 551 cqe_skb = ehea_poll_cq(pr->send_cq); 559 ehea_reset_cq_ep(pr->send_cq); 561 ehea_reset_cq_n1(pr->send_cq); 563 cqe_skb = ehea_poll_cq(pr->send_cq); [all...] |
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/include/rdma/ |
H A D | ib_verbs.h | 490 struct ib_cq *send_cq; member in struct:ib_qp_init_attr 775 struct ib_cq *send_cq; member in struct:ib_qp
|