/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/drivers/net/mlx4/ |
H A D | cq.c | 76 void mlx4_cq_completion(struct mlx4_dev *dev, u32 cqn) argument 81 cqn & (dev->caps.num_cqs - 1)); 83 mlx4_warn(dev, "Completion event for bogus CQ %08x\n", cqn); 92 void mlx4_cq_event(struct mlx4_dev *dev, u32 cqn, int event_type) argument 99 cq = radix_tree_lookup(&cq_table->tree, cqn & (dev->caps.num_cqs - 1)); 106 mlx4_warn(dev, "Async event for bogus CQ %08x\n", cqn); 155 err = mlx4_MODIFY_CQ(dev, mailbox, cq->cqn, 1); 183 err = mlx4_MODIFY_CQ(dev, mailbox, cq->cqn, 0); 206 cq->cqn = mlx4_bitmap_alloc(&cq_table->bitmap); 207 if (cq->cqn [all...] |
H A D | en_resources.c | 41 int is_tx, int rss, int qpn, int cqn, 61 context->cqn_send = cpu_to_be32(cqn); 62 context->cqn_recv = cpu_to_be32(cqn); 40 mlx4_en_fill_qp_context(struct mlx4_en_priv *priv, int size, int stride, int is_tx, int rss, int qpn, int cqn, struct mlx4_qp_context *context) argument
|
H A D | eq.c | 112 __be32 cqn; member in struct:mlx4_eqe::__anon12987::__anon12988 129 __be32 cqn; member in struct:mlx4_eqe::__anon12987::__anon12992 167 int cqn; local 181 cqn = be32_to_cpu(eqe->event.comp.cqn) & 0xffffff; 182 mlx4_cq_completion(dev, cqn); 227 be32_to_cpu(eqe->event.cq_err.cqn) & 0xffffff); 228 mlx4_cq_event(dev, be32_to_cpu(eqe->event.cq_err.cqn),
|
H A D | mlx4_en.h | 223 u16 cqn; /* index of port CQ associated with this ring */ member in struct:mlx4_en_tx_ring 258 u16 cqn; /* index of port CQ associated with this ring */ member in struct:mlx4_en_rx_ring 514 int is_tx, int rss, int qpn, int cqn,
|
H A D | mlx4.h | 380 void mlx4_cq_completion(struct mlx4_dev *dev, u32 cqn); 381 void mlx4_cq_event(struct mlx4_dev *dev, u32 cqn, int event_type);
|
H A D | en_rx.c | 361 ring->cqn = priv->rx_cq[ring_ind].mcq.cqn; 820 qpn, ring->cqn, context); 882 priv->rx_ring[0].cqn, &context);
|
H A D | en_netdev.c | 588 priv->rx_ring[i].cqn = cq->mcq.cqn; 618 err = mlx4_en_activate_tx_ring(priv, tx_ring, cq->mcq.cqn);
|
H A D | en_tx.c | 159 ring->cqn = cq; 172 ring->cqn, &ring->context);
|
/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/drivers/infiniband/hw/amso1100/ |
H A D | c2_user.h | 69 __u32 cqn; member in struct:c2_create_cq_resp
|
H A D | c2_cq.c | 46 static struct c2_cq *c2_cq_get(struct c2_dev *c2dev, int cqn) argument 52 cq = c2dev->qptr_array[cqn]; 369 cq->cqn = cq->mq.index; 370 c2dev->qptr_array[cq->cqn] = cq;
|
H A D | c2_provider.h | 95 int cqn; member in struct:c2_cq
|
H A D | c2_qp.c | 622 c2_cq_clean(c2dev, qp, send_cq->cqn); 624 c2_cq_clean(c2dev, qp, recv_cq->cqn);
|
/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/drivers/infiniband/hw/mlx4/ |
H A D | user.h | 70 __u32 cqn; member in struct:mlx4_ib_create_cq_resp
|
H A D | cq.c | 54 "on CQ %06x\n", type, cq->cqn); 234 if (ib_copy_to_udata(udata, &cq->mcq.cqn, sizeof (__u32))) { 603 cq->mcq.cqn, be32_to_cpu(cqe->vlan_my_qpn) & MLX4_CQE_QPN_MASK);
|
H A D | qp.c | 628 } else if (send_cq->mcq.cqn < recv_cq->mcq.cqn) { 643 } else if (send_cq->mcq.cqn < recv_cq->mcq.cqn) { 1032 context->cqn_send = cpu_to_be32(to_mcq(ibqp->send_cq)->mcq.cqn); 1056 context->cqn_recv = cpu_to_be32(to_mcq(ibqp->recv_cq)->mcq.cqn);
|
/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/drivers/infiniband/hw/mthca/ |
H A D | mthca_cq.c | 76 __be32 cqn; member in struct:mthca_cq_context 211 mthca_write64(MTHCA_TAVOR_CQ_DB_INC_CI | cq->cqn, incr - 1, 222 void mthca_cq_completion(struct mthca_dev *dev, u32 cqn) argument 226 cq = mthca_array_get(&dev->cq_table.cq, cqn & (dev->limits.num_cqs - 1)); 229 mthca_warn(dev, "Completion event for bogus CQ %08x\n", cqn); 238 void mthca_cq_event(struct mthca_dev *dev, u32 cqn, argument 246 cq = mthca_array_get(&dev->cq_table.cq, cqn & (dev->limits.num_cqs - 1)); 253 mthca_warn(dev, "Async event for bogus CQ %08x\n", cqn); 302 qpn, cq->cqn, cq->cons_index, prod_index); 389 cq->cqn, c [all...] |
H A D | mthca_user.h | 83 __u32 cqn; member in struct:mthca_create_cq_resp
|
H A D | mthca_eq.c | 132 __be32 cqn; member in struct:mthca_eqe::__anon11647::__anon11648 149 __be32 cqn; member in struct:mthca_eqe::__anon11647::__anon11652 219 static inline void disarm_cq(struct mthca_dev *dev, int eqn, int cqn) argument 222 mthca_write64(MTHCA_EQ_DB_DISARM_CQ | eqn, cqn, 276 disarm_cqn = be32_to_cpu(eqe->event.comp.cqn) & 0xffffff; 343 be32_to_cpu(eqe->event.cq_err.cqn) & 0xffffff); 344 mthca_cq_event(dev, be32_to_cpu(eqe->event.cq_err.cqn),
|
H A D | mthca_provider.h | 141 * a qp may be locked, with the cq with the lower cqn locked first. 205 int cqn; member in struct:mthca_cq
|
H A D | mthca_dev.h | 503 void mthca_cq_completion(struct mthca_dev *dev, u32 cqn); 504 void mthca_cq_event(struct mthca_dev *dev, u32 cqn,
|
H A D | mthca_qp.c | 725 qp_context->cqn_snd = cpu_to_be32(to_mcq(ibqp->send_cq)->cqn); 764 qp_context->cqn_rcv = cpu_to_be32(to_mcq(ibqp->recv_cq)->cqn); 1327 } else if (send_cq->cqn < recv_cq->cqn) { 1342 } else if (send_cq->cqn < recv_cq->cqn) {
|
H A D | mthca_provider.c | 727 if (context && ib_copy_to_udata(udata, &cq->cqn, sizeof (__u32))) { 830 ret = mthca_RESIZE_CQ(dev, cq->cqn, lkey, ilog2(entries), &status);
|
/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/include/linux/mlx4/ |
H A D | cq.h | 129 doorbell[0] = cpu_to_be32(sn << 28 | cmd | cq->cqn);
|
H A D | device.h | 335 int cqn; member in struct:mlx4_cq
|