/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/drivers/net/mlx4/ |
H A D | qp.c | 46 struct mlx4_qp_table *qp_table = &mlx4_priv(dev)->qp_table; local 49 spin_lock(&qp_table->lock); 55 spin_unlock(&qp_table->lock); 154 struct mlx4_qp_table *qp_table = &priv->qp_table; local 160 qp->qpn = mlx4_bitmap_alloc(&qp_table->bitmap); 165 err = mlx4_table_get(dev, &qp_table->qp_table, qp->qpn); 169 err = mlx4_table_get(dev, &qp_table 221 struct mlx4_qp_table *qp_table = &mlx4_priv(dev)->qp_table; local 232 struct mlx4_qp_table *qp_table = &mlx4_priv(dev)->qp_table; local 256 struct mlx4_qp_table *qp_table = &mlx4_priv(dev)->qp_table; local [all...] |
H A D | profile.c | 173 for (priv->qp_table.rdmarc_shift = 0; 174 request->num_qp << priv->qp_table.rdmarc_shift < profile[i].num; 175 ++priv->qp_table.rdmarc_shift) 177 dev->caps.max_qp_dest_rdma = 1 << priv->qp_table.rdmarc_shift; 178 priv->qp_table.rdmarc_base = (u32) profile[i].start; 180 init_hca->log_rd_per_qp = priv->qp_table.rdmarc_shift;
|
H A D | main.c | 204 err = mlx4_init_icm_table(dev, &priv->qp_table.cmpt_table, 251 mlx4_cleanup_icm_table(dev, &priv->qp_table.cmpt_table); 321 err = mlx4_init_icm_table(dev, &priv->qp_table.qp_table, 331 err = mlx4_init_icm_table(dev, &priv->qp_table.auxc_table, 341 err = mlx4_init_icm_table(dev, &priv->qp_table.altc_table, 351 err = mlx4_init_icm_table(dev, &priv->qp_table.rdmarc_table, 353 dev_cap->rdmarc_entry_sz << priv->qp_table.rdmarc_shift, 405 mlx4_cleanup_icm_table(dev, &priv->qp_table.rdmarc_table); 408 mlx4_cleanup_icm_table(dev, &priv->qp_table [all...] |
H A D | mlx4.h | 235 struct mlx4_icm_table qp_table; member in struct:mlx4_qp_table 269 struct mlx4_qp_table qp_table; member in struct:mlx4_priv
|
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/drivers/infiniband/hw/mthca/ |
H A D | mthca_qp.c | 199 return qp->qpn >= dev->qp_table.sqp_start && 200 qp->qpn <= dev->qp_table.sqp_start + 3; 205 return qp->qpn >= dev->qp_table.sqp_start && 206 qp->qpn <= dev->qp_table.sqp_start + 1; 245 spin_lock(&dev->qp_table.lock); 246 qp = mthca_array_get(&dev->qp_table.qp, qpn & (dev->limits.num_qps - 1)); 249 spin_unlock(&dev->qp_table.lock); 265 spin_lock(&dev->qp_table.lock); 268 spin_unlock(&dev->qp_table.lock); 754 cpu_to_be32(dev->qp_table [all...] |
H A D | mthca_main.c | 492 mdev->qp_table.qp_table = mthca_alloc_icm_table(mdev, init_hca->qpc_base, 497 if (!mdev->qp_table.qp_table) { 503 mdev->qp_table.eqp_table = mthca_alloc_icm_table(mdev, init_hca->eqpc_base, 508 if (!mdev->qp_table.eqp_table) { 514 mdev->qp_table.rdb_table = mthca_alloc_icm_table(mdev, init_hca->rdb_base, 517 mdev->qp_table.rdb_shift, 0, 519 if (!mdev->qp_table.rdb_table) { 579 mthca_free_icm_table(mdev, mdev->qp_table [all...] |
H A D | mthca_profile.c | 212 for (dev->qp_table.rdb_shift = 0; 213 request->num_qp << dev->qp_table.rdb_shift < profile[i].num; 214 ++dev->qp_table.rdb_shift) 216 dev->qp_table.rdb_base = (u32) profile[i].start;
|
H A D | mthca_dev.h | 262 struct mthca_icm_table *qp_table; member in struct:mthca_qp_table 348 struct mthca_qp_table qp_table; member in struct:mthca_dev
|
H A D | mthca_cq.c | 529 *cur_qp = mthca_array_get(&dev->qp_table.qp,
|
H A D | mthca_provider.c | 105 props->max_qp_rd_atom = 1 << mdev->qp_table.rdb_shift;
|
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/drivers/infiniband/hw/amso1100/ |
H A D | c2_qp.c | 385 spin_lock_irq(&c2dev->qp_table.lock); 386 ret = idr_get_new_above(&c2dev->qp_table.idr, qp, 387 c2dev->qp_table.last++, &qp->qpn); 388 spin_unlock_irq(&c2dev->qp_table.lock); 390 idr_pre_get(&c2dev->qp_table.idr, GFP_KERNEL)); 396 spin_lock_irq(&c2dev->qp_table.lock); 397 idr_remove(&c2dev->qp_table.idr, qpn); 398 spin_unlock_irq(&c2dev->qp_table.lock); 406 spin_lock_irqsave(&c2dev->qp_table.lock, flags); 407 qp = idr_find(&c2dev->qp_table [all...] |
H A D | c2.h | 315 struct c2_qp_table qp_table; member in struct:c2_dev
|
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/drivers/infiniband/hw/ipath/ |
H A D | ipath_qp.c | 818 err = ipath_alloc_qpn(&dev->qp_table, qp, 939 ipath_free_qp(&dev->qp_table, qp); 962 idev->qp_table.last = 1; /* QPN 0 and 1 are special. */ 963 idev->qp_table.max = size; 964 idev->qp_table.nmaps = 1; 965 idev->qp_table.table = kzalloc(size * sizeof(*idev->qp_table.table), 967 if (idev->qp_table.table == NULL) { 972 for (i = 0; i < ARRAY_SIZE(idev->qp_table.map); i++) { 973 atomic_set(&idev->qp_table [all...] |
H A D | ipath_verbs.c | 455 qp = ipath_lookup_qpn(&dev->qp_table, qp_num); 1450 spin_lock_init(&idev->qp_table.lock); 1622 kfree(idev->qp_table.table); 1655 ipath_free_all_qps(&dev->qp_table); 1656 kfree(dev->qp_table.table);
|
H A D | ipath_cq.c | 143 qp = ipath_lookup_qpn(&to_idev(cq->ibcq.device)->qp_table,
|
H A D | ipath_ud.c | 116 qp = ipath_lookup_qpn(&dev->qp_table, wr->wr.ud.remote_qpn);
|
H A D | ipath_verbs.h | 496 struct ipath_qp_table qp_table; member in struct:ipath_ibdev
|
H A D | ipath_ruc.c | 252 qp = ipath_lookup_qpn(&dev->qp_table, sqp->remote_qpn);
|