/linux-master/drivers/net/ethernet/mellanox/mlx4/ |
H A D | profile.c | 184 dev->caps.num_qps = profile[i].num;
|
/linux-master/drivers/infiniband/hw/bnxt_re/ |
H A D | roce_hsi.h | 843 u8 num_qps; member in struct:cmdq_query_qp_extend
|
/linux-master/drivers/infiniband/hw/hns/ |
H A D | hns_roce_device.h | 736 u32 num_qps; member in struct:hns_roce_caps
|
H A D | hns_roce_hw_v2.c | 1905 to_hr_hem_hopnum(caps->qpc_hop_num, caps->num_qps)); 1933 to_hr_hem_hopnum(caps->sccc_hop_num, caps->num_qps)); 2009 calc_pg_sz(caps->num_qps, caps->qpc_sz, caps->qpc_hop_num, 2014 calc_pg_sz(caps->num_qps, caps->sccc_sz, caps->sccc_hop_num, 2202 caps->num_qps = 1 << hr_reg_read(resp_c, PF_CAPS_C_NUM_QPS); 2466 size = hr_dev->caps.num_qps * HNS_ROCE_V2_EXT_LLM_ENTRY_SZ; 4685 *tail = (*tail == hr_dev->caps.num_qps - 1) ? 0 : (*tail + 1); 4705 *head = (*head == hr_dev->caps.num_qps - 1) ? 0 : (*head + 1);
|
H A D | hns_roce_main.c | 180 props->max_qp = hr_dev->caps.num_qps; 377 resp.qp_tab_size = hr_dev->caps.num_qps; 805 hr_dev->caps.num_qps); 815 hr_dev->caps.num_qps); 827 hr_dev->caps.num_qps); 860 hr_dev->caps.num_qps);
|
H A D | hns_roce_qp.c | 1547 qp_table->idx_table.spare_idx = kcalloc(hr_dev->caps.num_qps, 1565 hr_dev->qp_table.bank[i].max = hr_dev->caps.num_qps /
|
/linux-master/drivers/infiniband/hw/mlx4/ |
H A D | main.c | 1108 resp_v3.qp_tab_size = dev->dev->caps.num_qps; 1113 resp.qp_tab_size = dev->dev->caps.num_qps;
|
/linux-master/drivers/infiniband/hw/mlx5/ |
H A D | gsi.c | 95 int num_qps = 0; local 101 num_qps = pd->device->attrs.max_pkeys; 103 num_qps = dev->lag_ports; 107 gsi->tx_qps = kcalloc(num_qps, sizeof(*gsi->tx_qps), GFP_KERNEL); 125 gsi->num_qps = num_qps; 142 if (num_qps) { 185 for (qp_index = 0; qp_index < gsi->num_qps; ++qp_index) { 340 for (qp_index = 0; qp_index < gsi->num_qps; ++qp_index) 417 if (!gsi->num_qps) [all...] |
H A D | mlx5_ib.h | 485 int num_qps; member in struct:mlx5_ib_gsi_qp
|
/linux-master/drivers/infiniband/hw/mthca/ |
H A D | mthca_cq.c | 522 (dev->limits.num_qps - 1));
|
H A D | mthca_dev.h | 143 int num_qps; member in struct:mthca_limits
|
H A D | mthca_main.c | 445 mdev->limits.num_qps, 456 mdev->limits.num_qps, 467 mdev->limits.num_qps <<
|
H A D | mthca_profile.c | 173 dev->limits.num_qps = profile[i].num;
|
H A D | mthca_provider.c | 90 props->max_qp = mdev->limits.num_qps - mdev->limits.reserved_qps; 297 uresp.qp_tab_size = to_mdev(ibdev)->limits.num_qps;
|
H A D | mthca_qp.c | 245 qp = mthca_array_get(&dev->qp_table.qp, qpn & (dev->limits.num_qps - 1)); 771 ((qp->qpn & (dev->limits.num_qps - 1)) * MTHCA_RDB_ENTRY_SIZE << 1330 qp->qpn & (dev->limits.num_qps - 1), qp); 1462 qp->qpn & (dev->limits.num_qps - 1)); 2273 dev->limits.num_qps, 2281 dev->limits.num_qps); 2302 mthca_array_cleanup(&dev->qp_table.qp, dev->limits.num_qps); 2315 mthca_array_cleanup(&dev->qp_table.qp, dev->limits.num_qps);
|
/linux-master/drivers/infiniband/hw/vmw_pvrdma/ |
H A D | pvrdma.h | 241 atomic_t num_qps; member in struct:pvrdma_dev
|
H A D | pvrdma_main.c | 115 atomic_set(&dev->num_qps, 0);
|
H A D | pvrdma_qp.c | 228 if (!atomic_add_unless(&dev->num_qps, 1, dev->dsr->caps.max_qp)) 426 atomic_dec(&dev->num_qps); 448 atomic_dec(&dev->num_qps);
|
/linux-master/drivers/net/ethernet/huawei/hinic/ |
H A D | hinic_dev.h | 86 u16 num_qps; member in struct:hinic_dev
|
H A D | hinic_ethtool.c | 643 if (queue >= nic_dev->num_qps) { 697 q_id >= nic_dev->num_qps) 731 if (queue >= nic_dev->num_qps) { 907 nic_dev->hwdev->nic_cap.num_qps = count; 916 nic_dev->hwdev->nic_cap.num_qps = count; 1109 cmd->data = nic_dev->num_qps; 1375 for (qid = 0; qid < nic_dev->num_qps; qid++) { 1388 for (qid = 0; qid < nic_dev->num_qps; qid++) { 1457 q_num = nic_dev->num_qps; 1494 for (i = 0; i < nic_dev->num_qps; [all...] |
H A D | hinic_hw_dev.c | 65 nic_cap->num_qps = (num_irqs - (num_aeqs + num_ceqs)) / 2; 67 if (nic_cap->num_qps > HINIC_Q_CTXT_MAX) 68 nic_cap->num_qps = HINIC_Q_CTXT_MAX; 75 if (nic_cap->num_qps > nic_cap->max_qps) 76 nic_cap->num_qps = nic_cap->max_qps; 443 int err, num_aeqs, num_ceqs, num_qps; local 472 num_qps = nic_cap->num_qps; 474 rq_msix_entries = &hwdev->msix_entries[num_aeqs + num_ceqs + num_qps]; 476 err = hinic_io_create_qps(func_to_io, base_qpn, num_qps, [all...] |
H A D | hinic_hw_dev.h | 35 u16 num_qps; member in struct:hinic_cap
|
H A D | hinic_hw_io.c | 33 #define CI_TABLE_SIZE(num_qps) ((num_qps) * CI_Q_ADDR_SIZE) 197 * @num_qps: number of qps to write 202 u16 num_qps) 204 return (write_sq_ctxts(func_to_io, base_qpn, num_qps) || 205 write_rq_ctxts(func_to_io, base_qpn, num_qps)); 365 * @num_qps: number queue pairs to create 372 u16 base_qpn, int num_qps, 381 func_to_io->qps = devm_kcalloc(&pdev->dev, num_qps, 386 func_to_io->sq_wq = devm_kcalloc(&pdev->dev, num_qps, 201 write_qp_ctxts(struct hinic_func_to_io *func_to_io, u16 base_qpn, u16 num_qps) argument 371 hinic_io_create_qps(struct hinic_func_to_io *func_to_io, u16 base_qpn, int num_qps, struct msix_entry *sq_msix_entries, struct msix_entry *rq_msix_entries) argument 468 hinic_io_destroy_qps(struct hinic_func_to_io *func_to_io, int num_qps) argument [all...] |
H A D | hinic_hw_io.h | 110 u16 base_qpn, int num_qps, 115 int num_qps);
|
H A D | hinic_main.c | 114 int i, num_qps = hinic_hwdev_num_qps(nic_dev->hwdev); local 116 for (i = 0; i < num_qps; i++) 119 for (i = 0; i < num_qps; i++) 373 nic_dev->num_qps = nic_dev->max_qps; 386 nic_dev->num_qps = nic_dev->max_qps; 403 nic_dev->num_qps = hinic_hwdev_num_qps(hwdev); 404 nic_dev->num_qps = min_t(u16, nic_dev->num_qps, num_cpus); 406 nic_dev->rss_limit = nic_dev->num_qps; 407 nic_dev->num_rss = nic_dev->num_qps; 1166 int err, num_qps; local [all...] |