/linux-master/drivers/infiniband/hw/mana/ |
H A D | device.c | 60 dev = ib_alloc_device(mana_ib_dev, ib_dev); 64 ib_set_device_ops(&dev->ib_dev, &mana_ib_dev_ops); 66 dev->ib_dev.phys_port_cnt = mc->num_ports; 68 ibdev_dbg(&dev->ib_dev, "mdev=%p id=%d num_ports=%d\n", mdev, 69 mdev->dev_id.as_uint32, dev->ib_dev.phys_port_cnt); 71 dev->ib_dev.node_type = RDMA_NODE_IB_CA; 77 dev->ib_dev.num_comp_vectors = 1; 78 dev->ib_dev.dev.parent = mdev->gdma_context->dev; 82 ibdev_err(&dev->ib_dev, "Failed to register device, ret %d", 90 ibdev_err(&dev->ib_dev, "Faile [all...] |
H A D | wq.c | 13 container_of(pd->device, struct mana_ib_dev, ib_dev); 24 ibdev_dbg(&mdev->ib_dev, 33 ibdev_dbg(&mdev->ib_dev, "ucmd wq_buf_addr 0x%llx\n", ucmd.wq_buf_addr); 39 ibdev_dbg(&mdev->ib_dev, 51 ibdev_dbg(&mdev->ib_dev, 57 ibdev_dbg(&mdev->ib_dev, 84 struct ib_device *ib_dev = ibwq->device; local 87 mdev = container_of(ib_dev, struct mana_ib_dev, ib_dev);
|
H A D | qp.c | 56 ibdev_dbg(&dev->ib_dev, "ind table size %u\n", 1 << log_ind_tbl_size); 59 ibdev_dbg(&dev->ib_dev, "index %u handle 0x%llx\n", i, 69 ibdev_dbg(&dev->ib_dev, "vport handle %llu default_rxobj 0x%llx\n", 99 container_of(pd->device, struct mana_ib_dev, ib_dev); 123 ibdev_dbg(&mdev->ib_dev, 130 ibdev_dbg(&mdev->ib_dev, 137 ibdev_dbg(&mdev->ib_dev, 145 ibdev_dbg(&mdev->ib_dev, 152 ibdev_dbg(&mdev->ib_dev, 162 ibdev_dbg(&mdev->ib_dev, "Invali [all...] |
H A D | main.c | 14 ndev = mana_ib_get_netdev(&dev->ib_dev, port); 35 ndev = mana_ib_get_netdev(&dev->ib_dev, port); 42 ibdev_dbg(&dev->ib_dev, 53 ibdev_dbg(&dev->ib_dev, "Failed to configure vPort %d\n", err); 62 ibdev_dbg(&dev->ib_dev, "vport handle %llx pdid %x doorbell_id %x\n", 79 dev = container_of(ibdev, struct mana_ib_dev, ib_dev); 90 ibdev_dbg(&dev->ib_dev, 101 ibdev_dbg(&dev->ib_dev, "pd_handle 0x%llx pd_id %d\n", 119 dev = container_of(ibdev, struct mana_ib_dev, ib_dev); 130 ibdev_dbg(&dev->ib_dev, [all...] |
/linux-master/drivers/infiniband/core/ |
H A D | roce_gid_mgmt.c | 82 unsigned long roce_gid_type_mask_support(struct ib_device *ib_dev, u32 port) argument 87 if (!rdma_protocol_roce(ib_dev, port)) 91 if (PORT_CAP_TO_GID_TYPE[i].is_supported(ib_dev, port)) 98 static void update_gid(enum gid_op_type gid_op, struct ib_device *ib_dev, argument 103 unsigned long gid_type_mask = roce_gid_type_mask_support(ib_dev, port); 110 ib_cache_gid_add(ib_dev, port, 114 ib_cache_gid_del(ib_dev, port, 147 is_eth_port_of_netdev_filter(struct ib_device *ib_dev, u32 port, argument 171 is_eth_port_inactive_slave_filter(struct ib_device *ib_dev, u32 port, argument 192 * @ib_dev 201 is_ndev_for_default_gid_filter(struct ib_device *ib_dev, u32 port, struct net_device *rdma_ndev, void *cookie) argument 227 pass_all_filter(struct ib_device *ib_dev, u32 port, struct net_device *rdma_ndev, void *cookie) argument 233 upper_device_filter(struct ib_device *ib_dev, u32 port, struct net_device *rdma_ndev, void *cookie) argument 264 is_upper_ndev_bond_master_filter(struct ib_device *ib_dev, u32 port, struct net_device *rdma_ndev, void *cookie) argument 282 update_gid_ip(enum gid_op_type gid_op, struct ib_device *ib_dev, u32 port, struct net_device *ndev, struct sockaddr *addr) argument 297 bond_delete_netdev_default_gids(struct ib_device *ib_dev, u32 port, struct net_device *rdma_ndev, struct net_device *event_ndev) argument 331 enum_netdev_ipv4_ips(struct ib_device *ib_dev, u32 port, struct net_device *ndev) argument 375 enum_netdev_ipv6_ips(struct ib_device *ib_dev, u32 port, struct net_device *ndev) argument 421 _add_netdev_ips(struct ib_device *ib_dev, u32 port, struct net_device *ndev) argument 429 add_netdev_ips(struct ib_device *ib_dev, u32 port, struct net_device *rdma_ndev, void *cookie) argument 435 del_netdev_ips(struct ib_device *ib_dev, u32 port, struct net_device *rdma_ndev, void *cookie) argument 450 del_default_gids(struct ib_device *ib_dev, u32 port, struct net_device *rdma_ndev, void *cookie) argument 462 add_default_gids(struct ib_device *ib_dev, u32 port, struct net_device *rdma_ndev, void *cookie) argument 473 enum_all_gids_of_dev_cb(struct ib_device *ib_dev, u32 port, struct net_device *rdma_ndev, void *cookie) argument 511 rdma_roce_rescan_device(struct ib_device *ib_dev) argument 551 handle_netdev_upper(struct ib_device *ib_dev, u32 port, void *cookie, void (*handle_netdev)(struct ib_device *ib_dev, u32 port, struct net_device *ndev)) argument 578 _roce_del_all_netdev_gids(struct ib_device *ib_dev, u32 port, struct net_device *event_ndev) argument 584 del_netdev_upper_ips(struct ib_device *ib_dev, u32 port, struct net_device *rdma_ndev, void *cookie) argument 590 add_netdev_upper_ips(struct ib_device *ib_dev, u32 port, struct net_device *rdma_ndev, void *cookie) argument 596 del_netdev_default_ips_join(struct ib_device *ib_dev, u32 port, struct net_device *rdma_ndev, void *cookie) argument [all...] |
H A D | uverbs_std_types_dm.c | 56 struct ib_device *ib_dev = attrs->context->device; local 60 if (!ib_dev->ops.alloc_dm) 73 dm = ib_dev->ops.alloc_dm(ib_dev, attrs->context, &attr, attrs); 77 dm->device = ib_dev;
|
H A D | uverbs_std_types_counters.c | 60 struct ib_device *ib_dev = attrs->context->device; local 69 if (!ib_dev->ops.create_counters) 72 counters = rdma_zalloc_drv_obj(ib_dev, ib_counters); 76 counters->device = ib_dev; 81 ret = ib_dev->ops.create_counters(counters, attrs);
|
H A D | uverbs_std_types_device.c | 141 struct ib_device *ib_dev, u8 port_num) 153 if (rdma_is_grh_required(ib_dev, port_num)) 156 if (rdma_cap_opa_ah(ib_dev, port_num)) { 173 resp->link_layer = rdma_port_get_link_layer(ib_dev, port_num); 179 struct ib_device *ib_dev; local 189 ib_dev = ucontext->device; 192 if (!ib_dev->ops.query_port) 200 ret = ib_query_port(ib_dev, port_num, &attr); 204 copy_port_attr_to_resp(&attr, &resp.legacy_resp, ib_dev, port_num); 246 struct ib_device *ib_dev; local 139 copy_port_attr_to_resp(struct ib_port_attr *attr, struct ib_uverbs_query_port_resp *resp, struct ib_device *ib_dev, u8 port_num) argument 318 struct ib_device *ib_dev; local 374 struct ib_device *ib_dev; local [all...] |
H A D | cache.c | 124 static void dispatch_gid_change_event(struct ib_device *ib_dev, u32 port) argument 128 event.device = ib_dev; 376 * @ib_dev: IB device whose GID entry to be deleted 382 static void del_gid(struct ib_device *ib_dev, u32 port, argument 390 dev_dbg(&ib_dev->dev, "%s port=%u index=%d gid %pI6\n", __func__, port, 399 if (!rdma_protocol_roce(ib_dev, port)) 403 if (rdma_cap_roce_gid_table(ib_dev, port)) 404 ib_dev->ops.del_gid(&entry->attr, &entry->context); 546 static int __ib_cache_gid_add(struct ib_device *ib_dev, u32 port, argument 562 table = rdma_gid_table(ib_dev, por 590 ib_cache_gid_add(struct ib_device *ib_dev, u32 port, union ib_gid *gid, struct ib_gid_attr *attr) argument 601 _ib_cache_gid_del(struct ib_device *ib_dev, u32 port, union ib_gid *gid, struct ib_gid_attr *attr, unsigned long mask, bool default_gid) argument 630 ib_cache_gid_del(struct ib_device *ib_dev, u32 port, union ib_gid *gid, struct ib_gid_attr *attr) argument 641 ib_cache_gid_del_all_netdev_gids(struct ib_device *ib_dev, u32 port, struct net_device *ndev) argument 683 rdma_find_gid_by_port(struct ib_device *ib_dev, const union ib_gid *gid, enum ib_gid_type gid_type, u32 port, struct net_device *ndev) argument 736 rdma_find_gid_by_filter( struct ib_device *ib_dev, const union ib_gid *gid, u32 port, bool (*filter)(const union ib_gid *gid, const struct ib_gid_attr *, void *), void *context) argument 821 cleanup_gid_table_port(struct ib_device *ib_dev, u32 port, struct ib_gid_table *table) argument 837 ib_cache_gid_set_default_gid(struct ib_device *ib_dev, u32 port, struct net_device *ndev, unsigned long gid_type_mask, enum ib_cache_gid_default_mode mode) argument 870 gid_table_reserve_default(struct ib_device *ib_dev, u32 port, struct ib_gid_table *table) argument 885 gid_table_release_one(struct ib_device *ib_dev) argument 895 _gid_table_setup_one(struct ib_device *ib_dev) argument 916 gid_table_cleanup_one(struct ib_device *ib_dev) argument 925 gid_table_setup_one(struct ib_device *ib_dev) argument [all...] |
H A D | uverbs_std_types_mr.c | 53 struct ib_device *ib_dev = pd->device; local 60 if (!ib_dev->ops.advise_mr) 79 return ib_dev->ops.advise_mr(pd, advice, flags, sg_list, num_sge, 93 struct ib_device *ib_dev = pd->device; local 98 if (!ib_dev->ops.reg_dm_mr) 119 ret = ib_check_mr_access(ib_dev, attr.access_flags); 193 struct ib_device *ib_dev = pd->device; local 200 if (!ib_dev->ops.reg_user_mr_dmabuf) 236 ret = ib_check_mr_access(ib_dev, access_flags);
|
H A D | device.c | 186 static void free_netdevs(struct ib_device *ib_dev); 821 pdata->ib_dev = device; 1344 static void prevent_dealloc_device(struct ib_device *ib_dev) argument 1462 static void __ib_unregister_device(struct ib_device *ib_dev) argument 1471 mutex_lock(&ib_dev->unregistration_lock); 1472 if (!refcount_read(&ib_dev->refcount)) 1475 disable_device(ib_dev); 1478 free_netdevs(ib_dev); 1480 ib_free_port_attrs(&ib_dev->coredev); 1481 device_del(&ib_dev 1512 ib_unregister_device(struct ib_device *ib_dev) argument 1534 ib_unregister_device_and_put(struct ib_device *ib_dev) argument 1560 struct ib_device *ib_dev; local 1583 struct ib_device *ib_dev = local 1601 ib_unregister_device_queued(struct ib_device *ib_dev) argument 2141 ib_device_set_netdev(struct ib_device *ib_dev, struct net_device *ndev, u32 port) argument 2184 free_netdevs(struct ib_device *ib_dev) argument 2217 ib_device_get_netdev(struct ib_device *ib_dev, u32 port) argument 2299 ib_enum_roce_netdev(struct ib_device *ib_dev, roce_netdev_filter filter, void *filter_cookie, roce_netdev_callback cb, void *cookie) argument [all...] |
H A D | uverbs_std_types_cq.c | 67 struct ib_device *ib_dev = attrs->context->device; local 75 if (!ib_dev->ops.create_cq || !ib_dev->ops.destroy_cq) 115 cq = rdma_zalloc_drv_obj(ib_dev, ib_cq); 121 cq->device = ib_dev; 131 ret = ib_dev->ops.create_cq(cq, &attr, &attrs->driver_udata);
|
H A D | uverbs_main.c | 105 if (!srcu_dereference(ufile->device->ib_dev, 199 struct ib_device *ib_dev; local 205 ib_dev = srcu_dereference(file->device->ib_dev, 207 if (ib_dev && !ib_dev->ops.disassociate_ucontext) 208 module_put(ib_dev->ops.owner); 493 struct ib_device *ib_dev = async_file->uobj.context->device; local 506 INIT_IB_EVENT_HANDLER(&async_file->event_handler, ib_dev, 897 struct ib_device *ib_dev; local 1053 struct ib_device *ib_dev; local 1072 struct ib_device *ib_dev; local 1185 ib_uverbs_free_hw_resources(struct ib_uverbs_device *uverbs_dev, struct ib_device *ib_dev) argument [all...] |
H A D | core_priv.h | 90 struct net_device *ib_device_get_netdev(struct ib_device *ib_dev, 93 void ib_enum_roce_netdev(struct ib_device *ib_dev, 129 void ib_cache_gid_set_default_gid(struct ib_device *ib_dev, u32 port, 134 int ib_cache_gid_add(struct ib_device *ib_dev, u32 port, 137 int ib_cache_gid_del(struct ib_device *ib_dev, u32 port, 140 int ib_cache_gid_del_all_netdev_gids(struct ib_device *ib_dev, u32 port, 146 unsigned long roce_gid_type_mask_support(struct ib_device *ib_dev, u32 port);
|
/linux-master/drivers/target/ |
H A D | target_core_iblock.c | 60 struct iblock_dev *ib_dev = NULL; local 62 ib_dev = kzalloc(sizeof(struct iblock_dev), GFP_KERNEL); 63 if (!ib_dev) { 68 ib_dev->ibd_plug = kcalloc(nr_cpu_ids, sizeof(*ib_dev->ibd_plug), 70 if (!ib_dev->ibd_plug) 73 pr_debug( "IBLOCK: Allocated ib_dev for %s\n", name); 75 return &ib_dev->dev; 78 kfree(ib_dev); 84 struct iblock_dev *ib_dev local 92 struct iblock_dev *ib_dev = IBLOCK_DEV(dev); local 193 struct iblock_dev *ib_dev = IBLOCK_DEV(dev); local 206 struct iblock_dev *ib_dev = IBLOCK_DEV(dev); local 215 struct iblock_dev *ib_dev = IBLOCK_DEV(se_dev); local 242 struct iblock_dev *ib_dev = IBLOCK_DEV(dev); local 360 struct iblock_dev *ib_dev = IBLOCK_DEV(cmd->se_dev); local 419 struct iblock_dev *ib_dev = IBLOCK_DEV(cmd->se_dev); local 585 struct iblock_dev *ib_dev = IBLOCK_DEV(dev); local 649 struct iblock_dev *ib_dev = IBLOCK_DEV(dev); local 679 struct iblock_dev *ib_dev = IBLOCK_DEV(dev); local 744 struct iblock_dev *ib_dev = IBLOCK_DEV(dev); local 848 struct iblock_dev *ib_dev = IBLOCK_DEV(dev); local 967 struct iblock_dev *ib_dev = IBLOCK_DEV(dev); local 1036 struct iblock_dev *ib_dev = IBLOCK_DEV(dev); local 1098 struct iblock_dev *ib_dev = IBLOCK_DEV(dev); local 1112 struct iblock_dev *ib_dev = IBLOCK_DEV(dev); local 1122 struct iblock_dev *ib_dev = IBLOCK_DEV(dev); local 1130 struct iblock_dev *ib_dev = IBLOCK_DEV(dev); local [all...] |
/linux-master/drivers/infiniband/hw/usnic/ |
H A D | usnic_ib_main.c | 79 return scnprintf(buf, buf_sz, "PF: %s ", dev_name(&vf->pf->ib_dev.dev)); 147 usnic_info("PF Reset on %s\n", dev_name(&us_ibdev->ib_dev.dev)); 150 ib_event.device = &us_ibdev->ib_dev; 161 dev_name(&us_ibdev->ib_dev.dev)); 163 ib_event.device = &us_ibdev->ib_dev; 170 dev_name(&us_ibdev->ib_dev.dev)); 173 ib_event.device = &us_ibdev->ib_dev; 179 dev_name(&us_ibdev->ib_dev.dev)); 186 dev_name(&us_ibdev->ib_dev.dev)); 189 dev_name(&us_ibdev->ib_dev [all...] |
H A D | usnic_ib_sysfs.c | 52 rdma_device_to_drv_device(device, struct usnic_ib_dev, ib_dev); 70 rdma_device_to_drv_device(device, struct usnic_ib_dev, ib_dev); 87 dev_name(&us_ibdev->ib_dev.dev), 109 dev_name(&us_ibdev->ib_dev.dev)); 122 rdma_device_to_drv_device(device, struct usnic_ib_dev, ib_dev); 132 rdma_device_to_drv_device(device, struct usnic_ib_dev, ib_dev); 142 rdma_device_to_drv_device(device, struct usnic_ib_dev, ib_dev); 156 rdma_device_to_drv_device(device, struct usnic_ib_dev, ib_dev); 255 kobject_get(&us_ibdev->ib_dev.dev.kobj); 257 &us_ibdev->ib_dev [all...] |
H A D | usnic_ib.h | 74 struct ib_device ib_dev; member in struct:usnic_ib_dev 103 return container_of(ibdev, struct usnic_ib_dev, ib_dev);
|
/linux-master/drivers/infiniband/hw/hns/ |
H A D | hns_roce_pd.c | 47 struct ib_device *ib_dev = ibpd->device; local 48 struct hns_roce_dev *hr_dev = to_hr_dev(ib_dev); 57 ibdev_err(ib_dev, "failed to alloc pd, id = %d.\n", id); 69 ibdev_err(ib_dev, "failed to copy to udata, ret = %d\n", ret); 94 ibdev_err(&hr_dev->ib_dev, "failed to alloc uar id(%d).\n", id); 129 ibdev_err(&hr_dev->ib_dev, "failed to alloc xrcdn(%d).\n", id);
|
H A D | hns_roce_main.c | 163 static int hns_roce_query_device(struct ib_device *ib_dev, argument 167 struct hns_roce_dev *hr_dev = to_hr_dev(ib_dev); 213 static int hns_roce_query_port(struct ib_device *ib_dev, u32 port_num, argument 216 struct hns_roce_dev *hr_dev = to_hr_dev(ib_dev); 235 ret = ib_get_eth_speed(ib_dev, port_num, &props->active_speed, 238 ibdev_warn(ib_dev, "failed to get speed, ret = %d.\n", ret); 269 static int hns_roce_query_pkey(struct ib_device *ib_dev, u32 port, u16 index, argument 280 static int hns_roce_modify_device(struct ib_device *ib_dev, int mask, argument 289 spin_lock_irqsave(&to_hr_dev(ib_dev)->sm_lock, flags); 290 memcpy(ib_dev 491 hns_roce_port_immutable(struct ib_device *ib_dev, u32 port_num, struct ib_port_immutable *immutable) argument 697 struct ib_device *ib_dev = NULL; local [all...] |
/linux-master/include/rdma/ |
H A D | uverbs_std_types.h | 102 struct uverbs_attr_bundle *attrs, struct ib_device **ib_dev) 107 *ib_dev = attrs->context->device; 116 struct ib_device *ib_dev, 120 action->device = ib_dev; 101 __uobj_alloc(const struct uverbs_api_object *obj, struct uverbs_attr_bundle *attrs, struct ib_device **ib_dev) argument 114 uverbs_flow_action_fill_action(struct ib_flow_action *action, struct ib_uobject *uobj, struct ib_device *ib_dev, enum ib_flow_action_type type) argument
|
/linux-master/drivers/infiniband/hw/mlx5/ |
H A D | devx.h | 13 struct mlx5_ib_dev *ib_dev; member in struct:devx_obj
|
/linux-master/drivers/infiniband/ulp/isert/ |
H A D | ib_isert.c | 108 struct ib_device *ib_dev = device->ib_device; local 112 isert_conn->cq = ib_cq_pool_get(ib_dev, cq_size, -1, IB_POLL_WORKQUEUE); 152 struct ib_device *ib_dev = device->ib_device; local 167 dma_addr = ib_dma_map_single(ib_dev, rx_desc->buf, 169 if (ib_dma_mapping_error(ib_dev, dma_addr)) 186 ib_dma_unmap_single(ib_dev, rx_desc->dma_addr, 198 struct ib_device *ib_dev = isert_conn->device->ib_device; local 207 ib_dma_unmap_single(ib_dev, rx_desc->dma_addr, 218 struct ib_device *ib_dev = device->ib_device; local 222 ib_dev 322 struct ib_device *ib_dev = isert_conn->device->ib_device; local 334 isert_alloc_login_buf(struct isert_conn *isert_conn, struct ib_device *ib_dev) argument 785 struct ib_device *ib_dev = isert_conn->cm_id->device; local 830 struct ib_device *ib_dev = device->ib_device; local 843 struct ib_device *ib_dev = device->ib_device; local 922 struct ib_device *ib_dev = device->ib_device; local 1322 struct ib_device *ib_dev = isert_conn->cm_id->device; local 1381 struct ib_device *ib_dev = isert_conn->device->ib_device; local 1506 isert_unmap_tx_desc(struct iser_tx_desc *tx_desc, struct ib_device *ib_dev) argument 1517 isert_completion_put(struct iser_tx_desc *tx_desc, struct isert_cmd *isert_cmd, struct ib_device *ib_dev, bool comp_err) argument 1668 struct ib_device *ib_dev = isert_conn->cm_id->device; local 1697 struct ib_device *ib_dev = isert_conn->cm_id->device; local 1713 struct ib_device *ib_dev = isert_conn->cm_id->device; local 1780 struct ib_device *ib_dev = device->ib_device; local 1911 struct ib_device *ib_dev = device->ib_device; local 1960 struct ib_device *ib_dev = device->ib_device; local [all...] |
/linux-master/drivers/infiniband/hw/mlx4/ |
H A D | mad.c | 199 ah_attr.type = rdma_ah_find_type(&dev->ib_dev, port_num); 454 return ib_find_cached_pkey(&dev->ib_dev, port, pkey, ix); 464 ret = ib_get_cached_pkey(&dev->ib_dev, port, pkey_ix, &slot_pkey); 550 ret = ib_get_cached_pkey(&dev->ib_dev, port, wc->pkey_index, &cached_pkey); 577 attr.type = rdma_ah_find_type(&dev->ib_dev, port); 607 ib_dma_sync_single_for_cpu(&dev->ib_dev, 649 ib_dma_sync_single_for_device(&dev->ib_dev, 1031 ll = rdma_port_get_link_layer(&dev->ib_dev, p + 1); 1034 agent = ib_register_mad_agent(&dev->ib_dev, p + 1, 1159 mlx4_ib_warn(&dev->ib_dev, "Faile [all...] |
/linux-master/drivers/infiniband/hw/vmw_pvrdma/ |
H A D | pvrdma_main.c | 103 container_of(device, struct pvrdma_dev, ib_dev); 205 dev->ib_dev.node_guid = dev->dsr->caps.node_guid; 208 dev->ib_dev.num_comp_vectors = 1; 209 dev->ib_dev.dev.parent = &dev->pdev->dev; 211 dev->ib_dev.node_type = RDMA_NODE_IB_CA; 212 dev->ib_dev.phys_port_cnt = dev->dsr->caps.phys_port_cnt; 214 ib_set_device_ops(&dev->ib_dev, &pvrdma_dev_ops); 233 ib_set_device_ops(&dev->ib_dev, &pvrdma_dev_srq_ops); 241 ret = ib_device_set_netdev(&dev->ib_dev, dev->netdev, 1); 246 ret = ib_register_device(&dev->ib_dev, "vmw_pvrdm [all...] |