Searched refs:ib_dev (Results 1 - 25 of 106) sorted by relevance

12345

/linux-master/drivers/infiniband/hw/mana/
H A Ddevice.c60 dev = ib_alloc_device(mana_ib_dev, ib_dev);
64 ib_set_device_ops(&dev->ib_dev, &mana_ib_dev_ops);
66 dev->ib_dev.phys_port_cnt = mc->num_ports;
68 ibdev_dbg(&dev->ib_dev, "mdev=%p id=%d num_ports=%d\n", mdev,
69 mdev->dev_id.as_uint32, dev->ib_dev.phys_port_cnt);
71 dev->ib_dev.node_type = RDMA_NODE_IB_CA;
77 dev->ib_dev.num_comp_vectors = 1;
78 dev->ib_dev.dev.parent = mdev->gdma_context->dev;
82 ibdev_err(&dev->ib_dev, "Failed to register device, ret %d",
90 ibdev_err(&dev->ib_dev, "Faile
[all...]
H A Dwq.c13 container_of(pd->device, struct mana_ib_dev, ib_dev);
24 ibdev_dbg(&mdev->ib_dev,
33 ibdev_dbg(&mdev->ib_dev, "ucmd wq_buf_addr 0x%llx\n", ucmd.wq_buf_addr);
39 ibdev_dbg(&mdev->ib_dev,
51 ibdev_dbg(&mdev->ib_dev,
57 ibdev_dbg(&mdev->ib_dev,
84 struct ib_device *ib_dev = ibwq->device; local
87 mdev = container_of(ib_dev, struct mana_ib_dev, ib_dev);
H A Dqp.c56 ibdev_dbg(&dev->ib_dev, "ind table size %u\n", 1 << log_ind_tbl_size);
59 ibdev_dbg(&dev->ib_dev, "index %u handle 0x%llx\n", i,
69 ibdev_dbg(&dev->ib_dev, "vport handle %llu default_rxobj 0x%llx\n",
99 container_of(pd->device, struct mana_ib_dev, ib_dev);
123 ibdev_dbg(&mdev->ib_dev,
130 ibdev_dbg(&mdev->ib_dev,
137 ibdev_dbg(&mdev->ib_dev,
145 ibdev_dbg(&mdev->ib_dev,
152 ibdev_dbg(&mdev->ib_dev,
162 ibdev_dbg(&mdev->ib_dev, "Invali
[all...]
H A Dmain.c14 ndev = mana_ib_get_netdev(&dev->ib_dev, port);
35 ndev = mana_ib_get_netdev(&dev->ib_dev, port);
42 ibdev_dbg(&dev->ib_dev,
53 ibdev_dbg(&dev->ib_dev, "Failed to configure vPort %d\n", err);
62 ibdev_dbg(&dev->ib_dev, "vport handle %llx pdid %x doorbell_id %x\n",
79 dev = container_of(ibdev, struct mana_ib_dev, ib_dev);
90 ibdev_dbg(&dev->ib_dev,
101 ibdev_dbg(&dev->ib_dev, "pd_handle 0x%llx pd_id %d\n",
119 dev = container_of(ibdev, struct mana_ib_dev, ib_dev);
130 ibdev_dbg(&dev->ib_dev,
[all...]
/linux-master/drivers/infiniband/core/
H A Droce_gid_mgmt.c82 unsigned long roce_gid_type_mask_support(struct ib_device *ib_dev, u32 port) argument
87 if (!rdma_protocol_roce(ib_dev, port))
91 if (PORT_CAP_TO_GID_TYPE[i].is_supported(ib_dev, port))
98 static void update_gid(enum gid_op_type gid_op, struct ib_device *ib_dev, argument
103 unsigned long gid_type_mask = roce_gid_type_mask_support(ib_dev, port);
110 ib_cache_gid_add(ib_dev, port,
114 ib_cache_gid_del(ib_dev, port,
147 is_eth_port_of_netdev_filter(struct ib_device *ib_dev, u32 port, argument
171 is_eth_port_inactive_slave_filter(struct ib_device *ib_dev, u32 port, argument
192 * @ib_dev
201 is_ndev_for_default_gid_filter(struct ib_device *ib_dev, u32 port, struct net_device *rdma_ndev, void *cookie) argument
227 pass_all_filter(struct ib_device *ib_dev, u32 port, struct net_device *rdma_ndev, void *cookie) argument
233 upper_device_filter(struct ib_device *ib_dev, u32 port, struct net_device *rdma_ndev, void *cookie) argument
264 is_upper_ndev_bond_master_filter(struct ib_device *ib_dev, u32 port, struct net_device *rdma_ndev, void *cookie) argument
282 update_gid_ip(enum gid_op_type gid_op, struct ib_device *ib_dev, u32 port, struct net_device *ndev, struct sockaddr *addr) argument
297 bond_delete_netdev_default_gids(struct ib_device *ib_dev, u32 port, struct net_device *rdma_ndev, struct net_device *event_ndev) argument
331 enum_netdev_ipv4_ips(struct ib_device *ib_dev, u32 port, struct net_device *ndev) argument
375 enum_netdev_ipv6_ips(struct ib_device *ib_dev, u32 port, struct net_device *ndev) argument
421 _add_netdev_ips(struct ib_device *ib_dev, u32 port, struct net_device *ndev) argument
429 add_netdev_ips(struct ib_device *ib_dev, u32 port, struct net_device *rdma_ndev, void *cookie) argument
435 del_netdev_ips(struct ib_device *ib_dev, u32 port, struct net_device *rdma_ndev, void *cookie) argument
450 del_default_gids(struct ib_device *ib_dev, u32 port, struct net_device *rdma_ndev, void *cookie) argument
462 add_default_gids(struct ib_device *ib_dev, u32 port, struct net_device *rdma_ndev, void *cookie) argument
473 enum_all_gids_of_dev_cb(struct ib_device *ib_dev, u32 port, struct net_device *rdma_ndev, void *cookie) argument
511 rdma_roce_rescan_device(struct ib_device *ib_dev) argument
551 handle_netdev_upper(struct ib_device *ib_dev, u32 port, void *cookie, void (*handle_netdev)(struct ib_device *ib_dev, u32 port, struct net_device *ndev)) argument
578 _roce_del_all_netdev_gids(struct ib_device *ib_dev, u32 port, struct net_device *event_ndev) argument
584 del_netdev_upper_ips(struct ib_device *ib_dev, u32 port, struct net_device *rdma_ndev, void *cookie) argument
590 add_netdev_upper_ips(struct ib_device *ib_dev, u32 port, struct net_device *rdma_ndev, void *cookie) argument
596 del_netdev_default_ips_join(struct ib_device *ib_dev, u32 port, struct net_device *rdma_ndev, void *cookie) argument
[all...]
H A Duverbs_std_types_dm.c56 struct ib_device *ib_dev = attrs->context->device; local
60 if (!ib_dev->ops.alloc_dm)
73 dm = ib_dev->ops.alloc_dm(ib_dev, attrs->context, &attr, attrs);
77 dm->device = ib_dev;
H A Duverbs_std_types_counters.c60 struct ib_device *ib_dev = attrs->context->device; local
69 if (!ib_dev->ops.create_counters)
72 counters = rdma_zalloc_drv_obj(ib_dev, ib_counters);
76 counters->device = ib_dev;
81 ret = ib_dev->ops.create_counters(counters, attrs);
H A Duverbs_std_types_device.c141 struct ib_device *ib_dev, u8 port_num)
153 if (rdma_is_grh_required(ib_dev, port_num))
156 if (rdma_cap_opa_ah(ib_dev, port_num)) {
173 resp->link_layer = rdma_port_get_link_layer(ib_dev, port_num);
179 struct ib_device *ib_dev; local
189 ib_dev = ucontext->device;
192 if (!ib_dev->ops.query_port)
200 ret = ib_query_port(ib_dev, port_num, &attr);
204 copy_port_attr_to_resp(&attr, &resp.legacy_resp, ib_dev, port_num);
246 struct ib_device *ib_dev; local
139 copy_port_attr_to_resp(struct ib_port_attr *attr, struct ib_uverbs_query_port_resp *resp, struct ib_device *ib_dev, u8 port_num) argument
318 struct ib_device *ib_dev; local
374 struct ib_device *ib_dev; local
[all...]
H A Dcache.c124 static void dispatch_gid_change_event(struct ib_device *ib_dev, u32 port) argument
128 event.device = ib_dev;
376 * @ib_dev: IB device whose GID entry to be deleted
382 static void del_gid(struct ib_device *ib_dev, u32 port, argument
390 dev_dbg(&ib_dev->dev, "%s port=%u index=%d gid %pI6\n", __func__, port,
399 if (!rdma_protocol_roce(ib_dev, port))
403 if (rdma_cap_roce_gid_table(ib_dev, port))
404 ib_dev->ops.del_gid(&entry->attr, &entry->context);
546 static int __ib_cache_gid_add(struct ib_device *ib_dev, u32 port, argument
562 table = rdma_gid_table(ib_dev, por
590 ib_cache_gid_add(struct ib_device *ib_dev, u32 port, union ib_gid *gid, struct ib_gid_attr *attr) argument
601 _ib_cache_gid_del(struct ib_device *ib_dev, u32 port, union ib_gid *gid, struct ib_gid_attr *attr, unsigned long mask, bool default_gid) argument
630 ib_cache_gid_del(struct ib_device *ib_dev, u32 port, union ib_gid *gid, struct ib_gid_attr *attr) argument
641 ib_cache_gid_del_all_netdev_gids(struct ib_device *ib_dev, u32 port, struct net_device *ndev) argument
683 rdma_find_gid_by_port(struct ib_device *ib_dev, const union ib_gid *gid, enum ib_gid_type gid_type, u32 port, struct net_device *ndev) argument
736 rdma_find_gid_by_filter( struct ib_device *ib_dev, const union ib_gid *gid, u32 port, bool (*filter)(const union ib_gid *gid, const struct ib_gid_attr *, void *), void *context) argument
821 cleanup_gid_table_port(struct ib_device *ib_dev, u32 port, struct ib_gid_table *table) argument
837 ib_cache_gid_set_default_gid(struct ib_device *ib_dev, u32 port, struct net_device *ndev, unsigned long gid_type_mask, enum ib_cache_gid_default_mode mode) argument
870 gid_table_reserve_default(struct ib_device *ib_dev, u32 port, struct ib_gid_table *table) argument
885 gid_table_release_one(struct ib_device *ib_dev) argument
895 _gid_table_setup_one(struct ib_device *ib_dev) argument
916 gid_table_cleanup_one(struct ib_device *ib_dev) argument
925 gid_table_setup_one(struct ib_device *ib_dev) argument
[all...]
H A Duverbs_std_types_mr.c53 struct ib_device *ib_dev = pd->device; local
60 if (!ib_dev->ops.advise_mr)
79 return ib_dev->ops.advise_mr(pd, advice, flags, sg_list, num_sge,
93 struct ib_device *ib_dev = pd->device; local
98 if (!ib_dev->ops.reg_dm_mr)
119 ret = ib_check_mr_access(ib_dev, attr.access_flags);
193 struct ib_device *ib_dev = pd->device; local
200 if (!ib_dev->ops.reg_user_mr_dmabuf)
236 ret = ib_check_mr_access(ib_dev, access_flags);
H A Ddevice.c186 static void free_netdevs(struct ib_device *ib_dev);
821 pdata->ib_dev = device;
1344 static void prevent_dealloc_device(struct ib_device *ib_dev) argument
1462 static void __ib_unregister_device(struct ib_device *ib_dev) argument
1471 mutex_lock(&ib_dev->unregistration_lock);
1472 if (!refcount_read(&ib_dev->refcount))
1475 disable_device(ib_dev);
1478 free_netdevs(ib_dev);
1480 ib_free_port_attrs(&ib_dev->coredev);
1481 device_del(&ib_dev
1512 ib_unregister_device(struct ib_device *ib_dev) argument
1534 ib_unregister_device_and_put(struct ib_device *ib_dev) argument
1560 struct ib_device *ib_dev; local
1583 struct ib_device *ib_dev = local
1601 ib_unregister_device_queued(struct ib_device *ib_dev) argument
2141 ib_device_set_netdev(struct ib_device *ib_dev, struct net_device *ndev, u32 port) argument
2184 free_netdevs(struct ib_device *ib_dev) argument
2217 ib_device_get_netdev(struct ib_device *ib_dev, u32 port) argument
2299 ib_enum_roce_netdev(struct ib_device *ib_dev, roce_netdev_filter filter, void *filter_cookie, roce_netdev_callback cb, void *cookie) argument
[all...]
H A Duverbs_std_types_cq.c67 struct ib_device *ib_dev = attrs->context->device; local
75 if (!ib_dev->ops.create_cq || !ib_dev->ops.destroy_cq)
115 cq = rdma_zalloc_drv_obj(ib_dev, ib_cq);
121 cq->device = ib_dev;
131 ret = ib_dev->ops.create_cq(cq, &attr, &attrs->driver_udata);
H A Duverbs_main.c105 if (!srcu_dereference(ufile->device->ib_dev,
199 struct ib_device *ib_dev; local
205 ib_dev = srcu_dereference(file->device->ib_dev,
207 if (ib_dev && !ib_dev->ops.disassociate_ucontext)
208 module_put(ib_dev->ops.owner);
493 struct ib_device *ib_dev = async_file->uobj.context->device; local
506 INIT_IB_EVENT_HANDLER(&async_file->event_handler, ib_dev,
897 struct ib_device *ib_dev; local
1053 struct ib_device *ib_dev; local
1072 struct ib_device *ib_dev; local
1185 ib_uverbs_free_hw_resources(struct ib_uverbs_device *uverbs_dev, struct ib_device *ib_dev) argument
[all...]
H A Dcore_priv.h90 struct net_device *ib_device_get_netdev(struct ib_device *ib_dev,
93 void ib_enum_roce_netdev(struct ib_device *ib_dev,
129 void ib_cache_gid_set_default_gid(struct ib_device *ib_dev, u32 port,
134 int ib_cache_gid_add(struct ib_device *ib_dev, u32 port,
137 int ib_cache_gid_del(struct ib_device *ib_dev, u32 port,
140 int ib_cache_gid_del_all_netdev_gids(struct ib_device *ib_dev, u32 port,
146 unsigned long roce_gid_type_mask_support(struct ib_device *ib_dev, u32 port);
/linux-master/drivers/target/
H A Dtarget_core_iblock.c60 struct iblock_dev *ib_dev = NULL; local
62 ib_dev = kzalloc(sizeof(struct iblock_dev), GFP_KERNEL);
63 if (!ib_dev) {
68 ib_dev->ibd_plug = kcalloc(nr_cpu_ids, sizeof(*ib_dev->ibd_plug),
70 if (!ib_dev->ibd_plug)
73 pr_debug( "IBLOCK: Allocated ib_dev for %s\n", name);
75 return &ib_dev->dev;
78 kfree(ib_dev);
84 struct iblock_dev *ib_dev local
92 struct iblock_dev *ib_dev = IBLOCK_DEV(dev); local
193 struct iblock_dev *ib_dev = IBLOCK_DEV(dev); local
206 struct iblock_dev *ib_dev = IBLOCK_DEV(dev); local
215 struct iblock_dev *ib_dev = IBLOCK_DEV(se_dev); local
242 struct iblock_dev *ib_dev = IBLOCK_DEV(dev); local
360 struct iblock_dev *ib_dev = IBLOCK_DEV(cmd->se_dev); local
419 struct iblock_dev *ib_dev = IBLOCK_DEV(cmd->se_dev); local
585 struct iblock_dev *ib_dev = IBLOCK_DEV(dev); local
649 struct iblock_dev *ib_dev = IBLOCK_DEV(dev); local
679 struct iblock_dev *ib_dev = IBLOCK_DEV(dev); local
744 struct iblock_dev *ib_dev = IBLOCK_DEV(dev); local
848 struct iblock_dev *ib_dev = IBLOCK_DEV(dev); local
967 struct iblock_dev *ib_dev = IBLOCK_DEV(dev); local
1036 struct iblock_dev *ib_dev = IBLOCK_DEV(dev); local
1098 struct iblock_dev *ib_dev = IBLOCK_DEV(dev); local
1112 struct iblock_dev *ib_dev = IBLOCK_DEV(dev); local
1122 struct iblock_dev *ib_dev = IBLOCK_DEV(dev); local
1130 struct iblock_dev *ib_dev = IBLOCK_DEV(dev); local
[all...]
/linux-master/drivers/infiniband/hw/usnic/
H A Dusnic_ib_main.c79 return scnprintf(buf, buf_sz, "PF: %s ", dev_name(&vf->pf->ib_dev.dev));
147 usnic_info("PF Reset on %s\n", dev_name(&us_ibdev->ib_dev.dev));
150 ib_event.device = &us_ibdev->ib_dev;
161 dev_name(&us_ibdev->ib_dev.dev));
163 ib_event.device = &us_ibdev->ib_dev;
170 dev_name(&us_ibdev->ib_dev.dev));
173 ib_event.device = &us_ibdev->ib_dev;
179 dev_name(&us_ibdev->ib_dev.dev));
186 dev_name(&us_ibdev->ib_dev.dev));
189 dev_name(&us_ibdev->ib_dev
[all...]
H A Dusnic_ib_sysfs.c52 rdma_device_to_drv_device(device, struct usnic_ib_dev, ib_dev);
70 rdma_device_to_drv_device(device, struct usnic_ib_dev, ib_dev);
87 dev_name(&us_ibdev->ib_dev.dev),
109 dev_name(&us_ibdev->ib_dev.dev));
122 rdma_device_to_drv_device(device, struct usnic_ib_dev, ib_dev);
132 rdma_device_to_drv_device(device, struct usnic_ib_dev, ib_dev);
142 rdma_device_to_drv_device(device, struct usnic_ib_dev, ib_dev);
156 rdma_device_to_drv_device(device, struct usnic_ib_dev, ib_dev);
255 kobject_get(&us_ibdev->ib_dev.dev.kobj);
257 &us_ibdev->ib_dev
[all...]
H A Dusnic_ib.h74 struct ib_device ib_dev; member in struct:usnic_ib_dev
103 return container_of(ibdev, struct usnic_ib_dev, ib_dev);
/linux-master/drivers/infiniband/hw/hns/
H A Dhns_roce_pd.c47 struct ib_device *ib_dev = ibpd->device; local
48 struct hns_roce_dev *hr_dev = to_hr_dev(ib_dev);
57 ibdev_err(ib_dev, "failed to alloc pd, id = %d.\n", id);
69 ibdev_err(ib_dev, "failed to copy to udata, ret = %d\n", ret);
94 ibdev_err(&hr_dev->ib_dev, "failed to alloc uar id(%d).\n", id);
129 ibdev_err(&hr_dev->ib_dev, "failed to alloc xrcdn(%d).\n", id);
H A Dhns_roce_main.c163 static int hns_roce_query_device(struct ib_device *ib_dev, argument
167 struct hns_roce_dev *hr_dev = to_hr_dev(ib_dev);
213 static int hns_roce_query_port(struct ib_device *ib_dev, u32 port_num, argument
216 struct hns_roce_dev *hr_dev = to_hr_dev(ib_dev);
235 ret = ib_get_eth_speed(ib_dev, port_num, &props->active_speed,
238 ibdev_warn(ib_dev, "failed to get speed, ret = %d.\n", ret);
269 static int hns_roce_query_pkey(struct ib_device *ib_dev, u32 port, u16 index, argument
280 static int hns_roce_modify_device(struct ib_device *ib_dev, int mask, argument
289 spin_lock_irqsave(&to_hr_dev(ib_dev)->sm_lock, flags);
290 memcpy(ib_dev
491 hns_roce_port_immutable(struct ib_device *ib_dev, u32 port_num, struct ib_port_immutable *immutable) argument
697 struct ib_device *ib_dev = NULL; local
[all...]
/linux-master/include/rdma/
H A Duverbs_std_types.h102 struct uverbs_attr_bundle *attrs, struct ib_device **ib_dev)
107 *ib_dev = attrs->context->device;
116 struct ib_device *ib_dev,
120 action->device = ib_dev;
101 __uobj_alloc(const struct uverbs_api_object *obj, struct uverbs_attr_bundle *attrs, struct ib_device **ib_dev) argument
114 uverbs_flow_action_fill_action(struct ib_flow_action *action, struct ib_uobject *uobj, struct ib_device *ib_dev, enum ib_flow_action_type type) argument
/linux-master/drivers/infiniband/hw/mlx5/
H A Ddevx.h13 struct mlx5_ib_dev *ib_dev; member in struct:devx_obj
/linux-master/drivers/infiniband/ulp/isert/
H A Dib_isert.c108 struct ib_device *ib_dev = device->ib_device; local
112 isert_conn->cq = ib_cq_pool_get(ib_dev, cq_size, -1, IB_POLL_WORKQUEUE);
152 struct ib_device *ib_dev = device->ib_device; local
167 dma_addr = ib_dma_map_single(ib_dev, rx_desc->buf,
169 if (ib_dma_mapping_error(ib_dev, dma_addr))
186 ib_dma_unmap_single(ib_dev, rx_desc->dma_addr,
198 struct ib_device *ib_dev = isert_conn->device->ib_device; local
207 ib_dma_unmap_single(ib_dev, rx_desc->dma_addr,
218 struct ib_device *ib_dev = device->ib_device; local
222 ib_dev
322 struct ib_device *ib_dev = isert_conn->device->ib_device; local
334 isert_alloc_login_buf(struct isert_conn *isert_conn, struct ib_device *ib_dev) argument
785 struct ib_device *ib_dev = isert_conn->cm_id->device; local
830 struct ib_device *ib_dev = device->ib_device; local
843 struct ib_device *ib_dev = device->ib_device; local
922 struct ib_device *ib_dev = device->ib_device; local
1322 struct ib_device *ib_dev = isert_conn->cm_id->device; local
1381 struct ib_device *ib_dev = isert_conn->device->ib_device; local
1506 isert_unmap_tx_desc(struct iser_tx_desc *tx_desc, struct ib_device *ib_dev) argument
1517 isert_completion_put(struct iser_tx_desc *tx_desc, struct isert_cmd *isert_cmd, struct ib_device *ib_dev, bool comp_err) argument
1668 struct ib_device *ib_dev = isert_conn->cm_id->device; local
1697 struct ib_device *ib_dev = isert_conn->cm_id->device; local
1713 struct ib_device *ib_dev = isert_conn->cm_id->device; local
1780 struct ib_device *ib_dev = device->ib_device; local
1911 struct ib_device *ib_dev = device->ib_device; local
1960 struct ib_device *ib_dev = device->ib_device; local
[all...]
/linux-master/drivers/infiniband/hw/mlx4/
H A Dmad.c199 ah_attr.type = rdma_ah_find_type(&dev->ib_dev, port_num);
454 return ib_find_cached_pkey(&dev->ib_dev, port, pkey, ix);
464 ret = ib_get_cached_pkey(&dev->ib_dev, port, pkey_ix, &slot_pkey);
550 ret = ib_get_cached_pkey(&dev->ib_dev, port, wc->pkey_index, &cached_pkey);
577 attr.type = rdma_ah_find_type(&dev->ib_dev, port);
607 ib_dma_sync_single_for_cpu(&dev->ib_dev,
649 ib_dma_sync_single_for_device(&dev->ib_dev,
1031 ll = rdma_port_get_link_layer(&dev->ib_dev, p + 1);
1034 agent = ib_register_mad_agent(&dev->ib_dev, p + 1,
1159 mlx4_ib_warn(&dev->ib_dev, "Faile
[all...]
/linux-master/drivers/infiniband/hw/vmw_pvrdma/
H A Dpvrdma_main.c103 container_of(device, struct pvrdma_dev, ib_dev);
205 dev->ib_dev.node_guid = dev->dsr->caps.node_guid;
208 dev->ib_dev.num_comp_vectors = 1;
209 dev->ib_dev.dev.parent = &dev->pdev->dev;
211 dev->ib_dev.node_type = RDMA_NODE_IB_CA;
212 dev->ib_dev.phys_port_cnt = dev->dsr->caps.phys_port_cnt;
214 ib_set_device_ops(&dev->ib_dev, &pvrdma_dev_ops);
233 ib_set_device_ops(&dev->ib_dev, &pvrdma_dev_srq_ops);
241 ret = ib_device_set_netdev(&dev->ib_dev, dev->netdev, 1);
246 ret = ib_register_device(&dev->ib_dev, "vmw_pvrdm
[all...]

Completed in 494 milliseconds

12345