/linux-master/drivers/gpu/drm/xe/ |
H A D | xe_sriov.c | 50 xe_assert(xe, !xe->sriov.__mode); 51 xe->sriov.__mode = mode; 52 xe_assert(xe, xe->sriov.__mode); 63 destroy_workqueue(xe->sriov.wq); 64 xe->sriov.wq = NULL; 81 xe_assert(xe, !xe->sriov.wq); 82 xe->sriov.wq = alloc_workqueue("xe-sriov-wq", 0, 0); 83 if (!xe->sriov.wq)
|
H A D | xe_sriov.h | 20 xe_assert(xe, xe->sriov.__mode); 21 return xe->sriov.__mode;
|
H A D | xe_sriov_printk.h | 15 ((xe)->sriov.__mode == XE_SRIOV_MODE_PF ? "PF: " : \ 16 (xe)->sriov.__mode == XE_SRIOV_MODE_VF ? "VF: " : "")
|
H A D | xe_device_types.h | 190 /** @sriov: tile level virtualization data */ 193 /** @sriov.pf.lmtt: Local Memory Translation Table. */ 197 /** @sriov.vf.memirq: Memory Based Interrupts. */ 200 } sriov; member in struct:xe_tile 320 /** @sriov: device level virtualization data */ 322 /** @sriov.__mode: SR-IOV mode (Don't access directly!) */ 324 /** @sriov.wq: workqueue used by the virtualization workers */ 326 } sriov; member in struct:xe_device
|
H A D | xe_irq.c | 565 xe_memirq_reset(&tile->sriov.vf.memirq); 608 xe_memirq_postinstall(&tile->sriov.vf.memirq); 651 xe_memirq_handler(&tile->sriov.vf.memirq);
|
H A D | xe_gt.c | 359 xe_lmtt_init(>_to_tile(gt)->sriov.pf.lmtt); 474 xe_lmtt_init_hw(>_to_tile(gt)->sriov.pf.lmtt); 615 xe_lmtt_init_hw(>_to_tile(gt)->sriov.pf.lmtt);
|
H A D | xe_memirq.c | 30 return container_of(memirq, struct xe_tile, sriov.vf.memirq);
|
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/ |
H A D | sriov.c | 42 struct mlx5_core_sriov *sriov = &dev->priv.sriov; local 46 /* Restore sriov guid and policy settings */ 47 if (sriov->vfs_ctx[vf].node_guid || 48 sriov->vfs_ctx[vf].port_guid || 49 sriov->vfs_ctx[vf].policy != MLX5_POLICY_INVALID) { 54 in->node_guid = sriov->vfs_ctx[vf].node_guid; 55 in->port_guid = sriov->vfs_ctx[vf].port_guid; 56 in->policy = sriov->vfs_ctx[vf].policy; 74 struct mlx5_core_sriov *sriov local 128 struct mlx5_core_sriov *sriov = &dev->priv.sriov; local 216 struct mlx5_core_sriov *sriov = &dev->priv.sriov; local 234 struct mlx5_core_sriov *sriov; local 297 struct mlx5_core_sriov *sriov = &dev->priv.sriov; local 320 struct mlx5_core_sriov *sriov = &dev->priv.sriov; local 341 struct mlx5_core_sriov *sriov; local 367 struct mlx5_core_sriov *sriov; local [all...] |
/linux-master/drivers/infiniband/hw/mlx4/ |
H A D | cm.c | 156 struct rb_root *sl_id_map = &to_mdev(ibdev)->sriov.sl_id_map; 183 struct mlx4_ib_sriov *sriov = &dev->sriov; local 184 struct rb_root *sl_id_map = &sriov->sl_id_map; 186 spin_lock(&sriov->id_map_lock); 187 if (!xa_erase(&sriov->pv_id_table, ent->pv_cm_id)) 195 spin_unlock(&sriov->id_map_lock); 201 struct rb_root *sl_id_map = &to_mdev(ibdev)->sriov.sl_id_map; 236 struct mlx4_ib_sriov *sriov = &to_mdev(ibdev)->sriov; local 268 struct mlx4_ib_sriov *sriov = &to_mdev(ibdev)->sriov; local 284 struct mlx4_ib_sriov *sriov = &to_mdev(ibdev)->sriov; local 360 alloc_rej_tmout(struct mlx4_ib_sriov *sriov, u32 rem_pv_cm_id, int slave) argument 406 lookup_rej_tmout_slave(struct mlx4_ib_sriov *sriov, u32 rem_pv_cm_id) argument 429 struct mlx4_ib_sriov *sriov = &to_mdev(ibdev)->sriov; local 494 rej_tmout_xa_cleanup(struct mlx4_ib_sriov *sriov, int slave) argument 525 struct mlx4_ib_sriov *sriov = &dev->sriov; local [all...] |
H A D | alias_GUID.c | 85 guid_indexes = be64_to_cpu((__force __be64) dev->sriov.alias_guid. 101 memcpy(&dev->sriov.demux[port_index].guid_cache[slave_id], 116 return *(__be64 *)&dev->sriov.demux[port - 1].guid_cache[index]; 135 spin_lock_irqsave(&dev->sriov.alias_guid.ag_work_lock, flags); 136 if (dev->sriov.alias_guid.ports_guid[port_index].state_flags & 140 curr_guid = *(__be64 *)&dev->sriov. 153 *(__be64 *)&dev->sriov.alias_guid.ports_guid[port_index]. 156 dev->sriov.alias_guid.ports_guid[port_index]. 159 dev->sriov.alias_guid.ports_guid[port_index]. 163 dev->sriov 799 struct mlx4_ib_sriov *sriov = &dev->sriov; local [all...] |
H A D | mad.c | 284 if (!dev->sriov.is_going_down) 299 !dev->sriov.is_going_down) { 439 if (dev->sriov.demux[port - 1].guid_cache[i] == guid) 536 tun_ctx = dev->sriov.demux[port-1].tun[slave]; 758 atomic64_read(&dev->sriov.demux[port - 1].subnet_prefix))) { 1082 if (mlx4_is_master(dev->dev) && !dev->sriov.is_going_down) 1093 if (!dev->sriov.is_going_down) { 1094 mlx4_ib_mcg_port_cleanup(&dev->sriov.demux[port_num - 1], 0); 1220 (u64)atomic64_read(&dev->sriov.demux[port - 1].subnet_prefix), 1222 atomic64_set(&dev->sriov [all...] |
H A D | sysfs.c | 87 spin_lock_irqsave(&mdev->sriov.alias_guid.ag_work_lock, flags); 89 *(__be64 *)&mdev->sriov.alias_guid.ports_guid[port->num - 1]. 95 mdev->sriov.alias_guid.ports_guid[port->num - 1].all_rec_per_port[record_num].status 102 mdev->sriov.alias_guid.ports_guid[port->num - 1].all_rec_per_port[record_num].guid_indexes 105 spin_unlock_irqrestore(&mdev->sriov.alias_guid.ag_work_lock, flags);
|
/linux-master/drivers/net/ethernet/qlogic/qlcnic/ |
H A D | qlcnic_sriov_pf.c | 70 struct qlcnic_sriov *sriov = adapter->ahw->sriov; local 71 struct qlcnic_resources *res = &sriov->ff_max; 72 u16 num_macs = sriov->num_allowed_vlans + 1; 81 num_vfs = sriov->num_vfs; 103 info->max_tx_ques = res->num_tx_queues - sriov->num_vfs; 108 vp = sriov->vf_info[id].vp; 139 struct qlcnic_resources *ff_max = &adapter->ahw->sriov->ff_max; 158 struct qlcnic_sriov *sriov = adapter->ahw->sriov; local 235 struct qlcnic_sriov *sriov = adapter->ahw->sriov; local 253 struct qlcnic_sriov *sriov = adapter->ahw->sriov; local 271 struct qlcnic_sriov *sriov = adapter->ahw->sriov; local 408 struct qlcnic_sriov *sriov = adapter->ahw->sriov; local 746 struct qlcnic_sriov *sriov; local 859 struct qlcnic_sriov *sriov; local 1342 struct qlcnic_sriov *sriov; local 1372 struct qlcnic_sriov *sriov = adapter->ahw->sriov; local 1397 struct qlcnic_sriov *sriov = adapter->ahw->sriov; local 1647 qlcnic_sriov_add_act_list_irqsave(struct qlcnic_sriov *sriov, struct qlcnic_vf_info *vf, struct qlcnic_bc_trans *trans) argument 1695 qlcnic_sriov_schedule_flr(struct qlcnic_sriov *sriov, struct qlcnic_vf_info *vf, work_func_t func) argument 1710 struct qlcnic_sriov *sriov = adapter->ahw->sriov; local 1737 qlcnic_sriov_pf_handle_flr(struct qlcnic_sriov *sriov, struct qlcnic_vf_info *vf) argument 1765 struct qlcnic_sriov *sriov = ahw->sriov; local 1805 struct qlcnic_sriov *sriov = adapter->ahw->sriov; local 1853 struct qlcnic_sriov *sriov = adapter->ahw->sriov; local 1916 struct qlcnic_sriov *sriov = adapter->ahw->sriov; local 1995 struct qlcnic_sriov *sriov = adapter->ahw->sriov; local 2025 struct qlcnic_sriov *sriov = adapter->ahw->sriov; local [all...] |
H A D | qlcnic_sriov_common.c | 142 struct qlcnic_sriov *sriov; local 152 sriov = kzalloc(sizeof(struct qlcnic_sriov), GFP_KERNEL); 153 if (!sriov) 156 adapter->ahw->sriov = sriov; 157 sriov->num_vfs = num_vfs; 158 bc = &sriov->bc; 159 sriov->vf_info = kcalloc(num_vfs, sizeof(struct qlcnic_vf_info), 161 if (!sriov->vf_info) { 190 vf = &sriov 263 struct qlcnic_sriov *sriov = adapter->ahw->sriov; local 436 struct qlcnic_sriov *sriov = adapter->ahw->sriov; local 469 struct qlcnic_sriov *sriov = adapter->ahw->sriov; local 849 qlcnic_sriov_schedule_bc_cmd(struct qlcnic_sriov *sriov, struct qlcnic_vf_info *vf, work_func_t func) argument 1118 __qlcnic_sriov_add_act_list(struct qlcnic_sriov *sriov, struct qlcnic_vf_info *vf, struct qlcnic_bc_trans *trans) argument 1132 qlcnic_sriov_add_act_list(struct qlcnic_sriov *sriov, struct qlcnic_vf_info *vf, struct qlcnic_bc_trans *trans) argument 1146 qlcnic_sriov_handle_pending_trans(struct qlcnic_sriov *sriov, struct qlcnic_vf_info *vf, struct qlcnic_bc_hdr *hdr) argument 1190 qlcnic_sriov_handle_bc_cmd(struct qlcnic_sriov *sriov, struct qlcnic_bc_hdr *hdr, struct qlcnic_vf_info *vf) argument 1264 qlcnic_sriov_handle_msg_event(struct qlcnic_sriov *sriov, struct qlcnic_vf_info *vf) argument 1285 qlcnic_sriov_handle_flr_event(struct qlcnic_sriov *sriov, struct qlcnic_vf_info *vf) argument 1300 struct qlcnic_sriov *sriov; local 1501 struct qlcnic_sriov *sriov = adapter->ahw->sriov; local 1949 qlcnic_sriov_check_vlan_id(struct qlcnic_sriov *sriov, struct qlcnic_vf_info *vf, u16 vlan_id) argument 1970 qlcnic_sriov_validate_num_vlans(struct qlcnic_sriov *sriov, struct qlcnic_vf_info *vf) argument 1987 struct qlcnic_sriov *sriov = adapter->ahw->sriov; local 2026 struct qlcnic_sriov *sriov; local 2053 struct qlcnic_sriov *sriov = adapter->ahw->sriov; local 2162 struct qlcnic_sriov *sriov = adapter->ahw->sriov; local 2179 struct qlcnic_sriov *sriov = adapter->ahw->sriov; local 2190 qlcnic_sriov_add_vlan_id(struct qlcnic_sriov *sriov, struct qlcnic_vf_info *vf, u16 vlan_id) argument 2204 qlcnic_sriov_del_vlan_id(struct qlcnic_sriov *sriov, struct qlcnic_vf_info *vf, u16 vlan_id) argument [all...] |
H A D | qlcnic_sriov.h | 266 static inline void qlcnic_sriov_pf_handle_flr(struct qlcnic_sriov *sriov, argument
|
/linux-master/drivers/pci/ |
H A D | iov.c | 23 return dev->bus->number + ((dev->devfn + dev->sriov->offset + 24 dev->sriov->stride * vf_id) >> 8); 31 return (dev->devfn + dev->sriov->offset + 32 dev->sriov->stride * vf_id) & 0xff; 44 return (pci_dev_id(dev) - (pci_dev_id(pf) + pf->sriov->offset)) / 45 pf->sriov->stride; 86 struct pci_sriov *iov = dev->sriov; 102 struct pci_sriov *iov = dev->sriov; 153 return dev->sriov->barsz[resno - PCI_IOV_RESOURCES]; 170 &physfn->sriov [all...] |
/linux-master/tools/testing/selftests/powerpc/eeh/ |
H A D | eeh-functions.sh | 157 if [ ! -f /proc/device-tree/rtas/ibm,open-sriov-allow-unfreeze ] || 158 [ ! -f /proc/device-tree/rtas/ibm,open-sriov-map-pe-number ] ; then 173 [ ! -f "$sysfs/of_node/ibm,is-open-sriov-pf" ] && 174 [ ! -f "$sysfs/of_node/ibm,open-sriov-vf-bar-info" ] ; then
|
/linux-master/drivers/net/ethernet/sfc/ |
H A D | Makefile | 12 sfc-$(CONFIG_SFC_SRIOV) += sriov.o ef10_sriov.o ef100_sriov.o ef100_rep.o \
|
/linux-master/drivers/infiniband/hw/mlx5/ |
H A D | ib_virt.c | 98 struct mlx5_vf_context *vfs_ctx = mdev->priv.sriov.vfs_ctx; 157 struct mlx5_vf_context *vfs_ctx = mdev->priv.sriov.vfs_ctx; 181 struct mlx5_vf_context *vfs_ctx = mdev->priv.sriov.vfs_ctx; 216 struct mlx5_vf_context *vfs_ctx = mdev->priv.sriov.vfs_ctx;
|
/linux-master/arch/powerpc/platforms/powernv/ |
H A D | Makefile | 23 obj-$(CONFIG_PCI_IOV) += pci-sriov.o
|
/linux-master/drivers/net/ethernet/broadcom/bnx2x/ |
H A D | bnx2x_sriov.h | 219 #define BNX2X_NR_VIRTFN(bp) ((bp)->vfdb->sriov.nr_virtfn) 241 #define GET_NUM_VFS_PER_PF(bp) ((bp)->vfdb ? (bp)->vfdb->sriov.total \ 333 struct bnx2x_sriov sriov; member in struct:bnx2x_vfdb 541 return bp->vfdb->sriov.nr_virtfn * BNX2X_CIDS_PER_VF;
|
H A D | bnx2x_sriov.c | 1047 struct bnx2x_sriov *iov = &bp->vfdb->sriov; 1056 struct bnx2x_sriov *iov = &bp->vfdb->sriov; 1065 struct bnx2x_sriov *iov = &bp->vfdb->sriov; 1127 DP(BNX2X_MSG_IOV, "sriov ext pos %d\n", pos); 1182 /* verify sriov capability is present in configuration space */ 1228 /* get the sriov info - Linux already collected all the pertinent 1229 * information, however the sriov structure is for the private use 1233 iov = &(bp->vfdb->sriov); 1314 for (vf_idx = 0; vf_idx < bp->vfdb->sriov.total; vf_idx++) { 1317 bp->vfdb->sriov [all...] |
/linux-master/drivers/gpu/drm/xe/tests/ |
H A D | xe_pci.c | 159 xe->sriov.__mode = data && data->sriov_mode ?
|
/linux-master/drivers/net/ethernet/sfc/siena/ |
H A D | siena_sriov.c | 1245 unsigned index, devfn, sriov, buftbl_base; local 1250 sriov = pci_find_ext_capability(pci_dev, PCI_EXT_CAP_ID_SRIOV); 1251 if (!sriov) 1254 pci_read_config_word(pci_dev, sriov + PCI_SRIOV_VF_OFFSET, &offset); 1255 pci_read_config_word(pci_dev, sriov + PCI_SRIOV_VF_STRIDE, &stride);
|
/linux-master/include/linux/mlx5/ |
H A D | driver.h | 624 struct mlx5_core_sriov sriov; member in struct:mlx5_priv 1241 return dev->priv.sriov.max_vfs; 1258 return dev->priv.sriov.max_ec_vfs;
|