Lines Matching refs:mdev

36 	struct virtio_pci_modern_device *mdev;
48 struct virtio_pci_modern_device *mdev;
61 return vp_vdpa->mdev;
66 return vp_vdpa->mdev;
78 struct virtio_pci_modern_device *mdev = vdpa_to_mdev(vdpa);
80 vp_modern_set_features(mdev, features);
87 struct virtio_pci_modern_device *mdev = vdpa_to_mdev(vdpa);
89 return vp_modern_get_driver_features(mdev);
94 struct virtio_pci_modern_device *mdev = vdpa_to_mdev(vdpa);
96 return vp_modern_get_status(mdev);
112 struct virtio_pci_modern_device *mdev = vp_vdpa_to_mdev(vp_vdpa);
113 struct pci_dev *pdev = mdev->pci_dev;
118 vp_modern_queue_vector(mdev, i, VIRTIO_MSI_NO_VECTOR);
126 vp_modern_config_vector(mdev, VIRTIO_MSI_NO_VECTOR);
159 struct virtio_pci_modern_device *mdev = vp_vdpa_to_mdev(vp_vdpa);
160 struct pci_dev *pdev = mdev->pci_dev;
197 vp_modern_queue_vector(mdev, i, msix_vec);
212 vp_modern_config_vector(mdev, msix_vec);
224 struct virtio_pci_modern_device *mdev = vp_vdpa_to_mdev(vp_vdpa);
235 vp_modern_set_status(mdev, status);
241 struct virtio_pci_modern_device *mdev = vp_vdpa_to_mdev(vp_vdpa);
244 vp_modern_set_status(mdev, 0);
295 struct virtio_pci_modern_device *mdev = vdpa_to_mdev(vdpa);
301 if ((vp_modern_get_status(mdev) & VIRTIO_CONFIG_S_FEATURES_OK) &&
302 !vp_modern_get_queue_enable(mdev, qid)) {
303 if (vp_modern_get_driver_features(mdev) &
324 struct virtio_pci_modern_device *mdev = vdpa_to_mdev(vdpa);
326 vp_modern_set_queue_enable(mdev, qid, ready);
331 struct virtio_pci_modern_device *mdev = vdpa_to_mdev(vdpa);
333 return vp_modern_get_queue_enable(mdev, qid);
339 struct virtio_pci_modern_device *mdev = vdpa_to_mdev(vdpa);
341 vp_modern_set_queue_size(mdev, qid, num);
346 struct virtio_pci_modern_device *mdev = vdpa_to_mdev(vdpa);
348 return vp_modern_get_queue_size(mdev, qid);
355 struct virtio_pci_modern_device *mdev = vdpa_to_mdev(vdpa);
357 vp_modern_queue_address(mdev, qid, desc_area,
372 struct virtio_pci_modern_device *mdev = vdpa_to_mdev(vdpa);
374 return vp_modern_generation(mdev);
379 struct virtio_pci_modern_device *mdev = vdpa_to_mdev(vdpa);
381 return mdev->id.device;
386 struct virtio_pci_modern_device *mdev = vdpa_to_mdev(vdpa);
388 return mdev->id.vendor;
398 struct virtio_pci_modern_device *mdev = vdpa_to_mdev(vdpa);
400 return mdev->device_len;
408 struct virtio_pci_modern_device *mdev = vp_vdpa_to_mdev(vp_vdpa);
414 old = vp_ioread8(&mdev->common->config_generation);
417 *p++ = vp_ioread8(mdev->device + offset + i);
419 new = vp_ioread8(&mdev->common->config_generation);
428 struct virtio_pci_modern_device *mdev = vp_vdpa_to_mdev(vp_vdpa);
433 vp_iowrite8(*p++, mdev->device + offset + i);
448 struct virtio_pci_modern_device *mdev = vp_vdpa_to_mdev(vp_vdpa);
452 notify.size = mdev->notify_offset_multiplier;
497 struct virtio_pci_modern_device *mdev = vp_vdpa_mgtdev->mdev;
498 struct pci_dev *pdev = mdev->pci_dev;
515 vp_vdpa->queues = vp_modern_get_num_queues(mdev);
516 vp_vdpa->mdev = mdev;
518 device_features = vp_modern_get_features(mdev);
551 vp_modern_map_vq_notify(mdev, i,
561 vp_vdpa->vdpa.mdev = &vp_vdpa_mgtdev->mgtdev;
597 struct virtio_pci_modern_device *mdev = NULL;
609 mdev = kzalloc(sizeof(struct virtio_pci_modern_device), GFP_KERNEL);
610 if (!mdev) {
621 vp_vdpa_mgtdev->mdev = mdev;
622 mdev->pci_dev = pdev;
629 err = vp_modern_probe(mdev);
635 mdev_id->device = mdev->id.device;
636 mdev_id->vendor = mdev->id.vendor;
638 mgtdev->max_supported_vqs = vp_modern_get_num_queues(mdev);
639 mgtdev->supported_features = vp_modern_get_features(mdev);
653 vp_modern_remove(vp_vdpa_mgtdev->mdev);
657 kfree(mdev);
666 struct virtio_pci_modern_device *mdev = NULL;
668 mdev = vp_vdpa_mgtdev->mdev;
670 vp_modern_remove(mdev);
672 kfree(mdev);