Lines Matching refs:ep
183 struct viot_endpoint *ep;
196 ep = kzalloc(sizeof(*ep), GFP_KERNEL);
197 if (!ep)
207 ep->segment_start = node->pci.segment_start;
208 ep->segment_end = node->pci.segment_end;
209 ep->bdf_start = node->pci.bdf_start;
210 ep->bdf_end = node->pci.bdf_end;
211 ep->endpoint_id = node->pci.endpoint_start;
212 ep->viommu = viot_get_iommu(node->pci.output_node);
221 ep->address = node->mmio.base_address;
222 ep->endpoint_id = node->mmio.endpoint;
223 ep->viommu = viot_get_iommu(node->mmio.output_node);
232 if (!ep->viommu) {
242 list_add(&ep->list, list);
246 kfree(ep);
330 struct viot_endpoint *ep;
334 list_for_each_entry(ep, &viot_pci_ranges, list) {
335 if (domain_nr >= ep->segment_start &&
336 domain_nr <= ep->segment_end &&
337 dev_id >= ep->bdf_start &&
338 dev_id <= ep->bdf_end) {
339 epid = ((domain_nr - ep->segment_start) << 16) +
340 dev_id - ep->bdf_start + ep->endpoint_id;
342 return viot_dev_iommu_init(aliased_dev, ep->viommu,
352 struct viot_endpoint *ep;
358 list_for_each_entry(ep, &viot_mmio_endpoints, list) {
359 if (ep->address == mem->start)
360 return viot_dev_iommu_init(&pdev->dev, ep->viommu,
361 ep->endpoint_id);