/linux-master/drivers/pmdomain/st/ |
H A D | Makefile | 2 obj-$(CONFIG_UX500_PM_DOMAIN) += ste-ux500-pm-domain.o
|
/linux-master/drivers/pmdomain/imx/ |
H A D | imx93-pd.c | 37 struct imx93_power_domain *domain = to_imx93_pd(genpd); local 38 void __iomem *addr = domain->addr; 42 ret = clk_bulk_prepare_enable(domain->num_clks, domain->clks); 44 dev_err(domain->dev, "failed to enable clocks for domain: %s\n", genpd->name); 55 dev_err(domain->dev, "pd_on timeout: name: %s, stat: %x\n", genpd->name, val); 64 struct imx93_power_domain *domain = to_imx93_pd(genpd); local 65 void __iomem *addr = domain->addr; 77 dev_err(domain 88 struct imx93_power_domain *domain = platform_get_drvdata(pdev); local 103 struct imx93_power_domain *domain; local [all...] |
/linux-master/include/linux/ |
H A D | omap-iommu.h | 19 int omap_iommu_domain_deactivate(struct iommu_domain *domain); 20 int omap_iommu_domain_activate(struct iommu_domain *domain); 25 static inline int omap_iommu_domain_deactivate(struct iommu_domain *domain) argument 30 static inline int omap_iommu_domain_activate(struct iommu_domain *domain) argument
|
H A D | irqdomain.h | 14 * Interrupt controller "domain" data structure. This could be defined as a 15 * irq domain controller. That is, it handles the mapping between hardware 16 * and virtual interrupt numbers for a given interrupt domain. The domain 18 * (though a domain can cover more than one PIC if they have a flat number 19 * model). It's the domain callbacks that are responsible for setting the 23 * identify the domain. In some cases, and in order to preserve source 122 * @name: Name of interrupt domain 128 * @mutex: Domain lock, hierarchical domains use root domain's lock 129 * @root: Pointer to root domain, o 414 irq_resolve_mapping(struct irq_domain *domain, irq_hw_number_t hwirq) argument 425 irq_find_mapping(struct irq_domain *domain, irq_hw_number_t hwirq) argument 436 irq_linear_revmap(struct irq_domain *domain, irq_hw_number_t hwirq) argument 504 irq_domain_alloc_irqs(struct irq_domain *domain, unsigned int nr_irqs, int node, void *arg) argument 539 irq_domain_is_hierarchy(struct irq_domain *domain) argument 544 irq_domain_is_ipi(struct irq_domain *domain) argument 550 irq_domain_is_ipi_per_cpu(struct irq_domain *domain) argument 555 irq_domain_is_ipi_single(struct irq_domain *domain) argument 560 irq_domain_is_msi(struct irq_domain *domain) argument 565 irq_domain_is_msi_parent(struct irq_domain *domain) argument 570 irq_domain_is_msi_device(struct irq_domain *domain) argument 576 irq_domain_alloc_irqs(struct irq_domain *domain, unsigned int nr_irqs, int node, void *arg) argument 585 irq_domain_is_hierarchy(struct irq_domain *domain) argument 590 irq_domain_is_ipi(struct irq_domain *domain) argument 595 irq_domain_is_ipi_per_cpu(struct irq_domain *domain) argument 600 irq_domain_is_ipi_single(struct irq_domain *domain) argument 605 irq_domain_is_msi(struct irq_domain *domain) argument 610 irq_domain_is_msi_parent(struct irq_domain *domain) argument 615 irq_domain_is_msi_device(struct irq_domain *domain) argument 627 msi_device_domain_alloc_wired(struct irq_domain *domain, unsigned int hwirq, unsigned int type) argument 633 msi_device_domain_free_wired(struct irq_domain *domain, unsigned int virq) argument [all...] |
H A D | async.h | 24 * domain participates in global async_synchronize_full 31 * domain is free to go out of scope as soon as all pending work is 32 * complete, this domain does not participate in async_synchronize_full 42 struct async_domain *domain); 58 * async_schedule_domain - schedule a function for asynchronous execution within a certain domain 61 * @domain: the domain 64 * @domain may be used in the async_synchronize_*_domain() functions to 65 * wait within a certain synchronization domain rather than globally. 70 struct async_domain *domain) 69 async_schedule_domain(async_func_t func, void *data, struct async_domain *domain) argument 111 async_schedule_dev_domain(async_func_t func, struct device *dev, struct async_domain *domain) argument [all...] |
/linux-master/drivers/vdpa/vdpa_user/ |
H A D | iova_domain.c | 20 static int vduse_iotlb_add_range(struct vduse_iova_domain *domain, argument 35 ret = vhost_iotlb_add_range_ctx(domain->iotlb, start, last, 45 static void vduse_iotlb_del_range(struct vduse_iova_domain *domain, argument 51 while ((map = vhost_iotlb_itree_first(domain->iotlb, start, last))) { 55 vhost_iotlb_map_free(domain->iotlb, map); 59 int vduse_domain_set_map(struct vduse_iova_domain *domain, argument 67 spin_lock(&domain->iotlb_lock); 68 vduse_iotlb_del_range(domain, start, last); 73 ret = vduse_iotlb_add_range(domain, map->start, map->last, 80 spin_unlock(&domain 89 vduse_domain_clear_map(struct vduse_iova_domain *domain, struct vhost_iotlb *iotlb) argument 103 vduse_domain_map_bounce_page(struct vduse_iova_domain *domain, u64 iova, u64 size, u64 paddr) argument 123 vduse_domain_unmap_bounce_page(struct vduse_iova_domain *domain, u64 iova, u64 size) argument 160 vduse_domain_bounce(struct vduse_iova_domain *domain, dma_addr_t iova, size_t size, enum dma_data_direction dir) argument 190 vduse_domain_get_coherent_page(struct vduse_iova_domain *domain, u64 iova) argument 211 vduse_domain_get_bounce_page(struct vduse_iova_domain *domain, u64 iova) argument 230 vduse_domain_free_kernel_bounce_pages(struct vduse_iova_domain *domain) argument 250 vduse_domain_add_user_bounce_pages(struct vduse_iova_domain *domain, struct page **pages, int count) argument 286 vduse_domain_remove_user_bounce_pages(struct vduse_iova_domain *domain) argument 317 vduse_domain_reset_bounce_map(struct vduse_iova_domain *domain) argument 332 vduse_domain_init_bounce_map(struct vduse_iova_domain *domain) argument 376 vduse_domain_sync_single_for_device(struct vduse_iova_domain *domain, dma_addr_t dma_addr, size_t size, enum dma_data_direction dir) argument 386 vduse_domain_sync_single_for_cpu(struct vduse_iova_domain *domain, dma_addr_t dma_addr, size_t size, enum dma_data_direction dir) argument 396 vduse_domain_map_page(struct vduse_iova_domain *domain, struct page *page, unsigned long offset, size_t size, enum dma_data_direction dir, unsigned long attrs) argument 430 vduse_domain_unmap_page(struct vduse_iova_domain *domain, dma_addr_t dma_addr, size_t size, enum dma_data_direction dir, unsigned long attrs) argument 445 vduse_domain_alloc_coherent(struct vduse_iova_domain *domain, size_t size, dma_addr_t *dma_addr, gfp_t flag, unsigned long attrs) argument 479 vduse_domain_free_coherent(struct vduse_iova_domain *domain, size_t size, void *vaddr, dma_addr_t dma_addr, unsigned long attrs) argument 508 struct vduse_iova_domain *domain = vmf->vma->vm_private_data; local 534 struct vduse_iova_domain *domain = file->private_data; local 545 struct vduse_iova_domain *domain = file->private_data; local 567 vduse_domain_destroy(struct vduse_iova_domain *domain) argument 575 struct vduse_iova_domain *domain; local [all...] |
H A D | iova_domain.h | 41 int vduse_domain_set_map(struct vduse_iova_domain *domain, 44 void vduse_domain_clear_map(struct vduse_iova_domain *domain, 47 void vduse_domain_sync_single_for_device(struct vduse_iova_domain *domain, 51 void vduse_domain_sync_single_for_cpu(struct vduse_iova_domain *domain, 55 dma_addr_t vduse_domain_map_page(struct vduse_iova_domain *domain, 60 void vduse_domain_unmap_page(struct vduse_iova_domain *domain, 64 void *vduse_domain_alloc_coherent(struct vduse_iova_domain *domain, 68 void vduse_domain_free_coherent(struct vduse_iova_domain *domain, size_t size, 72 void vduse_domain_reset_bounce_map(struct vduse_iova_domain *domain); 74 int vduse_domain_add_user_bounce_pages(struct vduse_iova_domain *domain, [all...] |
/linux-master/kernel/irq/ |
H A D | irqdomain.c | 28 static int irq_domain_alloc_irqs_locked(struct irq_domain *domain, int irq_base, 31 static void irq_domain_check_hierarchy(struct irq_domain *domain); 32 static void irq_domain_free_one_irq(struct irq_domain *domain, unsigned int virq); 63 * identifying an irq domain 66 * @name: Optional user provided domain name 75 * domain struct. 138 struct irq_domain *domain; local 147 domain = kzalloc_node(struct_size(domain, revmap, size), 149 if (!domain) 233 __irq_domain_publish(struct irq_domain *domain) argument 261 struct irq_domain *domain; local 280 irq_domain_remove(struct irq_domain *domain) argument 307 irq_domain_update_bus_token(struct irq_domain *domain, enum irq_domain_bus_token bus_token) argument 363 struct irq_domain *domain; local 419 struct irq_domain *domain; local 480 irq_set_default_host(struct irq_domain *domain) argument 503 irq_domain_is_nomap(struct irq_domain *domain) argument 509 irq_domain_clear_mapping(struct irq_domain *domain, irq_hw_number_t hwirq) argument 523 irq_domain_set_mapping(struct irq_domain *domain, irq_hw_number_t hwirq, struct irq_data *irq_data) argument 542 irq_domain_disassociate(struct irq_domain *domain, unsigned int irq) argument 578 irq_domain_associate_locked(struct irq_domain *domain, unsigned int virq, irq_hw_number_t hwirq) argument 620 irq_domain_associate(struct irq_domain *domain, unsigned int virq, irq_hw_number_t hwirq) argument 633 irq_domain_associate_many(struct irq_domain *domain, unsigned int irq_base, irq_hw_number_t hwirq_base, int count) argument 659 irq_create_direct_mapping(struct irq_domain *domain) argument 691 irq_create_mapping_affinity_locked(struct irq_domain *domain, irq_hw_number_t hwirq, const struct irq_affinity_desc *affinity) argument 730 irq_create_mapping_affinity(struct irq_domain *domain, irq_hw_number_t hwirq, const struct irq_affinity_desc *affinity) argument 794 struct irq_domain *domain; local 913 struct irq_domain *domain; local 940 __irq_resolve_mapping(struct irq_domain *domain, irq_hw_number_t hwirq, unsigned int *irq) argument 1147 struct irq_domain *domain; local 1172 struct irq_domain *domain = data->domain; local 1191 struct irq_domain *domain = data->domain; local 1199 irq_domain_insert_irq_data(struct irq_domain *domain, struct irq_data *child) argument 1255 irq_domain_disconnect_hierarchy(struct irq_domain *domain, unsigned int virq) argument 1318 irq_domain_alloc_irq_data(struct irq_domain *domain, unsigned int virq, unsigned int nr_irqs) argument 1347 irq_domain_get_irq_data(struct irq_domain *domain, unsigned int virq) argument 1369 irq_domain_set_hwirq_and_chip(struct irq_domain *domain, unsigned int virq, irq_hw_number_t hwirq, const struct irq_chip *chip, void *chip_data) argument 1398 irq_domain_set_info(struct irq_domain *domain, unsigned int virq, irq_hw_number_t hwirq, const struct irq_chip *chip, void *chip_data, irq_flow_handler_t handler, void *handler_data, const char *handler_name) argument 1415 irq_domain_free_irqs_common(struct irq_domain *domain, unsigned int virq, unsigned int nr_irqs) argument 1436 irq_domain_free_irqs_top(struct irq_domain *domain, unsigned int virq, unsigned int nr_irqs) argument 1448 irq_domain_free_irqs_hierarchy(struct irq_domain *domain, unsigned int irq_base, unsigned int nr_irqs) argument 1463 irq_domain_alloc_irqs_hierarchy(struct irq_domain *domain, unsigned int irq_base, unsigned int nr_irqs, void *arg) argument 1475 irq_domain_alloc_irqs_locked(struct irq_domain *domain, int irq_base, unsigned int nr_irqs, int node, void *arg, bool realloc, const struct irq_affinity_desc *affinity) argument 1543 __irq_domain_alloc_irqs(struct irq_domain *domain, int irq_base, unsigned int nr_irqs, int node, void *arg, bool realloc, const struct irq_affinity_desc *affinity) argument 1596 irq_domain_push_irq(struct irq_domain *domain, int virq, void *arg) argument 1677 irq_domain_pop_irq(struct irq_domain *domain, int virq) argument 1746 struct irq_domain *domain; local 1765 irq_domain_free_one_irq(struct irq_domain *domain, unsigned int virq) argument 1780 irq_domain_alloc_irqs_parent(struct irq_domain *domain, unsigned int irq_base, unsigned int nr_irqs, void *arg) argument 1798 irq_domain_free_irqs_parent(struct irq_domain *domain, unsigned int irq_base, unsigned int nr_irqs) argument 1811 struct irq_domain *domain = irq_data->domain; local 1825 struct irq_domain *domain = irqd->domain; local 1876 irq_domain_check_hierarchy(struct irq_domain *domain) argument 1888 irq_domain_get_irq_data(struct irq_domain *domain, unsigned int virq) argument 1908 irq_domain_set_info(struct irq_domain *domain, unsigned int virq, irq_hw_number_t hwirq, const struct irq_chip *chip, void *chip_data, irq_flow_handler_t handler, void *handler_data, const char *handler_name) argument 1918 irq_domain_alloc_irqs_locked(struct irq_domain *domain, int irq_base, unsigned int nr_irqs, int node, void *arg, bool realloc, const struct irq_affinity_desc *affinity) argument 1925 irq_domain_check_hierarchy(struct irq_domain *domain) argument 1926 irq_domain_free_one_irq(struct irq_domain *domain, unsigned int virq) argument [all...] |
/linux-master/drivers/scsi/elx/libefc/ |
H A D | efc.h | 36 #define domain_sm_trace(domain) \ 37 efc_log_debug(domain->efc, "[domain:%s] %-20s %-20s\n", \ 38 domain->display_name, __func__, efc_sm_event_name(evt)) \ 40 #define domain_trace(domain, fmt, ...) \ 41 efc_log_debug(domain->efc, \ 42 "[%s]" fmt, domain->display_name, ##__VA_ARGS__) \
|
H A D | efc_domain.c | 17 struct efc_domain *domain = NULL; local 22 domain = data; 24 /* Accept domain callback events from the user driver */ 36 /* lookup domain, or allocate a new one */ 37 domain = efc->domain; 38 if (!domain) { 39 domain = efc_domain_alloc(efc, fcf_wwn); 40 if (!domain) { 45 efc_sm_transition(&domain 106 struct efc_domain *domain = container_of(arg, struct efc_domain, ref); local 116 efc_domain_free(struct efc_domain *domain) argument 135 struct efc_domain *domain; local 174 struct efc_domain *domain = ctx->app; local 196 struct efc_domain *domain = ctx->app; local 824 __efc_domain_attach_internal(struct efc_domain *domain, u32 s_id) argument 834 efc_domain_attach(struct efc_domain *domain, u32 s_id) argument 840 efc_domain_post_event(struct efc_domain *domain, enum efc_sm_event event, void *arg) argument 858 efct_domain_process_pending(struct efc_domain *domain) argument 907 struct efc_domain *domain = efc->domain; local 940 struct efc_domain *domain = (struct efc_domain *)arg; local [all...] |
H A D | efc_cmds.c | 132 nport->indicator, nport->domain->indicator); 207 struct efc_domain *domain, u8 *wwpn) 227 if (domain) { 237 /* domain NULL and wwpn non-NULL */ 278 nport->domain->indicator, false); 317 efc_domain_get_mbox_status(struct efc_domain *domain, u8 *mqe, int status) argument 319 struct efc *efc = domain->efc; 325 domain->indicator, status, 334 efc_domain_free_resources(struct efc_domain *domain, int evt, void *data) argument 336 struct efc *efc = domain 206 efc_cmd_nport_alloc(struct efc *efc, struct efc_nport *nport, struct efc_domain *domain, u8 *wwpn) argument 353 efc_domain_send_nport_evt(struct efc_domain *domain, int port_evt, int domain_evt, void *data) argument 369 struct efc_domain *domain = arg; local 383 efc_domain_alloc_read_sparm64(struct efc_domain *domain) argument 410 struct efc_domain *domain = arg; local 423 efc_domain_alloc_init_vfi(struct efc_domain *domain) argument 455 efc_cmd_domain_alloc(struct efc *efc, struct efc_domain *domain, u32 fcf) argument 499 struct efc_domain *domain = arg; local 513 efc_cmd_domain_attach(struct efc *efc, struct efc_domain *domain, u32 fc_id) argument 552 struct efc_domain *domain = arg; local 567 efc_domain_free_unreg_vfi(struct efc_domain *domain) argument 594 efc_cmd_domain_free(struct efc *efc, struct efc_domain *domain) argument [all...] |
/linux-master/drivers/iommu/ |
H A D | dma-iommu.h | 14 int iommu_get_dma_cookie(struct iommu_domain *domain); 15 void iommu_put_dma_cookie(struct iommu_domain *domain); 17 int iommu_dma_init_fq(struct iommu_domain *domain); 29 static inline int iommu_dma_init_fq(struct iommu_domain *domain) argument 34 static inline int iommu_get_dma_cookie(struct iommu_domain *domain) argument 39 static inline void iommu_put_dma_cookie(struct iommu_domain *domain) argument
|
H A D | iommu-sva.c | 73 struct iommu_domain *domain; local 100 /* Search for an existing domain. */ 101 list_for_each_entry(domain, &mm->iommu_mm->sva_domains, next) { 102 ret = iommu_attach_device_pasid(domain, dev, iommu_mm->pasid); 104 domain->users++; 109 /* Allocate a new domain and set it on device pasid. */ 110 domain = iommu_sva_domain_alloc(dev, mm); 111 if (IS_ERR(domain)) { 112 ret = PTR_ERR(domain); 116 ret = iommu_attach_device_pasid(domain, de 150 struct iommu_domain *domain = handle->domain; local 173 struct iommu_domain *domain = handle->domain; local 284 struct iommu_domain *domain; local [all...] |
H A D | ipmmu-vmsa.c | 211 static u32 ipmmu_ctx_read_root(struct ipmmu_vmsa_domain *domain, argument 214 return ipmmu_ctx_read(domain->mmu->root, domain->context_id, reg); 217 static void ipmmu_ctx_write_root(struct ipmmu_vmsa_domain *domain, argument 220 ipmmu_ctx_write(domain->mmu->root, domain->context_id, reg, data); 223 static void ipmmu_ctx_write_all(struct ipmmu_vmsa_domain *domain, argument 226 if (domain->mmu != domain->mmu->root) 227 ipmmu_ctx_write(domain 254 ipmmu_tlb_sync(struct ipmmu_vmsa_domain *domain) argument 265 ipmmu_tlb_invalidate(struct ipmmu_vmsa_domain *domain) argument 279 ipmmu_utlb_enable(struct ipmmu_vmsa_domain *domain, unsigned int utlb) argument 300 ipmmu_utlb_disable(struct ipmmu_vmsa_domain *domain, unsigned int utlb) argument 311 struct ipmmu_vmsa_domain *domain = cookie; local 331 ipmmu_domain_allocate_context(struct ipmmu_vmsa_device *mmu, struct ipmmu_vmsa_domain *domain) argument 364 ipmmu_domain_setup_context(struct ipmmu_vmsa_domain *domain) argument 417 ipmmu_domain_init_context(struct ipmmu_vmsa_domain *domain) argument 467 ipmmu_domain_destroy_context(struct ipmmu_vmsa_domain *domain) argument 487 ipmmu_domain_irq(struct ipmmu_vmsa_domain *domain) argument 567 struct ipmmu_vmsa_domain *domain; local 580 struct ipmmu_vmsa_domain *domain = to_vmsa_domain(io_domain); local 596 struct ipmmu_vmsa_domain *domain = to_vmsa_domain(io_domain); local 643 struct ipmmu_vmsa_domain *domain; local 672 struct ipmmu_vmsa_domain *domain = to_vmsa_domain(io_domain); local 682 struct ipmmu_vmsa_domain *domain = to_vmsa_domain(io_domain); local 689 struct ipmmu_vmsa_domain *domain = to_vmsa_domain(io_domain); local 704 struct ipmmu_vmsa_domain *domain = to_vmsa_domain(io_domain); local [all...] |
/linux-master/include/trace/events/ |
H A D | kyber.h | 16 TP_PROTO(dev_t dev, const char *domain, const char *type, 20 TP_ARGS(dev, domain, type, percentile, numerator, denominator, samples), 24 __array( char, domain, DOMAIN_LEN ) 34 strscpy(__entry->domain, domain, sizeof(__entry->domain)); 43 MAJOR(__entry->dev), MINOR(__entry->dev), __entry->domain, 50 TP_PROTO(dev_t dev, const char *domain, unsigned int depth), 52 TP_ARGS(dev, domain, depth), 56 __array( char, domain, DOMAIN_LE [all...] |
/linux-master/drivers/iommu/intel/ |
H A D | nested.c | 21 static int intel_nested_attach_dev(struct iommu_domain *domain, argument 25 struct dmar_domain *dmar_domain = to_dmar_domain(domain); 30 if (info->domain) 39 * Stage-1 domain cannot work alone, it is nested on a s2_domain. 43 ret = prepare_domain_attach_device(&dmar_domain->s2_domain->domain, dev); 45 dev_err_ratelimited(dev, "s2 domain is not compatible\n"); 51 dev_err_ratelimited(dev, "Failed to attach domain to iommu\n"); 64 info->domain = dmar_domain; 80 static void intel_nested_domain_free(struct iommu_domain *domain) argument 82 struct dmar_domain *dmar_domain = to_dmar_domain(domain); 91 intel_nested_cache_invalidate_user(struct iommu_domain *domain, struct iommu_user_data_array *array) argument 145 struct dmar_domain *domain; local [all...] |
H A D | cache.c | 42 /* Assign a cache tag with specified type to domain. */ 43 static int cache_tag_assign(struct dmar_domain *domain, u16 did, argument 67 spin_lock_irqsave(&domain->cache_lock, flags); 68 list_for_each_entry(temp, &domain->cache_tags, node) { 71 spin_unlock_irqrestore(&domain->cache_lock, flags); 77 list_add_tail(&tag->node, &domain->cache_tags); 78 spin_unlock_irqrestore(&domain->cache_lock, flags); 84 /* Unassign a cache tag with specified type from domain. */ 85 static void cache_tag_unassign(struct dmar_domain *domain, u16 did, argument 94 spin_lock_irqsave(&domain 108 __cache_tag_assign_domain(struct dmar_domain *domain, u16 did, struct device *dev, ioasid_t pasid) argument 125 __cache_tag_unassign_domain(struct dmar_domain *domain, u16 did, struct device *dev, ioasid_t pasid) argument 136 __cache_tag_assign_parent_domain(struct dmar_domain *domain, u16 did, struct device *dev, ioasid_t pasid) argument 153 __cache_tag_unassign_parent_domain(struct dmar_domain *domain, u16 did, struct device *dev, ioasid_t pasid) argument 164 domain_get_id_for_dev(struct dmar_domain *domain, struct device *dev) argument 187 cache_tag_assign_domain(struct dmar_domain *domain, struct device *dev, ioasid_t pasid) argument 211 cache_tag_unassign_domain(struct dmar_domain *domain, struct device *dev, ioasid_t pasid) argument 261 cache_tag_flush_range(struct dmar_domain *domain, unsigned long start, unsigned long end, int ih) argument 333 cache_tag_flush_all(struct dmar_domain *domain) argument 382 cache_tag_flush_range_np(struct dmar_domain *domain, unsigned long start, unsigned long end) argument [all...] |
/linux-master/drivers/gpu/drm/xe/ |
H A D | xe_force_wake.h | 25 enum xe_force_wake_domains domain) 27 xe_gt_assert(fw->gt, domain); 28 return fw->domains[ffs(domain) - 1].ref; 33 enum xe_force_wake_domains domain) 35 xe_gt_assert(fw->gt, fw->awake_domains & domain); 24 xe_force_wake_ref(struct xe_force_wake *fw, enum xe_force_wake_domains domain) argument 32 xe_force_wake_assert_held(struct xe_force_wake *fw, enum xe_force_wake_domains domain) argument
|
H A D | xe_force_wake.c | 29 static void domain_init(struct xe_force_wake_domain *domain, argument 33 domain->id = id; 34 domain->reg_ctl = reg; 35 domain->reg_ack = ack; 36 domain->val = val; 37 domain->mask = mask; 109 static void domain_wake(struct xe_gt *gt, struct xe_force_wake_domain *domain) argument 111 xe_mmio_write32(gt, domain->reg_ctl, domain->mask | domain 114 domain_wake_wait(struct xe_gt *gt, struct xe_force_wake_domain *domain) argument 122 domain_sleep(struct xe_gt *gt, struct xe_force_wake_domain *domain) argument 127 domain_sleep_wait(struct xe_gt *gt, struct xe_force_wake_domain *domain) argument 146 struct xe_force_wake_domain *domain; local 176 struct xe_force_wake_domain *domain; local [all...] |
/linux-master/arch/powerpc/include/asm/ |
H A D | fsl_pamu_stash.h | 19 int fsl_pamu_configure_l1_stash(struct iommu_domain *domain, u32 cpu);
|
/linux-master/drivers/dca/ |
H A D | dca-core.c | 44 struct dca_domain *domain; local 46 domain = kzalloc(sizeof(*domain), GFP_NOWAIT); 47 if (!domain) 50 INIT_LIST_HEAD(&domain->dca_providers); 51 domain->pci_rc = rc; 53 return domain; 56 static void dca_free_domain(struct dca_domain *domain) argument 58 list_del(&domain->node); 59 kfree(domain); 81 struct dca_domain *domain; local 114 struct dca_domain *domain; local 126 struct dca_domain *domain; local 143 struct dca_domain *domain; local 176 struct dca_domain *domain; local 342 struct dca_domain *domain, *newdomain = NULL; local 399 struct dca_domain *domain; local [all...] |
/linux-master/include/xen/ |
H A D | pci.h | 8 int xen_register_device_domain_owner(struct pci_dev *dev, uint16_t domain); 17 uint16_t domain) 16 xen_register_device_domain_owner(struct pci_dev *dev, uint16_t domain) argument
|
/linux-master/drivers/soc/rockchip/ |
H A D | Makefile | 6 obj-$(CONFIG_ROCKCHIP_IODOMAIN) += io-domain.o
|
/linux-master/arch/x86/include/asm/ |
H A D | irqdomain.h | 43 extern int mp_irqdomain_alloc(struct irq_domain *domain, unsigned int virq, 45 extern void mp_irqdomain_free(struct irq_domain *domain, unsigned int virq, 47 extern int mp_irqdomain_activate(struct irq_domain *domain, 49 extern void mp_irqdomain_deactivate(struct irq_domain *domain, 51 extern int mp_irqdomain_ioapic_idx(struct irq_domain *domain);
|
/linux-master/security/landlock/ |
H A D | cred.h | 20 struct landlock_ruleset *domain; member in struct:landlock_cred_security 31 return landlock_cred(current_cred())->domain; 40 return landlock_cred(__task_cred(task))->domain;
|