/linux-master/arch/powerpc/platforms/cell/spufs/ |
H A D | gang.c | 37 WARN_ON(gang->contexts || !list_empty(&gang->list)); 57 gang->contexts++; 70 gang->contexts--;
|
H A D | sched.c | 305 * contexts bigger than the amount of spus in the node, 320 available_spus -= spu->ctx->gang->contexts; 323 if (available_spus < ctx->gang->contexts) { 650 * higher priority contexts before lower priority 735 * @free_spu Whether to free the SPU for other contexts 738 * SPU is made available for other contexts (ie, may be returned by
|
H A D | spufs.h | 146 int contexts; member in struct:spu_gang
|
/linux-master/drivers/staging/media/omap4iss/ |
H A D | iss_csi2.h | 139 struct iss_csi2_ctx_cfg contexts[ISS_CSI2_MAX_CTX_NUM + 1]; member in struct:iss_csi2_device
|
/linux-master/drivers/gpu/drm/i915/gem/ |
H A D | i915_gem_context.c | 13 * supports contexts for the render ring. 15 * In software, there is a distinction between contexts created by the user, 24 * swap out, and/or destroy other contexts. 26 * All other contexts are created as a request by the GPU client. These contexts 217 * Only contexts that are short-lived [that will expire or be 652 /* Create contexts / engines */ 1278 spin_lock(&ctx->i915->gem.contexts.lock); 1280 spin_unlock(&ctx->i915->gem.contexts.lock); 1556 * Only contexts tha [all...] |
/linux-master/drivers/gpu/drm/i915/gem/selftests/ |
H A D | mock_context.c | 74 init_contexts(&i915->gem.contexts);
|
/linux-master/drivers/gpu/drm/i915/gt/ |
H A D | selftest_execlists.c | 214 * Setup the pair of contexts such that if we 2562 * Build a chain AB...BA between two contexts (A, B) and request 3079 /* All GPR are clear for new contexts. We use GPR(0) as a constant */ 3266 * a few preempting contexts thrown into the mix, we look for any 3311 /* Continuously preempt the set of 3 running contexts */ 3471 struct i915_gem_context **contexts; member in struct:preempt_smoke 3482 return smoke->contexts[i915_prandom_u32_max_state(smoke->ncontext, 3601 pr_info("Submitted %lu crescendo:%x requests across %d engines and %d contexts\n", 3630 pr_info("Submitted %lu random:%x requests across %d engines and %d contexts\n", 3648 smoke.contexts [all...] |
/linux-master/drivers/gpu/drm/i915/ |
H A D | i915_drv.h | 338 } contexts; member in struct:drm_i915_private::__anon439
|
H A D | i915_gem.c | 1185 * default HW state for all contexts. 1285 drm_WARN_ON(&dev_priv->drm, !list_empty(&dev_priv->gem.contexts.list));
|
H A D | i915_perf.c | 77 * the metrics of other processes/contexts. For these use cases a read() based 345 * @single_context: Whether a single or all gpu contexts should be monitored 834 * contexts while filtering since we can't stop userspace 843 * associated with other contexts. Note that the hardware 849 * other, unrelated contexts) we simply elect to forward them 863 * leaking the IDs of other contexts. 1401 * 0 - (NUM_CONTEXT_TAG -1) are used by other contexts 2658 * configuration across all contexts. 2673 * - Ensure that all existing contexts will have the correct per-context 2675 * - Ensure any new contexts wil [all...] |
H A D | i915_sysfs.c | 79 spin_lock(&i915->gem.contexts.lock); 84 spin_unlock(&i915->gem.contexts.lock); 112 spin_lock(&i915->gem.contexts.lock); 125 list_for_each_entry(ctx, &i915->gem.contexts.list, link) 128 spin_unlock(&i915->gem.contexts.lock);
|
/linux-master/drivers/gpu/drm/i915/pxp/ |
H A D | intel_pxp.c | 41 * session are marked as invalid and all contexts marked as using protected 495 /* ban all contexts marked as protected */ 496 spin_lock_irq(&i915->gem.contexts.lock); 497 list_for_each_entry_safe(ctx, cn, &i915->gem.contexts.list, link) { 509 spin_unlock_irq(&i915->gem.contexts.lock); 535 spin_lock_irq(&i915->gem.contexts.lock); 539 spin_unlock_irq(&i915->gem.contexts.lock);
|
/linux-master/drivers/gpu/drm/i915/selftests/ |
H A D | i915_request.c | 284 struct i915_gem_context **contexts; member in struct:smoketest 366 t->contexts[order[n] % t->ncontexts]; 478 t.contexts = kcalloc(t.ncontexts, sizeof(*t.contexts), GFP_KERNEL); 479 if (!t.contexts) { 485 t.contexts[n] = mock_context(t.engine->i915, "mock"); 486 if (!t.contexts[n]) { 532 if (!t.contexts[n]) 534 mock_context_close(t.contexts[n]); 536 kfree(t.contexts); [all...] |
H A D | i915_vma.c | 109 struct list_head *contexts) 117 list_for_each_entry(ctx, contexts, link) { 158 LIST_HEAD(contexts); 183 list_move(&ctx->link, &contexts); 186 err = create_vmas(i915, &objects, &contexts); 191 "%s timed out: after %lu objects in %lu contexts\n", 196 list_for_each_entry_safe(ctx, cn, &contexts, link) { 205 /* Final pass to lookup all created contexts */ 206 err = create_vmas(i915, &objects, &contexts); 208 list_for_each_entry_safe(ctx, cn, &contexts, lin 107 create_vmas(struct drm_i915_private *i915, struct list_head *objects, struct list_head *contexts) argument [all...] |
/linux-master/drivers/gpu/drm/tegra/ |
H A D | drm.c | 111 xa_init_flags(&fpriv->contexts, XA_FLAGS_ALLOC1);
|
H A D | submit.c | 524 context = xa_load(&fpriv->contexts, args->context);
|
H A D | uapi.c | 55 xa_for_each(&file->contexts, id, context) 61 xa_destroy(&file->contexts); 127 * OK, HW does not support contexts or contexts 135 err = xa_alloc(&fpriv->contexts, &args->context, context, XA_LIMIT(1, U32_MAX), 170 context = xa_load(&fpriv->contexts, args->context); 176 xa_erase(&fpriv->contexts, args->context); 200 context = xa_load(&fpriv->contexts, args->context); 281 context = xa_load(&fpriv->contexts, args->context);
|
H A D | uapi.h | 23 struct xarray contexts; member in struct:tegra_drm_file
|
/linux-master/drivers/md/dm-vdo/ |
H A D | dedupe.c | 94 * hash_lock's agent, added to the list of pending contexts, and then sent to the index. The 103 * pending list will be searched for any contexts in the pending state which have been running for 104 * too long. Those contexts are transitioned to the DEDUPE_CONTEXT_TIMED_OUT state and moved to the 106 * data_vios associated with timed out contexts are sent to continue processing their write 114 * be searched for any contexts which are timed out and complete. One of these will be used 2290 VDO_ASSERT_LOG_ONLY(READ_ONCE(zone->active) == 0, "all contexts inactive"); 2383 struct dedupe_context *context = &zone->contexts[i];
|
H A D | dedupe.h | 57 /* These fields are used to manage the dedupe contexts */ 66 /* The dedupe contexts for querying the index from this zone */ 67 struct dedupe_context contexts[MAXIMUM_VDO_USER_VIOS]; member in struct:hash_zone
|
H A D | vdo.c | 408 * @contexts: An array of queue_count contexts, one for each individual queue; may be NULL. 418 unsigned int queue_count, void *contexts[]) 436 type, queue_count, contexts, &thread->queue); 416 vdo_make_thread(struct vdo *vdo, thread_id_t thread_id, const struct vdo_work_queue_type *type, unsigned int queue_count, void *contexts[]) argument
|
H A D | vdo.h | 292 unsigned int queue_count, void *contexts[]);
|
/linux-master/drivers/media/platform/ti/omap3isp/ |
H A D | ispcsi2.c | 220 struct isp_csi2_ctx_cfg *ctx = &csi2->contexts[0]; 250 struct isp_csi2_ctx_cfg *ctx = &csi2->contexts[ctxnum]; 561 if (csi2->contexts[0].enabled || csi2->ctrl.if_enable) 599 csi2->contexts[0].format_id = csi2_ctx_map_format(csi2); 602 csi2->contexts[0].data_offset = 0; 604 csi2->contexts[0].data_offset = csi2->video_out.bpl_value; 613 csi2->contexts[0].eof_enabled = 1; 614 csi2->contexts[0].eol_enabled = 1; 623 csi2_ctx_config(isp, csi2, &csi2->contexts[0]); 785 csi2_isr_ctx(csi2, &csi2->contexts[ [all...] |
H A D | ispcsi2.h | 137 struct isp_csi2_ctx_cfg contexts[ISP_CSI2_MAX_CTX_NUM + 1]; member in struct:isp_csi2_device
|
/linux-master/drivers/misc/ocxl/ |
H A D | sysfs.c | 92 __ATTR_RO(contexts),
|