Searched refs:contexts (Results 1 - 25 of 51) sorted by path

123

/linux-master/arch/powerpc/platforms/cell/spufs/
H A Dgang.c37 WARN_ON(gang->contexts || !list_empty(&gang->list));
57 gang->contexts++;
70 gang->contexts--;
H A Dsched.c305 * contexts bigger than the amount of spus in the node,
320 available_spus -= spu->ctx->gang->contexts;
323 if (available_spus < ctx->gang->contexts) {
650 * higher priority contexts before lower priority
735 * @free_spu Whether to free the SPU for other contexts
738 * SPU is made available for other contexts (ie, may be returned by
H A Dspufs.h146 int contexts; member in struct:spu_gang
/linux-master/drivers/staging/media/omap4iss/
H A Diss_csi2.h139 struct iss_csi2_ctx_cfg contexts[ISS_CSI2_MAX_CTX_NUM + 1]; member in struct:iss_csi2_device
/linux-master/drivers/gpu/drm/i915/gem/
H A Di915_gem_context.c13 * supports contexts for the render ring.
15 * In software, there is a distinction between contexts created by the user,
24 * swap out, and/or destroy other contexts.
26 * All other contexts are created as a request by the GPU client. These contexts
217 * Only contexts that are short-lived [that will expire or be
652 /* Create contexts / engines */
1278 spin_lock(&ctx->i915->gem.contexts.lock);
1280 spin_unlock(&ctx->i915->gem.contexts.lock);
1556 * Only contexts tha
[all...]
/linux-master/drivers/gpu/drm/i915/gem/selftests/
H A Dmock_context.c74 init_contexts(&i915->gem.contexts);
/linux-master/drivers/gpu/drm/i915/gt/
H A Dselftest_execlists.c214 * Setup the pair of contexts such that if we
2562 * Build a chain AB...BA between two contexts (A, B) and request
3079 /* All GPR are clear for new contexts. We use GPR(0) as a constant */
3266 * a few preempting contexts thrown into the mix, we look for any
3311 /* Continuously preempt the set of 3 running contexts */
3471 struct i915_gem_context **contexts; member in struct:preempt_smoke
3482 return smoke->contexts[i915_prandom_u32_max_state(smoke->ncontext,
3601 pr_info("Submitted %lu crescendo:%x requests across %d engines and %d contexts\n",
3630 pr_info("Submitted %lu random:%x requests across %d engines and %d contexts\n",
3648 smoke.contexts
[all...]
/linux-master/drivers/gpu/drm/i915/
H A Di915_drv.h338 } contexts; member in struct:drm_i915_private::__anon439
H A Di915_gem.c1185 * default HW state for all contexts.
1285 drm_WARN_ON(&dev_priv->drm, !list_empty(&dev_priv->gem.contexts.list));
H A Di915_perf.c77 * the metrics of other processes/contexts. For these use cases a read() based
345 * @single_context: Whether a single or all gpu contexts should be monitored
834 * contexts while filtering since we can't stop userspace
843 * associated with other contexts. Note that the hardware
849 * other, unrelated contexts) we simply elect to forward them
863 * leaking the IDs of other contexts.
1401 * 0 - (NUM_CONTEXT_TAG -1) are used by other contexts
2658 * configuration across all contexts.
2673 * - Ensure that all existing contexts will have the correct per-context
2675 * - Ensure any new contexts wil
[all...]
H A Di915_sysfs.c79 spin_lock(&i915->gem.contexts.lock);
84 spin_unlock(&i915->gem.contexts.lock);
112 spin_lock(&i915->gem.contexts.lock);
125 list_for_each_entry(ctx, &i915->gem.contexts.list, link)
128 spin_unlock(&i915->gem.contexts.lock);
/linux-master/drivers/gpu/drm/i915/pxp/
H A Dintel_pxp.c41 * session are marked as invalid and all contexts marked as using protected
495 /* ban all contexts marked as protected */
496 spin_lock_irq(&i915->gem.contexts.lock);
497 list_for_each_entry_safe(ctx, cn, &i915->gem.contexts.list, link) {
509 spin_unlock_irq(&i915->gem.contexts.lock);
535 spin_lock_irq(&i915->gem.contexts.lock);
539 spin_unlock_irq(&i915->gem.contexts.lock);
/linux-master/drivers/gpu/drm/i915/selftests/
H A Di915_request.c284 struct i915_gem_context **contexts; member in struct:smoketest
366 t->contexts[order[n] % t->ncontexts];
478 t.contexts = kcalloc(t.ncontexts, sizeof(*t.contexts), GFP_KERNEL);
479 if (!t.contexts) {
485 t.contexts[n] = mock_context(t.engine->i915, "mock");
486 if (!t.contexts[n]) {
532 if (!t.contexts[n])
534 mock_context_close(t.contexts[n]);
536 kfree(t.contexts);
[all...]
H A Di915_vma.c109 struct list_head *contexts)
117 list_for_each_entry(ctx, contexts, link) {
158 LIST_HEAD(contexts);
183 list_move(&ctx->link, &contexts);
186 err = create_vmas(i915, &objects, &contexts);
191 "%s timed out: after %lu objects in %lu contexts\n",
196 list_for_each_entry_safe(ctx, cn, &contexts, link) {
205 /* Final pass to lookup all created contexts */
206 err = create_vmas(i915, &objects, &contexts);
208 list_for_each_entry_safe(ctx, cn, &contexts, lin
107 create_vmas(struct drm_i915_private *i915, struct list_head *objects, struct list_head *contexts) argument
[all...]
/linux-master/drivers/gpu/drm/tegra/
H A Ddrm.c111 xa_init_flags(&fpriv->contexts, XA_FLAGS_ALLOC1);
H A Dsubmit.c524 context = xa_load(&fpriv->contexts, args->context);
H A Duapi.c55 xa_for_each(&file->contexts, id, context)
61 xa_destroy(&file->contexts);
127 * OK, HW does not support contexts or contexts
135 err = xa_alloc(&fpriv->contexts, &args->context, context, XA_LIMIT(1, U32_MAX),
170 context = xa_load(&fpriv->contexts, args->context);
176 xa_erase(&fpriv->contexts, args->context);
200 context = xa_load(&fpriv->contexts, args->context);
281 context = xa_load(&fpriv->contexts, args->context);
H A Duapi.h23 struct xarray contexts; member in struct:tegra_drm_file
/linux-master/drivers/md/dm-vdo/
H A Ddedupe.c94 * hash_lock's agent, added to the list of pending contexts, and then sent to the index. The
103 * pending list will be searched for any contexts in the pending state which have been running for
104 * too long. Those contexts are transitioned to the DEDUPE_CONTEXT_TIMED_OUT state and moved to the
106 * data_vios associated with timed out contexts are sent to continue processing their write
114 * be searched for any contexts which are timed out and complete. One of these will be used
2290 VDO_ASSERT_LOG_ONLY(READ_ONCE(zone->active) == 0, "all contexts inactive");
2383 struct dedupe_context *context = &zone->contexts[i];
H A Ddedupe.h57 /* These fields are used to manage the dedupe contexts */
66 /* The dedupe contexts for querying the index from this zone */
67 struct dedupe_context contexts[MAXIMUM_VDO_USER_VIOS]; member in struct:hash_zone
H A Dvdo.c408 * @contexts: An array of queue_count contexts, one for each individual queue; may be NULL.
418 unsigned int queue_count, void *contexts[])
436 type, queue_count, contexts, &thread->queue);
416 vdo_make_thread(struct vdo *vdo, thread_id_t thread_id, const struct vdo_work_queue_type *type, unsigned int queue_count, void *contexts[]) argument
H A Dvdo.h292 unsigned int queue_count, void *contexts[]);
/linux-master/drivers/media/platform/ti/omap3isp/
H A Dispcsi2.c220 struct isp_csi2_ctx_cfg *ctx = &csi2->contexts[0];
250 struct isp_csi2_ctx_cfg *ctx = &csi2->contexts[ctxnum];
561 if (csi2->contexts[0].enabled || csi2->ctrl.if_enable)
599 csi2->contexts[0].format_id = csi2_ctx_map_format(csi2);
602 csi2->contexts[0].data_offset = 0;
604 csi2->contexts[0].data_offset = csi2->video_out.bpl_value;
613 csi2->contexts[0].eof_enabled = 1;
614 csi2->contexts[0].eol_enabled = 1;
623 csi2_ctx_config(isp, csi2, &csi2->contexts[0]);
785 csi2_isr_ctx(csi2, &csi2->contexts[
[all...]
H A Dispcsi2.h137 struct isp_csi2_ctx_cfg contexts[ISP_CSI2_MAX_CTX_NUM + 1]; member in struct:isp_csi2_device
/linux-master/drivers/misc/ocxl/
H A Dsysfs.c92 __ATTR_RO(contexts),

Completed in 317 milliseconds

123