/linux-master/drivers/gpu/drm/i915/gt/uc/ |
H A D | selftest_guc_hangcheck.c | 34 struct intel_context *ce; local 55 ce = intel_context_create(engine); 56 if (IS_ERR(ce)) { 57 ret = PTR_ERR(ce); 58 gt_err(gt, "Failed to create spinner request: %pe\n", ce); 77 rq = igt_spinner_create_request(&spin, ce, MI_ARB_CHECK); 78 intel_context_put(ce);
|
H A D | intel_gsc_uc.c | 215 struct intel_context *ce; local 226 ce = intel_engine_create_pinned_context(engine, engine->gt->vm, SZ_4K, 229 if (IS_ERR(ce)) { 231 err = PTR_ERR(ce); 235 gsc->ce = ce; 266 if (gsc->ce) 267 intel_engine_destroy_pinned_context(fetch_and_zero(&gsc->ce));
|
/linux-master/drivers/clocksource/ |
H A D | timer-riscv.c | 47 struct clock_event_device *ce) 109 struct clock_event_device *ce = per_cpu_ptr(&riscv_clock_event, cpu); local 114 ce->cpumask = cpumask_of(cpu); 115 ce->irq = riscv_clock_event_irq; 117 ce->features |= CLOCK_EVT_FEAT_C3STOP; 119 ce->rating = 450; 120 clockevents_config_and_register(ce, riscv_timebase, 100, ULONG_MAX); 46 riscv_clock_next_event(unsigned long delta, struct clock_event_device *ce) argument
|
H A D | timer-cadence-ttc.c | 96 struct clock_event_device ce; member in struct:ttc_timer_clockevent 100 container_of(x, struct ttc_timer_clockevent, ce) 147 ttce->ce.event_handler(&ttce->ce); 404 clockevents_update_freq(&ttcce->ce, ndata->new_rate / PRESCALE); 444 ttcce->ce.name = "ttc_clockevent"; 445 ttcce->ce.features = CLOCK_EVT_FEAT_PERIODIC | CLOCK_EVT_FEAT_ONESHOT; 446 ttcce->ce.set_next_event = ttc_set_next_event; 447 ttcce->ce.set_state_shutdown = ttc_shutdown; 448 ttcce->ce [all...] |
H A D | timer-mp-csky.c | 21 struct clock_event_device *ce) 28 static int csky_mptimer_shutdown(struct clock_event_device *ce) argument 35 static int csky_mptimer_oneshot(struct clock_event_device *ce) argument 42 static int csky_mptimer_oneshot_stopped(struct clock_event_device *ce) argument 20 csky_mptimer_set_next_event(unsigned long delta, struct clock_event_device *ce) argument
|
/linux-master/drivers/gpu/drm/i915/gt/ |
H A D | selftest_execlists.c | 124 struct intel_context *ce; local 127 ce = intel_context_create(engine); 128 if (IS_ERR(ce)) { 129 err = PTR_ERR(ce); 133 rq = igt_spinner_create_request(&spin, ce, MI_NOOP); 155 intel_context_put(ce); 181 struct intel_context *ce[2] = {}; local 198 for (n = 0; n < ARRAY_SIZE(ce); n++) { 215 * lite-restore using the RING_TAIL from ce[1] it 216 * will execute garbage from ce[ 343 struct intel_context *ce[2] = {}; local 492 struct intel_context *ce; local 600 struct intel_context *ce; local 719 struct intel_context *ce; local 859 struct intel_context *ce; local 1049 create_rewinder(struct intel_context *ce, struct i915_request *wait, void *slot, int idx) argument 1131 struct intel_context *ce; local 1429 struct intel_context *ce; local 1720 struct intel_context *ce; local 2705 struct intel_context *ce; local 2800 struct intel_context *ce[2] = {}; local 3152 struct intel_context *ce; local 4026 struct intel_context *ce; local 4093 struct intel_context *ce; local [all...] |
H A D | selftest_engine_cs.c | 79 static struct i915_vma *create_empty_batch(struct intel_context *ce) argument 86 obj = i915_gem_object_create_internal(ce->engine->i915, PAGE_SIZE); 100 vma = i915_vma_instance(obj, ce->vm, NULL); 146 struct intel_context *ce = engine->kernel_context; local 156 batch = create_empty_batch(ce); 173 rq = i915_request_create(ce); 219 static struct i915_vma *create_nop_batch(struct intel_context *ce) argument 226 obj = i915_gem_object_create_internal(ce->engine->i915, SZ_64K); 241 vma = i915_vma_instance(obj, ce->vm, NULL); 274 struct intel_context *ce local [all...] |
H A D | selftest_rc6.c | 125 static const u32 *__live_rc6_ctx(struct intel_context *ce) argument 132 rq = intel_context_create_request(ce); 148 *cs++ = ce->timeline->hwsp_offset + 8; 209 struct intel_context *ce; local 215 ce = intel_context_create(engine); 216 if (IS_ERR(ce)) { 217 err = PTR_ERR(ce); 222 res = __live_rc6_ctx(ce); 224 intel_context_put(ce);
|
H A D | selftest_timeline.c | 877 struct intel_context *ce; local 879 ce = intel_context_create(engine); 880 if (IS_ERR(ce)) 881 return PTR_ERR(ce); 883 ce->ring_size = ringsz; 884 w->rq = intel_context_create_request(ce); 885 intel_context_put(ce); 958 struct intel_context *ce = rq->context; local 959 struct intel_timeline *tl = ce->timeline; 964 rq = intel_context_create_request(ce); 1020 struct intel_context *ce; local 1192 struct intel_context *ce = engine->kernel_context; local 1272 struct intel_context *ce; local [all...] |
H A D | intel_migrate.h | 35 int intel_context_migrate_copy(struct intel_context *ce, 55 intel_context_migrate_clear(struct intel_context *ce,
|
H A D | selftest_migrate.c | 137 static int intel_context_copy_ccs(struct intel_context *ce, argument 151 GEM_BUG_ON(ce->vm != ce->engine->gt->migrate.context->vm); 154 GEM_BUG_ON(ce->ring->size < SZ_64K); 157 if (HAS_64K_PAGES(ce->engine->i915)) 163 rq = i915_request_create(ce); 230 struct intel_context *ce; local 237 ce = intel_migrate_create_context(m); 238 if (IS_ERR(ce)) 239 ce 552 struct intel_context *ce; local 843 __perf_clear_blt(struct intel_context *ce, struct scatterlist *sg, unsigned int pat_index, bool is_lmem, size_t sz) argument 919 __perf_copy_blt(struct intel_context *ce, struct scatterlist *src, unsigned int src_pat_index, bool src_is_lmem, struct scatterlist *dst, unsigned int dst_pat_index, bool dst_is_lmem, size_t sz) argument [all...] |
H A D | intel_execlists_submission.c | 434 struct intel_context * const ce = rq->context; local 459 head = __active_request(ce->timeline, rq, -EIO)->head; 460 head = intel_ring_wrap(ce->ring, head); 463 lrc_init_regs(ce, engine, true); 466 ce->lrc.lrca = lrc_update_regs(ce, engine, head); 478 struct intel_context * const ce = rq->context; local 480 intel_context_get(ce); 482 if (unlikely(intel_context_is_closed(ce) && 484 intel_context_set_exiting(ce); 532 struct intel_context * const ce = rq->context; local 560 kick_siblings(struct i915_request *rq, struct intel_context *ce) argument 589 __execlists_schedule_out(struct i915_request * const rq, struct intel_context * const ce) argument 653 struct intel_context * const ce = rq->context; local 677 struct intel_context *ce = rq->context; local 785 struct intel_context *ce = NULL; local 944 ctx_single_port_submission(const struct intel_context *ce) argument 2606 __execlists_context_pre_pin(struct intel_context *ce, struct intel_engine_cs *engine, struct i915_gem_ww_ctx *ww, void **vaddr) argument 2625 execlists_context_pre_pin(struct intel_context *ce, struct i915_gem_ww_ctx *ww, void **vaddr) argument 2632 execlists_context_pin(struct intel_context *ce, void *vaddr) argument 2637 execlists_context_alloc(struct intel_context *ce) argument 2642 execlists_context_cancel_request(struct intel_context *ce, struct i915_request *rq) argument 2660 struct intel_context *parent = NULL, *ce, *err; local 3029 struct intel_context *ce; local 3712 virtual_context_alloc(struct intel_context *ce) argument 3719 virtual_context_pre_pin(struct intel_context *ce, struct i915_gem_ww_ctx *ww, void **vaddr) argument 3729 virtual_context_pin(struct intel_context *ce, void *vaddr) argument 3736 virtual_context_enter(struct intel_context *ce) argument 3747 virtual_context_exit(struct intel_context *ce) argument [all...] |
H A D | selftest_engine_pm.c | 73 static int __measure_timestamps(struct intel_context *ce, argument 76 struct intel_engine_cs *engine = ce->engine; 82 rq = intel_context_create_request(ce); 142 struct intel_context *ce; local 145 ce = intel_context_create(engine); 146 if (IS_ERR(ce)) 147 return PTR_ERR(ce); 150 err = __measure_timestamps(ce, &st[i], &s_ring[i], &s_ctx[i]); 154 intel_context_put(ce);
|
/linux-master/drivers/gpu/drm/i915/gem/selftests/ |
H A D | mock_context.c | 116 struct intel_context *ce; local 128 ce = intel_context_create(engine); 129 if (IS_ERR(ce)) { 131 return ERR_CAST(ce); 134 intel_context_set_gem(ce, ctx, null_sseu); 135 engines->engines[0] = ce;
|
H A D | igt_gem_utils.h | 30 int igt_gpu_fill_dw(struct intel_context *ce,
|
H A D | i915_gem_client_blt.c | 100 struct intel_context *ce; member in struct:tiled_blits 154 struct intel_gt *gt = t->ce->engine->gt; 162 *cs++ = i915_mmio_reg_offset(BLIT_CCTL(t->ce->engine->mmio_base)); 271 struct drm_i915_private *i915 = t->ce->vm->i915; 282 vma = i915_vma_instance(obj, t->ce->vm, NULL); 298 struct drm_i915_private *i915 = t->ce->engine->i915; 499 rq = intel_context_create_request(t->ce); 543 t->ce = intel_context_create(engine); 544 if (IS_ERR(t->ce)) { 545 err = PTR_ERR(t->ce); [all...] |
/linux-master/drivers/gpu/drm/i915/pxp/ |
H A D | intel_pxp.c | 57 return IS_ENABLED(CONFIG_DRM_I915_PXP) && pxp && pxp->ce; 88 struct intel_context *ce; local 100 ce = intel_engine_create_pinned_context(engine, engine->gt->vm, SZ_4K, 103 if (IS_ERR(ce)) { 105 return PTR_ERR(ce); 108 pxp->ce = ce; 115 if (pxp->ce) 116 intel_engine_destroy_pinned_context(fetch_and_zero(&pxp->ce)); 499 struct intel_context *ce; local [all...] |
H A D | intel_pxp_gsccs.c | 68 if (!exec_res->ce) 114 exec_res->ce, &pkt, exec_res->bb_vaddr, 307 if (exec_res->ce) 308 intel_context_put(exec_res->ce); 374 struct intel_context *ce; local 401 ce = intel_context_create(engine); 402 if (IS_ERR(ce)) { 404 err = PTR_ERR(ce); 408 i915_vm_put(ce->vm); 409 ce [all...] |
/linux-master/include/linux/ |
H A D | clockchips.h | 190 extern int clockevents_update_freq(struct clock_event_device *ce, u32 freq); 193 clockevents_calc_mult_shift(struct clock_event_device *ce, u32 freq, u32 maxsec) argument 195 return clocks_calc_mult_shift(&ce->mult, &ce->shift, NSEC_PER_SEC, freq, maxsec);
|
/linux-master/drivers/gpu/drm/i915/gem/ |
H A D | i915_gem_context.c | 968 static int intel_context_set_gem(struct intel_context *ce, argument 974 GEM_BUG_ON(rcu_access_pointer(ce->gem_context)); 975 RCU_INIT_POINTER(ce->gem_context, ctx); 977 GEM_BUG_ON(intel_context_is_pinned(ce)); 979 if (ce->engine->class == COMPUTE_CLASS) 980 ce->ring_size = SZ_512K; 982 ce->ring_size = SZ_16K; 984 i915_vm_put(ce->vm); 985 ce->vm = i915_gem_context_get_eb_vm(ctx); 988 intel_engine_has_timeslices(ce 1012 struct intel_context *ce = e->engines[count], *child; local 1057 struct intel_context *ce; local 1126 struct intel_context *ce; local 1163 perma_pin_contexts(struct intel_context *ce) argument 1210 struct intel_context *ce, *child; local 1354 active_engine(struct intel_context *ce) argument 1396 struct intel_context *ce; local 1462 struct intel_context *ce; local 2010 struct intel_context *ce; local 2074 struct intel_context *ce; local 2389 struct intel_context *ce; local [all...] |
H A D | i915_gem_context.h | 208 struct intel_context *ce; local 213 ce = ERR_PTR(-ENOENT); 215 ce = intel_context_get(e->engines[idx]); 217 ce = ERR_PTR(-EINVAL); 220 return ce; 234 #define for_each_gem_engine(ce, engines, it) \ 236 ((ce) = i915_gem_engines_iter_next(&(it)));)
|
/linux-master/drivers/gpu/drm/nouveau/nvkm/engine/ce/ |
H A D | gt215.c | 43 gt215_ce_intr(struct nvkm_falcon *ce, struct nvkm_chan *chan) argument 45 struct nvkm_subdev *subdev = &ce->engine.subdev;
|
/linux-master/drivers/gpu/drm/i915/ |
H A D | i915_drm_client.h | 73 struct intel_context *ce); 87 struct intel_context *ce) 86 i915_drm_client_add_context_objects(struct i915_drm_client *client, struct intel_context *ce) argument
|
/linux-master/drivers/crypto/allwinner/sun8i-ce/ |
H A D | sun8i-ce.h | 3 * sun8i-ce.h - hardware cryptographic offloader for 113 * struct ce_clock - Describe clocks used by sun8i-ce 270 * @ce: pointer to the private data of driver handling this TFM 276 struct sun8i_ce_dev *ce; member in struct:sun8i_cipher_tfm_ctx 282 * @ce: pointer to the private data of driver handling this TFM 286 struct sun8i_ce_dev *ce; member in struct:sun8i_ce_hash_tfm_ctx 315 * @ce: pointer to the sun8i_ce_dev structure associated with 326 struct sun8i_ce_dev *ce; member in struct:sun8i_ce_alg_template 356 int sun8i_ce_get_engine_number(struct sun8i_ce_dev *ce); 358 int sun8i_ce_run_task(struct sun8i_ce_dev *ce, in [all...] |
/linux-master/drivers/net/wireless/ath/ath11k/ |
H A D | ce.c | 268 ath11k_dp_shadow_stop_timer(ab, &ab->ce.hp_timer[i]); 282 lockdep_assert_held(&ab->ce.ce_lock); 330 spin_lock_bh(&ab->ce.ce_lock); 344 ath11k_warn(ab, "failed to dma map ce rx buf\n"); 365 spin_unlock_bh(&ab->ce.ce_lock); 379 spin_lock_bh(&ab->ce.ce_lock); 414 spin_unlock_bh(&ab->ce.ce_lock); 445 ath11k_dbg(ab, ATH11K_DBG_CE, "rx ce pipe %d len %d\n", 468 spin_lock_bh(&ab->ce.ce_lock); 495 spin_unlock_bh(&ab->ce [all...] |