/linux-master/drivers/gpu/drm/i915/gt/ |
H A D | intel_context.h | 22 #define CE_TRACE(ce, fmt, ...) do { \ 23 const struct intel_context *ce__ = (ce); \ 33 void intel_context_init(struct intel_context *ce, 35 void intel_context_fini(struct intel_context *ce); 43 int intel_context_alloc_state(struct intel_context *ce); 45 void intel_context_free(struct intel_context *ce); 47 int intel_context_reconfigure_sseu(struct intel_context *ce, 52 static inline bool intel_context_is_child(struct intel_context *ce) argument 54 return !!ce->parallel.parent; 57 static inline bool intel_context_is_parent(struct intel_context *ce) argument 65 intel_context_to_parent(struct intel_context *ce) argument 83 intel_context_is_parallel(struct intel_context *ce) argument 122 intel_context_is_pinned(struct intel_context *ce) argument 127 intel_context_cancel_request(struct intel_context *ce, struct i915_request *rq) argument 150 intel_context_pin_if_active(struct intel_context *ce) argument 155 intel_context_pin(struct intel_context *ce) argument 163 intel_context_pin_ww(struct intel_context *ce, struct i915_gem_ww_ctx *ww) argument 172 __intel_context_pin(struct intel_context *ce) argument 180 intel_context_sched_disable_unpin(struct intel_context *ce) argument 185 intel_context_unpin(struct intel_context *ce) argument 208 intel_context_enter(struct intel_context *ce) argument 218 intel_context_mark_active(struct intel_context *ce) argument 225 intel_context_exit(struct intel_context *ce) argument 236 intel_context_get(struct intel_context *ce) argument 242 intel_context_put(struct intel_context *ce) argument 280 intel_context_is_barrier(const struct intel_context *ce) argument 285 intel_context_close(struct intel_context *ce) argument 293 intel_context_is_closed(const struct intel_context *ce) argument 298 intel_context_has_inflight(const struct intel_context *ce) argument 303 intel_context_use_semaphores(const struct intel_context *ce) argument 308 intel_context_set_use_semaphores(struct intel_context *ce) argument 313 intel_context_clear_use_semaphores(struct intel_context *ce) argument 318 intel_context_is_banned(const struct intel_context *ce) argument 323 intel_context_set_banned(struct intel_context *ce) argument 330 intel_context_is_schedulable(const struct intel_context *ce) argument 336 intel_context_is_exiting(const struct intel_context *ce) argument 341 intel_context_set_exiting(struct intel_context *ce) argument 349 intel_context_force_single_submission(const struct intel_context *ce) argument 355 intel_context_set_single_submission(struct intel_context *ce) argument 361 intel_context_nopreempt(const struct intel_context *ce) argument 367 intel_context_set_nopreempt(struct intel_context *ce) argument 373 intel_context_clear_nopreempt(struct intel_context *ce) argument [all...] |
H A D | intel_context.c | 27 struct intel_context *ce = container_of(rcu, typeof(*ce), rcu); local 29 trace_intel_context_free(ce); 30 kmem_cache_free(slab_ce, ce); 33 void intel_context_free(struct intel_context *ce) argument 35 call_rcu(&ce->rcu, rcu_context_free); 41 struct intel_context *ce; local 43 ce = intel_context_alloc(); 44 if (!ce) 47 intel_context_init(ce, engin 52 intel_context_alloc_state(struct intel_context *ce) argument 90 intel_context_active_acquire(struct intel_context *ce) argument 109 intel_context_active_release(struct intel_context *ce) argument 176 intel_context_pre_pin(struct intel_context *ce, struct i915_gem_ww_ctx *ww) argument 208 intel_context_post_unpin(struct intel_context *ce) argument 217 __intel_context_do_pin_ww(struct intel_context *ce, struct i915_gem_ww_ctx *ww) argument 310 __intel_context_do_pin(struct intel_context *ce) argument 327 __intel_context_do_unpin(struct intel_context *ce, int sub) argument 350 struct intel_context *ce = container_of(active, typeof(*ce), active); local 363 struct intel_context *ce = container_of(active, typeof(*ce), active); local 390 intel_context_init(struct intel_context *ce, struct intel_engine_cs *engine) argument 436 intel_context_fini(struct intel_context *ce) argument 468 intel_context_enter_engine(struct intel_context *ce) argument 474 intel_context_exit_engine(struct intel_context *ce) argument 480 intel_context_prepare_remote_request(struct intel_context *ce, struct i915_request *rq) argument 507 intel_context_create_request(struct intel_context *ce) argument 545 intel_context_get_active_request(struct intel_context *ce) argument 595 intel_context_get_total_runtime_ns(struct intel_context *ce) argument 613 intel_context_get_avg_runtime_ns(struct intel_context *ce) argument 623 intel_context_ban(struct intel_context *ce, struct i915_request *rq) argument 636 intel_context_revoke(struct intel_context *ce) argument [all...] |
H A D | intel_context_param.h | 14 intel_context_set_watchdog_us(struct intel_context *ce, u64 timeout_us) argument 16 ce->watchdog.timeout_us = timeout_us;
|
H A D | intel_context_sseu.c | 17 const struct intel_context *ce, 27 offset = i915_ggtt_offset(ce->state) + 41 gen8_modify_rpcs(struct intel_context *ce, const struct intel_sseu sseu) argument 46 lockdep_assert_held(&ce->pin_mutex); 54 if (!intel_context_pin_if_active(ce)) 57 rq = intel_engine_create_kernel_request(ce->engine); 64 ret = intel_context_prepare_remote_request(ce, rq); 66 ret = gen8_emit_rpcs_config(rq, ce, sseu); 70 intel_context_unpin(ce); 75 intel_context_reconfigure_sseu(struct intel_context *ce, argument 16 gen8_emit_rpcs_config(struct i915_request *rq, const struct intel_context *ce, const struct intel_sseu sseu) argument [all...] |
H A D | intel_lrc.h | 36 int lrc_alloc(struct intel_context *ce, 38 void lrc_reset(struct intel_context *ce); 39 void lrc_fini(struct intel_context *ce); 43 lrc_pre_pin(struct intel_context *ce, 48 lrc_pin(struct intel_context *ce, 51 void lrc_unpin(struct intel_context *ce); 52 void lrc_post_unpin(struct intel_context *ce); 54 void lrc_init_state(struct intel_context *ce, 58 void lrc_init_regs(const struct intel_context *ce, 61 void lrc_reset_regs(const struct intel_context *ce, 116 lrc_runtime_start(struct intel_context *ce) argument 129 lrc_runtime_stop(struct intel_context *ce) argument [all...] |
H A D | intel_engine_pm.c | 35 static void dbg_poison_ce(struct intel_context *ce) argument 40 if (ce->state) { 41 struct drm_i915_gem_object *obj = ce->state->obj; 42 int type = intel_gt_coherent_map_type(ce->engine->gt, obj, true); 62 struct intel_context *ce; local 69 ce = engine->kernel_context; 70 if (ce) { 71 GEM_BUG_ON(test_bit(CONTEXT_VALID_BIT, &ce->flags)); 74 while (unlikely(intel_context_inflight(ce))) 78 dbg_poison_ce(ce); 153 struct intel_context *ce = engine->kernel_context; local 313 struct intel_context *ce; local [all...] |
/linux-master/drivers/crypto/allwinner/sun8i-ce/ |
H A D | Makefile | 1 obj-$(CONFIG_CRYPTO_DEV_SUN8I_CE) += sun8i-ce.o 2 sun8i-ce-y += sun8i-ce-core.o sun8i-ce-cipher.o 3 sun8i-ce-$(CONFIG_CRYPTO_DEV_SUN8I_CE_HASH) += sun8i-ce-hash.o 4 sun8i-ce-$(CONFIG_CRYPTO_DEV_SUN8I_CE_PRNG) += sun8i-ce-prng.o 5 sun8i-ce-$(CONFIG_CRYPTO_DEV_SUN8I_CE_TRNG) += sun8i-ce [all...] |
H A D | sun8i-ce-trng.c | 3 * sun8i-ce-trng.c - hardware cryptographic offloader for 12 #include "sun8i-ce.h" 25 struct sun8i_ce_dev *ce; local 35 ce = container_of(rng, struct sun8i_ce_dev, trng); 46 ce->hwrng_stat_req++; 47 ce->hwrng_stat_bytes += todo; 50 dma_dst = dma_map_single(ce->dev, d, todo, DMA_FROM_DEVICE); 51 if (dma_mapping_error(ce->dev, dma_dst)) { 52 dev_err(ce->dev, "Cannot DMA MAP DST\n"); 57 err = pm_runtime_resume_and_get(ce 101 sun8i_ce_hwrng_register(struct sun8i_ce_dev *ce) argument 118 sun8i_ce_hwrng_unregister(struct sun8i_ce_dev *ce) argument [all...] |
H A D | sun8i-ce-core.c | 3 * sun8i-ce-core.c - hardware cryptographic offloader for 31 #include "sun8i-ce.h" 153 int sun8i_ce_get_engine_number(struct sun8i_ce_dev *ce) argument 155 return atomic_inc_return(&ce->flow) % (MAXFLOW - 1); 158 int sun8i_ce_run_task(struct sun8i_ce_dev *ce, int flow, const char *name) argument 162 struct ce_task *cet = ce->chanlist[flow].tl; 165 ce->chanlist[flow].stat_req++; 168 mutex_lock(&ce->mlock); 170 v = readl(ce->base + CE_ICR); 172 writel(v, ce 262 struct sun8i_ce_dev *ce = (struct sun8i_ce_dev *)data; local 688 sun8i_ce_free_chanlist(struct sun8i_ce_dev *ce, int i) argument 703 sun8i_ce_allocate_chanlist(struct sun8i_ce_dev *ce) argument 762 struct sun8i_ce_dev *ce = dev_get_drvdata(dev); local 773 struct sun8i_ce_dev *ce = dev_get_drvdata(dev); local 801 sun8i_ce_pm_init(struct sun8i_ce_dev *ce) argument 815 sun8i_ce_pm_exit(struct sun8i_ce_dev *ce) argument 820 sun8i_ce_get_clks(struct sun8i_ce_dev *ce) argument 860 sun8i_ce_register_algs(struct sun8i_ce_dev *ce) argument 941 sun8i_ce_unregister_algs(struct sun8i_ce_dev *ce) argument 970 struct sun8i_ce_dev *ce; local 1070 struct sun8i_ce_dev *ce = platform_get_drvdata(pdev); local [all...] |
H A D | sun8i-ce-prng.c | 3 * sun8i-ce-prng.c - hardware cryptographic offloader for 12 #include "sun8i-ce.h" 61 struct sun8i_ce_dev *ce; local 72 ce = algt->ce; 75 dev_err(ce->dev, "not seeded\n"); 89 dev_dbg(ce->dev, "%s PRNG slen=%u dlen=%u todo=%u multi=%u\n", __func__, 97 dma_iv = dma_map_single(ce->dev, ctx->seed, ctx->slen, DMA_TO_DEVICE); 98 if (dma_mapping_error(ce->dev, dma_iv)) { 99 dev_err(ce [all...] |
/linux-master/drivers/crypto/gemini/ |
H A D | Makefile | 1 obj-$(CONFIG_CRYPTO_DEV_SL3516) += sl3516-ce.o 2 sl3516-ce-y += sl3516-ce-core.o sl3516-ce-cipher.o sl3516-ce-rng.o
|
H A D | sl3516-ce-core.c | 3 * sl3516-ce-core.c - hardware cryptographic offloader for Storlink SL3516 SoC 28 #include "sl3516-ce.h" 30 static int sl3516_ce_desc_init(struct sl3516_ce_dev *ce) argument 35 ce->tx = dma_alloc_coherent(ce->dev, sz, &ce->dtx, GFP_KERNEL); 36 if (!ce->tx) 38 ce->rx = dma_alloc_coherent(ce->dev, sz, &ce 67 sl3516_ce_free_descs(struct sl3516_ce_dev *ce) argument 77 start_dma_tx(struct sl3516_ce_dev *ce) argument 87 start_dma_rx(struct sl3516_ce_dev *ce) argument 99 get_desc_tx(struct sl3516_ce_dev *ce) argument 110 get_desc_rx(struct sl3516_ce_dev *ce) argument 121 sl3516_ce_run_task(struct sl3516_ce_dev *ce, struct sl3516_ce_cipher_req_ctx *rctx, const char *name) argument 189 struct sl3516_ce_dev *ce = (struct sl3516_ce_dev *)data; local 250 struct sl3516_ce_dev *ce = seq->private; local 282 sl3516_ce_register_algs(struct sl3516_ce_dev *ce) argument 309 sl3516_ce_unregister_algs(struct sl3516_ce_dev *ce) argument 326 sl3516_ce_start(struct sl3516_ce_dev *ce) argument 341 struct sl3516_ce_dev *ce = dev_get_drvdata(dev); local 350 struct sl3516_ce_dev *ce = dev_get_drvdata(dev); local 376 sl3516_ce_pm_init(struct sl3516_ce_dev *ce) argument 390 sl3516_ce_pm_exit(struct sl3516_ce_dev *ce) argument 397 struct sl3516_ce_dev *ce; local 510 struct sl3516_ce_dev *ce = platform_get_drvdata(pdev); local [all...] |
H A D | sl3516-ce-rng.c | 3 * sl3516-ce-rng.c - hardware cryptographic offloader for SL3516 SoC. 9 #include "sl3516-ce.h" 15 struct sl3516_ce_dev *ce; local 20 ce = container_of(rng, struct sl3516_ce_dev, trng); 23 ce->hwrng_stat_req++; 24 ce->hwrng_stat_bytes += max; 27 err = pm_runtime_get_sync(ce->dev); 29 pm_runtime_put_noidle(ce->dev); 34 *data = readl(ce->base + IPSEC_RAND_NUM_REG); 39 pm_runtime_put(ce 44 sl3516_ce_rng_register(struct sl3516_ce_dev *ce) argument 58 sl3516_ce_rng_unregister(struct sl3516_ce_dev *ce) argument [all...] |
/linux-master/arch/arm64/crypto/ |
H A D | Makefile | 8 obj-$(CONFIG_CRYPTO_SHA1_ARM64_CE) += sha1-ce.o 9 sha1-ce-y := sha1-ce-glue.o sha1-ce-core.o 11 obj-$(CONFIG_CRYPTO_SHA2_ARM64_CE) += sha2-ce.o 12 sha2-ce-y := sha2-ce-glue.o sha2-ce-core.o 14 obj-$(CONFIG_CRYPTO_SHA512_ARM64_CE) += sha512-ce.o 15 sha512-ce [all...] |
/linux-master/drivers/gpu/drm/i915/gt/uc/ |
H A D | intel_guc_submission.c | 112 * ce->guc_state.lock 113 * Protects everything under ce->guc_state. Ensures that a context is in the 121 * sched_engine->lock -> ce->guc_state.lock 122 * guc->submission_state.lock -> ce->guc_state.lock 181 static inline void init_sched_state(struct intel_context *ce) argument 183 lockdep_assert_held(&ce->guc_state.lock); 184 ce->guc_state.sched_state &= SCHED_STATE_BLOCKED_MASK; 198 static bool sched_state_is_init(struct intel_context *ce) argument 200 return !(ce->guc_state.sched_state & ~SCHED_STATE_VALID_INIT); 204 context_wait_for_deregister_to_register(struct intel_context *ce) argument 211 set_context_wait_for_deregister_to_register(struct intel_context *ce) argument 219 clr_context_wait_for_deregister_to_register(struct intel_context *ce) argument 227 context_destroyed(struct intel_context *ce) argument 233 set_context_destroyed(struct intel_context *ce) argument 240 clr_context_destroyed(struct intel_context *ce) argument 246 context_pending_disable(struct intel_context *ce) argument 251 set_context_pending_disable(struct intel_context *ce) argument 257 clr_context_pending_disable(struct intel_context *ce) argument 263 context_banned(struct intel_context *ce) argument 268 set_context_banned(struct intel_context *ce) argument 274 clr_context_banned(struct intel_context *ce) argument 280 context_enabled(struct intel_context *ce) argument 285 set_context_enabled(struct intel_context *ce) argument 291 clr_context_enabled(struct intel_context *ce) argument 297 context_pending_enable(struct intel_context *ce) argument 302 set_context_pending_enable(struct intel_context *ce) argument 308 clr_context_pending_enable(struct intel_context *ce) argument 314 context_registered(struct intel_context *ce) argument 319 set_context_registered(struct intel_context *ce) argument 325 clr_context_registered(struct intel_context *ce) argument 331 context_policy_required(struct intel_context *ce) argument 336 set_context_policy_required(struct intel_context *ce) argument 342 clr_context_policy_required(struct intel_context *ce) argument 348 context_close_done(struct intel_context *ce) argument 353 set_context_close_done(struct intel_context *ce) argument 359 context_blocked(struct intel_context *ce) argument 365 incr_context_blocked(struct intel_context *ce) argument 374 decr_context_blocked(struct intel_context *ce) argument 389 context_guc_id_invalid(struct intel_context *ce) argument 394 set_context_guc_id_invalid(struct intel_context *ce) argument 399 ce_to_guc(struct intel_context *ce) argument 449 __get_parent_scratch_offset(struct intel_context *ce) argument 456 __get_wq_offset(struct intel_context *ce) argument 464 __get_parent_scratch(struct intel_context *ce) argument 481 __get_process_desc_v69(struct intel_context *ce) argument 489 __get_wq_desc_v70(struct intel_context *ce) argument 496 get_wq_pointer(struct intel_context *ce, u32 wqi_size) argument 518 struct intel_context *ce = xa_load(&guc->context_lookup, id); local 579 set_ctx_id_mapping(struct intel_guc *guc, u32 id, struct intel_context *ce) argument 699 struct intel_context *ce = request_to_scheduling_context(rq); local 814 wq_space_until_wrap(struct intel_context *ce) argument 819 write_wqi(struct intel_context *ce, u32 wqi_size) argument 833 guc_wq_noop_append(struct intel_context *ce) argument 852 struct intel_context *ce = request_to_scheduling_context(rq); local 896 struct intel_context *ce = request_to_scheduling_context(rq); local 913 struct intel_context *ce = request_to_scheduling_context(rq); local 990 struct intel_context *ce = request_to_scheduling_context(last); local 1079 struct intel_context *ce; local 1460 __guc_context_update_stats(struct intel_context *ce) argument 1470 guc_context_update_stats(struct intel_context *ce) argument 1485 struct intel_context *ce; local 1713 __context_to_physical_engine(struct intel_context *ce) argument 1723 guc_reset_state(struct intel_context *ce, u32 head, bool scrub) argument 1768 __unwind_incomplete_requests(struct intel_context *ce) argument 1803 __guc_reset_context(struct intel_context *ce, intel_engine_mask_t stalled) argument 1875 struct intel_context *ce; local 1905 guc_cancel_context_requests(struct intel_context *ce) argument 1973 struct intel_context *ce; local 2182 struct intel_context *ce = request_to_scheduling_context(rq); local 2206 new_guc_id(struct intel_guc *guc, struct intel_context *ce) argument 2233 __release_guc_id(struct intel_guc *guc, struct intel_context *ce) argument 2255 release_guc_id(struct intel_guc *guc, struct intel_context *ce) argument 2264 steal_guc_id(struct intel_guc *guc, struct intel_context *ce) argument 2301 assign_guc_id(struct intel_guc *guc, struct intel_context *ce) argument 2330 pin_guc_id(struct intel_guc *guc, struct intel_context *ce) argument 2380 unpin_guc_id(struct intel_guc *guc, struct intel_context *ce) argument 2399 __guc_action_register_multi_lrc_v69(struct intel_guc *guc, struct intel_context *ce, u32 guc_id, u32 offset, bool loop) argument 2423 __guc_action_register_multi_lrc_v70(struct intel_guc *guc, struct intel_context *ce, struct guc_ctxt_registration_info *info, bool loop) argument 2509 register_context_v69(struct intel_guc *guc, struct intel_context *ce, bool loop) argument 2525 register_context_v70(struct intel_guc *guc, struct intel_context *ce, bool loop) argument 2537 register_context(struct intel_context *ce, bool loop) argument 2577 deregister_context(struct intel_context *ce, u32 guc_id) argument 2587 clear_children_join_go_memory(struct intel_context *ce) argument 2597 get_children_go_value(struct intel_context *ce) argument 2602 get_children_join_value(struct intel_context *ce, u8 child_index) argument 2655 guc_context_policy_init_v70(struct intel_context *ce, bool loop) argument 2731 prepare_context_registration_info_v69(struct intel_context *ce) argument 2799 prepare_context_registration_info_v70(struct intel_context *ce, struct guc_ctxt_registration_info *info) argument 2862 try_context_registration(struct intel_context *ce, bool loop) argument 2929 __guc_context_pre_pin(struct intel_context *ce, struct intel_engine_cs *engine, struct i915_gem_ww_ctx *ww, void **vaddr) argument 2937 __guc_context_pin(struct intel_context *ce, struct intel_engine_cs *engine, void *vaddr) argument 2953 guc_context_pre_pin(struct intel_context *ce, struct i915_gem_ww_ctx *ww, void **vaddr) argument 2960 guc_context_pin(struct intel_context *ce, void *vaddr) argument 2970 guc_context_unpin(struct intel_context *ce) argument 2982 guc_context_post_unpin(struct intel_context *ce) argument 2987 __guc_context_sched_enable(struct intel_guc *guc, struct intel_context *ce) argument 3002 __guc_context_sched_disable(struct intel_guc *guc, struct intel_context *ce, u16 guc_id) argument 3021 guc_blocked_fence_complete(struct intel_context *ce) argument 3029 guc_blocked_fence_reinit(struct intel_context *ce) argument 3045 prep_context_pending_disable(struct intel_context *ce) argument 3057 guc_context_block(struct intel_context *ce) argument 3103 context_cant_unblock(struct intel_context *ce) argument 3113 guc_context_unblock(struct intel_context *ce) argument 3146 guc_context_cancel_request(struct intel_context *ce, struct i915_request *rq) argument 3191 guc_context_revoke(struct intel_context *ce, struct i915_request *rq, unsigned int preempt_timeout_ms) argument 3262 bypass_sched_disable(struct intel_guc *guc, struct intel_context *ce) argument 3279 struct intel_context *ce = local 3294 guc_id_pressure(struct intel_guc *guc, struct intel_context *ce) argument 3310 guc_context_sched_disable(struct intel_context *ce) argument 3332 guc_context_close(struct intel_context *ce) argument 3345 guc_lrc_desc_unpin(struct intel_context *ce) argument 3399 __guc_context_destroy(struct intel_context *ce) argument 3424 struct intel_context *ce; local 3449 struct intel_context *ce; local 3507 struct intel_context *ce = container_of(kref, typeof(*ce), ref); local 3542 guc_context_alloc(struct intel_context *ce) argument 3547 __guc_context_set_prio(struct intel_guc *guc, struct intel_context *ce) argument 3567 guc_context_set_prio(struct intel_guc *guc, struct intel_context *ce, u8 prio) argument 3599 add_context_inflight_prio(struct intel_context *ce, u8 guc_prio) argument 3611 sub_context_inflight_prio(struct intel_context *ce, u8 guc_prio) argument 3623 update_context_prio(struct intel_context *ce) argument 3649 struct intel_context *ce = request_to_scheduling_context(rq); local 3671 guc_prio_fini(struct i915_request *rq, struct intel_context *ce) argument 3685 struct intel_context *ce = request_to_scheduling_context(rq); local 3742 __guc_signal_context_fence(struct intel_context *ce) argument 3764 guc_signal_context_fence(struct intel_context *ce) argument 3776 context_needs_register(struct intel_context *ce, bool new_guc_id) argument 3783 guc_context_init(struct intel_context *ce) argument 3804 struct intel_context *ce = request_to_scheduling_context(rq); local 3914 guc_virtual_context_pre_pin(struct intel_context *ce, struct i915_gem_ww_ctx *ww, void **vaddr) argument 3923 guc_virtual_context_pin(struct intel_context *ce, void *vaddr) argument 3936 guc_virtual_context_unpin(struct intel_context *ce) argument 3952 guc_virtual_context_enter(struct intel_context *ce) argument 3963 guc_virtual_context_exit(struct intel_context *ce) argument 3974 guc_virtual_context_alloc(struct intel_context *ce) argument 4007 guc_parent_context_pin(struct intel_context *ce, void *vaddr) argument 4023 guc_child_context_pin(struct intel_context *ce, void *vaddr) argument 4034 guc_parent_context_unpin(struct intel_context *ce) argument 4047 guc_child_context_unpin(struct intel_context *ce) argument 4057 guc_child_context_post_unpin(struct intel_context *ce) argument 4069 struct intel_context *ce = container_of(kref, typeof(*ce), ref); local 4147 struct intel_context *parent = NULL, *ce, *err; local 4261 struct intel_context *ce = request_to_scheduling_context(rq); local 4284 struct intel_context *ce = request_to_scheduling_context(rq); local 4372 guc_kernel_context_pin(struct intel_guc *guc, struct intel_context *ce) argument 4425 struct intel_context *ce; local 4798 struct intel_context *ce; local 4971 struct intel_context *ce; local 5022 struct intel_context *ce; local 5096 capture_error_state(struct intel_guc *guc, struct intel_context *ce) argument 5135 guc_context_replay(struct intel_context *ce) argument 5143 guc_handle_context_reset(struct intel_guc *guc, struct intel_context *ce) argument 5165 struct intel_context *ce; local 5307 struct intel_context *ce; local 5369 struct intel_context *ce; local 5440 guc_log_context_priority(struct drm_printer *p, struct intel_context *ce) argument 5455 guc_log_context(struct drm_printer *p, struct intel_context *ce) argument 5477 struct intel_context *ce; local 5521 get_children_go_addr(struct intel_context *ce) argument 5530 get_children_join_addr(struct intel_context *ce, u8 child_index) argument 5548 struct intel_context *ce = rq->context; local 5595 struct intel_context *ce = rq->context; local 5639 struct intel_context *ce = rq->context; local 5689 struct intel_context *ce = rq->context; local 5733 struct intel_context *ce = rq->context; local 5765 struct intel_context *ce = rq->context; local [all...] |
H A D | selftest_guc.c | 23 static struct i915_request *nop_user_request(struct intel_context *ce, argument 29 rq = intel_context_create_request(ce); 57 struct intel_context *ce; local 67 ce = intel_context_create(engine); 68 if (IS_ERR(ce)) { 69 ret = PTR_ERR(ce); 70 gt_err(gt, "Failed to create context %d: %pe\n", i, ce); 76 ce->drop_schedule_enable = true; 79 ce->drop_schedule_disable = true; 82 ce 151 struct intel_context **ce; local 313 struct intel_context *ce; local [all...] |
/linux-master/drivers/clocksource/ |
H A D | timer-digicolor.c | 58 struct clock_event_device ce; member in struct:digicolor_timer 64 static struct digicolor_timer *dc_timer(struct clock_event_device *ce) argument 66 return container_of(ce, struct digicolor_timer, ce); 69 static inline void dc_timer_disable(struct clock_event_device *ce) argument 71 struct digicolor_timer *dt = dc_timer(ce); 75 static inline void dc_timer_enable(struct clock_event_device *ce, u32 mode) argument 77 struct digicolor_timer *dt = dc_timer(ce); 81 static inline void dc_timer_set_count(struct clock_event_device *ce, argument 84 struct digicolor_timer *dt = dc_timer(ce); 88 digicolor_clkevt_shutdown(struct clock_event_device *ce) argument 94 digicolor_clkevt_set_oneshot(struct clock_event_device *ce) argument 101 digicolor_clkevt_set_periodic(struct clock_event_device *ce) argument 111 digicolor_clkevt_next_event(unsigned long evt, struct clock_event_device *ce) argument [all...] |
H A D | mps2-timer.c | 54 static int mps2_timer_shutdown(struct clock_event_device *ce) argument 56 clockevent_mps2_writel(0, ce, TIMER_RELOAD); 57 clockevent_mps2_writel(0, ce, TIMER_CTRL); 62 static int mps2_timer_set_next_event(unsigned long next, struct clock_event_device *ce) argument 64 clockevent_mps2_writel(next, ce, TIMER_VALUE); 65 clockevent_mps2_writel(TIMER_CTRL_IE | TIMER_CTRL_ENABLE, ce, TIMER_CTRL); 70 static int mps2_timer_set_periodic(struct clock_event_device *ce) argument 72 u32 clock_count_per_tick = to_mps2_clkevt(ce)->clock_count_per_tick; 74 clockevent_mps2_writel(clock_count_per_tick, ce, TIMER_RELOAD); 75 clockevent_mps2_writel(clock_count_per_tick, ce, TIMER_VALU 83 struct clockevent_mps2 *ce = dev_id; local 102 struct clockevent_mps2 *ce; local [all...] |
H A D | timer-sun5i.c | 58 static void sun5i_clkevt_sync(struct sun5i_timer *ce) argument 60 u32 old = readl(ce->base + TIMER_CNTVAL_LO_REG(1)); 62 while ((old - readl(ce->base + TIMER_CNTVAL_LO_REG(1))) < TIMER_SYNC_TICKS) 66 static void sun5i_clkevt_time_stop(struct sun5i_timer *ce, u8 timer) argument 68 u32 val = readl(ce->base + TIMER_CTL_REG(timer)); 69 writel(val & ~TIMER_CTL_ENABLE, ce->base + TIMER_CTL_REG(timer)); 71 sun5i_clkevt_sync(ce); 74 static void sun5i_clkevt_time_setup(struct sun5i_timer *ce, u8 timer, u32 delay) argument 76 writel(delay, ce->base + TIMER_INTVAL_LO_REG(timer)); 79 static void sun5i_clkevt_time_start(struct sun5i_timer *ce, u argument 94 struct sun5i_timer *ce = clkevt_to_sun5i_timer(clkevt); local 102 struct sun5i_timer *ce = clkevt_to_sun5i_timer(clkevt); local 111 struct sun5i_timer *ce = clkevt_to_sun5i_timer(clkevt); local 122 struct sun5i_timer *ce = clkevt_to_sun5i_timer(clkevt); local 133 struct sun5i_timer *ce = dev_id; local 202 struct sun5i_timer *ce = platform_get_drvdata(pdev); local [all...] |
/linux-master/drivers/base/power/ |
H A D | clock_ops.c | 145 * @ce: PM clock entry corresponding to the clock. 147 static inline void __pm_clk_enable(struct device *dev, struct pm_clock_entry *ce) argument 151 switch (ce->status) { 153 ret = clk_prepare_enable(ce->clk); 156 ret = clk_enable(ce->clk); 162 ce->status = PCE_STATUS_ENABLED; 165 __func__, ce->clk, ret); 171 * @ce: PM clock entry corresponding to the clock. 173 static void pm_clk_acquire(struct device *dev, struct pm_clock_entry *ce) argument 175 if (!ce 199 struct pm_clock_entry *ce; local 354 __pm_clk_remove(struct pm_clock_entry *ce) argument 390 struct pm_clock_entry *ce; local 430 struct pm_clock_entry *ce; local 497 struct pm_clock_entry *ce, *c; local 546 struct pm_clock_entry *ce; local 584 struct pm_clock_entry *ce; local [all...] |
/linux-master/fs/smb/client/ |
H A D | dfs_cache.c | 108 static inline bool cache_entry_expired(const struct cache_entry *ce) argument 113 return timespec64_compare(&ts, &ce->etime) >= 0; 116 static inline void free_tgts(struct cache_entry *ce) argument 120 list_for_each_entry_safe(t, n, &ce->tlist, list) { 127 static inline void flush_cache_ent(struct cache_entry *ce) argument 129 hlist_del_init(&ce->hlist); 130 kfree(ce->path); 131 free_tgts(ce); 133 kmem_cache_free(cache_slab, ce); 143 struct cache_entry *ce; local 158 struct cache_entry *ce; local 226 dump_tgts(const struct cache_entry *ce) argument 237 dump_ce(const struct cache_entry *ce) argument 340 get_tgt_name(const struct cache_entry *ce) argument 382 copy_ref_data(const struct dfs_info3_param *refs, int numrefs, struct cache_entry *ce, const char *tgthint) argument 422 struct cache_entry *ce; local 447 struct cache_entry *ce; local 479 struct cache_entry *ce; local 537 struct cache_entry *ce; local 559 struct cache_entry *ce; local 620 update_cache_entry_locked(struct cache_entry *ce, const struct dfs_info3_param *refs, int numrefs) argument 689 struct cache_entry *ce; local 756 setup_referral(const char *path, struct cache_entry *ce, struct dfs_info3_param *ref, const char *target) argument 790 get_targets(struct cache_entry *ce, struct dfs_cache_tgt_list *tl) argument 862 struct cache_entry *ce; local 908 struct cache_entry *ce; local 949 struct cache_entry *ce; local 994 struct cache_entry *ce; local 1182 struct cache_entry *ce; local [all...] |
/linux-master/drivers/gpu/drm/i915/selftests/ |
H A D | mock_request.h | 33 mock_request(struct intel_context *ce, unsigned long delay);
|
/linux-master/drivers/crypto/allwinner/ |
H A D | Makefile | 2 obj-$(CONFIG_CRYPTO_DEV_SUN8I_CE) += sun8i-ce/
|
/linux-master/arch/sparc/kernel/ |
H A D | time_32.c | 120 struct clock_event_device *ce = &timer_ce; local 124 ce->name = "timer_ce"; 125 ce->rating = 100; 126 ce->features = CLOCK_EVT_FEAT_PERIODIC; 127 ce->set_state_shutdown = timer_ce_shutdown; 128 ce->set_state_periodic = timer_ce_set_periodic; 129 ce->tick_resume = timer_ce_set_periodic; 130 ce->cpumask = cpu_possible_mask; 131 ce->shift = 32; 132 ce 213 struct clock_event_device *ce = &per_cpu(sparc32_clockevent, cpu); local [all...] |
/linux-master/drivers/of/ |
H A D | dynamic.c | 520 static void __of_changeset_entry_destroy(struct of_changeset_entry *ce) argument 522 if (ce->action == OF_RECONFIG_ATTACH_NODE && 523 of_node_check_flag(ce->np, OF_OVERLAY)) { 524 if (kref_read(&ce->np->kobj.kref) > 1) { 526 kref_read(&ce->np->kobj.kref), ce->np); 528 of_node_set_flag(ce->np, OF_OVERLAY_FREE_CSET); 532 of_node_put(ce->np); 533 list_del(&ce->node); 534 kfree(ce); 537 __of_changeset_entry_invert(struct of_changeset_entry *ce, struct of_changeset_entry *rce) argument 567 __of_changeset_entry_notify(struct of_changeset_entry *ce, bool revert) argument 602 __of_changeset_entry_apply(struct of_changeset_entry *ce) argument 637 __of_changeset_entry_revert(struct of_changeset_entry *ce) argument 669 struct of_changeset_entry *ce, *cen; local 699 struct of_changeset_entry *ce; local 728 struct of_changeset_entry *ce; local 803 struct of_changeset_entry *ce; local 829 struct of_changeset_entry *ce; local 904 struct of_changeset_entry *ce; local [all...] |