Searched refs:ce (Results 1 - 25 of 212) sorted by relevance

123456789

/linux-master/drivers/gpu/drm/i915/gt/
H A Dintel_context.h22 #define CE_TRACE(ce, fmt, ...) do { \
23 const struct intel_context *ce__ = (ce); \
33 void intel_context_init(struct intel_context *ce,
35 void intel_context_fini(struct intel_context *ce);
43 int intel_context_alloc_state(struct intel_context *ce);
45 void intel_context_free(struct intel_context *ce);
47 int intel_context_reconfigure_sseu(struct intel_context *ce,
52 static inline bool intel_context_is_child(struct intel_context *ce) argument
54 return !!ce->parallel.parent;
57 static inline bool intel_context_is_parent(struct intel_context *ce) argument
65 intel_context_to_parent(struct intel_context *ce) argument
83 intel_context_is_parallel(struct intel_context *ce) argument
122 intel_context_is_pinned(struct intel_context *ce) argument
127 intel_context_cancel_request(struct intel_context *ce, struct i915_request *rq) argument
150 intel_context_pin_if_active(struct intel_context *ce) argument
155 intel_context_pin(struct intel_context *ce) argument
163 intel_context_pin_ww(struct intel_context *ce, struct i915_gem_ww_ctx *ww) argument
172 __intel_context_pin(struct intel_context *ce) argument
180 intel_context_sched_disable_unpin(struct intel_context *ce) argument
185 intel_context_unpin(struct intel_context *ce) argument
208 intel_context_enter(struct intel_context *ce) argument
218 intel_context_mark_active(struct intel_context *ce) argument
225 intel_context_exit(struct intel_context *ce) argument
236 intel_context_get(struct intel_context *ce) argument
242 intel_context_put(struct intel_context *ce) argument
280 intel_context_is_barrier(const struct intel_context *ce) argument
285 intel_context_close(struct intel_context *ce) argument
293 intel_context_is_closed(const struct intel_context *ce) argument
298 intel_context_has_inflight(const struct intel_context *ce) argument
303 intel_context_use_semaphores(const struct intel_context *ce) argument
308 intel_context_set_use_semaphores(struct intel_context *ce) argument
313 intel_context_clear_use_semaphores(struct intel_context *ce) argument
318 intel_context_is_banned(const struct intel_context *ce) argument
323 intel_context_set_banned(struct intel_context *ce) argument
330 intel_context_is_schedulable(const struct intel_context *ce) argument
336 intel_context_is_exiting(const struct intel_context *ce) argument
341 intel_context_set_exiting(struct intel_context *ce) argument
349 intel_context_force_single_submission(const struct intel_context *ce) argument
355 intel_context_set_single_submission(struct intel_context *ce) argument
361 intel_context_nopreempt(const struct intel_context *ce) argument
367 intel_context_set_nopreempt(struct intel_context *ce) argument
373 intel_context_clear_nopreempt(struct intel_context *ce) argument
[all...]
H A Dintel_context.c27 struct intel_context *ce = container_of(rcu, typeof(*ce), rcu); local
29 trace_intel_context_free(ce);
30 kmem_cache_free(slab_ce, ce);
33 void intel_context_free(struct intel_context *ce) argument
35 call_rcu(&ce->rcu, rcu_context_free);
41 struct intel_context *ce; local
43 ce = intel_context_alloc();
44 if (!ce)
47 intel_context_init(ce, engin
52 intel_context_alloc_state(struct intel_context *ce) argument
90 intel_context_active_acquire(struct intel_context *ce) argument
109 intel_context_active_release(struct intel_context *ce) argument
176 intel_context_pre_pin(struct intel_context *ce, struct i915_gem_ww_ctx *ww) argument
208 intel_context_post_unpin(struct intel_context *ce) argument
217 __intel_context_do_pin_ww(struct intel_context *ce, struct i915_gem_ww_ctx *ww) argument
310 __intel_context_do_pin(struct intel_context *ce) argument
327 __intel_context_do_unpin(struct intel_context *ce, int sub) argument
350 struct intel_context *ce = container_of(active, typeof(*ce), active); local
363 struct intel_context *ce = container_of(active, typeof(*ce), active); local
390 intel_context_init(struct intel_context *ce, struct intel_engine_cs *engine) argument
436 intel_context_fini(struct intel_context *ce) argument
468 intel_context_enter_engine(struct intel_context *ce) argument
474 intel_context_exit_engine(struct intel_context *ce) argument
480 intel_context_prepare_remote_request(struct intel_context *ce, struct i915_request *rq) argument
507 intel_context_create_request(struct intel_context *ce) argument
545 intel_context_get_active_request(struct intel_context *ce) argument
595 intel_context_get_total_runtime_ns(struct intel_context *ce) argument
613 intel_context_get_avg_runtime_ns(struct intel_context *ce) argument
623 intel_context_ban(struct intel_context *ce, struct i915_request *rq) argument
636 intel_context_revoke(struct intel_context *ce) argument
[all...]
H A Dintel_context_param.h14 intel_context_set_watchdog_us(struct intel_context *ce, u64 timeout_us) argument
16 ce->watchdog.timeout_us = timeout_us;
H A Dintel_context_sseu.c17 const struct intel_context *ce,
27 offset = i915_ggtt_offset(ce->state) +
41 gen8_modify_rpcs(struct intel_context *ce, const struct intel_sseu sseu) argument
46 lockdep_assert_held(&ce->pin_mutex);
54 if (!intel_context_pin_if_active(ce))
57 rq = intel_engine_create_kernel_request(ce->engine);
64 ret = intel_context_prepare_remote_request(ce, rq);
66 ret = gen8_emit_rpcs_config(rq, ce, sseu);
70 intel_context_unpin(ce);
75 intel_context_reconfigure_sseu(struct intel_context *ce, argument
16 gen8_emit_rpcs_config(struct i915_request *rq, const struct intel_context *ce, const struct intel_sseu sseu) argument
[all...]
H A Dintel_lrc.h36 int lrc_alloc(struct intel_context *ce,
38 void lrc_reset(struct intel_context *ce);
39 void lrc_fini(struct intel_context *ce);
43 lrc_pre_pin(struct intel_context *ce,
48 lrc_pin(struct intel_context *ce,
51 void lrc_unpin(struct intel_context *ce);
52 void lrc_post_unpin(struct intel_context *ce);
54 void lrc_init_state(struct intel_context *ce,
58 void lrc_init_regs(const struct intel_context *ce,
61 void lrc_reset_regs(const struct intel_context *ce,
116 lrc_runtime_start(struct intel_context *ce) argument
129 lrc_runtime_stop(struct intel_context *ce) argument
[all...]
H A Dintel_engine_pm.c35 static void dbg_poison_ce(struct intel_context *ce) argument
40 if (ce->state) {
41 struct drm_i915_gem_object *obj = ce->state->obj;
42 int type = intel_gt_coherent_map_type(ce->engine->gt, obj, true);
62 struct intel_context *ce; local
69 ce = engine->kernel_context;
70 if (ce) {
71 GEM_BUG_ON(test_bit(CONTEXT_VALID_BIT, &ce->flags));
74 while (unlikely(intel_context_inflight(ce)))
78 dbg_poison_ce(ce);
153 struct intel_context *ce = engine->kernel_context; local
313 struct intel_context *ce; local
[all...]
/linux-master/drivers/crypto/allwinner/sun8i-ce/
H A DMakefile1 obj-$(CONFIG_CRYPTO_DEV_SUN8I_CE) += sun8i-ce.o
2 sun8i-ce-y += sun8i-ce-core.o sun8i-ce-cipher.o
3 sun8i-ce-$(CONFIG_CRYPTO_DEV_SUN8I_CE_HASH) += sun8i-ce-hash.o
4 sun8i-ce-$(CONFIG_CRYPTO_DEV_SUN8I_CE_PRNG) += sun8i-ce-prng.o
5 sun8i-ce-$(CONFIG_CRYPTO_DEV_SUN8I_CE_TRNG) += sun8i-ce
[all...]
H A Dsun8i-ce-trng.c3 * sun8i-ce-trng.c - hardware cryptographic offloader for
12 #include "sun8i-ce.h"
25 struct sun8i_ce_dev *ce; local
35 ce = container_of(rng, struct sun8i_ce_dev, trng);
46 ce->hwrng_stat_req++;
47 ce->hwrng_stat_bytes += todo;
50 dma_dst = dma_map_single(ce->dev, d, todo, DMA_FROM_DEVICE);
51 if (dma_mapping_error(ce->dev, dma_dst)) {
52 dev_err(ce->dev, "Cannot DMA MAP DST\n");
57 err = pm_runtime_resume_and_get(ce
101 sun8i_ce_hwrng_register(struct sun8i_ce_dev *ce) argument
118 sun8i_ce_hwrng_unregister(struct sun8i_ce_dev *ce) argument
[all...]
H A Dsun8i-ce-core.c3 * sun8i-ce-core.c - hardware cryptographic offloader for
31 #include "sun8i-ce.h"
153 int sun8i_ce_get_engine_number(struct sun8i_ce_dev *ce) argument
155 return atomic_inc_return(&ce->flow) % (MAXFLOW - 1);
158 int sun8i_ce_run_task(struct sun8i_ce_dev *ce, int flow, const char *name) argument
162 struct ce_task *cet = ce->chanlist[flow].tl;
165 ce->chanlist[flow].stat_req++;
168 mutex_lock(&ce->mlock);
170 v = readl(ce->base + CE_ICR);
172 writel(v, ce
262 struct sun8i_ce_dev *ce = (struct sun8i_ce_dev *)data; local
688 sun8i_ce_free_chanlist(struct sun8i_ce_dev *ce, int i) argument
703 sun8i_ce_allocate_chanlist(struct sun8i_ce_dev *ce) argument
762 struct sun8i_ce_dev *ce = dev_get_drvdata(dev); local
773 struct sun8i_ce_dev *ce = dev_get_drvdata(dev); local
801 sun8i_ce_pm_init(struct sun8i_ce_dev *ce) argument
815 sun8i_ce_pm_exit(struct sun8i_ce_dev *ce) argument
820 sun8i_ce_get_clks(struct sun8i_ce_dev *ce) argument
860 sun8i_ce_register_algs(struct sun8i_ce_dev *ce) argument
941 sun8i_ce_unregister_algs(struct sun8i_ce_dev *ce) argument
970 struct sun8i_ce_dev *ce; local
1070 struct sun8i_ce_dev *ce = platform_get_drvdata(pdev); local
[all...]
H A Dsun8i-ce-prng.c3 * sun8i-ce-prng.c - hardware cryptographic offloader for
12 #include "sun8i-ce.h"
61 struct sun8i_ce_dev *ce; local
72 ce = algt->ce;
75 dev_err(ce->dev, "not seeded\n");
89 dev_dbg(ce->dev, "%s PRNG slen=%u dlen=%u todo=%u multi=%u\n", __func__,
97 dma_iv = dma_map_single(ce->dev, ctx->seed, ctx->slen, DMA_TO_DEVICE);
98 if (dma_mapping_error(ce->dev, dma_iv)) {
99 dev_err(ce
[all...]
/linux-master/drivers/crypto/gemini/
H A DMakefile1 obj-$(CONFIG_CRYPTO_DEV_SL3516) += sl3516-ce.o
2 sl3516-ce-y += sl3516-ce-core.o sl3516-ce-cipher.o sl3516-ce-rng.o
H A Dsl3516-ce-core.c3 * sl3516-ce-core.c - hardware cryptographic offloader for Storlink SL3516 SoC
28 #include "sl3516-ce.h"
30 static int sl3516_ce_desc_init(struct sl3516_ce_dev *ce) argument
35 ce->tx = dma_alloc_coherent(ce->dev, sz, &ce->dtx, GFP_KERNEL);
36 if (!ce->tx)
38 ce->rx = dma_alloc_coherent(ce->dev, sz, &ce
67 sl3516_ce_free_descs(struct sl3516_ce_dev *ce) argument
77 start_dma_tx(struct sl3516_ce_dev *ce) argument
87 start_dma_rx(struct sl3516_ce_dev *ce) argument
99 get_desc_tx(struct sl3516_ce_dev *ce) argument
110 get_desc_rx(struct sl3516_ce_dev *ce) argument
121 sl3516_ce_run_task(struct sl3516_ce_dev *ce, struct sl3516_ce_cipher_req_ctx *rctx, const char *name) argument
189 struct sl3516_ce_dev *ce = (struct sl3516_ce_dev *)data; local
250 struct sl3516_ce_dev *ce = seq->private; local
282 sl3516_ce_register_algs(struct sl3516_ce_dev *ce) argument
309 sl3516_ce_unregister_algs(struct sl3516_ce_dev *ce) argument
326 sl3516_ce_start(struct sl3516_ce_dev *ce) argument
341 struct sl3516_ce_dev *ce = dev_get_drvdata(dev); local
350 struct sl3516_ce_dev *ce = dev_get_drvdata(dev); local
376 sl3516_ce_pm_init(struct sl3516_ce_dev *ce) argument
390 sl3516_ce_pm_exit(struct sl3516_ce_dev *ce) argument
397 struct sl3516_ce_dev *ce; local
510 struct sl3516_ce_dev *ce = platform_get_drvdata(pdev); local
[all...]
H A Dsl3516-ce-rng.c3 * sl3516-ce-rng.c - hardware cryptographic offloader for SL3516 SoC.
9 #include "sl3516-ce.h"
15 struct sl3516_ce_dev *ce; local
20 ce = container_of(rng, struct sl3516_ce_dev, trng);
23 ce->hwrng_stat_req++;
24 ce->hwrng_stat_bytes += max;
27 err = pm_runtime_get_sync(ce->dev);
29 pm_runtime_put_noidle(ce->dev);
34 *data = readl(ce->base + IPSEC_RAND_NUM_REG);
39 pm_runtime_put(ce
44 sl3516_ce_rng_register(struct sl3516_ce_dev *ce) argument
58 sl3516_ce_rng_unregister(struct sl3516_ce_dev *ce) argument
[all...]
/linux-master/arch/arm64/crypto/
H A DMakefile8 obj-$(CONFIG_CRYPTO_SHA1_ARM64_CE) += sha1-ce.o
9 sha1-ce-y := sha1-ce-glue.o sha1-ce-core.o
11 obj-$(CONFIG_CRYPTO_SHA2_ARM64_CE) += sha2-ce.o
12 sha2-ce-y := sha2-ce-glue.o sha2-ce-core.o
14 obj-$(CONFIG_CRYPTO_SHA512_ARM64_CE) += sha512-ce.o
15 sha512-ce
[all...]
/linux-master/drivers/gpu/drm/i915/gt/uc/
H A Dintel_guc_submission.c112 * ce->guc_state.lock
113 * Protects everything under ce->guc_state. Ensures that a context is in the
121 * sched_engine->lock -> ce->guc_state.lock
122 * guc->submission_state.lock -> ce->guc_state.lock
181 static inline void init_sched_state(struct intel_context *ce) argument
183 lockdep_assert_held(&ce->guc_state.lock);
184 ce->guc_state.sched_state &= SCHED_STATE_BLOCKED_MASK;
198 static bool sched_state_is_init(struct intel_context *ce) argument
200 return !(ce->guc_state.sched_state & ~SCHED_STATE_VALID_INIT);
204 context_wait_for_deregister_to_register(struct intel_context *ce) argument
211 set_context_wait_for_deregister_to_register(struct intel_context *ce) argument
219 clr_context_wait_for_deregister_to_register(struct intel_context *ce) argument
227 context_destroyed(struct intel_context *ce) argument
233 set_context_destroyed(struct intel_context *ce) argument
240 clr_context_destroyed(struct intel_context *ce) argument
246 context_pending_disable(struct intel_context *ce) argument
251 set_context_pending_disable(struct intel_context *ce) argument
257 clr_context_pending_disable(struct intel_context *ce) argument
263 context_banned(struct intel_context *ce) argument
268 set_context_banned(struct intel_context *ce) argument
274 clr_context_banned(struct intel_context *ce) argument
280 context_enabled(struct intel_context *ce) argument
285 set_context_enabled(struct intel_context *ce) argument
291 clr_context_enabled(struct intel_context *ce) argument
297 context_pending_enable(struct intel_context *ce) argument
302 set_context_pending_enable(struct intel_context *ce) argument
308 clr_context_pending_enable(struct intel_context *ce) argument
314 context_registered(struct intel_context *ce) argument
319 set_context_registered(struct intel_context *ce) argument
325 clr_context_registered(struct intel_context *ce) argument
331 context_policy_required(struct intel_context *ce) argument
336 set_context_policy_required(struct intel_context *ce) argument
342 clr_context_policy_required(struct intel_context *ce) argument
348 context_close_done(struct intel_context *ce) argument
353 set_context_close_done(struct intel_context *ce) argument
359 context_blocked(struct intel_context *ce) argument
365 incr_context_blocked(struct intel_context *ce) argument
374 decr_context_blocked(struct intel_context *ce) argument
389 context_guc_id_invalid(struct intel_context *ce) argument
394 set_context_guc_id_invalid(struct intel_context *ce) argument
399 ce_to_guc(struct intel_context *ce) argument
449 __get_parent_scratch_offset(struct intel_context *ce) argument
456 __get_wq_offset(struct intel_context *ce) argument
464 __get_parent_scratch(struct intel_context *ce) argument
481 __get_process_desc_v69(struct intel_context *ce) argument
489 __get_wq_desc_v70(struct intel_context *ce) argument
496 get_wq_pointer(struct intel_context *ce, u32 wqi_size) argument
518 struct intel_context *ce = xa_load(&guc->context_lookup, id); local
579 set_ctx_id_mapping(struct intel_guc *guc, u32 id, struct intel_context *ce) argument
699 struct intel_context *ce = request_to_scheduling_context(rq); local
814 wq_space_until_wrap(struct intel_context *ce) argument
819 write_wqi(struct intel_context *ce, u32 wqi_size) argument
833 guc_wq_noop_append(struct intel_context *ce) argument
852 struct intel_context *ce = request_to_scheduling_context(rq); local
896 struct intel_context *ce = request_to_scheduling_context(rq); local
913 struct intel_context *ce = request_to_scheduling_context(rq); local
990 struct intel_context *ce = request_to_scheduling_context(last); local
1079 struct intel_context *ce; local
1460 __guc_context_update_stats(struct intel_context *ce) argument
1470 guc_context_update_stats(struct intel_context *ce) argument
1485 struct intel_context *ce; local
1713 __context_to_physical_engine(struct intel_context *ce) argument
1723 guc_reset_state(struct intel_context *ce, u32 head, bool scrub) argument
1768 __unwind_incomplete_requests(struct intel_context *ce) argument
1803 __guc_reset_context(struct intel_context *ce, intel_engine_mask_t stalled) argument
1875 struct intel_context *ce; local
1905 guc_cancel_context_requests(struct intel_context *ce) argument
1973 struct intel_context *ce; local
2182 struct intel_context *ce = request_to_scheduling_context(rq); local
2206 new_guc_id(struct intel_guc *guc, struct intel_context *ce) argument
2233 __release_guc_id(struct intel_guc *guc, struct intel_context *ce) argument
2255 release_guc_id(struct intel_guc *guc, struct intel_context *ce) argument
2264 steal_guc_id(struct intel_guc *guc, struct intel_context *ce) argument
2301 assign_guc_id(struct intel_guc *guc, struct intel_context *ce) argument
2330 pin_guc_id(struct intel_guc *guc, struct intel_context *ce) argument
2380 unpin_guc_id(struct intel_guc *guc, struct intel_context *ce) argument
2399 __guc_action_register_multi_lrc_v69(struct intel_guc *guc, struct intel_context *ce, u32 guc_id, u32 offset, bool loop) argument
2423 __guc_action_register_multi_lrc_v70(struct intel_guc *guc, struct intel_context *ce, struct guc_ctxt_registration_info *info, bool loop) argument
2509 register_context_v69(struct intel_guc *guc, struct intel_context *ce, bool loop) argument
2525 register_context_v70(struct intel_guc *guc, struct intel_context *ce, bool loop) argument
2537 register_context(struct intel_context *ce, bool loop) argument
2577 deregister_context(struct intel_context *ce, u32 guc_id) argument
2587 clear_children_join_go_memory(struct intel_context *ce) argument
2597 get_children_go_value(struct intel_context *ce) argument
2602 get_children_join_value(struct intel_context *ce, u8 child_index) argument
2655 guc_context_policy_init_v70(struct intel_context *ce, bool loop) argument
2731 prepare_context_registration_info_v69(struct intel_context *ce) argument
2799 prepare_context_registration_info_v70(struct intel_context *ce, struct guc_ctxt_registration_info *info) argument
2862 try_context_registration(struct intel_context *ce, bool loop) argument
2929 __guc_context_pre_pin(struct intel_context *ce, struct intel_engine_cs *engine, struct i915_gem_ww_ctx *ww, void **vaddr) argument
2937 __guc_context_pin(struct intel_context *ce, struct intel_engine_cs *engine, void *vaddr) argument
2953 guc_context_pre_pin(struct intel_context *ce, struct i915_gem_ww_ctx *ww, void **vaddr) argument
2960 guc_context_pin(struct intel_context *ce, void *vaddr) argument
2970 guc_context_unpin(struct intel_context *ce) argument
2982 guc_context_post_unpin(struct intel_context *ce) argument
2987 __guc_context_sched_enable(struct intel_guc *guc, struct intel_context *ce) argument
3002 __guc_context_sched_disable(struct intel_guc *guc, struct intel_context *ce, u16 guc_id) argument
3021 guc_blocked_fence_complete(struct intel_context *ce) argument
3029 guc_blocked_fence_reinit(struct intel_context *ce) argument
3045 prep_context_pending_disable(struct intel_context *ce) argument
3057 guc_context_block(struct intel_context *ce) argument
3103 context_cant_unblock(struct intel_context *ce) argument
3113 guc_context_unblock(struct intel_context *ce) argument
3146 guc_context_cancel_request(struct intel_context *ce, struct i915_request *rq) argument
3191 guc_context_revoke(struct intel_context *ce, struct i915_request *rq, unsigned int preempt_timeout_ms) argument
3262 bypass_sched_disable(struct intel_guc *guc, struct intel_context *ce) argument
3279 struct intel_context *ce = local
3294 guc_id_pressure(struct intel_guc *guc, struct intel_context *ce) argument
3310 guc_context_sched_disable(struct intel_context *ce) argument
3332 guc_context_close(struct intel_context *ce) argument
3345 guc_lrc_desc_unpin(struct intel_context *ce) argument
3399 __guc_context_destroy(struct intel_context *ce) argument
3424 struct intel_context *ce; local
3449 struct intel_context *ce; local
3507 struct intel_context *ce = container_of(kref, typeof(*ce), ref); local
3542 guc_context_alloc(struct intel_context *ce) argument
3547 __guc_context_set_prio(struct intel_guc *guc, struct intel_context *ce) argument
3567 guc_context_set_prio(struct intel_guc *guc, struct intel_context *ce, u8 prio) argument
3599 add_context_inflight_prio(struct intel_context *ce, u8 guc_prio) argument
3611 sub_context_inflight_prio(struct intel_context *ce, u8 guc_prio) argument
3623 update_context_prio(struct intel_context *ce) argument
3649 struct intel_context *ce = request_to_scheduling_context(rq); local
3671 guc_prio_fini(struct i915_request *rq, struct intel_context *ce) argument
3685 struct intel_context *ce = request_to_scheduling_context(rq); local
3742 __guc_signal_context_fence(struct intel_context *ce) argument
3764 guc_signal_context_fence(struct intel_context *ce) argument
3776 context_needs_register(struct intel_context *ce, bool new_guc_id) argument
3783 guc_context_init(struct intel_context *ce) argument
3804 struct intel_context *ce = request_to_scheduling_context(rq); local
3914 guc_virtual_context_pre_pin(struct intel_context *ce, struct i915_gem_ww_ctx *ww, void **vaddr) argument
3923 guc_virtual_context_pin(struct intel_context *ce, void *vaddr) argument
3936 guc_virtual_context_unpin(struct intel_context *ce) argument
3952 guc_virtual_context_enter(struct intel_context *ce) argument
3963 guc_virtual_context_exit(struct intel_context *ce) argument
3974 guc_virtual_context_alloc(struct intel_context *ce) argument
4007 guc_parent_context_pin(struct intel_context *ce, void *vaddr) argument
4023 guc_child_context_pin(struct intel_context *ce, void *vaddr) argument
4034 guc_parent_context_unpin(struct intel_context *ce) argument
4047 guc_child_context_unpin(struct intel_context *ce) argument
4057 guc_child_context_post_unpin(struct intel_context *ce) argument
4069 struct intel_context *ce = container_of(kref, typeof(*ce), ref); local
4147 struct intel_context *parent = NULL, *ce, *err; local
4261 struct intel_context *ce = request_to_scheduling_context(rq); local
4284 struct intel_context *ce = request_to_scheduling_context(rq); local
4372 guc_kernel_context_pin(struct intel_guc *guc, struct intel_context *ce) argument
4425 struct intel_context *ce; local
4798 struct intel_context *ce; local
4971 struct intel_context *ce; local
5022 struct intel_context *ce; local
5096 capture_error_state(struct intel_guc *guc, struct intel_context *ce) argument
5135 guc_context_replay(struct intel_context *ce) argument
5143 guc_handle_context_reset(struct intel_guc *guc, struct intel_context *ce) argument
5165 struct intel_context *ce; local
5307 struct intel_context *ce; local
5369 struct intel_context *ce; local
5440 guc_log_context_priority(struct drm_printer *p, struct intel_context *ce) argument
5455 guc_log_context(struct drm_printer *p, struct intel_context *ce) argument
5477 struct intel_context *ce; local
5521 get_children_go_addr(struct intel_context *ce) argument
5530 get_children_join_addr(struct intel_context *ce, u8 child_index) argument
5548 struct intel_context *ce = rq->context; local
5595 struct intel_context *ce = rq->context; local
5639 struct intel_context *ce = rq->context; local
5689 struct intel_context *ce = rq->context; local
5733 struct intel_context *ce = rq->context; local
5765 struct intel_context *ce = rq->context; local
[all...]
H A Dselftest_guc.c23 static struct i915_request *nop_user_request(struct intel_context *ce, argument
29 rq = intel_context_create_request(ce);
57 struct intel_context *ce; local
67 ce = intel_context_create(engine);
68 if (IS_ERR(ce)) {
69 ret = PTR_ERR(ce);
70 gt_err(gt, "Failed to create context %d: %pe\n", i, ce);
76 ce->drop_schedule_enable = true;
79 ce->drop_schedule_disable = true;
82 ce
151 struct intel_context **ce; local
313 struct intel_context *ce; local
[all...]
/linux-master/drivers/clocksource/
H A Dtimer-digicolor.c58 struct clock_event_device ce; member in struct:digicolor_timer
64 static struct digicolor_timer *dc_timer(struct clock_event_device *ce) argument
66 return container_of(ce, struct digicolor_timer, ce);
69 static inline void dc_timer_disable(struct clock_event_device *ce) argument
71 struct digicolor_timer *dt = dc_timer(ce);
75 static inline void dc_timer_enable(struct clock_event_device *ce, u32 mode) argument
77 struct digicolor_timer *dt = dc_timer(ce);
81 static inline void dc_timer_set_count(struct clock_event_device *ce, argument
84 struct digicolor_timer *dt = dc_timer(ce);
88 digicolor_clkevt_shutdown(struct clock_event_device *ce) argument
94 digicolor_clkevt_set_oneshot(struct clock_event_device *ce) argument
101 digicolor_clkevt_set_periodic(struct clock_event_device *ce) argument
111 digicolor_clkevt_next_event(unsigned long evt, struct clock_event_device *ce) argument
[all...]
H A Dmps2-timer.c54 static int mps2_timer_shutdown(struct clock_event_device *ce) argument
56 clockevent_mps2_writel(0, ce, TIMER_RELOAD);
57 clockevent_mps2_writel(0, ce, TIMER_CTRL);
62 static int mps2_timer_set_next_event(unsigned long next, struct clock_event_device *ce) argument
64 clockevent_mps2_writel(next, ce, TIMER_VALUE);
65 clockevent_mps2_writel(TIMER_CTRL_IE | TIMER_CTRL_ENABLE, ce, TIMER_CTRL);
70 static int mps2_timer_set_periodic(struct clock_event_device *ce) argument
72 u32 clock_count_per_tick = to_mps2_clkevt(ce)->clock_count_per_tick;
74 clockevent_mps2_writel(clock_count_per_tick, ce, TIMER_RELOAD);
75 clockevent_mps2_writel(clock_count_per_tick, ce, TIMER_VALU
83 struct clockevent_mps2 *ce = dev_id; local
102 struct clockevent_mps2 *ce; local
[all...]
H A Dtimer-sun5i.c58 static void sun5i_clkevt_sync(struct sun5i_timer *ce) argument
60 u32 old = readl(ce->base + TIMER_CNTVAL_LO_REG(1));
62 while ((old - readl(ce->base + TIMER_CNTVAL_LO_REG(1))) < TIMER_SYNC_TICKS)
66 static void sun5i_clkevt_time_stop(struct sun5i_timer *ce, u8 timer) argument
68 u32 val = readl(ce->base + TIMER_CTL_REG(timer));
69 writel(val & ~TIMER_CTL_ENABLE, ce->base + TIMER_CTL_REG(timer));
71 sun5i_clkevt_sync(ce);
74 static void sun5i_clkevt_time_setup(struct sun5i_timer *ce, u8 timer, u32 delay) argument
76 writel(delay, ce->base + TIMER_INTVAL_LO_REG(timer));
79 static void sun5i_clkevt_time_start(struct sun5i_timer *ce, u argument
94 struct sun5i_timer *ce = clkevt_to_sun5i_timer(clkevt); local
102 struct sun5i_timer *ce = clkevt_to_sun5i_timer(clkevt); local
111 struct sun5i_timer *ce = clkevt_to_sun5i_timer(clkevt); local
122 struct sun5i_timer *ce = clkevt_to_sun5i_timer(clkevt); local
133 struct sun5i_timer *ce = dev_id; local
202 struct sun5i_timer *ce = platform_get_drvdata(pdev); local
[all...]
/linux-master/drivers/base/power/
H A Dclock_ops.c145 * @ce: PM clock entry corresponding to the clock.
147 static inline void __pm_clk_enable(struct device *dev, struct pm_clock_entry *ce) argument
151 switch (ce->status) {
153 ret = clk_prepare_enable(ce->clk);
156 ret = clk_enable(ce->clk);
162 ce->status = PCE_STATUS_ENABLED;
165 __func__, ce->clk, ret);
171 * @ce: PM clock entry corresponding to the clock.
173 static void pm_clk_acquire(struct device *dev, struct pm_clock_entry *ce) argument
175 if (!ce
199 struct pm_clock_entry *ce; local
354 __pm_clk_remove(struct pm_clock_entry *ce) argument
390 struct pm_clock_entry *ce; local
430 struct pm_clock_entry *ce; local
497 struct pm_clock_entry *ce, *c; local
546 struct pm_clock_entry *ce; local
584 struct pm_clock_entry *ce; local
[all...]
/linux-master/fs/smb/client/
H A Ddfs_cache.c108 static inline bool cache_entry_expired(const struct cache_entry *ce) argument
113 return timespec64_compare(&ts, &ce->etime) >= 0;
116 static inline void free_tgts(struct cache_entry *ce) argument
120 list_for_each_entry_safe(t, n, &ce->tlist, list) {
127 static inline void flush_cache_ent(struct cache_entry *ce) argument
129 hlist_del_init(&ce->hlist);
130 kfree(ce->path);
131 free_tgts(ce);
133 kmem_cache_free(cache_slab, ce);
143 struct cache_entry *ce; local
158 struct cache_entry *ce; local
226 dump_tgts(const struct cache_entry *ce) argument
237 dump_ce(const struct cache_entry *ce) argument
340 get_tgt_name(const struct cache_entry *ce) argument
382 copy_ref_data(const struct dfs_info3_param *refs, int numrefs, struct cache_entry *ce, const char *tgthint) argument
422 struct cache_entry *ce; local
447 struct cache_entry *ce; local
479 struct cache_entry *ce; local
537 struct cache_entry *ce; local
559 struct cache_entry *ce; local
620 update_cache_entry_locked(struct cache_entry *ce, const struct dfs_info3_param *refs, int numrefs) argument
689 struct cache_entry *ce; local
756 setup_referral(const char *path, struct cache_entry *ce, struct dfs_info3_param *ref, const char *target) argument
790 get_targets(struct cache_entry *ce, struct dfs_cache_tgt_list *tl) argument
862 struct cache_entry *ce; local
908 struct cache_entry *ce; local
949 struct cache_entry *ce; local
994 struct cache_entry *ce; local
1182 struct cache_entry *ce; local
[all...]
/linux-master/drivers/gpu/drm/i915/selftests/
H A Dmock_request.h33 mock_request(struct intel_context *ce, unsigned long delay);
/linux-master/drivers/crypto/allwinner/
H A DMakefile2 obj-$(CONFIG_CRYPTO_DEV_SUN8I_CE) += sun8i-ce/
/linux-master/arch/sparc/kernel/
H A Dtime_32.c120 struct clock_event_device *ce = &timer_ce; local
124 ce->name = "timer_ce";
125 ce->rating = 100;
126 ce->features = CLOCK_EVT_FEAT_PERIODIC;
127 ce->set_state_shutdown = timer_ce_shutdown;
128 ce->set_state_periodic = timer_ce_set_periodic;
129 ce->tick_resume = timer_ce_set_periodic;
130 ce->cpumask = cpu_possible_mask;
131 ce->shift = 32;
132 ce
213 struct clock_event_device *ce = &per_cpu(sparc32_clockevent, cpu); local
[all...]
/linux-master/drivers/of/
H A Ddynamic.c520 static void __of_changeset_entry_destroy(struct of_changeset_entry *ce) argument
522 if (ce->action == OF_RECONFIG_ATTACH_NODE &&
523 of_node_check_flag(ce->np, OF_OVERLAY)) {
524 if (kref_read(&ce->np->kobj.kref) > 1) {
526 kref_read(&ce->np->kobj.kref), ce->np);
528 of_node_set_flag(ce->np, OF_OVERLAY_FREE_CSET);
532 of_node_put(ce->np);
533 list_del(&ce->node);
534 kfree(ce);
537 __of_changeset_entry_invert(struct of_changeset_entry *ce, struct of_changeset_entry *rce) argument
567 __of_changeset_entry_notify(struct of_changeset_entry *ce, bool revert) argument
602 __of_changeset_entry_apply(struct of_changeset_entry *ce) argument
637 __of_changeset_entry_revert(struct of_changeset_entry *ce) argument
669 struct of_changeset_entry *ce, *cen; local
699 struct of_changeset_entry *ce; local
728 struct of_changeset_entry *ce; local
803 struct of_changeset_entry *ce; local
829 struct of_changeset_entry *ce; local
904 struct of_changeset_entry *ce; local
[all...]

Completed in 214 milliseconds

123456789