Searched refs:guc (Results 26 - 50 of 83) sorted by relevance

1234

/linux-master/drivers/gpu/drm/xe/
H A Dxe_gt_pagefault.h16 int xe_guc_pagefault_handler(struct xe_guc *guc, u32 *msg, u32 len);
17 int xe_guc_access_counter_notify_handler(struct xe_guc *guc, u32 *msg, u32 len);
H A Dxe_uc.c41 ret = xe_guc_init(&uc->guc);
60 ret = xe_guc_submit_init(&uc->guc);
64 ret = xe_guc_db_mgr_init(&uc->guc.dbm, ~0);
88 err = xe_guc_init_post_hwconfig(&uc->guc);
104 ret = xe_guc_reset(&uc->guc);
116 xe_guc_sanitize(&uc->guc);
140 ret = xe_guc_min_load_for_hwconfig(&uc->guc);
163 ret = xe_guc_upload(&uc->guc);
167 ret = xe_guc_enable_communication(&uc->guc);
175 ret = xe_guc_post_load_init(&uc->guc);
[all...]
H A Dxe_gt_tlb_invalidation.c96 struct xe_guc *guc = &gt->uc.guc; local
104 mutex_lock(&gt->uc.guc.ct.lock);
119 wake_up_all(&guc->ct.wq);
125 mutex_unlock(&gt->uc.guc.ct.lock);
141 static int send_tlb_invalidation(struct xe_guc *guc, argument
145 struct xe_gt *gt = guc_to_gt(guc);
155 mutex_lock(&guc->ct.lock);
162 ret = xe_guc_ct_send_locked(&guc->ct, action, len,
195 mutex_unlock(&guc
364 struct xe_guc *guc = &gt->uc.guc; local
402 xe_guc_tlb_invalidation_done_handler(struct xe_guc *guc, u32 *msg, u32 len) argument
[all...]
H A Dxe_uc_debugfs.c26 xe_guc_debugfs_register(&uc->guc, root);
H A Dxe_gt_sriov_pf_policy.c21 static int guc_action_update_vgt_policy(struct xe_guc *guc, u64 addr, u32 size) argument
30 return xe_guc_ct_send_block(&guc->ct, request, ARRAY_SIZE(request));
42 struct xe_guc *guc = &gt->uc.guc; local
56 ret = guc_action_update_vgt_policy(guc, xe_bo_ggtt_addr(bo), num_dwords);
144 &gt->sriov.pf.policy.guc.sched_if_idle,
153 return pf_provision_sched_if_idle(gt, gt->sriov.pf.policy.guc.sched_if_idle);
161 gt->sriov.pf.policy.guc.sched_if_idle = false;
199 enable = gt->sriov.pf.policy.guc.sched_if_idle;
211 &gt->sriov.pf.policy.guc
[all...]
H A Dxe_memirq.h24 int xe_memirq_init_guc(struct xe_memirq *memirq, struct xe_guc *guc);
H A Dxe_gt_tlb_invalidation.h24 int xe_guc_tlb_invalidation_done_handler(struct xe_guc *guc, u32 *msg, u32 len);
H A Dxe_memirq.c38 static const char *guc_name(struct xe_guc *guc) argument
40 return xe_gt_is_media_type(guc_to_gt(guc)) ? "media GuC" : "GuC";
265 * @guc: the &xe_guc to setup
275 int xe_memirq_init_guc(struct xe_memirq *memirq, struct xe_guc *guc) argument
277 bool is_media = xe_gt_is_media_type(guc_to_gt(guc));
289 err = xe_guc_self_cfg64(guc, GUC_KLV_SELF_CFG_MEMIRQ_SOURCE_ADDR_KEY,
294 err = xe_guc_self_cfg64(guc, GUC_KLV_SELF_CFG_MEMIRQ_STATUS_ADDR_KEY,
304 guc_name(guc), ERR_PTR(err));
371 struct xe_guc *guc)
373 const char *name = guc_name(guc);
370 memirq_dispatch_guc(struct xe_memirq *memirq, struct iosys_map *status, struct xe_guc *guc) argument
[all...]
H A Dxe_wopcm.c141 u32 base = wopcm->guc.base;
142 u32 size = wopcm->guc.size;
201 u32 guc_fw_size = xe_uc_fw_get_upload_size(&gt->uc.guc.fw);
257 wopcm->guc.base = guc_wopcm_base;
258 wopcm->guc.size = guc_wopcm_size;
259 XE_WARN_ON(!wopcm->guc.base);
260 XE_WARN_ON(!wopcm->guc.size);
H A Dxe_gt_idle.c41 return &gtidle_to_gt(gtidle)->uc.guc.pc;
47 struct xe_guc *guc = container_of(pc, struct xe_guc, pc); local
48 struct xe_gt *gt = container_of(guc, struct xe_gt, uc.guc);
H A Dxe_gt_sriov_pf_control.c33 static int guc_action_vf_control_cmd(struct xe_guc *guc, u32 vfid, u32 cmd) argument
46 return xe_guc_ct_send_g2h_handler(&guc->ct, request, ARRAY_SIZE(request));
48 ret = xe_guc_ct_send_block(&guc->ct, request, ARRAY_SIZE(request));
58 err = guc_action_vf_control_cmd(&gt->uc.guc, vfid, cmd);
/linux-master/drivers/gpu/drm/i915/gt/uc/
H A Dintel_guc_submission.c108 * guc->submission_state.lock
122 * guc->submission_state.lock -> ce->guc_state.lock
161 #define NUMBER_MULTI_LRC_GUC_ID(guc) \
162 ((guc)->submission_state.num_guc_ids / 16)
451 GEM_BUG_ON(!ce->parallel.guc.parent_page);
453 return ce->parallel.guc.parent_page * PAGE_SIZE;
471 * parallel.guc.parent_page is the offset into ce->state while
504 CIRC_SPACE(ce->parallel.guc.wqi_tail, ce->parallel.guc.wqi_head, WQ_SIZE)
506 ce->parallel.guc
516 __get_context(struct intel_guc *guc, u32 id) argument
525 __get_lrc_desc_v69(struct intel_guc *guc, u32 index) argument
537 guc_lrc_desc_pool_create_v69(struct intel_guc *guc) argument
552 guc_lrc_desc_pool_destroy_v69(struct intel_guc *guc) argument
561 guc_submission_initialized(struct intel_guc *guc) argument
566 _reset_lrc_desc_v69(struct intel_guc *guc, u32 id) argument
574 ctx_id_mapped(struct intel_guc *guc, u32 id) argument
579 set_ctx_id_mapping(struct intel_guc *guc, u32 id, struct intel_context *ce) argument
593 clr_ctx_id_mapping(struct intel_guc *guc, u32 id) argument
611 decr_outstanding_submission_g2h(struct intel_guc *guc) argument
617 guc_submission_send_busy_loop(struct intel_guc *guc, const u32 *action, u32 len, u32 g2h_len_dw, bool loop) argument
642 intel_guc_wait_for_pending_msg(struct intel_guc *guc, atomic_t *wait_var, bool interruptible, long timeout) argument
683 intel_guc_wait_for_idle(struct intel_guc *guc, long timeout) argument
696 __guc_add_request(struct intel_guc *guc, struct i915_request *rq) argument
779 guc_add_request(struct intel_guc *guc, struct i915_request *rq) argument
893 guc_wq_item_append(struct intel_guc *guc, struct i915_request *rq) argument
927 guc_dequeue_one_context(struct intel_guc *guc) argument
1077 scrub_guc_desc_for_outstanding_g2h(struct intel_guc *guc) argument
1182 __extend_last_switch(struct intel_guc *guc, u64 *prev_start, u32 new_start) argument
1249 struct intel_guc *guc = gt_to_guc(engine->gt); local
1284 guc_update_pm_timestamp(struct intel_guc *guc, ktime_t *now) argument
1314 struct intel_guc *guc = gt_to_guc(gt); local
1371 guc_enable_busyness_worker(struct intel_guc *guc) argument
1376 guc_cancel_busyness_worker(struct intel_guc *guc) argument
1422 __reset_guc_busyness_stats(struct intel_guc *guc) argument
1441 __update_guc_busyness_stats(struct intel_guc *guc) argument
1462 struct intel_guc *guc = ce_to_guc(ce); local
1481 struct intel_guc *guc = container_of(wrk, typeof(*guc), local
1544 guc_action_enable_usage_stats(struct intel_guc *guc) argument
1556 guc_init_engine_stats(struct intel_guc *guc) argument
1573 guc_fini_engine_stats(struct intel_guc *guc) argument
1580 struct intel_guc *guc = gt_to_guc(gt); local
1607 struct intel_guc *guc = gt_to_guc(gt); local
1621 submission_disabled(struct intel_guc *guc) argument
1630 disable_submission(struct intel_guc *guc) argument
1641 enable_submission(struct intel_guc *guc) argument
1659 guc_flush_submissions(struct intel_guc *guc) argument
1668 intel_guc_submission_flush_work(struct intel_guc *guc) argument
1675 intel_guc_submission_reset_prepare(struct intel_guc *guc) argument
1859 wake_up_all_tlb_invalidate(struct intel_guc *guc) argument
1873 intel_guc_submission_reset(struct intel_guc *guc, intel_engine_mask_t stalled) argument
1971 intel_guc_submission_cancel_requests(struct intel_guc *guc) argument
2006 intel_guc_submission_reset_finish(struct intel_guc *guc) argument
2038 intel_guc_tlb_invalidation_is_available(struct intel_guc *guc) argument
2044 init_tlb_lookup(struct intel_guc *guc) argument
2071 fini_tlb_lookup(struct intel_guc *guc) argument
2090 intel_guc_submission_init(struct intel_guc *guc) argument
2128 intel_guc_submission_fini(struct intel_guc *guc) argument
2153 guc_bypass_tasklet_submit(struct intel_guc *guc, struct i915_request *rq) argument
2179 need_tasklet(struct intel_guc *guc, struct i915_request *rq) argument
2192 struct intel_guc *guc = gt_to_guc(rq->engine->gt); local
2206 new_guc_id(struct intel_guc *guc, struct intel_context *ce) argument
2232 __release_guc_id(struct intel_guc *guc, struct intel_context *ce) argument
2254 release_guc_id(struct intel_guc *guc, struct intel_context *ce) argument
2263 steal_guc_id(struct intel_guc *guc, struct intel_context *ce) argument
2300 assign_guc_id(struct intel_guc *guc, struct intel_context *ce) argument
2329 pin_guc_id(struct intel_guc *guc, struct intel_context *ce) argument
2379 unpin_guc_id(struct intel_guc *guc, struct intel_context *ce) argument
2398 __guc_action_register_multi_lrc_v69(struct intel_guc *guc, struct intel_context *ce, u32 guc_id, u32 offset, bool loop) argument
2422 __guc_action_register_multi_lrc_v70(struct intel_guc *guc, struct intel_context *ce, struct guc_ctxt_registration_info *info, bool loop) argument
2465 __guc_action_register_context_v69(struct intel_guc *guc, u32 guc_id, u32 offset, bool loop) argument
2480 __guc_action_register_context_v70(struct intel_guc *guc, struct guc_ctxt_registration_info *info, bool loop) argument
2508 register_context_v69(struct intel_guc *guc, struct intel_context *ce, bool loop) argument
2524 register_context_v70(struct intel_guc *guc, struct intel_context *ce, bool loop) argument
2538 struct intel_guc *guc = ce_to_guc(ce); local
2563 __guc_action_deregister_context(struct intel_guc *guc, u32 guc_id) argument
2578 struct intel_guc *guc = ce_to_guc(ce); local
2646 __guc_context_set_context_policies(struct intel_guc *guc, struct context_policy *policy, bool loop) argument
2658 struct intel_guc *guc = gt_to_guc(engine->gt); local
2739 struct intel_guc *guc = gt_to_guc(engine->gt); local
2808 struct intel_guc *guc = gt_to_guc(engine->gt); local
2871 struct intel_guc *guc = gt_to_guc(engine->gt); local
2977 struct intel_guc *guc = ce_to_guc(ce); local
2992 __guc_context_sched_enable(struct intel_guc *guc, struct intel_context *ce) argument
3007 __guc_context_sched_disable(struct intel_guc *guc, struct intel_context *ce, u16 guc_id) argument
3064 struct intel_guc *guc = ce_to_guc(ce); local
3120 struct intel_guc *guc = ce_to_guc(ce); local
3174 __guc_context_set_preemption_timeout(struct intel_guc *guc, u16 guc_id, u32 preemption_timeout) argument
3199 struct intel_guc *guc = ce_to_guc(ce); local
3267 bypass_sched_disable(struct intel_guc *guc, struct intel_context *ce) argument
3286 struct intel_guc *guc = ce_to_guc(ce); local
3299 guc_id_pressure(struct intel_guc *guc, struct intel_context *ce) argument
3317 struct intel_guc *guc = ce_to_guc(ce); local
3352 struct intel_guc *guc = ce_to_guc(ce); local
3427 guc_flush_destroyed_contexts(struct intel_guc *guc) argument
3452 deregister_destroyed_contexts(struct intel_guc *guc) argument
3490 struct intel_guc *guc = container_of(w, struct intel_guc, local
3513 struct intel_guc *guc = ce_to_guc(ce); local
3552 __guc_context_set_prio(struct intel_guc *guc, struct intel_context *ce) argument
3572 guc_context_set_prio(struct intel_guc *guc, struct intel_context *ce, u8 prio) argument
3630 struct intel_guc *guc = &ce->engine->gt->uc.guc; local
3810 struct intel_guc *guc = ce_to_guc(ce); local
3945 struct intel_guc *guc = ce_to_guc(ce); local
4015 struct intel_guc *guc = ce_to_guc(ce); local
4041 struct intel_guc *guc = ce_to_guc(ce); local
4377 guc_kernel_context_pin(struct intel_guc *guc, struct intel_context *ce) argument
4406 guc_init_submission(struct intel_guc *guc) argument
4548 struct intel_guc *guc = sched_engine->private_data; local
4558 struct intel_guc *guc = gt_to_guc(engine->gt); local
4641 __guc_action_set_scheduling_policies(struct intel_guc *guc, struct scheduling_policy *policy) argument
4664 guc_init_global_schedule_policy(struct intel_guc *guc) argument
4692 guc_route_semaphores(struct intel_guc *guc, bool to_guc) argument
4708 intel_guc_submission_enable(struct intel_guc *guc) argument
4737 intel_guc_submission_disable(struct intel_guc *guc) argument
4745 __guc_submission_supported(struct intel_guc *guc) argument
4752 __guc_submission_selected(struct intel_guc *guc) argument
4762 intel_guc_sched_disable_gucid_threshold_max(struct intel_guc *guc) argument
4782 intel_guc_submission_init_early(struct intel_guc *guc) argument
4807 g2h_context_lookup(struct intel_guc *guc, u32 ctx_id) argument
4830 wait_wake_outstanding_tlb_g2h(struct intel_guc *guc, u32 seqno) argument
4847 intel_guc_tlb_invalidation_done(struct intel_guc *guc, const u32 *payload, u32 len) argument
4889 guc_send_invalidate_tlb(struct intel_guc *guc, enum intel_guc_tlb_invalidation_type type) argument
4967 intel_guc_invalidate_tlb_engines(struct intel_guc *guc) argument
4973 intel_guc_invalidate_tlb_guc(struct intel_guc *guc) argument
4978 intel_guc_deregister_done_process_msg(struct intel_guc *guc, const u32 *msg, u32 len) argument
5029 intel_guc_sched_done_process_msg(struct intel_guc *guc, const u32 *msg, u32 len) argument
5107 capture_error_state(struct intel_guc *guc, struct intel_context *ce) argument
5154 guc_handle_context_reset(struct intel_guc *guc, struct intel_context *ce) argument
5173 intel_guc_context_reset_process_msg(struct intel_guc *guc, const u32 *msg, u32 len) argument
5208 intel_guc_error_capture_process_msg(struct intel_guc *guc, const u32 *msg, u32 len) argument
5228 intel_guc_lookup_engine(struct intel_guc *guc, u8 guc_class, u8 instance) argument
5241 struct intel_guc *guc = container_of(w, struct intel_guc, local
5272 intel_guc_engine_failure_process_msg(struct intel_guc *guc, const u32 *msg, u32 len) argument
5317 struct intel_guc *guc = gt_to_guc(engine->gt); local
5379 struct intel_guc *guc = gt_to_guc(engine->gt); local
5418 intel_guc_submission_print_info(struct intel_guc *guc, struct drm_printer *p) argument
5485 intel_guc_submission_print_context_info(struct intel_guc *guc, struct drm_printer *p) argument
5823 struct intel_guc *guc; local
[all...]
H A Dintel_guc_capture.c292 guc_capture_alloc_steered_lists(struct intel_guc *guc, argument
295 struct intel_gt *gt = guc_to_gt(guc);
307 if (!list || guc->capture->extlists)
347 guc_dbg(guc, "capture found %d ext-regs.\n", num_tot_regs);
348 guc->capture->extlists = extlists;
352 guc_capture_get_device_reglist(struct intel_guc *guc) argument
354 struct drm_i915_private *i915 = guc_to_i915(guc);
368 guc_capture_alloc_steered_lists(guc, lists);
412 guc_capture_list_init(struct intel_guc *guc, u32 owner, u32 type, u32 classid, argument
416 const struct __guc_mmio_reg_descr_group *reglists = guc
473 guc_capture_getlistsize(struct intel_guc *guc, u32 owner, u32 type, u32 classid, size_t *size, bool is_purpose_est) argument
514 intel_guc_capture_getlistsize(struct intel_guc *guc, u32 owner, u32 type, u32 classid, size_t *size) argument
523 intel_guc_capture_getlist(struct intel_guc *guc, u32 owner, u32 type, u32 classid, void **outptr) argument
584 intel_guc_capture_getnullheader(struct intel_guc *guc, void **outptr, size_t *size) argument
611 guc_capture_output_min_size_est(struct intel_guc *guc) argument
659 check_guc_capture_size(struct intel_guc *guc) argument
778 guc_capture_log_remove_dw(struct intel_guc *guc, struct __guc_capture_bufstate *buf, u32 *dw) argument
817 guc_capture_log_get_group_hdr(struct intel_guc *guc, struct __guc_capture_bufstate *buf, struct guc_state_capture_group_header_t *ghdr) argument
838 guc_capture_log_get_data_hdr(struct intel_guc *guc, struct __guc_capture_bufstate *buf, struct guc_state_capture_header_t *hdr) argument
862 guc_capture_log_get_register(struct intel_guc *guc, struct __guc_capture_bufstate *buf, struct guc_mmio_reg *reg) argument
885 guc_capture_delete_one_node(struct intel_guc *guc, struct __guc_capture_parsed_output *node) argument
896 guc_capture_delete_prealloc_nodes(struct intel_guc *guc) argument
934 guc_capture_init_node(struct intel_guc *guc, struct __guc_capture_parsed_output *node) argument
952 guc_capture_get_prealloc_node(struct intel_guc *guc) argument
982 guc_capture_alloc_one_node(struct intel_guc *guc) argument
1007 guc_capture_clone_node(struct intel_guc *guc, struct __guc_capture_parsed_output *original, u32 keep_reglist_mask) argument
1047 __guc_capture_create_prealloc_nodes(struct intel_guc *guc) argument
1064 guc_get_max_reglist_count(struct intel_guc *guc) argument
1087 guc_capture_create_prealloc_nodes(struct intel_guc *guc) argument
1098 guc_capture_extract_reglists(struct intel_guc *guc, struct __guc_capture_bufstate *buf) argument
1282 __guc_capture_flushlog_complete(struct intel_guc *guc) argument
1293 __guc_capture_process_output(struct intel_guc *guc) argument
1359 guc_capture_reg_to_str(const struct intel_guc *guc, u32 owner, u32 type, u32 class, u32 id, u32 offset, u32 *is_ext) argument
1429 struct intel_guc *guc; local
1537 struct intel_guc *guc; local
1567 struct intel_guc *guc; local
1600 intel_guc_capture_process(struct intel_guc *guc) argument
1624 intel_guc_capture_destroy(struct intel_guc *guc) argument
1640 intel_guc_capture_init(struct intel_guc *guc) argument
[all...]
H A Dintel_guc_slpc.h17 static inline bool intel_guc_slpc_is_supported(struct intel_guc *guc) argument
19 return guc->slpc.supported;
22 static inline bool intel_guc_slpc_is_wanted(struct intel_guc *guc) argument
24 return guc->slpc.selected;
27 static inline bool intel_guc_slpc_is_used(struct intel_guc *guc) argument
29 return intel_guc_submission_is_used(guc) && intel_guc_slpc_is_wanted(guc);
H A Dintel_guc_fw.c71 struct intel_guc *guc = container_of(guc_fw, struct intel_guc, fw); local
74 intel_guc_ggtt_offset(guc, guc_fw->rsa_data));
154 static int guc_wait_ucode(struct intel_guc *guc) argument
156 struct intel_gt *gt = guc_to_gt(guc);
195 guc_dbg(guc, "load still in progress, count = %d, freq = %dMHz, status = 0x%08X [0x%02X/%02X]\n",
207 guc_info(guc, "load failed: status = 0x%08X, time = %lldms, freq = %dMHz, ret = %d\n",
209 guc_info(guc, "load failed: status: Reset = %d, BootROM = 0x%02X, UKernel = 0x%02X, MIA = 0x%02X, Auth = 0x%02X\n",
217 guc_info(guc, "invalid key requested, header = 0x%08X\n",
223 guc_info(guc, "firmware signature verification failed\n");
228 guc_info(guc, "firmwar
286 intel_guc_fw_upload(struct intel_guc *guc) argument
[all...]
H A Dselftest_guc.c147 struct intel_guc *guc = gt_to_guc(gt); local
154 int number_guc_id_stolen = guc->number_guc_id_stolen;
158 guc_err(guc, "Context array allocation failed\n");
164 sv = guc->submission_state.num_guc_ids;
165 guc->submission_state.num_guc_ids = 512;
171 guc_err(guc, "Failed to create context: %pe\n", ce[context_index]);
177 guc_err(guc, "Failed to create spinner: %pe\n", ERR_PTR(ret));
184 guc_err(guc, "Failed to create spinner request: %pe\n", spin_rq);
189 guc_err(guc, "Failed to add Spinner request: %pe\n", ERR_PTR(ret));
198 guc_err(guc, "Faile
293 bad_h2g(struct intel_guc *guc) argument
[all...]
H A Dintel_guc_log.c42 struct intel_guc *guc = log_to_guc(log); local
85 guc_err(guc, "Mis-aligned log %s size: 0x%X vs 0x%X!\n",
90 guc_err(guc, "Zero log %s size!\n", sections[i].name);
98 guc_err(guc, "log %s size too large: %d vs %d!\n",
105 guc_err(guc, "Unit mismatch for crash and debug sections: %d vs %d!\n",
182 static int guc_action_flush_log_complete(struct intel_guc *guc) argument
189 return intel_guc_send_nb(guc, action, ARRAY_SIZE(action), 0);
192 static int guc_action_flush_log(struct intel_guc *guc) argument
199 return intel_guc_send(guc, action, ARRAY_SIZE(action));
202 static int guc_action_control_log(struct intel_guc *guc, boo argument
376 struct intel_guc *guc = log_to_guc(log); local
522 struct intel_guc *guc = log_to_guc(log); local
575 struct intel_guc *guc = log_to_guc(log); local
591 struct intel_guc *guc = log_to_guc(log); local
616 struct intel_guc *guc = log_to_guc(log); local
666 struct intel_guc *guc = log_to_guc(log); local
772 struct intel_guc *guc = log_to_guc(log); local
798 struct intel_guc *guc = log_to_guc(log); local
884 struct intel_guc *guc = log_to_guc(log); local
[all...]
H A Dintel_guc_slpc.c33 static bool __detect_slpc_supported(struct intel_guc *guc) argument
36 return guc->submission_supported &&
37 GRAPHICS_VER(guc_to_i915(guc)) >= 12;
40 static bool __guc_slpc_selected(struct intel_guc *guc) argument
42 if (!intel_guc_slpc_is_supported(guc))
45 return guc->submission_selected;
50 struct intel_guc *guc = slpc_to_guc(slpc); local
52 slpc->supported = __detect_slpc_supported(guc);
53 slpc->selected = __guc_slpc_selected(guc);
102 static int guc_action_slpc_set_param_nb(struct intel_guc *guc, u argument
119 struct intel_guc *guc = slpc_to_guc(slpc); local
126 guc_action_slpc_set_param(struct intel_guc *guc, u8 id, u32 value) argument
146 guc_action_slpc_query(struct intel_guc *guc, u32 offset) argument
163 struct intel_guc *guc = slpc_to_guc(slpc); local
178 struct intel_guc *guc = slpc_to_guc(slpc); local
193 struct intel_guc *guc = slpc_to_guc(slpc); local
246 struct intel_guc *guc = slpc_to_guc(slpc); local
299 guc_action_slpc_reset(struct intel_guc *guc, u32 offset) argument
316 struct intel_guc *guc = slpc_to_guc(slpc); local
684 struct intel_guc *guc = slpc_to_guc(slpc); local
[all...]
H A Dintel_uc.h33 struct intel_guc guc; member in struct:intel_uc
89 uc_state_checkers(guc, guc);
91 uc_state_checkers(guc, guc_submission);
92 uc_state_checkers(guc, guc_slpc);
93 uc_state_checkers(guc, guc_rc);
101 return intel_guc_wait_for_idle(&uc->guc, timeout);
H A Dintel_guc_log_debugfs.c33 struct intel_guc *guc = log_to_guc(log); local
35 if (!intel_guc_is_supported(guc))
61 struct intel_guc *guc = log_to_guc(log); local
62 struct intel_uc *uc = container_of(guc, struct intel_uc, guc);
64 if (!intel_guc_is_supported(guc))
135 * Enable and start the guc log relay on value of 1.
H A Dintel_uc_debugfs.c23 drm_printf(&p, "[guc] supported:%s wanted:%s used:%s\n",
58 uc->guc.dbgfs_node = root;
63 intel_guc_debugfs_register(&uc->guc, root);
/linux-master/drivers/gpu/drm/i915/gt/
H A Dintel_wopcm.c237 u32 guc_fw_size = intel_uc_fw_get_upload_size(&gt->uc.guc.fw);
248 GEM_BUG_ON(wopcm->guc.base);
249 GEM_BUG_ON(wopcm->guc.size);
317 wopcm->guc.base = guc_wopcm_base;
318 wopcm->guc.size = guc_wopcm_size;
319 GEM_BUG_ON(!wopcm->guc.base);
320 GEM_BUG_ON(!wopcm->guc.size);
H A Dintel_tlb.c135 struct intel_guc *guc = gt_to_guc(gt); local
148 if (intel_guc_is_ready(guc))
149 intel_guc_invalidate_tlb_engines(guc);
H A Dintel_gt.h102 static inline struct intel_gt *guc_to_gt(struct intel_guc *guc) argument
104 return container_of(guc, struct intel_gt, uc.guc);
122 static inline struct drm_i915_private *guc_to_i915(struct intel_guc *guc) argument
124 return guc_to_gt(guc)->i915;
129 return &gt->uc.guc;
/linux-master/drivers/gpu/drm/i915/selftests/
H A Dintel_scheduler_helpers.c70 err = intel_guc_global_policies_update(&engine->gt->uc.guc);
89 return intel_guc_global_policies_update(&engine->gt->uc.guc);

Completed in 224 milliseconds

1234