/linux-master/drivers/gpu/drm/xe/ |
H A D | xe_gt_pagefault.h | 16 int xe_guc_pagefault_handler(struct xe_guc *guc, u32 *msg, u32 len); 17 int xe_guc_access_counter_notify_handler(struct xe_guc *guc, u32 *msg, u32 len);
|
H A D | xe_uc.c | 41 ret = xe_guc_init(&uc->guc); 60 ret = xe_guc_submit_init(&uc->guc); 64 ret = xe_guc_db_mgr_init(&uc->guc.dbm, ~0); 88 err = xe_guc_init_post_hwconfig(&uc->guc); 104 ret = xe_guc_reset(&uc->guc); 116 xe_guc_sanitize(&uc->guc); 140 ret = xe_guc_min_load_for_hwconfig(&uc->guc); 163 ret = xe_guc_upload(&uc->guc); 167 ret = xe_guc_enable_communication(&uc->guc); 175 ret = xe_guc_post_load_init(&uc->guc); [all...] |
H A D | xe_gt_tlb_invalidation.c | 96 struct xe_guc *guc = >->uc.guc; local 104 mutex_lock(>->uc.guc.ct.lock); 119 wake_up_all(&guc->ct.wq); 125 mutex_unlock(>->uc.guc.ct.lock); 141 static int send_tlb_invalidation(struct xe_guc *guc, argument 145 struct xe_gt *gt = guc_to_gt(guc); 155 mutex_lock(&guc->ct.lock); 162 ret = xe_guc_ct_send_locked(&guc->ct, action, len, 195 mutex_unlock(&guc 364 struct xe_guc *guc = >->uc.guc; local 402 xe_guc_tlb_invalidation_done_handler(struct xe_guc *guc, u32 *msg, u32 len) argument [all...] |
H A D | xe_uc_debugfs.c | 26 xe_guc_debugfs_register(&uc->guc, root);
|
H A D | xe_gt_sriov_pf_policy.c | 21 static int guc_action_update_vgt_policy(struct xe_guc *guc, u64 addr, u32 size) argument 30 return xe_guc_ct_send_block(&guc->ct, request, ARRAY_SIZE(request)); 42 struct xe_guc *guc = >->uc.guc; local 56 ret = guc_action_update_vgt_policy(guc, xe_bo_ggtt_addr(bo), num_dwords); 144 >->sriov.pf.policy.guc.sched_if_idle, 153 return pf_provision_sched_if_idle(gt, gt->sriov.pf.policy.guc.sched_if_idle); 161 gt->sriov.pf.policy.guc.sched_if_idle = false; 199 enable = gt->sriov.pf.policy.guc.sched_if_idle; 211 >->sriov.pf.policy.guc [all...] |
H A D | xe_memirq.h | 24 int xe_memirq_init_guc(struct xe_memirq *memirq, struct xe_guc *guc);
|
H A D | xe_gt_tlb_invalidation.h | 24 int xe_guc_tlb_invalidation_done_handler(struct xe_guc *guc, u32 *msg, u32 len);
|
H A D | xe_memirq.c | 38 static const char *guc_name(struct xe_guc *guc) argument 40 return xe_gt_is_media_type(guc_to_gt(guc)) ? "media GuC" : "GuC"; 265 * @guc: the &xe_guc to setup 275 int xe_memirq_init_guc(struct xe_memirq *memirq, struct xe_guc *guc) argument 277 bool is_media = xe_gt_is_media_type(guc_to_gt(guc)); 289 err = xe_guc_self_cfg64(guc, GUC_KLV_SELF_CFG_MEMIRQ_SOURCE_ADDR_KEY, 294 err = xe_guc_self_cfg64(guc, GUC_KLV_SELF_CFG_MEMIRQ_STATUS_ADDR_KEY, 304 guc_name(guc), ERR_PTR(err)); 371 struct xe_guc *guc) 373 const char *name = guc_name(guc); 370 memirq_dispatch_guc(struct xe_memirq *memirq, struct iosys_map *status, struct xe_guc *guc) argument [all...] |
H A D | xe_wopcm.c | 141 u32 base = wopcm->guc.base; 142 u32 size = wopcm->guc.size; 201 u32 guc_fw_size = xe_uc_fw_get_upload_size(>->uc.guc.fw); 257 wopcm->guc.base = guc_wopcm_base; 258 wopcm->guc.size = guc_wopcm_size; 259 XE_WARN_ON(!wopcm->guc.base); 260 XE_WARN_ON(!wopcm->guc.size);
|
H A D | xe_gt_idle.c | 41 return >idle_to_gt(gtidle)->uc.guc.pc; 47 struct xe_guc *guc = container_of(pc, struct xe_guc, pc); local 48 struct xe_gt *gt = container_of(guc, struct xe_gt, uc.guc);
|
H A D | xe_gt_sriov_pf_control.c | 33 static int guc_action_vf_control_cmd(struct xe_guc *guc, u32 vfid, u32 cmd) argument 46 return xe_guc_ct_send_g2h_handler(&guc->ct, request, ARRAY_SIZE(request)); 48 ret = xe_guc_ct_send_block(&guc->ct, request, ARRAY_SIZE(request)); 58 err = guc_action_vf_control_cmd(>->uc.guc, vfid, cmd);
|
/linux-master/drivers/gpu/drm/i915/gt/uc/ |
H A D | intel_guc_submission.c | 108 * guc->submission_state.lock 122 * guc->submission_state.lock -> ce->guc_state.lock 161 #define NUMBER_MULTI_LRC_GUC_ID(guc) \ 162 ((guc)->submission_state.num_guc_ids / 16) 451 GEM_BUG_ON(!ce->parallel.guc.parent_page); 453 return ce->parallel.guc.parent_page * PAGE_SIZE; 471 * parallel.guc.parent_page is the offset into ce->state while 504 CIRC_SPACE(ce->parallel.guc.wqi_tail, ce->parallel.guc.wqi_head, WQ_SIZE) 506 ce->parallel.guc 516 __get_context(struct intel_guc *guc, u32 id) argument 525 __get_lrc_desc_v69(struct intel_guc *guc, u32 index) argument 537 guc_lrc_desc_pool_create_v69(struct intel_guc *guc) argument 552 guc_lrc_desc_pool_destroy_v69(struct intel_guc *guc) argument 561 guc_submission_initialized(struct intel_guc *guc) argument 566 _reset_lrc_desc_v69(struct intel_guc *guc, u32 id) argument 574 ctx_id_mapped(struct intel_guc *guc, u32 id) argument 579 set_ctx_id_mapping(struct intel_guc *guc, u32 id, struct intel_context *ce) argument 593 clr_ctx_id_mapping(struct intel_guc *guc, u32 id) argument 611 decr_outstanding_submission_g2h(struct intel_guc *guc) argument 617 guc_submission_send_busy_loop(struct intel_guc *guc, const u32 *action, u32 len, u32 g2h_len_dw, bool loop) argument 642 intel_guc_wait_for_pending_msg(struct intel_guc *guc, atomic_t *wait_var, bool interruptible, long timeout) argument 683 intel_guc_wait_for_idle(struct intel_guc *guc, long timeout) argument 696 __guc_add_request(struct intel_guc *guc, struct i915_request *rq) argument 779 guc_add_request(struct intel_guc *guc, struct i915_request *rq) argument 893 guc_wq_item_append(struct intel_guc *guc, struct i915_request *rq) argument 927 guc_dequeue_one_context(struct intel_guc *guc) argument 1077 scrub_guc_desc_for_outstanding_g2h(struct intel_guc *guc) argument 1182 __extend_last_switch(struct intel_guc *guc, u64 *prev_start, u32 new_start) argument 1249 struct intel_guc *guc = gt_to_guc(engine->gt); local 1284 guc_update_pm_timestamp(struct intel_guc *guc, ktime_t *now) argument 1314 struct intel_guc *guc = gt_to_guc(gt); local 1371 guc_enable_busyness_worker(struct intel_guc *guc) argument 1376 guc_cancel_busyness_worker(struct intel_guc *guc) argument 1422 __reset_guc_busyness_stats(struct intel_guc *guc) argument 1441 __update_guc_busyness_stats(struct intel_guc *guc) argument 1462 struct intel_guc *guc = ce_to_guc(ce); local 1481 struct intel_guc *guc = container_of(wrk, typeof(*guc), local 1544 guc_action_enable_usage_stats(struct intel_guc *guc) argument 1556 guc_init_engine_stats(struct intel_guc *guc) argument 1573 guc_fini_engine_stats(struct intel_guc *guc) argument 1580 struct intel_guc *guc = gt_to_guc(gt); local 1607 struct intel_guc *guc = gt_to_guc(gt); local 1621 submission_disabled(struct intel_guc *guc) argument 1630 disable_submission(struct intel_guc *guc) argument 1641 enable_submission(struct intel_guc *guc) argument 1659 guc_flush_submissions(struct intel_guc *guc) argument 1668 intel_guc_submission_flush_work(struct intel_guc *guc) argument 1675 intel_guc_submission_reset_prepare(struct intel_guc *guc) argument 1859 wake_up_all_tlb_invalidate(struct intel_guc *guc) argument 1873 intel_guc_submission_reset(struct intel_guc *guc, intel_engine_mask_t stalled) argument 1971 intel_guc_submission_cancel_requests(struct intel_guc *guc) argument 2006 intel_guc_submission_reset_finish(struct intel_guc *guc) argument 2038 intel_guc_tlb_invalidation_is_available(struct intel_guc *guc) argument 2044 init_tlb_lookup(struct intel_guc *guc) argument 2071 fini_tlb_lookup(struct intel_guc *guc) argument 2090 intel_guc_submission_init(struct intel_guc *guc) argument 2128 intel_guc_submission_fini(struct intel_guc *guc) argument 2153 guc_bypass_tasklet_submit(struct intel_guc *guc, struct i915_request *rq) argument 2179 need_tasklet(struct intel_guc *guc, struct i915_request *rq) argument 2192 struct intel_guc *guc = gt_to_guc(rq->engine->gt); local 2206 new_guc_id(struct intel_guc *guc, struct intel_context *ce) argument 2232 __release_guc_id(struct intel_guc *guc, struct intel_context *ce) argument 2254 release_guc_id(struct intel_guc *guc, struct intel_context *ce) argument 2263 steal_guc_id(struct intel_guc *guc, struct intel_context *ce) argument 2300 assign_guc_id(struct intel_guc *guc, struct intel_context *ce) argument 2329 pin_guc_id(struct intel_guc *guc, struct intel_context *ce) argument 2379 unpin_guc_id(struct intel_guc *guc, struct intel_context *ce) argument 2398 __guc_action_register_multi_lrc_v69(struct intel_guc *guc, struct intel_context *ce, u32 guc_id, u32 offset, bool loop) argument 2422 __guc_action_register_multi_lrc_v70(struct intel_guc *guc, struct intel_context *ce, struct guc_ctxt_registration_info *info, bool loop) argument 2465 __guc_action_register_context_v69(struct intel_guc *guc, u32 guc_id, u32 offset, bool loop) argument 2480 __guc_action_register_context_v70(struct intel_guc *guc, struct guc_ctxt_registration_info *info, bool loop) argument 2508 register_context_v69(struct intel_guc *guc, struct intel_context *ce, bool loop) argument 2524 register_context_v70(struct intel_guc *guc, struct intel_context *ce, bool loop) argument 2538 struct intel_guc *guc = ce_to_guc(ce); local 2563 __guc_action_deregister_context(struct intel_guc *guc, u32 guc_id) argument 2578 struct intel_guc *guc = ce_to_guc(ce); local 2646 __guc_context_set_context_policies(struct intel_guc *guc, struct context_policy *policy, bool loop) argument 2658 struct intel_guc *guc = gt_to_guc(engine->gt); local 2739 struct intel_guc *guc = gt_to_guc(engine->gt); local 2808 struct intel_guc *guc = gt_to_guc(engine->gt); local 2871 struct intel_guc *guc = gt_to_guc(engine->gt); local 2977 struct intel_guc *guc = ce_to_guc(ce); local 2992 __guc_context_sched_enable(struct intel_guc *guc, struct intel_context *ce) argument 3007 __guc_context_sched_disable(struct intel_guc *guc, struct intel_context *ce, u16 guc_id) argument 3064 struct intel_guc *guc = ce_to_guc(ce); local 3120 struct intel_guc *guc = ce_to_guc(ce); local 3174 __guc_context_set_preemption_timeout(struct intel_guc *guc, u16 guc_id, u32 preemption_timeout) argument 3199 struct intel_guc *guc = ce_to_guc(ce); local 3267 bypass_sched_disable(struct intel_guc *guc, struct intel_context *ce) argument 3286 struct intel_guc *guc = ce_to_guc(ce); local 3299 guc_id_pressure(struct intel_guc *guc, struct intel_context *ce) argument 3317 struct intel_guc *guc = ce_to_guc(ce); local 3352 struct intel_guc *guc = ce_to_guc(ce); local 3427 guc_flush_destroyed_contexts(struct intel_guc *guc) argument 3452 deregister_destroyed_contexts(struct intel_guc *guc) argument 3490 struct intel_guc *guc = container_of(w, struct intel_guc, local 3513 struct intel_guc *guc = ce_to_guc(ce); local 3552 __guc_context_set_prio(struct intel_guc *guc, struct intel_context *ce) argument 3572 guc_context_set_prio(struct intel_guc *guc, struct intel_context *ce, u8 prio) argument 3630 struct intel_guc *guc = &ce->engine->gt->uc.guc; local 3810 struct intel_guc *guc = ce_to_guc(ce); local 3945 struct intel_guc *guc = ce_to_guc(ce); local 4015 struct intel_guc *guc = ce_to_guc(ce); local 4041 struct intel_guc *guc = ce_to_guc(ce); local 4377 guc_kernel_context_pin(struct intel_guc *guc, struct intel_context *ce) argument 4406 guc_init_submission(struct intel_guc *guc) argument 4548 struct intel_guc *guc = sched_engine->private_data; local 4558 struct intel_guc *guc = gt_to_guc(engine->gt); local 4641 __guc_action_set_scheduling_policies(struct intel_guc *guc, struct scheduling_policy *policy) argument 4664 guc_init_global_schedule_policy(struct intel_guc *guc) argument 4692 guc_route_semaphores(struct intel_guc *guc, bool to_guc) argument 4708 intel_guc_submission_enable(struct intel_guc *guc) argument 4737 intel_guc_submission_disable(struct intel_guc *guc) argument 4745 __guc_submission_supported(struct intel_guc *guc) argument 4752 __guc_submission_selected(struct intel_guc *guc) argument 4762 intel_guc_sched_disable_gucid_threshold_max(struct intel_guc *guc) argument 4782 intel_guc_submission_init_early(struct intel_guc *guc) argument 4807 g2h_context_lookup(struct intel_guc *guc, u32 ctx_id) argument 4830 wait_wake_outstanding_tlb_g2h(struct intel_guc *guc, u32 seqno) argument 4847 intel_guc_tlb_invalidation_done(struct intel_guc *guc, const u32 *payload, u32 len) argument 4889 guc_send_invalidate_tlb(struct intel_guc *guc, enum intel_guc_tlb_invalidation_type type) argument 4967 intel_guc_invalidate_tlb_engines(struct intel_guc *guc) argument 4973 intel_guc_invalidate_tlb_guc(struct intel_guc *guc) argument 4978 intel_guc_deregister_done_process_msg(struct intel_guc *guc, const u32 *msg, u32 len) argument 5029 intel_guc_sched_done_process_msg(struct intel_guc *guc, const u32 *msg, u32 len) argument 5107 capture_error_state(struct intel_guc *guc, struct intel_context *ce) argument 5154 guc_handle_context_reset(struct intel_guc *guc, struct intel_context *ce) argument 5173 intel_guc_context_reset_process_msg(struct intel_guc *guc, const u32 *msg, u32 len) argument 5208 intel_guc_error_capture_process_msg(struct intel_guc *guc, const u32 *msg, u32 len) argument 5228 intel_guc_lookup_engine(struct intel_guc *guc, u8 guc_class, u8 instance) argument 5241 struct intel_guc *guc = container_of(w, struct intel_guc, local 5272 intel_guc_engine_failure_process_msg(struct intel_guc *guc, const u32 *msg, u32 len) argument 5317 struct intel_guc *guc = gt_to_guc(engine->gt); local 5379 struct intel_guc *guc = gt_to_guc(engine->gt); local 5418 intel_guc_submission_print_info(struct intel_guc *guc, struct drm_printer *p) argument 5485 intel_guc_submission_print_context_info(struct intel_guc *guc, struct drm_printer *p) argument 5823 struct intel_guc *guc; local [all...] |
H A D | intel_guc_capture.c | 292 guc_capture_alloc_steered_lists(struct intel_guc *guc, argument 295 struct intel_gt *gt = guc_to_gt(guc); 307 if (!list || guc->capture->extlists) 347 guc_dbg(guc, "capture found %d ext-regs.\n", num_tot_regs); 348 guc->capture->extlists = extlists; 352 guc_capture_get_device_reglist(struct intel_guc *guc) argument 354 struct drm_i915_private *i915 = guc_to_i915(guc); 368 guc_capture_alloc_steered_lists(guc, lists); 412 guc_capture_list_init(struct intel_guc *guc, u32 owner, u32 type, u32 classid, argument 416 const struct __guc_mmio_reg_descr_group *reglists = guc 473 guc_capture_getlistsize(struct intel_guc *guc, u32 owner, u32 type, u32 classid, size_t *size, bool is_purpose_est) argument 514 intel_guc_capture_getlistsize(struct intel_guc *guc, u32 owner, u32 type, u32 classid, size_t *size) argument 523 intel_guc_capture_getlist(struct intel_guc *guc, u32 owner, u32 type, u32 classid, void **outptr) argument 584 intel_guc_capture_getnullheader(struct intel_guc *guc, void **outptr, size_t *size) argument 611 guc_capture_output_min_size_est(struct intel_guc *guc) argument 659 check_guc_capture_size(struct intel_guc *guc) argument 778 guc_capture_log_remove_dw(struct intel_guc *guc, struct __guc_capture_bufstate *buf, u32 *dw) argument 817 guc_capture_log_get_group_hdr(struct intel_guc *guc, struct __guc_capture_bufstate *buf, struct guc_state_capture_group_header_t *ghdr) argument 838 guc_capture_log_get_data_hdr(struct intel_guc *guc, struct __guc_capture_bufstate *buf, struct guc_state_capture_header_t *hdr) argument 862 guc_capture_log_get_register(struct intel_guc *guc, struct __guc_capture_bufstate *buf, struct guc_mmio_reg *reg) argument 885 guc_capture_delete_one_node(struct intel_guc *guc, struct __guc_capture_parsed_output *node) argument 896 guc_capture_delete_prealloc_nodes(struct intel_guc *guc) argument 934 guc_capture_init_node(struct intel_guc *guc, struct __guc_capture_parsed_output *node) argument 952 guc_capture_get_prealloc_node(struct intel_guc *guc) argument 982 guc_capture_alloc_one_node(struct intel_guc *guc) argument 1007 guc_capture_clone_node(struct intel_guc *guc, struct __guc_capture_parsed_output *original, u32 keep_reglist_mask) argument 1047 __guc_capture_create_prealloc_nodes(struct intel_guc *guc) argument 1064 guc_get_max_reglist_count(struct intel_guc *guc) argument 1087 guc_capture_create_prealloc_nodes(struct intel_guc *guc) argument 1098 guc_capture_extract_reglists(struct intel_guc *guc, struct __guc_capture_bufstate *buf) argument 1282 __guc_capture_flushlog_complete(struct intel_guc *guc) argument 1293 __guc_capture_process_output(struct intel_guc *guc) argument 1359 guc_capture_reg_to_str(const struct intel_guc *guc, u32 owner, u32 type, u32 class, u32 id, u32 offset, u32 *is_ext) argument 1429 struct intel_guc *guc; local 1537 struct intel_guc *guc; local 1567 struct intel_guc *guc; local 1600 intel_guc_capture_process(struct intel_guc *guc) argument 1624 intel_guc_capture_destroy(struct intel_guc *guc) argument 1640 intel_guc_capture_init(struct intel_guc *guc) argument [all...] |
H A D | intel_guc_slpc.h | 17 static inline bool intel_guc_slpc_is_supported(struct intel_guc *guc) argument 19 return guc->slpc.supported; 22 static inline bool intel_guc_slpc_is_wanted(struct intel_guc *guc) argument 24 return guc->slpc.selected; 27 static inline bool intel_guc_slpc_is_used(struct intel_guc *guc) argument 29 return intel_guc_submission_is_used(guc) && intel_guc_slpc_is_wanted(guc);
|
H A D | intel_guc_fw.c | 71 struct intel_guc *guc = container_of(guc_fw, struct intel_guc, fw); local 74 intel_guc_ggtt_offset(guc, guc_fw->rsa_data)); 154 static int guc_wait_ucode(struct intel_guc *guc) argument 156 struct intel_gt *gt = guc_to_gt(guc); 195 guc_dbg(guc, "load still in progress, count = %d, freq = %dMHz, status = 0x%08X [0x%02X/%02X]\n", 207 guc_info(guc, "load failed: status = 0x%08X, time = %lldms, freq = %dMHz, ret = %d\n", 209 guc_info(guc, "load failed: status: Reset = %d, BootROM = 0x%02X, UKernel = 0x%02X, MIA = 0x%02X, Auth = 0x%02X\n", 217 guc_info(guc, "invalid key requested, header = 0x%08X\n", 223 guc_info(guc, "firmware signature verification failed\n"); 228 guc_info(guc, "firmwar 286 intel_guc_fw_upload(struct intel_guc *guc) argument [all...] |
H A D | selftest_guc.c | 147 struct intel_guc *guc = gt_to_guc(gt); local 154 int number_guc_id_stolen = guc->number_guc_id_stolen; 158 guc_err(guc, "Context array allocation failed\n"); 164 sv = guc->submission_state.num_guc_ids; 165 guc->submission_state.num_guc_ids = 512; 171 guc_err(guc, "Failed to create context: %pe\n", ce[context_index]); 177 guc_err(guc, "Failed to create spinner: %pe\n", ERR_PTR(ret)); 184 guc_err(guc, "Failed to create spinner request: %pe\n", spin_rq); 189 guc_err(guc, "Failed to add Spinner request: %pe\n", ERR_PTR(ret)); 198 guc_err(guc, "Faile 293 bad_h2g(struct intel_guc *guc) argument [all...] |
H A D | intel_guc_log.c | 42 struct intel_guc *guc = log_to_guc(log); local 85 guc_err(guc, "Mis-aligned log %s size: 0x%X vs 0x%X!\n", 90 guc_err(guc, "Zero log %s size!\n", sections[i].name); 98 guc_err(guc, "log %s size too large: %d vs %d!\n", 105 guc_err(guc, "Unit mismatch for crash and debug sections: %d vs %d!\n", 182 static int guc_action_flush_log_complete(struct intel_guc *guc) argument 189 return intel_guc_send_nb(guc, action, ARRAY_SIZE(action), 0); 192 static int guc_action_flush_log(struct intel_guc *guc) argument 199 return intel_guc_send(guc, action, ARRAY_SIZE(action)); 202 static int guc_action_control_log(struct intel_guc *guc, boo argument 376 struct intel_guc *guc = log_to_guc(log); local 522 struct intel_guc *guc = log_to_guc(log); local 575 struct intel_guc *guc = log_to_guc(log); local 591 struct intel_guc *guc = log_to_guc(log); local 616 struct intel_guc *guc = log_to_guc(log); local 666 struct intel_guc *guc = log_to_guc(log); local 772 struct intel_guc *guc = log_to_guc(log); local 798 struct intel_guc *guc = log_to_guc(log); local 884 struct intel_guc *guc = log_to_guc(log); local [all...] |
H A D | intel_guc_slpc.c | 33 static bool __detect_slpc_supported(struct intel_guc *guc) argument 36 return guc->submission_supported && 37 GRAPHICS_VER(guc_to_i915(guc)) >= 12; 40 static bool __guc_slpc_selected(struct intel_guc *guc) argument 42 if (!intel_guc_slpc_is_supported(guc)) 45 return guc->submission_selected; 50 struct intel_guc *guc = slpc_to_guc(slpc); local 52 slpc->supported = __detect_slpc_supported(guc); 53 slpc->selected = __guc_slpc_selected(guc); 102 static int guc_action_slpc_set_param_nb(struct intel_guc *guc, u argument 119 struct intel_guc *guc = slpc_to_guc(slpc); local 126 guc_action_slpc_set_param(struct intel_guc *guc, u8 id, u32 value) argument 146 guc_action_slpc_query(struct intel_guc *guc, u32 offset) argument 163 struct intel_guc *guc = slpc_to_guc(slpc); local 178 struct intel_guc *guc = slpc_to_guc(slpc); local 193 struct intel_guc *guc = slpc_to_guc(slpc); local 246 struct intel_guc *guc = slpc_to_guc(slpc); local 299 guc_action_slpc_reset(struct intel_guc *guc, u32 offset) argument 316 struct intel_guc *guc = slpc_to_guc(slpc); local 684 struct intel_guc *guc = slpc_to_guc(slpc); local [all...] |
H A D | intel_uc.h | 33 struct intel_guc guc; member in struct:intel_uc 89 uc_state_checkers(guc, guc); 91 uc_state_checkers(guc, guc_submission); 92 uc_state_checkers(guc, guc_slpc); 93 uc_state_checkers(guc, guc_rc); 101 return intel_guc_wait_for_idle(&uc->guc, timeout);
|
H A D | intel_guc_log_debugfs.c | 33 struct intel_guc *guc = log_to_guc(log); local 35 if (!intel_guc_is_supported(guc)) 61 struct intel_guc *guc = log_to_guc(log); local 62 struct intel_uc *uc = container_of(guc, struct intel_uc, guc); 64 if (!intel_guc_is_supported(guc)) 135 * Enable and start the guc log relay on value of 1.
|
H A D | intel_uc_debugfs.c | 23 drm_printf(&p, "[guc] supported:%s wanted:%s used:%s\n", 58 uc->guc.dbgfs_node = root; 63 intel_guc_debugfs_register(&uc->guc, root);
|
/linux-master/drivers/gpu/drm/i915/gt/ |
H A D | intel_wopcm.c | 237 u32 guc_fw_size = intel_uc_fw_get_upload_size(>->uc.guc.fw); 248 GEM_BUG_ON(wopcm->guc.base); 249 GEM_BUG_ON(wopcm->guc.size); 317 wopcm->guc.base = guc_wopcm_base; 318 wopcm->guc.size = guc_wopcm_size; 319 GEM_BUG_ON(!wopcm->guc.base); 320 GEM_BUG_ON(!wopcm->guc.size);
|
H A D | intel_tlb.c | 135 struct intel_guc *guc = gt_to_guc(gt); local 148 if (intel_guc_is_ready(guc)) 149 intel_guc_invalidate_tlb_engines(guc);
|
H A D | intel_gt.h | 102 static inline struct intel_gt *guc_to_gt(struct intel_guc *guc) argument 104 return container_of(guc, struct intel_gt, uc.guc); 122 static inline struct drm_i915_private *guc_to_i915(struct intel_guc *guc) argument 124 return guc_to_gt(guc)->i915; 129 return >->uc.guc;
|
/linux-master/drivers/gpu/drm/i915/selftests/ |
H A D | intel_scheduler_helpers.c | 70 err = intel_guc_global_policies_update(&engine->gt->uc.guc); 89 return intel_guc_global_policies_update(&engine->gt->uc.guc);
|