Lines Matching defs:xe

261 	struct xe_device *xe = guc_to_xe(guc);
265 err = drmm_mutex_init(&xe->drm, &guc->submission_state.lock);
286 return drmm_add_action_or_reset(&xe->drm, guc_submit_fini, guc);
399 struct xe_device *xe = guc_to_xe(guc);
404 xe_assert(xe, exec_queue_registered(q));
438 struct xe_device *xe = guc_to_xe(guc);
443 xe_assert(xe, xe_exec_queue_is_parallel(q));
466 xe_assert(xe, len <= MAX_MLRC_REG_SIZE);
496 struct xe_device *xe = guc_to_xe(guc);
500 xe_assert(xe, !exec_queue_registered(q));
526 xe_map_memset(xe, &map, 0, 0, PARALLEL_SCRATCH_SIZE - WQ_SIZE);
527 parallel_write(xe, map, wq_desc.wq_status, WQ_STATUS_ACTIVE);
555 struct xe_device *xe = guc_to_xe(guc);
563 q->guc->wqi_head = parallel_read(xe, map, wq_desc.head);
583 struct xe_device *xe = guc_to_xe(guc);
590 xe_assert(xe, FIELD_FIT(WQ_LEN_MASK, len_dw));
592 parallel_write(xe, map, wq[q->guc->wqi_tail / sizeof(u32)],
603 struct xe_device *xe = guc_to_xe(guc);
630 xe_assert(xe, i == wqi_size / sizeof(u32));
634 xe_map_memcpy_to(xe, &map, 0, wqi, wqi_size);
636 xe_assert(xe, q->guc->wqi_tail <= WQ_SIZE);
638 xe_device_wmb(xe);
641 parallel_write(xe, map, wq_desc.tail, q->guc->wqi_tail);
648 struct xe_device *xe = guc_to_xe(guc);
656 xe_assert(xe, exec_queue_registered(q));
703 struct xe_device *xe = guc_to_xe(guc);
706 xe_assert(xe, !(exec_queue_destroyed(q) || exec_queue_pending_disable(q)) ||
753 struct xe_device *xe = guc_to_xe(guc);
763 drm_warn(&xe->drm, "Pending enable failed to respond");
790 struct xe_device *xe = guc_to_xe(guc);
791 struct drm_printer p = drm_err_printer(&xe->drm, NULL);
836 struct xe_device *xe = guc_to_xe(guc);
839 wake_up_all(&xe->ufence_wq);
853 struct xe_device *xe = guc_to_xe(guc);
856 xe_assert(xe, xe_exec_queue_is_lr(q));
888 drm_warn(&xe->drm, "Schedule disable failed to respond");
905 struct xe_device *xe = guc_to_xe(exec_queue_to_guc(q));
919 drm_notice(&xe->drm, "Timedout job: seqno=%u, guc_id=%d, flags=0x%lx",
974 drm_warn(&xe->drm, "Schedule disable failed to respond");
1053 struct xe_device *xe = guc_to_xe(guc);
1055 xe_assert(xe, !(q->flags & EXEC_QUEUE_FLAG_PERMANENT));
1082 struct xe_device *xe = guc_to_xe(guc);
1084 xe_assert(xe, exec_queue_suspended(q) || exec_queue_killed(q) ||
1086 xe_assert(xe, q->guc->suspend_pending);
1190 struct xe_device *xe = guc_to_xe(guc);
1195 xe_assert(xe, xe_device_uc_enabled(guc_to_xe(guc)));
1363 struct xe_device *xe = guc_to_xe(guc);
1365 xe_assert(xe, !q->guc->suspend_pending);
1465 struct xe_device *xe = guc_to_xe(guc);
1467 xe_assert(xe, guc_read_stopped(guc) == 1);
1504 struct xe_device *xe = guc_to_xe(guc);
1506 xe_assert(xe, guc_read_stopped(guc) == 1);
1522 struct xe_device *xe = guc_to_xe(guc);
1526 drm_err(&xe->drm, "Invalid guc_id %u", guc_id);
1532 drm_err(&xe->drm, "Not engine present for guc_id %u", guc_id);
1536 xe_assert(xe, guc_id >= q->guc->id);
1537 xe_assert(xe, guc_id < (q->guc->id + q->width));
1579 struct xe_device *xe = guc_to_xe(guc);
1584 drm_err(&xe->drm, "Invalid length %u", len);
1594 drm_err(&xe->drm, "Unexpected engine state 0x%04x",
1618 struct xe_device *xe = guc_to_xe(guc);
1623 drm_err(&xe->drm, "Invalid length %u", len);
1633 drm_err(&xe->drm, "Unexpected engine state 0x%04x",
1645 struct xe_device *xe = guc_to_xe(guc);
1650 drm_err(&xe->drm, "Invalid length %u", len);
1658 drm_info(&xe->drm, "Engine reset: guc_id=%d", guc_id);
1680 struct xe_device *xe = guc_to_xe(guc);
1685 drm_err(&xe->drm, "Invalid length %u", len);
1693 drm_dbg(&xe->drm, "Engine memory cat error: guc_id=%d", guc_id);
1706 struct xe_device *xe = guc_to_xe(guc);
1711 drm_err(&xe->drm, "Invalid length %u", len);
1720 drm_err(&xe->drm, "GuC engine reset request failed on %d:%d because 0x%08X",
1733 struct xe_device *xe = guc_to_xe(guc);
1739 snapshot->parallel.wq_desc.head = parallel_read(xe, map, wq_desc.head);
1740 snapshot->parallel.wq_desc.tail = parallel_read(xe, map, wq_desc.tail);
1741 snapshot->parallel.wq_desc.status = parallel_read(xe, map,
1750 parallel_read(xe, map, wq[i / sizeof(u32)]);