Searched refs:sched_engine (Results 1 - 20 of 20) sorted by relevance

/linux-master/drivers/gpu/drm/i915/
H A Di915_scheduler.h43 i915_sched_lookup_priolist(struct i915_sched_engine *sched_engine, int prio);
56 i915_sched_engine_get(struct i915_sched_engine *sched_engine) argument
58 kref_get(&sched_engine->ref);
59 return sched_engine;
63 i915_sched_engine_put(struct i915_sched_engine *sched_engine) argument
65 kref_put(&sched_engine->ref, sched_engine->destroy);
69 i915_sched_engine_is_empty(struct i915_sched_engine *sched_engine) argument
71 return RB_EMPTY_ROOT(&sched_engine->queue.rb_root);
75 i915_sched_engine_reset_on_empty(struct i915_sched_engine *sched_engine) argument
82 i915_sched_engine_active_lock_bh(struct i915_sched_engine *sched_engine) argument
89 i915_sched_engine_active_unlock_bh(struct i915_sched_engine *sched_engine) argument
101 i915_sched_engine_disabled(struct i915_sched_engine *sched_engine) argument
[all...]
H A Di915_scheduler.c39 static void assert_priolists(struct i915_sched_engine * const sched_engine) argument
47 GEM_BUG_ON(rb_first_cached(&sched_engine->queue) !=
48 rb_first(&sched_engine->queue.rb_root));
51 for (rb = rb_first_cached(&sched_engine->queue); rb; rb = rb_next(rb)) {
60 i915_sched_lookup_priolist(struct i915_sched_engine *sched_engine, int prio) argument
66 lockdep_assert_held(&sched_engine->lock);
67 assert_priolists(sched_engine);
69 if (unlikely(sched_engine->no_priolist))
75 parent = &sched_engine->queue.rb_root.rb_node;
90 p = &sched_engine
134 struct i915_sched_engine *sched_engine; local
159 struct i915_sched_engine *sched_engine; local
439 struct i915_sched_engine *sched_engine = local
446 default_disabled(struct i915_sched_engine *sched_engine) argument
454 struct i915_sched_engine *sched_engine; local
[all...]
H A Di915_scheduler_types.h174 bool (*disabled)(struct i915_sched_engine *sched_engine);
H A Di915_request.c258 spin_lock_irq(&locked->sched_engine->lock);
260 spin_unlock(&locked->sched_engine->lock);
262 spin_lock(&locked->sched_engine->lock);
271 spin_unlock_irq(&locked->sched_engine->lock);
612 lockdep_assert_held(&engine->sched_engine->lock);
624 * priority queue, under the sched_engine->lock. This ensures that the
703 spin_lock_irqsave(&engine->sched_engine->lock, flags);
707 spin_unlock_irqrestore(&engine->sched_engine->lock, flags);
721 lockdep_assert_held(&engine->sched_engine->lock);
754 spin_lock_irqsave(&engine->sched_engine
[all...]
H A Di915_request.h689 lockdep_is_held(&rq->engine->sched_engine->lock));
/linux-master/drivers/gpu/drm/i915/gt/
H A Dintel_execlists_submission.c279 static int queue_prio(const struct i915_sched_engine *sched_engine) argument
283 rb = rb_first_cached(&sched_engine->queue);
324 if (engine->sched_engine->queue_priority_hint <= last_prio)
331 if (!list_is_last(&rq->sched.link, &engine->sched_engine->requests) &&
346 queue_prio(engine->sched_engine)) > last_prio;
373 lockdep_assert_held(&engine->sched_engine->lock);
376 &engine->sched_engine->requests,
388 pl = i915_sched_lookup_priolist(engine->sched_engine,
391 GEM_BUG_ON(i915_sched_engine_is_empty(engine->sched_engine));
551 spin_lock_irq(&engine->sched_engine
1276 struct i915_sched_engine * const sched_engine = engine->sched_engine; local
2432 struct i915_sched_engine *sched_engine = local
2563 struct i915_sched_engine *sched_engine = engine->sched_engine; local
3139 struct i915_sched_engine *sched_engine = local
3150 struct i915_sched_engine * const sched_engine = engine->sched_engine; local
3328 struct i915_sched_engine *sched_engine = engine->sched_engine; local
3815 struct i915_sched_engine *sched_engine = local
4089 struct i915_sched_engine *sched_engine = engine->sched_engine; local
[all...]
H A Dmock_engine.c258 lockdep_assert_held(&rq->engine->sched_engine->lock);
259 list_move_tail(&rq->sched.link, &rq->engine->sched_engine->requests);
274 spin_lock_irq(&locked->sched_engine->lock);
276 spin_unlock(&locked->sched_engine->lock);
277 spin_lock(&engine->sched_engine->lock);
281 spin_unlock_irq(&locked->sched_engine->lock);
302 spin_lock_irqsave(&engine->sched_engine->lock, flags);
305 list_for_each_entry(rq, &engine->sched_engine->requests, sched.link)
318 spin_unlock_irqrestore(&engine->sched_engine->lock, flags);
332 i915_sched_engine_put(engine->sched_engine);
[all...]
H A Dintel_ring_submission.c359 spin_lock_irqsave(&engine->sched_engine->lock, flags);
361 list_for_each_entry(pos, &engine->sched_engine->requests, sched.link) {
416 spin_unlock_irqrestore(&engine->sched_engine->lock, flags);
428 spin_lock_irqsave(&engine->sched_engine->lock, flags);
431 list_for_each_entry(request, &engine->sched_engine->requests, sched.link)
437 spin_unlock_irqrestore(&engine->sched_engine->lock, flags);
613 lockdep_assert_held(&engine->sched_engine->lock);
614 list_for_each_entry_continue(rq, &engine->sched_engine->requests,
1102 lockdep_assert_held(&rq->engine->sched_engine->lock);
1103 list_move_tail(&rq->sched.link, &rq->engine->sched_engine
[all...]
H A Dintel_engine_heartbeat.c160 if (i915_sched_engine_disabled(engine->sched_engine)) {
184 } else if (engine->sched_engine->schedule &&
199 engine->sched_engine->schedule(rq, &attr);
H A Dintel_engine_cs.c1293 engine->sched_engine = i915_sched_engine_create(ENGINE_PHYSICAL);
1294 if (!engine->sched_engine) {
1298 engine->sched_engine->private_data = engine;
1322 i915_sched_engine_put(engine->sched_engine);
1363 spin_lock_irq(&engine->sched_engine->lock);
1367 spin_unlock_irq(&engine->sched_engine->lock);
1570 GEM_BUG_ON(!list_empty(&engine->sched_engine->requests));
1572 i915_sched_engine_put(engine->sched_engine);
1896 struct tasklet_struct *t = &engine->sched_engine->tasklet;
1936 if (!i915_sched_engine_is_empty(engine->sched_engine))
[all...]
H A Dselftest_execlists.c47 tasklet_hi_schedule(&engine->sched_engine->tasklet);
275 engine->sched_engine->schedule(rq[1], &attr);
555 tasklet_disable(&engine->sched_engine->tasklet);
561 tasklet_enable(&engine->sched_engine->tasklet);
576 tasklet_enable(&engine->sched_engine->tasklet);
630 engine->sched_engine->tasklet.callback(&engine->sched_engine->tasklet);
919 engine->sched_engine->schedule(rq, &attr);
1202 tasklet_hi_schedule(&engine->sched_engine->tasklet);
1344 engine->sched_engine
[all...]
H A Dintel_breadcrumbs.c254 if (rq->engine->sched_engine->retire_inflight_request_prio)
255 rq->engine->sched_engine->retire_inflight_request_prio(rq);
H A Dintel_engine_user.c114 if (engine->sched_engine->schedule)
H A Dintel_engine_types.h417 struct i915_sched_engine *sched_engine; member in struct:intel_engine_cs
H A Dselftest_reset.c330 struct tasklet_struct *t = &engine->sched_engine->tasklet;
H A Dselftest_hangcheck.c929 if (engine->sched_engine->schedule && arg->flags & TEST_PRIORITY) {
934 engine->sched_engine->schedule(rq[idx], &attr);
1880 struct tasklet_struct * const t = &engine->sched_engine->tasklet;
H A Dselftest_lrc.c57 tasklet_hi_schedule(&engine->sched_engine->tasklet);
1749 tasklet_disable(&engine->sched_engine->tasklet);
1754 tasklet_enable(&engine->sched_engine->tasklet);
/linux-master/drivers/gpu/drm/i915/gt/uc/
H A Dintel_guc_submission.c102 * sched_engine->lock
104 * engine (sched_engine), thus only one of the contexts which share a
105 * sched_engine can be submitting at a time. Currently only one sched_engine is
121 * sched_engine->lock -> ce->guc_state.lock
705 lockdep_assert_held(&rq->engine->sched_engine->lock);
929 struct i915_sched_engine * const sched_engine = guc->sched_engine; local
935 lockdep_assert_held(&sched_engine->lock);
953 while ((rb = rb_first_cached(&sched_engine
1049 struct i915_sched_engine *sched_engine = local
1623 struct i915_sched_engine * const sched_engine = guc->sched_engine; local
1632 struct i915_sched_engine * const sched_engine = guc->sched_engine; local
1643 struct i915_sched_engine * const sched_engine = guc->sched_engine; local
1661 struct i915_sched_engine * const sched_engine = guc->sched_engine; local
1773 struct i915_sched_engine * const sched_engine = local
1907 struct i915_sched_engine *sched_engine = ce_to_guc(ce)->sched_engine; local
1921 guc_cancel_sched_engine_requests(struct i915_sched_engine *sched_engine) argument
2142 queue_request(struct i915_sched_engine *sched_engine, struct i915_request *rq, int prio) argument
2181 struct i915_sched_engine *sched_engine = rq->engine->sched_engine; local
2191 struct i915_sched_engine *sched_engine = rq->engine->sched_engine; local
4362 guc_sched_engine_disabled(struct i915_sched_engine *sched_engine) argument
4535 struct i915_sched_engine *sched_engine = local
5137 struct i915_sched_engine *sched_engine = ce->engine->sched_engine; local
5410 struct i915_sched_engine *sched_engine = guc->sched_engine; local
[all...]
H A Dintel_guc.h48 /** @sched_engine: Global engine used to submit requests to GuC */
49 struct i915_sched_engine *sched_engine; member in struct:intel_guc
/linux-master/drivers/gpu/drm/i915/gem/
H A Di915_gem_wait.c105 if (engine->sched_engine->schedule)
106 engine->sched_engine->schedule(rq, attr);

Completed in 187 milliseconds