/linux-master/drivers/gpu/drm/i915/ |
H A D | i915_scheduler.h | 43 i915_sched_lookup_priolist(struct i915_sched_engine *sched_engine, int prio); 56 i915_sched_engine_get(struct i915_sched_engine *sched_engine) argument 58 kref_get(&sched_engine->ref); 59 return sched_engine; 63 i915_sched_engine_put(struct i915_sched_engine *sched_engine) argument 65 kref_put(&sched_engine->ref, sched_engine->destroy); 69 i915_sched_engine_is_empty(struct i915_sched_engine *sched_engine) argument 71 return RB_EMPTY_ROOT(&sched_engine->queue.rb_root); 75 i915_sched_engine_reset_on_empty(struct i915_sched_engine *sched_engine) argument 82 i915_sched_engine_active_lock_bh(struct i915_sched_engine *sched_engine) argument 89 i915_sched_engine_active_unlock_bh(struct i915_sched_engine *sched_engine) argument 101 i915_sched_engine_disabled(struct i915_sched_engine *sched_engine) argument [all...] |
H A D | i915_scheduler.c | 39 static void assert_priolists(struct i915_sched_engine * const sched_engine) argument 47 GEM_BUG_ON(rb_first_cached(&sched_engine->queue) != 48 rb_first(&sched_engine->queue.rb_root)); 51 for (rb = rb_first_cached(&sched_engine->queue); rb; rb = rb_next(rb)) { 60 i915_sched_lookup_priolist(struct i915_sched_engine *sched_engine, int prio) argument 66 lockdep_assert_held(&sched_engine->lock); 67 assert_priolists(sched_engine); 69 if (unlikely(sched_engine->no_priolist)) 75 parent = &sched_engine->queue.rb_root.rb_node; 90 p = &sched_engine 134 struct i915_sched_engine *sched_engine; local 159 struct i915_sched_engine *sched_engine; local 439 struct i915_sched_engine *sched_engine = local 446 default_disabled(struct i915_sched_engine *sched_engine) argument 454 struct i915_sched_engine *sched_engine; local [all...] |
H A D | i915_scheduler_types.h | 174 bool (*disabled)(struct i915_sched_engine *sched_engine);
|
H A D | i915_request.c | 258 spin_lock_irq(&locked->sched_engine->lock); 260 spin_unlock(&locked->sched_engine->lock); 262 spin_lock(&locked->sched_engine->lock); 271 spin_unlock_irq(&locked->sched_engine->lock); 612 lockdep_assert_held(&engine->sched_engine->lock); 624 * priority queue, under the sched_engine->lock. This ensures that the 703 spin_lock_irqsave(&engine->sched_engine->lock, flags); 707 spin_unlock_irqrestore(&engine->sched_engine->lock, flags); 721 lockdep_assert_held(&engine->sched_engine->lock); 754 spin_lock_irqsave(&engine->sched_engine [all...] |
H A D | i915_request.h | 689 lockdep_is_held(&rq->engine->sched_engine->lock));
|
/linux-master/drivers/gpu/drm/i915/gt/ |
H A D | intel_execlists_submission.c | 279 static int queue_prio(const struct i915_sched_engine *sched_engine) argument 283 rb = rb_first_cached(&sched_engine->queue); 324 if (engine->sched_engine->queue_priority_hint <= last_prio) 331 if (!list_is_last(&rq->sched.link, &engine->sched_engine->requests) && 346 queue_prio(engine->sched_engine)) > last_prio; 373 lockdep_assert_held(&engine->sched_engine->lock); 376 &engine->sched_engine->requests, 388 pl = i915_sched_lookup_priolist(engine->sched_engine, 391 GEM_BUG_ON(i915_sched_engine_is_empty(engine->sched_engine)); 551 spin_lock_irq(&engine->sched_engine 1276 struct i915_sched_engine * const sched_engine = engine->sched_engine; local 2432 struct i915_sched_engine *sched_engine = local 2563 struct i915_sched_engine *sched_engine = engine->sched_engine; local 3139 struct i915_sched_engine *sched_engine = local 3150 struct i915_sched_engine * const sched_engine = engine->sched_engine; local 3328 struct i915_sched_engine *sched_engine = engine->sched_engine; local 3815 struct i915_sched_engine *sched_engine = local 4089 struct i915_sched_engine *sched_engine = engine->sched_engine; local [all...] |
H A D | mock_engine.c | 258 lockdep_assert_held(&rq->engine->sched_engine->lock); 259 list_move_tail(&rq->sched.link, &rq->engine->sched_engine->requests); 274 spin_lock_irq(&locked->sched_engine->lock); 276 spin_unlock(&locked->sched_engine->lock); 277 spin_lock(&engine->sched_engine->lock); 281 spin_unlock_irq(&locked->sched_engine->lock); 302 spin_lock_irqsave(&engine->sched_engine->lock, flags); 305 list_for_each_entry(rq, &engine->sched_engine->requests, sched.link) 318 spin_unlock_irqrestore(&engine->sched_engine->lock, flags); 332 i915_sched_engine_put(engine->sched_engine); [all...] |
H A D | intel_ring_submission.c | 359 spin_lock_irqsave(&engine->sched_engine->lock, flags); 361 list_for_each_entry(pos, &engine->sched_engine->requests, sched.link) { 416 spin_unlock_irqrestore(&engine->sched_engine->lock, flags); 428 spin_lock_irqsave(&engine->sched_engine->lock, flags); 431 list_for_each_entry(request, &engine->sched_engine->requests, sched.link) 437 spin_unlock_irqrestore(&engine->sched_engine->lock, flags); 613 lockdep_assert_held(&engine->sched_engine->lock); 614 list_for_each_entry_continue(rq, &engine->sched_engine->requests, 1102 lockdep_assert_held(&rq->engine->sched_engine->lock); 1103 list_move_tail(&rq->sched.link, &rq->engine->sched_engine [all...] |
H A D | intel_engine_heartbeat.c | 160 if (i915_sched_engine_disabled(engine->sched_engine)) { 184 } else if (engine->sched_engine->schedule && 199 engine->sched_engine->schedule(rq, &attr);
|
H A D | intel_engine_cs.c | 1293 engine->sched_engine = i915_sched_engine_create(ENGINE_PHYSICAL); 1294 if (!engine->sched_engine) { 1298 engine->sched_engine->private_data = engine; 1322 i915_sched_engine_put(engine->sched_engine); 1363 spin_lock_irq(&engine->sched_engine->lock); 1367 spin_unlock_irq(&engine->sched_engine->lock); 1570 GEM_BUG_ON(!list_empty(&engine->sched_engine->requests)); 1572 i915_sched_engine_put(engine->sched_engine); 1896 struct tasklet_struct *t = &engine->sched_engine->tasklet; 1936 if (!i915_sched_engine_is_empty(engine->sched_engine)) [all...] |
H A D | selftest_execlists.c | 47 tasklet_hi_schedule(&engine->sched_engine->tasklet); 275 engine->sched_engine->schedule(rq[1], &attr); 555 tasklet_disable(&engine->sched_engine->tasklet); 561 tasklet_enable(&engine->sched_engine->tasklet); 576 tasklet_enable(&engine->sched_engine->tasklet); 630 engine->sched_engine->tasklet.callback(&engine->sched_engine->tasklet); 919 engine->sched_engine->schedule(rq, &attr); 1202 tasklet_hi_schedule(&engine->sched_engine->tasklet); 1344 engine->sched_engine [all...] |
H A D | intel_breadcrumbs.c | 254 if (rq->engine->sched_engine->retire_inflight_request_prio) 255 rq->engine->sched_engine->retire_inflight_request_prio(rq);
|
H A D | intel_engine_user.c | 114 if (engine->sched_engine->schedule)
|
H A D | intel_engine_types.h | 417 struct i915_sched_engine *sched_engine; member in struct:intel_engine_cs
|
H A D | selftest_reset.c | 330 struct tasklet_struct *t = &engine->sched_engine->tasklet;
|
H A D | selftest_hangcheck.c | 929 if (engine->sched_engine->schedule && arg->flags & TEST_PRIORITY) { 934 engine->sched_engine->schedule(rq[idx], &attr); 1880 struct tasklet_struct * const t = &engine->sched_engine->tasklet;
|
H A D | selftest_lrc.c | 57 tasklet_hi_schedule(&engine->sched_engine->tasklet); 1749 tasklet_disable(&engine->sched_engine->tasklet); 1754 tasklet_enable(&engine->sched_engine->tasklet);
|
/linux-master/drivers/gpu/drm/i915/gt/uc/ |
H A D | intel_guc_submission.c | 102 * sched_engine->lock 104 * engine (sched_engine), thus only one of the contexts which share a 105 * sched_engine can be submitting at a time. Currently only one sched_engine is 121 * sched_engine->lock -> ce->guc_state.lock 705 lockdep_assert_held(&rq->engine->sched_engine->lock); 929 struct i915_sched_engine * const sched_engine = guc->sched_engine; local 935 lockdep_assert_held(&sched_engine->lock); 953 while ((rb = rb_first_cached(&sched_engine 1049 struct i915_sched_engine *sched_engine = local 1623 struct i915_sched_engine * const sched_engine = guc->sched_engine; local 1632 struct i915_sched_engine * const sched_engine = guc->sched_engine; local 1643 struct i915_sched_engine * const sched_engine = guc->sched_engine; local 1661 struct i915_sched_engine * const sched_engine = guc->sched_engine; local 1773 struct i915_sched_engine * const sched_engine = local 1907 struct i915_sched_engine *sched_engine = ce_to_guc(ce)->sched_engine; local 1921 guc_cancel_sched_engine_requests(struct i915_sched_engine *sched_engine) argument 2142 queue_request(struct i915_sched_engine *sched_engine, struct i915_request *rq, int prio) argument 2181 struct i915_sched_engine *sched_engine = rq->engine->sched_engine; local 2191 struct i915_sched_engine *sched_engine = rq->engine->sched_engine; local 4362 guc_sched_engine_disabled(struct i915_sched_engine *sched_engine) argument 4535 struct i915_sched_engine *sched_engine = local 5137 struct i915_sched_engine *sched_engine = ce->engine->sched_engine; local 5410 struct i915_sched_engine *sched_engine = guc->sched_engine; local [all...] |
H A D | intel_guc.h | 48 /** @sched_engine: Global engine used to submit requests to GuC */ 49 struct i915_sched_engine *sched_engine; member in struct:intel_guc
|
/linux-master/drivers/gpu/drm/i915/gem/ |
H A D | i915_gem_wait.c | 105 if (engine->sched_engine->schedule) 106 engine->sched_engine->schedule(rq, attr);
|