Lines Matching refs:rq

74 		struct i915_request *rq = NULL;
87 if (rq) {
88 i915_request_await_dma_fence(this, &rq->fence);
89 i915_request_put(rq);
91 rq = i915_request_get(this);
94 if (i915_request_wait(rq, 0, 10 * HZ) < 0) {
97 i915_request_put(rq);
101 i915_request_put(rq);
116 rq = NULL;
126 if (rq) { /* Force submission order */
127 i915_request_await_dma_fence(this, &rq->fence);
128 i915_request_put(rq);
145 rq = i915_request_get(this);
148 GEM_BUG_ON(!rq);
149 if (i915_request_wait(rq, 0, HZ / 5) < 0) {
153 i915_request_put(rq);
156 i915_request_put(rq);
200 struct i915_request *rq = NULL;
204 struct i915_request *prev = rq;
206 rq = i915_request_create(arg->ce[n]);
207 if (IS_ERR(rq)) {
209 arg->result = PTR_ERR(rq);
213 i915_request_get(rq);
216 i915_request_await_dma_fence(rq,
221 i915_request_add(rq);
224 if (IS_ERR_OR_NULL(rq))
227 if (i915_request_wait(rq, 0, HZ) < 0)
230 i915_request_put(rq);
243 struct i915_request *rq = NULL;
252 struct i915_request *prev = rq;
254 rq = i915_request_create(arg->ce[n]);
255 if (IS_ERR(rq)) {
257 arg->result = PTR_ERR(rq);
261 i915_request_get(rq);
264 i915_request_await_dma_fence(rq,
269 i915_request_add(rq);
275 if (!IS_ERR_OR_NULL(rq))
276 i915_request_put(rq);
945 struct i915_request *rq;
991 rq = i915_request_create(ce);
992 if (IS_ERR(rq)) {
993 err = PTR_ERR(rq);
997 err = i915_vma_move_to_active(batch, rq, 0);
1001 err = i915_vma_move_to_active(vma, rq, EXEC_OBJECT_WRITE);
1005 if (rq->engine->emit_init_breadcrumb) {
1006 err = rq->engine->emit_init_breadcrumb(rq);
1011 err = rq->engine->emit_bb_start(rq,
1018 *rq_out = i915_request_get(rq);
1022 i915_request_set_error_once(rq, err);
1023 i915_request_add(rq);
1049 struct i915_request *rq;
1064 rq = igt_spinner_create_request(*spin, ce, MI_NOOP);
1065 if (IS_ERR(rq)) {
1066 ret = PTR_ERR(rq);
1070 i915_request_add(rq);
1072 if (!igt_wait_for_spinner(*spin, rq)) {
1095 struct i915_request *rq = NULL;
1101 ret = emit_rpcs_query(obj, ce, &rq);
1108 ret = i915_request_wait(rq, 0, MAX_SCHEDULE_TIMEOUT);
1109 i915_request_put(rq);
1515 struct i915_request *rq;
1556 rq = igt_request_alloc(ctx, engine);
1557 if (IS_ERR(rq)) {
1558 err = PTR_ERR(rq);
1562 err = igt_vma_move_to_active_unlocked(vma, rq, 0);
1566 if (rq->engine->emit_init_breadcrumb) {
1567 err = rq->engine->emit_init_breadcrumb(rq);
1572 err = engine->emit_bb_start(rq, i915_vma_offset(vma),
1579 i915_request_add(rq);
1583 i915_request_set_error_once(rq, err);
1584 i915_request_add(rq);
1604 struct i915_request *rq;
1689 rq = igt_request_alloc(ctx, engine);
1690 if (IS_ERR(rq)) {
1691 err = PTR_ERR(rq);
1695 err = igt_vma_move_to_active_unlocked(vma, rq, EXEC_OBJECT_WRITE);
1699 if (rq->engine->emit_init_breadcrumb) {
1700 err = rq->engine->emit_init_breadcrumb(rq);
1705 err = engine->emit_bb_start(rq, i915_vma_offset(vma),
1712 i915_request_add(rq);
1731 i915_request_set_error_once(rq, err);
1732 i915_request_add(rq);