Lines Matching defs:seqno

53 i915_trace_irq_get(struct intel_ring_buffer *ring, uint32_t seqno)
59 ring->trace_irq_seqno = seqno;
352 DRM_ERROR("Failed to allocate seqno page\n");
460 u32 seqno,
467 intel_ring_emit(ring, seqno);
475 * @seqno - return seqno stuck into the ring
477 * Update the mailbox registers in the *other* rings with the current seqno.
482 u32 *seqno)
495 *seqno = i915_gem_next_request_seqno(ring);
497 update_mboxes(ring, *seqno, mbox1_reg);
498 update_mboxes(ring, *seqno, mbox2_reg);
501 intel_ring_emit(ring, *seqno);
509 * intel_ring_sync - sync the waiter to the signaller on seqno
513 * @seqno - seqno which the waiter will block on
519 u32 seqno)
531 intel_ring_emit(waiter, seqno);
540 struct intel_ring_buffer *signaller, u32 seqno);
542 struct intel_ring_buffer *signaller, u32 seqno);
544 struct intel_ring_buffer *signaller, u32 seqno);
550 u32 seqno)
557 seqno);
564 u32 seqno)
571 seqno);
578 u32 seqno)
585 seqno);
601 u32 seqno = i915_gem_next_request_seqno(ring);
622 intel_ring_emit(ring, seqno);
640 intel_ring_emit(ring, seqno);
644 *result = seqno;
652 u32 seqno = i915_gem_next_request_seqno(ring);
661 intel_ring_emit(ring, seqno);
665 *result = seqno;
674 /* Workaround to force correct ordering between irq and seqno writes on
822 uint32_t seqno;
829 seqno = i915_gem_next_request_seqno(ring);
833 intel_ring_emit(ring, seqno);
837 *result = seqno;
1158 static int intel_ring_wait_seqno(struct intel_ring_buffer *ring, u32 seqno)
1171 ret = i915_wait_request(ring, seqno, true);
1181 u32 seqno = 0;
1204 seqno = request->seqno;
1216 if (seqno == 0)
1219 ret = intel_ring_wait_seqno(ring, seqno);