Lines Matching refs:ring

21 unsigned int intel_ring_update_space(struct intel_ring *ring);
23 void __intel_ring_pin(struct intel_ring *ring);
24 int intel_ring_pin(struct intel_ring *ring, struct i915_gem_ww_ctx *ww);
25 void intel_ring_unpin(struct intel_ring *ring);
26 void intel_ring_reset(struct intel_ring *ring, u32 tail);
30 static inline struct intel_ring *intel_ring_get(struct intel_ring *ring)
32 kref_get(&ring->ref);
33 return ring;
36 static inline void intel_ring_put(struct intel_ring *ring)
38 kref_put(&ring->ref, intel_ring_free);
51 GEM_BUG_ON((rq->ring->vaddr + rq->ring->emit) != cs);
52 GEM_BUG_ON(!IS_ALIGNED(rq->ring->emit, 8)); /* RING_TAIL qword align */
55 static inline u32 intel_ring_wrap(const struct intel_ring *ring, u32 pos)
57 return pos & (ring->size - 1);
60 static inline int intel_ring_direction(const struct intel_ring *ring,
63 typecheck(typeof(ring->size), next);
64 typecheck(typeof(ring->size), prev);
65 return (next - prev) << ring->wrap;
69 intel_ring_offset_valid(const struct intel_ring *ring,
72 if (pos & -ring->size) /* must be strictly within the ring */
83 /* Don't write ring->size (equivalent to 0) as that hangs some GPUs. */
84 u32 offset = addr - rq->ring->vaddr;
86 GEM_BUG_ON(offset > rq->ring->size);
87 return intel_ring_wrap(rq->ring, offset);
91 assert_ring_tail_valid(const struct intel_ring *ring, unsigned int tail)
93 unsigned int head = READ_ONCE(ring->head);
95 GEM_BUG_ON(!intel_ring_offset_valid(ring, tail));
106 * We use ring->head as the last known location of the actual RING_HEAD,
108 * as ring->head and so we should never program RING_TAIL to advance
109 * into the same cacheline as ring->head.
117 intel_ring_set_tail(struct intel_ring *ring, unsigned int tail)
125 assert_ring_tail_valid(ring, tail);
126 ring->tail = tail;