intel_ringbuffer.c revision 290238
1/*
2 * Copyright �� 2008-2010 Intel Corporation
3 *
4 * Permission is hereby granted, free of charge, to any person obtaining a
5 * copy of this software and associated documentation files (the "Software"),
6 * to deal in the Software without restriction, including without limitation
7 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
8 * and/or sell copies of the Software, and to permit persons to whom the
9 * Software is furnished to do so, subject to the following conditions:
10 *
11 * The above copyright notice and this permission notice (including the next
12 * paragraph) shall be included in all copies or substantial portions of the
13 * Software.
14 *
15 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
16 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
17 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
18 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
19 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
20 * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
21 * IN THE SOFTWARE.
22 *
23 * Authors:
24 *    Eric Anholt <eric@anholt.net>
25 *    Zou Nan hai <nanhai.zou@intel.com>
26 *    Xiang Hai hao<haihao.xiang@intel.com>
27 *
28 */
29
30#include <sys/cdefs.h>
31__FBSDID("$FreeBSD: head/sys/dev/drm2/i915/intel_ringbuffer.c 290238 2015-11-01 19:55:32Z dumbbell $");
32
33#include <dev/drm2/drmP.h>
34#include <dev/drm2/drm.h>
35#include <dev/drm2/i915/i915_drm.h>
36#include <dev/drm2/i915/i915_drv.h>
37#include <dev/drm2/i915/intel_drv.h>
38#include <dev/drm2/i915/intel_ringbuffer.h>
39#include <sys/sched.h>
40#include <sys/sf_buf.h>
41
42/*
43 * 965+ support PIPE_CONTROL commands, which provide finer grained control
44 * over cache flushing.
45 */
46struct pipe_control {
47	struct drm_i915_gem_object *obj;
48	volatile u32 *cpu_page;
49	u32 gtt_offset;
50};
51
52void
53i915_trace_irq_get(struct intel_ring_buffer *ring, uint32_t seqno)
54{
55	struct drm_i915_private *dev_priv;
56
57	if (ring->trace_irq_seqno == 0) {
58		dev_priv = ring->dev->dev_private;
59		mtx_lock(&dev_priv->irq_lock);
60		if (ring->irq_get(ring))
61			ring->trace_irq_seqno = seqno;
62		mtx_unlock(&dev_priv->irq_lock);
63	}
64}
65
66static inline int ring_space(struct intel_ring_buffer *ring)
67{
68	int space = (ring->head & HEAD_ADDR) - (ring->tail + 8);
69	if (space < 0)
70		space += ring->size;
71	return space;
72}
73
74static int
75gen2_render_ring_flush(struct intel_ring_buffer *ring,
76		       u32	invalidate_domains,
77		       u32	flush_domains)
78{
79	u32 cmd;
80	int ret;
81
82	cmd = MI_FLUSH;
83	if (((invalidate_domains|flush_domains) & I915_GEM_DOMAIN_RENDER) == 0)
84		cmd |= MI_NO_WRITE_FLUSH;
85
86	if (invalidate_domains & I915_GEM_DOMAIN_SAMPLER)
87		cmd |= MI_READ_FLUSH;
88
89	ret = intel_ring_begin(ring, 2);
90	if (ret)
91		return ret;
92
93	intel_ring_emit(ring, cmd);
94	intel_ring_emit(ring, MI_NOOP);
95	intel_ring_advance(ring);
96
97	return 0;
98}
99
100static int
101gen4_render_ring_flush(struct intel_ring_buffer *ring,
102		       u32	invalidate_domains,
103		       u32	flush_domains)
104{
105	struct drm_device *dev = ring->dev;
106	u32 cmd;
107	int ret;
108
109	/*
110	 * read/write caches:
111	 *
112	 * I915_GEM_DOMAIN_RENDER is always invalidated, but is
113	 * only flushed if MI_NO_WRITE_FLUSH is unset.  On 965, it is
114	 * also flushed at 2d versus 3d pipeline switches.
115	 *
116	 * read-only caches:
117	 *
118	 * I915_GEM_DOMAIN_SAMPLER is flushed on pre-965 if
119	 * MI_READ_FLUSH is set, and is always flushed on 965.
120	 *
121	 * I915_GEM_DOMAIN_COMMAND may not exist?
122	 *
123	 * I915_GEM_DOMAIN_INSTRUCTION, which exists on 965, is
124	 * invalidated when MI_EXE_FLUSH is set.
125	 *
126	 * I915_GEM_DOMAIN_VERTEX, which exists on 965, is
127	 * invalidated with every MI_FLUSH.
128	 *
129	 * TLBs:
130	 *
131	 * On 965, TLBs associated with I915_GEM_DOMAIN_COMMAND
132	 * and I915_GEM_DOMAIN_CPU in are invalidated at PTE write and
133	 * I915_GEM_DOMAIN_RENDER and I915_GEM_DOMAIN_SAMPLER
134	 * are flushed at any MI_FLUSH.
135	 */
136
137	cmd = MI_FLUSH | MI_NO_WRITE_FLUSH;
138	if ((invalidate_domains|flush_domains) & I915_GEM_DOMAIN_RENDER)
139		cmd &= ~MI_NO_WRITE_FLUSH;
140	if (invalidate_domains & I915_GEM_DOMAIN_INSTRUCTION)
141		cmd |= MI_EXE_FLUSH;
142
143	if (invalidate_domains & I915_GEM_DOMAIN_COMMAND &&
144	    (IS_G4X(dev) || IS_GEN5(dev)))
145		cmd |= MI_INVALIDATE_ISP;
146
147	ret = intel_ring_begin(ring, 2);
148	if (ret)
149		return ret;
150
151	intel_ring_emit(ring, cmd);
152	intel_ring_emit(ring, MI_NOOP);
153	intel_ring_advance(ring);
154
155	return 0;
156}
157
158/**
159 * Emits a PIPE_CONTROL with a non-zero post-sync operation, for
160 * implementing two workarounds on gen6.  From section 1.4.7.1
161 * "PIPE_CONTROL" of the Sandy Bridge PRM volume 2 part 1:
162 *
163 * [DevSNB-C+{W/A}] Before any depth stall flush (including those
164 * produced by non-pipelined state commands), software needs to first
165 * send a PIPE_CONTROL with no bits set except Post-Sync Operation !=
166 * 0.
167 *
168 * [Dev-SNB{W/A}]: Before a PIPE_CONTROL with Write Cache Flush Enable
169 * =1, a PIPE_CONTROL with any non-zero post-sync-op is required.
170 *
171 * And the workaround for these two requires this workaround first:
172 *
173 * [Dev-SNB{W/A}]: Pipe-control with CS-stall bit set must be sent
174 * BEFORE the pipe-control with a post-sync op and no write-cache
175 * flushes.
176 *
177 * And this last workaround is tricky because of the requirements on
178 * that bit.  From section 1.4.7.2.3 "Stall" of the Sandy Bridge PRM
179 * volume 2 part 1:
180 *
181 *     "1 of the following must also be set:
182 *      - Render Target Cache Flush Enable ([12] of DW1)
183 *      - Depth Cache Flush Enable ([0] of DW1)
184 *      - Stall at Pixel Scoreboard ([1] of DW1)
185 *      - Depth Stall ([13] of DW1)
186 *      - Post-Sync Operation ([13] of DW1)
187 *      - Notify Enable ([8] of DW1)"
188 *
189 * The cache flushes require the workaround flush that triggered this
190 * one, so we can't use it.  Depth stall would trigger the same.
191 * Post-sync nonzero is what triggered this second workaround, so we
192 * can't use that one either.  Notify enable is IRQs, which aren't
193 * really our business.  That leaves only stall at scoreboard.
194 */
195static int
196intel_emit_post_sync_nonzero_flush(struct intel_ring_buffer *ring)
197{
198	struct pipe_control *pc = ring->private;
199	u32 scratch_addr = pc->gtt_offset + 128;
200	int ret;
201
202
203	ret = intel_ring_begin(ring, 6);
204	if (ret)
205		return ret;
206
207	intel_ring_emit(ring, GFX_OP_PIPE_CONTROL(5));
208	intel_ring_emit(ring, PIPE_CONTROL_CS_STALL |
209			PIPE_CONTROL_STALL_AT_SCOREBOARD);
210	intel_ring_emit(ring, scratch_addr | PIPE_CONTROL_GLOBAL_GTT); /* address */
211	intel_ring_emit(ring, 0); /* low dword */
212	intel_ring_emit(ring, 0); /* high dword */
213	intel_ring_emit(ring, MI_NOOP);
214	intel_ring_advance(ring);
215
216	ret = intel_ring_begin(ring, 6);
217	if (ret)
218		return ret;
219
220	intel_ring_emit(ring, GFX_OP_PIPE_CONTROL(5));
221	intel_ring_emit(ring, PIPE_CONTROL_QW_WRITE);
222	intel_ring_emit(ring, scratch_addr | PIPE_CONTROL_GLOBAL_GTT); /* address */
223	intel_ring_emit(ring, 0);
224	intel_ring_emit(ring, 0);
225	intel_ring_emit(ring, MI_NOOP);
226	intel_ring_advance(ring);
227
228	return 0;
229}
230
231static int
232gen6_render_ring_flush(struct intel_ring_buffer *ring,
233                         u32 invalidate_domains, u32 flush_domains)
234{
235	u32 flags = 0;
236	struct pipe_control *pc = ring->private;
237	u32 scratch_addr = pc->gtt_offset + 128;
238	int ret;
239
240	/* Force SNB workarounds for PIPE_CONTROL flushes */
241	intel_emit_post_sync_nonzero_flush(ring);
242
243	/* Just flush everything.  Experiments have shown that reducing the
244	 * number of bits based on the write domains has little performance
245	 * impact.
246	 */
247	flags |= PIPE_CONTROL_RENDER_TARGET_CACHE_FLUSH;
248	flags |= PIPE_CONTROL_INSTRUCTION_CACHE_INVALIDATE;
249	flags |= PIPE_CONTROL_TEXTURE_CACHE_INVALIDATE;
250	flags |= PIPE_CONTROL_DEPTH_CACHE_FLUSH;
251	flags |= PIPE_CONTROL_VF_CACHE_INVALIDATE;
252	flags |= PIPE_CONTROL_CONST_CACHE_INVALIDATE;
253	flags |= PIPE_CONTROL_STATE_CACHE_INVALIDATE;
254
255	ret = intel_ring_begin(ring, 6);
256	if (ret)
257		return ret;
258
259	intel_ring_emit(ring, GFX_OP_PIPE_CONTROL(5));
260	intel_ring_emit(ring, flags);
261	intel_ring_emit(ring, scratch_addr | PIPE_CONTROL_GLOBAL_GTT);
262	intel_ring_emit(ring, 0); /* lower dword */
263	intel_ring_emit(ring, 0); /* uppwer dword */
264	intel_ring_emit(ring, MI_NOOP);
265	intel_ring_advance(ring);
266
267	return 0;
268}
269
270static void ring_write_tail(struct intel_ring_buffer *ring,
271			    u32 value)
272{
273	drm_i915_private_t *dev_priv = ring->dev->dev_private;
274	I915_WRITE_TAIL(ring, value);
275}
276
277u32 intel_ring_get_active_head(struct intel_ring_buffer *ring)
278{
279	drm_i915_private_t *dev_priv = ring->dev->dev_private;
280	u32 acthd_reg = INTEL_INFO(ring->dev)->gen >= 4 ?
281			RING_ACTHD(ring->mmio_base) : ACTHD;
282
283	return I915_READ(acthd_reg);
284}
285
286static int init_ring_common(struct intel_ring_buffer *ring)
287{
288	drm_i915_private_t *dev_priv = ring->dev->dev_private;
289	struct drm_i915_gem_object *obj = ring->obj;
290	u32 head;
291
292	/* Stop the ring if it's running. */
293	I915_WRITE_CTL(ring, 0);
294	I915_WRITE_HEAD(ring, 0);
295	ring->write_tail(ring, 0);
296
297	/* Initialize the ring. */
298	I915_WRITE_START(ring, obj->gtt_offset);
299	head = I915_READ_HEAD(ring) & HEAD_ADDR;
300
301	/* G45 ring initialization fails to reset head to zero */
302	if (head != 0) {
303		DRM_DEBUG_KMS("%s head not reset to zero "
304			      "ctl %08x head %08x tail %08x start %08x\n",
305			      ring->name,
306			      I915_READ_CTL(ring),
307			      I915_READ_HEAD(ring),
308			      I915_READ_TAIL(ring),
309			      I915_READ_START(ring));
310
311		I915_WRITE_HEAD(ring, 0);
312
313		if (I915_READ_HEAD(ring) & HEAD_ADDR) {
314			DRM_ERROR("failed to set %s head to zero "
315				  "ctl %08x head %08x tail %08x start %08x\n",
316				  ring->name,
317				  I915_READ_CTL(ring),
318				  I915_READ_HEAD(ring),
319				  I915_READ_TAIL(ring),
320				  I915_READ_START(ring));
321		}
322	}
323
324	I915_WRITE_CTL(ring,
325			((ring->size - PAGE_SIZE) & RING_NR_PAGES)
326			| RING_VALID);
327
328	/* If the head is still not zero, the ring is dead */
329	if (_intel_wait_for(ring->dev,
330	    (I915_READ_CTL(ring) & RING_VALID) != 0 &&
331	     I915_READ_START(ring) == obj->gtt_offset &&
332	     (I915_READ_HEAD(ring) & HEAD_ADDR) == 0,
333	    50, 1, "915rii")) {
334		DRM_ERROR("%s initialization failed "
335				"ctl %08x head %08x tail %08x start %08x\n",
336				ring->name,
337				I915_READ_CTL(ring),
338				I915_READ_HEAD(ring),
339				I915_READ_TAIL(ring),
340				I915_READ_START(ring));
341		return -EIO;
342	}
343
344	if (!drm_core_check_feature(ring->dev, DRIVER_MODESET))
345		i915_kernel_lost_context(ring->dev);
346	else {
347		ring->head = I915_READ_HEAD(ring);
348		ring->tail = I915_READ_TAIL(ring) & TAIL_ADDR;
349		ring->space = ring_space(ring);
350	}
351
352	return 0;
353}
354
355static int
356init_pipe_control(struct intel_ring_buffer *ring)
357{
358	struct pipe_control *pc;
359	struct drm_i915_gem_object *obj;
360	int ret;
361
362	if (ring->private)
363		return 0;
364
365	pc = malloc(sizeof(*pc), DRM_I915_GEM, M_WAITOK);
366	if (!pc)
367		return -ENOMEM;
368
369	obj = i915_gem_alloc_object(ring->dev, 4096);
370	if (obj == NULL) {
371		DRM_ERROR("Failed to allocate seqno page\n");
372		ret = -ENOMEM;
373		goto err;
374	}
375
376	i915_gem_object_set_cache_level(obj, I915_CACHE_LLC);
377
378	ret = i915_gem_object_pin(obj, 4096, true);
379	if (ret)
380		goto err_unref;
381
382	pc->gtt_offset = obj->gtt_offset;
383	pc->cpu_page = (uint32_t *)kva_alloc(PAGE_SIZE);
384	if (pc->cpu_page == NULL)
385		goto err_unpin;
386	pmap_qenter((uintptr_t)pc->cpu_page, &obj->pages[0], 1);
387	pmap_invalidate_cache_range((vm_offset_t)pc->cpu_page,
388	    (vm_offset_t)pc->cpu_page + PAGE_SIZE, FALSE);
389
390	pc->obj = obj;
391	ring->private = pc;
392	return 0;
393
394err_unpin:
395	i915_gem_object_unpin(obj);
396err_unref:
397	drm_gem_object_unreference(&obj->base);
398err:
399	free(pc, DRM_I915_GEM);
400	return ret;
401}
402
403static void
404cleanup_pipe_control(struct intel_ring_buffer *ring)
405{
406	struct pipe_control *pc = ring->private;
407	struct drm_i915_gem_object *obj;
408
409	if (!ring->private)
410		return;
411
412	obj = pc->obj;
413
414	pmap_qremove((vm_offset_t)pc->cpu_page, 1);
415	kva_free((uintptr_t)pc->cpu_page, PAGE_SIZE);
416	i915_gem_object_unpin(obj);
417	drm_gem_object_unreference(&obj->base);
418
419	free(pc, DRM_I915_GEM);
420	ring->private = NULL;
421}
422
423static int init_render_ring(struct intel_ring_buffer *ring)
424{
425	struct drm_device *dev = ring->dev;
426	struct drm_i915_private *dev_priv = dev->dev_private;
427	int ret = init_ring_common(ring);
428
429	if (INTEL_INFO(dev)->gen > 3) {
430		I915_WRITE(MI_MODE, _MASKED_BIT_ENABLE(VS_TIMER_DISPATCH));
431		if (IS_GEN7(dev))
432			I915_WRITE(GFX_MODE_GEN7,
433				   _MASKED_BIT_DISABLE(GFX_TLB_INVALIDATE_ALWAYS) |
434				   _MASKED_BIT_ENABLE(GFX_REPLAY_MODE));
435	}
436
437	if (INTEL_INFO(dev)->gen >= 5) {
438		ret = init_pipe_control(ring);
439		if (ret)
440			return ret;
441	}
442
443	if (IS_GEN6(dev)) {
444		/* From the Sandybridge PRM, volume 1 part 3, page 24:
445		 * "If this bit is set, STCunit will have LRA as replacement
446		 *  policy. [...] This bit must be reset.  LRA replacement
447		 *  policy is not supported."
448		 */
449		I915_WRITE(CACHE_MODE_0,
450			   _MASKED_BIT_DISABLE(CM0_STC_EVICT_DISABLE_LRA_SNB));
451
452		/* This is not explicitly set for GEN6, so read the register.
453		 * see intel_ring_mi_set_context() for why we care.
454		 * TODO: consider explicitly setting the bit for GEN5
455		 */
456		ring->itlb_before_ctx_switch =
457			!!(I915_READ(GFX_MODE) & GFX_TLB_INVALIDATE_ALWAYS);
458	}
459
460	if (INTEL_INFO(dev)->gen >= 6)
461		I915_WRITE(INSTPM, _MASKED_BIT_ENABLE(INSTPM_FORCE_ORDERING));
462
463	return ret;
464}
465
466static void render_ring_cleanup(struct intel_ring_buffer *ring)
467{
468	if (!ring->private)
469		return;
470
471	cleanup_pipe_control(ring);
472}
473
474static void
475update_mboxes(struct intel_ring_buffer *ring,
476	      u32 seqno,
477	      u32 mmio_offset)
478{
479	intel_ring_emit(ring, MI_SEMAPHORE_MBOX |
480			      MI_SEMAPHORE_GLOBAL_GTT |
481			      MI_SEMAPHORE_REGISTER |
482			      MI_SEMAPHORE_UPDATE);
483	intel_ring_emit(ring, seqno);
484	intel_ring_emit(ring, mmio_offset);
485}
486
487/**
488 * gen6_add_request - Update the semaphore mailbox registers
489 *
490 * @ring - ring that is adding a request
491 * @seqno - return seqno stuck into the ring
492 *
493 * Update the mailbox registers in the *other* rings with the current seqno.
494 * This acts like a signal in the canonical semaphore.
495 */
496static int
497gen6_add_request(struct intel_ring_buffer *ring,
498		 u32 *seqno)
499{
500	u32 mbox1_reg;
501	u32 mbox2_reg;
502	int ret;
503
504	ret = intel_ring_begin(ring, 10);
505	if (ret)
506		return ret;
507
508	mbox1_reg = ring->signal_mbox[0];
509	mbox2_reg = ring->signal_mbox[1];
510
511	*seqno = i915_gem_next_request_seqno(ring);
512
513	update_mboxes(ring, *seqno, mbox1_reg);
514	update_mboxes(ring, *seqno, mbox2_reg);
515	intel_ring_emit(ring, MI_STORE_DWORD_INDEX);
516	intel_ring_emit(ring, I915_GEM_HWS_INDEX << MI_STORE_DWORD_INDEX_SHIFT);
517	intel_ring_emit(ring, *seqno);
518	intel_ring_emit(ring, MI_USER_INTERRUPT);
519	intel_ring_advance(ring);
520
521	return 0;
522}
523
524/**
525 * intel_ring_sync - sync the waiter to the signaller on seqno
526 *
527 * @waiter - ring that is waiting
528 * @signaller - ring which has, or will signal
529 * @seqno - seqno which the waiter will block on
530 */
531static int
532gen6_ring_sync(struct intel_ring_buffer *waiter,
533	       struct intel_ring_buffer *signaller,
534	       u32 seqno)
535{
536	int ret;
537	u32 dw1 = MI_SEMAPHORE_MBOX |
538		  MI_SEMAPHORE_COMPARE |
539		  MI_SEMAPHORE_REGISTER;
540
541	/* Throughout all of the GEM code, seqno passed implies our current
542	 * seqno is >= the last seqno executed. However for hardware the
543	 * comparison is strictly greater than.
544	 */
545	seqno -= 1;
546
547	if (signaller->semaphore_register[waiter->id] ==
548	    MI_SEMAPHORE_SYNC_INVALID)
549		printf("gen6_ring_sync semaphore_register %d invalid\n",
550		    waiter->id);
551
552	ret = intel_ring_begin(waiter, 4);
553	if (ret)
554		return ret;
555
556	intel_ring_emit(waiter,
557			dw1 | signaller->semaphore_register[waiter->id]);
558	intel_ring_emit(waiter, seqno);
559	intel_ring_emit(waiter, 0);
560	intel_ring_emit(waiter, MI_NOOP);
561	intel_ring_advance(waiter);
562
563	return 0;
564}
565
566int render_ring_sync_to(struct intel_ring_buffer *waiter,
567    struct intel_ring_buffer *signaller, u32 seqno);
568int gen6_bsd_ring_sync_to(struct intel_ring_buffer *waiter,
569    struct intel_ring_buffer *signaller, u32 seqno);
570int gen6_blt_ring_sync_to(struct intel_ring_buffer *waiter,
571    struct intel_ring_buffer *signaller, u32 seqno);
572
573#define PIPE_CONTROL_FLUSH(ring__, addr__)					\
574do {									\
575	intel_ring_emit(ring__, GFX_OP_PIPE_CONTROL(4) | PIPE_CONTROL_QW_WRITE |		\
576		 PIPE_CONTROL_DEPTH_STALL);				\
577	intel_ring_emit(ring__, (addr__) | PIPE_CONTROL_GLOBAL_GTT);			\
578	intel_ring_emit(ring__, 0);							\
579	intel_ring_emit(ring__, 0);							\
580} while (0)
581
582static int
583pc_render_add_request(struct intel_ring_buffer *ring,
584		      uint32_t *result)
585{
586	u32 seqno = i915_gem_next_request_seqno(ring);
587	struct pipe_control *pc = ring->private;
588	u32 scratch_addr = pc->gtt_offset + 128;
589	int ret;
590
591	/* For Ironlake, MI_USER_INTERRUPT was deprecated and apparently
592	 * incoherent with writes to memory, i.e. completely fubar,
593	 * so we need to use PIPE_NOTIFY instead.
594	 *
595	 * However, we also need to workaround the qword write
596	 * incoherence by flushing the 6 PIPE_NOTIFY buffers out to
597	 * memory before requesting an interrupt.
598	 */
599	ret = intel_ring_begin(ring, 32);
600	if (ret)
601		return ret;
602
603	intel_ring_emit(ring, GFX_OP_PIPE_CONTROL(4) | PIPE_CONTROL_QW_WRITE |
604			PIPE_CONTROL_WRITE_FLUSH |
605			PIPE_CONTROL_TEXTURE_CACHE_INVALIDATE);
606	intel_ring_emit(ring, pc->gtt_offset | PIPE_CONTROL_GLOBAL_GTT);
607	intel_ring_emit(ring, seqno);
608	intel_ring_emit(ring, 0);
609	PIPE_CONTROL_FLUSH(ring, scratch_addr);
610	scratch_addr += 128; /* write to separate cachelines */
611	PIPE_CONTROL_FLUSH(ring, scratch_addr);
612	scratch_addr += 128;
613	PIPE_CONTROL_FLUSH(ring, scratch_addr);
614	scratch_addr += 128;
615	PIPE_CONTROL_FLUSH(ring, scratch_addr);
616	scratch_addr += 128;
617	PIPE_CONTROL_FLUSH(ring, scratch_addr);
618	scratch_addr += 128;
619	PIPE_CONTROL_FLUSH(ring, scratch_addr);
620	intel_ring_emit(ring, GFX_OP_PIPE_CONTROL(4) | PIPE_CONTROL_QW_WRITE |
621			PIPE_CONTROL_WRITE_FLUSH |
622			PIPE_CONTROL_TEXTURE_CACHE_INVALIDATE |
623			PIPE_CONTROL_NOTIFY);
624	intel_ring_emit(ring, pc->gtt_offset | PIPE_CONTROL_GLOBAL_GTT);
625	intel_ring_emit(ring, seqno);
626	intel_ring_emit(ring, 0);
627	intel_ring_advance(ring);
628
629	*result = seqno;
630	return 0;
631}
632
633static u32
634gen6_ring_get_seqno(struct intel_ring_buffer *ring)
635{
636	struct drm_device *dev = ring->dev;
637
638	/* Workaround to force correct ordering between irq and seqno writes on
639	 * ivb (and maybe also on snb) by reading from a CS register (like
640	 * ACTHD) before reading the status page. */
641	if (/* IS_GEN6(dev) || */IS_GEN7(dev))
642		intel_ring_get_active_head(ring);
643	return intel_read_status_page(ring, I915_GEM_HWS_INDEX);
644}
645
646static u32
647ring_get_seqno(struct intel_ring_buffer *ring)
648{
649	if (ring->status_page.page_addr == NULL)
650		return (-1);
651	return intel_read_status_page(ring, I915_GEM_HWS_INDEX);
652}
653
654static u32
655pc_render_get_seqno(struct intel_ring_buffer *ring)
656{
657	struct pipe_control *pc = ring->private;
658	if (pc != NULL)
659		return pc->cpu_page[0];
660	else
661		return (-1);
662}
663
664static bool
665gen5_ring_get_irq(struct intel_ring_buffer *ring)
666{
667	struct drm_device *dev = ring->dev;
668	drm_i915_private_t *dev_priv = dev->dev_private;
669
670	if (!dev->irq_enabled)
671		return false;
672
673	mtx_assert(&dev_priv->irq_lock, MA_OWNED);
674	if (ring->irq_refcount++ == 0) {
675		dev_priv->gt_irq_mask &= ~ring->irq_enable_mask;
676		I915_WRITE(GTIMR, dev_priv->gt_irq_mask);
677		POSTING_READ(GTIMR);
678	}
679
680	return true;
681}
682
683static void
684gen5_ring_put_irq(struct intel_ring_buffer *ring)
685{
686	struct drm_device *dev = ring->dev;
687	drm_i915_private_t *dev_priv = dev->dev_private;
688
689	mtx_assert(&dev_priv->irq_lock, MA_OWNED);
690	if (--ring->irq_refcount == 0) {
691		dev_priv->gt_irq_mask |= ring->irq_enable_mask;
692		I915_WRITE(GTIMR, dev_priv->gt_irq_mask);
693		POSTING_READ(GTIMR);
694	}
695}
696
697static bool
698i9xx_ring_get_irq(struct intel_ring_buffer *ring)
699{
700	struct drm_device *dev = ring->dev;
701	drm_i915_private_t *dev_priv = dev->dev_private;
702
703	if (!dev->irq_enabled)
704		return false;
705
706	mtx_assert(&dev_priv->irq_lock, MA_OWNED);
707	if (ring->irq_refcount++ == 0) {
708		dev_priv->irq_mask &= ~ring->irq_enable_mask;
709		I915_WRITE(IMR, dev_priv->irq_mask);
710		POSTING_READ(IMR);
711	}
712
713	return true;
714}
715
716static void
717i9xx_ring_put_irq(struct intel_ring_buffer *ring)
718{
719	struct drm_device *dev = ring->dev;
720	drm_i915_private_t *dev_priv = dev->dev_private;
721
722	mtx_assert(&dev_priv->irq_lock, MA_OWNED);
723	if (--ring->irq_refcount == 0) {
724		dev_priv->irq_mask |= ring->irq_enable_mask;
725		I915_WRITE(IMR, dev_priv->irq_mask);
726		POSTING_READ(IMR);
727	}
728}
729
730static bool
731i8xx_ring_get_irq(struct intel_ring_buffer *ring)
732{
733	struct drm_device *dev = ring->dev;
734	drm_i915_private_t *dev_priv = dev->dev_private;
735
736	if (!dev->irq_enabled)
737		return false;
738
739	mtx_assert(&dev_priv->irq_lock, MA_OWNED);
740	if (ring->irq_refcount++ == 0) {
741		dev_priv->irq_mask &= ~ring->irq_enable_mask;
742		I915_WRITE16(IMR, dev_priv->irq_mask);
743		POSTING_READ16(IMR);
744	}
745
746	return true;
747}
748
749static void
750i8xx_ring_put_irq(struct intel_ring_buffer *ring)
751{
752	struct drm_device *dev = ring->dev;
753	drm_i915_private_t *dev_priv = dev->dev_private;
754
755	mtx_assert(&dev_priv->irq_lock, MA_OWNED);
756	if (--ring->irq_refcount == 0) {
757		dev_priv->irq_mask |= ring->irq_enable_mask;
758		I915_WRITE16(IMR, dev_priv->irq_mask);
759		POSTING_READ16(IMR);
760	}
761}
762
763void intel_ring_setup_status_page(struct intel_ring_buffer *ring)
764{
765	struct drm_device *dev = ring->dev;
766	drm_i915_private_t *dev_priv = dev->dev_private;
767	u32 mmio = 0;
768
769	/* The ring status page addresses are no longer next to the rest of
770	 * the ring registers as of gen7.
771	 */
772	if (IS_GEN7(dev)) {
773		switch (ring->id) {
774		case RCS:
775			mmio = RENDER_HWS_PGA_GEN7;
776			break;
777		case BCS:
778			mmio = BLT_HWS_PGA_GEN7;
779			break;
780		case VCS:
781			mmio = BSD_HWS_PGA_GEN7;
782			break;
783		}
784	} else if (IS_GEN6(dev)) {
785		mmio = RING_HWS_PGA_GEN6(ring->mmio_base);
786	} else {
787		mmio = RING_HWS_PGA(ring->mmio_base);
788	}
789
790	I915_WRITE(mmio, (u32)ring->status_page.gfx_addr);
791	POSTING_READ(mmio);
792}
793
794static int
795bsd_ring_flush(struct intel_ring_buffer *ring,
796	       u32     invalidate_domains,
797	       u32     flush_domains)
798{
799	int ret;
800
801	ret = intel_ring_begin(ring, 2);
802	if (ret)
803		return ret;
804
805	intel_ring_emit(ring, MI_FLUSH);
806	intel_ring_emit(ring, MI_NOOP);
807	intel_ring_advance(ring);
808	return 0;
809}
810
811static int
812i9xx_add_request(struct intel_ring_buffer *ring,
813		 u32 *result)
814{
815	u32 seqno;
816	int ret;
817
818	ret = intel_ring_begin(ring, 4);
819	if (ret)
820		return ret;
821
822	seqno = i915_gem_next_request_seqno(ring);
823
824	intel_ring_emit(ring, MI_STORE_DWORD_INDEX);
825	intel_ring_emit(ring, I915_GEM_HWS_INDEX << MI_STORE_DWORD_INDEX_SHIFT);
826	intel_ring_emit(ring, seqno);
827	intel_ring_emit(ring, MI_USER_INTERRUPT);
828	intel_ring_advance(ring);
829
830	*result = seqno;
831	return 0;
832}
833
834static bool
835gen6_ring_get_irq(struct intel_ring_buffer *ring)
836{
837	struct drm_device *dev = ring->dev;
838	drm_i915_private_t *dev_priv = dev->dev_private;
839
840	if (!dev->irq_enabled)
841	       return false;
842
843	gen6_gt_force_wake_get(dev_priv);
844
845	mtx_assert(&dev_priv->irq_lock, MA_OWNED);
846	if (ring->irq_refcount++ == 0) {
847		I915_WRITE_IMR(ring, ~ring->irq_enable_mask);
848		dev_priv->gt_irq_mask &= ~ring->irq_enable_mask;
849		I915_WRITE(GTIMR, dev_priv->gt_irq_mask);
850		POSTING_READ(GTIMR);
851	}
852
853	return true;
854}
855
856static void
857gen6_ring_put_irq(struct intel_ring_buffer *ring)
858{
859	struct drm_device *dev = ring->dev;
860	drm_i915_private_t *dev_priv = dev->dev_private;
861
862	mtx_assert(&dev_priv->irq_lock, MA_OWNED);
863	if (--ring->irq_refcount == 0) {
864		I915_WRITE_IMR(ring, ~0);
865		dev_priv->gt_irq_mask |= ring->irq_enable_mask;
866		I915_WRITE(GTIMR, dev_priv->gt_irq_mask);
867		POSTING_READ(GTIMR);
868	}
869
870	gen6_gt_force_wake_put(dev_priv);
871}
872
873static int
874i965_dispatch_execbuffer(struct intel_ring_buffer *ring,
875			 u32 offset, u32 length)
876{
877	int ret;
878
879	ret = intel_ring_begin(ring, 2);
880	if (ret)
881		return ret;
882
883	intel_ring_emit(ring,
884			MI_BATCH_BUFFER_START |
885			MI_BATCH_GTT |
886			MI_BATCH_NON_SECURE_I965);
887	intel_ring_emit(ring, offset);
888	intel_ring_advance(ring);
889
890	return 0;
891}
892
893static int
894i830_dispatch_execbuffer(struct intel_ring_buffer *ring,
895				u32 offset, u32 len)
896{
897	int ret;
898
899	ret = intel_ring_begin(ring, 4);
900	if (ret)
901		return ret;
902
903	intel_ring_emit(ring, MI_BATCH_BUFFER);
904	intel_ring_emit(ring, offset | MI_BATCH_NON_SECURE);
905	intel_ring_emit(ring, offset + len - 8);
906	intel_ring_emit(ring, 0);
907	intel_ring_advance(ring);
908
909	return 0;
910}
911
912static int
913i915_dispatch_execbuffer(struct intel_ring_buffer *ring,
914			 u32 offset, u32 len)
915{
916	int ret;
917
918	ret = intel_ring_begin(ring, 2);
919	if (ret)
920		return ret;
921
922	intel_ring_emit(ring, MI_BATCH_BUFFER_START | MI_BATCH_GTT);
923	intel_ring_emit(ring, offset | MI_BATCH_NON_SECURE);
924	intel_ring_advance(ring);
925
926	return 0;
927}
928
929static void cleanup_status_page(struct intel_ring_buffer *ring)
930{
931	struct drm_i915_gem_object *obj;
932
933	obj = ring->status_page.obj;
934	if (obj == NULL)
935		return;
936
937	pmap_qremove((vm_offset_t)ring->status_page.page_addr, 1);
938	kva_free((vm_offset_t)ring->status_page.page_addr,
939	    PAGE_SIZE);
940	i915_gem_object_unpin(obj);
941	drm_gem_object_unreference(&obj->base);
942	ring->status_page.obj = NULL;
943}
944
945static int init_status_page(struct intel_ring_buffer *ring)
946{
947	struct drm_device *dev = ring->dev;
948	struct drm_i915_gem_object *obj;
949	int ret;
950
951	obj = i915_gem_alloc_object(dev, 4096);
952	if (obj == NULL) {
953		DRM_ERROR("Failed to allocate status page\n");
954		ret = -ENOMEM;
955		goto err;
956	}
957
958	i915_gem_object_set_cache_level(obj, I915_CACHE_LLC);
959
960	ret = i915_gem_object_pin(obj, 4096, true);
961	if (ret != 0) {
962		goto err_unref;
963	}
964
965	ring->status_page.gfx_addr = obj->gtt_offset;
966	ring->status_page.page_addr = (void *)kva_alloc(PAGE_SIZE);
967	if (ring->status_page.page_addr == NULL) {
968		goto err_unpin;
969	}
970	pmap_qenter((vm_offset_t)ring->status_page.page_addr, &obj->pages[0],
971	    1);
972	pmap_invalidate_cache_range((vm_offset_t)ring->status_page.page_addr,
973	    (vm_offset_t)ring->status_page.page_addr + PAGE_SIZE, FALSE);
974	ring->status_page.obj = obj;
975	memset(ring->status_page.page_addr, 0, PAGE_SIZE);
976
977	intel_ring_setup_status_page(ring);
978	DRM_DEBUG_DRIVER("%s hws offset: 0x%08x\n",
979			ring->name, ring->status_page.gfx_addr);
980
981	return 0;
982
983err_unpin:
984	i915_gem_object_unpin(obj);
985err_unref:
986	drm_gem_object_unreference(&obj->base);
987err:
988	return ret;
989}
990
991static int intel_init_ring_buffer(struct drm_device *dev,
992			   struct intel_ring_buffer *ring)
993{
994	struct drm_i915_gem_object *obj;
995	int ret;
996
997	ring->dev = dev;
998	INIT_LIST_HEAD(&ring->active_list);
999	INIT_LIST_HEAD(&ring->request_list);
1000	INIT_LIST_HEAD(&ring->gpu_write_list);
1001	ring->size = 32 * PAGE_SIZE;
1002
1003	if (I915_NEED_GFX_HWS(dev)) {
1004		ret = init_status_page(ring);
1005		if (ret)
1006			return ret;
1007	}
1008
1009	obj = i915_gem_alloc_object(dev, ring->size);
1010	if (obj == NULL) {
1011		DRM_ERROR("Failed to allocate ringbuffer\n");
1012		ret = -ENOMEM;
1013		goto err_hws;
1014	}
1015
1016	ring->obj = obj;
1017
1018	ret = i915_gem_object_pin(obj, PAGE_SIZE, true);
1019	if (ret)
1020		goto err_unref;
1021
1022	ring->virtual_start = pmap_mapdev_attr(
1023	    dev->agp->base + obj->gtt_offset, ring->size,
1024	    VM_MEMATTR_WRITE_COMBINING);
1025	if (ring->virtual_start == NULL) {
1026		DRM_ERROR("Failed to map ringbuffer.\n");
1027		ret = -EINVAL;
1028		goto err_unpin;
1029	}
1030
1031	ret = ring->init(ring);
1032	if (ret)
1033		goto err_unmap;
1034
1035	/* Workaround an erratum on the i830 which causes a hang if
1036	 * the TAIL pointer points to within the last 2 cachelines
1037	 * of the buffer.
1038	 */
1039	ring->effective_size = ring->size;
1040	if (IS_I830(ring->dev) || IS_845G(ring->dev))
1041		ring->effective_size -= 128;
1042
1043	return 0;
1044
1045err_unmap:
1046	pmap_unmapdev((vm_offset_t)ring->virtual_start, ring->size);
1047err_unpin:
1048	i915_gem_object_unpin(obj);
1049err_unref:
1050	drm_gem_object_unreference(&obj->base);
1051	ring->obj = NULL;
1052err_hws:
1053	cleanup_status_page(ring);
1054	return ret;
1055}
1056
1057void intel_cleanup_ring_buffer(struct intel_ring_buffer *ring)
1058{
1059	struct drm_i915_private *dev_priv;
1060	int ret;
1061
1062	if (ring->obj == NULL)
1063		return;
1064
1065	/* Disable the ring buffer. The ring must be idle at this point */
1066	dev_priv = ring->dev->dev_private;
1067	ret = intel_wait_ring_idle(ring);
1068	I915_WRITE_CTL(ring, 0);
1069
1070	pmap_unmapdev((vm_offset_t)ring->virtual_start, ring->size);
1071
1072	i915_gem_object_unpin(ring->obj);
1073	drm_gem_object_unreference(&ring->obj->base);
1074	ring->obj = NULL;
1075
1076	if (ring->cleanup)
1077		ring->cleanup(ring);
1078
1079	cleanup_status_page(ring);
1080}
1081
1082static int intel_ring_wait_seqno(struct intel_ring_buffer *ring, u32 seqno)
1083{
1084	struct drm_i915_private *dev_priv = ring->dev->dev_private;
1085	bool was_interruptible;
1086	int ret;
1087
1088	/* XXX As we have not yet audited all the paths to check that
1089	 * they are ready for ERESTARTSYS from intel_ring_begin, do not
1090	 * allow us to be interruptible by a signal.
1091	 */
1092	was_interruptible = dev_priv->mm.interruptible;
1093	dev_priv->mm.interruptible = false;
1094
1095	ret = i915_wait_request(ring, seqno);
1096
1097	dev_priv->mm.interruptible = was_interruptible;
1098	if (!ret)
1099		i915_gem_retire_requests_ring(ring);
1100
1101	return ret;
1102}
1103
1104static int intel_ring_wait_request(struct intel_ring_buffer *ring, int n)
1105{
1106	struct drm_i915_gem_request *request;
1107	u32 seqno = 0;
1108	int ret;
1109
1110	i915_gem_retire_requests_ring(ring);
1111
1112	if (ring->last_retired_head != -1) {
1113		ring->head = ring->last_retired_head;
1114		ring->last_retired_head = -1;
1115		ring->space = ring_space(ring);
1116		if (ring->space >= n)
1117			return 0;
1118	}
1119
1120	list_for_each_entry(request, &ring->request_list, list) {
1121		int space;
1122
1123		if (request->tail == -1)
1124			continue;
1125
1126		space = request->tail - (ring->tail + 8);
1127		if (space < 0)
1128			space += ring->size;
1129		if (space >= n) {
1130			seqno = request->seqno;
1131			break;
1132		}
1133
1134		/* Consume this request in case we need more space than
1135		 * is available and so need to prevent a race between
1136		 * updating last_retired_head and direct reads of
1137		 * I915_RING_HEAD. It also provides a nice sanity check.
1138		 */
1139		request->tail = -1;
1140	}
1141
1142	if (seqno == 0)
1143		return -ENOSPC;
1144
1145	ret = intel_ring_wait_seqno(ring, seqno);
1146	if (ret)
1147		return ret;
1148
1149	if (ring->last_retired_head == -1)
1150		return -ENOSPC;
1151
1152	ring->head = ring->last_retired_head;
1153	ring->last_retired_head = -1;
1154	ring->space = ring_space(ring);
1155	if (ring->space < n)
1156		return -ENOSPC;
1157
1158	return 0;
1159}
1160
1161int intel_wait_ring_buffer(struct intel_ring_buffer *ring, int n)
1162{
1163	struct drm_device *dev = ring->dev;
1164	struct drm_i915_private *dev_priv = dev->dev_private;
1165	int end;
1166	int ret;
1167
1168	ret = intel_ring_wait_request(ring, n);
1169	if (ret != -ENOSPC)
1170		return ret;
1171
1172	CTR1(KTR_DRM, "ring_wait_begin %s", ring->name);
1173	/* With GEM the hangcheck timer should kick us out of the loop,
1174	 * leaving it early runs the risk of corrupting GEM state (due
1175	 * to running on almost untested codepaths). But on resume
1176	 * timers don't work yet, so prevent a complete hang in that
1177	 * case by choosing an insanely large timeout. */
1178	end = ticks + hz * 60;
1179
1180	do {
1181		ring->head = I915_READ_HEAD(ring);
1182		ring->space = ring_space(ring);
1183		if (ring->space >= n) {
1184			CTR1(KTR_DRM, "ring_wait_end %s", ring->name);
1185			return 0;
1186		}
1187
1188		if (dev->primary->master) {
1189			struct drm_i915_master_private *master_priv = dev->primary->master->driver_priv;
1190			if (master_priv->sarea_priv)
1191				master_priv->sarea_priv->perf_boxes |= I915_BOX_WAIT;
1192		}
1193
1194		pause("915rng", 1);
1195		if (atomic_load_acq_32(&dev_priv->mm.wedged) != 0) {
1196			CTR1(KTR_DRM, "ring_wait_end %s wedged", ring->name);
1197			return -EAGAIN;
1198		}
1199	} while (!time_after(ticks, end));
1200	CTR1(KTR_DRM, "ring_wait_end %s busy", ring->name);
1201	return -EBUSY;
1202}
1203
1204static int intel_wrap_ring_buffer(struct intel_ring_buffer *ring)
1205{
1206	uint32_t *virt;
1207	int rem = ring->size - ring->tail;
1208
1209	if (ring->space < rem) {
1210		int ret = intel_wait_ring_buffer(ring, rem);
1211		if (ret)
1212			return ret;
1213	}
1214
1215	virt = (uint32_t *)((char *)ring->virtual_start + ring->tail);
1216	rem /= 4;
1217	while (rem--)
1218		*virt++ = MI_NOOP;
1219
1220	ring->tail = 0;
1221	ring->space = ring_space(ring);
1222
1223	return 0;
1224}
1225
1226int intel_ring_begin(struct intel_ring_buffer *ring,
1227		     int num_dwords)
1228{
1229	struct drm_i915_private *dev_priv = ring->dev->dev_private;
1230	int n = 4*num_dwords;
1231	int ret;
1232
1233	if (atomic_load_acq_int(&dev_priv->mm.wedged))
1234		return -EIO;
1235
1236	if (ring->tail + n > ring->effective_size) {
1237		ret = intel_wrap_ring_buffer(ring);
1238		if (ret != 0)
1239			return ret;
1240	}
1241
1242	if (ring->space < n) {
1243		ret = intel_wait_ring_buffer(ring, n);
1244		if (ret != 0)
1245			return ret;
1246	}
1247
1248	ring->space -= n;
1249	return 0;
1250}
1251
1252void intel_ring_advance(struct intel_ring_buffer *ring)
1253{
1254	struct drm_i915_private *dev_priv = ring->dev->dev_private;
1255
1256	ring->tail &= ring->size - 1;
1257	if (dev_priv->stop_rings & intel_ring_flag(ring))
1258		return;
1259	ring->write_tail(ring, ring->tail);
1260}
1261
1262
1263static void gen6_bsd_ring_write_tail(struct intel_ring_buffer *ring,
1264				     u32 value)
1265{
1266	drm_i915_private_t *dev_priv = ring->dev->dev_private;
1267
1268	/* Every tail move must follow the sequence below */
1269	I915_WRITE(GEN6_BSD_SLEEP_PSMI_CONTROL,
1270	    GEN6_BSD_SLEEP_PSMI_CONTROL_RC_ILDL_MESSAGE_MODIFY_MASK |
1271	    GEN6_BSD_SLEEP_PSMI_CONTROL_RC_ILDL_MESSAGE_DISABLE);
1272	I915_WRITE(GEN6_BSD_RNCID, 0x0);
1273
1274	if (_intel_wait_for(ring->dev,
1275	    (I915_READ(GEN6_BSD_SLEEP_PSMI_CONTROL) &
1276	     GEN6_BSD_SLEEP_PSMI_CONTROL_IDLE_INDICATOR) == 0, 50,
1277	    true, "915g6i") != 0)
1278		DRM_ERROR("timed out waiting for IDLE Indicator\n");
1279
1280	I915_WRITE_TAIL(ring, value);
1281	I915_WRITE(GEN6_BSD_SLEEP_PSMI_CONTROL,
1282	    GEN6_BSD_SLEEP_PSMI_CONTROL_RC_ILDL_MESSAGE_MODIFY_MASK |
1283	    GEN6_BSD_SLEEP_PSMI_CONTROL_RC_ILDL_MESSAGE_ENABLE);
1284}
1285
1286static int gen6_ring_flush(struct intel_ring_buffer *ring,
1287			   u32 invalidate, u32 flush)
1288{
1289	uint32_t cmd;
1290	int ret;
1291
1292	ret = intel_ring_begin(ring, 4);
1293	if (ret)
1294		return ret;
1295
1296	cmd = MI_FLUSH_DW;
1297	if (invalidate & I915_GEM_GPU_DOMAINS)
1298		cmd |= MI_INVALIDATE_TLB | MI_INVALIDATE_BSD;
1299	intel_ring_emit(ring, cmd);
1300	intel_ring_emit(ring, 0);
1301	intel_ring_emit(ring, 0);
1302	intel_ring_emit(ring, MI_NOOP);
1303	intel_ring_advance(ring);
1304	return 0;
1305}
1306
1307static int
1308gen6_ring_dispatch_execbuffer(struct intel_ring_buffer *ring,
1309			      u32 offset, u32 len)
1310{
1311	int ret;
1312
1313	ret = intel_ring_begin(ring, 2);
1314	if (ret)
1315		return ret;
1316
1317	intel_ring_emit(ring, MI_BATCH_BUFFER_START | MI_BATCH_NON_SECURE_I965);
1318	/* bit0-7 is the length on GEN6+ */
1319	intel_ring_emit(ring, offset);
1320	intel_ring_advance(ring);
1321
1322	return 0;
1323}
1324
1325/* Blitter support (SandyBridge+) */
1326
1327static int blt_ring_flush(struct intel_ring_buffer *ring,
1328			  u32 invalidate, u32 flush)
1329{
1330	uint32_t cmd;
1331	int ret;
1332
1333	ret = intel_ring_begin(ring, 4);
1334	if (ret)
1335		return ret;
1336
1337	cmd = MI_FLUSH_DW;
1338	if (invalidate & I915_GEM_DOMAIN_RENDER)
1339		cmd |= MI_INVALIDATE_TLB;
1340	intel_ring_emit(ring, cmd);
1341	intel_ring_emit(ring, 0);
1342	intel_ring_emit(ring, 0);
1343	intel_ring_emit(ring, MI_NOOP);
1344	intel_ring_advance(ring);
1345	return 0;
1346}
1347
1348int intel_init_render_ring_buffer(struct drm_device *dev)
1349{
1350	drm_i915_private_t *dev_priv = dev->dev_private;
1351	struct intel_ring_buffer *ring = &dev_priv->rings[RCS];
1352
1353	ring->name = "render ring";
1354	ring->id = RCS;
1355	ring->mmio_base = RENDER_RING_BASE;
1356
1357	if (INTEL_INFO(dev)->gen >= 6) {
1358		ring->add_request = gen6_add_request;
1359		ring->flush = gen6_render_ring_flush;
1360		ring->irq_get = gen6_ring_get_irq;
1361		ring->irq_put = gen6_ring_put_irq;
1362		ring->irq_enable_mask = GT_USER_INTERRUPT;
1363		ring->get_seqno = gen6_ring_get_seqno;
1364		ring->sync_to = gen6_ring_sync;
1365		ring->semaphore_register[0] = MI_SEMAPHORE_SYNC_INVALID;
1366		ring->semaphore_register[1] = MI_SEMAPHORE_SYNC_RV;
1367		ring->semaphore_register[2] = MI_SEMAPHORE_SYNC_RB;
1368		ring->signal_mbox[0] = GEN6_VRSYNC;
1369		ring->signal_mbox[1] = GEN6_BRSYNC;
1370	} else if (IS_GEN5(dev)) {
1371		ring->add_request = pc_render_add_request;
1372		ring->flush = gen4_render_ring_flush;
1373		ring->get_seqno = pc_render_get_seqno;
1374		ring->irq_get = gen5_ring_get_irq;
1375		ring->irq_put = gen5_ring_put_irq;
1376		ring->irq_enable_mask = GT_USER_INTERRUPT | GT_PIPE_NOTIFY;
1377	} else {
1378		ring->add_request = i9xx_add_request;
1379		if (INTEL_INFO(dev)->gen < 4)
1380			ring->flush = gen2_render_ring_flush;
1381		else
1382			ring->flush = gen4_render_ring_flush;
1383		ring->get_seqno = ring_get_seqno;
1384		if (IS_GEN2(dev)) {
1385			ring->irq_get = i8xx_ring_get_irq;
1386			ring->irq_put = i8xx_ring_put_irq;
1387		} else {
1388			ring->irq_get = i9xx_ring_get_irq;
1389			ring->irq_put = i9xx_ring_put_irq;
1390		}
1391		ring->irq_enable_mask = I915_USER_INTERRUPT;
1392	}
1393	ring->write_tail = ring_write_tail;
1394	if (INTEL_INFO(dev)->gen >= 6)
1395		ring->dispatch_execbuffer = gen6_ring_dispatch_execbuffer;
1396	else if (INTEL_INFO(dev)->gen >= 4)
1397		ring->dispatch_execbuffer = i965_dispatch_execbuffer;
1398	else if (IS_I830(dev) || IS_845G(dev))
1399		ring->dispatch_execbuffer = i830_dispatch_execbuffer;
1400	else
1401		ring->dispatch_execbuffer = i915_dispatch_execbuffer;
1402	ring->init = init_render_ring;
1403	ring->cleanup = render_ring_cleanup;
1404
1405
1406	if (!I915_NEED_GFX_HWS(dev)) {
1407		ring->status_page.page_addr = dev_priv->status_page_dmah->vaddr;
1408		memset(ring->status_page.page_addr, 0, PAGE_SIZE);
1409	}
1410
1411	return intel_init_ring_buffer(dev, ring);
1412}
1413
1414int intel_render_ring_init_dri(struct drm_device *dev, u64 start, u32 size)
1415{
1416	drm_i915_private_t *dev_priv = dev->dev_private;
1417	struct intel_ring_buffer *ring = &dev_priv->rings[RCS];
1418
1419	ring->name = "render ring";
1420	ring->id = RCS;
1421	ring->mmio_base = RENDER_RING_BASE;
1422
1423	if (INTEL_INFO(dev)->gen >= 6) {
1424		/* non-kms not supported on gen6+ */
1425		return -ENODEV;
1426	}
1427
1428	/* Note: gem is not supported on gen5/ilk without kms (the corresponding
1429	 * gem_init ioctl returns with -ENODEV). Hence we do not need to set up
1430	 * the special gen5 functions. */
1431	ring->add_request = i9xx_add_request;
1432	if (INTEL_INFO(dev)->gen < 4)
1433		ring->flush = gen2_render_ring_flush;
1434	else
1435		ring->flush = gen4_render_ring_flush;
1436	ring->get_seqno = ring_get_seqno;
1437	if (IS_GEN2(dev)) {
1438		ring->irq_get = i8xx_ring_get_irq;
1439		ring->irq_put = i8xx_ring_put_irq;
1440	} else {
1441		ring->irq_get = i9xx_ring_get_irq;
1442		ring->irq_put = i9xx_ring_put_irq;
1443	}
1444	ring->irq_enable_mask = I915_USER_INTERRUPT;
1445	ring->write_tail = ring_write_tail;
1446	if (INTEL_INFO(dev)->gen >= 4)
1447		ring->dispatch_execbuffer = i965_dispatch_execbuffer;
1448	else if (IS_I830(dev) || IS_845G(dev))
1449		ring->dispatch_execbuffer = i830_dispatch_execbuffer;
1450	else
1451		ring->dispatch_execbuffer = i915_dispatch_execbuffer;
1452	ring->init = init_render_ring;
1453	ring->cleanup = render_ring_cleanup;
1454
1455	ring->dev = dev;
1456	INIT_LIST_HEAD(&ring->active_list);
1457	INIT_LIST_HEAD(&ring->request_list);
1458	INIT_LIST_HEAD(&ring->gpu_write_list);
1459
1460	ring->size = size;
1461	ring->effective_size = ring->size;
1462	if (IS_I830(ring->dev))
1463		ring->effective_size -= 128;
1464
1465	ring->virtual_start = pmap_mapdev_attr(start, size,
1466	    VM_MEMATTR_WRITE_COMBINING);
1467	if (ring->virtual_start == NULL) {
1468		DRM_ERROR("can not ioremap virtual address for"
1469			  " ring buffer\n");
1470		return -ENOMEM;
1471	}
1472
1473	return 0;
1474}
1475
1476int intel_init_bsd_ring_buffer(struct drm_device *dev)
1477{
1478	drm_i915_private_t *dev_priv = dev->dev_private;
1479	struct intel_ring_buffer *ring = &dev_priv->rings[VCS];
1480
1481	ring->name = "bsd ring";
1482	ring->id = VCS;
1483
1484	ring->write_tail = ring_write_tail;
1485	if (IS_GEN6(dev) || IS_GEN7(dev)) {
1486		ring->mmio_base = GEN6_BSD_RING_BASE;
1487		/* gen6 bsd needs a special wa for tail updates */
1488		if (IS_GEN6(dev))
1489			ring->write_tail = gen6_bsd_ring_write_tail;
1490		ring->flush = gen6_ring_flush;
1491		ring->add_request = gen6_add_request;
1492		ring->get_seqno = gen6_ring_get_seqno;
1493		ring->irq_enable_mask = GEN6_BSD_USER_INTERRUPT;
1494		ring->irq_get = gen6_ring_get_irq;
1495		ring->irq_put = gen6_ring_put_irq;
1496		ring->dispatch_execbuffer = gen6_ring_dispatch_execbuffer;
1497		ring->sync_to = gen6_ring_sync;
1498		ring->semaphore_register[0] = MI_SEMAPHORE_SYNC_VR;
1499		ring->semaphore_register[1] = MI_SEMAPHORE_SYNC_INVALID;
1500		ring->semaphore_register[2] = MI_SEMAPHORE_SYNC_VB;
1501		ring->signal_mbox[0] = GEN6_RVSYNC;
1502		ring->signal_mbox[1] = GEN6_BVSYNC;
1503	} else {
1504		ring->mmio_base = BSD_RING_BASE;
1505		ring->flush = bsd_ring_flush;
1506		ring->add_request = i9xx_add_request;
1507		ring->get_seqno = ring_get_seqno;
1508		if (IS_GEN5(dev)) {
1509			ring->irq_enable_mask = GT_BSD_USER_INTERRUPT;
1510			ring->irq_get = gen5_ring_get_irq;
1511			ring->irq_put = gen5_ring_put_irq;
1512		} else {
1513			ring->irq_enable_mask = I915_BSD_USER_INTERRUPT;
1514			ring->irq_get = i9xx_ring_get_irq;
1515			ring->irq_put = i9xx_ring_put_irq;
1516		}
1517		ring->dispatch_execbuffer = i965_dispatch_execbuffer;
1518	}
1519	ring->init = init_ring_common;
1520
1521
1522	return intel_init_ring_buffer(dev, ring);
1523}
1524
1525int intel_init_blt_ring_buffer(struct drm_device *dev)
1526{
1527	drm_i915_private_t *dev_priv = dev->dev_private;
1528	struct intel_ring_buffer *ring = &dev_priv->rings[BCS];
1529
1530	ring->name = "blitter ring";
1531	ring->id = BCS;
1532
1533	ring->mmio_base = BLT_RING_BASE;
1534	ring->write_tail = ring_write_tail;
1535	ring->flush = blt_ring_flush;
1536	ring->add_request = gen6_add_request;
1537	ring->get_seqno = gen6_ring_get_seqno;
1538	ring->irq_enable_mask = GEN6_BLITTER_USER_INTERRUPT;
1539	ring->irq_get = gen6_ring_get_irq;
1540	ring->irq_put = gen6_ring_put_irq;
1541	ring->dispatch_execbuffer = gen6_ring_dispatch_execbuffer;
1542	ring->sync_to = gen6_ring_sync;
1543	ring->semaphore_register[0] = MI_SEMAPHORE_SYNC_BR;
1544	ring->semaphore_register[1] = MI_SEMAPHORE_SYNC_BV;
1545	ring->semaphore_register[2] = MI_SEMAPHORE_SYNC_INVALID;
1546	ring->signal_mbox[0] = GEN6_RBSYNC;
1547	ring->signal_mbox[1] = GEN6_VBSYNC;
1548	ring->init = init_ring_common;
1549
1550	return intel_init_ring_buffer(dev, ring);
1551}
1552