1/*
2 * Copyright 2016 Advanced Micro Devices, Inc.
3 *
4 * Permission is hereby granted, free of charge, to any person obtaining a
5 * copy of this software and associated documentation files (the "Software"),
6 * to deal in the Software without restriction, including without limitation
7 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
8 * and/or sell copies of the Software, and to permit persons to whom the
9 * Software is furnished to do so, subject to the following conditions:
10 *
11 * The above copyright notice and this permission notice shall be included in
12 * all copies or substantial portions of the Software.
13 *
14 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
15 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
16 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
17 * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
18 * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
19 * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
20 * OTHER DEALINGS IN THE SOFTWARE.
21 *
22 * Authors: AMD
23 *
24 */
25
26
27#include "dcn30/dcn30_hubbub.h"
28#include "dcn31_hubbub.h"
29#include "dm_services.h"
30#include "reg_helper.h"
31
32
33#define CTX \
34	hubbub2->base.ctx
35#define DC_LOGGER \
36	hubbub2->base.ctx->logger
37#define REG(reg)\
38	hubbub2->regs->reg
39
40#undef FN
41#define FN(reg_name, field_name) \
42	hubbub2->shifts->field_name, hubbub2->masks->field_name
43
44#ifdef NUM_VMID
45#undef NUM_VMID
46#endif
47#define NUM_VMID 16
48
49#define DCN31_CRB_SEGMENT_SIZE_KB 64
50
51static void dcn31_init_crb(struct hubbub *hubbub)
52{
53	struct dcn20_hubbub *hubbub2 = TO_DCN20_HUBBUB(hubbub);
54
55	REG_GET(DCHUBBUB_DET0_CTRL, DET0_SIZE_CURRENT,
56		&hubbub2->det0_size);
57
58	REG_GET(DCHUBBUB_DET1_CTRL, DET1_SIZE_CURRENT,
59		&hubbub2->det1_size);
60
61	REG_GET(DCHUBBUB_DET2_CTRL, DET2_SIZE_CURRENT,
62		&hubbub2->det2_size);
63
64	REG_GET(DCHUBBUB_DET3_CTRL, DET3_SIZE_CURRENT,
65		&hubbub2->det3_size);
66
67	REG_GET(DCHUBBUB_COMPBUF_CTRL, COMPBUF_SIZE_CURRENT,
68		&hubbub2->compbuf_size_segments);
69
70	REG_SET_2(COMPBUF_RESERVED_SPACE, 0,
71			COMPBUF_RESERVED_SPACE_64B, hubbub2->pixel_chunk_size / 32,
72			COMPBUF_RESERVED_SPACE_ZS, hubbub2->pixel_chunk_size / 128);
73	REG_UPDATE(DCHUBBUB_DEBUG_CTRL_0, DET_DEPTH, 0x17F);
74}
75
76static void dcn31_program_det_size(struct hubbub *hubbub, int hubp_inst, unsigned int det_buffer_size_in_kbyte)
77{
78	struct dcn20_hubbub *hubbub2 = TO_DCN20_HUBBUB(hubbub);
79
80	unsigned int det_size_segments = (det_buffer_size_in_kbyte + DCN31_CRB_SEGMENT_SIZE_KB - 1) / DCN31_CRB_SEGMENT_SIZE_KB;
81
82	switch (hubp_inst) {
83	case 0:
84		REG_UPDATE(DCHUBBUB_DET0_CTRL,
85					DET0_SIZE, det_size_segments);
86		hubbub2->det0_size = det_size_segments;
87		break;
88	case 1:
89		REG_UPDATE(DCHUBBUB_DET1_CTRL,
90					DET1_SIZE, det_size_segments);
91		hubbub2->det1_size = det_size_segments;
92		break;
93	case 2:
94		REG_UPDATE(DCHUBBUB_DET2_CTRL,
95					DET2_SIZE, det_size_segments);
96		hubbub2->det2_size = det_size_segments;
97		break;
98	case 3:
99		REG_UPDATE(DCHUBBUB_DET3_CTRL,
100					DET3_SIZE, det_size_segments);
101		hubbub2->det3_size = det_size_segments;
102		break;
103	default:
104		break;
105	}
106	DC_LOG_DEBUG("Set DET%d to %d segments\n", hubp_inst, det_size_segments);
107	/* Should never be hit, if it is we have an erroneous hw config*/
108	ASSERT(hubbub2->det0_size + hubbub2->det1_size + hubbub2->det2_size
109			+ hubbub2->det3_size + hubbub2->compbuf_size_segments <= hubbub2->crb_size_segs);
110}
111
112static void dcn31_wait_for_det_apply(struct hubbub *hubbub, int hubp_inst)
113{
114	struct dcn20_hubbub *hubbub2 = TO_DCN20_HUBBUB(hubbub);
115
116	switch (hubp_inst) {
117	case 0:
118		REG_WAIT(DCHUBBUB_DET0_CTRL, DET0_SIZE_CURRENT, hubbub2->det0_size, 1000, 30);
119		break;
120	case 1:
121		REG_WAIT(DCHUBBUB_DET1_CTRL, DET1_SIZE_CURRENT, hubbub2->det1_size, 1000, 30);
122		break;
123	case 2:
124		REG_WAIT(DCHUBBUB_DET2_CTRL, DET2_SIZE_CURRENT, hubbub2->det2_size, 1000, 30);
125		break;
126	case 3:
127		REG_WAIT(DCHUBBUB_DET3_CTRL, DET3_SIZE_CURRENT, hubbub2->det3_size, 1000, 30);
128		break;
129	default:
130		break;
131	}
132}
133
134static void dcn31_program_compbuf_size(struct hubbub *hubbub, unsigned int compbuf_size_kb, bool safe_to_increase)
135{
136	struct dcn20_hubbub *hubbub2 = TO_DCN20_HUBBUB(hubbub);
137	unsigned int compbuf_size_segments = (compbuf_size_kb + DCN31_CRB_SEGMENT_SIZE_KB - 1) / DCN31_CRB_SEGMENT_SIZE_KB;
138
139	if (safe_to_increase || compbuf_size_segments <= hubbub2->compbuf_size_segments) {
140		if (compbuf_size_segments > hubbub2->compbuf_size_segments) {
141			REG_WAIT(DCHUBBUB_DET0_CTRL, DET0_SIZE_CURRENT, hubbub2->det0_size, 1, 100);
142			REG_WAIT(DCHUBBUB_DET1_CTRL, DET1_SIZE_CURRENT, hubbub2->det1_size, 1, 100);
143			REG_WAIT(DCHUBBUB_DET2_CTRL, DET2_SIZE_CURRENT, hubbub2->det2_size, 1, 100);
144			REG_WAIT(DCHUBBUB_DET3_CTRL, DET3_SIZE_CURRENT, hubbub2->det3_size, 1, 100);
145		}
146		/* Should never be hit, if it is we have an erroneous hw config*/
147		ASSERT(hubbub2->det0_size + hubbub2->det1_size + hubbub2->det2_size
148				+ hubbub2->det3_size + compbuf_size_segments <= hubbub2->crb_size_segs);
149		REG_UPDATE(DCHUBBUB_COMPBUF_CTRL, COMPBUF_SIZE, compbuf_size_segments);
150		hubbub2->compbuf_size_segments = compbuf_size_segments;
151		ASSERT(REG_GET(DCHUBBUB_COMPBUF_CTRL, CONFIG_ERROR, &compbuf_size_segments) && !compbuf_size_segments);
152	}
153}
154
155static uint32_t convert_and_clamp(
156	uint32_t wm_ns,
157	uint32_t refclk_mhz,
158	uint32_t clamp_value)
159{
160	uint32_t ret_val = 0;
161	ret_val = wm_ns * refclk_mhz;
162	ret_val /= 1000;
163
164	if (ret_val > clamp_value) {
165		/* clamping WMs is abnormal, unexpected and may lead to underflow*/
166		ASSERT(0);
167		ret_val = clamp_value;
168	}
169
170	return ret_val;
171}
172
173static bool hubbub31_program_urgent_watermarks(
174		struct hubbub *hubbub,
175		struct dcn_watermark_set *watermarks,
176		unsigned int refclk_mhz,
177		bool safe_to_lower)
178{
179	struct dcn20_hubbub *hubbub2 = TO_DCN20_HUBBUB(hubbub);
180	uint32_t prog_wm_value;
181	bool wm_pending = false;
182
183	/* Repeat for water mark set A, B, C and D. */
184	/* clock state A */
185	if (safe_to_lower || watermarks->a.urgent_ns > hubbub2->watermarks.a.urgent_ns) {
186		hubbub2->watermarks.a.urgent_ns = watermarks->a.urgent_ns;
187		prog_wm_value = convert_and_clamp(watermarks->a.urgent_ns,
188				refclk_mhz, 0x3fff);
189		REG_SET(DCHUBBUB_ARB_DATA_URGENCY_WATERMARK_A, 0,
190				DCHUBBUB_ARB_DATA_URGENCY_WATERMARK_A, prog_wm_value);
191
192		DC_LOG_BANDWIDTH_CALCS("URGENCY_WATERMARK_A calculated =%d\n"
193			"HW register value = 0x%x\n",
194			watermarks->a.urgent_ns, prog_wm_value);
195	} else if (watermarks->a.urgent_ns < hubbub2->watermarks.a.urgent_ns)
196		wm_pending = true;
197
198	/* determine the transfer time for a quantity of data for a particular requestor.*/
199	if (safe_to_lower || watermarks->a.frac_urg_bw_flip
200			> hubbub2->watermarks.a.frac_urg_bw_flip) {
201		hubbub2->watermarks.a.frac_urg_bw_flip = watermarks->a.frac_urg_bw_flip;
202
203		REG_SET(DCHUBBUB_ARB_FRAC_URG_BW_FLIP_A, 0,
204				DCHUBBUB_ARB_FRAC_URG_BW_FLIP_A, watermarks->a.frac_urg_bw_flip);
205	} else if (watermarks->a.frac_urg_bw_flip
206			< hubbub2->watermarks.a.frac_urg_bw_flip)
207		wm_pending = true;
208
209	if (safe_to_lower || watermarks->a.frac_urg_bw_nom
210			> hubbub2->watermarks.a.frac_urg_bw_nom) {
211		hubbub2->watermarks.a.frac_urg_bw_nom = watermarks->a.frac_urg_bw_nom;
212
213		REG_SET(DCHUBBUB_ARB_FRAC_URG_BW_NOM_A, 0,
214				DCHUBBUB_ARB_FRAC_URG_BW_NOM_A, watermarks->a.frac_urg_bw_nom);
215	} else if (watermarks->a.frac_urg_bw_nom
216			< hubbub2->watermarks.a.frac_urg_bw_nom)
217		wm_pending = true;
218
219	if (safe_to_lower || watermarks->a.urgent_latency_ns > hubbub2->watermarks.a.urgent_latency_ns) {
220		hubbub2->watermarks.a.urgent_latency_ns = watermarks->a.urgent_latency_ns;
221		prog_wm_value = convert_and_clamp(watermarks->a.urgent_latency_ns,
222				refclk_mhz, 0x3fff);
223		REG_SET(DCHUBBUB_ARB_REFCYC_PER_TRIP_TO_MEMORY_A, 0,
224				DCHUBBUB_ARB_REFCYC_PER_TRIP_TO_MEMORY_A, prog_wm_value);
225	} else if (watermarks->a.urgent_latency_ns < hubbub2->watermarks.a.urgent_latency_ns)
226		wm_pending = true;
227
228	/* clock state B */
229	if (safe_to_lower || watermarks->b.urgent_ns > hubbub2->watermarks.b.urgent_ns) {
230		hubbub2->watermarks.b.urgent_ns = watermarks->b.urgent_ns;
231		prog_wm_value = convert_and_clamp(watermarks->b.urgent_ns,
232				refclk_mhz, 0x3fff);
233		REG_SET(DCHUBBUB_ARB_DATA_URGENCY_WATERMARK_B, 0,
234				DCHUBBUB_ARB_DATA_URGENCY_WATERMARK_B, prog_wm_value);
235
236		DC_LOG_BANDWIDTH_CALCS("URGENCY_WATERMARK_B calculated =%d\n"
237			"HW register value = 0x%x\n",
238			watermarks->b.urgent_ns, prog_wm_value);
239	} else if (watermarks->b.urgent_ns < hubbub2->watermarks.b.urgent_ns)
240		wm_pending = true;
241
242	/* determine the transfer time for a quantity of data for a particular requestor.*/
243	if (safe_to_lower || watermarks->b.frac_urg_bw_flip
244			> hubbub2->watermarks.b.frac_urg_bw_flip) {
245		hubbub2->watermarks.b.frac_urg_bw_flip = watermarks->b.frac_urg_bw_flip;
246
247		REG_SET(DCHUBBUB_ARB_FRAC_URG_BW_FLIP_B, 0,
248				DCHUBBUB_ARB_FRAC_URG_BW_FLIP_B, watermarks->b.frac_urg_bw_flip);
249	} else if (watermarks->b.frac_urg_bw_flip
250			< hubbub2->watermarks.b.frac_urg_bw_flip)
251		wm_pending = true;
252
253	if (safe_to_lower || watermarks->b.frac_urg_bw_nom
254			> hubbub2->watermarks.b.frac_urg_bw_nom) {
255		hubbub2->watermarks.b.frac_urg_bw_nom = watermarks->b.frac_urg_bw_nom;
256
257		REG_SET(DCHUBBUB_ARB_FRAC_URG_BW_NOM_B, 0,
258				DCHUBBUB_ARB_FRAC_URG_BW_NOM_B, watermarks->b.frac_urg_bw_nom);
259	} else if (watermarks->b.frac_urg_bw_nom
260			< hubbub2->watermarks.b.frac_urg_bw_nom)
261		wm_pending = true;
262
263	if (safe_to_lower || watermarks->b.urgent_latency_ns > hubbub2->watermarks.b.urgent_latency_ns) {
264		hubbub2->watermarks.b.urgent_latency_ns = watermarks->b.urgent_latency_ns;
265		prog_wm_value = convert_and_clamp(watermarks->b.urgent_latency_ns,
266				refclk_mhz, 0x3fff);
267		REG_SET(DCHUBBUB_ARB_REFCYC_PER_TRIP_TO_MEMORY_B, 0,
268				DCHUBBUB_ARB_REFCYC_PER_TRIP_TO_MEMORY_B, prog_wm_value);
269	} else if (watermarks->b.urgent_latency_ns < hubbub2->watermarks.b.urgent_latency_ns)
270		wm_pending = true;
271
272	/* clock state C */
273	if (safe_to_lower || watermarks->c.urgent_ns > hubbub2->watermarks.c.urgent_ns) {
274		hubbub2->watermarks.c.urgent_ns = watermarks->c.urgent_ns;
275		prog_wm_value = convert_and_clamp(watermarks->c.urgent_ns,
276				refclk_mhz, 0x3fff);
277		REG_SET(DCHUBBUB_ARB_DATA_URGENCY_WATERMARK_C, 0,
278				DCHUBBUB_ARB_DATA_URGENCY_WATERMARK_C, prog_wm_value);
279
280		DC_LOG_BANDWIDTH_CALCS("URGENCY_WATERMARK_C calculated =%d\n"
281			"HW register value = 0x%x\n",
282			watermarks->c.urgent_ns, prog_wm_value);
283	} else if (watermarks->c.urgent_ns < hubbub2->watermarks.c.urgent_ns)
284		wm_pending = true;
285
286	/* determine the transfer time for a quantity of data for a particular requestor.*/
287	if (safe_to_lower || watermarks->c.frac_urg_bw_flip
288			> hubbub2->watermarks.c.frac_urg_bw_flip) {
289		hubbub2->watermarks.c.frac_urg_bw_flip = watermarks->c.frac_urg_bw_flip;
290
291		REG_SET(DCHUBBUB_ARB_FRAC_URG_BW_FLIP_C, 0,
292				DCHUBBUB_ARB_FRAC_URG_BW_FLIP_C, watermarks->c.frac_urg_bw_flip);
293	} else if (watermarks->c.frac_urg_bw_flip
294			< hubbub2->watermarks.c.frac_urg_bw_flip)
295		wm_pending = true;
296
297	if (safe_to_lower || watermarks->c.frac_urg_bw_nom
298			> hubbub2->watermarks.c.frac_urg_bw_nom) {
299		hubbub2->watermarks.c.frac_urg_bw_nom = watermarks->c.frac_urg_bw_nom;
300
301		REG_SET(DCHUBBUB_ARB_FRAC_URG_BW_NOM_C, 0,
302				DCHUBBUB_ARB_FRAC_URG_BW_NOM_C, watermarks->c.frac_urg_bw_nom);
303	} else if (watermarks->c.frac_urg_bw_nom
304			< hubbub2->watermarks.c.frac_urg_bw_nom)
305		wm_pending = true;
306
307	if (safe_to_lower || watermarks->c.urgent_latency_ns > hubbub2->watermarks.c.urgent_latency_ns) {
308		hubbub2->watermarks.c.urgent_latency_ns = watermarks->c.urgent_latency_ns;
309		prog_wm_value = convert_and_clamp(watermarks->c.urgent_latency_ns,
310				refclk_mhz, 0x3fff);
311		REG_SET(DCHUBBUB_ARB_REFCYC_PER_TRIP_TO_MEMORY_C, 0,
312				DCHUBBUB_ARB_REFCYC_PER_TRIP_TO_MEMORY_C, prog_wm_value);
313	} else if (watermarks->c.urgent_latency_ns < hubbub2->watermarks.c.urgent_latency_ns)
314		wm_pending = true;
315
316	/* clock state D */
317	if (safe_to_lower || watermarks->d.urgent_ns > hubbub2->watermarks.d.urgent_ns) {
318		hubbub2->watermarks.d.urgent_ns = watermarks->d.urgent_ns;
319		prog_wm_value = convert_and_clamp(watermarks->d.urgent_ns,
320				refclk_mhz, 0x3fff);
321		REG_SET(DCHUBBUB_ARB_DATA_URGENCY_WATERMARK_D, 0,
322				DCHUBBUB_ARB_DATA_URGENCY_WATERMARK_D, prog_wm_value);
323
324		DC_LOG_BANDWIDTH_CALCS("URGENCY_WATERMARK_D calculated =%d\n"
325			"HW register value = 0x%x\n",
326			watermarks->d.urgent_ns, prog_wm_value);
327	} else if (watermarks->d.urgent_ns < hubbub2->watermarks.d.urgent_ns)
328		wm_pending = true;
329
330	/* determine the transfer time for a quantity of data for a particular requestor.*/
331	if (safe_to_lower || watermarks->d.frac_urg_bw_flip
332			> hubbub2->watermarks.d.frac_urg_bw_flip) {
333		hubbub2->watermarks.d.frac_urg_bw_flip = watermarks->d.frac_urg_bw_flip;
334
335		REG_SET(DCHUBBUB_ARB_FRAC_URG_BW_FLIP_D, 0,
336				DCHUBBUB_ARB_FRAC_URG_BW_FLIP_D, watermarks->d.frac_urg_bw_flip);
337	} else if (watermarks->d.frac_urg_bw_flip
338			< hubbub2->watermarks.d.frac_urg_bw_flip)
339		wm_pending = true;
340
341	if (safe_to_lower || watermarks->d.frac_urg_bw_nom
342			> hubbub2->watermarks.d.frac_urg_bw_nom) {
343		hubbub2->watermarks.d.frac_urg_bw_nom = watermarks->d.frac_urg_bw_nom;
344
345		REG_SET(DCHUBBUB_ARB_FRAC_URG_BW_NOM_D, 0,
346				DCHUBBUB_ARB_FRAC_URG_BW_NOM_D, watermarks->d.frac_urg_bw_nom);
347	} else if (watermarks->d.frac_urg_bw_nom
348			< hubbub2->watermarks.d.frac_urg_bw_nom)
349		wm_pending = true;
350
351	if (safe_to_lower || watermarks->d.urgent_latency_ns > hubbub2->watermarks.d.urgent_latency_ns) {
352		hubbub2->watermarks.d.urgent_latency_ns = watermarks->d.urgent_latency_ns;
353		prog_wm_value = convert_and_clamp(watermarks->d.urgent_latency_ns,
354				refclk_mhz, 0x3fff);
355		REG_SET(DCHUBBUB_ARB_REFCYC_PER_TRIP_TO_MEMORY_D, 0,
356				DCHUBBUB_ARB_REFCYC_PER_TRIP_TO_MEMORY_D, prog_wm_value);
357	} else if (watermarks->d.urgent_latency_ns < hubbub2->watermarks.d.urgent_latency_ns)
358		wm_pending = true;
359
360	return wm_pending;
361}
362
363static bool hubbub31_program_stutter_watermarks(
364		struct hubbub *hubbub,
365		struct dcn_watermark_set *watermarks,
366		unsigned int refclk_mhz,
367		bool safe_to_lower)
368{
369	struct dcn20_hubbub *hubbub2 = TO_DCN20_HUBBUB(hubbub);
370	uint32_t prog_wm_value;
371	bool wm_pending = false;
372
373	/* clock state A */
374	if (safe_to_lower || watermarks->a.cstate_pstate.cstate_enter_plus_exit_ns
375			> hubbub2->watermarks.a.cstate_pstate.cstate_enter_plus_exit_ns) {
376		hubbub2->watermarks.a.cstate_pstate.cstate_enter_plus_exit_ns =
377				watermarks->a.cstate_pstate.cstate_enter_plus_exit_ns;
378		prog_wm_value = convert_and_clamp(
379				watermarks->a.cstate_pstate.cstate_enter_plus_exit_ns,
380				refclk_mhz, 0xfffff);
381		REG_SET(DCHUBBUB_ARB_ALLOW_SR_ENTER_WATERMARK_A, 0,
382				DCHUBBUB_ARB_ALLOW_SR_ENTER_WATERMARK_A, prog_wm_value);
383		DC_LOG_BANDWIDTH_CALCS("SR_ENTER_EXIT_WATERMARK_A calculated =%d\n"
384			"HW register value = 0x%x\n",
385			watermarks->a.cstate_pstate.cstate_enter_plus_exit_ns, prog_wm_value);
386	} else if (watermarks->a.cstate_pstate.cstate_enter_plus_exit_ns
387			< hubbub2->watermarks.a.cstate_pstate.cstate_enter_plus_exit_ns)
388		wm_pending = true;
389
390	if (safe_to_lower || watermarks->a.cstate_pstate.cstate_exit_ns
391			> hubbub2->watermarks.a.cstate_pstate.cstate_exit_ns) {
392		hubbub2->watermarks.a.cstate_pstate.cstate_exit_ns =
393				watermarks->a.cstate_pstate.cstate_exit_ns;
394		prog_wm_value = convert_and_clamp(
395				watermarks->a.cstate_pstate.cstate_exit_ns,
396				refclk_mhz, 0xfffff);
397		REG_SET(DCHUBBUB_ARB_ALLOW_SR_EXIT_WATERMARK_A, 0,
398				DCHUBBUB_ARB_ALLOW_SR_EXIT_WATERMARK_A, prog_wm_value);
399		DC_LOG_BANDWIDTH_CALCS("SR_EXIT_WATERMARK_A calculated =%d\n"
400			"HW register value = 0x%x\n",
401			watermarks->a.cstate_pstate.cstate_exit_ns, prog_wm_value);
402	} else if (watermarks->a.cstate_pstate.cstate_exit_ns
403			< hubbub2->watermarks.a.cstate_pstate.cstate_exit_ns)
404		wm_pending = true;
405
406	if (safe_to_lower || watermarks->a.cstate_pstate.cstate_enter_plus_exit_z8_ns
407			> hubbub2->watermarks.a.cstate_pstate.cstate_enter_plus_exit_z8_ns) {
408		hubbub2->watermarks.a.cstate_pstate.cstate_enter_plus_exit_z8_ns =
409				watermarks->a.cstate_pstate.cstate_enter_plus_exit_z8_ns;
410		prog_wm_value = convert_and_clamp(
411				watermarks->a.cstate_pstate.cstate_enter_plus_exit_z8_ns,
412				refclk_mhz, 0xfffff);
413		REG_SET(DCHUBBUB_ARB_ALLOW_SR_ENTER_WATERMARK_Z8_A, 0,
414				DCHUBBUB_ARB_ALLOW_SR_ENTER_WATERMARK_Z8_A, prog_wm_value);
415		DC_LOG_BANDWIDTH_CALCS("SR_ENTER_WATERMARK_Z8_A calculated =%d\n"
416			"HW register value = 0x%x\n",
417			watermarks->a.cstate_pstate.cstate_enter_plus_exit_z8_ns, prog_wm_value);
418	} else if (watermarks->a.cstate_pstate.cstate_enter_plus_exit_z8_ns
419			< hubbub2->watermarks.a.cstate_pstate.cstate_enter_plus_exit_z8_ns)
420		wm_pending = true;
421
422	if (safe_to_lower || watermarks->a.cstate_pstate.cstate_exit_z8_ns
423			> hubbub2->watermarks.a.cstate_pstate.cstate_exit_z8_ns) {
424		hubbub2->watermarks.a.cstate_pstate.cstate_exit_z8_ns =
425				watermarks->a.cstate_pstate.cstate_exit_z8_ns;
426		prog_wm_value = convert_and_clamp(
427				watermarks->a.cstate_pstate.cstate_exit_z8_ns,
428				refclk_mhz, 0xfffff);
429		REG_SET(DCHUBBUB_ARB_ALLOW_SR_EXIT_WATERMARK_Z8_A, 0,
430				DCHUBBUB_ARB_ALLOW_SR_EXIT_WATERMARK_Z8_A, prog_wm_value);
431		DC_LOG_BANDWIDTH_CALCS("SR_EXIT_WATERMARK_Z8_A calculated =%d\n"
432			"HW register value = 0x%x\n",
433			watermarks->a.cstate_pstate.cstate_exit_z8_ns, prog_wm_value);
434	} else if (watermarks->a.cstate_pstate.cstate_exit_z8_ns
435			< hubbub2->watermarks.a.cstate_pstate.cstate_exit_z8_ns)
436		wm_pending = true;
437
438	/* clock state B */
439	if (safe_to_lower || watermarks->b.cstate_pstate.cstate_enter_plus_exit_ns
440			> hubbub2->watermarks.b.cstate_pstate.cstate_enter_plus_exit_ns) {
441		hubbub2->watermarks.b.cstate_pstate.cstate_enter_plus_exit_ns =
442				watermarks->b.cstate_pstate.cstate_enter_plus_exit_ns;
443		prog_wm_value = convert_and_clamp(
444				watermarks->b.cstate_pstate.cstate_enter_plus_exit_ns,
445				refclk_mhz, 0xfffff);
446		REG_SET(DCHUBBUB_ARB_ALLOW_SR_ENTER_WATERMARK_B, 0,
447				DCHUBBUB_ARB_ALLOW_SR_ENTER_WATERMARK_B, prog_wm_value);
448		DC_LOG_BANDWIDTH_CALCS("SR_ENTER_EXIT_WATERMARK_B calculated =%d\n"
449			"HW register value = 0x%x\n",
450			watermarks->b.cstate_pstate.cstate_enter_plus_exit_ns, prog_wm_value);
451	} else if (watermarks->b.cstate_pstate.cstate_enter_plus_exit_ns
452			< hubbub2->watermarks.b.cstate_pstate.cstate_enter_plus_exit_ns)
453		wm_pending = true;
454
455	if (safe_to_lower || watermarks->b.cstate_pstate.cstate_exit_ns
456			> hubbub2->watermarks.b.cstate_pstate.cstate_exit_ns) {
457		hubbub2->watermarks.b.cstate_pstate.cstate_exit_ns =
458				watermarks->b.cstate_pstate.cstate_exit_ns;
459		prog_wm_value = convert_and_clamp(
460				watermarks->b.cstate_pstate.cstate_exit_ns,
461				refclk_mhz, 0xfffff);
462		REG_SET(DCHUBBUB_ARB_ALLOW_SR_EXIT_WATERMARK_B, 0,
463				DCHUBBUB_ARB_ALLOW_SR_EXIT_WATERMARK_B, prog_wm_value);
464		DC_LOG_BANDWIDTH_CALCS("SR_EXIT_WATERMARK_B calculated =%d\n"
465			"HW register value = 0x%x\n",
466			watermarks->b.cstate_pstate.cstate_exit_ns, prog_wm_value);
467	} else if (watermarks->b.cstate_pstate.cstate_exit_ns
468			< hubbub2->watermarks.b.cstate_pstate.cstate_exit_ns)
469		wm_pending = true;
470
471	if (safe_to_lower || watermarks->b.cstate_pstate.cstate_enter_plus_exit_z8_ns
472			> hubbub2->watermarks.b.cstate_pstate.cstate_enter_plus_exit_z8_ns) {
473		hubbub2->watermarks.b.cstate_pstate.cstate_enter_plus_exit_z8_ns =
474				watermarks->b.cstate_pstate.cstate_enter_plus_exit_z8_ns;
475		prog_wm_value = convert_and_clamp(
476				watermarks->b.cstate_pstate.cstate_enter_plus_exit_z8_ns,
477				refclk_mhz, 0xfffff);
478		REG_SET(DCHUBBUB_ARB_ALLOW_SR_ENTER_WATERMARK_Z8_B, 0,
479				DCHUBBUB_ARB_ALLOW_SR_ENTER_WATERMARK_Z8_B, prog_wm_value);
480		DC_LOG_BANDWIDTH_CALCS("SR_ENTER_WATERMARK_Z8_B calculated =%d\n"
481			"HW register value = 0x%x\n",
482			watermarks->b.cstate_pstate.cstate_enter_plus_exit_z8_ns, prog_wm_value);
483	} else if (watermarks->b.cstate_pstate.cstate_enter_plus_exit_z8_ns
484			< hubbub2->watermarks.b.cstate_pstate.cstate_enter_plus_exit_z8_ns)
485		wm_pending = true;
486
487	if (safe_to_lower || watermarks->b.cstate_pstate.cstate_exit_z8_ns
488			> hubbub2->watermarks.b.cstate_pstate.cstate_exit_z8_ns) {
489		hubbub2->watermarks.b.cstate_pstate.cstate_exit_z8_ns =
490				watermarks->b.cstate_pstate.cstate_exit_z8_ns;
491		prog_wm_value = convert_and_clamp(
492				watermarks->b.cstate_pstate.cstate_exit_z8_ns,
493				refclk_mhz, 0xfffff);
494		REG_SET(DCHUBBUB_ARB_ALLOW_SR_EXIT_WATERMARK_Z8_B, 0,
495				DCHUBBUB_ARB_ALLOW_SR_EXIT_WATERMARK_Z8_B, prog_wm_value);
496		DC_LOG_BANDWIDTH_CALCS("SR_EXIT_WATERMARK_Z8_B calculated =%d\n"
497			"HW register value = 0x%x\n",
498			watermarks->b.cstate_pstate.cstate_exit_z8_ns, prog_wm_value);
499	} else if (watermarks->b.cstate_pstate.cstate_exit_z8_ns
500			< hubbub2->watermarks.b.cstate_pstate.cstate_exit_z8_ns)
501		wm_pending = true;
502
503	/* clock state C */
504	if (safe_to_lower || watermarks->c.cstate_pstate.cstate_enter_plus_exit_ns
505			> hubbub2->watermarks.c.cstate_pstate.cstate_enter_plus_exit_ns) {
506		hubbub2->watermarks.c.cstate_pstate.cstate_enter_plus_exit_ns =
507				watermarks->c.cstate_pstate.cstate_enter_plus_exit_ns;
508		prog_wm_value = convert_and_clamp(
509				watermarks->c.cstate_pstate.cstate_enter_plus_exit_ns,
510				refclk_mhz, 0xfffff);
511		REG_SET(DCHUBBUB_ARB_ALLOW_SR_ENTER_WATERMARK_C, 0,
512				DCHUBBUB_ARB_ALLOW_SR_ENTER_WATERMARK_C, prog_wm_value);
513		DC_LOG_BANDWIDTH_CALCS("SR_ENTER_EXIT_WATERMARK_C calculated =%d\n"
514			"HW register value = 0x%x\n",
515			watermarks->c.cstate_pstate.cstate_enter_plus_exit_ns, prog_wm_value);
516	} else if (watermarks->c.cstate_pstate.cstate_enter_plus_exit_ns
517			< hubbub2->watermarks.c.cstate_pstate.cstate_enter_plus_exit_ns)
518		wm_pending = true;
519
520	if (safe_to_lower || watermarks->c.cstate_pstate.cstate_exit_ns
521			> hubbub2->watermarks.c.cstate_pstate.cstate_exit_ns) {
522		hubbub2->watermarks.c.cstate_pstate.cstate_exit_ns =
523				watermarks->c.cstate_pstate.cstate_exit_ns;
524		prog_wm_value = convert_and_clamp(
525				watermarks->c.cstate_pstate.cstate_exit_ns,
526				refclk_mhz, 0xfffff);
527		REG_SET(DCHUBBUB_ARB_ALLOW_SR_EXIT_WATERMARK_C, 0,
528				DCHUBBUB_ARB_ALLOW_SR_EXIT_WATERMARK_C, prog_wm_value);
529		DC_LOG_BANDWIDTH_CALCS("SR_EXIT_WATERMARK_C calculated =%d\n"
530			"HW register value = 0x%x\n",
531			watermarks->c.cstate_pstate.cstate_exit_ns, prog_wm_value);
532	} else if (watermarks->c.cstate_pstate.cstate_exit_ns
533			< hubbub2->watermarks.c.cstate_pstate.cstate_exit_ns)
534		wm_pending = true;
535
536	if (safe_to_lower || watermarks->c.cstate_pstate.cstate_enter_plus_exit_z8_ns
537			> hubbub2->watermarks.c.cstate_pstate.cstate_enter_plus_exit_z8_ns) {
538		hubbub2->watermarks.c.cstate_pstate.cstate_enter_plus_exit_z8_ns =
539				watermarks->c.cstate_pstate.cstate_enter_plus_exit_z8_ns;
540		prog_wm_value = convert_and_clamp(
541				watermarks->c.cstate_pstate.cstate_enter_plus_exit_z8_ns,
542				refclk_mhz, 0xfffff);
543		REG_SET(DCHUBBUB_ARB_ALLOW_SR_ENTER_WATERMARK_Z8_C, 0,
544				DCHUBBUB_ARB_ALLOW_SR_ENTER_WATERMARK_Z8_C, prog_wm_value);
545		DC_LOG_BANDWIDTH_CALCS("SR_ENTER_WATERMARK_Z8_C calculated =%d\n"
546			"HW register value = 0x%x\n",
547			watermarks->c.cstate_pstate.cstate_enter_plus_exit_z8_ns, prog_wm_value);
548	} else if (watermarks->c.cstate_pstate.cstate_enter_plus_exit_z8_ns
549			< hubbub2->watermarks.c.cstate_pstate.cstate_enter_plus_exit_z8_ns)
550		wm_pending = true;
551
552	if (safe_to_lower || watermarks->c.cstate_pstate.cstate_exit_z8_ns
553			> hubbub2->watermarks.c.cstate_pstate.cstate_exit_z8_ns) {
554		hubbub2->watermarks.c.cstate_pstate.cstate_exit_z8_ns =
555				watermarks->c.cstate_pstate.cstate_exit_z8_ns;
556		prog_wm_value = convert_and_clamp(
557				watermarks->c.cstate_pstate.cstate_exit_z8_ns,
558				refclk_mhz, 0xfffff);
559		REG_SET(DCHUBBUB_ARB_ALLOW_SR_EXIT_WATERMARK_Z8_C, 0,
560				DCHUBBUB_ARB_ALLOW_SR_EXIT_WATERMARK_Z8_C, prog_wm_value);
561		DC_LOG_BANDWIDTH_CALCS("SR_EXIT_WATERMARK_Z8_C calculated =%d\n"
562			"HW register value = 0x%x\n",
563			watermarks->c.cstate_pstate.cstate_exit_z8_ns, prog_wm_value);
564	} else if (watermarks->c.cstate_pstate.cstate_exit_z8_ns
565			< hubbub2->watermarks.c.cstate_pstate.cstate_exit_z8_ns)
566		wm_pending = true;
567
568	/* clock state D */
569	if (safe_to_lower || watermarks->d.cstate_pstate.cstate_enter_plus_exit_ns
570			> hubbub2->watermarks.d.cstate_pstate.cstate_enter_plus_exit_ns) {
571		hubbub2->watermarks.d.cstate_pstate.cstate_enter_plus_exit_ns =
572				watermarks->d.cstate_pstate.cstate_enter_plus_exit_ns;
573		prog_wm_value = convert_and_clamp(
574				watermarks->d.cstate_pstate.cstate_enter_plus_exit_ns,
575				refclk_mhz, 0xfffff);
576		REG_SET(DCHUBBUB_ARB_ALLOW_SR_ENTER_WATERMARK_D, 0,
577				DCHUBBUB_ARB_ALLOW_SR_ENTER_WATERMARK_D, prog_wm_value);
578		DC_LOG_BANDWIDTH_CALCS("SR_ENTER_EXIT_WATERMARK_D calculated =%d\n"
579			"HW register value = 0x%x\n",
580			watermarks->d.cstate_pstate.cstate_enter_plus_exit_ns, prog_wm_value);
581	} else if (watermarks->d.cstate_pstate.cstate_enter_plus_exit_ns
582			< hubbub2->watermarks.d.cstate_pstate.cstate_enter_plus_exit_ns)
583		wm_pending = true;
584
585	if (safe_to_lower || watermarks->d.cstate_pstate.cstate_exit_ns
586			> hubbub2->watermarks.d.cstate_pstate.cstate_exit_ns) {
587		hubbub2->watermarks.d.cstate_pstate.cstate_exit_ns =
588				watermarks->d.cstate_pstate.cstate_exit_ns;
589		prog_wm_value = convert_and_clamp(
590				watermarks->d.cstate_pstate.cstate_exit_ns,
591				refclk_mhz, 0xfffff);
592		REG_SET(DCHUBBUB_ARB_ALLOW_SR_EXIT_WATERMARK_D, 0,
593				DCHUBBUB_ARB_ALLOW_SR_EXIT_WATERMARK_D, prog_wm_value);
594		DC_LOG_BANDWIDTH_CALCS("SR_EXIT_WATERMARK_D calculated =%d\n"
595			"HW register value = 0x%x\n",
596			watermarks->d.cstate_pstate.cstate_exit_ns, prog_wm_value);
597	} else if (watermarks->d.cstate_pstate.cstate_exit_ns
598			< hubbub2->watermarks.d.cstate_pstate.cstate_exit_ns)
599		wm_pending = true;
600
601	if (safe_to_lower || watermarks->d.cstate_pstate.cstate_enter_plus_exit_z8_ns
602			> hubbub2->watermarks.d.cstate_pstate.cstate_enter_plus_exit_z8_ns) {
603		hubbub2->watermarks.d.cstate_pstate.cstate_enter_plus_exit_z8_ns =
604				watermarks->d.cstate_pstate.cstate_enter_plus_exit_z8_ns;
605		prog_wm_value = convert_and_clamp(
606				watermarks->d.cstate_pstate.cstate_enter_plus_exit_z8_ns,
607				refclk_mhz, 0xfffff);
608		REG_SET(DCHUBBUB_ARB_ALLOW_SR_ENTER_WATERMARK_Z8_D, 0,
609				DCHUBBUB_ARB_ALLOW_SR_ENTER_WATERMARK_Z8_D, prog_wm_value);
610		DC_LOG_BANDWIDTH_CALCS("SR_ENTER_WATERMARK_Z8_D calculated =%d\n"
611			"HW register value = 0x%x\n",
612			watermarks->d.cstate_pstate.cstate_enter_plus_exit_z8_ns, prog_wm_value);
613	} else if (watermarks->d.cstate_pstate.cstate_enter_plus_exit_z8_ns
614			< hubbub2->watermarks.d.cstate_pstate.cstate_enter_plus_exit_z8_ns)
615		wm_pending = true;
616
617	if (safe_to_lower || watermarks->d.cstate_pstate.cstate_exit_z8_ns
618			> hubbub2->watermarks.d.cstate_pstate.cstate_exit_z8_ns) {
619		hubbub2->watermarks.d.cstate_pstate.cstate_exit_z8_ns =
620				watermarks->d.cstate_pstate.cstate_exit_z8_ns;
621		prog_wm_value = convert_and_clamp(
622				watermarks->d.cstate_pstate.cstate_exit_z8_ns,
623				refclk_mhz, 0xfffff);
624		REG_SET(DCHUBBUB_ARB_ALLOW_SR_EXIT_WATERMARK_Z8_D, 0,
625				DCHUBBUB_ARB_ALLOW_SR_EXIT_WATERMARK_Z8_D, prog_wm_value);
626		DC_LOG_BANDWIDTH_CALCS("SR_EXIT_WATERMARK_Z8_D calculated =%d\n"
627			"HW register value = 0x%x\n",
628			watermarks->d.cstate_pstate.cstate_exit_z8_ns, prog_wm_value);
629	} else if (watermarks->d.cstate_pstate.cstate_exit_z8_ns
630			< hubbub2->watermarks.d.cstate_pstate.cstate_exit_z8_ns)
631		wm_pending = true;
632
633	return wm_pending;
634}
635
636static bool hubbub31_program_pstate_watermarks(
637		struct hubbub *hubbub,
638		struct dcn_watermark_set *watermarks,
639		unsigned int refclk_mhz,
640		bool safe_to_lower)
641{
642	struct dcn20_hubbub *hubbub2 = TO_DCN20_HUBBUB(hubbub);
643	uint32_t prog_wm_value;
644
645	bool wm_pending = false;
646
647	/* clock state A */
648	if (safe_to_lower || watermarks->a.cstate_pstate.pstate_change_ns
649			> hubbub2->watermarks.a.cstate_pstate.pstate_change_ns) {
650		hubbub2->watermarks.a.cstate_pstate.pstate_change_ns =
651				watermarks->a.cstate_pstate.pstate_change_ns;
652		prog_wm_value = convert_and_clamp(
653				watermarks->a.cstate_pstate.pstate_change_ns,
654				refclk_mhz, 0xffff);
655		REG_SET(DCHUBBUB_ARB_ALLOW_DRAM_CLK_CHANGE_WATERMARK_A, 0,
656				DCHUBBUB_ARB_ALLOW_DRAM_CLK_CHANGE_WATERMARK_A, prog_wm_value);
657		DC_LOG_BANDWIDTH_CALCS("DRAM_CLK_CHANGE_WATERMARK_A calculated =%d\n"
658			"HW register value = 0x%x\n\n",
659			watermarks->a.cstate_pstate.pstate_change_ns, prog_wm_value);
660	} else if (watermarks->a.cstate_pstate.pstate_change_ns
661			< hubbub2->watermarks.a.cstate_pstate.pstate_change_ns)
662		wm_pending = true;
663
664	/* clock state B */
665	if (safe_to_lower || watermarks->b.cstate_pstate.pstate_change_ns
666			> hubbub2->watermarks.b.cstate_pstate.pstate_change_ns) {
667		hubbub2->watermarks.b.cstate_pstate.pstate_change_ns =
668				watermarks->b.cstate_pstate.pstate_change_ns;
669		prog_wm_value = convert_and_clamp(
670				watermarks->b.cstate_pstate.pstate_change_ns,
671				refclk_mhz, 0xffff);
672		REG_SET(DCHUBBUB_ARB_ALLOW_DRAM_CLK_CHANGE_WATERMARK_B, 0,
673				DCHUBBUB_ARB_ALLOW_DRAM_CLK_CHANGE_WATERMARK_B, prog_wm_value);
674		DC_LOG_BANDWIDTH_CALCS("DRAM_CLK_CHANGE_WATERMARK_B calculated =%d\n"
675			"HW register value = 0x%x\n\n",
676			watermarks->b.cstate_pstate.pstate_change_ns, prog_wm_value);
677	} else if (watermarks->b.cstate_pstate.pstate_change_ns
678			< hubbub2->watermarks.b.cstate_pstate.pstate_change_ns)
679		wm_pending = false;
680
681	/* clock state C */
682	if (safe_to_lower || watermarks->c.cstate_pstate.pstate_change_ns
683			> hubbub2->watermarks.c.cstate_pstate.pstate_change_ns) {
684		hubbub2->watermarks.c.cstate_pstate.pstate_change_ns =
685				watermarks->c.cstate_pstate.pstate_change_ns;
686		prog_wm_value = convert_and_clamp(
687				watermarks->c.cstate_pstate.pstate_change_ns,
688				refclk_mhz, 0xffff);
689		REG_SET(DCHUBBUB_ARB_ALLOW_DRAM_CLK_CHANGE_WATERMARK_C, 0,
690				DCHUBBUB_ARB_ALLOW_DRAM_CLK_CHANGE_WATERMARK_C, prog_wm_value);
691		DC_LOG_BANDWIDTH_CALCS("DRAM_CLK_CHANGE_WATERMARK_C calculated =%d\n"
692			"HW register value = 0x%x\n\n",
693			watermarks->c.cstate_pstate.pstate_change_ns, prog_wm_value);
694	} else if (watermarks->c.cstate_pstate.pstate_change_ns
695			< hubbub2->watermarks.c.cstate_pstate.pstate_change_ns)
696		wm_pending = true;
697
698	/* clock state D */
699	if (safe_to_lower || watermarks->d.cstate_pstate.pstate_change_ns
700			> hubbub2->watermarks.d.cstate_pstate.pstate_change_ns) {
701		hubbub2->watermarks.d.cstate_pstate.pstate_change_ns =
702				watermarks->d.cstate_pstate.pstate_change_ns;
703		prog_wm_value = convert_and_clamp(
704				watermarks->d.cstate_pstate.pstate_change_ns,
705				refclk_mhz, 0xffff);
706		REG_SET(DCHUBBUB_ARB_ALLOW_DRAM_CLK_CHANGE_WATERMARK_D, 0,
707				DCHUBBUB_ARB_ALLOW_DRAM_CLK_CHANGE_WATERMARK_D, prog_wm_value);
708		DC_LOG_BANDWIDTH_CALCS("DRAM_CLK_CHANGE_WATERMARK_D calculated =%d\n"
709			"HW register value = 0x%x\n\n",
710			watermarks->d.cstate_pstate.pstate_change_ns, prog_wm_value);
711	} else if (watermarks->d.cstate_pstate.pstate_change_ns
712			< hubbub2->watermarks.d.cstate_pstate.pstate_change_ns)
713		wm_pending = true;
714
715	return wm_pending;
716}
717
718static bool hubbub31_program_watermarks(
719		struct hubbub *hubbub,
720		struct dcn_watermark_set *watermarks,
721		unsigned int refclk_mhz,
722		bool safe_to_lower)
723{
724	bool wm_pending = false;
725
726	if (hubbub31_program_urgent_watermarks(hubbub, watermarks, refclk_mhz, safe_to_lower))
727		wm_pending = true;
728
729	if (hubbub31_program_stutter_watermarks(hubbub, watermarks, refclk_mhz, safe_to_lower))
730		wm_pending = true;
731
732	if (hubbub31_program_pstate_watermarks(hubbub, watermarks, refclk_mhz, safe_to_lower))
733		wm_pending = true;
734
735	/*
736	 * The DCHub arbiter has a mechanism to dynamically rate limit the DCHub request stream to the fabric.
737	 * If the memory controller is fully utilized and the DCHub requestors are
738	 * well ahead of their amortized schedule, then it is safe to prevent the next winner
739	 * from being committed and sent to the fabric.
740	 * The utilization of the memory controller is approximated by ensuring that
741	 * the number of outstanding requests is greater than a threshold specified
742	 * by the ARB_MIN_REQ_OUTSTANDING. To determine that the DCHub requestors are well ahead of the amortized schedule,
743	 * the slack of the next winner is compared with the ARB_SAT_LEVEL in DLG RefClk cycles.
744	 *
745	 * TODO: Revisit request limit after figure out right number. request limit for RM isn't decided yet, set maximum value (0x1FF)
746	 * to turn off it for now.
747	 */
748	/*REG_SET(DCHUBBUB_ARB_SAT_LEVEL, 0,
749			DCHUBBUB_ARB_SAT_LEVEL, 60 * refclk_mhz);
750	REG_UPDATE(DCHUBBUB_ARB_DF_REQ_OUTSTAND,
751			DCHUBBUB_ARB_MIN_REQ_OUTSTAND, 0x1FF);*/
752
753	hubbub1_allow_self_refresh_control(hubbub, !hubbub->ctx->dc->debug.disable_stutter);
754	return wm_pending;
755}
756
757static void hubbub3_get_blk256_size(unsigned int *blk256_width, unsigned int *blk256_height,
758		unsigned int bytes_per_element)
759{
760	/* copied from DML.  might want to refactor DML to leverage from DML */
761	/* DML : get_blk256_size */
762	if (bytes_per_element == 1) {
763		*blk256_width = 16;
764		*blk256_height = 16;
765	} else if (bytes_per_element == 2) {
766		*blk256_width = 16;
767		*blk256_height = 8;
768	} else if (bytes_per_element == 4) {
769		*blk256_width = 8;
770		*blk256_height = 8;
771	} else if (bytes_per_element == 8) {
772		*blk256_width = 8;
773		*blk256_height = 4;
774	}
775}
776
777static void hubbub31_det_request_size(
778		unsigned int detile_buf_size,
779		unsigned int height,
780		unsigned int width,
781		unsigned int bpe,
782		bool *req128_horz_wc,
783		bool *req128_vert_wc)
784{
785	unsigned int blk256_height = 0;
786	unsigned int blk256_width = 0;
787	unsigned int swath_bytes_horz_wc, swath_bytes_vert_wc;
788
789	hubbub3_get_blk256_size(&blk256_width, &blk256_height, bpe);
790
791	swath_bytes_horz_wc = width * blk256_height * bpe;
792	swath_bytes_vert_wc = height * blk256_width * bpe;
793
794	*req128_horz_wc = (2 * swath_bytes_horz_wc <= detile_buf_size) ?
795			false : /* full 256B request */
796			true; /* half 128b request */
797
798	*req128_vert_wc = (2 * swath_bytes_vert_wc <= detile_buf_size) ?
799			false : /* full 256B request */
800			true; /* half 128b request */
801}
802
803static bool hubbub31_get_dcc_compression_cap(struct hubbub *hubbub,
804		const struct dc_dcc_surface_param *input,
805		struct dc_surface_dcc_cap *output)
806{
807	struct dc *dc = hubbub->ctx->dc;
808	enum dcc_control dcc_control;
809	unsigned int bpe;
810	enum segment_order segment_order_horz, segment_order_vert;
811	bool req128_horz_wc, req128_vert_wc;
812
813	memset(output, 0, sizeof(*output));
814
815	if (dc->debug.disable_dcc == DCC_DISABLE)
816		return false;
817
818	if (!hubbub->funcs->dcc_support_pixel_format(input->format,
819			&bpe))
820		return false;
821
822	if (!hubbub->funcs->dcc_support_swizzle(input->swizzle_mode, bpe,
823			&segment_order_horz, &segment_order_vert))
824		return false;
825
826	hubbub31_det_request_size(TO_DCN20_HUBBUB(hubbub)->detile_buf_size,
827			input->surface_size.height,  input->surface_size.width,
828			bpe, &req128_horz_wc, &req128_vert_wc);
829
830	if (!req128_horz_wc && !req128_vert_wc) {
831		dcc_control = dcc_control__256_256_xxx;
832	} else if (input->scan == SCAN_DIRECTION_HORIZONTAL) {
833		if (!req128_horz_wc)
834			dcc_control = dcc_control__256_256_xxx;
835		else if (segment_order_horz == segment_order__contiguous)
836			dcc_control = dcc_control__128_128_xxx;
837		else
838			dcc_control = dcc_control__256_64_64;
839	} else if (input->scan == SCAN_DIRECTION_VERTICAL) {
840		if (!req128_vert_wc)
841			dcc_control = dcc_control__256_256_xxx;
842		else if (segment_order_vert == segment_order__contiguous)
843			dcc_control = dcc_control__128_128_xxx;
844		else
845			dcc_control = dcc_control__256_64_64;
846	} else {
847		if ((req128_horz_wc &&
848			segment_order_horz == segment_order__non_contiguous) ||
849			(req128_vert_wc &&
850			segment_order_vert == segment_order__non_contiguous))
851			/* access_dir not known, must use most constraining */
852			dcc_control = dcc_control__256_64_64;
853		else
854			/* reg128 is true for either horz and vert
855			 * but segment_order is contiguous
856			 */
857			dcc_control = dcc_control__128_128_xxx;
858	}
859
860	/* Exception for 64KB_R_X */
861	if ((bpe == 2) && (input->swizzle_mode == DC_SW_64KB_R_X))
862		dcc_control = dcc_control__128_128_xxx;
863
864	if (dc->debug.disable_dcc == DCC_HALF_REQ_DISALBE &&
865		dcc_control != dcc_control__256_256_xxx)
866		return false;
867
868	switch (dcc_control) {
869	case dcc_control__256_256_xxx:
870		output->grph.rgb.max_uncompressed_blk_size = 256;
871		output->grph.rgb.max_compressed_blk_size = 256;
872		output->grph.rgb.independent_64b_blks = false;
873		output->grph.rgb.dcc_controls.dcc_256_256_unconstrained = 1;
874		output->grph.rgb.dcc_controls.dcc_256_128_128 = 1;
875		break;
876	case dcc_control__128_128_xxx:
877		output->grph.rgb.max_uncompressed_blk_size = 128;
878		output->grph.rgb.max_compressed_blk_size = 128;
879		output->grph.rgb.independent_64b_blks = false;
880		output->grph.rgb.dcc_controls.dcc_128_128_uncontrained = 1;
881		output->grph.rgb.dcc_controls.dcc_256_128_128 = 1;
882		break;
883	case dcc_control__256_64_64:
884		output->grph.rgb.max_uncompressed_blk_size = 256;
885		output->grph.rgb.max_compressed_blk_size = 64;
886		output->grph.rgb.independent_64b_blks = true;
887		output->grph.rgb.dcc_controls.dcc_256_64_64 = 1;
888		break;
889	case dcc_control__256_128_128:
890		output->grph.rgb.max_uncompressed_blk_size = 256;
891		output->grph.rgb.max_compressed_blk_size = 128;
892		output->grph.rgb.independent_64b_blks = false;
893		output->grph.rgb.dcc_controls.dcc_256_128_128 = 1;
894		break;
895	}
896	output->capable = true;
897	output->const_color_support = true;
898
899	return true;
900}
901
902int hubbub31_init_dchub_sys_ctx(struct hubbub *hubbub,
903		struct dcn_hubbub_phys_addr_config *pa_config)
904{
905	struct dcn20_hubbub *hubbub2 = TO_DCN20_HUBBUB(hubbub);
906	struct dcn_vmid_page_table_config phys_config;
907
908	REG_SET(DCN_VM_FB_LOCATION_BASE, 0,
909			FB_BASE, pa_config->system_aperture.fb_base >> 24);
910	REG_SET(DCN_VM_FB_LOCATION_TOP, 0,
911			FB_TOP, pa_config->system_aperture.fb_top >> 24);
912	REG_SET(DCN_VM_FB_OFFSET, 0,
913			FB_OFFSET, pa_config->system_aperture.fb_offset >> 24);
914	REG_SET(DCN_VM_AGP_BOT, 0,
915			AGP_BOT, pa_config->system_aperture.agp_bot >> 24);
916	REG_SET(DCN_VM_AGP_TOP, 0,
917			AGP_TOP, pa_config->system_aperture.agp_top >> 24);
918	REG_SET(DCN_VM_AGP_BASE, 0,
919			AGP_BASE, pa_config->system_aperture.agp_base >> 24);
920
921	if (pa_config->gart_config.page_table_start_addr != pa_config->gart_config.page_table_end_addr) {
922		phys_config.page_table_start_addr = pa_config->gart_config.page_table_start_addr >> 12;
923		phys_config.page_table_end_addr = pa_config->gart_config.page_table_end_addr >> 12;
924		phys_config.page_table_base_addr = pa_config->gart_config.page_table_base_addr;
925		phys_config.depth = 0;
926		phys_config.block_size = 0;
927		// Init VMID 0 based on PA config
928		dcn20_vmid_setup(&hubbub2->vmid[0], &phys_config);
929
930		dcn20_vmid_setup(&hubbub2->vmid[15], &phys_config);
931	}
932
933	dcn21_dchvm_init(hubbub);
934
935	return NUM_VMID;
936}
937
938static void hubbub31_get_dchub_ref_freq(struct hubbub *hubbub,
939		unsigned int dccg_ref_freq_inKhz,
940		unsigned int *dchub_ref_freq_inKhz)
941{
942	struct dcn20_hubbub *hubbub2 = TO_DCN20_HUBBUB(hubbub);
943	uint32_t ref_div = 0;
944	uint32_t ref_en = 0;
945	unsigned int dc_refclk_khz = 24000;
946
947	REG_GET_2(DCHUBBUB_GLOBAL_TIMER_CNTL, DCHUBBUB_GLOBAL_TIMER_REFDIV, &ref_div,
948			DCHUBBUB_GLOBAL_TIMER_ENABLE, &ref_en);
949
950	if (ref_en) {
951		if (ref_div == 2)
952			*dchub_ref_freq_inKhz = dc_refclk_khz / 2;
953		else
954			*dchub_ref_freq_inKhz = dc_refclk_khz;
955
956		/*
957		 * The external Reference Clock may change based on the board or
958		 * platform requirements and the programmable integer divide must
959		 * be programmed to provide a suitable DLG RefClk frequency between
960		 * a minimum of 20MHz and maximum of 50MHz
961		 */
962		if (*dchub_ref_freq_inKhz < 20000 || *dchub_ref_freq_inKhz > 50000)
963			ASSERT_CRITICAL(false);
964
965		return;
966	} else {
967		*dchub_ref_freq_inKhz = dc_refclk_khz;
968
969		// HUBBUB global timer must be enabled.
970		ASSERT_CRITICAL(false);
971		return;
972	}
973}
974
975static bool hubbub31_verify_allow_pstate_change_high(struct hubbub *hubbub)
976{
977	struct dcn20_hubbub *hubbub2 = TO_DCN20_HUBBUB(hubbub);
978
979	/*
980	 * Pstate latency is ~20us so if we wait over 40us and pstate allow
981	 * still not asserted, we are probably stuck and going to hang
982	 */
983	const unsigned int pstate_wait_timeout_us = 100;
984	const unsigned int pstate_wait_expected_timeout_us = 40;
985
986	static unsigned int max_sampled_pstate_wait_us; /* data collection */
987	static bool forced_pstate_allow; /* help with revert wa */
988
989	unsigned int debug_data = 0;
990	unsigned int i;
991
992	if (forced_pstate_allow) {
993		/* we hacked to force pstate allow to prevent hang last time
994		 * we verify_allow_pstate_change_high.  so disable force
995		 * here so we can check status
996		 */
997		REG_UPDATE_2(DCHUBBUB_ARB_DRAM_STATE_CNTL,
998			     DCHUBBUB_ARB_ALLOW_PSTATE_CHANGE_FORCE_VALUE, 0,
999			     DCHUBBUB_ARB_ALLOW_PSTATE_CHANGE_FORCE_ENABLE, 0);
1000		forced_pstate_allow = false;
1001	}
1002
1003	REG_WRITE(DCHUBBUB_TEST_DEBUG_INDEX, hubbub2->debug_test_index_pstate);
1004
1005	for (i = 0; i < pstate_wait_timeout_us; i++) {
1006		debug_data = REG_READ(DCHUBBUB_TEST_DEBUG_DATA);
1007
1008		/* Debug bit is specific to ASIC. */
1009		if (debug_data & (1 << 26)) {
1010			if (i > pstate_wait_expected_timeout_us)
1011				DC_LOG_WARNING("pstate took longer than expected ~%dus\n", i);
1012			return true;
1013		}
1014		if (max_sampled_pstate_wait_us < i)
1015			max_sampled_pstate_wait_us = i;
1016
1017		udelay(1);
1018	}
1019
1020	/* force pstate allow to prevent system hang
1021	 * and break to debugger to investigate
1022	 */
1023	REG_UPDATE_2(DCHUBBUB_ARB_DRAM_STATE_CNTL,
1024		     DCHUBBUB_ARB_ALLOW_PSTATE_CHANGE_FORCE_VALUE, 1,
1025		     DCHUBBUB_ARB_ALLOW_PSTATE_CHANGE_FORCE_ENABLE, 1);
1026	forced_pstate_allow = true;
1027
1028	DC_LOG_WARNING("pstate TEST_DEBUG_DATA: 0x%X\n",
1029			debug_data);
1030
1031	return false;
1032}
1033
1034void hubbub31_init(struct hubbub *hubbub)
1035{
1036	struct dcn20_hubbub *hubbub2 = TO_DCN20_HUBBUB(hubbub);
1037
1038	/*Enable clock gate*/
1039	if (hubbub->ctx->dc->debug.disable_clock_gate) {
1040		/*done in hwseq*/
1041		/*REG_UPDATE(DCFCLK_CNTL, DCFCLK_GATE_DIS, 0);*/
1042		REG_UPDATE_2(DCHUBBUB_CLOCK_CNTL,
1043				DISPCLK_R_DCHUBBUB_GATE_DIS, 1,
1044				DCFCLK_R_DCHUBBUB_GATE_DIS, 1);
1045	}
1046
1047	/*
1048	only the DCN will determine when to connect the SDP port
1049	*/
1050	REG_UPDATE(DCHUBBUB_SDPIF_CFG0,	SDPIF_PORT_CONTROL, 1);
1051}
1052static const struct hubbub_funcs hubbub31_funcs = {
1053	.update_dchub = hubbub2_update_dchub,
1054	.init_dchub_sys_ctx = hubbub31_init_dchub_sys_ctx,
1055	.init_vm_ctx = hubbub2_init_vm_ctx,
1056	.dcc_support_swizzle = hubbub3_dcc_support_swizzle,
1057	.dcc_support_pixel_format = hubbub2_dcc_support_pixel_format,
1058	.get_dcc_compression_cap = hubbub31_get_dcc_compression_cap,
1059	.wm_read_state = hubbub21_wm_read_state,
1060	.get_dchub_ref_freq = hubbub31_get_dchub_ref_freq,
1061	.program_watermarks = hubbub31_program_watermarks,
1062	.allow_self_refresh_control = hubbub1_allow_self_refresh_control,
1063	.is_allow_self_refresh_enabled = hubbub1_is_allow_self_refresh_enabled,
1064	.verify_allow_pstate_change_high = hubbub31_verify_allow_pstate_change_high,
1065	.program_det_size = dcn31_program_det_size,
1066	.wait_for_det_apply = dcn31_wait_for_det_apply,
1067	.program_compbuf_size = dcn31_program_compbuf_size,
1068	.init_crb = dcn31_init_crb,
1069	.hubbub_read_state = hubbub2_read_state,
1070};
1071
1072void hubbub31_construct(struct dcn20_hubbub *hubbub31,
1073	struct dc_context *ctx,
1074	const struct dcn_hubbub_registers *hubbub_regs,
1075	const struct dcn_hubbub_shift *hubbub_shift,
1076	const struct dcn_hubbub_mask *hubbub_mask,
1077	int det_size_kb,
1078	int pixel_chunk_size_kb,
1079	int config_return_buffer_size_kb)
1080{
1081
1082	hubbub3_construct(hubbub31, ctx, hubbub_regs, hubbub_shift, hubbub_mask);
1083	hubbub31->base.funcs = &hubbub31_funcs;
1084	hubbub31->detile_buf_size = det_size_kb * 1024;
1085	hubbub31->pixel_chunk_size = pixel_chunk_size_kb * 1024;
1086	hubbub31->crb_size_segs = config_return_buffer_size_kb / DCN31_CRB_SEGMENT_SIZE_KB;
1087
1088	hubbub31->debug_test_index_pstate = 0x6;
1089}
1090
1091