1/*
2 * Copyright 2020-2021 Advanced Micro Devices, Inc.
3 *
4 * Permission is hereby granted, free of charge, to any person obtaining a
5 * copy of this software and associated documentation files (the "Software"),
6 * to deal in the Software without restriction, including without limitation
7 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
8 * and/or sell copies of the Software, and to permit persons to whom the
9 * Software is furnished to do so, subject to the following conditions:
10 *
11 * The above copyright notice and this permission notice shall be included in
12 * all copies or substantial portions of the Software.
13 *
14 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
15 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
16 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
17 * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
18 * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
19 * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
20 * OTHER DEALINGS IN THE SOFTWARE.
21 *
22 * Authors: AMD
23 *
24 */
25#include "resource.h"
26#include "clk_mgr.h"
27#include "reg_helper.h"
28#include "dcn_calc_math.h"
29#include "dcn20/dcn20_resource.h"
30#include "dcn30/dcn30_resource.h"
31
32#include "clk_mgr/dcn30/dcn30_smu11_driver_if.h"
33#include "display_mode_vba_30.h"
34#include "dcn30_fpu.h"
35
36#define REG(reg)\
37	optc1->tg_regs->reg
38
39#define CTX \
40	optc1->base.ctx
41
42#undef FN
43#define FN(reg_name, field_name) \
44	optc1->tg_shift->field_name, optc1->tg_mask->field_name
45
46
47struct _vcs_dpi_ip_params_st dcn3_0_ip = {
48	.use_min_dcfclk = 0,
49	.clamp_min_dcfclk = 0,
50	.odm_capable = 1,
51	.gpuvm_enable = 0,
52	.hostvm_enable = 0,
53	.gpuvm_max_page_table_levels = 4,
54	.hostvm_max_page_table_levels = 4,
55	.hostvm_cached_page_table_levels = 0,
56	.pte_group_size_bytes = 2048,
57	.num_dsc = 6,
58	.rob_buffer_size_kbytes = 184,
59	.det_buffer_size_kbytes = 184,
60	.dpte_buffer_size_in_pte_reqs_luma = 84,
61	.pde_proc_buffer_size_64k_reqs = 48,
62	.dpp_output_buffer_pixels = 2560,
63	.opp_output_buffer_lines = 1,
64	.pixel_chunk_size_kbytes = 8,
65	.pte_enable = 1,
66	.max_page_table_levels = 2,
67	.pte_chunk_size_kbytes = 2,  // ?
68	.meta_chunk_size_kbytes = 2,
69	.writeback_chunk_size_kbytes = 8,
70	.line_buffer_size_bits = 789504,
71	.is_line_buffer_bpp_fixed = 0,  // ?
72	.line_buffer_fixed_bpp = 0,     // ?
73	.dcc_supported = true,
74	.writeback_interface_buffer_size_kbytes = 90,
75	.writeback_line_buffer_buffer_size = 0,
76	.max_line_buffer_lines = 12,
77	.writeback_luma_buffer_size_kbytes = 12,  // writeback_line_buffer_buffer_size = 656640
78	.writeback_chroma_buffer_size_kbytes = 8,
79	.writeback_chroma_line_buffer_width_pixels = 4,
80	.writeback_max_hscl_ratio = 1,
81	.writeback_max_vscl_ratio = 1,
82	.writeback_min_hscl_ratio = 1,
83	.writeback_min_vscl_ratio = 1,
84	.writeback_max_hscl_taps = 1,
85	.writeback_max_vscl_taps = 1,
86	.writeback_line_buffer_luma_buffer_size = 0,
87	.writeback_line_buffer_chroma_buffer_size = 14643,
88	.cursor_buffer_size = 8,
89	.cursor_chunk_size = 2,
90	.max_num_otg = 6,
91	.max_num_dpp = 6,
92	.max_num_wb = 1,
93	.max_dchub_pscl_bw_pix_per_clk = 4,
94	.max_pscl_lb_bw_pix_per_clk = 2,
95	.max_lb_vscl_bw_pix_per_clk = 4,
96	.max_vscl_hscl_bw_pix_per_clk = 4,
97	.max_hscl_ratio = 6,
98	.max_vscl_ratio = 6,
99	.hscl_mults = 4,
100	.vscl_mults = 4,
101	.max_hscl_taps = 8,
102	.max_vscl_taps = 8,
103	.dispclk_ramp_margin_percent = 1,
104	.underscan_factor = 1.11,
105	.min_vblank_lines = 32,
106	.dppclk_delay_subtotal = 46,
107	.dynamic_metadata_vm_enabled = true,
108	.dppclk_delay_scl_lb_only = 16,
109	.dppclk_delay_scl = 50,
110	.dppclk_delay_cnvc_formatter = 27,
111	.dppclk_delay_cnvc_cursor = 6,
112	.dispclk_delay_subtotal = 119,
113	.dcfclk_cstate_latency = 5.2, // SRExitTime
114	.max_inter_dcn_tile_repeaters = 8,
115	.max_num_hdmi_frl_outputs = 1,
116	.odm_combine_4to1_supported = true,
117
118	.xfc_supported = false,
119	.xfc_fill_bw_overhead_percent = 10.0,
120	.xfc_fill_constant_bytes = 0,
121	.gfx7_compat_tiling_supported = 0,
122	.number_of_cursors = 1,
123};
124
125struct _vcs_dpi_soc_bounding_box_st dcn3_0_soc = {
126	.clock_limits = {
127			{
128				.state = 0,
129				.dispclk_mhz = 562.0,
130				.dppclk_mhz = 300.0,
131				.phyclk_mhz = 300.0,
132				.phyclk_d18_mhz = 667.0,
133				.dscclk_mhz = 405.6,
134			},
135		},
136
137	.min_dcfclk = 500.0, /* TODO: set this to actual min DCFCLK */
138	.num_states = 1,
139	.sr_exit_time_us = 15.5,
140	.sr_enter_plus_exit_time_us = 20,
141	.urgent_latency_us = 4.0,
142	.urgent_latency_pixel_data_only_us = 4.0,
143	.urgent_latency_pixel_mixed_with_vm_data_us = 4.0,
144	.urgent_latency_vm_data_only_us = 4.0,
145	.urgent_out_of_order_return_per_channel_pixel_only_bytes = 4096,
146	.urgent_out_of_order_return_per_channel_pixel_and_vm_bytes = 4096,
147	.urgent_out_of_order_return_per_channel_vm_only_bytes = 4096,
148	.pct_ideal_dram_sdp_bw_after_urgent_pixel_only = 80.0,
149	.pct_ideal_dram_sdp_bw_after_urgent_pixel_and_vm = 60.0,
150	.pct_ideal_dram_sdp_bw_after_urgent_vm_only = 40.0,
151	.max_avg_sdp_bw_use_normal_percent = 60.0,
152	.max_avg_dram_bw_use_normal_percent = 40.0,
153	.writeback_latency_us = 12.0,
154	.max_request_size_bytes = 256,
155	.fabric_datapath_to_dcn_data_return_bytes = 64,
156	.dcn_downspread_percent = 0.5,
157	.downspread_percent = 0.38,
158	.dram_page_open_time_ns = 50.0,
159	.dram_rw_turnaround_time_ns = 17.5,
160	.dram_return_buffer_per_channel_bytes = 8192,
161	.round_trip_ping_latency_dcfclk_cycles = 191,
162	.urgent_out_of_order_return_per_channel_bytes = 4096,
163	.channel_interleave_bytes = 256,
164	.num_banks = 8,
165	.gpuvm_min_page_size_bytes = 4096,
166	.hostvm_min_page_size_bytes = 4096,
167	.dram_clock_change_latency_us = 404,
168	.dummy_pstate_latency_us = 5,
169	.writeback_dram_clock_change_latency_us = 23.0,
170	.return_bus_width_bytes = 64,
171	.dispclk_dppclk_vco_speed_mhz = 3650,
172	.xfc_bus_transport_time_us = 20,      // ?
173	.xfc_xbuf_latency_tolerance_us = 4,  // ?
174	.use_urgent_burst_bw = 1,            // ?
175	.do_urgent_latency_adjustment = true,
176	.urgent_latency_adjustment_fabric_clock_component_us = 1.0,
177	.urgent_latency_adjustment_fabric_clock_reference_mhz = 1000,
178};
179
180
181void optc3_fpu_set_vrr_m_const(struct timing_generator *optc,
182		double vtotal_avg)
183{
184	struct optc *optc1 = DCN10TG_FROM_TG(optc);
185	double vtotal_min, vtotal_max;
186	double ratio, modulo, phase;
187	uint32_t vblank_start;
188	uint32_t v_total_mask_value = 0;
189
190	dc_assert_fp_enabled();
191
192	/* Compute VTOTAL_MIN and VTOTAL_MAX, so that
193	 * VOTAL_MAX - VTOTAL_MIN = 1
194	 */
195	v_total_mask_value = 16;
196	vtotal_min = dcn_bw_floor(vtotal_avg);
197	vtotal_max = dcn_bw_ceil(vtotal_avg);
198
199	/* Check that bottom VBLANK is at least 2 lines tall when running with
200	 * VTOTAL_MIN. Note that VTOTAL registers are defined as 'total number
201	 * of lines in a frame - 1'.
202	 */
203	REG_GET(OTG_V_BLANK_START_END, OTG_V_BLANK_START,
204		&vblank_start);
205	ASSERT(vtotal_min >= vblank_start + 1);
206
207	/* Special case where the average frame rate can be achieved
208	 * without using the DTO
209	 */
210	if (vtotal_min == vtotal_max) {
211		REG_SET(OTG_V_TOTAL, 0, OTG_V_TOTAL, (uint32_t)vtotal_min);
212
213		optc->funcs->set_vtotal_min_max(optc, 0, 0);
214		REG_SET(OTG_M_CONST_DTO0, 0, OTG_M_CONST_DTO_PHASE, 0);
215		REG_SET(OTG_M_CONST_DTO1, 0, OTG_M_CONST_DTO_MODULO, 0);
216		REG_UPDATE_3(OTG_V_TOTAL_CONTROL,
217			OTG_V_TOTAL_MIN_SEL, 0,
218			OTG_V_TOTAL_MAX_SEL, 0,
219			OTG_SET_V_TOTAL_MIN_MASK_EN, 0);
220		return;
221	}
222
223	ratio = vtotal_max - vtotal_avg;
224	modulo = 65536.0 * 65536.0 - 1.0; /* 2^32 - 1 */
225	phase = ratio * modulo;
226
227	/* Special cases where the DTO phase gets rounded to 0 or
228	 * to DTO modulo
229	 */
230	if (phase <= 0 || phase >= modulo) {
231		REG_SET(OTG_V_TOTAL, 0, OTG_V_TOTAL,
232			phase <= 0 ?
233				(uint32_t)vtotal_max : (uint32_t)vtotal_min);
234		REG_SET(OTG_V_TOTAL_MIN, 0, OTG_V_TOTAL_MIN, 0);
235		REG_SET(OTG_V_TOTAL_MAX, 0, OTG_V_TOTAL_MAX, 0);
236		REG_SET(OTG_M_CONST_DTO0, 0, OTG_M_CONST_DTO_PHASE, 0);
237		REG_SET(OTG_M_CONST_DTO1, 0, OTG_M_CONST_DTO_MODULO, 0);
238		REG_UPDATE_3(OTG_V_TOTAL_CONTROL,
239			OTG_V_TOTAL_MIN_SEL, 0,
240			OTG_V_TOTAL_MAX_SEL, 0,
241			OTG_SET_V_TOTAL_MIN_MASK_EN, 0);
242		return;
243	}
244	REG_UPDATE_6(OTG_V_TOTAL_CONTROL,
245		OTG_V_TOTAL_MIN_SEL, 1,
246		OTG_V_TOTAL_MAX_SEL, 1,
247		OTG_SET_V_TOTAL_MIN_MASK_EN, 1,
248		OTG_SET_V_TOTAL_MIN_MASK, v_total_mask_value,
249		OTG_VTOTAL_MID_REPLACING_MIN_EN, 0,
250		OTG_VTOTAL_MID_REPLACING_MAX_EN, 0);
251	REG_SET(OTG_V_TOTAL, 0, OTG_V_TOTAL, (uint32_t)vtotal_min);
252	optc->funcs->set_vtotal_min_max(optc, vtotal_min, vtotal_max);
253	REG_SET(OTG_M_CONST_DTO0, 0, OTG_M_CONST_DTO_PHASE, (uint32_t)phase);
254	REG_SET(OTG_M_CONST_DTO1, 0, OTG_M_CONST_DTO_MODULO, (uint32_t)modulo);
255}
256
257void dcn30_fpu_populate_dml_writeback_from_context(
258		struct dc *dc, struct resource_context *res_ctx, display_e2e_pipe_params_st *pipes)
259{
260	int pipe_cnt, i, j;
261	double max_calc_writeback_dispclk;
262	double writeback_dispclk;
263	struct writeback_st dout_wb;
264
265	dc_assert_fp_enabled();
266
267	for (i = 0, pipe_cnt = 0; i < dc->res_pool->pipe_count; i++) {
268		struct dc_stream_state *stream = res_ctx->pipe_ctx[i].stream;
269
270		if (!stream)
271			continue;
272		max_calc_writeback_dispclk = 0;
273
274		/* Set writeback information */
275		pipes[pipe_cnt].dout.wb_enable = 0;
276		pipes[pipe_cnt].dout.num_active_wb = 0;
277		for (j = 0; j < stream->num_wb_info; j++) {
278			struct dc_writeback_info *wb_info = &stream->writeback_info[j];
279
280			if (wb_info->wb_enabled && wb_info->writeback_source_plane &&
281					(wb_info->writeback_source_plane == res_ctx->pipe_ctx[i].plane_state)) {
282				pipes[pipe_cnt].dout.wb_enable = 1;
283				pipes[pipe_cnt].dout.num_active_wb++;
284				dout_wb.wb_src_height = wb_info->dwb_params.cnv_params.crop_en ?
285					wb_info->dwb_params.cnv_params.crop_height :
286					wb_info->dwb_params.cnv_params.src_height;
287				dout_wb.wb_src_width = wb_info->dwb_params.cnv_params.crop_en ?
288					wb_info->dwb_params.cnv_params.crop_width :
289					wb_info->dwb_params.cnv_params.src_width;
290				dout_wb.wb_dst_width = wb_info->dwb_params.dest_width;
291				dout_wb.wb_dst_height = wb_info->dwb_params.dest_height;
292
293				/* For IP that doesn't support WB scaling, set h/v taps to 1 to avoid DML validation failure */
294				if (dc->dml.ip.writeback_max_hscl_taps > 1) {
295					dout_wb.wb_htaps_luma = wb_info->dwb_params.scaler_taps.h_taps;
296					dout_wb.wb_vtaps_luma = wb_info->dwb_params.scaler_taps.v_taps;
297				} else {
298					dout_wb.wb_htaps_luma = 1;
299					dout_wb.wb_vtaps_luma = 1;
300				}
301				dout_wb.wb_htaps_chroma = 0;
302				dout_wb.wb_vtaps_chroma = 0;
303				dout_wb.wb_hratio = wb_info->dwb_params.cnv_params.crop_en ?
304					(double)wb_info->dwb_params.cnv_params.crop_width /
305						(double)wb_info->dwb_params.dest_width :
306					(double)wb_info->dwb_params.cnv_params.src_width /
307						(double)wb_info->dwb_params.dest_width;
308				dout_wb.wb_vratio = wb_info->dwb_params.cnv_params.crop_en ?
309					(double)wb_info->dwb_params.cnv_params.crop_height /
310						(double)wb_info->dwb_params.dest_height :
311					(double)wb_info->dwb_params.cnv_params.src_height /
312						(double)wb_info->dwb_params.dest_height;
313				if (wb_info->dwb_params.cnv_params.fc_out_format == DWB_OUT_FORMAT_64BPP_ARGB ||
314					wb_info->dwb_params.cnv_params.fc_out_format == DWB_OUT_FORMAT_64BPP_RGBA)
315					dout_wb.wb_pixel_format = dm_444_64;
316				else
317					dout_wb.wb_pixel_format = dm_444_32;
318
319				/* Workaround for cases where multiple writebacks are connected to same plane
320				 * In which case, need to compute worst case and set the associated writeback parameters
321				 * This workaround is necessary due to DML computation assuming only 1 set of writeback
322				 * parameters per pipe
323				 */
324				writeback_dispclk = dml30_CalculateWriteBackDISPCLK(
325						dout_wb.wb_pixel_format,
326						pipes[pipe_cnt].pipe.dest.pixel_rate_mhz,
327						dout_wb.wb_hratio,
328						dout_wb.wb_vratio,
329						dout_wb.wb_htaps_luma,
330						dout_wb.wb_vtaps_luma,
331						dout_wb.wb_src_width,
332						dout_wb.wb_dst_width,
333						pipes[pipe_cnt].pipe.dest.htotal,
334						dc->current_state->bw_ctx.dml.ip.writeback_line_buffer_buffer_size);
335
336				if (writeback_dispclk > max_calc_writeback_dispclk) {
337					max_calc_writeback_dispclk = writeback_dispclk;
338					pipes[pipe_cnt].dout.wb = dout_wb;
339				}
340			}
341		}
342
343		pipe_cnt++;
344	}
345}
346
347void dcn30_fpu_set_mcif_arb_params(struct mcif_arb_params *wb_arb_params,
348	struct display_mode_lib *dml,
349	display_e2e_pipe_params_st *pipes,
350	int pipe_cnt,
351	int cur_pipe)
352{
353	int i;
354
355	dc_assert_fp_enabled();
356
357	for (i = 0; i < ARRAY_SIZE(wb_arb_params->cli_watermark); i++) {
358		wb_arb_params->cli_watermark[i] = get_wm_writeback_urgent(dml, pipes, pipe_cnt) * 1000;
359		wb_arb_params->pstate_watermark[i] = get_wm_writeback_dram_clock_change(dml, pipes, pipe_cnt) * 1000;
360	}
361
362	wb_arb_params->dram_speed_change_duration = dml->vba.WritebackAllowDRAMClockChangeEndPosition[cur_pipe] * pipes[0].clks_cfg.refclk_mhz; /* num_clock_cycles = us * MHz */
363}
364
365void dcn30_fpu_update_soc_for_wm_a(struct dc *dc, struct dc_state *context)
366{
367
368	dc_assert_fp_enabled();
369
370	if (dc->clk_mgr->bw_params->wm_table.nv_entries[WM_A].valid) {
371		if (!context->bw_ctx.bw.dcn.clk.fw_based_mclk_switching ||
372				context->bw_ctx.dml.soc.dram_clock_change_latency_us == 0)
373			context->bw_ctx.dml.soc.dram_clock_change_latency_us = dc->clk_mgr->bw_params->wm_table.nv_entries[WM_A].dml_input.pstate_latency_us;
374		context->bw_ctx.dml.soc.sr_enter_plus_exit_time_us = dc->clk_mgr->bw_params->wm_table.nv_entries[WM_A].dml_input.sr_enter_plus_exit_time_us;
375		context->bw_ctx.dml.soc.sr_exit_time_us = dc->clk_mgr->bw_params->wm_table.nv_entries[WM_A].dml_input.sr_exit_time_us;
376	}
377}
378
379void dcn30_fpu_calculate_wm_and_dlg(
380		struct dc *dc, struct dc_state *context,
381		display_e2e_pipe_params_st *pipes,
382		int pipe_cnt,
383		int vlevel)
384{
385	int maxMpcComb = context->bw_ctx.dml.vba.maxMpcComb;
386	int i, pipe_idx;
387	double dcfclk = context->bw_ctx.dml.vba.DCFCLKState[vlevel][maxMpcComb];
388	bool pstate_en = context->bw_ctx.dml.vba.DRAMClockChangeSupport[vlevel][maxMpcComb] != dm_dram_clock_change_unsupported;
389	unsigned int dummy_latency_index = 0;
390
391	dc_assert_fp_enabled();
392
393	context->bw_ctx.bw.dcn.clk.fw_based_mclk_switching = false;
394    for (i = 0; i < context->stream_count; i++) {
395		if (context->streams[i])
396			context->streams[i]->fpo_in_use = false;
397	}
398
399	if (!pstate_en) {
400		/* only when the mclk switch can not be natural, is the fw based vblank stretch attempted */
401		context->bw_ctx.bw.dcn.clk.fw_based_mclk_switching =
402			dcn30_can_support_mclk_switch_using_fw_based_vblank_stretch(dc, context);
403
404		if (context->bw_ctx.bw.dcn.clk.fw_based_mclk_switching) {
405			dummy_latency_index = dcn30_find_dummy_latency_index_for_fw_based_mclk_switch(dc,
406				context, pipes, pipe_cnt, vlevel);
407
408			/* After calling dcn30_find_dummy_latency_index_for_fw_based_mclk_switch
409			 * we reinstate the original dram_clock_change_latency_us on the context
410			 * and all variables that may have changed up to this point, except the
411			 * newly found dummy_latency_index
412			 */
413			context->bw_ctx.dml.soc.dram_clock_change_latency_us = dc->clk_mgr->bw_params->wm_table.nv_entries[WM_A].dml_input.pstate_latency_us;
414			dcn30_internal_validate_bw(dc, context, pipes, &pipe_cnt, &vlevel, false, true);
415			maxMpcComb = context->bw_ctx.dml.vba.maxMpcComb;
416			dcfclk = context->bw_ctx.dml.vba.DCFCLKState[vlevel][context->bw_ctx.dml.vba.maxMpcComb];
417			pstate_en = context->bw_ctx.dml.vba.DRAMClockChangeSupport[vlevel][maxMpcComb] != dm_dram_clock_change_unsupported;
418		}
419	}
420
421	if (context->bw_ctx.dml.soc.min_dcfclk > dcfclk)
422		dcfclk = context->bw_ctx.dml.soc.min_dcfclk;
423
424	pipes[0].clks_cfg.voltage = vlevel;
425	pipes[0].clks_cfg.dcfclk_mhz = dcfclk;
426	pipes[0].clks_cfg.socclk_mhz = context->bw_ctx.dml.soc.clock_limits[vlevel].socclk_mhz;
427
428	/* Set B:
429	 * DCFCLK: 1GHz or min required above 1GHz
430	 * FCLK/UCLK: Max
431	 */
432	if (dc->clk_mgr->bw_params->wm_table.nv_entries[WM_B].valid) {
433		if (vlevel == 0) {
434			pipes[0].clks_cfg.voltage = 1;
435			pipes[0].clks_cfg.dcfclk_mhz = context->bw_ctx.dml.soc.clock_limits[0].dcfclk_mhz;
436		}
437		context->bw_ctx.dml.soc.dram_clock_change_latency_us = dc->clk_mgr->bw_params->wm_table.nv_entries[WM_B].dml_input.pstate_latency_us;
438		context->bw_ctx.dml.soc.sr_enter_plus_exit_time_us = dc->clk_mgr->bw_params->wm_table.nv_entries[WM_B].dml_input.sr_enter_plus_exit_time_us;
439		context->bw_ctx.dml.soc.sr_exit_time_us = dc->clk_mgr->bw_params->wm_table.nv_entries[WM_B].dml_input.sr_exit_time_us;
440	}
441	context->bw_ctx.bw.dcn.watermarks.b.urgent_ns = get_wm_urgent(&context->bw_ctx.dml, pipes, pipe_cnt) * 1000;
442	context->bw_ctx.bw.dcn.watermarks.b.cstate_pstate.cstate_enter_plus_exit_ns = get_wm_stutter_enter_exit(&context->bw_ctx.dml, pipes, pipe_cnt) * 1000;
443	context->bw_ctx.bw.dcn.watermarks.b.cstate_pstate.cstate_exit_ns = get_wm_stutter_exit(&context->bw_ctx.dml, pipes, pipe_cnt) * 1000;
444	context->bw_ctx.bw.dcn.watermarks.b.cstate_pstate.pstate_change_ns = get_wm_dram_clock_change(&context->bw_ctx.dml, pipes, pipe_cnt) * 1000;
445	context->bw_ctx.bw.dcn.watermarks.b.pte_meta_urgent_ns = get_wm_memory_trip(&context->bw_ctx.dml, pipes, pipe_cnt) * 1000;
446	context->bw_ctx.bw.dcn.watermarks.b.frac_urg_bw_nom = get_fraction_of_urgent_bandwidth(&context->bw_ctx.dml, pipes, pipe_cnt) * 1000;
447	context->bw_ctx.bw.dcn.watermarks.b.frac_urg_bw_flip = get_fraction_of_urgent_bandwidth_imm_flip(&context->bw_ctx.dml, pipes, pipe_cnt) * 1000;
448	context->bw_ctx.bw.dcn.watermarks.b.urgent_latency_ns = get_urgent_latency(&context->bw_ctx.dml, pipes, pipe_cnt) * 1000;
449
450	pipes[0].clks_cfg.voltage = vlevel;
451	pipes[0].clks_cfg.dcfclk_mhz = dcfclk;
452
453	/* Set D:
454	 * DCFCLK: Min Required
455	 * FCLK(proportional to UCLK): 1GHz or Max
456	 * MALL stutter, sr_enter_exit = 4, sr_exit = 2us
457	 */
458	/*
459	if (dc->clk_mgr->bw_params->wm_table.nv_entries[WM_D].valid) {
460		context->bw_ctx.dml.soc.dram_clock_change_latency_us = dc->clk_mgr->bw_params->wm_table.nv_entries[WM_D].dml_input.pstate_latency_us;
461		context->bw_ctx.dml.soc.sr_enter_plus_exit_time_us = dc->clk_mgr->bw_params->wm_table.nv_entries[WM_D].dml_input.sr_enter_plus_exit_time_us;
462		context->bw_ctx.dml.soc.sr_exit_time_us = dc->clk_mgr->bw_params->wm_table.nv_entries[WM_D].dml_input.sr_exit_time_us;
463	}
464	context->bw_ctx.bw.dcn.watermarks.d.urgent_ns = get_wm_urgent(&context->bw_ctx.dml, pipes, pipe_cnt) * 1000;
465	context->bw_ctx.bw.dcn.watermarks.d.cstate_pstate.cstate_enter_plus_exit_ns = get_wm_stutter_enter_exit(&context->bw_ctx.dml, pipes, pipe_cnt) * 1000;
466	context->bw_ctx.bw.dcn.watermarks.d.cstate_pstate.cstate_exit_ns = get_wm_stutter_exit(&context->bw_ctx.dml, pipes, pipe_cnt) * 1000;
467	context->bw_ctx.bw.dcn.watermarks.d.cstate_pstate.pstate_change_ns = get_wm_dram_clock_change(&context->bw_ctx.dml, pipes, pipe_cnt) * 1000;
468	context->bw_ctx.bw.dcn.watermarks.d.pte_meta_urgent_ns = get_wm_memory_trip(&context->bw_ctx.dml, pipes, pipe_cnt) * 1000;
469	context->bw_ctx.bw.dcn.watermarks.d.frac_urg_bw_nom = get_fraction_of_urgent_bandwidth(&context->bw_ctx.dml, pipes, pipe_cnt) * 1000;
470	context->bw_ctx.bw.dcn.watermarks.d.frac_urg_bw_flip = get_fraction_of_urgent_bandwidth_imm_flip(&context->bw_ctx.dml, pipes, pipe_cnt) * 1000;
471	context->bw_ctx.bw.dcn.watermarks.d.urgent_latency_ns = get_urgent_latency(&context->bw_ctx.dml, pipes, pipe_cnt) * 1000;
472	*/
473
474	/* Set C:
475	 * DCFCLK: Min Required
476	 * FCLK(proportional to UCLK): 1GHz or Max
477	 * pstate latency overridden to 5us
478	 */
479	if (dc->clk_mgr->bw_params->wm_table.nv_entries[WM_C].valid) {
480		unsigned int min_dram_speed_mts = context->bw_ctx.dml.vba.DRAMSpeed;
481		unsigned int min_dram_speed_mts_margin = 160;
482
483		context->bw_ctx.dml.soc.dram_clock_change_latency_us =
484			dc->clk_mgr->bw_params->dummy_pstate_table[0].dummy_pstate_latency_us;
485
486		if (context->bw_ctx.dml.vba.DRAMClockChangeSupport[vlevel][maxMpcComb] ==
487			dm_dram_clock_change_unsupported) {
488			int min_dram_speed_mts_offset = dc->clk_mgr->bw_params->clk_table.num_entries - 1;
489
490			min_dram_speed_mts =
491				dc->clk_mgr->bw_params->clk_table.entries[min_dram_speed_mts_offset].memclk_mhz * 16;
492		}
493
494		if (!context->bw_ctx.bw.dcn.clk.fw_based_mclk_switching) {
495			/* find largest table entry that is lower than dram speed,
496			 * but lower than DPM0 still uses DPM0
497			 */
498			for (dummy_latency_index = 3; dummy_latency_index > 0; dummy_latency_index--)
499				if (min_dram_speed_mts + min_dram_speed_mts_margin >
500					dc->clk_mgr->bw_params->dummy_pstate_table[dummy_latency_index].dram_speed_mts)
501					break;
502		}
503
504		context->bw_ctx.dml.soc.dram_clock_change_latency_us =
505			dc->clk_mgr->bw_params->dummy_pstate_table[dummy_latency_index].dummy_pstate_latency_us;
506
507		context->bw_ctx.dml.soc.sr_enter_plus_exit_time_us = dc->clk_mgr->bw_params->wm_table.nv_entries[WM_C].dml_input.sr_enter_plus_exit_time_us;
508		context->bw_ctx.dml.soc.sr_exit_time_us = dc->clk_mgr->bw_params->wm_table.nv_entries[WM_C].dml_input.sr_exit_time_us;
509	}
510
511	context->bw_ctx.bw.dcn.watermarks.c.urgent_ns = get_wm_urgent(&context->bw_ctx.dml, pipes, pipe_cnt) * 1000;
512	context->bw_ctx.bw.dcn.watermarks.c.cstate_pstate.cstate_enter_plus_exit_ns = get_wm_stutter_enter_exit(&context->bw_ctx.dml, pipes, pipe_cnt) * 1000;
513	context->bw_ctx.bw.dcn.watermarks.c.cstate_pstate.cstate_exit_ns = get_wm_stutter_exit(&context->bw_ctx.dml, pipes, pipe_cnt) * 1000;
514	context->bw_ctx.bw.dcn.watermarks.c.cstate_pstate.pstate_change_ns = get_wm_dram_clock_change(&context->bw_ctx.dml, pipes, pipe_cnt) * 1000;
515	context->bw_ctx.bw.dcn.watermarks.c.pte_meta_urgent_ns = get_wm_memory_trip(&context->bw_ctx.dml, pipes, pipe_cnt) * 1000;
516	context->bw_ctx.bw.dcn.watermarks.c.frac_urg_bw_nom = get_fraction_of_urgent_bandwidth(&context->bw_ctx.dml, pipes, pipe_cnt) * 1000;
517	context->bw_ctx.bw.dcn.watermarks.c.frac_urg_bw_flip = get_fraction_of_urgent_bandwidth_imm_flip(&context->bw_ctx.dml, pipes, pipe_cnt) * 1000;
518	context->bw_ctx.bw.dcn.watermarks.c.urgent_latency_ns = get_urgent_latency(&context->bw_ctx.dml, pipes, pipe_cnt) * 1000;
519
520	if (!pstate_en) {
521		/* The only difference between A and C is p-state latency, if p-state is not supported we want to
522		 * calculate DLG based on dummy p-state latency, and max out the set A p-state watermark
523		 */
524		context->bw_ctx.bw.dcn.watermarks.a = context->bw_ctx.bw.dcn.watermarks.c;
525		context->bw_ctx.bw.dcn.watermarks.a.cstate_pstate.pstate_change_ns = 0;
526	} else {
527		/* Set A:
528		 * DCFCLK: Min Required
529		 * FCLK(proportional to UCLK): 1GHz or Max
530		 *
531		 * Set A calculated last so that following calculations are based on Set A
532		 */
533		dc->res_pool->funcs->update_soc_for_wm_a(dc, context);
534		context->bw_ctx.bw.dcn.watermarks.a.urgent_ns = get_wm_urgent(&context->bw_ctx.dml, pipes, pipe_cnt) * 1000;
535		context->bw_ctx.bw.dcn.watermarks.a.cstate_pstate.cstate_enter_plus_exit_ns = get_wm_stutter_enter_exit(&context->bw_ctx.dml, pipes, pipe_cnt) * 1000;
536		context->bw_ctx.bw.dcn.watermarks.a.cstate_pstate.cstate_exit_ns = get_wm_stutter_exit(&context->bw_ctx.dml, pipes, pipe_cnt) * 1000;
537		context->bw_ctx.bw.dcn.watermarks.a.cstate_pstate.pstate_change_ns = get_wm_dram_clock_change(&context->bw_ctx.dml, pipes, pipe_cnt) * 1000;
538		context->bw_ctx.bw.dcn.watermarks.a.pte_meta_urgent_ns = get_wm_memory_trip(&context->bw_ctx.dml, pipes, pipe_cnt) * 1000;
539		context->bw_ctx.bw.dcn.watermarks.a.frac_urg_bw_nom = get_fraction_of_urgent_bandwidth(&context->bw_ctx.dml, pipes, pipe_cnt) * 1000;
540		context->bw_ctx.bw.dcn.watermarks.a.frac_urg_bw_flip = get_fraction_of_urgent_bandwidth_imm_flip(&context->bw_ctx.dml, pipes, pipe_cnt) * 1000;
541		context->bw_ctx.bw.dcn.watermarks.a.urgent_latency_ns = get_urgent_latency(&context->bw_ctx.dml, pipes, pipe_cnt) * 1000;
542	}
543
544	context->perf_params.stutter_period_us = context->bw_ctx.dml.vba.StutterPeriod;
545
546	/* Make set D = set A until set D is enabled */
547	context->bw_ctx.bw.dcn.watermarks.d = context->bw_ctx.bw.dcn.watermarks.a;
548
549	for (i = 0, pipe_idx = 0; i < dc->res_pool->pipe_count; i++) {
550		if (!context->res_ctx.pipe_ctx[i].stream)
551			continue;
552
553		pipes[pipe_idx].clks_cfg.dispclk_mhz = get_dispclk_calculated(&context->bw_ctx.dml, pipes, pipe_cnt);
554		pipes[pipe_idx].clks_cfg.dppclk_mhz = get_dppclk_calculated(&context->bw_ctx.dml, pipes, pipe_cnt, pipe_idx);
555
556		if (dc->config.forced_clocks) {
557			pipes[pipe_idx].clks_cfg.dispclk_mhz = context->bw_ctx.dml.soc.clock_limits[0].dispclk_mhz;
558			pipes[pipe_idx].clks_cfg.dppclk_mhz = context->bw_ctx.dml.soc.clock_limits[0].dppclk_mhz;
559		}
560		if (dc->debug.min_disp_clk_khz > pipes[pipe_idx].clks_cfg.dispclk_mhz * 1000)
561			pipes[pipe_idx].clks_cfg.dispclk_mhz = dc->debug.min_disp_clk_khz / 1000.0;
562		if (dc->debug.min_dpp_clk_khz > pipes[pipe_idx].clks_cfg.dppclk_mhz * 1000)
563			pipes[pipe_idx].clks_cfg.dppclk_mhz = dc->debug.min_dpp_clk_khz / 1000.0;
564
565		pipe_idx++;
566	}
567
568	// WA: restrict FPO to use first non-strobe mode (NV24 BW issue)
569	if (context->bw_ctx.bw.dcn.clk.fw_based_mclk_switching &&
570			dc->dml.soc.num_chans <= 4 &&
571			context->bw_ctx.dml.vba.DRAMSpeed <= 1700 &&
572			context->bw_ctx.dml.vba.DRAMSpeed >= 1500) {
573
574		for (i = 0; i < dc->dml.soc.num_states; i++) {
575			if (dc->dml.soc.clock_limits[i].dram_speed_mts > 1700) {
576				context->bw_ctx.dml.vba.DRAMSpeed = dc->dml.soc.clock_limits[i].dram_speed_mts;
577				break;
578			}
579		}
580	}
581
582	dcn20_calculate_dlg_params(dc, context, pipes, pipe_cnt, vlevel);
583
584	if (!pstate_en)
585		/* Restore full p-state latency */
586		context->bw_ctx.dml.soc.dram_clock_change_latency_us =
587				dc->clk_mgr->bw_params->wm_table.nv_entries[WM_A].dml_input.pstate_latency_us;
588
589	if (context->bw_ctx.bw.dcn.clk.fw_based_mclk_switching)
590		dcn30_setup_mclk_switch_using_fw_based_vblank_stretch(dc, context);
591}
592
593void dcn30_fpu_update_dram_channel_width_bytes(struct dc *dc)
594{
595	dc_assert_fp_enabled();
596
597	if (dc->ctx->dc_bios->vram_info.dram_channel_width_bytes)
598		dcn3_0_soc.dram_channel_width_bytes = dc->ctx->dc_bios->vram_info.dram_channel_width_bytes;
599}
600
601void dcn30_fpu_update_max_clk(struct dc_bounding_box_max_clk *dcn30_bb_max_clk)
602{
603		dc_assert_fp_enabled();
604
605		if (!dcn30_bb_max_clk->max_dcfclk_mhz)
606			dcn30_bb_max_clk->max_dcfclk_mhz = dcn3_0_soc.clock_limits[0].dcfclk_mhz;
607		if (!dcn30_bb_max_clk->max_dispclk_mhz)
608			dcn30_bb_max_clk->max_dispclk_mhz = dcn3_0_soc.clock_limits[0].dispclk_mhz;
609		if (!dcn30_bb_max_clk->max_dppclk_mhz)
610			dcn30_bb_max_clk->max_dppclk_mhz = dcn3_0_soc.clock_limits[0].dppclk_mhz;
611		if (!dcn30_bb_max_clk->max_phyclk_mhz)
612			dcn30_bb_max_clk->max_phyclk_mhz = dcn3_0_soc.clock_limits[0].phyclk_mhz;
613}
614
615void dcn30_fpu_get_optimal_dcfclk_fclk_for_uclk(unsigned int uclk_mts,
616		unsigned int *optimal_dcfclk,
617		unsigned int *optimal_fclk)
618{
619	double bw_from_dram, bw_from_dram1, bw_from_dram2;
620
621	dc_assert_fp_enabled();
622
623	bw_from_dram1 = uclk_mts * dcn3_0_soc.num_chans *
624		dcn3_0_soc.dram_channel_width_bytes * (dcn3_0_soc.max_avg_dram_bw_use_normal_percent / 100);
625	bw_from_dram2 = uclk_mts * dcn3_0_soc.num_chans *
626		dcn3_0_soc.dram_channel_width_bytes * (dcn3_0_soc.max_avg_sdp_bw_use_normal_percent / 100);
627
628	bw_from_dram = (bw_from_dram1 < bw_from_dram2) ? bw_from_dram1 : bw_from_dram2;
629
630	if (optimal_fclk)
631		*optimal_fclk = bw_from_dram /
632		(dcn3_0_soc.fabric_datapath_to_dcn_data_return_bytes * (dcn3_0_soc.max_avg_sdp_bw_use_normal_percent / 100));
633
634	if (optimal_dcfclk)
635		*optimal_dcfclk =  bw_from_dram /
636		(dcn3_0_soc.return_bus_width_bytes * (dcn3_0_soc.max_avg_sdp_bw_use_normal_percent / 100));
637}
638
639void dcn30_fpu_update_bw_bounding_box(struct dc *dc,
640	struct clk_bw_params *bw_params,
641	struct dc_bounding_box_max_clk *dcn30_bb_max_clk,
642	unsigned int *dcfclk_mhz,
643	unsigned int *dram_speed_mts)
644{
645	unsigned int i;
646
647	dc_assert_fp_enabled();
648
649	dcn3_0_soc.dispclk_dppclk_vco_speed_mhz = dc->clk_mgr->dentist_vco_freq_khz / 1000.0;
650	dc->dml.soc.dispclk_dppclk_vco_speed_mhz = dc->clk_mgr->dentist_vco_freq_khz / 1000.0;
651
652	for (i = 0; i < dcn3_0_soc.num_states; i++) {
653		dcn3_0_soc.clock_limits[i].state = i;
654		dcn3_0_soc.clock_limits[i].dcfclk_mhz = dcfclk_mhz[i];
655		dcn3_0_soc.clock_limits[i].fabricclk_mhz = dcfclk_mhz[i];
656		dcn3_0_soc.clock_limits[i].dram_speed_mts = dram_speed_mts[i];
657
658		/* Fill all states with max values of all other clocks */
659		dcn3_0_soc.clock_limits[i].dispclk_mhz = dcn30_bb_max_clk->max_dispclk_mhz;
660		dcn3_0_soc.clock_limits[i].dppclk_mhz  = dcn30_bb_max_clk->max_dppclk_mhz;
661		dcn3_0_soc.clock_limits[i].phyclk_mhz  = dcn30_bb_max_clk->max_phyclk_mhz;
662		dcn3_0_soc.clock_limits[i].dtbclk_mhz = dcn3_0_soc.clock_limits[0].dtbclk_mhz;
663		/* These clocks cannot come from bw_params, always fill from dcn3_0_soc[1] */
664		/* FCLK, PHYCLK_D18, SOCCLK, DSCCLK */
665		dcn3_0_soc.clock_limits[i].phyclk_d18_mhz = dcn3_0_soc.clock_limits[0].phyclk_d18_mhz;
666		dcn3_0_soc.clock_limits[i].socclk_mhz = dcn3_0_soc.clock_limits[0].socclk_mhz;
667		dcn3_0_soc.clock_limits[i].dscclk_mhz = dcn3_0_soc.clock_limits[0].dscclk_mhz;
668	}
669	/* re-init DML with updated bb */
670	dml_init_instance(&dc->dml, &dcn3_0_soc, &dcn3_0_ip, DML_PROJECT_DCN30);
671	if (dc->current_state)
672		dml_init_instance(&dc->current_state->bw_ctx.dml, &dcn3_0_soc, &dcn3_0_ip, DML_PROJECT_DCN30);
673
674}
675
676/**
677 * dcn30_find_dummy_latency_index_for_fw_based_mclk_switch() - Finds
678 * dummy_latency_index when MCLK switching using firmware based vblank stretch
679 * is enabled. This function will iterate through the table of dummy pstate
680 * latencies until the lowest value that allows
681 * dm_allow_self_refresh_and_mclk_switch to happen is found
682 *
683 * @dc: Current DC state
684 * @context: new dc state
685 * @pipes: DML pipe params
686 * @pipe_cnt: number of DML pipes
687 * @vlevel: Voltage level calculated by DML
688 *
689 * Return: lowest dummy_latency_index value
690 */
691int dcn30_find_dummy_latency_index_for_fw_based_mclk_switch(struct dc *dc,
692							    struct dc_state *context,
693							    display_e2e_pipe_params_st *pipes,
694							    int pipe_cnt,
695							    int vlevel)
696{
697	const int max_latency_table_entries = 4;
698	int dummy_latency_index = 0;
699
700	dc_assert_fp_enabled();
701
702	while (dummy_latency_index < max_latency_table_entries) {
703		context->bw_ctx.dml.soc.dram_clock_change_latency_us =
704				dc->clk_mgr->bw_params->dummy_pstate_table[dummy_latency_index].dummy_pstate_latency_us;
705		dcn30_internal_validate_bw(dc, context, pipes, &pipe_cnt, &vlevel, false, true);
706
707		if (context->bw_ctx.dml.soc.allow_dram_self_refresh_or_dram_clock_change_in_vblank ==
708			dm_allow_self_refresh_and_mclk_switch)
709			break;
710
711		dummy_latency_index++;
712	}
713
714	if (dummy_latency_index == max_latency_table_entries) {
715		ASSERT(dummy_latency_index != max_latency_table_entries);
716		/* If the execution gets here, it means dummy p_states are
717		 * not possible. This should never happen and would mean
718		 * something is severely wrong.
719		 * Here we reset dummy_latency_index to 3, because it is
720		 * better to have underflows than system crashes.
721		 */
722		dummy_latency_index = 3;
723	}
724
725	return dummy_latency_index;
726}
727
728void dcn3_fpu_build_wm_range_table(struct clk_mgr *base)
729{
730	/* defaults */
731	double pstate_latency_us = base->ctx->dc->dml.soc.dram_clock_change_latency_us;
732	double sr_exit_time_us = base->ctx->dc->dml.soc.sr_exit_time_us;
733	double sr_enter_plus_exit_time_us = base->ctx->dc->dml.soc.sr_enter_plus_exit_time_us;
734	uint16_t min_uclk_mhz = base->bw_params->clk_table.entries[0].memclk_mhz;
735
736	dc_assert_fp_enabled();
737
738	/* Set A - Normal - default values*/
739	base->bw_params->wm_table.nv_entries[WM_A].valid = true;
740	base->bw_params->wm_table.nv_entries[WM_A].dml_input.pstate_latency_us = pstate_latency_us;
741	base->bw_params->wm_table.nv_entries[WM_A].dml_input.sr_exit_time_us = sr_exit_time_us;
742	base->bw_params->wm_table.nv_entries[WM_A].dml_input.sr_enter_plus_exit_time_us = sr_enter_plus_exit_time_us;
743	base->bw_params->wm_table.nv_entries[WM_A].pmfw_breakdown.wm_type = WATERMARKS_CLOCK_RANGE;
744	base->bw_params->wm_table.nv_entries[WM_A].pmfw_breakdown.min_dcfclk = 0;
745	base->bw_params->wm_table.nv_entries[WM_A].pmfw_breakdown.max_dcfclk = 0xFFFF;
746	base->bw_params->wm_table.nv_entries[WM_A].pmfw_breakdown.min_uclk = min_uclk_mhz;
747	base->bw_params->wm_table.nv_entries[WM_A].pmfw_breakdown.max_uclk = 0xFFFF;
748
749	/* Set B - Performance - higher minimum clocks */
750//	base->bw_params->wm_table.nv_entries[WM_B].valid = true;
751//	base->bw_params->wm_table.nv_entries[WM_B].dml_input.pstate_latency_us = pstate_latency_us;
752//	base->bw_params->wm_table.nv_entries[WM_B].dml_input.sr_exit_time_us = sr_exit_time_us;
753//	base->bw_params->wm_table.nv_entries[WM_B].dml_input.sr_enter_plus_exit_time_us = sr_enter_plus_exit_time_us;
754//	base->bw_params->wm_table.nv_entries[WM_B].pmfw_breakdown.wm_type = WATERMARKS_CLOCK_RANGE;
755//	base->bw_params->wm_table.nv_entries[WM_B].pmfw_breakdown.min_dcfclk = TUNED VALUE;
756//	base->bw_params->wm_table.nv_entries[WM_B].pmfw_breakdown.max_dcfclk = 0xFFFF;
757//	base->bw_params->wm_table.nv_entries[WM_B].pmfw_breakdown.min_uclk = TUNED VALUE;
758//	base->bw_params->wm_table.nv_entries[WM_B].pmfw_breakdown.max_uclk = 0xFFFF;
759
760	/* Set C - Dummy P-State - P-State latency set to "dummy p-state" value */
761	base->bw_params->wm_table.nv_entries[WM_C].valid = true;
762	base->bw_params->wm_table.nv_entries[WM_C].dml_input.pstate_latency_us = 0;
763	base->bw_params->wm_table.nv_entries[WM_C].dml_input.sr_exit_time_us = sr_exit_time_us;
764	base->bw_params->wm_table.nv_entries[WM_C].dml_input.sr_enter_plus_exit_time_us = sr_enter_plus_exit_time_us;
765	base->bw_params->wm_table.nv_entries[WM_C].pmfw_breakdown.wm_type = WATERMARKS_DUMMY_PSTATE;
766	base->bw_params->wm_table.nv_entries[WM_C].pmfw_breakdown.min_dcfclk = 0;
767	base->bw_params->wm_table.nv_entries[WM_C].pmfw_breakdown.max_dcfclk = 0xFFFF;
768	base->bw_params->wm_table.nv_entries[WM_C].pmfw_breakdown.min_uclk = min_uclk_mhz;
769	base->bw_params->wm_table.nv_entries[WM_C].pmfw_breakdown.max_uclk = 0xFFFF;
770	base->bw_params->dummy_pstate_table[0].dram_speed_mts = 1600;
771	base->bw_params->dummy_pstate_table[0].dummy_pstate_latency_us = 38;
772	base->bw_params->dummy_pstate_table[1].dram_speed_mts = 8000;
773	base->bw_params->dummy_pstate_table[1].dummy_pstate_latency_us = 9;
774	base->bw_params->dummy_pstate_table[2].dram_speed_mts = 10000;
775	base->bw_params->dummy_pstate_table[2].dummy_pstate_latency_us = 8;
776	base->bw_params->dummy_pstate_table[3].dram_speed_mts = 16000;
777	base->bw_params->dummy_pstate_table[3].dummy_pstate_latency_us = 5;
778
779	/* Set D - MALL - SR enter and exit times adjusted for MALL */
780	base->bw_params->wm_table.nv_entries[WM_D].valid = true;
781	base->bw_params->wm_table.nv_entries[WM_D].dml_input.pstate_latency_us = pstate_latency_us;
782	base->bw_params->wm_table.nv_entries[WM_D].dml_input.sr_exit_time_us = 2;
783	base->bw_params->wm_table.nv_entries[WM_D].dml_input.sr_enter_plus_exit_time_us = 4;
784	base->bw_params->wm_table.nv_entries[WM_D].pmfw_breakdown.wm_type = WATERMARKS_MALL;
785	base->bw_params->wm_table.nv_entries[WM_D].pmfw_breakdown.min_dcfclk = 0;
786	base->bw_params->wm_table.nv_entries[WM_D].pmfw_breakdown.max_dcfclk = 0xFFFF;
787	base->bw_params->wm_table.nv_entries[WM_D].pmfw_breakdown.min_uclk = min_uclk_mhz;
788	base->bw_params->wm_table.nv_entries[WM_D].pmfw_breakdown.max_uclk = 0xFFFF;
789}
790
791void patch_dcn30_soc_bounding_box(struct dc *dc, struct _vcs_dpi_soc_bounding_box_st *dcn3_0_ip)
792{
793	dc_assert_fp_enabled();
794
795	if (dc->ctx->dc_bios->funcs->get_soc_bb_info) {
796		struct bp_soc_bb_info bb_info = {0};
797
798		if (dc->ctx->dc_bios->funcs->get_soc_bb_info(dc->ctx->dc_bios, &bb_info) == BP_RESULT_OK) {
799			if (bb_info.dram_clock_change_latency_100ns > 0)
800				dcn3_0_soc.dram_clock_change_latency_us = bb_info.dram_clock_change_latency_100ns * 10;
801
802			if (bb_info.dram_sr_enter_exit_latency_100ns > 0)
803				dcn3_0_soc.sr_enter_plus_exit_time_us = bb_info.dram_sr_enter_exit_latency_100ns * 10;
804
805			if (bb_info.dram_sr_exit_latency_100ns > 0)
806				dcn3_0_soc.sr_exit_time_us = bb_info.dram_sr_exit_latency_100ns * 10;
807		}
808	}
809}
810