1/*
2 * Copyright 2019 Advanced Micro Devices, Inc.
3 *
4 * Permission is hereby granted, free of charge, to any person obtaining a
5 * copy of this software and associated documentation files (the "Software"),
6 * to deal in the Software without restriction, including without limitation
7 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
8 * and/or sell copies of the Software, and to permit persons to whom the
9 * Software is furnished to do so, subject to the following conditions:
10 *
11 * The above copyright notice and this permission notice shall be included in
12 * all copies or substantial portions of the Software.
13 *
14 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
15 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
16 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
17 * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
18 * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
19 * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
20 * OTHER DEALINGS IN THE SOFTWARE.
21 *
22 */
23
24#define SWSMU_CODE_LAYER_L2
25
26#include "amdgpu.h"
27#include "amdgpu_smu.h"
28#include "smu_v12_0_ppsmc.h"
29#include "smu12_driver_if.h"
30#include "smu_v12_0.h"
31#include "renoir_ppt.h"
32#include "smu_cmn.h"
33
34/*
35 * DO NOT use these for err/warn/info/debug messages.
36 * Use dev_err, dev_warn, dev_info and dev_dbg instead.
37 * They are more MGPU friendly.
38 */
39#undef pr_err
40#undef pr_warn
41#undef pr_info
42#undef pr_debug
43
44#define mmMP1_SMN_C2PMSG_66                                                                            0x0282
45#define mmMP1_SMN_C2PMSG_66_BASE_IDX                                                                   0
46
47#define mmMP1_SMN_C2PMSG_82                                                                            0x0292
48#define mmMP1_SMN_C2PMSG_82_BASE_IDX                                                                   0
49
50#define mmMP1_SMN_C2PMSG_90                                                                            0x029a
51#define mmMP1_SMN_C2PMSG_90_BASE_IDX                                                                   0
52
53static struct cmn2asic_msg_mapping renoir_message_map[SMU_MSG_MAX_COUNT] = {
54	MSG_MAP(TestMessage,                    PPSMC_MSG_TestMessage,                  1),
55	MSG_MAP(GetSmuVersion,                  PPSMC_MSG_GetSmuVersion,                1),
56	MSG_MAP(GetDriverIfVersion,             PPSMC_MSG_GetDriverIfVersion,           1),
57	MSG_MAP(PowerUpGfx,                     PPSMC_MSG_PowerUpGfx,                   1),
58	MSG_MAP(AllowGfxOff,                    PPSMC_MSG_EnableGfxOff,                 1),
59	MSG_MAP(DisallowGfxOff,                 PPSMC_MSG_DisableGfxOff,                1),
60	MSG_MAP(PowerDownIspByTile,             PPSMC_MSG_PowerDownIspByTile,           1),
61	MSG_MAP(PowerUpIspByTile,               PPSMC_MSG_PowerUpIspByTile,             1),
62	MSG_MAP(PowerDownVcn,                   PPSMC_MSG_PowerDownVcn,                 1),
63	MSG_MAP(PowerUpVcn,                     PPSMC_MSG_PowerUpVcn,                   1),
64	MSG_MAP(PowerDownSdma,                  PPSMC_MSG_PowerDownSdma,                1),
65	MSG_MAP(PowerUpSdma,                    PPSMC_MSG_PowerUpSdma,                  1),
66	MSG_MAP(SetHardMinIspclkByFreq,         PPSMC_MSG_SetHardMinIspclkByFreq,       1),
67	MSG_MAP(SetHardMinVcn,                  PPSMC_MSG_SetHardMinVcn,                1),
68	MSG_MAP(SetAllowFclkSwitch,             PPSMC_MSG_SetAllowFclkSwitch,           1),
69	MSG_MAP(SetMinVideoGfxclkFreq,          PPSMC_MSG_SetMinVideoGfxclkFreq,        1),
70	MSG_MAP(ActiveProcessNotify,            PPSMC_MSG_ActiveProcessNotify,          1),
71	MSG_MAP(SetCustomPolicy,                PPSMC_MSG_SetCustomPolicy,              1),
72	MSG_MAP(SetVideoFps,                    PPSMC_MSG_SetVideoFps,                  1),
73	MSG_MAP(NumOfDisplays,                  PPSMC_MSG_SetDisplayCount,              1),
74	MSG_MAP(QueryPowerLimit,                PPSMC_MSG_QueryPowerLimit,              1),
75	MSG_MAP(SetDriverDramAddrHigh,          PPSMC_MSG_SetDriverDramAddrHigh,        1),
76	MSG_MAP(SetDriverDramAddrLow,           PPSMC_MSG_SetDriverDramAddrLow,         1),
77	MSG_MAP(TransferTableSmu2Dram,          PPSMC_MSG_TransferTableSmu2Dram,        1),
78	MSG_MAP(TransferTableDram2Smu,          PPSMC_MSG_TransferTableDram2Smu,        1),
79	MSG_MAP(GfxDeviceDriverReset,           PPSMC_MSG_GfxDeviceDriverReset,         1),
80	MSG_MAP(SetGfxclkOverdriveByFreqVid,    PPSMC_MSG_SetGfxclkOverdriveByFreqVid,  1),
81	MSG_MAP(SetHardMinDcfclkByFreq,         PPSMC_MSG_SetHardMinDcfclkByFreq,       1),
82	MSG_MAP(SetHardMinSocclkByFreq,         PPSMC_MSG_SetHardMinSocclkByFreq,       1),
83	MSG_MAP(ControlIgpuATS,                 PPSMC_MSG_ControlIgpuATS,               1),
84	MSG_MAP(SetMinVideoFclkFreq,            PPSMC_MSG_SetMinVideoFclkFreq,          1),
85	MSG_MAP(SetMinDeepSleepDcfclk,          PPSMC_MSG_SetMinDeepSleepDcfclk,        1),
86	MSG_MAP(ForcePowerDownGfx,              PPSMC_MSG_ForcePowerDownGfx,            1),
87	MSG_MAP(SetPhyclkVoltageByFreq,         PPSMC_MSG_SetPhyclkVoltageByFreq,       1),
88	MSG_MAP(SetDppclkVoltageByFreq,         PPSMC_MSG_SetDppclkVoltageByFreq,       1),
89	MSG_MAP(SetSoftMinVcn,                  PPSMC_MSG_SetSoftMinVcn,                1),
90	MSG_MAP(EnablePostCode,                 PPSMC_MSG_EnablePostCode,               1),
91	MSG_MAP(GetGfxclkFrequency,             PPSMC_MSG_GetGfxclkFrequency,           1),
92	MSG_MAP(GetFclkFrequency,               PPSMC_MSG_GetFclkFrequency,             1),
93	MSG_MAP(GetMinGfxclkFrequency,          PPSMC_MSG_GetMinGfxclkFrequency,        1),
94	MSG_MAP(GetMaxGfxclkFrequency,          PPSMC_MSG_GetMaxGfxclkFrequency,        1),
95	MSG_MAP(SoftReset,                      PPSMC_MSG_SoftReset,                    1),
96	MSG_MAP(SetGfxCGPG,                     PPSMC_MSG_SetGfxCGPG,                   1),
97	MSG_MAP(SetSoftMaxGfxClk,               PPSMC_MSG_SetSoftMaxGfxClk,             1),
98	MSG_MAP(SetHardMinGfxClk,               PPSMC_MSG_SetHardMinGfxClk,             1),
99	MSG_MAP(SetSoftMaxSocclkByFreq,         PPSMC_MSG_SetSoftMaxSocclkByFreq,       1),
100	MSG_MAP(SetSoftMaxFclkByFreq,           PPSMC_MSG_SetSoftMaxFclkByFreq,         1),
101	MSG_MAP(SetSoftMaxVcn,                  PPSMC_MSG_SetSoftMaxVcn,                1),
102	MSG_MAP(PowerGateMmHub,                 PPSMC_MSG_PowerGateMmHub,               1),
103	MSG_MAP(UpdatePmeRestore,               PPSMC_MSG_UpdatePmeRestore,             1),
104	MSG_MAP(GpuChangeState,                 PPSMC_MSG_GpuChangeState,               1),
105	MSG_MAP(SetPowerLimitPercentage,        PPSMC_MSG_SetPowerLimitPercentage,      1),
106	MSG_MAP(ForceGfxContentSave,            PPSMC_MSG_ForceGfxContentSave,          1),
107	MSG_MAP(EnableTmdp48MHzRefclkPwrDown,   PPSMC_MSG_EnableTmdp48MHzRefclkPwrDown, 1),
108	MSG_MAP(PowerDownJpeg,                  PPSMC_MSG_PowerDownJpeg,                1),
109	MSG_MAP(PowerUpJpeg,                    PPSMC_MSG_PowerUpJpeg,                  1),
110	MSG_MAP(PowerGateAtHub,                 PPSMC_MSG_PowerGateAtHub,               1),
111	MSG_MAP(SetSoftMinJpeg,                 PPSMC_MSG_SetSoftMinJpeg,               1),
112	MSG_MAP(SetHardMinFclkByFreq,           PPSMC_MSG_SetHardMinFclkByFreq,         1),
113};
114
115static struct cmn2asic_mapping renoir_clk_map[SMU_CLK_COUNT] = {
116	CLK_MAP(GFXCLK, CLOCK_GFXCLK),
117	CLK_MAP(SCLK,	CLOCK_GFXCLK),
118	CLK_MAP(SOCCLK, CLOCK_SOCCLK),
119	CLK_MAP(UCLK, CLOCK_FCLK),
120	CLK_MAP(MCLK, CLOCK_FCLK),
121	CLK_MAP(VCLK, CLOCK_VCLK),
122	CLK_MAP(DCLK, CLOCK_DCLK),
123};
124
125static struct cmn2asic_mapping renoir_table_map[SMU_TABLE_COUNT] = {
126	TAB_MAP_VALID(WATERMARKS),
127	TAB_MAP_INVALID(CUSTOM_DPM),
128	TAB_MAP_VALID(DPMCLOCKS),
129	TAB_MAP_VALID(SMU_METRICS),
130};
131
132static struct cmn2asic_mapping renoir_workload_map[PP_SMC_POWER_PROFILE_COUNT] = {
133	WORKLOAD_MAP(PP_SMC_POWER_PROFILE_FULLSCREEN3D,		WORKLOAD_PPLIB_FULL_SCREEN_3D_BIT),
134	WORKLOAD_MAP(PP_SMC_POWER_PROFILE_VIDEO,		WORKLOAD_PPLIB_VIDEO_BIT),
135	WORKLOAD_MAP(PP_SMC_POWER_PROFILE_VR,			WORKLOAD_PPLIB_VR_BIT),
136	WORKLOAD_MAP(PP_SMC_POWER_PROFILE_COMPUTE,		WORKLOAD_PPLIB_COMPUTE_BIT),
137	WORKLOAD_MAP(PP_SMC_POWER_PROFILE_CUSTOM,		WORKLOAD_PPLIB_CUSTOM_BIT),
138};
139
140static const uint8_t renoir_throttler_map[] = {
141	[THROTTLER_STATUS_BIT_SPL]		= (SMU_THROTTLER_SPL_BIT),
142	[THROTTLER_STATUS_BIT_FPPT]		= (SMU_THROTTLER_FPPT_BIT),
143	[THROTTLER_STATUS_BIT_SPPT]		= (SMU_THROTTLER_SPPT_BIT),
144	[THROTTLER_STATUS_BIT_SPPT_APU]		= (SMU_THROTTLER_SPPT_APU_BIT),
145	[THROTTLER_STATUS_BIT_THM_CORE]		= (SMU_THROTTLER_TEMP_CORE_BIT),
146	[THROTTLER_STATUS_BIT_THM_GFX]		= (SMU_THROTTLER_TEMP_GPU_BIT),
147	[THROTTLER_STATUS_BIT_THM_SOC]		= (SMU_THROTTLER_TEMP_SOC_BIT),
148	[THROTTLER_STATUS_BIT_TDC_VDD]		= (SMU_THROTTLER_TDC_VDD_BIT),
149	[THROTTLER_STATUS_BIT_TDC_SOC]		= (SMU_THROTTLER_TDC_SOC_BIT),
150	[THROTTLER_STATUS_BIT_PROCHOT_CPU]	= (SMU_THROTTLER_PROCHOT_CPU_BIT),
151	[THROTTLER_STATUS_BIT_PROCHOT_GFX]	= (SMU_THROTTLER_PROCHOT_GFX_BIT),
152	[THROTTLER_STATUS_BIT_EDC_CPU]		= (SMU_THROTTLER_EDC_CPU_BIT),
153	[THROTTLER_STATUS_BIT_EDC_GFX]		= (SMU_THROTTLER_EDC_GFX_BIT),
154};
155
156static int renoir_init_smc_tables(struct smu_context *smu)
157{
158	struct smu_table_context *smu_table = &smu->smu_table;
159	struct smu_table *tables = smu_table->tables;
160
161	SMU_TABLE_INIT(tables, SMU_TABLE_WATERMARKS, sizeof(Watermarks_t),
162		PAGE_SIZE, AMDGPU_GEM_DOMAIN_VRAM);
163	SMU_TABLE_INIT(tables, SMU_TABLE_DPMCLOCKS, sizeof(DpmClocks_t),
164		PAGE_SIZE, AMDGPU_GEM_DOMAIN_VRAM);
165	SMU_TABLE_INIT(tables, SMU_TABLE_SMU_METRICS, sizeof(SmuMetrics_t),
166		PAGE_SIZE, AMDGPU_GEM_DOMAIN_VRAM);
167
168	smu_table->clocks_table = kzalloc(sizeof(DpmClocks_t), GFP_KERNEL);
169	if (!smu_table->clocks_table)
170		goto err0_out;
171
172	smu_table->metrics_table = kzalloc(sizeof(SmuMetrics_t), GFP_KERNEL);
173	if (!smu_table->metrics_table)
174		goto err1_out;
175	smu_table->metrics_time = 0;
176
177	smu_table->watermarks_table = kzalloc(sizeof(Watermarks_t), GFP_KERNEL);
178	if (!smu_table->watermarks_table)
179		goto err2_out;
180
181	smu_table->gpu_metrics_table_size = sizeof(struct gpu_metrics_v2_2);
182	smu_table->gpu_metrics_table = kzalloc(smu_table->gpu_metrics_table_size, GFP_KERNEL);
183	if (!smu_table->gpu_metrics_table)
184		goto err3_out;
185
186	return 0;
187
188err3_out:
189	kfree(smu_table->watermarks_table);
190err2_out:
191	kfree(smu_table->metrics_table);
192err1_out:
193	kfree(smu_table->clocks_table);
194err0_out:
195	return -ENOMEM;
196}
197
198/*
199 * This interface just for getting uclk ultimate freq and should't introduce
200 * other likewise function result in overmuch callback.
201 */
202static int renoir_get_dpm_clk_limited(struct smu_context *smu, enum smu_clk_type clk_type,
203						uint32_t dpm_level, uint32_t *freq)
204{
205	DpmClocks_t *clk_table = smu->smu_table.clocks_table;
206
207	if (!clk_table || clk_type >= SMU_CLK_COUNT)
208		return -EINVAL;
209
210	switch (clk_type) {
211	case SMU_SOCCLK:
212		if (dpm_level >= NUM_SOCCLK_DPM_LEVELS)
213			return -EINVAL;
214		*freq = clk_table->SocClocks[dpm_level].Freq;
215		break;
216	case SMU_UCLK:
217	case SMU_MCLK:
218		if (dpm_level >= NUM_FCLK_DPM_LEVELS)
219			return -EINVAL;
220		*freq = clk_table->FClocks[dpm_level].Freq;
221		break;
222	case SMU_DCEFCLK:
223		if (dpm_level >= NUM_DCFCLK_DPM_LEVELS)
224			return -EINVAL;
225		*freq = clk_table->DcfClocks[dpm_level].Freq;
226		break;
227	case SMU_FCLK:
228		if (dpm_level >= NUM_FCLK_DPM_LEVELS)
229			return -EINVAL;
230		*freq = clk_table->FClocks[dpm_level].Freq;
231		break;
232	case SMU_VCLK:
233		if (dpm_level >= NUM_VCN_DPM_LEVELS)
234			return -EINVAL;
235		*freq = clk_table->VClocks[dpm_level].Freq;
236		break;
237	case SMU_DCLK:
238		if (dpm_level >= NUM_VCN_DPM_LEVELS)
239			return -EINVAL;
240		*freq = clk_table->DClocks[dpm_level].Freq;
241		break;
242
243	default:
244		return -EINVAL;
245	}
246
247	return 0;
248}
249
250static int renoir_get_profiling_clk_mask(struct smu_context *smu,
251					 enum amd_dpm_forced_level level,
252					 uint32_t *sclk_mask,
253					 uint32_t *mclk_mask,
254					 uint32_t *soc_mask)
255{
256
257	if (level == AMD_DPM_FORCED_LEVEL_PROFILE_MIN_SCLK) {
258		if (sclk_mask)
259			*sclk_mask = 0;
260	} else if (level == AMD_DPM_FORCED_LEVEL_PROFILE_MIN_MCLK) {
261		if (mclk_mask)
262			/* mclk levels are in reverse order */
263			*mclk_mask = NUM_MEMCLK_DPM_LEVELS - 1;
264	} else if (level == AMD_DPM_FORCED_LEVEL_PROFILE_PEAK) {
265		if (sclk_mask)
266			/* The sclk as gfxclk and has three level about max/min/current */
267			*sclk_mask = 3 - 1;
268
269		if (mclk_mask)
270			/* mclk levels are in reverse order */
271			*mclk_mask = 0;
272
273		if (soc_mask)
274			*soc_mask = NUM_SOCCLK_DPM_LEVELS - 1;
275	}
276
277	return 0;
278}
279
280static int renoir_get_dpm_ultimate_freq(struct smu_context *smu,
281					enum smu_clk_type clk_type,
282					uint32_t *min,
283					uint32_t *max)
284{
285	int ret = 0;
286	uint32_t mclk_mask, soc_mask;
287	uint32_t clock_limit;
288
289	if (!smu_cmn_clk_dpm_is_enabled(smu, clk_type)) {
290		switch (clk_type) {
291		case SMU_MCLK:
292		case SMU_UCLK:
293			clock_limit = smu->smu_table.boot_values.uclk;
294			break;
295		case SMU_GFXCLK:
296		case SMU_SCLK:
297			clock_limit = smu->smu_table.boot_values.gfxclk;
298			break;
299		case SMU_SOCCLK:
300			clock_limit = smu->smu_table.boot_values.socclk;
301			break;
302		default:
303			clock_limit = 0;
304			break;
305		}
306
307		/* clock in Mhz unit */
308		if (min)
309			*min = clock_limit / 100;
310		if (max)
311			*max = clock_limit / 100;
312
313		return 0;
314	}
315
316	if (max) {
317		ret = renoir_get_profiling_clk_mask(smu,
318						    AMD_DPM_FORCED_LEVEL_PROFILE_PEAK,
319						    NULL,
320						    &mclk_mask,
321						    &soc_mask);
322		if (ret)
323			goto failed;
324
325		switch (clk_type) {
326		case SMU_GFXCLK:
327		case SMU_SCLK:
328			ret = smu_cmn_send_smc_msg(smu, SMU_MSG_GetMaxGfxclkFrequency, max);
329			if (ret) {
330				dev_err(smu->adev->dev, "Attempt to get max GX frequency from SMC Failed !\n");
331				goto failed;
332			}
333			break;
334		case SMU_UCLK:
335		case SMU_FCLK:
336		case SMU_MCLK:
337			ret = renoir_get_dpm_clk_limited(smu, clk_type, mclk_mask, max);
338			if (ret)
339				goto failed;
340			break;
341		case SMU_SOCCLK:
342			ret = renoir_get_dpm_clk_limited(smu, clk_type, soc_mask, max);
343			if (ret)
344				goto failed;
345			break;
346		default:
347			ret = -EINVAL;
348			goto failed;
349		}
350	}
351
352	if (min) {
353		switch (clk_type) {
354		case SMU_GFXCLK:
355		case SMU_SCLK:
356			ret = smu_cmn_send_smc_msg(smu, SMU_MSG_GetMinGfxclkFrequency, min);
357			if (ret) {
358				dev_err(smu->adev->dev, "Attempt to get min GX frequency from SMC Failed !\n");
359				goto failed;
360			}
361			break;
362		case SMU_UCLK:
363		case SMU_FCLK:
364		case SMU_MCLK:
365			ret = renoir_get_dpm_clk_limited(smu, clk_type, NUM_MEMCLK_DPM_LEVELS - 1, min);
366			if (ret)
367				goto failed;
368			break;
369		case SMU_SOCCLK:
370			ret = renoir_get_dpm_clk_limited(smu, clk_type, 0, min);
371			if (ret)
372				goto failed;
373			break;
374		default:
375			ret = -EINVAL;
376			goto failed;
377		}
378	}
379failed:
380	return ret;
381}
382
383static int renoir_od_edit_dpm_table(struct smu_context *smu,
384							enum PP_OD_DPM_TABLE_COMMAND type,
385							long input[], uint32_t size)
386{
387	int ret = 0;
388	struct smu_dpm_context *smu_dpm_ctx = &(smu->smu_dpm);
389
390	if (!(smu_dpm_ctx->dpm_level == AMD_DPM_FORCED_LEVEL_MANUAL)) {
391		dev_warn(smu->adev->dev,
392			"pp_od_clk_voltage is not accessible if power_dpm_force_performance_level is not in manual mode!\n");
393		return -EINVAL;
394	}
395
396	switch (type) {
397	case PP_OD_EDIT_SCLK_VDDC_TABLE:
398		if (size != 2) {
399			dev_err(smu->adev->dev, "Input parameter number not correct\n");
400			return -EINVAL;
401		}
402
403		if (input[0] == 0) {
404			if (input[1] < smu->gfx_default_hard_min_freq) {
405				dev_warn(smu->adev->dev,
406					"Fine grain setting minimum sclk (%ld) MHz is less than the minimum allowed (%d) MHz\n",
407					input[1], smu->gfx_default_hard_min_freq);
408				return -EINVAL;
409			}
410			smu->gfx_actual_hard_min_freq = input[1];
411		} else if (input[0] == 1) {
412			if (input[1] > smu->gfx_default_soft_max_freq) {
413				dev_warn(smu->adev->dev,
414					"Fine grain setting maximum sclk (%ld) MHz is greater than the maximum allowed (%d) MHz\n",
415					input[1], smu->gfx_default_soft_max_freq);
416				return -EINVAL;
417			}
418			smu->gfx_actual_soft_max_freq = input[1];
419		} else {
420			return -EINVAL;
421		}
422		break;
423	case PP_OD_RESTORE_DEFAULT_TABLE:
424		if (size != 0) {
425			dev_err(smu->adev->dev, "Input parameter number not correct\n");
426			return -EINVAL;
427		}
428		smu->gfx_actual_hard_min_freq = smu->gfx_default_hard_min_freq;
429		smu->gfx_actual_soft_max_freq = smu->gfx_default_soft_max_freq;
430		break;
431	case PP_OD_COMMIT_DPM_TABLE:
432		if (size != 0) {
433			dev_err(smu->adev->dev, "Input parameter number not correct\n");
434			return -EINVAL;
435		} else {
436			if (smu->gfx_actual_hard_min_freq > smu->gfx_actual_soft_max_freq) {
437				dev_err(smu->adev->dev,
438					"The setting minimum sclk (%d) MHz is greater than the setting maximum sclk (%d) MHz\n",
439					smu->gfx_actual_hard_min_freq,
440					smu->gfx_actual_soft_max_freq);
441				return -EINVAL;
442			}
443
444			ret = smu_cmn_send_smc_msg_with_param(smu,
445								SMU_MSG_SetHardMinGfxClk,
446								smu->gfx_actual_hard_min_freq,
447								NULL);
448			if (ret) {
449				dev_err(smu->adev->dev, "Set hard min sclk failed!");
450				return ret;
451			}
452
453			ret = smu_cmn_send_smc_msg_with_param(smu,
454								SMU_MSG_SetSoftMaxGfxClk,
455								smu->gfx_actual_soft_max_freq,
456								NULL);
457			if (ret) {
458				dev_err(smu->adev->dev, "Set soft max sclk failed!");
459				return ret;
460			}
461		}
462		break;
463	default:
464		return -ENOSYS;
465	}
466
467	return ret;
468}
469
470static int renoir_set_fine_grain_gfx_freq_parameters(struct smu_context *smu)
471{
472	uint32_t min = 0, max = 0;
473	uint32_t ret = 0;
474
475	ret = smu_cmn_send_smc_msg_with_param(smu,
476								SMU_MSG_GetMinGfxclkFrequency,
477								0, &min);
478	if (ret)
479		return ret;
480	ret = smu_cmn_send_smc_msg_with_param(smu,
481								SMU_MSG_GetMaxGfxclkFrequency,
482								0, &max);
483	if (ret)
484		return ret;
485
486	smu->gfx_default_hard_min_freq = min;
487	smu->gfx_default_soft_max_freq = max;
488	smu->gfx_actual_hard_min_freq = 0;
489	smu->gfx_actual_soft_max_freq = 0;
490
491	return 0;
492}
493
494static int renoir_print_clk_levels(struct smu_context *smu,
495			enum smu_clk_type clk_type, char *buf)
496{
497	int i, idx, size = 0, ret = 0;
498	uint32_t cur_value = 0, value = 0, count = 0, min = 0, max = 0;
499	SmuMetrics_t metrics;
500	struct smu_dpm_context *smu_dpm_ctx = &(smu->smu_dpm);
501	bool cur_value_match_level = false;
502
503	memset(&metrics, 0, sizeof(metrics));
504
505	ret = smu_cmn_get_metrics_table(smu, &metrics, false);
506	if (ret)
507		return ret;
508
509	smu_cmn_get_sysfs_buf(&buf, &size);
510
511	switch (clk_type) {
512	case SMU_OD_RANGE:
513		if (smu_dpm_ctx->dpm_level == AMD_DPM_FORCED_LEVEL_MANUAL) {
514			ret = smu_cmn_send_smc_msg_with_param(smu,
515						SMU_MSG_GetMinGfxclkFrequency,
516						0, &min);
517			if (ret)
518				return ret;
519			ret = smu_cmn_send_smc_msg_with_param(smu,
520						SMU_MSG_GetMaxGfxclkFrequency,
521						0, &max);
522			if (ret)
523				return ret;
524			size += sysfs_emit_at(buf, size, "OD_RANGE\nSCLK: %10uMhz %10uMhz\n", min, max);
525		}
526		break;
527	case SMU_OD_SCLK:
528		if (smu_dpm_ctx->dpm_level == AMD_DPM_FORCED_LEVEL_MANUAL) {
529			min = (smu->gfx_actual_hard_min_freq > 0) ? smu->gfx_actual_hard_min_freq : smu->gfx_default_hard_min_freq;
530			max = (smu->gfx_actual_soft_max_freq > 0) ? smu->gfx_actual_soft_max_freq : smu->gfx_default_soft_max_freq;
531			size += sysfs_emit_at(buf, size, "OD_SCLK\n");
532			size += sysfs_emit_at(buf, size, "0:%10uMhz\n", min);
533			size += sysfs_emit_at(buf, size, "1:%10uMhz\n", max);
534		}
535		break;
536	case SMU_GFXCLK:
537	case SMU_SCLK:
538		/* retirve table returned paramters unit is MHz */
539		cur_value = metrics.ClockFrequency[CLOCK_GFXCLK];
540		ret = renoir_get_dpm_ultimate_freq(smu, SMU_GFXCLK, &min, &max);
541		if (!ret) {
542			/* driver only know min/max gfx_clk, Add level 1 for all other gfx clks */
543			if (cur_value  == max)
544				i = 2;
545			else if (cur_value == min)
546				i = 0;
547			else
548				i = 1;
549
550			size += sysfs_emit_at(buf, size, "0: %uMhz %s\n", min,
551					i == 0 ? "*" : "");
552			size += sysfs_emit_at(buf, size, "1: %uMhz %s\n",
553					i == 1 ? cur_value : RENOIR_UMD_PSTATE_GFXCLK,
554					i == 1 ? "*" : "");
555			size += sysfs_emit_at(buf, size, "2: %uMhz %s\n", max,
556					i == 2 ? "*" : "");
557		}
558		return size;
559	case SMU_SOCCLK:
560		count = NUM_SOCCLK_DPM_LEVELS;
561		cur_value = metrics.ClockFrequency[CLOCK_SOCCLK];
562		break;
563	case SMU_MCLK:
564		count = NUM_MEMCLK_DPM_LEVELS;
565		cur_value = metrics.ClockFrequency[CLOCK_FCLK];
566		break;
567	case SMU_DCEFCLK:
568		count = NUM_DCFCLK_DPM_LEVELS;
569		cur_value = metrics.ClockFrequency[CLOCK_DCFCLK];
570		break;
571	case SMU_FCLK:
572		count = NUM_FCLK_DPM_LEVELS;
573		cur_value = metrics.ClockFrequency[CLOCK_FCLK];
574		break;
575	case SMU_VCLK:
576		count = NUM_VCN_DPM_LEVELS;
577		cur_value = metrics.ClockFrequency[CLOCK_VCLK];
578		break;
579	case SMU_DCLK:
580		count = NUM_VCN_DPM_LEVELS;
581		cur_value = metrics.ClockFrequency[CLOCK_DCLK];
582		break;
583	default:
584		break;
585	}
586
587	switch (clk_type) {
588	case SMU_GFXCLK:
589	case SMU_SCLK:
590	case SMU_SOCCLK:
591	case SMU_MCLK:
592	case SMU_DCEFCLK:
593	case SMU_FCLK:
594	case SMU_VCLK:
595	case SMU_DCLK:
596		for (i = 0; i < count; i++) {
597			idx = (clk_type == SMU_FCLK || clk_type == SMU_MCLK) ? (count - i - 1) : i;
598			ret = renoir_get_dpm_clk_limited(smu, clk_type, idx, &value);
599			if (ret)
600				return ret;
601			if (!value)
602				continue;
603			size += sysfs_emit_at(buf, size, "%d: %uMhz %s\n", i, value,
604					cur_value == value ? "*" : "");
605			if (cur_value == value)
606				cur_value_match_level = true;
607		}
608
609		if (!cur_value_match_level)
610			size += sysfs_emit_at(buf, size, "   %uMhz *\n", cur_value);
611
612		break;
613	default:
614		break;
615	}
616
617	return size;
618}
619
620static enum amd_pm_state_type renoir_get_current_power_state(struct smu_context *smu)
621{
622	enum amd_pm_state_type pm_type;
623	struct smu_dpm_context *smu_dpm_ctx = &(smu->smu_dpm);
624
625	if (!smu_dpm_ctx->dpm_context ||
626	    !smu_dpm_ctx->dpm_current_power_state)
627		return -EINVAL;
628
629	switch (smu_dpm_ctx->dpm_current_power_state->classification.ui_label) {
630	case SMU_STATE_UI_LABEL_BATTERY:
631		pm_type = POWER_STATE_TYPE_BATTERY;
632		break;
633	case SMU_STATE_UI_LABEL_BALLANCED:
634		pm_type = POWER_STATE_TYPE_BALANCED;
635		break;
636	case SMU_STATE_UI_LABEL_PERFORMANCE:
637		pm_type = POWER_STATE_TYPE_PERFORMANCE;
638		break;
639	default:
640		if (smu_dpm_ctx->dpm_current_power_state->classification.flags & SMU_STATE_CLASSIFICATION_FLAG_BOOT)
641			pm_type = POWER_STATE_TYPE_INTERNAL_BOOT;
642		else
643			pm_type = POWER_STATE_TYPE_DEFAULT;
644		break;
645	}
646
647	return pm_type;
648}
649
650static int renoir_dpm_set_vcn_enable(struct smu_context *smu, bool enable)
651{
652	int ret = 0;
653
654	if (enable) {
655		/* vcn dpm on is a prerequisite for vcn power gate messages */
656		if (smu_cmn_feature_is_enabled(smu, SMU_FEATURE_VCN_PG_BIT)) {
657			ret = smu_cmn_send_smc_msg_with_param(smu, SMU_MSG_PowerUpVcn, 0, NULL);
658			if (ret)
659				return ret;
660		}
661	} else {
662		if (smu_cmn_feature_is_enabled(smu, SMU_FEATURE_VCN_PG_BIT)) {
663			ret = smu_cmn_send_smc_msg(smu, SMU_MSG_PowerDownVcn, NULL);
664			if (ret)
665				return ret;
666		}
667	}
668
669	return ret;
670}
671
672static int renoir_dpm_set_jpeg_enable(struct smu_context *smu, bool enable)
673{
674	int ret = 0;
675
676	if (enable) {
677		if (smu_cmn_feature_is_enabled(smu, SMU_FEATURE_JPEG_PG_BIT)) {
678			ret = smu_cmn_send_smc_msg_with_param(smu, SMU_MSG_PowerUpJpeg, 0, NULL);
679			if (ret)
680				return ret;
681		}
682	} else {
683		if (smu_cmn_feature_is_enabled(smu, SMU_FEATURE_JPEG_PG_BIT)) {
684			ret = smu_cmn_send_smc_msg_with_param(smu, SMU_MSG_PowerDownJpeg, 0, NULL);
685			if (ret)
686				return ret;
687		}
688	}
689
690	return ret;
691}
692
693static int renoir_force_dpm_limit_value(struct smu_context *smu, bool highest)
694{
695	int ret = 0, i = 0;
696	uint32_t min_freq, max_freq, force_freq;
697	enum smu_clk_type clk_type;
698
699	enum smu_clk_type clks[] = {
700		SMU_GFXCLK,
701		SMU_MCLK,
702		SMU_SOCCLK,
703	};
704
705	for (i = 0; i < ARRAY_SIZE(clks); i++) {
706		clk_type = clks[i];
707		ret = renoir_get_dpm_ultimate_freq(smu, clk_type, &min_freq, &max_freq);
708		if (ret)
709			return ret;
710
711		force_freq = highest ? max_freq : min_freq;
712		ret = smu_v12_0_set_soft_freq_limited_range(smu, clk_type, force_freq, force_freq);
713		if (ret)
714			return ret;
715	}
716
717	return ret;
718}
719
720static int renoir_unforce_dpm_levels(struct smu_context *smu) {
721
722	int ret = 0, i = 0;
723	uint32_t min_freq, max_freq;
724	enum smu_clk_type clk_type;
725
726	struct clk_feature_map {
727		enum smu_clk_type clk_type;
728		uint32_t	feature;
729	} clk_feature_map[] = {
730		{SMU_GFXCLK, SMU_FEATURE_DPM_GFXCLK_BIT},
731		{SMU_MCLK,   SMU_FEATURE_DPM_UCLK_BIT},
732		{SMU_SOCCLK, SMU_FEATURE_DPM_SOCCLK_BIT},
733	};
734
735	for (i = 0; i < ARRAY_SIZE(clk_feature_map); i++) {
736		if (!smu_cmn_feature_is_enabled(smu, clk_feature_map[i].feature))
737		    continue;
738
739		clk_type = clk_feature_map[i].clk_type;
740
741		ret = renoir_get_dpm_ultimate_freq(smu, clk_type, &min_freq, &max_freq);
742		if (ret)
743			return ret;
744
745		ret = smu_v12_0_set_soft_freq_limited_range(smu, clk_type, min_freq, max_freq);
746		if (ret)
747			return ret;
748	}
749
750	return ret;
751}
752
753/*
754 * This interface get dpm clock table for dc
755 */
756static int renoir_get_dpm_clock_table(struct smu_context *smu, struct dpm_clocks *clock_table)
757{
758	DpmClocks_t *table = smu->smu_table.clocks_table;
759	int i;
760
761	if (!clock_table || !table)
762		return -EINVAL;
763
764	for (i = 0; i < NUM_DCFCLK_DPM_LEVELS; i++) {
765		clock_table->DcfClocks[i].Freq = table->DcfClocks[i].Freq;
766		clock_table->DcfClocks[i].Vol = table->DcfClocks[i].Vol;
767	}
768
769	for (i = 0; i < NUM_SOCCLK_DPM_LEVELS; i++) {
770		clock_table->SocClocks[i].Freq = table->SocClocks[i].Freq;
771		clock_table->SocClocks[i].Vol = table->SocClocks[i].Vol;
772	}
773
774	for (i = 0; i < NUM_FCLK_DPM_LEVELS; i++) {
775		clock_table->FClocks[i].Freq = table->FClocks[i].Freq;
776		clock_table->FClocks[i].Vol = table->FClocks[i].Vol;
777	}
778
779	for (i = 0; i<  NUM_MEMCLK_DPM_LEVELS; i++) {
780		clock_table->MemClocks[i].Freq = table->MemClocks[i].Freq;
781		clock_table->MemClocks[i].Vol = table->MemClocks[i].Vol;
782	}
783
784	for (i = 0; i < NUM_VCN_DPM_LEVELS; i++) {
785		clock_table->VClocks[i].Freq = table->VClocks[i].Freq;
786		clock_table->VClocks[i].Vol = table->VClocks[i].Vol;
787	}
788
789	for (i = 0; i < NUM_VCN_DPM_LEVELS; i++) {
790		clock_table->DClocks[i].Freq = table->DClocks[i].Freq;
791		clock_table->DClocks[i].Vol = table->DClocks[i].Vol;
792	}
793
794	return 0;
795}
796
797static int renoir_force_clk_levels(struct smu_context *smu,
798				   enum smu_clk_type clk_type, uint32_t mask)
799{
800
801	int ret = 0 ;
802	uint32_t soft_min_level = 0, soft_max_level = 0, min_freq = 0, max_freq = 0;
803
804	soft_min_level = mask ? (ffs(mask) - 1) : 0;
805	soft_max_level = mask ? (fls(mask) - 1) : 0;
806
807	switch (clk_type) {
808	case SMU_GFXCLK:
809	case SMU_SCLK:
810		if (soft_min_level > 2 || soft_max_level > 2) {
811			dev_info(smu->adev->dev, "Currently sclk only support 3 levels on APU\n");
812			return -EINVAL;
813		}
814
815		ret = renoir_get_dpm_ultimate_freq(smu, SMU_GFXCLK, &min_freq, &max_freq);
816		if (ret)
817			return ret;
818		ret = smu_cmn_send_smc_msg_with_param(smu, SMU_MSG_SetSoftMaxGfxClk,
819					soft_max_level == 0 ? min_freq :
820					soft_max_level == 1 ? RENOIR_UMD_PSTATE_GFXCLK : max_freq,
821					NULL);
822		if (ret)
823			return ret;
824		ret = smu_cmn_send_smc_msg_with_param(smu, SMU_MSG_SetHardMinGfxClk,
825					soft_min_level == 2 ? max_freq :
826					soft_min_level == 1 ? RENOIR_UMD_PSTATE_GFXCLK : min_freq,
827					NULL);
828		if (ret)
829			return ret;
830		break;
831	case SMU_SOCCLK:
832		ret = renoir_get_dpm_clk_limited(smu, clk_type, soft_min_level, &min_freq);
833		if (ret)
834			return ret;
835		ret = renoir_get_dpm_clk_limited(smu, clk_type, soft_max_level, &max_freq);
836		if (ret)
837			return ret;
838		ret = smu_cmn_send_smc_msg_with_param(smu, SMU_MSG_SetSoftMaxSocclkByFreq, max_freq, NULL);
839		if (ret)
840			return ret;
841		ret = smu_cmn_send_smc_msg_with_param(smu, SMU_MSG_SetHardMinSocclkByFreq, min_freq, NULL);
842		if (ret)
843			return ret;
844		break;
845	case SMU_MCLK:
846	case SMU_FCLK:
847		ret = renoir_get_dpm_clk_limited(smu, clk_type, soft_min_level, &min_freq);
848		if (ret)
849			return ret;
850		ret = renoir_get_dpm_clk_limited(smu, clk_type, soft_max_level, &max_freq);
851		if (ret)
852			return ret;
853		ret = smu_cmn_send_smc_msg_with_param(smu, SMU_MSG_SetSoftMaxFclkByFreq, max_freq, NULL);
854		if (ret)
855			return ret;
856		ret = smu_cmn_send_smc_msg_with_param(smu, SMU_MSG_SetHardMinFclkByFreq, min_freq, NULL);
857		if (ret)
858			return ret;
859		break;
860	default:
861		break;
862	}
863
864	return ret;
865}
866
867static int renoir_set_power_profile_mode(struct smu_context *smu, long *input, uint32_t size)
868{
869	int workload_type, ret;
870	uint32_t profile_mode = input[size];
871
872	if (profile_mode > PP_SMC_POWER_PROFILE_CUSTOM) {
873		dev_err(smu->adev->dev, "Invalid power profile mode %d\n", profile_mode);
874		return -EINVAL;
875	}
876
877	if (profile_mode == PP_SMC_POWER_PROFILE_BOOTUP_DEFAULT ||
878			profile_mode == PP_SMC_POWER_PROFILE_POWERSAVING)
879		return 0;
880
881	/* conv PP_SMC_POWER_PROFILE* to WORKLOAD_PPLIB_*_BIT */
882	workload_type = smu_cmn_to_asic_specific_index(smu,
883						       CMN2ASIC_MAPPING_WORKLOAD,
884						       profile_mode);
885	if (workload_type < 0) {
886		/*
887		 * TODO: If some case need switch to powersave/default power mode
888		 * then can consider enter WORKLOAD_COMPUTE/WORKLOAD_CUSTOM for power saving.
889		 */
890		dev_dbg(smu->adev->dev, "Unsupported power profile mode %d on RENOIR\n", profile_mode);
891		return -EINVAL;
892	}
893
894	ret = smu_cmn_send_smc_msg_with_param(smu, SMU_MSG_ActiveProcessNotify,
895				    1 << workload_type,
896				    NULL);
897	if (ret) {
898		dev_err_once(smu->adev->dev, "Fail to set workload type %d\n", workload_type);
899		return ret;
900	}
901
902	smu->power_profile_mode = profile_mode;
903
904	return 0;
905}
906
907static int renoir_set_peak_clock_by_device(struct smu_context *smu)
908{
909	int ret = 0;
910	uint32_t sclk_freq = 0, uclk_freq = 0;
911
912	ret = renoir_get_dpm_ultimate_freq(smu, SMU_SCLK, NULL, &sclk_freq);
913	if (ret)
914		return ret;
915
916	ret = smu_v12_0_set_soft_freq_limited_range(smu, SMU_SCLK, sclk_freq, sclk_freq);
917	if (ret)
918		return ret;
919
920	ret = renoir_get_dpm_ultimate_freq(smu, SMU_UCLK, NULL, &uclk_freq);
921	if (ret)
922		return ret;
923
924	ret = smu_v12_0_set_soft_freq_limited_range(smu, SMU_UCLK, uclk_freq, uclk_freq);
925	if (ret)
926		return ret;
927
928	return ret;
929}
930
931static int renoir_set_performance_level(struct smu_context *smu,
932					enum amd_dpm_forced_level level)
933{
934	int ret = 0;
935	uint32_t sclk_mask, mclk_mask, soc_mask;
936
937	switch (level) {
938	case AMD_DPM_FORCED_LEVEL_HIGH:
939		smu->gfx_actual_hard_min_freq = smu->gfx_default_hard_min_freq;
940		smu->gfx_actual_soft_max_freq = smu->gfx_default_soft_max_freq;
941
942		ret = renoir_force_dpm_limit_value(smu, true);
943		break;
944	case AMD_DPM_FORCED_LEVEL_LOW:
945		smu->gfx_actual_hard_min_freq = smu->gfx_default_hard_min_freq;
946		smu->gfx_actual_soft_max_freq = smu->gfx_default_soft_max_freq;
947
948		ret = renoir_force_dpm_limit_value(smu, false);
949		break;
950	case AMD_DPM_FORCED_LEVEL_AUTO:
951		smu->gfx_actual_hard_min_freq = smu->gfx_default_hard_min_freq;
952		smu->gfx_actual_soft_max_freq = smu->gfx_default_soft_max_freq;
953
954		ret = renoir_unforce_dpm_levels(smu);
955		break;
956	case AMD_DPM_FORCED_LEVEL_PROFILE_STANDARD:
957		smu->gfx_actual_hard_min_freq = smu->gfx_default_hard_min_freq;
958		smu->gfx_actual_soft_max_freq = smu->gfx_default_soft_max_freq;
959
960		ret = smu_cmn_send_smc_msg_with_param(smu,
961						      SMU_MSG_SetHardMinGfxClk,
962						      RENOIR_UMD_PSTATE_GFXCLK,
963						      NULL);
964		if (ret)
965			return ret;
966		ret = smu_cmn_send_smc_msg_with_param(smu,
967						      SMU_MSG_SetHardMinFclkByFreq,
968						      RENOIR_UMD_PSTATE_FCLK,
969						      NULL);
970		if (ret)
971			return ret;
972		ret = smu_cmn_send_smc_msg_with_param(smu,
973						      SMU_MSG_SetHardMinSocclkByFreq,
974						      RENOIR_UMD_PSTATE_SOCCLK,
975						      NULL);
976		if (ret)
977			return ret;
978		ret = smu_cmn_send_smc_msg_with_param(smu,
979						      SMU_MSG_SetHardMinVcn,
980						      RENOIR_UMD_PSTATE_VCNCLK,
981						      NULL);
982		if (ret)
983			return ret;
984
985		ret = smu_cmn_send_smc_msg_with_param(smu,
986						      SMU_MSG_SetSoftMaxGfxClk,
987						      RENOIR_UMD_PSTATE_GFXCLK,
988						      NULL);
989		if (ret)
990			return ret;
991		ret = smu_cmn_send_smc_msg_with_param(smu,
992						      SMU_MSG_SetSoftMaxFclkByFreq,
993						      RENOIR_UMD_PSTATE_FCLK,
994						      NULL);
995		if (ret)
996			return ret;
997		ret = smu_cmn_send_smc_msg_with_param(smu,
998						      SMU_MSG_SetSoftMaxSocclkByFreq,
999						      RENOIR_UMD_PSTATE_SOCCLK,
1000						      NULL);
1001		if (ret)
1002			return ret;
1003		ret = smu_cmn_send_smc_msg_with_param(smu,
1004						      SMU_MSG_SetSoftMaxVcn,
1005						      RENOIR_UMD_PSTATE_VCNCLK,
1006						      NULL);
1007		if (ret)
1008			return ret;
1009		break;
1010	case AMD_DPM_FORCED_LEVEL_PROFILE_MIN_SCLK:
1011	case AMD_DPM_FORCED_LEVEL_PROFILE_MIN_MCLK:
1012		smu->gfx_actual_hard_min_freq = smu->gfx_default_hard_min_freq;
1013		smu->gfx_actual_soft_max_freq = smu->gfx_default_soft_max_freq;
1014
1015		ret = renoir_get_profiling_clk_mask(smu, level,
1016						    &sclk_mask,
1017						    &mclk_mask,
1018						    &soc_mask);
1019		if (ret)
1020			return ret;
1021		renoir_force_clk_levels(smu, SMU_SCLK, 1 << sclk_mask);
1022		renoir_force_clk_levels(smu, SMU_MCLK, 1 << mclk_mask);
1023		renoir_force_clk_levels(smu, SMU_SOCCLK, 1 << soc_mask);
1024		break;
1025	case AMD_DPM_FORCED_LEVEL_PROFILE_PEAK:
1026		smu->gfx_actual_hard_min_freq = smu->gfx_default_hard_min_freq;
1027		smu->gfx_actual_soft_max_freq = smu->gfx_default_soft_max_freq;
1028
1029		ret = renoir_set_peak_clock_by_device(smu);
1030		break;
1031	case AMD_DPM_FORCED_LEVEL_MANUAL:
1032	case AMD_DPM_FORCED_LEVEL_PROFILE_EXIT:
1033	default:
1034		break;
1035	}
1036	return ret;
1037}
1038
1039/* save watermark settings into pplib smu structure,
1040 * also pass data to smu controller
1041 */
1042static int renoir_set_watermarks_table(
1043		struct smu_context *smu,
1044		struct pp_smu_wm_range_sets *clock_ranges)
1045{
1046	Watermarks_t *table = smu->smu_table.watermarks_table;
1047	int ret = 0;
1048	int i;
1049
1050	if (clock_ranges) {
1051		if (clock_ranges->num_reader_wm_sets > NUM_WM_RANGES ||
1052		    clock_ranges->num_writer_wm_sets > NUM_WM_RANGES)
1053			return -EINVAL;
1054
1055		/* save into smu->smu_table.tables[SMU_TABLE_WATERMARKS]->cpu_addr*/
1056		for (i = 0; i < clock_ranges->num_reader_wm_sets; i++) {
1057			table->WatermarkRow[WM_DCFCLK][i].MinClock =
1058				clock_ranges->reader_wm_sets[i].min_drain_clk_mhz;
1059			table->WatermarkRow[WM_DCFCLK][i].MaxClock =
1060				clock_ranges->reader_wm_sets[i].max_drain_clk_mhz;
1061			table->WatermarkRow[WM_DCFCLK][i].MinMclk =
1062				clock_ranges->reader_wm_sets[i].min_fill_clk_mhz;
1063			table->WatermarkRow[WM_DCFCLK][i].MaxMclk =
1064				clock_ranges->reader_wm_sets[i].max_fill_clk_mhz;
1065
1066			table->WatermarkRow[WM_DCFCLK][i].WmSetting =
1067				clock_ranges->reader_wm_sets[i].wm_inst;
1068			table->WatermarkRow[WM_DCFCLK][i].WmType =
1069				clock_ranges->reader_wm_sets[i].wm_type;
1070		}
1071
1072		for (i = 0; i < clock_ranges->num_writer_wm_sets; i++) {
1073			table->WatermarkRow[WM_SOCCLK][i].MinClock =
1074				clock_ranges->writer_wm_sets[i].min_fill_clk_mhz;
1075			table->WatermarkRow[WM_SOCCLK][i].MaxClock =
1076				clock_ranges->writer_wm_sets[i].max_fill_clk_mhz;
1077			table->WatermarkRow[WM_SOCCLK][i].MinMclk =
1078				clock_ranges->writer_wm_sets[i].min_drain_clk_mhz;
1079			table->WatermarkRow[WM_SOCCLK][i].MaxMclk =
1080				clock_ranges->writer_wm_sets[i].max_drain_clk_mhz;
1081
1082			table->WatermarkRow[WM_SOCCLK][i].WmSetting =
1083				clock_ranges->writer_wm_sets[i].wm_inst;
1084			table->WatermarkRow[WM_SOCCLK][i].WmType =
1085				clock_ranges->writer_wm_sets[i].wm_type;
1086		}
1087
1088		smu->watermarks_bitmap |= WATERMARKS_EXIST;
1089	}
1090
1091	/* pass data to smu controller */
1092	if ((smu->watermarks_bitmap & WATERMARKS_EXIST) &&
1093	     !(smu->watermarks_bitmap & WATERMARKS_LOADED)) {
1094		ret = smu_cmn_write_watermarks_table(smu);
1095		if (ret) {
1096			dev_err(smu->adev->dev, "Failed to update WMTABLE!");
1097			return ret;
1098		}
1099		smu->watermarks_bitmap |= WATERMARKS_LOADED;
1100	}
1101
1102	return 0;
1103}
1104
1105static int renoir_get_power_profile_mode(struct smu_context *smu,
1106					   char *buf)
1107{
1108	uint32_t i, size = 0;
1109	int16_t workload_type = 0;
1110
1111	if (!buf)
1112		return -EINVAL;
1113
1114	for (i = 0; i <= PP_SMC_POWER_PROFILE_CUSTOM; i++) {
1115		/*
1116		 * Conv PP_SMC_POWER_PROFILE* to WORKLOAD_PPLIB_*_BIT
1117		 * Not all profile modes are supported on arcturus.
1118		 */
1119		workload_type = smu_cmn_to_asic_specific_index(smu,
1120							       CMN2ASIC_MAPPING_WORKLOAD,
1121							       i);
1122		if (workload_type < 0)
1123			continue;
1124
1125		size += sysfs_emit_at(buf, size, "%2d %14s%s\n",
1126			i, amdgpu_pp_profile_name[i], (i == smu->power_profile_mode) ? "*" : " ");
1127	}
1128
1129	return size;
1130}
1131
1132static void renoir_get_ss_power_percent(SmuMetrics_t *metrics,
1133					uint32_t *apu_percent, uint32_t *dgpu_percent)
1134{
1135	uint32_t apu_boost = 0;
1136	uint32_t dgpu_boost = 0;
1137	uint16_t apu_limit = 0;
1138	uint16_t dgpu_limit = 0;
1139	uint16_t apu_power = 0;
1140	uint16_t dgpu_power = 0;
1141
1142	apu_power = metrics->ApuPower;
1143	apu_limit = metrics->StapmOriginalLimit;
1144	if (apu_power > apu_limit && apu_limit != 0)
1145		apu_boost =  ((apu_power - apu_limit) * 100) / apu_limit;
1146	apu_boost = (apu_boost > 100) ? 100 : apu_boost;
1147
1148	dgpu_power = metrics->dGpuPower;
1149	if (metrics->StapmCurrentLimit > metrics->StapmOriginalLimit)
1150		dgpu_limit = metrics->StapmCurrentLimit - metrics->StapmOriginalLimit;
1151	if (dgpu_power > dgpu_limit && dgpu_limit != 0)
1152		dgpu_boost = ((dgpu_power - dgpu_limit) * 100) / dgpu_limit;
1153	dgpu_boost = (dgpu_boost > 100) ? 100 : dgpu_boost;
1154
1155	if (dgpu_boost >= apu_boost)
1156		apu_boost = 0;
1157	else
1158		dgpu_boost = 0;
1159
1160	*apu_percent = apu_boost;
1161	*dgpu_percent = dgpu_boost;
1162}
1163
1164
1165static int renoir_get_smu_metrics_data(struct smu_context *smu,
1166				       MetricsMember_t member,
1167				       uint32_t *value)
1168{
1169	struct smu_table_context *smu_table = &smu->smu_table;
1170
1171	SmuMetrics_t *metrics = (SmuMetrics_t *)smu_table->metrics_table;
1172	int ret = 0;
1173	uint32_t apu_percent = 0;
1174	uint32_t dgpu_percent = 0;
1175	struct amdgpu_device *adev = smu->adev;
1176
1177
1178	ret = smu_cmn_get_metrics_table(smu,
1179					NULL,
1180					false);
1181	if (ret)
1182		return ret;
1183
1184	switch (member) {
1185	case METRICS_AVERAGE_GFXCLK:
1186		*value = metrics->ClockFrequency[CLOCK_GFXCLK];
1187		break;
1188	case METRICS_AVERAGE_SOCCLK:
1189		*value = metrics->ClockFrequency[CLOCK_SOCCLK];
1190		break;
1191	case METRICS_AVERAGE_UCLK:
1192		*value = metrics->ClockFrequency[CLOCK_FCLK];
1193		break;
1194	case METRICS_AVERAGE_GFXACTIVITY:
1195		*value = metrics->AverageGfxActivity / 100;
1196		break;
1197	case METRICS_AVERAGE_VCNACTIVITY:
1198		*value = metrics->AverageUvdActivity / 100;
1199		break;
1200	case METRICS_CURR_SOCKETPOWER:
1201		if (((amdgpu_ip_version(adev, MP1_HWIP, 0) ==
1202		      IP_VERSION(12, 0, 1)) &&
1203		     (adev->pm.fw_version >= 0x40000f)) ||
1204		    ((amdgpu_ip_version(adev, MP1_HWIP, 0) ==
1205		      IP_VERSION(12, 0, 0)) &&
1206		     (adev->pm.fw_version >= 0x373200)))
1207			*value = metrics->CurrentSocketPower << 8;
1208		else
1209			*value = (metrics->CurrentSocketPower << 8) / 1000;
1210		break;
1211	case METRICS_TEMPERATURE_EDGE:
1212		*value = (metrics->GfxTemperature / 100) *
1213			SMU_TEMPERATURE_UNITS_PER_CENTIGRADES;
1214		break;
1215	case METRICS_TEMPERATURE_HOTSPOT:
1216		*value = (metrics->SocTemperature / 100) *
1217			SMU_TEMPERATURE_UNITS_PER_CENTIGRADES;
1218		break;
1219	case METRICS_THROTTLER_STATUS:
1220		*value = metrics->ThrottlerStatus;
1221		break;
1222	case METRICS_VOLTAGE_VDDGFX:
1223		*value = metrics->Voltage[0];
1224		break;
1225	case METRICS_VOLTAGE_VDDSOC:
1226		*value = metrics->Voltage[1];
1227		break;
1228	case METRICS_SS_APU_SHARE:
1229		/* return the percentage of APU power boost
1230		 * with respect to APU's power limit.
1231		 */
1232		renoir_get_ss_power_percent(metrics, &apu_percent, &dgpu_percent);
1233		*value = apu_percent;
1234		break;
1235	case METRICS_SS_DGPU_SHARE:
1236		/* return the percentage of dGPU power boost
1237		 * with respect to dGPU's power limit.
1238		 */
1239		renoir_get_ss_power_percent(metrics, &apu_percent, &dgpu_percent);
1240		*value = dgpu_percent;
1241		break;
1242	default:
1243		*value = UINT_MAX;
1244		break;
1245	}
1246
1247	return ret;
1248}
1249
1250static int renoir_read_sensor(struct smu_context *smu,
1251				 enum amd_pp_sensors sensor,
1252				 void *data, uint32_t *size)
1253{
1254	int ret = 0;
1255
1256	if (!data || !size)
1257		return -EINVAL;
1258
1259	switch (sensor) {
1260	case AMDGPU_PP_SENSOR_GPU_LOAD:
1261		ret = renoir_get_smu_metrics_data(smu,
1262						  METRICS_AVERAGE_GFXACTIVITY,
1263						  (uint32_t *)data);
1264		*size = 4;
1265		break;
1266	case AMDGPU_PP_SENSOR_EDGE_TEMP:
1267		ret = renoir_get_smu_metrics_data(smu,
1268						  METRICS_TEMPERATURE_EDGE,
1269						  (uint32_t *)data);
1270		*size = 4;
1271		break;
1272	case AMDGPU_PP_SENSOR_HOTSPOT_TEMP:
1273		ret = renoir_get_smu_metrics_data(smu,
1274						  METRICS_TEMPERATURE_HOTSPOT,
1275						  (uint32_t *)data);
1276		*size = 4;
1277		break;
1278	case AMDGPU_PP_SENSOR_GFX_MCLK:
1279		ret = renoir_get_smu_metrics_data(smu,
1280						  METRICS_AVERAGE_UCLK,
1281						  (uint32_t *)data);
1282		*(uint32_t *)data *= 100;
1283		*size = 4;
1284		break;
1285	case AMDGPU_PP_SENSOR_GFX_SCLK:
1286		ret = renoir_get_smu_metrics_data(smu,
1287						  METRICS_AVERAGE_GFXCLK,
1288						  (uint32_t *)data);
1289		*(uint32_t *)data *= 100;
1290		*size = 4;
1291		break;
1292	case AMDGPU_PP_SENSOR_VDDGFX:
1293		ret = renoir_get_smu_metrics_data(smu,
1294						  METRICS_VOLTAGE_VDDGFX,
1295						  (uint32_t *)data);
1296		*size = 4;
1297		break;
1298	case AMDGPU_PP_SENSOR_VDDNB:
1299		ret = renoir_get_smu_metrics_data(smu,
1300						  METRICS_VOLTAGE_VDDSOC,
1301						  (uint32_t *)data);
1302		*size = 4;
1303		break;
1304	case AMDGPU_PP_SENSOR_GPU_INPUT_POWER:
1305		ret = renoir_get_smu_metrics_data(smu,
1306						  METRICS_CURR_SOCKETPOWER,
1307						  (uint32_t *)data);
1308		*size = 4;
1309		break;
1310	case AMDGPU_PP_SENSOR_SS_APU_SHARE:
1311		ret = renoir_get_smu_metrics_data(smu,
1312						  METRICS_SS_APU_SHARE,
1313						  (uint32_t *)data);
1314		*size = 4;
1315		break;
1316	case AMDGPU_PP_SENSOR_SS_DGPU_SHARE:
1317		ret = renoir_get_smu_metrics_data(smu,
1318						  METRICS_SS_DGPU_SHARE,
1319						  (uint32_t *)data);
1320		*size = 4;
1321		break;
1322	case AMDGPU_PP_SENSOR_GPU_AVG_POWER:
1323	default:
1324		ret = -EOPNOTSUPP;
1325		break;
1326	}
1327
1328	return ret;
1329}
1330
1331static bool renoir_is_dpm_running(struct smu_context *smu)
1332{
1333	struct amdgpu_device *adev = smu->adev;
1334
1335	/*
1336	 * Until now, the pmfw hasn't exported the interface of SMU
1337	 * feature mask to APU SKU so just force on all the feature
1338	 * at early initial stage.
1339	 */
1340	if (adev->in_suspend)
1341		return false;
1342	else
1343		return true;
1344
1345}
1346
1347static ssize_t renoir_get_gpu_metrics(struct smu_context *smu,
1348				      void **table)
1349{
1350	struct smu_table_context *smu_table = &smu->smu_table;
1351	struct gpu_metrics_v2_2 *gpu_metrics =
1352		(struct gpu_metrics_v2_2 *)smu_table->gpu_metrics_table;
1353	SmuMetrics_t metrics;
1354	int ret = 0;
1355
1356	ret = smu_cmn_get_metrics_table(smu, &metrics, true);
1357	if (ret)
1358		return ret;
1359
1360	smu_cmn_init_soft_gpu_metrics(gpu_metrics, 2, 2);
1361
1362	gpu_metrics->temperature_gfx = metrics.GfxTemperature;
1363	gpu_metrics->temperature_soc = metrics.SocTemperature;
1364	memcpy(&gpu_metrics->temperature_core[0],
1365		&metrics.CoreTemperature[0],
1366		sizeof(uint16_t) * 8);
1367	gpu_metrics->temperature_l3[0] = metrics.L3Temperature[0];
1368	gpu_metrics->temperature_l3[1] = metrics.L3Temperature[1];
1369
1370	gpu_metrics->average_gfx_activity = metrics.AverageGfxActivity;
1371	gpu_metrics->average_mm_activity = metrics.AverageUvdActivity;
1372
1373	gpu_metrics->average_socket_power = metrics.CurrentSocketPower;
1374	gpu_metrics->average_cpu_power = metrics.Power[0];
1375	gpu_metrics->average_soc_power = metrics.Power[1];
1376	memcpy(&gpu_metrics->average_core_power[0],
1377		&metrics.CorePower[0],
1378		sizeof(uint16_t) * 8);
1379
1380	gpu_metrics->average_gfxclk_frequency = metrics.AverageGfxclkFrequency;
1381	gpu_metrics->average_socclk_frequency = metrics.AverageSocclkFrequency;
1382	gpu_metrics->average_fclk_frequency = metrics.AverageFclkFrequency;
1383	gpu_metrics->average_vclk_frequency = metrics.AverageVclkFrequency;
1384
1385	gpu_metrics->current_gfxclk = metrics.ClockFrequency[CLOCK_GFXCLK];
1386	gpu_metrics->current_socclk = metrics.ClockFrequency[CLOCK_SOCCLK];
1387	gpu_metrics->current_uclk = metrics.ClockFrequency[CLOCK_UMCCLK];
1388	gpu_metrics->current_fclk = metrics.ClockFrequency[CLOCK_FCLK];
1389	gpu_metrics->current_vclk = metrics.ClockFrequency[CLOCK_VCLK];
1390	gpu_metrics->current_dclk = metrics.ClockFrequency[CLOCK_DCLK];
1391	memcpy(&gpu_metrics->current_coreclk[0],
1392		&metrics.CoreFrequency[0],
1393		sizeof(uint16_t) * 8);
1394	gpu_metrics->current_l3clk[0] = metrics.L3Frequency[0];
1395	gpu_metrics->current_l3clk[1] = metrics.L3Frequency[1];
1396
1397	gpu_metrics->throttle_status = metrics.ThrottlerStatus;
1398	gpu_metrics->indep_throttle_status =
1399		smu_cmn_get_indep_throttler_status(metrics.ThrottlerStatus,
1400						   renoir_throttler_map);
1401
1402	gpu_metrics->fan_pwm = metrics.FanPwm;
1403
1404	gpu_metrics->system_clock_counter = ktime_get_boottime_ns();
1405
1406	*table = (void *)gpu_metrics;
1407
1408	return sizeof(struct gpu_metrics_v2_2);
1409}
1410
1411static int renoir_gfx_state_change_set(struct smu_context *smu, uint32_t state)
1412{
1413
1414	return 0;
1415}
1416
1417static int renoir_get_enabled_mask(struct smu_context *smu,
1418				   uint64_t *feature_mask)
1419{
1420	if (!feature_mask)
1421		return -EINVAL;
1422	memset(feature_mask, 0xff, sizeof(*feature_mask));
1423
1424	return 0;
1425}
1426
1427static const struct pptable_funcs renoir_ppt_funcs = {
1428	.set_power_state = NULL,
1429	.print_clk_levels = renoir_print_clk_levels,
1430	.get_current_power_state = renoir_get_current_power_state,
1431	.dpm_set_vcn_enable = renoir_dpm_set_vcn_enable,
1432	.dpm_set_jpeg_enable = renoir_dpm_set_jpeg_enable,
1433	.force_clk_levels = renoir_force_clk_levels,
1434	.set_power_profile_mode = renoir_set_power_profile_mode,
1435	.set_performance_level = renoir_set_performance_level,
1436	.get_dpm_clock_table = renoir_get_dpm_clock_table,
1437	.set_watermarks_table = renoir_set_watermarks_table,
1438	.get_power_profile_mode = renoir_get_power_profile_mode,
1439	.read_sensor = renoir_read_sensor,
1440	.check_fw_status = smu_v12_0_check_fw_status,
1441	.check_fw_version = smu_v12_0_check_fw_version,
1442	.powergate_sdma = smu_v12_0_powergate_sdma,
1443	.send_smc_msg_with_param = smu_cmn_send_smc_msg_with_param,
1444	.send_smc_msg = smu_cmn_send_smc_msg,
1445	.set_gfx_cgpg = smu_v12_0_set_gfx_cgpg,
1446	.gfx_off_control = smu_v12_0_gfx_off_control,
1447	.get_gfx_off_status = smu_v12_0_get_gfxoff_status,
1448	.init_smc_tables = renoir_init_smc_tables,
1449	.fini_smc_tables = smu_v12_0_fini_smc_tables,
1450	.set_default_dpm_table = smu_v12_0_set_default_dpm_tables,
1451	.get_enabled_mask = renoir_get_enabled_mask,
1452	.feature_is_enabled = smu_cmn_feature_is_enabled,
1453	.disable_all_features_with_exception = smu_cmn_disable_all_features_with_exception,
1454	.get_dpm_ultimate_freq = renoir_get_dpm_ultimate_freq,
1455	.mode2_reset = smu_v12_0_mode2_reset,
1456	.set_soft_freq_limited_range = smu_v12_0_set_soft_freq_limited_range,
1457	.set_driver_table_location = smu_v12_0_set_driver_table_location,
1458	.is_dpm_running = renoir_is_dpm_running,
1459	.get_pp_feature_mask = smu_cmn_get_pp_feature_mask,
1460	.set_pp_feature_mask = smu_cmn_set_pp_feature_mask,
1461	.get_gpu_metrics = renoir_get_gpu_metrics,
1462	.gfx_state_change_set = renoir_gfx_state_change_set,
1463	.set_fine_grain_gfx_freq_parameters = renoir_set_fine_grain_gfx_freq_parameters,
1464	.od_edit_dpm_table = renoir_od_edit_dpm_table,
1465	.get_vbios_bootup_values = smu_v12_0_get_vbios_bootup_values,
1466};
1467
1468void renoir_set_ppt_funcs(struct smu_context *smu)
1469{
1470	struct amdgpu_device *adev = smu->adev;
1471
1472	smu->ppt_funcs = &renoir_ppt_funcs;
1473	smu->message_map = renoir_message_map;
1474	smu->clock_map = renoir_clk_map;
1475	smu->table_map = renoir_table_map;
1476	smu->workload_map = renoir_workload_map;
1477	smu->smc_driver_if_version = SMU12_DRIVER_IF_VERSION;
1478	smu->is_apu = true;
1479	smu->param_reg = SOC15_REG_OFFSET(MP1, 0, mmMP1_SMN_C2PMSG_82);
1480	smu->msg_reg = SOC15_REG_OFFSET(MP1, 0, mmMP1_SMN_C2PMSG_66);
1481	smu->resp_reg = SOC15_REG_OFFSET(MP1, 0, mmMP1_SMN_C2PMSG_90);
1482}
1483