1/*
2 * Copyright 2019 Advanced Micro Devices, Inc.
3 *
4 * Permission is hereby granted, free of charge, to any person obtaining a
5 * copy of this software and associated documentation files (the "Software"),
6 * to deal in the Software without restriction, including without limitation
7 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
8 * and/or sell copies of the Software, and to permit persons to whom the
9 * Software is furnished to do so, subject to the following conditions:
10 *
11 * The above copyright notice and this permission notice shall be included in
12 * all copies or substantial portions of the Software.
13 *
14 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
15 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
16 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
17 * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
18 * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
19 * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
20 * OTHER DEALINGS IN THE SOFTWARE.
21 */
22
23#define SWSMU_CODE_LAYER_L3
24
25#include <linux/firmware.h>
26#include "amdgpu.h"
27#include "amdgpu_smu.h"
28#include "atomfirmware.h"
29#include "amdgpu_atomfirmware.h"
30#include "amdgpu_atombios.h"
31#include "smu_v12_0.h"
32#include "soc15_common.h"
33#include "atom.h"
34#include "smu_cmn.h"
35
36#include "asic_reg/mp/mp_12_0_0_offset.h"
37#include "asic_reg/mp/mp_12_0_0_sh_mask.h"
38#include "asic_reg/smuio/smuio_12_0_0_offset.h"
39#include "asic_reg/smuio/smuio_12_0_0_sh_mask.h"
40
41/*
42 * DO NOT use these for err/warn/info/debug messages.
43 * Use dev_err, dev_warn, dev_info and dev_dbg instead.
44 * They are more MGPU friendly.
45 */
46#undef pr_err
47#undef pr_warn
48#undef pr_info
49#undef pr_debug
50
51// because some SMU12 based ASICs use older ip offset tables
52// we should undefine this register from the smuio12 header
53// to prevent confusion down the road
54#undef mmPWR_MISC_CNTL_STATUS
55
56#define smnMP1_FIRMWARE_FLAGS                                0x3010024
57
58int smu_v12_0_check_fw_status(struct smu_context *smu)
59{
60	struct amdgpu_device *adev = smu->adev;
61	uint32_t mp1_fw_flags;
62
63	mp1_fw_flags = RREG32_PCIE(MP1_Public |
64		(smnMP1_FIRMWARE_FLAGS & 0xffffffff));
65
66	if ((mp1_fw_flags & MP1_FIRMWARE_FLAGS__INTERRUPTS_ENABLED_MASK) >>
67		MP1_FIRMWARE_FLAGS__INTERRUPTS_ENABLED__SHIFT)
68		return 0;
69
70	return -EIO;
71}
72
73int smu_v12_0_check_fw_version(struct smu_context *smu)
74{
75	struct amdgpu_device *adev = smu->adev;
76	uint32_t if_version = 0xff, smu_version = 0xff;
77	uint8_t smu_program, smu_major, smu_minor, smu_debug;
78	int ret = 0;
79
80	ret = smu_cmn_get_smc_version(smu, &if_version, &smu_version);
81	if (ret)
82		return ret;
83
84	smu_program = (smu_version >> 24) & 0xff;
85	smu_major = (smu_version >> 16) & 0xff;
86	smu_minor = (smu_version >> 8) & 0xff;
87	smu_debug = (smu_version >> 0) & 0xff;
88	if (smu->is_apu)
89		adev->pm.fw_version = smu_version;
90
91	/*
92	 * 1. if_version mismatch is not critical as our fw is designed
93	 * to be backward compatible.
94	 * 2. New fw usually brings some optimizations. But that's visible
95	 * only on the paired driver.
96	 * Considering above, we just leave user a verbal message instead
97	 * of halt driver loading.
98	 */
99	if (if_version != smu->smc_driver_if_version) {
100		dev_info(smu->adev->dev, "smu driver if version = 0x%08x, smu fw if version = 0x%08x, "
101			"smu fw program = %d, smu fw version = 0x%08x (%d.%d.%d)\n",
102			smu->smc_driver_if_version, if_version,
103			smu_program, smu_version, smu_major, smu_minor, smu_debug);
104		dev_info(smu->adev->dev, "SMU driver if version not matched\n");
105	}
106
107	return ret;
108}
109
110int smu_v12_0_powergate_sdma(struct smu_context *smu, bool gate)
111{
112	if (!smu->is_apu)
113		return 0;
114
115	if (gate)
116		return smu_cmn_send_smc_msg(smu, SMU_MSG_PowerDownSdma, NULL);
117	else
118		return smu_cmn_send_smc_msg(smu, SMU_MSG_PowerUpSdma, NULL);
119}
120
121int smu_v12_0_set_gfx_cgpg(struct smu_context *smu, bool enable)
122{
123	/* Until now the SMU12 only implemented for Renoir series so here neen't do APU check. */
124	if (!(smu->adev->pg_flags & AMD_PG_SUPPORT_GFX_PG) || smu->adev->in_s0ix)
125		return 0;
126
127	return smu_cmn_send_smc_msg_with_param(smu,
128		SMU_MSG_SetGfxCGPG,
129		enable ? 1 : 0,
130		NULL);
131}
132
133/**
134 * smu_v12_0_get_gfxoff_status - get gfxoff status
135 *
136 * @smu: amdgpu_device pointer
137 *
138 * This function will be used to get gfxoff status
139 *
140 * Returns 0=GFXOFF(default).
141 * Returns 1=Transition out of GFX State.
142 * Returns 2=Not in GFXOFF.
143 * Returns 3=Transition into GFXOFF.
144 */
145uint32_t smu_v12_0_get_gfxoff_status(struct smu_context *smu)
146{
147	uint32_t reg;
148	uint32_t gfxOff_Status = 0;
149	struct amdgpu_device *adev = smu->adev;
150
151	reg = RREG32_SOC15(SMUIO, 0, mmSMUIO_GFX_MISC_CNTL);
152	gfxOff_Status = (reg & SMUIO_GFX_MISC_CNTL__PWR_GFXOFF_STATUS_MASK)
153		>> SMUIO_GFX_MISC_CNTL__PWR_GFXOFF_STATUS__SHIFT;
154
155	return gfxOff_Status;
156}
157
158int smu_v12_0_gfx_off_control(struct smu_context *smu, bool enable)
159{
160	int ret = 0, timeout = 500;
161
162	if (enable) {
163		ret = smu_cmn_send_smc_msg(smu, SMU_MSG_AllowGfxOff, NULL);
164
165	} else {
166		ret = smu_cmn_send_smc_msg(smu, SMU_MSG_DisallowGfxOff, NULL);
167
168		/* confirm gfx is back to "on" state, timeout is 0.5 second */
169		while (!(smu_v12_0_get_gfxoff_status(smu) == 2)) {
170			drm_msleep(1);
171			timeout--;
172			if (timeout == 0) {
173				DRM_ERROR("disable gfxoff timeout and failed!\n");
174				break;
175			}
176		}
177	}
178
179	return ret;
180}
181
182int smu_v12_0_fini_smc_tables(struct smu_context *smu)
183{
184	struct smu_table_context *smu_table = &smu->smu_table;
185
186	kfree(smu_table->clocks_table);
187	smu_table->clocks_table = NULL;
188
189	kfree(smu_table->metrics_table);
190	smu_table->metrics_table = NULL;
191
192	kfree(smu_table->watermarks_table);
193	smu_table->watermarks_table = NULL;
194
195	kfree(smu_table->gpu_metrics_table);
196	smu_table->gpu_metrics_table = NULL;
197
198	return 0;
199}
200
201int smu_v12_0_set_default_dpm_tables(struct smu_context *smu)
202{
203	struct smu_table_context *smu_table = &smu->smu_table;
204
205	return smu_cmn_update_table(smu, SMU_TABLE_DPMCLOCKS, 0, smu_table->clocks_table, false);
206}
207
208int smu_v12_0_mode2_reset(struct smu_context *smu)
209{
210	return smu_cmn_send_smc_msg_with_param(smu, SMU_MSG_GfxDeviceDriverReset, SMU_RESET_MODE_2, NULL);
211}
212
213int smu_v12_0_set_soft_freq_limited_range(struct smu_context *smu, enum smu_clk_type clk_type,
214			    uint32_t min, uint32_t max)
215{
216	int ret = 0;
217
218	if (!smu_cmn_clk_dpm_is_enabled(smu, clk_type))
219		return 0;
220
221	switch (clk_type) {
222	case SMU_GFXCLK:
223	case SMU_SCLK:
224		ret = smu_cmn_send_smc_msg_with_param(smu, SMU_MSG_SetHardMinGfxClk, min, NULL);
225		if (ret)
226			return ret;
227
228		ret = smu_cmn_send_smc_msg_with_param(smu, SMU_MSG_SetSoftMaxGfxClk, max, NULL);
229		if (ret)
230			return ret;
231	break;
232	case SMU_FCLK:
233	case SMU_MCLK:
234	case SMU_UCLK:
235		ret = smu_cmn_send_smc_msg_with_param(smu, SMU_MSG_SetHardMinFclkByFreq, min, NULL);
236		if (ret)
237			return ret;
238
239		ret = smu_cmn_send_smc_msg_with_param(smu, SMU_MSG_SetSoftMaxFclkByFreq, max, NULL);
240		if (ret)
241			return ret;
242	break;
243	case SMU_SOCCLK:
244		ret = smu_cmn_send_smc_msg_with_param(smu, SMU_MSG_SetHardMinSocclkByFreq, min, NULL);
245		if (ret)
246			return ret;
247
248		ret = smu_cmn_send_smc_msg_with_param(smu, SMU_MSG_SetSoftMaxSocclkByFreq, max, NULL);
249		if (ret)
250			return ret;
251	break;
252	case SMU_VCLK:
253		ret = smu_cmn_send_smc_msg_with_param(smu, SMU_MSG_SetHardMinVcn, min, NULL);
254		if (ret)
255			return ret;
256
257		ret = smu_cmn_send_smc_msg_with_param(smu, SMU_MSG_SetSoftMaxVcn, max, NULL);
258		if (ret)
259			return ret;
260	break;
261	default:
262		return -EINVAL;
263	}
264
265	return ret;
266}
267
268int smu_v12_0_set_driver_table_location(struct smu_context *smu)
269{
270	struct smu_table *driver_table = &smu->smu_table.driver_table;
271	int ret = 0;
272
273	if (driver_table->mc_address) {
274		ret = smu_cmn_send_smc_msg_with_param(smu,
275				SMU_MSG_SetDriverDramAddrHigh,
276				upper_32_bits(driver_table->mc_address),
277				NULL);
278		if (!ret)
279			ret = smu_cmn_send_smc_msg_with_param(smu,
280				SMU_MSG_SetDriverDramAddrLow,
281				lower_32_bits(driver_table->mc_address),
282				NULL);
283	}
284
285	return ret;
286}
287
288static int smu_v12_0_atom_get_smu_clockinfo(struct amdgpu_device *adev,
289					    uint8_t clk_id,
290					    uint8_t syspll_id,
291					    uint32_t *clk_freq)
292{
293	struct atom_get_smu_clock_info_parameters_v3_1 input = {0};
294	struct atom_get_smu_clock_info_output_parameters_v3_1 *output;
295	int ret, index;
296
297	input.clk_id = clk_id;
298	input.syspll_id = syspll_id;
299	input.command = GET_SMU_CLOCK_INFO_V3_1_GET_CLOCK_FREQ;
300	index = get_index_into_master_table(atom_master_list_of_command_functions_v2_1,
301					    getsmuclockinfo);
302
303	ret = amdgpu_atom_execute_table(adev->mode_info.atom_context, index,
304					(uint32_t *)&input);
305	if (ret)
306		return -EINVAL;
307
308	output = (struct atom_get_smu_clock_info_output_parameters_v3_1 *)&input;
309	*clk_freq = le32_to_cpu(output->atom_smu_outputclkfreq.smu_clock_freq_hz) / 10000;
310
311	return 0;
312}
313
314int smu_v12_0_get_vbios_bootup_values(struct smu_context *smu)
315{
316	int ret, index;
317	uint16_t size;
318	uint8_t frev, crev;
319	struct atom_common_table_header *header;
320	struct atom_firmware_info_v3_1 *v_3_1;
321	struct atom_firmware_info_v3_3 *v_3_3;
322
323	index = get_index_into_master_table(atom_master_list_of_data_tables_v2_1,
324					    firmwareinfo);
325
326	ret = amdgpu_atombios_get_data_table(smu->adev, index, &size, &frev, &crev,
327				      (uint8_t **)&header);
328	if (ret)
329		return ret;
330
331	if (header->format_revision != 3) {
332		dev_err(smu->adev->dev, "unknown atom_firmware_info version! for smu12\n");
333		return -EINVAL;
334	}
335
336	switch (header->content_revision) {
337	case 0:
338	case 1:
339	case 2:
340		v_3_1 = (struct atom_firmware_info_v3_1 *)header;
341		smu->smu_table.boot_values.revision = v_3_1->firmware_revision;
342		smu->smu_table.boot_values.gfxclk = v_3_1->bootup_sclk_in10khz;
343		smu->smu_table.boot_values.uclk = v_3_1->bootup_mclk_in10khz;
344		smu->smu_table.boot_values.socclk = 0;
345		smu->smu_table.boot_values.dcefclk = 0;
346		smu->smu_table.boot_values.vddc = v_3_1->bootup_vddc_mv;
347		smu->smu_table.boot_values.vddci = v_3_1->bootup_vddci_mv;
348		smu->smu_table.boot_values.mvddc = v_3_1->bootup_mvddc_mv;
349		smu->smu_table.boot_values.vdd_gfx = v_3_1->bootup_vddgfx_mv;
350		smu->smu_table.boot_values.cooling_id = v_3_1->coolingsolution_id;
351		smu->smu_table.boot_values.pp_table_id = 0;
352		smu->smu_table.boot_values.firmware_caps = v_3_1->firmware_capability;
353		break;
354	case 3:
355	case 4:
356	default:
357		v_3_3 = (struct atom_firmware_info_v3_3 *)header;
358		smu->smu_table.boot_values.revision = v_3_3->firmware_revision;
359		smu->smu_table.boot_values.gfxclk = v_3_3->bootup_sclk_in10khz;
360		smu->smu_table.boot_values.uclk = v_3_3->bootup_mclk_in10khz;
361		smu->smu_table.boot_values.socclk = 0;
362		smu->smu_table.boot_values.dcefclk = 0;
363		smu->smu_table.boot_values.vddc = v_3_3->bootup_vddc_mv;
364		smu->smu_table.boot_values.vddci = v_3_3->bootup_vddci_mv;
365		smu->smu_table.boot_values.mvddc = v_3_3->bootup_mvddc_mv;
366		smu->smu_table.boot_values.vdd_gfx = v_3_3->bootup_vddgfx_mv;
367		smu->smu_table.boot_values.cooling_id = v_3_3->coolingsolution_id;
368		smu->smu_table.boot_values.pp_table_id = v_3_3->pplib_pptable_id;
369		smu->smu_table.boot_values.firmware_caps = v_3_3->firmware_capability;
370	}
371
372	smu->smu_table.boot_values.format_revision = header->format_revision;
373	smu->smu_table.boot_values.content_revision = header->content_revision;
374
375	smu_v12_0_atom_get_smu_clockinfo(smu->adev,
376					 (uint8_t)SMU12_SYSPLL0_SOCCLK_ID,
377					 (uint8_t)SMU12_SYSPLL0_ID,
378					 &smu->smu_table.boot_values.socclk);
379
380	smu_v12_0_atom_get_smu_clockinfo(smu->adev,
381					 (uint8_t)SMU12_SYSPLL1_DCFCLK_ID,
382					 (uint8_t)SMU12_SYSPLL1_ID,
383					 &smu->smu_table.boot_values.dcefclk);
384
385	smu_v12_0_atom_get_smu_clockinfo(smu->adev,
386					 (uint8_t)SMU12_SYSPLL0_VCLK_ID,
387					 (uint8_t)SMU12_SYSPLL0_ID,
388					 &smu->smu_table.boot_values.vclk);
389
390	smu_v12_0_atom_get_smu_clockinfo(smu->adev,
391					 (uint8_t)SMU12_SYSPLL0_DCLK_ID,
392					 (uint8_t)SMU12_SYSPLL0_ID,
393					 &smu->smu_table.boot_values.dclk);
394
395	if ((smu->smu_table.boot_values.format_revision == 3) &&
396	    (smu->smu_table.boot_values.content_revision >= 2))
397		smu_v12_0_atom_get_smu_clockinfo(smu->adev,
398						 (uint8_t)SMU12_SYSPLL3_0_FCLK_ID,
399						 (uint8_t)SMU12_SYSPLL3_0_ID,
400						 &smu->smu_table.boot_values.fclk);
401
402	smu_v12_0_atom_get_smu_clockinfo(smu->adev,
403					 (uint8_t)SMU12_SYSPLL0_LCLK_ID,
404					 (uint8_t)SMU12_SYSPLL0_ID,
405					 &smu->smu_table.boot_values.lclk);
406
407	return 0;
408}
409