1// SPDX-License-Identifier: GPL-2.0 OR BSD-3-Clause
2/*
3 * Copyright(c) 2015, 2016 Intel Corporation.
4 */
5
6#include <linux/firmware.h>
7
8#include "hfi.h"
9#include "efivar.h"
10#include "eprom.h"
11
12#define DEFAULT_PLATFORM_CONFIG_NAME "hfi1_platform.dat"
13
14static int validate_scratch_checksum(struct hfi1_devdata *dd)
15{
16	u64 checksum = 0, temp_scratch = 0;
17	int i, j, version;
18
19	temp_scratch = read_csr(dd, ASIC_CFG_SCRATCH);
20	version = (temp_scratch & BITMAP_VERSION_SMASK) >> BITMAP_VERSION_SHIFT;
21
22	/* Prevent power on default of all zeroes from passing checksum */
23	if (!version) {
24		dd_dev_err(dd, "%s: Config bitmap uninitialized\n", __func__);
25		dd_dev_err(dd,
26			   "%s: Please update your BIOS to support active channels\n",
27			   __func__);
28		return 0;
29	}
30
31	/*
32	 * ASIC scratch 0 only contains the checksum and bitmap version as
33	 * fields of interest, both of which are handled separately from the
34	 * loop below, so skip it
35	 */
36	checksum += version;
37	for (i = 1; i < ASIC_NUM_SCRATCH; i++) {
38		temp_scratch = read_csr(dd, ASIC_CFG_SCRATCH + (8 * i));
39		for (j = sizeof(u64); j != 0; j -= 2) {
40			checksum += (temp_scratch & 0xFFFF);
41			temp_scratch >>= 16;
42		}
43	}
44
45	while (checksum >> 16)
46		checksum = (checksum & CHECKSUM_MASK) + (checksum >> 16);
47
48	temp_scratch = read_csr(dd, ASIC_CFG_SCRATCH);
49	temp_scratch &= CHECKSUM_SMASK;
50	temp_scratch >>= CHECKSUM_SHIFT;
51
52	if (checksum + temp_scratch == 0xFFFF)
53		return 1;
54
55	dd_dev_err(dd, "%s: Configuration bitmap corrupted\n", __func__);
56	return 0;
57}
58
59static void save_platform_config_fields(struct hfi1_devdata *dd)
60{
61	struct hfi1_pportdata *ppd = dd->pport;
62	u64 temp_scratch = 0, temp_dest = 0;
63
64	temp_scratch = read_csr(dd, ASIC_CFG_SCRATCH_1);
65
66	temp_dest = temp_scratch &
67		    (dd->hfi1_id ? PORT1_PORT_TYPE_SMASK :
68		     PORT0_PORT_TYPE_SMASK);
69	ppd->port_type = temp_dest >>
70			 (dd->hfi1_id ? PORT1_PORT_TYPE_SHIFT :
71			  PORT0_PORT_TYPE_SHIFT);
72
73	temp_dest = temp_scratch &
74		    (dd->hfi1_id ? PORT1_LOCAL_ATTEN_SMASK :
75		     PORT0_LOCAL_ATTEN_SMASK);
76	ppd->local_atten = temp_dest >>
77			   (dd->hfi1_id ? PORT1_LOCAL_ATTEN_SHIFT :
78			    PORT0_LOCAL_ATTEN_SHIFT);
79
80	temp_dest = temp_scratch &
81		    (dd->hfi1_id ? PORT1_REMOTE_ATTEN_SMASK :
82		     PORT0_REMOTE_ATTEN_SMASK);
83	ppd->remote_atten = temp_dest >>
84			    (dd->hfi1_id ? PORT1_REMOTE_ATTEN_SHIFT :
85			     PORT0_REMOTE_ATTEN_SHIFT);
86
87	temp_dest = temp_scratch &
88		    (dd->hfi1_id ? PORT1_DEFAULT_ATTEN_SMASK :
89		     PORT0_DEFAULT_ATTEN_SMASK);
90	ppd->default_atten = temp_dest >>
91			     (dd->hfi1_id ? PORT1_DEFAULT_ATTEN_SHIFT :
92			      PORT0_DEFAULT_ATTEN_SHIFT);
93
94	temp_scratch = read_csr(dd, dd->hfi1_id ? ASIC_CFG_SCRATCH_3 :
95				ASIC_CFG_SCRATCH_2);
96
97	ppd->tx_preset_eq = (temp_scratch & TX_EQ_SMASK) >> TX_EQ_SHIFT;
98	ppd->tx_preset_noeq = (temp_scratch & TX_NO_EQ_SMASK) >> TX_NO_EQ_SHIFT;
99	ppd->rx_preset = (temp_scratch & RX_SMASK) >> RX_SHIFT;
100
101	ppd->max_power_class = (temp_scratch & QSFP_MAX_POWER_SMASK) >>
102				QSFP_MAX_POWER_SHIFT;
103
104	ppd->config_from_scratch = true;
105}
106
107void get_platform_config(struct hfi1_devdata *dd)
108{
109	int ret = 0;
110	u8 *temp_platform_config = NULL;
111	u32 esize;
112	const struct firmware *platform_config_file = NULL;
113
114	if (is_integrated(dd)) {
115		if (validate_scratch_checksum(dd)) {
116			save_platform_config_fields(dd);
117			return;
118		}
119	} else {
120		ret = eprom_read_platform_config(dd,
121						 (void **)&temp_platform_config,
122						 &esize);
123		if (!ret) {
124			/* success */
125			dd->platform_config.data = temp_platform_config;
126			dd->platform_config.size = esize;
127			return;
128		}
129	}
130	dd_dev_err(dd,
131		   "%s: Failed to get platform config, falling back to sub-optimal default file\n",
132		   __func__);
133
134	ret = request_firmware(&platform_config_file,
135			       DEFAULT_PLATFORM_CONFIG_NAME,
136			       &dd->pcidev->dev);
137	if (ret) {
138		dd_dev_err(dd,
139			   "%s: No default platform config file found\n",
140			   __func__);
141		return;
142	}
143
144	/*
145	 * Allocate separate memory block to store data and free firmware
146	 * structure. This allows free_platform_config to treat EPROM and
147	 * fallback configs in the same manner.
148	 */
149	dd->platform_config.data = kmemdup(platform_config_file->data,
150					   platform_config_file->size,
151					   GFP_KERNEL);
152	dd->platform_config.size = platform_config_file->size;
153	release_firmware(platform_config_file);
154}
155
156void free_platform_config(struct hfi1_devdata *dd)
157{
158	/* Release memory allocated for eprom or fallback file read. */
159	kfree(dd->platform_config.data);
160	dd->platform_config.data = NULL;
161}
162
163void get_port_type(struct hfi1_pportdata *ppd)
164{
165	int ret;
166	u32 temp;
167
168	ret = get_platform_config_field(ppd->dd, PLATFORM_CONFIG_PORT_TABLE, 0,
169					PORT_TABLE_PORT_TYPE, &temp,
170					4);
171	if (ret) {
172		ppd->port_type = PORT_TYPE_UNKNOWN;
173		return;
174	}
175	ppd->port_type = temp;
176}
177
178int set_qsfp_tx(struct hfi1_pportdata *ppd, int on)
179{
180	u8 tx_ctrl_byte = on ? 0x0 : 0xF;
181	int ret = 0;
182
183	ret = qsfp_write(ppd, ppd->dd->hfi1_id, QSFP_TX_CTRL_BYTE_OFFS,
184			 &tx_ctrl_byte, 1);
185	/* we expected 1, so consider 0 an error */
186	if (ret == 0)
187		ret = -EIO;
188	else if (ret == 1)
189		ret = 0;
190	return ret;
191}
192
193static int qual_power(struct hfi1_pportdata *ppd)
194{
195	u32 cable_power_class = 0, power_class_max = 0;
196	u8 *cache = ppd->qsfp_info.cache;
197	int ret = 0;
198
199	ret = get_platform_config_field(
200		ppd->dd, PLATFORM_CONFIG_SYSTEM_TABLE, 0,
201		SYSTEM_TABLE_QSFP_POWER_CLASS_MAX, &power_class_max, 4);
202	if (ret)
203		return ret;
204
205	cable_power_class = get_qsfp_power_class(cache[QSFP_MOD_PWR_OFFS]);
206
207	if (cable_power_class > power_class_max)
208		ppd->offline_disabled_reason =
209			HFI1_ODR_MASK(OPA_LINKDOWN_REASON_POWER_POLICY);
210
211	if (ppd->offline_disabled_reason ==
212			HFI1_ODR_MASK(OPA_LINKDOWN_REASON_POWER_POLICY)) {
213		dd_dev_err(
214			ppd->dd,
215			"%s: Port disabled due to system power restrictions\n",
216			__func__);
217		ret = -EPERM;
218	}
219	return ret;
220}
221
222static int qual_bitrate(struct hfi1_pportdata *ppd)
223{
224	u16 lss = ppd->link_speed_supported, lse = ppd->link_speed_enabled;
225	u8 *cache = ppd->qsfp_info.cache;
226
227	if ((lss & OPA_LINK_SPEED_25G) && (lse & OPA_LINK_SPEED_25G) &&
228	    cache[QSFP_NOM_BIT_RATE_250_OFFS] < 0x64)
229		ppd->offline_disabled_reason =
230			   HFI1_ODR_MASK(OPA_LINKDOWN_REASON_LINKSPEED_POLICY);
231
232	if ((lss & OPA_LINK_SPEED_12_5G) && (lse & OPA_LINK_SPEED_12_5G) &&
233	    cache[QSFP_NOM_BIT_RATE_100_OFFS] < 0x7D)
234		ppd->offline_disabled_reason =
235			   HFI1_ODR_MASK(OPA_LINKDOWN_REASON_LINKSPEED_POLICY);
236
237	if (ppd->offline_disabled_reason ==
238			HFI1_ODR_MASK(OPA_LINKDOWN_REASON_LINKSPEED_POLICY)) {
239		dd_dev_err(
240			ppd->dd,
241			"%s: Cable failed bitrate check, disabling port\n",
242			__func__);
243		return -EPERM;
244	}
245	return 0;
246}
247
248static int set_qsfp_high_power(struct hfi1_pportdata *ppd)
249{
250	u8 cable_power_class = 0, power_ctrl_byte = 0;
251	u8 *cache = ppd->qsfp_info.cache;
252	int ret;
253
254	cable_power_class = get_qsfp_power_class(cache[QSFP_MOD_PWR_OFFS]);
255
256	if (cable_power_class > QSFP_POWER_CLASS_1) {
257		power_ctrl_byte = cache[QSFP_PWR_CTRL_BYTE_OFFS];
258
259		power_ctrl_byte |= 1;
260		power_ctrl_byte &= ~(0x2);
261
262		ret = qsfp_write(ppd, ppd->dd->hfi1_id,
263				 QSFP_PWR_CTRL_BYTE_OFFS,
264				 &power_ctrl_byte, 1);
265		if (ret != 1)
266			return -EIO;
267
268		if (cable_power_class > QSFP_POWER_CLASS_4) {
269			power_ctrl_byte |= (1 << 2);
270			ret = qsfp_write(ppd, ppd->dd->hfi1_id,
271					 QSFP_PWR_CTRL_BYTE_OFFS,
272					 &power_ctrl_byte, 1);
273			if (ret != 1)
274				return -EIO;
275		}
276
277		/* SFF 8679 rev 1.7 LPMode Deassert time */
278		msleep(300);
279	}
280	return 0;
281}
282
283static void apply_rx_cdr(struct hfi1_pportdata *ppd,
284			 u32 rx_preset_index,
285			 u8 *cdr_ctrl_byte)
286{
287	u32 rx_preset;
288	u8 *cache = ppd->qsfp_info.cache;
289	int cable_power_class;
290
291	if (!((cache[QSFP_MOD_PWR_OFFS] & 0x4) &&
292	      (cache[QSFP_CDR_INFO_OFFS] & 0x40)))
293		return;
294
295	/* RX CDR present, bypass supported */
296	cable_power_class = get_qsfp_power_class(cache[QSFP_MOD_PWR_OFFS]);
297
298	if (cable_power_class <= QSFP_POWER_CLASS_3) {
299		/* Power class <= 3, ignore config & turn RX CDR on */
300		*cdr_ctrl_byte |= 0xF;
301		return;
302	}
303
304	get_platform_config_field(
305		ppd->dd, PLATFORM_CONFIG_RX_PRESET_TABLE,
306		rx_preset_index, RX_PRESET_TABLE_QSFP_RX_CDR_APPLY,
307		&rx_preset, 4);
308
309	if (!rx_preset) {
310		dd_dev_info(
311			ppd->dd,
312			"%s: RX_CDR_APPLY is set to disabled\n",
313			__func__);
314		return;
315	}
316	get_platform_config_field(
317		ppd->dd, PLATFORM_CONFIG_RX_PRESET_TABLE,
318		rx_preset_index, RX_PRESET_TABLE_QSFP_RX_CDR,
319		&rx_preset, 4);
320
321	/* Expand cdr setting to all 4 lanes */
322	rx_preset = (rx_preset | (rx_preset << 1) |
323			(rx_preset << 2) | (rx_preset << 3));
324
325	if (rx_preset) {
326		*cdr_ctrl_byte |= rx_preset;
327	} else {
328		*cdr_ctrl_byte &= rx_preset;
329		/* Preserve current TX CDR status */
330		*cdr_ctrl_byte |= (cache[QSFP_CDR_CTRL_BYTE_OFFS] & 0xF0);
331	}
332}
333
334static void apply_tx_cdr(struct hfi1_pportdata *ppd,
335			 u32 tx_preset_index,
336			 u8 *cdr_ctrl_byte)
337{
338	u32 tx_preset;
339	u8 *cache = ppd->qsfp_info.cache;
340	int cable_power_class;
341
342	if (!((cache[QSFP_MOD_PWR_OFFS] & 0x8) &&
343	      (cache[QSFP_CDR_INFO_OFFS] & 0x80)))
344		return;
345
346	/* TX CDR present, bypass supported */
347	cable_power_class = get_qsfp_power_class(cache[QSFP_MOD_PWR_OFFS]);
348
349	if (cable_power_class <= QSFP_POWER_CLASS_3) {
350		/* Power class <= 3, ignore config & turn TX CDR on */
351		*cdr_ctrl_byte |= 0xF0;
352		return;
353	}
354
355	get_platform_config_field(
356		ppd->dd,
357		PLATFORM_CONFIG_TX_PRESET_TABLE, tx_preset_index,
358		TX_PRESET_TABLE_QSFP_TX_CDR_APPLY, &tx_preset, 4);
359
360	if (!tx_preset) {
361		dd_dev_info(
362			ppd->dd,
363			"%s: TX_CDR_APPLY is set to disabled\n",
364			__func__);
365		return;
366	}
367	get_platform_config_field(
368		ppd->dd,
369		PLATFORM_CONFIG_TX_PRESET_TABLE,
370		tx_preset_index,
371		TX_PRESET_TABLE_QSFP_TX_CDR, &tx_preset, 4);
372
373	/* Expand cdr setting to all 4 lanes */
374	tx_preset = (tx_preset | (tx_preset << 1) |
375			(tx_preset << 2) | (tx_preset << 3));
376
377	if (tx_preset)
378		*cdr_ctrl_byte |= (tx_preset << 4);
379	else
380		/* Preserve current/determined RX CDR status */
381		*cdr_ctrl_byte &= ((tx_preset << 4) | 0xF);
382}
383
384static void apply_cdr_settings(
385		struct hfi1_pportdata *ppd, u32 rx_preset_index,
386		u32 tx_preset_index)
387{
388	u8 *cache = ppd->qsfp_info.cache;
389	u8 cdr_ctrl_byte = cache[QSFP_CDR_CTRL_BYTE_OFFS];
390
391	apply_rx_cdr(ppd, rx_preset_index, &cdr_ctrl_byte);
392
393	apply_tx_cdr(ppd, tx_preset_index, &cdr_ctrl_byte);
394
395	qsfp_write(ppd, ppd->dd->hfi1_id, QSFP_CDR_CTRL_BYTE_OFFS,
396		   &cdr_ctrl_byte, 1);
397}
398
399static void apply_tx_eq_auto(struct hfi1_pportdata *ppd)
400{
401	u8 *cache = ppd->qsfp_info.cache;
402	u8 tx_eq;
403
404	if (!(cache[QSFP_EQ_INFO_OFFS] & 0x8))
405		return;
406	/* Disable adaptive TX EQ if present */
407	tx_eq = cache[(128 * 3) + 241];
408	tx_eq &= 0xF0;
409	qsfp_write(ppd, ppd->dd->hfi1_id, (256 * 3) + 241, &tx_eq, 1);
410}
411
412static void apply_tx_eq_prog(struct hfi1_pportdata *ppd, u32 tx_preset_index)
413{
414	u8 *cache = ppd->qsfp_info.cache;
415	u32 tx_preset;
416	u8 tx_eq;
417
418	if (!(cache[QSFP_EQ_INFO_OFFS] & 0x4))
419		return;
420
421	get_platform_config_field(
422		ppd->dd, PLATFORM_CONFIG_TX_PRESET_TABLE,
423		tx_preset_index, TX_PRESET_TABLE_QSFP_TX_EQ_APPLY,
424		&tx_preset, 4);
425	if (!tx_preset) {
426		dd_dev_info(
427			ppd->dd,
428			"%s: TX_EQ_APPLY is set to disabled\n",
429			__func__);
430		return;
431	}
432	get_platform_config_field(
433			ppd->dd, PLATFORM_CONFIG_TX_PRESET_TABLE,
434			tx_preset_index, TX_PRESET_TABLE_QSFP_TX_EQ,
435			&tx_preset, 4);
436
437	if (((cache[(128 * 3) + 224] & 0xF0) >> 4) < tx_preset) {
438		dd_dev_info(
439			ppd->dd,
440			"%s: TX EQ %x unsupported\n",
441			__func__, tx_preset);
442
443		dd_dev_info(
444			ppd->dd,
445			"%s: Applying EQ %x\n",
446			__func__, cache[608] & 0xF0);
447
448		tx_preset = (cache[608] & 0xF0) >> 4;
449	}
450
451	tx_eq = tx_preset | (tx_preset << 4);
452	qsfp_write(ppd, ppd->dd->hfi1_id, (256 * 3) + 234, &tx_eq, 1);
453	qsfp_write(ppd, ppd->dd->hfi1_id, (256 * 3) + 235, &tx_eq, 1);
454}
455
456static void apply_rx_eq_emp(struct hfi1_pportdata *ppd, u32 rx_preset_index)
457{
458	u32 rx_preset;
459	u8 rx_eq, *cache = ppd->qsfp_info.cache;
460
461	if (!(cache[QSFP_EQ_INFO_OFFS] & 0x2))
462		return;
463	get_platform_config_field(
464			ppd->dd, PLATFORM_CONFIG_RX_PRESET_TABLE,
465			rx_preset_index, RX_PRESET_TABLE_QSFP_RX_EMP_APPLY,
466			&rx_preset, 4);
467
468	if (!rx_preset) {
469		dd_dev_info(
470			ppd->dd,
471			"%s: RX_EMP_APPLY is set to disabled\n",
472			__func__);
473		return;
474	}
475	get_platform_config_field(
476		ppd->dd, PLATFORM_CONFIG_RX_PRESET_TABLE,
477		rx_preset_index, RX_PRESET_TABLE_QSFP_RX_EMP,
478		&rx_preset, 4);
479
480	if ((cache[(128 * 3) + 224] & 0xF) < rx_preset) {
481		dd_dev_info(
482			ppd->dd,
483			"%s: Requested RX EMP %x\n",
484			__func__, rx_preset);
485
486		dd_dev_info(
487			ppd->dd,
488			"%s: Applying supported EMP %x\n",
489			__func__, cache[608] & 0xF);
490
491		rx_preset = cache[608] & 0xF;
492	}
493
494	rx_eq = rx_preset | (rx_preset << 4);
495
496	qsfp_write(ppd, ppd->dd->hfi1_id, (256 * 3) + 236, &rx_eq, 1);
497	qsfp_write(ppd, ppd->dd->hfi1_id, (256 * 3) + 237, &rx_eq, 1);
498}
499
500static void apply_eq_settings(struct hfi1_pportdata *ppd,
501			      u32 rx_preset_index, u32 tx_preset_index)
502{
503	u8 *cache = ppd->qsfp_info.cache;
504
505	/* no point going on w/o a page 3 */
506	if (cache[2] & 4) {
507		dd_dev_info(ppd->dd,
508			    "%s: Upper page 03 not present\n",
509			    __func__);
510		return;
511	}
512
513	apply_tx_eq_auto(ppd);
514
515	apply_tx_eq_prog(ppd, tx_preset_index);
516
517	apply_rx_eq_emp(ppd, rx_preset_index);
518}
519
520static void apply_rx_amplitude_settings(
521		struct hfi1_pportdata *ppd, u32 rx_preset_index,
522		u32 tx_preset_index)
523{
524	u32 rx_preset;
525	u8 rx_amp = 0, i = 0, preferred = 0, *cache = ppd->qsfp_info.cache;
526
527	/* no point going on w/o a page 3 */
528	if (cache[2] & 4) {
529		dd_dev_info(ppd->dd,
530			    "%s: Upper page 03 not present\n",
531			    __func__);
532		return;
533	}
534	if (!(cache[QSFP_EQ_INFO_OFFS] & 0x1)) {
535		dd_dev_info(ppd->dd,
536			    "%s: RX_AMP_APPLY is set to disabled\n",
537			    __func__);
538		return;
539	}
540
541	get_platform_config_field(ppd->dd,
542				  PLATFORM_CONFIG_RX_PRESET_TABLE,
543				  rx_preset_index,
544				  RX_PRESET_TABLE_QSFP_RX_AMP_APPLY,
545				  &rx_preset, 4);
546
547	if (!rx_preset) {
548		dd_dev_info(ppd->dd,
549			    "%s: RX_AMP_APPLY is set to disabled\n",
550			    __func__);
551		return;
552	}
553	get_platform_config_field(ppd->dd,
554				  PLATFORM_CONFIG_RX_PRESET_TABLE,
555				  rx_preset_index,
556				  RX_PRESET_TABLE_QSFP_RX_AMP,
557				  &rx_preset, 4);
558
559	dd_dev_info(ppd->dd,
560		    "%s: Requested RX AMP %x\n",
561		    __func__,
562		    rx_preset);
563
564	for (i = 0; i < 4; i++) {
565		if (cache[(128 * 3) + 225] & (1 << i)) {
566			preferred = i;
567			if (preferred == rx_preset)
568				break;
569		}
570	}
571
572	/*
573	 * Verify that preferred RX amplitude is not just a
574	 * fall through of the default
575	 */
576	if (!preferred && !(cache[(128 * 3) + 225] & 0x1)) {
577		dd_dev_info(ppd->dd, "No supported RX AMP, not applying\n");
578		return;
579	}
580
581	dd_dev_info(ppd->dd,
582		    "%s: Applying RX AMP %x\n", __func__, preferred);
583
584	rx_amp = preferred | (preferred << 4);
585	qsfp_write(ppd, ppd->dd->hfi1_id, (256 * 3) + 238, &rx_amp, 1);
586	qsfp_write(ppd, ppd->dd->hfi1_id, (256 * 3) + 239, &rx_amp, 1);
587}
588
589#define OPA_INVALID_INDEX 0xFFF
590
591static void apply_tx_lanes(struct hfi1_pportdata *ppd, u8 field_id,
592			   u32 config_data, const char *message)
593{
594	u8 i;
595	int ret;
596
597	for (i = 0; i < 4; i++) {
598		ret = load_8051_config(ppd->dd, field_id, i, config_data);
599		if (ret != HCMD_SUCCESS) {
600			dd_dev_err(
601				ppd->dd,
602				"%s: %s for lane %u failed\n",
603				message, __func__, i);
604		}
605	}
606}
607
608/*
609 * Return a special SerDes setting for low power AOC cables.  The power class
610 * threshold and setting being used were all found by empirical testing.
611 *
612 * Summary of the logic:
613 *
614 * if (QSFP and QSFP_TYPE == AOC and QSFP_POWER_CLASS < 4)
615 *     return 0xe
616 * return 0; // leave at default
617 */
618static u8 aoc_low_power_setting(struct hfi1_pportdata *ppd)
619{
620	u8 *cache = ppd->qsfp_info.cache;
621	int power_class;
622
623	/* QSFP only */
624	if (ppd->port_type != PORT_TYPE_QSFP)
625		return 0; /* leave at default */
626
627	/* active optical cables only */
628	switch ((cache[QSFP_MOD_TECH_OFFS] & 0xF0) >> 4) {
629	case 0x0 ... 0x9: fallthrough;
630	case 0xC: fallthrough;
631	case 0xE:
632		/* active AOC */
633		power_class = get_qsfp_power_class(cache[QSFP_MOD_PWR_OFFS]);
634		if (power_class < QSFP_POWER_CLASS_4)
635			return 0xe;
636	}
637	return 0; /* leave at default */
638}
639
640static void apply_tunings(
641		struct hfi1_pportdata *ppd, u32 tx_preset_index,
642		u8 tuning_method, u32 total_atten, u8 limiting_active)
643{
644	int ret = 0;
645	u32 config_data = 0, tx_preset = 0;
646	u8 precur = 0, attn = 0, postcur = 0, external_device_config = 0;
647	u8 *cache = ppd->qsfp_info.cache;
648
649	/* Pass tuning method to 8051 */
650	read_8051_config(ppd->dd, LINK_TUNING_PARAMETERS, GENERAL_CONFIG,
651			 &config_data);
652	config_data &= ~(0xff << TUNING_METHOD_SHIFT);
653	config_data |= ((u32)tuning_method << TUNING_METHOD_SHIFT);
654	ret = load_8051_config(ppd->dd, LINK_TUNING_PARAMETERS, GENERAL_CONFIG,
655			       config_data);
656	if (ret != HCMD_SUCCESS)
657		dd_dev_err(ppd->dd, "%s: Failed to set tuning method\n",
658			   __func__);
659
660	/* Set same channel loss for both TX and RX */
661	config_data = 0 | (total_atten << 16) | (total_atten << 24);
662	apply_tx_lanes(ppd, CHANNEL_LOSS_SETTINGS, config_data,
663		       "Setting channel loss");
664
665	/* Inform 8051 of cable capabilities */
666	if (ppd->qsfp_info.cache_valid) {
667		external_device_config =
668			((cache[QSFP_MOD_PWR_OFFS] & 0x4) << 3) |
669			((cache[QSFP_MOD_PWR_OFFS] & 0x8) << 2) |
670			((cache[QSFP_EQ_INFO_OFFS] & 0x2) << 1) |
671			(cache[QSFP_EQ_INFO_OFFS] & 0x4);
672		ret = read_8051_config(ppd->dd, DC_HOST_COMM_SETTINGS,
673				       GENERAL_CONFIG, &config_data);
674		/* Clear, then set the external device config field */
675		config_data &= ~(u32)0xFF;
676		config_data |= external_device_config;
677		ret = load_8051_config(ppd->dd, DC_HOST_COMM_SETTINGS,
678				       GENERAL_CONFIG, config_data);
679		if (ret != HCMD_SUCCESS)
680			dd_dev_err(ppd->dd,
681				   "%s: Failed set ext device config params\n",
682				   __func__);
683	}
684
685	if (tx_preset_index == OPA_INVALID_INDEX) {
686		if (ppd->port_type == PORT_TYPE_QSFP && limiting_active)
687			dd_dev_err(ppd->dd, "%s: Invalid Tx preset index\n",
688				   __func__);
689		return;
690	}
691
692	/* Following for limiting active channels only */
693	get_platform_config_field(
694		ppd->dd, PLATFORM_CONFIG_TX_PRESET_TABLE, tx_preset_index,
695		TX_PRESET_TABLE_PRECUR, &tx_preset, 4);
696	precur = tx_preset;
697
698	get_platform_config_field(
699		ppd->dd, PLATFORM_CONFIG_TX_PRESET_TABLE,
700		tx_preset_index, TX_PRESET_TABLE_ATTN, &tx_preset, 4);
701	attn = tx_preset;
702
703	get_platform_config_field(
704		ppd->dd, PLATFORM_CONFIG_TX_PRESET_TABLE,
705		tx_preset_index, TX_PRESET_TABLE_POSTCUR, &tx_preset, 4);
706	postcur = tx_preset;
707
708	/*
709	 * NOTES:
710	 * o The aoc_low_power_setting is applied to all lanes even
711	 *   though only lane 0's value is examined by the firmware.
712	 * o A lingering low power setting after a cable swap does
713	 *   not occur.  On cable unplug the 8051 is reset and
714	 *   restarted on cable insert.  This resets all settings to
715	 *   their default, erasing any previous low power setting.
716	 */
717	config_data = precur | (attn << 8) | (postcur << 16) |
718			(aoc_low_power_setting(ppd) << 24);
719
720	apply_tx_lanes(ppd, TX_EQ_SETTINGS, config_data,
721		       "Applying TX settings");
722}
723
724/* Must be holding the QSFP i2c resource */
725static int tune_active_qsfp(struct hfi1_pportdata *ppd, u32 *ptr_tx_preset,
726			    u32 *ptr_rx_preset, u32 *ptr_total_atten)
727{
728	int ret;
729	u16 lss = ppd->link_speed_supported, lse = ppd->link_speed_enabled;
730	u8 *cache = ppd->qsfp_info.cache;
731
732	ppd->qsfp_info.limiting_active = 1;
733
734	ret = set_qsfp_tx(ppd, 0);
735	if (ret)
736		return ret;
737
738	ret = qual_power(ppd);
739	if (ret)
740		return ret;
741
742	ret = qual_bitrate(ppd);
743	if (ret)
744		return ret;
745
746	/*
747	 * We'll change the QSFP memory contents from here on out, thus we set a
748	 * flag here to remind ourselves to reset the QSFP module. This prevents
749	 * reuse of stale settings established in our previous pass through.
750	 */
751	if (ppd->qsfp_info.reset_needed) {
752		ret = reset_qsfp(ppd);
753		if (ret)
754			return ret;
755		refresh_qsfp_cache(ppd, &ppd->qsfp_info);
756	} else {
757		ppd->qsfp_info.reset_needed = 1;
758	}
759
760	ret = set_qsfp_high_power(ppd);
761	if (ret)
762		return ret;
763
764	if (cache[QSFP_EQ_INFO_OFFS] & 0x4) {
765		ret = get_platform_config_field(
766			ppd->dd,
767			PLATFORM_CONFIG_PORT_TABLE, 0,
768			PORT_TABLE_TX_PRESET_IDX_ACTIVE_EQ,
769			ptr_tx_preset, 4);
770		if (ret) {
771			*ptr_tx_preset = OPA_INVALID_INDEX;
772			return ret;
773		}
774	} else {
775		ret = get_platform_config_field(
776			ppd->dd,
777			PLATFORM_CONFIG_PORT_TABLE, 0,
778			PORT_TABLE_TX_PRESET_IDX_ACTIVE_NO_EQ,
779			ptr_tx_preset, 4);
780		if (ret) {
781			*ptr_tx_preset = OPA_INVALID_INDEX;
782			return ret;
783		}
784	}
785
786	ret = get_platform_config_field(
787		ppd->dd, PLATFORM_CONFIG_PORT_TABLE, 0,
788		PORT_TABLE_RX_PRESET_IDX, ptr_rx_preset, 4);
789	if (ret) {
790		*ptr_rx_preset = OPA_INVALID_INDEX;
791		return ret;
792	}
793
794	if ((lss & OPA_LINK_SPEED_25G) && (lse & OPA_LINK_SPEED_25G))
795		get_platform_config_field(
796			ppd->dd, PLATFORM_CONFIG_PORT_TABLE, 0,
797			PORT_TABLE_LOCAL_ATTEN_25G, ptr_total_atten, 4);
798	else if ((lss & OPA_LINK_SPEED_12_5G) && (lse & OPA_LINK_SPEED_12_5G))
799		get_platform_config_field(
800			ppd->dd, PLATFORM_CONFIG_PORT_TABLE, 0,
801			PORT_TABLE_LOCAL_ATTEN_12G, ptr_total_atten, 4);
802
803	apply_cdr_settings(ppd, *ptr_rx_preset, *ptr_tx_preset);
804
805	apply_eq_settings(ppd, *ptr_rx_preset, *ptr_tx_preset);
806
807	apply_rx_amplitude_settings(ppd, *ptr_rx_preset, *ptr_tx_preset);
808
809	ret = set_qsfp_tx(ppd, 1);
810
811	return ret;
812}
813
814static int tune_qsfp(struct hfi1_pportdata *ppd,
815		     u32 *ptr_tx_preset, u32 *ptr_rx_preset,
816		     u8 *ptr_tuning_method, u32 *ptr_total_atten)
817{
818	u32 cable_atten = 0, remote_atten = 0, platform_atten = 0;
819	u16 lss = ppd->link_speed_supported, lse = ppd->link_speed_enabled;
820	int ret = 0;
821	u8 *cache = ppd->qsfp_info.cache;
822
823	switch ((cache[QSFP_MOD_TECH_OFFS] & 0xF0) >> 4) {
824	case 0xA ... 0xB:
825		ret = get_platform_config_field(
826			ppd->dd,
827			PLATFORM_CONFIG_PORT_TABLE, 0,
828			PORT_TABLE_LOCAL_ATTEN_25G,
829			&platform_atten, 4);
830		if (ret)
831			return ret;
832
833		if ((lss & OPA_LINK_SPEED_25G) && (lse & OPA_LINK_SPEED_25G))
834			cable_atten = cache[QSFP_CU_ATTEN_12G_OFFS];
835		else if ((lss & OPA_LINK_SPEED_12_5G) &&
836			 (lse & OPA_LINK_SPEED_12_5G))
837			cable_atten = cache[QSFP_CU_ATTEN_7G_OFFS];
838
839		/* Fallback to configured attenuation if cable memory is bad */
840		if (cable_atten == 0 || cable_atten > 36) {
841			ret = get_platform_config_field(
842				ppd->dd,
843				PLATFORM_CONFIG_SYSTEM_TABLE, 0,
844				SYSTEM_TABLE_QSFP_ATTENUATION_DEFAULT_25G,
845				&cable_atten, 4);
846			if (ret)
847				return ret;
848		}
849
850		ret = get_platform_config_field(
851			ppd->dd, PLATFORM_CONFIG_PORT_TABLE, 0,
852			PORT_TABLE_REMOTE_ATTEN_25G, &remote_atten, 4);
853		if (ret)
854			return ret;
855
856		*ptr_total_atten = platform_atten + cable_atten + remote_atten;
857
858		*ptr_tuning_method = OPA_PASSIVE_TUNING;
859		break;
860	case 0x0 ... 0x9: fallthrough;
861	case 0xC: fallthrough;
862	case 0xE:
863		ret = tune_active_qsfp(ppd, ptr_tx_preset, ptr_rx_preset,
864				       ptr_total_atten);
865		if (ret)
866			return ret;
867
868		*ptr_tuning_method = OPA_ACTIVE_TUNING;
869		break;
870	case 0xD: fallthrough;
871	case 0xF:
872	default:
873		dd_dev_warn(ppd->dd, "%s: Unknown/unsupported cable\n",
874			    __func__);
875		break;
876	}
877	return ret;
878}
879
880/*
881 * This function communicates its success or failure via ppd->driver_link_ready
882 * Thus, it depends on its association with start_link(...) which checks
883 * driver_link_ready before proceeding with the link negotiation and
884 * initialization process.
885 */
886void tune_serdes(struct hfi1_pportdata *ppd)
887{
888	int ret = 0;
889	u32 total_atten = 0;
890	u32 remote_atten = 0, platform_atten = 0;
891	u32 rx_preset_index, tx_preset_index;
892	u8 tuning_method = 0, limiting_active = 0;
893	struct hfi1_devdata *dd = ppd->dd;
894
895	rx_preset_index = OPA_INVALID_INDEX;
896	tx_preset_index = OPA_INVALID_INDEX;
897
898	/* the link defaults to enabled */
899	ppd->link_enabled = 1;
900	/* the driver link ready state defaults to not ready */
901	ppd->driver_link_ready = 0;
902	ppd->offline_disabled_reason = HFI1_ODR_MASK(OPA_LINKDOWN_REASON_NONE);
903
904	/* Skip the tuning for testing (loopback != none) and simulations */
905	if (loopback != LOOPBACK_NONE ||
906	    ppd->dd->icode == ICODE_FUNCTIONAL_SIMULATOR) {
907		ppd->driver_link_ready = 1;
908
909		if (qsfp_mod_present(ppd)) {
910			ret = acquire_chip_resource(ppd->dd,
911						    qsfp_resource(ppd->dd),
912						    QSFP_WAIT);
913			if (ret) {
914				dd_dev_err(ppd->dd, "%s: hfi%d: cannot lock i2c chain\n",
915					   __func__, (int)ppd->dd->hfi1_id);
916				goto bail;
917			}
918
919			refresh_qsfp_cache(ppd, &ppd->qsfp_info);
920			release_chip_resource(ppd->dd, qsfp_resource(ppd->dd));
921		}
922
923		return;
924	}
925
926	switch (ppd->port_type) {
927	case PORT_TYPE_DISCONNECTED:
928		ppd->offline_disabled_reason =
929			HFI1_ODR_MASK(OPA_LINKDOWN_REASON_DISCONNECTED);
930		dd_dev_warn(dd, "%s: Port disconnected, disabling port\n",
931			    __func__);
932		goto bail;
933	case PORT_TYPE_FIXED:
934		/* platform_atten, remote_atten pre-zeroed to catch error */
935		get_platform_config_field(
936			ppd->dd, PLATFORM_CONFIG_PORT_TABLE, 0,
937			PORT_TABLE_LOCAL_ATTEN_25G, &platform_atten, 4);
938
939		get_platform_config_field(
940			ppd->dd, PLATFORM_CONFIG_PORT_TABLE, 0,
941			PORT_TABLE_REMOTE_ATTEN_25G, &remote_atten, 4);
942
943		total_atten = platform_atten + remote_atten;
944
945		tuning_method = OPA_PASSIVE_TUNING;
946		break;
947	case PORT_TYPE_VARIABLE:
948		if (qsfp_mod_present(ppd)) {
949			/*
950			 * platform_atten, remote_atten pre-zeroed to
951			 * catch error
952			 */
953			get_platform_config_field(
954				ppd->dd, PLATFORM_CONFIG_PORT_TABLE, 0,
955				PORT_TABLE_LOCAL_ATTEN_25G,
956				&platform_atten, 4);
957
958			get_platform_config_field(
959				ppd->dd, PLATFORM_CONFIG_PORT_TABLE, 0,
960				PORT_TABLE_REMOTE_ATTEN_25G,
961				&remote_atten, 4);
962
963			total_atten = platform_atten + remote_atten;
964
965			tuning_method = OPA_PASSIVE_TUNING;
966		} else {
967			ppd->offline_disabled_reason =
968			     HFI1_ODR_MASK(OPA_LINKDOWN_REASON_CHASSIS_CONFIG);
969			goto bail;
970		}
971		break;
972	case PORT_TYPE_QSFP:
973		if (qsfp_mod_present(ppd)) {
974			ret = acquire_chip_resource(ppd->dd,
975						    qsfp_resource(ppd->dd),
976						    QSFP_WAIT);
977			if (ret) {
978				dd_dev_err(ppd->dd, "%s: hfi%d: cannot lock i2c chain\n",
979					   __func__, (int)ppd->dd->hfi1_id);
980				goto bail;
981			}
982			refresh_qsfp_cache(ppd, &ppd->qsfp_info);
983
984			if (ppd->qsfp_info.cache_valid) {
985				ret = tune_qsfp(ppd,
986						&tx_preset_index,
987						&rx_preset_index,
988						&tuning_method,
989						&total_atten);
990
991				/*
992				 * We may have modified the QSFP memory, so
993				 * update the cache to reflect the changes
994				 */
995				refresh_qsfp_cache(ppd, &ppd->qsfp_info);
996				limiting_active =
997						ppd->qsfp_info.limiting_active;
998			} else {
999				dd_dev_err(dd,
1000					   "%s: Reading QSFP memory failed\n",
1001					   __func__);
1002				ret = -EINVAL; /* a fail indication */
1003			}
1004			release_chip_resource(ppd->dd, qsfp_resource(ppd->dd));
1005			if (ret)
1006				goto bail;
1007		} else {
1008			ppd->offline_disabled_reason =
1009			   HFI1_ODR_MASK(
1010				OPA_LINKDOWN_REASON_LOCAL_MEDIA_NOT_INSTALLED);
1011			goto bail;
1012		}
1013		break;
1014	default:
1015		dd_dev_warn(ppd->dd, "%s: Unknown port type\n", __func__);
1016		ppd->port_type = PORT_TYPE_UNKNOWN;
1017		tuning_method = OPA_UNKNOWN_TUNING;
1018		total_atten = 0;
1019		limiting_active = 0;
1020		tx_preset_index = OPA_INVALID_INDEX;
1021		break;
1022	}
1023
1024	if (ppd->offline_disabled_reason ==
1025			HFI1_ODR_MASK(OPA_LINKDOWN_REASON_NONE))
1026		apply_tunings(ppd, tx_preset_index, tuning_method,
1027			      total_atten, limiting_active);
1028
1029	if (!ret)
1030		ppd->driver_link_ready = 1;
1031
1032	return;
1033bail:
1034	ppd->driver_link_ready = 0;
1035}
1036