1// SPDX-License-Identifier: GPL-2.0
2/*
3 * Copyright (C) Marvell International Ltd. and its affiliates
4 */
5
6#include <common.h>
7#include <i2c.h>
8#include <log.h>
9#include <spl.h>
10#include <asm/io.h>
11#include <asm/arch/cpu.h>
12#include <asm/arch/soc.h>
13#include <linux/delay.h>
14
15#include "ddr3_init.h"
16#include "ddr3_hw_training.h"
17#include "xor.h"
18
19#ifdef MV88F78X60
20#include "ddr3_patterns_64bit.h"
21#else
22#include "ddr3_patterns_16bit.h"
23#if defined(MV88F672X)
24#include "ddr3_patterns_16bit.h"
25#endif
26#endif
27
28/*
29 * Debug
30 */
31
32#define DEBUG_MAIN_C(s, d, l) \
33	DEBUG_MAIN_S(s); DEBUG_MAIN_D(d, l); DEBUG_MAIN_S("\n")
34#define DEBUG_MAIN_FULL_C(s, d, l) \
35	DEBUG_MAIN_FULL_S(s); DEBUG_MAIN_FULL_D(d, l); DEBUG_MAIN_FULL_S("\n")
36
37#ifdef MV_DEBUG_MAIN
38#define DEBUG_MAIN_S(s)			puts(s)
39#define DEBUG_MAIN_D(d, l)		printf("%x", d)
40#else
41#define DEBUG_MAIN_S(s)
42#define DEBUG_MAIN_D(d, l)
43#endif
44
45#ifdef MV_DEBUG_MAIN_FULL
46#define DEBUG_MAIN_FULL_S(s)		puts(s)
47#define DEBUG_MAIN_FULL_D(d, l)		printf("%x", d)
48#else
49#define DEBUG_MAIN_FULL_S(s)
50#define DEBUG_MAIN_FULL_D(d, l)
51#endif
52
53#ifdef MV_DEBUG_SUSPEND_RESUME
54#define DEBUG_SUSPEND_RESUME_S(s)	puts(s)
55#define DEBUG_SUSPEND_RESUME_D(d, l)	printf("%x", d)
56#else
57#define DEBUG_SUSPEND_RESUME_S(s)
58#define DEBUG_SUSPEND_RESUME_D(d, l)
59#endif
60
61static u32 ddr3_sw_wl_rl_debug;
62static u32 ddr3_run_pbs = 1;
63
64void ddr3_print_version(void)
65{
66	puts("DDR3 Training Sequence - Ver 5.7.");
67}
68
69void ddr3_set_sw_wl_rl_debug(u32 val)
70{
71	ddr3_sw_wl_rl_debug = val;
72}
73
74void ddr3_set_pbs(u32 val)
75{
76	ddr3_run_pbs = val;
77}
78
79int ddr3_hw_training(u32 target_freq, u32 ddr_width, int xor_bypass,
80		     u32 scrub_offs, u32 scrub_size, int dqs_clk_aligned,
81		     int debug_mode, int reg_dimm_skip_wl)
82{
83	/* A370 has no PBS mechanism */
84	__maybe_unused u32 first_loop_flag = 0;
85	u32 freq, reg;
86	MV_DRAM_INFO dram_info;
87	int ratio_2to1 = 0;
88	int tmp_ratio = 1;
89	int status;
90
91	if (debug_mode)
92		DEBUG_MAIN_S("DDR3 Training Sequence - DEBUG - 1\n");
93
94	memset(&dram_info, 0, sizeof(dram_info));
95	dram_info.num_cs = ddr3_get_cs_num_from_reg();
96	dram_info.cs_ena = ddr3_get_cs_ena_from_reg();
97	dram_info.target_frequency = target_freq;
98	dram_info.ddr_width = ddr_width;
99	dram_info.num_of_std_pups = ddr_width / PUP_SIZE;
100	dram_info.rl400_bug = 0;
101	dram_info.multi_cs_mr_support = 0;
102#ifdef MV88F67XX
103	dram_info.rl400_bug = 1;
104#endif
105
106	/* Ignore ECC errors - if ECC is enabled */
107	reg = reg_read(REG_SDRAM_CONFIG_ADDR);
108	if (reg & (1 << REG_SDRAM_CONFIG_ECC_OFFS)) {
109		dram_info.ecc_ena = 1;
110		reg |= (1 << REG_SDRAM_CONFIG_IERR_OFFS);
111		reg_write(REG_SDRAM_CONFIG_ADDR, reg);
112	} else {
113		dram_info.ecc_ena = 0;
114	}
115
116	reg = reg_read(REG_SDRAM_CONFIG_ADDR);
117	if (reg & (1 << REG_SDRAM_CONFIG_REGDIMM_OFFS))
118		dram_info.reg_dimm = 1;
119	else
120		dram_info.reg_dimm = 0;
121
122	dram_info.num_of_total_pups = ddr_width / PUP_SIZE + dram_info.ecc_ena;
123
124	/* Get target 2T value */
125	reg = reg_read(REG_DUNIT_CTRL_LOW_ADDR);
126	dram_info.mode_2t = (reg >> REG_DUNIT_CTRL_LOW_2T_OFFS) &
127		REG_DUNIT_CTRL_LOW_2T_MASK;
128
129	/* Get target CL value */
130#ifdef MV88F67XX
131	reg = reg_read(REG_DDR3_MR0_ADDR) >> 2;
132#else
133	reg = reg_read(REG_DDR3_MR0_CS_ADDR) >> 2;
134#endif
135
136	reg = (((reg >> 1) & 0xE) | (reg & 0x1)) & 0xF;
137	dram_info.cl = ddr3_valid_cl_to_cl(reg);
138
139	/* Get target CWL value */
140#ifdef MV88F67XX
141	reg = reg_read(REG_DDR3_MR2_ADDR) >> REG_DDR3_MR2_CWL_OFFS;
142#else
143	reg = reg_read(REG_DDR3_MR2_CS_ADDR) >> REG_DDR3_MR2_CWL_OFFS;
144#endif
145
146	reg &= REG_DDR3_MR2_CWL_MASK;
147	dram_info.cwl = reg;
148#if !defined(MV88F67XX)
149	/* A370 has no PBS mechanism */
150#if defined(MV88F78X60)
151	if ((dram_info.target_frequency > DDR_400) && (ddr3_run_pbs))
152		first_loop_flag = 1;
153#else
154	/* first_loop_flag = 1; skip mid freq at ALP/A375 */
155	if ((dram_info.target_frequency > DDR_400) && (ddr3_run_pbs) &&
156	    (mv_ctrl_revision_get() >= UMC_A0))
157		first_loop_flag = 1;
158	else
159		first_loop_flag = 0;
160#endif
161#endif
162
163	freq = dram_info.target_frequency;
164
165	/* Set ODT to always on */
166	ddr3_odt_activate(1);
167
168	/* Init XOR */
169	mv_sys_xor_init(&dram_info);
170
171	/* Get DRAM/HCLK ratio */
172	if (reg_read(REG_DDR_IO_ADDR) & (1 << REG_DDR_IO_CLK_RATIO_OFFS))
173		ratio_2to1 = 1;
174
175	/*
176	 * Xor Bypass - ECC support in AXP is currently available for 1:1
177	 * modes frequency modes.
178	 * Not all frequency modes support the ddr3 training sequence
179	 * (Only 1200/300).
180	 * Xor Bypass allows using the Xor initializations and scrubbing
181	 * inside the ddr3 training sequence without running the training
182	 * itself.
183	 */
184	if (xor_bypass == 0) {
185		if (ddr3_run_pbs) {
186			DEBUG_MAIN_S("DDR3 Training Sequence - Run with PBS.\n");
187		} else {
188			DEBUG_MAIN_S("DDR3 Training Sequence - Run without PBS.\n");
189		}
190
191		if (dram_info.target_frequency > DFS_MARGIN) {
192			tmp_ratio = 0;
193			freq = DDR_100;
194
195			if (dram_info.reg_dimm == 1)
196				freq = DDR_300;
197
198			if (MV_OK != ddr3_dfs_high_2_low(freq, &dram_info)) {
199				/* Set low - 100Mhz DDR Frequency by HW */
200				DEBUG_MAIN_S("DDR3 Training Sequence - FAILED (Dfs High2Low)\n");
201				return MV_DDR3_TRAINING_ERR_DFS_H2L;
202			}
203
204			if ((dram_info.reg_dimm == 1) &&
205			    (reg_dimm_skip_wl == 0)) {
206				if (MV_OK !=
207				    ddr3_write_leveling_hw_reg_dimm(freq,
208								    &dram_info))
209					DEBUG_MAIN_S("DDR3 Training Sequence - Registered DIMM Low WL - SKIP\n");
210			}
211
212			if (ddr3_get_log_level() >= MV_LOG_LEVEL_1)
213				ddr3_print_freq(freq);
214
215			if (debug_mode)
216				DEBUG_MAIN_S("DDR3 Training Sequence - DEBUG - 2\n");
217		} else {
218			if (!dqs_clk_aligned) {
219#ifdef MV88F67XX
220				/*
221				 * If running training sequence without DFS,
222				 * we must run Write leveling before writing
223				 * the patterns
224				 */
225
226				/*
227				 * ODT - Multi CS system use SW WL,
228				 * Single CS System use HW WL
229				 */
230				if (dram_info.cs_ena > 1) {
231					if (MV_OK !=
232					    ddr3_write_leveling_sw(
233						    freq, tmp_ratio,
234						    &dram_info)) {
235						DEBUG_MAIN_S("DDR3 Training Sequence - FAILED (Write Leveling Sw)\n");
236						return MV_DDR3_TRAINING_ERR_WR_LVL_SW;
237					}
238				} else {
239					if (MV_OK !=
240					    ddr3_write_leveling_hw(freq,
241								   &dram_info)) {
242						DEBUG_MAIN_S("DDR3 Training Sequence - FAILED (Write Leveling Hw)\n");
243						return MV_DDR3_TRAINING_ERR_WR_LVL_HW;
244					}
245				}
246#else
247				if (MV_OK != ddr3_write_leveling_hw(
248					    freq, &dram_info)) {
249					DEBUG_MAIN_S("DDR3 Training Sequence - FAILED (Write Leveling Hw)\n");
250					if (ddr3_sw_wl_rl_debug) {
251						if (MV_OK !=
252						    ddr3_write_leveling_sw(
253							    freq, tmp_ratio,
254							    &dram_info)) {
255							DEBUG_MAIN_S("DDR3 Training Sequence - FAILED (Write Leveling Sw)\n");
256							return MV_DDR3_TRAINING_ERR_WR_LVL_SW;
257						}
258					} else {
259						return MV_DDR3_TRAINING_ERR_WR_LVL_HW;
260					}
261				}
262#endif
263			}
264
265			if (debug_mode)
266				DEBUG_MAIN_S("DDR3 Training Sequence - DEBUG - 3\n");
267		}
268
269		if (MV_OK != ddr3_load_patterns(&dram_info, 0)) {
270			DEBUG_MAIN_S("DDR3 Training Sequence - FAILED (Loading Patterns)\n");
271			return MV_DDR3_TRAINING_ERR_LOAD_PATTERNS;
272		}
273
274		/*
275		 * TODO:
276		 * The mainline U-Boot port of the bin_hdr DDR training code
277		 * needs a delay of minimum 20ms here (10ms is a bit too short
278		 * and the CPU hangs). The bin_hdr code doesn't have this delay.
279		 * To be save here, lets add a delay of 50ms here.
280		 *
281		 * Tested on the Marvell DB-MV784MP-GP board
282		 */
283		mdelay(50);
284
285		do {
286			freq = dram_info.target_frequency;
287			tmp_ratio = ratio_2to1;
288			DEBUG_MAIN_FULL_S("DDR3 Training Sequence - DEBUG - 4\n");
289
290#if defined(MV88F78X60)
291			/*
292			 * There is a difference on the DFS frequency at the
293			 * first iteration of this loop
294			 */
295			if (first_loop_flag) {
296				freq = DDR_400;
297				tmp_ratio = 0;
298			}
299#endif
300
301			if (MV_OK != ddr3_dfs_low_2_high(freq, tmp_ratio,
302							 &dram_info)) {
303				DEBUG_MAIN_S("DDR3 Training Sequence - FAILED (Dfs Low2High)\n");
304				return MV_DDR3_TRAINING_ERR_DFS_H2L;
305			}
306
307			if (ddr3_get_log_level() >= MV_LOG_LEVEL_1) {
308				ddr3_print_freq(freq);
309			}
310
311			if (debug_mode)
312				DEBUG_MAIN_S("DDR3 Training Sequence - DEBUG - 5\n");
313
314			/* Write leveling */
315			if (!dqs_clk_aligned) {
316#ifdef MV88F67XX
317				/*
318				 * ODT - Multi CS system that not support Multi
319				 * CS MRS commands must use SW WL
320				 */
321				if (dram_info.cs_ena > 1) {
322					if (MV_OK != ddr3_write_leveling_sw(
323						    freq, tmp_ratio, &dram_info)) {
324						DEBUG_MAIN_S("DDR3 Training Sequence - FAILED (Write Leveling Sw)\n");
325						return MV_DDR3_TRAINING_ERR_WR_LVL_SW;
326					}
327				} else {
328					if (MV_OK != ddr3_write_leveling_hw(
329						    freq, &dram_info)) {
330						DEBUG_MAIN_S("DDR3 Training Sequence - FAILED (Write Leveling Hw)\n");
331						return MV_DDR3_TRAINING_ERR_WR_LVL_HW;
332					}
333				}
334#else
335				if ((dram_info.reg_dimm == 1) &&
336				    (freq == DDR_400)) {
337					if (reg_dimm_skip_wl == 0) {
338						if (MV_OK != ddr3_write_leveling_hw_reg_dimm(
339							    freq, &dram_info))
340							DEBUG_MAIN_S("DDR3 Training Sequence - Registered DIMM WL - SKIP\n");
341					}
342				} else {
343					if (MV_OK != ddr3_write_leveling_hw(
344						    freq, &dram_info)) {
345						DEBUG_MAIN_S("DDR3 Training Sequence - FAILED (Write Leveling Hw)\n");
346						if (ddr3_sw_wl_rl_debug) {
347							if (MV_OK != ddr3_write_leveling_sw(
348								    freq, tmp_ratio, &dram_info)) {
349								DEBUG_MAIN_S("DDR3 Training Sequence - FAILED (Write Leveling Sw)\n");
350								return MV_DDR3_TRAINING_ERR_WR_LVL_SW;
351							}
352						} else {
353							return MV_DDR3_TRAINING_ERR_WR_LVL_HW;
354						}
355					}
356				}
357#endif
358				if (debug_mode)
359					DEBUG_MAIN_S
360					    ("DDR3 Training Sequence - DEBUG - 6\n");
361			}
362
363			/* Read Leveling */
364			/*
365			 * Armada 370 - Support for HCLK @ 400MHZ - must use
366			 * SW read leveling
367			 */
368			if (freq == DDR_400 && dram_info.rl400_bug) {
369				status = ddr3_read_leveling_sw(freq, tmp_ratio,
370						       &dram_info);
371				if (MV_OK != status) {
372					DEBUG_MAIN_S
373					    ("DDR3 Training Sequence - FAILED (Read Leveling Sw)\n");
374					return status;
375				}
376			} else {
377				if (MV_OK != ddr3_read_leveling_hw(
378					    freq, &dram_info)) {
379					DEBUG_MAIN_S("DDR3 Training Sequence - FAILED (Read Leveling Hw)\n");
380					if (ddr3_sw_wl_rl_debug) {
381						if (MV_OK != ddr3_read_leveling_sw(
382							    freq, tmp_ratio,
383							    &dram_info)) {
384							DEBUG_MAIN_S("DDR3 Training Sequence - FAILED (Read Leveling Sw)\n");
385							return MV_DDR3_TRAINING_ERR_WR_LVL_SW;
386						}
387					} else {
388						return MV_DDR3_TRAINING_ERR_WR_LVL_HW;
389					}
390				}
391			}
392
393			if (debug_mode)
394				DEBUG_MAIN_S("DDR3 Training Sequence - DEBUG - 7\n");
395
396			if (MV_OK != ddr3_wl_supplement(&dram_info)) {
397				DEBUG_MAIN_S("DDR3 Training Sequence - FAILED (Write Leveling Hi-Freq Sup)\n");
398				return MV_DDR3_TRAINING_ERR_WR_LVL_HI_FREQ;
399			}
400
401			if (debug_mode)
402				DEBUG_MAIN_S("DDR3 Training Sequence - DEBUG - 8\n");
403#if !defined(MV88F67XX)
404			/* A370 has no PBS mechanism */
405#if defined(MV88F78X60) || defined(MV88F672X)
406			if (first_loop_flag == 1) {
407				first_loop_flag = 0;
408
409				status = MV_OK;
410				status = ddr3_pbs_rx(&dram_info);
411				if (MV_OK != status) {
412					DEBUG_MAIN_S("DDR3 Training Sequence - FAILED (PBS RX)\n");
413					return status;
414				}
415
416				if (debug_mode)
417					DEBUG_MAIN_S("DDR3 Training Sequence - DEBUG - 9\n");
418
419				status = ddr3_pbs_tx(&dram_info);
420				if (MV_OK != status) {
421					DEBUG_MAIN_S("DDR3 Training Sequence - FAILED (PBS TX)\n");
422					return status;
423				}
424
425				if (debug_mode)
426					DEBUG_MAIN_S("DDR3 Training Sequence - DEBUG - 10\n");
427			}
428#endif
429#endif
430		} while (freq != dram_info.target_frequency);
431
432		status = ddr3_dqs_centralization_rx(&dram_info);
433		if (MV_OK != status) {
434			DEBUG_MAIN_S("DDR3 Training Sequence - FAILED (DQS Centralization RX)\n");
435			return status;
436		}
437
438		if (debug_mode)
439			DEBUG_MAIN_S("DDR3 Training Sequence - DEBUG - 11\n");
440
441		status = ddr3_dqs_centralization_tx(&dram_info);
442		if (MV_OK != status) {
443			DEBUG_MAIN_S("DDR3 Training Sequence - FAILED (DQS Centralization TX)\n");
444			return status;
445		}
446
447		if (debug_mode)
448			DEBUG_MAIN_S("DDR3 Training Sequence - DEBUG - 12\n");
449	}
450
451	ddr3_set_performance_params(&dram_info);
452
453	if (dram_info.ecc_ena) {
454		/* Need to SCRUB the DRAM memory area to load U-Boot */
455		mv_sys_xor_finish();
456		dram_info.num_cs = 1;
457		dram_info.cs_ena = 1;
458		mv_sys_xor_init(&dram_info);
459		mv_xor_mem_init(0, scrub_offs, scrub_size, 0xdeadbeef,
460				0xdeadbeef);
461
462		/* Wait for previous transfer completion */
463		while (mv_xor_state_get(0) != MV_IDLE)
464			;
465
466		if (debug_mode)
467			DEBUG_MAIN_S("DDR3 Training Sequence - DEBUG - 13\n");
468	}
469
470	/* Return XOR State */
471	mv_sys_xor_finish();
472
473#if defined(MV88F78X60)
474	/* Save training results in memeory for resume state */
475	ddr3_save_training(&dram_info);
476#endif
477	/* Clear ODT always on */
478	ddr3_odt_activate(0);
479
480	/* Configure Dynamic read ODT */
481	ddr3_odt_read_dynamic_config(&dram_info);
482
483	return MV_OK;
484}
485
486void ddr3_set_performance_params(MV_DRAM_INFO *dram_info)
487{
488	u32 twr2wr, trd2rd, trd2wr_wr2rd;
489	u32 tmp1, tmp2, reg;
490
491	DEBUG_MAIN_FULL_C("Max WL Phase: ", dram_info->wl_max_phase, 2);
492	DEBUG_MAIN_FULL_C("Min WL Phase: ", dram_info->wl_min_phase, 2);
493	DEBUG_MAIN_FULL_C("Max RL Phase: ", dram_info->rl_max_phase, 2);
494	DEBUG_MAIN_FULL_C("Min RL Phase: ", dram_info->rl_min_phase, 2);
495
496	if (dram_info->wl_max_phase < 2)
497		twr2wr = 0x2;
498	else
499		twr2wr = 0x3;
500
501	trd2rd = 0x1 + (dram_info->rl_max_phase + 1) / 2 +
502		(dram_info->rl_max_phase + 1) % 2;
503
504	tmp1 = (dram_info->rl_max_phase - dram_info->wl_min_phase) / 2 +
505		(((dram_info->rl_max_phase - dram_info->wl_min_phase) % 2) >
506		 0 ? 1 : 0);
507	tmp2 = (dram_info->wl_max_phase - dram_info->rl_min_phase) / 2 +
508		((dram_info->wl_max_phase - dram_info->rl_min_phase) % 2 >
509		 0 ? 1 : 0);
510	trd2wr_wr2rd = (tmp1 >= tmp2) ? tmp1 : tmp2;
511
512	trd2wr_wr2rd += 2;
513	trd2rd += 2;
514	twr2wr += 2;
515
516	DEBUG_MAIN_FULL_C("WR 2 WR: ", twr2wr, 2);
517	DEBUG_MAIN_FULL_C("RD 2 RD: ", trd2rd, 2);
518	DEBUG_MAIN_FULL_C("RD 2 WR / WR 2 RD: ", trd2wr_wr2rd, 2);
519
520	reg = reg_read(REG_SDRAM_TIMING_HIGH_ADDR);
521
522	reg &= ~(REG_SDRAM_TIMING_H_W2W_MASK << REG_SDRAM_TIMING_H_W2W_OFFS);
523	reg |= ((twr2wr & REG_SDRAM_TIMING_H_W2W_MASK) <<
524		REG_SDRAM_TIMING_H_W2W_OFFS);
525
526	reg &= ~(REG_SDRAM_TIMING_H_R2R_MASK << REG_SDRAM_TIMING_H_R2R_OFFS);
527	reg &= ~(REG_SDRAM_TIMING_H_R2R_H_MASK <<
528		 REG_SDRAM_TIMING_H_R2R_H_OFFS);
529	reg |= ((trd2rd & REG_SDRAM_TIMING_H_R2R_MASK) <<
530		REG_SDRAM_TIMING_H_R2R_OFFS);
531	reg |= (((trd2rd >> 2) & REG_SDRAM_TIMING_H_R2R_H_MASK) <<
532		REG_SDRAM_TIMING_H_R2R_H_OFFS);
533
534	reg &= ~(REG_SDRAM_TIMING_H_R2W_W2R_MASK <<
535		 REG_SDRAM_TIMING_H_R2W_W2R_OFFS);
536	reg &= ~(REG_SDRAM_TIMING_H_R2W_W2R_H_MASK <<
537		 REG_SDRAM_TIMING_H_R2W_W2R_H_OFFS);
538	reg |= ((trd2wr_wr2rd & REG_SDRAM_TIMING_H_R2W_W2R_MASK) <<
539		REG_SDRAM_TIMING_H_R2W_W2R_OFFS);
540	reg |= (((trd2wr_wr2rd >> 2) & REG_SDRAM_TIMING_H_R2W_W2R_H_MASK) <<
541		REG_SDRAM_TIMING_H_R2W_W2R_H_OFFS);
542
543	reg_write(REG_SDRAM_TIMING_HIGH_ADDR, reg);
544}
545
546/*
547 * Perform DDR3 PUP Indirect Write
548 */
549void ddr3_write_pup_reg(u32 mode, u32 cs, u32 pup, u32 phase, u32 delay)
550{
551	u32 reg = 0;
552
553	if (pup == PUP_BC)
554		reg |= (1 << REG_PHY_BC_OFFS);
555	else
556		reg |= (pup << REG_PHY_PUP_OFFS);
557
558	reg |= ((0x4 * cs + mode) << REG_PHY_CS_OFFS);
559	reg |= (phase << REG_PHY_PHASE_OFFS) | delay;
560
561	if (mode == PUP_WL_MODE)
562		reg |= ((INIT_WL_DELAY + delay) << REG_PHY_DQS_REF_DLY_OFFS);
563
564	reg_write(REG_PHY_REGISTRY_FILE_ACCESS_ADDR, reg);	/* 0x16A0 */
565	reg |= REG_PHY_REGISTRY_FILE_ACCESS_OP_WR;
566	reg_write(REG_PHY_REGISTRY_FILE_ACCESS_ADDR, reg);	/* 0x16A0 */
567
568	do {
569		reg = reg_read(REG_PHY_REGISTRY_FILE_ACCESS_ADDR) &
570			REG_PHY_REGISTRY_FILE_ACCESS_OP_DONE;
571	} while (reg);	/* Wait for '0' to mark the end of the transaction */
572
573	/* If read Leveling mode - need to write to register 3 separetly */
574	if (mode == PUP_RL_MODE) {
575		reg = 0;
576
577		if (pup == PUP_BC)
578			reg |= (1 << REG_PHY_BC_OFFS);
579		else
580			reg |= (pup << REG_PHY_PUP_OFFS);
581
582		reg |= ((0x4 * cs + mode + 1) << REG_PHY_CS_OFFS);
583		reg |= (INIT_RL_DELAY);
584
585		reg_write(REG_PHY_REGISTRY_FILE_ACCESS_ADDR, reg); /* 0x16A0 */
586		reg |= REG_PHY_REGISTRY_FILE_ACCESS_OP_WR;
587		reg_write(REG_PHY_REGISTRY_FILE_ACCESS_ADDR, reg); /* 0x16A0 */
588
589		do {
590			reg = reg_read(REG_PHY_REGISTRY_FILE_ACCESS_ADDR) &
591				REG_PHY_REGISTRY_FILE_ACCESS_OP_DONE;
592		} while (reg);
593	}
594}
595
596/*
597 * Perform DDR3 PUP Indirect Read
598 */
599u32 ddr3_read_pup_reg(u32 mode, u32 cs, u32 pup)
600{
601	u32 reg;
602
603	reg = (pup << REG_PHY_PUP_OFFS) |
604		((0x4 * cs + mode) << REG_PHY_CS_OFFS);
605	reg_write(REG_PHY_REGISTRY_FILE_ACCESS_ADDR, reg);	/* 0x16A0 */
606
607	reg |= REG_PHY_REGISTRY_FILE_ACCESS_OP_RD;
608	reg_write(REG_PHY_REGISTRY_FILE_ACCESS_ADDR, reg);	/* 0x16A0 */
609
610	do {
611		reg = reg_read(REG_PHY_REGISTRY_FILE_ACCESS_ADDR) &
612			REG_PHY_REGISTRY_FILE_ACCESS_OP_DONE;
613	} while (reg);	/* Wait for '0' to mark the end of the transaction */
614
615	return reg_read(REG_PHY_REGISTRY_FILE_ACCESS_ADDR);	/* 0x16A0 */
616}
617
618/*
619 * Set training patterns
620 */
621int ddr3_load_patterns(MV_DRAM_INFO *dram_info, int resume)
622{
623	u32 reg;
624
625	/* Enable SW override - Required for the ECC Pup */
626	reg = reg_read(REG_DRAM_TRAINING_2_ADDR) |
627		(1 << REG_DRAM_TRAINING_2_SW_OVRD_OFFS);
628
629	/* [0] = 1 - Enable SW override  */
630	/* 0x15B8 - Training SW 2 Register */
631	reg_write(REG_DRAM_TRAINING_2_ADDR, reg);
632
633	reg = (1 << REG_DRAM_TRAINING_AUTO_OFFS);
634	reg_write(REG_DRAM_TRAINING_ADDR, reg);	/* 0x15B0 - Training Register */
635
636	if (resume == 0) {
637#if defined(MV88F78X60) || defined(MV88F672X)
638		ddr3_load_pbs_patterns(dram_info);
639#endif
640		ddr3_load_dqs_patterns(dram_info);
641	}
642
643	/* Disable SW override - Must be in a different stage */
644	/* [0]=0 - Enable SW override  */
645	reg = reg_read(REG_DRAM_TRAINING_2_ADDR);
646	reg &= ~(1 << REG_DRAM_TRAINING_2_SW_OVRD_OFFS);
647	/* 0x15B8 - Training SW 2 Register */
648	reg_write(REG_DRAM_TRAINING_2_ADDR, reg);
649
650	reg = reg_read(REG_DRAM_TRAINING_1_ADDR) |
651		(1 << REG_DRAM_TRAINING_1_TRNBPOINT_OFFS);
652	reg_write(REG_DRAM_TRAINING_1_ADDR, reg);
653
654	/* Set Base Addr */
655#if defined(MV88F67XX)
656	reg_write(REG_DRAM_TRAINING_PATTERN_BASE_ADDR, 0);
657#else
658	if (resume == 0)
659		reg_write(REG_DRAM_TRAINING_PATTERN_BASE_ADDR, 0);
660	else
661		reg_write(REG_DRAM_TRAINING_PATTERN_BASE_ADDR,
662			  RESUME_RL_PATTERNS_ADDR);
663#endif
664
665	/* Set Patterns */
666	if (resume == 0) {
667		reg = (dram_info->cs_ena << REG_DRAM_TRAINING_CS_OFFS) |
668			(1 << REG_DRAM_TRAINING_PATTERNS_OFFS);
669	} else {
670		reg = (0x1 << REG_DRAM_TRAINING_CS_OFFS) |
671			(1 << REG_DRAM_TRAINING_PATTERNS_OFFS);
672	}
673
674	reg |= (1 << REG_DRAM_TRAINING_AUTO_OFFS);
675
676	reg_write(REG_DRAM_TRAINING_ADDR, reg);
677
678	udelay(100);
679
680	/* Check if Successful */
681	if (reg_read(REG_DRAM_TRAINING_ADDR) &
682	    (1 << REG_DRAM_TRAINING_ERROR_OFFS))
683		return MV_OK;
684	else
685		return MV_FAIL;
686}
687
688#if !defined(MV88F67XX)
689/*
690 * Name:     ddr3_save_training(MV_DRAM_INFO *dram_info)
691 * Desc:     saves the training results to memeory (RL,WL,PBS,Rx/Tx
692 *           Centeralization)
693 * Args:     MV_DRAM_INFO *dram_info
694 * Notes:
695 * Returns:  None.
696 */
697void ddr3_save_training(MV_DRAM_INFO *dram_info)
698{
699	u32 val, pup, tmp_cs, cs, i, dq;
700	u32 crc = 0;
701	u32 regs = 0;
702	u32 *sdram_offset = (u32 *)RESUME_TRAINING_VALUES_ADDR;
703	u32 mode_config[MAX_TRAINING_MODE];
704
705	mode_config[DQS_WR_MODE] = PUP_DQS_WR;
706	mode_config[WL_MODE_] = PUP_WL_MODE;
707	mode_config[RL_MODE_] = PUP_RL_MODE;
708	mode_config[DQS_RD_MODE] = PUP_DQS_RD;
709	mode_config[PBS_TX_DM_MODE] = PUP_PBS_TX_DM;
710	mode_config[PBS_TX_MODE] = PUP_PBS_TX;
711	mode_config[PBS_RX_MODE] = PUP_PBS_RX;
712
713	/* num of training modes */
714	for (i = 0; i < MAX_TRAINING_MODE; i++) {
715		tmp_cs = dram_info->cs_ena;
716		/* num of CS */
717		for (cs = 0; cs < MAX_CS; cs++) {
718			if (tmp_cs & (1 << cs)) {
719				/* num of PUPs */
720				for (pup = 0; pup < dram_info->num_of_total_pups;
721				     pup++) {
722					if (pup == dram_info->num_of_std_pups &&
723					    dram_info->ecc_ena)
724						pup = ECC_PUP;
725					if (i == PBS_TX_DM_MODE) {
726						/*
727						 * Change CS bitmask because
728						 * PBS works only with CS0
729						 */
730						tmp_cs = 0x1;
731						val = ddr3_read_pup_reg(
732							mode_config[i], CS0, pup);
733					} else if (i == PBS_TX_MODE ||
734						   i == PBS_RX_MODE) {
735						/*
736						 * Change CS bitmask because
737						 * PBS works only with CS0
738						 */
739						tmp_cs = 0x1;
740						for (dq = 0; dq <= DQ_NUM;
741						     dq++) {
742							val = ddr3_read_pup_reg(
743								mode_config[i] + dq,
744								CS0,
745								pup);
746							(*sdram_offset) = val;
747							crc += *sdram_offset;
748							sdram_offset++;
749							regs++;
750						}
751						continue;
752					} else {
753						val = ddr3_read_pup_reg(
754							mode_config[i], cs, pup);
755					}
756
757					*sdram_offset = val;
758					crc += *sdram_offset;
759					sdram_offset++;
760					regs++;
761				}
762			}
763		}
764	}
765
766	*sdram_offset = reg_read(REG_READ_DATA_SAMPLE_DELAYS_ADDR);
767	crc += *sdram_offset;
768	sdram_offset++;
769	regs++;
770	*sdram_offset = reg_read(REG_READ_DATA_READY_DELAYS_ADDR);
771	crc += *sdram_offset;
772	sdram_offset++;
773	regs++;
774	sdram_offset = (u32 *)NUM_OF_REGISTER_ADDR;
775	*sdram_offset = regs;
776	DEBUG_SUSPEND_RESUME_S("Training Results CheckSum write= ");
777	DEBUG_SUSPEND_RESUME_D(crc, 8);
778	DEBUG_SUSPEND_RESUME_S("\n");
779	sdram_offset = (u32 *)CHECKSUM_RESULT_ADDR;
780	*sdram_offset = crc;
781}
782
783/*
784 * Name:     ddr3_read_training_results()
785 * Desc:     Reads the training results from memeory (RL,WL,PBS,Rx/Tx
786 *           Centeralization)
787 *           and writes them to the relevant registers
788 * Args:     MV_DRAM_INFO *dram_info
789 * Notes:
790 * Returns:  None.
791 */
792int ddr3_read_training_results(void)
793{
794	u32 val, reg, idx, dqs_wr_idx = 0, crc = 0;
795	u32 *sdram_offset = (u32 *)RESUME_TRAINING_VALUES_ADDR;
796	u32 training_val[RESUME_TRAINING_VALUES_MAX] = { 0 };
797	u32 regs = *((u32 *)NUM_OF_REGISTER_ADDR);
798
799	/*
800	 * Read Training results & Dunit registers from memory and write
801	 * it to an array
802	 */
803	for (idx = 0; idx < regs; idx++) {
804		training_val[idx] = *sdram_offset;
805		crc += *sdram_offset;
806		sdram_offset++;
807	}
808
809	sdram_offset = (u32 *)CHECKSUM_RESULT_ADDR;
810
811	if ((*sdram_offset) == crc) {
812		DEBUG_SUSPEND_RESUME_S("Training Results CheckSum read PASS= ");
813		DEBUG_SUSPEND_RESUME_D(crc, 8);
814		DEBUG_SUSPEND_RESUME_S("\n");
815	} else {
816		DEBUG_MAIN_S("Wrong Training Results CheckSum\n");
817		return MV_FAIL;
818	}
819
820	/*
821	 * We iterate through all the registers except for the last 2 since
822	 * they are Dunit registers (and not PHY registers)
823	 */
824	for (idx = 0; idx < (regs - 2); idx++) {
825		val = training_val[idx];
826		reg = (val >> REG_PHY_CS_OFFS) & 0x3F; /*read the phy address */
827
828		/* Check if the values belongs to the DQS WR */
829		if (reg == PUP_WL_MODE) {
830			/* bit[5:0] in DQS_WR are delay */
831			val = (training_val[dqs_wr_idx++] & 0x3F);
832			/*
833			 * bit[15:10] are DQS_WR delay & bit[9:0] are
834			 * WL phase & delay
835			 */
836			val = (val << REG_PHY_DQS_REF_DLY_OFFS) |
837				(training_val[idx] & 0x3C003FF);
838			/* Add Request pending and write operation bits */
839			val |= REG_PHY_REGISTRY_FILE_ACCESS_OP_WR;
840		} else if (reg == PUP_DQS_WR) {
841			/*
842			 * Do nothing since DQS_WR will be done in PUP_WL_MODE
843			 */
844			continue;
845		}
846
847		val |= REG_PHY_REGISTRY_FILE_ACCESS_OP_WR;
848		reg_write(REG_PHY_REGISTRY_FILE_ACCESS_ADDR, val);
849		do {
850			val = (reg_read(REG_PHY_REGISTRY_FILE_ACCESS_ADDR)) &
851				REG_PHY_REGISTRY_FILE_ACCESS_OP_DONE;
852		} while (val);	/* Wait for '0' to mark the end of the transaction */
853	}
854
855	/* write last 2 Dunit configurations */
856	val = training_val[idx];
857	reg_write(REG_READ_DATA_SAMPLE_DELAYS_ADDR, val);	/* reg 0x1538 */
858	val = training_val[idx + 1];
859	reg_write(REG_READ_DATA_READY_DELAYS_ADDR, val);	/* reg 0x153c */
860
861	return MV_OK;
862}
863
864/*
865 * Name:     ddr3_check_if_resume_mode()
866 * Desc:     Reads the address (0x3000) of the Resume Magic word (0xDEADB002)
867 * Args:     MV_DRAM_INFO *dram_info
868 * Notes:
869 * Returns:  return (magic_word == SUSPEND_MAGIC_WORD)
870 */
871int ddr3_check_if_resume_mode(MV_DRAM_INFO *dram_info, u32 freq)
872{
873	u32 magic_word;
874	u32 *sdram_offset = (u32 *)BOOT_INFO_ADDR;
875
876	if (dram_info->reg_dimm != 1) {
877		/*
878		 * Perform write levleling in order initiate the phy with
879		 * low frequency
880		 */
881		if (MV_OK != ddr3_write_leveling_hw(freq, dram_info)) {
882			DEBUG_MAIN_S("DDR3 Training Sequence - FAILED (Write Leveling Hw)\n");
883			return MV_DDR3_TRAINING_ERR_WR_LVL_HW;
884		}
885	}
886
887	if (MV_OK != ddr3_load_patterns(dram_info, 1)) {
888		DEBUG_MAIN_S("DDR3 Training Sequence - FAILED (Loading Patterns)\n");
889		return MV_DDR3_TRAINING_ERR_LOAD_PATTERNS;
890	}
891
892	/* Enable CS0 only for RL */
893	dram_info->cs_ena = 0x1;
894
895	/* Perform Read levleling in order to get stable memory */
896	if (MV_OK != ddr3_read_leveling_hw(freq, dram_info)) {
897		DEBUG_MAIN_S("DDR3 Training Sequence - FAILED (Read Leveling Hw)\n");
898		return MV_DDR3_TRAINING_ERR_WR_LVL_HW;
899	}
900
901	/* Back to relevant CS */
902	dram_info->cs_ena = ddr3_get_cs_ena_from_reg();
903
904	magic_word = *sdram_offset;
905	return magic_word == SUSPEND_MAGIC_WORD;
906}
907
908/*
909 * Name:     ddr3_training_suspend_resume()
910 * Desc:     Execute the Resume state
911 * Args:     MV_DRAM_INFO *dram_info
912 * Notes:
913 * Returns:  return (magic_word == SUSPEND_MAGIC_WORD)
914 */
915int ddr3_training_suspend_resume(MV_DRAM_INFO *dram_info)
916{
917	u32 freq, reg;
918	int tmp_ratio;
919
920	/* Configure DDR */
921	if (MV_OK != ddr3_read_training_results())
922		return MV_FAIL;
923
924	/* Reset read FIFO */
925	reg = reg_read(REG_DRAM_TRAINING_ADDR);
926
927	/* Start Auto Read Leveling procedure */
928	reg |= (1 << REG_DRAM_TRAINING_RL_OFFS);
929	reg_write(REG_DRAM_TRAINING_ADDR, reg);	/* 0x15B0 - Training Register */
930
931	reg = reg_read(REG_DRAM_TRAINING_2_ADDR);
932	reg |= ((1 << REG_DRAM_TRAINING_2_FIFO_RST_OFFS) +
933		(1 << REG_DRAM_TRAINING_2_SW_OVRD_OFFS));
934
935	/* [0] = 1 - Enable SW override, [4] = 1 - FIFO reset  */
936	/* 0x15B8 - Training SW 2 Register */
937	reg_write(REG_DRAM_TRAINING_2_ADDR, reg);
938
939	udelay(2);
940
941	reg = reg_read(REG_DRAM_TRAINING_ADDR);
942	/* Clear Auto Read Leveling procedure */
943	reg &= ~(1 << REG_DRAM_TRAINING_RL_OFFS);
944	reg_write(REG_DRAM_TRAINING_ADDR, reg);	/* 0x15B0 - Training Register */
945
946	/* Return to target frequency */
947	freq = dram_info->target_frequency;
948	tmp_ratio = 1;
949	if (MV_OK != ddr3_dfs_low_2_high(freq, tmp_ratio, dram_info)) {
950		DEBUG_MAIN_S("DDR3 Training Sequence - FAILED (Dfs Low2High)\n");
951		return MV_DDR3_TRAINING_ERR_DFS_H2L;
952	}
953
954	if (dram_info->ecc_ena) {
955		/* Scabbling the RL area pattern and the training area */
956		mv_sys_xor_finish();
957		dram_info->num_cs = 1;
958		dram_info->cs_ena = 1;
959		mv_sys_xor_init(dram_info);
960		mv_xor_mem_init(0, RESUME_RL_PATTERNS_ADDR,
961				RESUME_RL_PATTERNS_SIZE, 0xFFFFFFFF, 0xFFFFFFFF);
962
963		/* Wait for previous transfer completion */
964
965		while (mv_xor_state_get(0) != MV_IDLE)
966			;
967
968		/* Return XOR State */
969		mv_sys_xor_finish();
970	}
971
972	return MV_OK;
973}
974#endif
975
976void ddr3_print_freq(u32 freq)
977{
978	u32 tmp_freq;
979
980	switch (freq) {
981	case 0:
982		tmp_freq = 100;
983		break;
984	case 1:
985		tmp_freq = 300;
986		break;
987	case 2:
988		tmp_freq = 360;
989		break;
990	case 3:
991		tmp_freq = 400;
992		break;
993	case 4:
994		tmp_freq = 444;
995		break;
996	case 5:
997		tmp_freq = 500;
998		break;
999	case 6:
1000		tmp_freq = 533;
1001		break;
1002	case 7:
1003		tmp_freq = 600;
1004		break;
1005	case 8:
1006		tmp_freq = 666;
1007		break;
1008	case 9:
1009		tmp_freq = 720;
1010		break;
1011	case 10:
1012		tmp_freq = 800;
1013		break;
1014	default:
1015		tmp_freq = 100;
1016	}
1017
1018	printf("Current frequency is: %dMHz\n", tmp_freq);
1019}
1020
1021int ddr3_get_min_max_read_sample_delay(u32 cs_enable, u32 reg, u32 *min,
1022				       u32 *max, u32 *cs_max)
1023{
1024	u32 cs, delay;
1025
1026	*min = 0xFFFFFFFF;
1027	*max = 0x0;
1028
1029	for (cs = 0; cs < MAX_CS; cs++) {
1030		if ((cs_enable & (1 << cs)) == 0)
1031			continue;
1032
1033		delay = ((reg >> (cs * 8)) & 0x1F);
1034
1035		if (delay < *min)
1036			*min = delay;
1037
1038		if (delay > *max) {
1039			*max = delay;
1040			*cs_max = cs;
1041		}
1042	}
1043
1044	return MV_OK;
1045}
1046
1047int ddr3_get_min_max_rl_phase(MV_DRAM_INFO *dram_info, u32 *min, u32 *max,
1048			      u32 cs)
1049{
1050	u32 pup, reg, phase;
1051
1052	*min = 0xFFFFFFFF;
1053	*max = 0x0;
1054
1055	for (pup = 0; pup < dram_info->num_of_total_pups; pup++) {
1056		reg = ddr3_read_pup_reg(PUP_RL_MODE, cs, pup);
1057		phase = ((reg >> 8) & 0x7);
1058
1059		if (phase < *min)
1060			*min = phase;
1061
1062		if (phase > *max)
1063			*max = phase;
1064	}
1065
1066	return MV_OK;
1067}
1068
1069int ddr3_odt_activate(int activate)
1070{
1071	u32 reg, mask;
1072
1073	mask = (1 << REG_DUNIT_ODT_CTRL_OVRD_OFFS) |
1074		(1 << REG_DUNIT_ODT_CTRL_OVRD_VAL_OFFS);
1075	/* {0x0000149C}  -   DDR Dunit ODT Control Register */
1076	reg = reg_read(REG_DUNIT_ODT_CTRL_ADDR);
1077	if (activate)
1078		reg |= mask;
1079	else
1080		reg &= ~mask;
1081
1082	reg_write(REG_DUNIT_ODT_CTRL_ADDR, reg);
1083
1084	return MV_OK;
1085}
1086
1087int ddr3_odt_read_dynamic_config(MV_DRAM_INFO *dram_info)
1088{
1089	u32 min_read_sample_delay, max_read_sample_delay, max_rl_phase;
1090	u32 min, max, cs_max;
1091	u32 cs_ena, reg;
1092
1093	reg = reg_read(REG_READ_DATA_SAMPLE_DELAYS_ADDR);
1094	cs_ena = ddr3_get_cs_ena_from_reg();
1095
1096	/* Get minimum and maximum of read sample delay of all CS */
1097	ddr3_get_min_max_read_sample_delay(cs_ena, reg, &min_read_sample_delay,
1098					   &max_read_sample_delay, &cs_max);
1099
1100	/*
1101	 * Get minimum and maximum read leveling phase which belongs to the
1102	 * maximal read sample delay
1103	 */
1104	ddr3_get_min_max_rl_phase(dram_info, &min, &max, cs_max);
1105	max_rl_phase = max;
1106
1107	/* DDR ODT Timing (Low) Register calculation */
1108	reg = reg_read(REG_ODT_TIME_LOW_ADDR);
1109	reg &= ~(0x1FF << REG_ODT_ON_CTL_RD_OFFS);
1110	reg |= (((min_read_sample_delay - 1) & 0xF) << REG_ODT_ON_CTL_RD_OFFS);
1111	reg |= (((max_read_sample_delay + 4 + (((max_rl_phase + 1) / 2) + 1)) &
1112		 0x1F) << REG_ODT_OFF_CTL_RD_OFFS);
1113	reg_write(REG_ODT_TIME_LOW_ADDR, reg);
1114
1115	return MV_OK;
1116}
1117