docs: Add 24.12 release notes
[coreboot2.git] / src / soc / rockchip / rk3399 / sdram.c
blobbe1da598a8c82098e7429d7094db02542a3ee98c
1 /* SPDX-License-Identifier: GPL-2.0-only */
3 #include <device/mmio.h>
4 #include <console/console.h>
5 #include <delay.h>
6 #include <reset.h>
7 #include <soc/addressmap.h>
8 #include <soc/clock.h>
9 #include <soc/sdram.h>
10 #include <soc/grf.h>
11 #include <soc/soc.h>
12 #include <timer.h>
13 #include <types.h>
15 #define DDR_PI_OFFSET 0x800
16 #define DDR_PHY_OFFSET 0x2000
17 #define DDRC0_PI_BASE_ADDR (DDRC0_BASE_ADDR + DDR_PI_OFFSET)
18 #define DDRC0_PHY_BASE_ADDR (DDRC0_BASE_ADDR + DDR_PHY_OFFSET)
19 #define DDRC1_PI_BASE_ADDR (DDRC1_BASE_ADDR + DDR_PI_OFFSET)
20 #define DDRC1_PHY_BASE_ADDR (DDRC1_BASE_ADDR + DDR_PHY_OFFSET)
22 static struct rk3399_ddr_pctl_regs * const rk3399_ddr_pctl[2] = {
23 (void *)DDRC0_BASE_ADDR, (void *)DDRC1_BASE_ADDR };
24 static struct rk3399_ddr_pi_regs * const rk3399_ddr_pi[2] = {
25 (void *)DDRC0_PI_BASE_ADDR, (void *)DDRC1_PI_BASE_ADDR };
26 static struct rk3399_ddr_publ_regs * const rk3399_ddr_publ[2] = {
27 (void *)DDRC0_PHY_BASE_ADDR, (void *)DDRC1_PHY_BASE_ADDR };
28 static struct rk3399_msch_regs * const rk3399_msch[2] = {
29 (void *)SERVER_MSCH0_BASE_ADDR, (void *)SERVER_MSCH1_BASE_ADDR };
30 static struct rk3399_ddr_cic_regs *const rk3399_ddr_cic = (void *)CIC_BASE_ADDR;
33 * sys_reg bitfield struct
34 * [31] row_3_4_ch1
35 * [30] row_3_4_ch0
36 * [29:28] chinfo
37 * [27] rank_ch1
38 * [26:25] col_ch1
39 * [24] bk_ch1
40 * [23:22] cs0_row_ch1
41 * [21:20] cs1_row_ch1
42 * [19:18] bw_ch1
43 * [17:16] dbw_ch1;
44 * [15:13] ddrtype
45 * [12] channelnum
46 * [11] rank_ch0
47 * [10:9] col_ch0
48 * [8] bk_ch0
49 * [7:6] cs0_row_ch0
50 * [5:4] cs1_row_ch0
51 * [3:2] bw_ch0
52 * [1:0] dbw_ch0
54 #define SYS_REG_ENC_ROW_3_4(n, ch) ((n) << (30 + (ch)))
55 #define SYS_REG_DEC_ROW_3_4(n, ch) ((n >> (30 + ch)) & 0x1)
56 #define SYS_REG_ENC_CHINFO(ch) (1 << (28 + (ch)))
57 #define SYS_REG_ENC_DDRTYPE(n) ((n) << 13)
58 #define SYS_REG_ENC_NUM_CH(n) (((n) - 1) << 12)
59 #define SYS_REG_DEC_NUM_CH(n) (1 + ((n >> 12) & 0x1))
60 #define SYS_REG_ENC_RANK(n, ch) (((n) - 1) << (11 + ((ch) * 16)))
61 #define SYS_REG_DEC_RANK(n, ch) (1 + ((n >> (11 + 16 * ch)) & 0x1))
62 #define SYS_REG_ENC_COL(n, ch) (((n) - 9) << (9 + ((ch) * 16)))
63 #define SYS_REG_DEC_COL(n, ch) (9 + ((n >> (9 + 16 * ch)) & 0x3))
64 #define SYS_REG_ENC_BK(n, ch) (((n) == 3 ? 0 : 1) \
65 << (8 + ((ch) * 16)))
66 #define SYS_REG_DEC_BK(n, ch) (3 - ((n >> (8 + 16 * ch)) & 0x1))
67 #define SYS_REG_ENC_CS0_ROW(n, ch) (((n) - 13) << (6 + ((ch) * 16)))
68 #define SYS_REG_DEC_CS0_ROW(n, ch) (13 + ((n >> (6 + 16 * ch)) & 0x3))
69 #define SYS_REG_ENC_CS1_ROW(n, ch) (((n) - 13) << (4 + ((ch) * 16)))
70 #define SYS_REG_DEC_CS1_ROW(n, ch) (13 + ((n >> (4 + 16 * ch)) & 0x3))
71 #define SYS_REG_ENC_BW(n, ch) ((2 >> (n)) << (2 + ((ch) * 16)))
72 #define SYS_REG_DEC_BW(n, ch) (2 >> ((n >> (2 + 16 * ch)) & 0x3))
73 #define SYS_REG_ENC_DBW(n, ch) ((2 >> (n)) << (0 + ((ch) * 16)))
74 #define SYS_REG_DEC_DBW(n, ch) (2 >> ((n >> (0 + 16 * ch)) & 0x3))
76 #define DDR_STRIDE(n) write32(&rk3399_pmusgrf->soc_con4,\
77 (0x1F << (10 + 16)) | (n << 10))
79 #define PRESET_SGRF_HOLD(n) ((0x1 << (6+16)) | ((n) << 6))
80 #define PRESET_GPIO0_HOLD(n) ((0x1 << (7+16)) | ((n) << 7))
81 #define PRESET_GPIO1_HOLD(n) ((0x1 << (8+16)) | ((n) << 8))
83 #define PHY_DRV_ODT_Hi_Z (0x0)
84 #define PHY_DRV_ODT_240 (0x1)
85 #define PHY_DRV_ODT_120 (0x8)
86 #define PHY_DRV_ODT_80 (0x9)
87 #define PHY_DRV_ODT_60 (0xc)
88 #define PHY_DRV_ODT_48 (0xd)
89 #define PHY_DRV_ODT_40 (0xe)
90 #define PHY_DRV_ODT_34_3 (0xf)
92 #define MAX_RANKS_PER_CHANNEL 4
94 static void copy_to_reg(u32 *dest, const u32 *src, u32 n)
96 int i;
98 for (i = 0; i < n / sizeof(u32); i++) {
99 write32(dest, *src);
100 src++;
101 dest++;
105 static void phy_pctrl_reset(u32 channel)
107 rkclk_ddr_reset(channel, 1, 1);
108 udelay(10);
110 rkclk_ddr_reset(channel, 1, 0);
111 udelay(10);
113 rkclk_ddr_reset(channel, 0, 0);
114 udelay(10);
117 static void phy_dll_bypass_set(struct rk3399_ddr_publ_regs *ddr_publ_regs,
118 u32 freq)
120 u32 *denali_phy = ddr_publ_regs->denali_phy;
122 if (freq <= 125*MHz) {
123 /* phy_sw_master_mode_X PHY_86/214/342/470 4bits offset_8 */
124 setbits32(&denali_phy[86], (0x3 << 2) << 8);
125 setbits32(&denali_phy[214], (0x3 << 2) << 8);
126 setbits32(&denali_phy[342], (0x3 << 2) << 8);
127 setbits32(&denali_phy[470], (0x3 << 2) << 8);
129 /* phy_adrctl_sw_master_mode PHY_547/675/803 4bits offset_16 */
130 setbits32(&denali_phy[547], (0x3 << 2) << 16);
131 setbits32(&denali_phy[675], (0x3 << 2) << 16);
132 setbits32(&denali_phy[803], (0x3 << 2) << 16);
133 } else {
134 /* phy_sw_master_mode_X PHY_86/214/342/470 4bits offset_8 */
135 clrbits32(&denali_phy[86], (0x3 << 2) << 8);
136 clrbits32(&denali_phy[214], (0x3 << 2) << 8);
137 clrbits32(&denali_phy[342], (0x3 << 2) << 8);
138 clrbits32(&denali_phy[470], (0x3 << 2) << 8);
140 /* phy_adrctl_sw_master_mode PHY_547/675/803 4bits offset_16 */
141 clrbits32(&denali_phy[547], (0x3 << 2) << 16);
142 clrbits32(&denali_phy[675], (0x3 << 2) << 16);
143 clrbits32(&denali_phy[803], (0x3 << 2) << 16);
147 static void set_memory_map(u32 channel, const struct rk3399_sdram_params *params)
149 const struct rk3399_sdram_channel *sdram_ch = &params->ch[channel];
150 u32 *denali_ctl = rk3399_ddr_pctl[channel]->denali_ctl;
151 u32 *denali_pi = rk3399_ddr_pi[channel]->denali_pi;
152 u32 cs_map;
153 u32 reduc;
154 u32 row;
156 if ((sdram_ch->ddrconfig < 2) || (sdram_ch->ddrconfig == 4))
157 row = 16;
158 else if (sdram_ch->ddrconfig == 3)
159 row = 14;
160 else
161 row = 15;
163 cs_map = (sdram_ch->rank > 1) ? 3 : 1;
164 reduc = (sdram_ch->bw == 2) ? 0 : 1;
166 clrsetbits32(&denali_ctl[191], 0xF, (12 - sdram_ch->col));
167 clrsetbits32(&denali_ctl[190], (0x3 << 16) | (0x7 << 24),
168 ((3 - sdram_ch->bk) << 16) |
169 ((16 - row) << 24));
171 clrsetbits32(&denali_ctl[196], 0x3 | (1 << 16),
172 cs_map | (reduc << 16));
174 /* PI_199 PI_COL_DIFF:RW:0:4 */
175 clrsetbits32(&denali_pi[199], 0xF, (12 - sdram_ch->col));
177 /* PI_155 PI_ROW_DIFF:RW:24:3 PI_BANK_DIFF:RW:16:2 */
178 clrsetbits32(&denali_pi[155], (0x3 << 16) | (0x7 << 24),
179 ((3 - sdram_ch->bk) << 16) |
180 ((16 - row) << 24));
181 /* PI_41 PI_CS_MAP:RW:24:4 */
182 clrsetbits32(&denali_pi[41], 0xf << 24, cs_map << 24);
183 if ((sdram_ch->rank == 1) && (params->dramtype == DDR3))
184 write32(&denali_pi[34], 0x2EC7FFFF);
187 static void set_ds_odt(u32 channel, const struct rk3399_sdram_params *params)
189 u32 *denali_phy = rk3399_ddr_publ[channel]->denali_phy;
191 u32 tsel_idle_en, tsel_wr_en, tsel_rd_en;
192 u32 tsel_idle_select_p, tsel_wr_select_p, tsel_rd_select_p;
193 u32 ca_tsel_wr_select_p, ca_tsel_wr_select_n;
194 u32 tsel_idle_select_n, tsel_wr_select_n, tsel_rd_select_n;
195 u32 reg_value;
197 if (params->dramtype == LPDDR4) {
198 tsel_rd_select_p = PHY_DRV_ODT_Hi_Z;
199 tsel_rd_select_n = PHY_DRV_ODT_240;
201 tsel_wr_select_p = PHY_DRV_ODT_40;
202 tsel_wr_select_n = PHY_DRV_ODT_40;
204 tsel_idle_select_p = PHY_DRV_ODT_Hi_Z;
205 tsel_idle_select_n = PHY_DRV_ODT_240;
207 ca_tsel_wr_select_p = PHY_DRV_ODT_40;
208 ca_tsel_wr_select_n = PHY_DRV_ODT_40;
209 } else if (params->dramtype == LPDDR3) {
210 tsel_rd_select_p = PHY_DRV_ODT_240;
211 tsel_rd_select_n = PHY_DRV_ODT_Hi_Z;
213 tsel_wr_select_p = PHY_DRV_ODT_34_3;
214 tsel_wr_select_n = PHY_DRV_ODT_34_3;
216 tsel_idle_select_p = PHY_DRV_ODT_240;
217 tsel_idle_select_n = PHY_DRV_ODT_Hi_Z;
219 ca_tsel_wr_select_p = PHY_DRV_ODT_48;
220 ca_tsel_wr_select_n = PHY_DRV_ODT_48;
221 } else {
222 tsel_rd_select_p = PHY_DRV_ODT_240;
223 tsel_rd_select_n = PHY_DRV_ODT_240;
225 tsel_wr_select_p = PHY_DRV_ODT_34_3;
226 tsel_wr_select_n = PHY_DRV_ODT_34_3;
228 tsel_idle_select_p = PHY_DRV_ODT_240;
229 tsel_idle_select_n = PHY_DRV_ODT_240;
231 ca_tsel_wr_select_p = PHY_DRV_ODT_34_3;
232 ca_tsel_wr_select_n = PHY_DRV_ODT_34_3;
235 if (params->odt == 1)
236 tsel_rd_en = 1;
237 else
238 tsel_rd_en = 0;
240 tsel_wr_en = 0;
241 tsel_idle_en = 0;
244 * phy_dq_tsel_select_X 24bits DENALI_PHY_6/134/262/390 offset_0
245 * sets termination values for read/idle cycles and drive strength
246 * for write cycles for DQ/DM
248 reg_value = tsel_rd_select_n | (tsel_rd_select_p << 0x4) |
249 (tsel_wr_select_n << 8) | (tsel_wr_select_p << 12) |
250 (tsel_idle_select_n << 16) | (tsel_idle_select_p << 20);
251 clrsetbits32(&denali_phy[6], 0xffffff, reg_value);
252 clrsetbits32(&denali_phy[134], 0xffffff, reg_value);
253 clrsetbits32(&denali_phy[262], 0xffffff, reg_value);
254 clrsetbits32(&denali_phy[390], 0xffffff, reg_value);
257 * phy_dqs_tsel_select_X 24bits DENALI_PHY_7/135/263/391 offset_0
258 * sets termination values for read/idle cycles and drive strength
259 * for write cycles for DQS
261 clrsetbits32(&denali_phy[7], 0xffffff, reg_value);
262 clrsetbits32(&denali_phy[135], 0xffffff, reg_value);
263 clrsetbits32(&denali_phy[263], 0xffffff, reg_value);
264 clrsetbits32(&denali_phy[391], 0xffffff, reg_value);
266 /* phy_adr_tsel_select_ 8bits DENALI_PHY_544/672/800 offset_0 */
267 reg_value = ca_tsel_wr_select_n | (ca_tsel_wr_select_p << 0x4);
268 clrsetbits32(&denali_phy[544], 0xff, reg_value);
269 clrsetbits32(&denali_phy[672], 0xff, reg_value);
270 clrsetbits32(&denali_phy[800], 0xff, reg_value);
272 /* phy_pad_addr_drive 8bits DENALI_PHY_928 offset_0 */
273 clrsetbits32(&denali_phy[928], 0xff, reg_value);
275 /* phy_pad_rst_drive 8bits DENALI_PHY_937 offset_0 */
276 clrsetbits32(&denali_phy[937], 0xff, reg_value);
278 /* phy_pad_cke_drive 8bits DENALI_PHY_935 offset_0 */
279 clrsetbits32(&denali_phy[935], 0xff, reg_value);
281 /* phy_pad_cs_drive 8bits DENALI_PHY_939 offset_0 */
282 clrsetbits32(&denali_phy[939], 0xff, reg_value);
284 /* phy_pad_clk_drive 8bits DENALI_PHY_929 offset_0 */
285 clrsetbits32(&denali_phy[929], 0xff, reg_value);
287 /* phy_pad_fdbk_drive 23bit DENALI_PHY_924/925 */
288 clrsetbits32(&denali_phy[924], 0xff,
289 tsel_wr_select_n | (tsel_wr_select_p << 4));
290 clrsetbits32(&denali_phy[925], 0xff,
291 tsel_rd_select_n | (tsel_rd_select_p << 4));
293 /* phy_dq_tsel_enable_X 3bits DENALI_PHY_5/133/261/389 offset_16 */
294 reg_value = (tsel_rd_en | (tsel_wr_en << 1) | (tsel_idle_en << 2))
295 << 16;
296 clrsetbits32(&denali_phy[5], 0x7 << 16, reg_value);
297 clrsetbits32(&denali_phy[133], 0x7 << 16, reg_value);
298 clrsetbits32(&denali_phy[261], 0x7 << 16, reg_value);
299 clrsetbits32(&denali_phy[389], 0x7 << 16, reg_value);
301 /* phy_dqs_tsel_enable_X 3bits DENALI_PHY_6/134/262/390 offset_24 */
302 reg_value = (tsel_rd_en | (tsel_wr_en << 1) | (tsel_idle_en << 2))
303 << 24;
304 clrsetbits32(&denali_phy[6], 0x7 << 24, reg_value);
305 clrsetbits32(&denali_phy[134], 0x7 << 24, reg_value);
306 clrsetbits32(&denali_phy[262], 0x7 << 24, reg_value);
307 clrsetbits32(&denali_phy[390], 0x7 << 24, reg_value);
309 /* phy_adr_tsel_enable_ 1bit DENALI_PHY_518/646/774 offset_8 */
310 reg_value = tsel_wr_en << 8;
311 clrsetbits32(&denali_phy[518], 0x1 << 8, reg_value);
312 clrsetbits32(&denali_phy[646], 0x1 << 8, reg_value);
313 clrsetbits32(&denali_phy[774], 0x1 << 8, reg_value);
315 /* phy_pad_addr_term tsel 1bit DENALI_PHY_933 offset_17 */
316 reg_value = tsel_wr_en << 17;
317 clrsetbits32(&denali_phy[933], 0x1 << 17, reg_value);
319 * pad_rst/cke/cs/clk_term tsel 1bits
320 * DENALI_PHY_938/936/940/934 offset_17
322 clrsetbits32(&denali_phy[938], 0x1 << 17, reg_value);
323 clrsetbits32(&denali_phy[936], 0x1 << 17, reg_value);
324 clrsetbits32(&denali_phy[940], 0x1 << 17, reg_value);
325 clrsetbits32(&denali_phy[934], 0x1 << 17, reg_value);
327 /* phy_pad_fdbk_term 1bit DENALI_PHY_930 offset_17 */
328 clrsetbits32(&denali_phy[930], 0x1 << 17, reg_value);
331 static void phy_io_config(u32 channel, const struct rk3399_sdram_params *params)
333 u32 *denali_phy = rk3399_ddr_publ[channel]->denali_phy;
334 u32 vref_mode_dq, vref_value_dq, vref_mode_ac, vref_value_ac;
335 u32 mode_sel = 0;
336 u32 reg_value;
337 u32 drv_value, odt_value;
338 u32 speed;
340 /* vref setting */
341 if (params->dramtype == LPDDR4) {
342 /* LPDDR4 */
343 vref_mode_dq = 0x6;
344 vref_value_dq = 0x1f;
345 vref_mode_ac = 0x6;
346 vref_value_ac = 0x1f;
347 } else if (params->dramtype == LPDDR3) {
348 if (params->odt == 1) {
349 vref_mode_dq = 0x5; /* LPDDR3 ODT */
350 drv_value = (read32(&denali_phy[6]) >> 12) & 0xf;
351 odt_value = (read32(&denali_phy[6]) >> 4) & 0xf;
352 if (drv_value == PHY_DRV_ODT_48) {
353 switch (odt_value) {
354 case PHY_DRV_ODT_240:
355 vref_value_dq = 0x16;
356 break;
357 case PHY_DRV_ODT_120:
358 vref_value_dq = 0x26;
359 break;
360 case PHY_DRV_ODT_60:
361 vref_value_dq = 0x36;
362 break;
363 default:
364 die("Halting: Invalid ODT value.\n");
366 } else if (drv_value == PHY_DRV_ODT_40) {
367 switch (odt_value) {
368 case PHY_DRV_ODT_240:
369 vref_value_dq = 0x19;
370 break;
371 case PHY_DRV_ODT_120:
372 vref_value_dq = 0x23;
373 break;
374 case PHY_DRV_ODT_60:
375 vref_value_dq = 0x31;
376 break;
377 default:
378 die("Halting: Invalid ODT value.\n");
380 } else if (drv_value == PHY_DRV_ODT_34_3) {
381 switch (odt_value) {
382 case PHY_DRV_ODT_240:
383 vref_value_dq = 0x17;
384 break;
385 case PHY_DRV_ODT_120:
386 vref_value_dq = 0x20;
387 break;
388 case PHY_DRV_ODT_60:
389 vref_value_dq = 0x2e;
390 break;
391 default:
392 die("Halting: Invalid ODT value.\n");
394 } else {
395 die("Halting: Invalid DRV value.\n");
397 } else {
398 vref_mode_dq = 0x2; /* LPDDR3 */
399 vref_value_dq = 0x1f;
401 vref_mode_ac = 0x2;
402 vref_value_ac = 0x1f;
403 } else if (params->dramtype == DDR3) {
404 /* DDR3L */
405 vref_mode_dq = 0x1;
406 vref_value_dq = 0x1f;
407 vref_mode_ac = 0x1;
408 vref_value_ac = 0x1f;
409 } else {
410 die("Halting: Unknown DRAM type.\n");
413 reg_value = (vref_mode_dq << 9) | (0x1 << 8) | vref_value_dq;
415 /* PHY_913 PHY_PAD_VREF_CTRL_DQ_0 12bits offset_8 */
416 clrsetbits32(&denali_phy[913], 0xfff << 8, reg_value << 8);
417 /* PHY_914 PHY_PAD_VREF_CTRL_DQ_1 12bits offset_0 */
418 clrsetbits32(&denali_phy[914], 0xfff, reg_value);
419 /* PHY_914 PHY_PAD_VREF_CTRL_DQ_2 12bits offset_16 */
420 clrsetbits32(&denali_phy[914], 0xfff << 16, reg_value << 16);
421 /* PHY_915 PHY_PAD_VREF_CTRL_DQ_3 12bits offset_0 */
422 clrsetbits32(&denali_phy[915], 0xfff, reg_value);
424 reg_value = (vref_mode_ac << 9) | (0x1 << 8) | vref_value_ac;
426 /* PHY_915 PHY_PAD_VREF_CTRL_AC 12bits offset_16 */
427 clrsetbits32(&denali_phy[915], 0xfff << 16, reg_value << 16);
429 if (params->dramtype == LPDDR4)
430 mode_sel = 0x6;
431 else if (params->dramtype == LPDDR3)
432 mode_sel = 0x0;
433 else if (params->dramtype == DDR3)
434 mode_sel = 0x1;
436 /* PHY_924 PHY_PAD_FDBK_DRIVE */
437 clrsetbits32(&denali_phy[924], 0x7 << 15, mode_sel << 15);
438 /* PHY_926 PHY_PAD_DATA_DRIVE */
439 clrsetbits32(&denali_phy[926], 0x7 << 6, mode_sel << 6);
440 /* PHY_927 PHY_PAD_DQS_DRIVE */
441 clrsetbits32(&denali_phy[927], 0x7 << 6, mode_sel << 6);
442 /* PHY_928 PHY_PAD_ADDR_DRIVE */
443 clrsetbits32(&denali_phy[928], 0x7 << 14, mode_sel << 14);
444 /* PHY_929 PHY_PAD_CLK_DRIVE */
445 clrsetbits32(&denali_phy[929], 0x7 << 14, mode_sel << 14);
446 /* PHY_935 PHY_PAD_CKE_DRIVE */
447 clrsetbits32(&denali_phy[935], 0x7 << 14, mode_sel << 14);
448 /* PHY_937 PHY_PAD_RST_DRIVE */
449 clrsetbits32(&denali_phy[937], 0x7 << 14, mode_sel << 14);
450 /* PHY_939 PHY_PAD_CS_DRIVE */
451 clrsetbits32(&denali_phy[939], 0x7 << 14, mode_sel << 14);
453 /* speed setting */
454 if (params->ddr_freq < 400 * MHz)
455 speed = 0x0;
456 else if (params->ddr_freq < 800 * MHz)
457 speed = 0x1;
458 else if (params->ddr_freq < 1200 * MHz)
459 speed = 0x2;
460 else
461 speed = 0x3;
463 /* PHY_924 PHY_PAD_FDBK_DRIVE */
464 clrsetbits32(&denali_phy[924], 0x3 << 21, speed << 21);
465 /* PHY_926 PHY_PAD_DATA_DRIVE */
466 clrsetbits32(&denali_phy[926], 0x3 << 9, speed << 9);
467 /* PHY_927 PHY_PAD_DQS_DRIVE */
468 clrsetbits32(&denali_phy[927], 0x3 << 9, speed << 9);
469 /* PHY_928 PHY_PAD_ADDR_DRIVE */
470 clrsetbits32(&denali_phy[928], 0x3 << 17, speed << 17);
471 /* PHY_929 PHY_PAD_CLK_DRIVE */
472 clrsetbits32(&denali_phy[929], 0x3 << 17, speed << 17);
473 /* PHY_935 PHY_PAD_CKE_DRIVE */
474 clrsetbits32(&denali_phy[935], 0x3 << 17, speed << 17);
475 /* PHY_937 PHY_PAD_RST_DRIVE */
476 clrsetbits32(&denali_phy[937], 0x3 << 17, speed << 17);
477 /* PHY_939 PHY_PAD_CS_DRIVE */
478 clrsetbits32(&denali_phy[939], 0x3 << 17, speed << 17);
481 static int pctl_cfg(u32 channel, const struct rk3399_sdram_params *params)
483 u32 *denali_ctl = rk3399_ddr_pctl[channel]->denali_ctl;
484 u32 *denali_pi = rk3399_ddr_pi[channel]->denali_pi;
485 u32 *denali_phy = rk3399_ddr_publ[channel]->denali_phy;
486 const u32 *params_ctl = params->pctl_regs.denali_ctl;
487 const u32 *params_phy = params->phy_regs.denali_phy;
488 u32 tmp, tmp1, tmp2;
489 u32 pwrup_srefresh_exit;
490 struct stopwatch sw;
493 * work around controller bug:
494 * Do not program DRAM_CLASS until NO_PHY_IND_TRAIN_INT is programmed
496 copy_to_reg(&denali_ctl[1], &params_ctl[1],
497 sizeof(struct rk3399_ddr_pctl_regs) - 4);
498 write32(&denali_ctl[0], params_ctl[0]);
499 copy_to_reg(denali_pi, &params->pi_regs.denali_pi[0],
500 sizeof(struct rk3399_ddr_pi_regs));
501 /* rank count need to set for init */
502 set_memory_map(channel, params);
504 write32(&denali_phy[910], params->phy_regs.denali_phy[910]);
505 write32(&denali_phy[911], params->phy_regs.denali_phy[911]);
506 write32(&denali_phy[912], params->phy_regs.denali_phy[912]);
508 pwrup_srefresh_exit = read32(&denali_ctl[68]) & PWRUP_SREFRESH_EXIT;
509 clrbits32(&denali_ctl[68], PWRUP_SREFRESH_EXIT);
511 /* PHY_DLL_RST_EN */
512 clrsetbits32(&denali_phy[957], 0x3 << 24, 1 << 24);
514 setbits32(&denali_pi[0], START);
515 setbits32(&denali_ctl[0], START);
517 while (1) {
518 tmp = read32(&denali_phy[920]);
519 tmp1 = read32(&denali_phy[921]);
520 tmp2 = read32(&denali_phy[922]);
521 if ((((tmp >> 16) & 0x1) == 0x1) &&
522 (((tmp1 >> 16) & 0x1) == 0x1) &&
523 (((tmp1 >> 0) & 0x1) == 0x1) &&
524 (((tmp2 >> 0) & 0x1) == 0x1))
525 break;
528 copy_to_reg(&denali_phy[896], &params_phy[896], (958 - 895) * 4);
529 copy_to_reg(&denali_phy[0], &params_phy[0], (90 - 0 + 1) * 4);
530 copy_to_reg(&denali_phy[128], &params_phy[128], (218 - 128 + 1) * 4);
531 copy_to_reg(&denali_phy[256], &params_phy[256], (346 - 256 + 1) * 4);
532 copy_to_reg(&denali_phy[384], &params_phy[384], (474 - 384 + 1) * 4);
533 copy_to_reg(&denali_phy[512], &params_phy[512], (549 - 512 + 1) * 4);
534 copy_to_reg(&denali_phy[640], &params_phy[640], (677 - 640 + 1) * 4);
535 copy_to_reg(&denali_phy[768], &params_phy[768], (805 - 768 + 1) * 4);
536 set_ds_odt(channel, params);
539 * phy_dqs_tsel_wr_timing_X 8bits DENALI_PHY_84/212/340/468 offset_8
540 * dqs_tsel_wr_end[7:4] add Half cycle
542 tmp = (read32(&denali_phy[84]) >> 8) & 0xff;
543 clrsetbits32(&denali_phy[84], 0xff << 8, (tmp + 0x10) << 8);
544 tmp = (read32(&denali_phy[212]) >> 8) & 0xff;
545 clrsetbits32(&denali_phy[212], 0xff << 8, (tmp + 0x10) << 8);
546 tmp = (read32(&denali_phy[340]) >> 8) & 0xff;
547 clrsetbits32(&denali_phy[340], 0xff << 8, (tmp + 0x10) << 8);
548 tmp = (read32(&denali_phy[468]) >> 8) & 0xff;
549 clrsetbits32(&denali_phy[468], 0xff << 8, (tmp + 0x10) << 8);
552 * phy_dqs_tsel_wr_timing_X 8bits DENALI_PHY_83/211/339/467 offset_8
553 * dq_tsel_wr_end[7:4] add Half cycle
555 tmp = (read32(&denali_phy[83]) >> 16) & 0xff;
556 clrsetbits32(&denali_phy[83], 0xff << 16, (tmp + 0x10) << 16);
557 tmp = (read32(&denali_phy[211]) >> 16) & 0xff;
558 clrsetbits32(&denali_phy[211], 0xff << 16, (tmp + 0x10) << 16);
559 tmp = (read32(&denali_phy[339]) >> 16) & 0xff;
560 clrsetbits32(&denali_phy[339], 0xff << 16, (tmp + 0x10) << 16);
561 tmp = (read32(&denali_phy[467]) >> 16) & 0xff;
562 clrsetbits32(&denali_phy[467], 0xff << 16, (tmp + 0x10) << 16);
564 phy_io_config(channel, params);
566 /* PHY_DLL_RST_EN */
567 clrsetbits32(&denali_phy[957], 0x3 << 24, 0x2 << 24);
569 /* FIXME: need to care ERROR bit */
570 stopwatch_init_msecs_expire(&sw, 100);
571 while (!(read32(&denali_ctl[203]) & (1 << 3))) {
572 if (stopwatch_expired(&sw))
573 return -1;
576 clrsetbits32(&denali_ctl[68], PWRUP_SREFRESH_EXIT,
577 pwrup_srefresh_exit);
578 return 0;
581 static void select_per_cs_training_index(u32 channel, u32 rank)
583 u32 *denali_phy = rk3399_ddr_publ[channel]->denali_phy;
585 /* PHY_84 PHY_PER_CS_TRAINING_EN_0 1bit offset_16 */
586 if ((read32(&denali_phy[84])>>16) & 1) {
588 * PHY_8/136/264/392
589 * phy_per_cs_training_index_X 1bit offset_24
591 clrsetbits32(&denali_phy[8], 0x1 << 24, rank << 24);
592 clrsetbits32(&denali_phy[136], 0x1 << 24, rank << 24);
593 clrsetbits32(&denali_phy[264], 0x1 << 24, rank << 24);
594 clrsetbits32(&denali_phy[392], 0x1 << 24, rank << 24);
598 static void override_write_leveling_value(u32 channel)
600 u32 *denali_ctl = rk3399_ddr_pctl[channel]->denali_ctl;
601 u32 *denali_phy = rk3399_ddr_publ[channel]->denali_phy;
602 u32 byte;
604 /* PHY_896 PHY_FREQ_SEL_MULTICAST_EN 1bit offset_0 */
605 setbits32(&denali_phy[896], 1);
608 * PHY_8/136/264/392
609 * phy_per_cs_training_multicast_en_X 1bit offset_16
611 clrsetbits32(&denali_phy[8], 0x1 << 16, 1 << 16);
612 clrsetbits32(&denali_phy[136], 0x1 << 16, 1 << 16);
613 clrsetbits32(&denali_phy[264], 0x1 << 16, 1 << 16);
614 clrsetbits32(&denali_phy[392], 0x1 << 16, 1 << 16);
616 for (byte = 0; byte < 4; byte++)
617 clrsetbits32(&denali_phy[63 + (128 * byte)], 0xffff << 16,
618 0x200 << 16);
620 /* PHY_896 PHY_FREQ_SEL_MULTICAST_EN 1bit offset_0 */
621 clrbits32(&denali_phy[896], 1);
623 /* CTL_200 ctrlupd_req 1bit offset_8 */
624 clrsetbits32(&denali_ctl[200], 0x1 << 8, 0x1 << 8);
627 static u32 get_rank_mask(u32 channel, const struct rk3399_sdram_params *params)
629 const u32 rank = params->ch[channel].rank;
631 /* required rank mask is different for LPDDR4 */
632 if (params->dramtype == LPDDR4)
633 return (rank == 1) ? 0x5 : 0xf;
634 else
635 return (rank == 1) ? 0x1 : 0x3;
638 static int data_training_ca(u32 channel, const struct rk3399_sdram_params *params)
640 u32 *denali_pi = rk3399_ddr_pi[channel]->denali_pi;
641 u32 *denali_phy = rk3399_ddr_publ[channel]->denali_phy;
642 u32 obs_0, obs_1, obs_2;
643 const u32 rank_mask = get_rank_mask(channel, params);
644 u32 i, tmp;
646 /* clear interrupt,PI_175 PI_INT_ACK:WR:0:17 */
647 write32(&denali_pi[175], 0x00003f7c);
649 for (i = 0; i < MAX_RANKS_PER_CHANNEL; i++) {
650 if (!(rank_mask & (1 << i)))
651 continue;
653 select_per_cs_training_index(channel, i);
654 /* PI_100 PI_CALVL_EN:RW:8:2 */
655 clrsetbits32(&denali_pi[100], 0x3 << 8, 0x2 << 8);
656 /* PI_92 PI_CALVL_REQ:WR:16:1,PI_CALVL_CS:RW:24:2 */
657 clrsetbits32(&denali_pi[92], (0x1 << 16) | (0x3 << 24),
658 (0x1 << 16) | (i << 24));
660 while (1) {
661 /* PI_174 PI_INT_STATUS:RD:8:18 */
662 tmp = read32(&denali_pi[174]) >> 8;
665 * check status obs
666 * PHY_532/660/789 phy_adr_calvl_obs1_:0:32
668 obs_0 = read32(&denali_phy[532]);
669 obs_1 = read32(&denali_phy[660]);
670 obs_2 = read32(&denali_phy[788]);
671 if (((obs_0 >> 30) & 0x3) ||
672 ((obs_1 >> 30) & 0x3) ||
673 ((obs_2 >> 30) & 0x3))
674 return -1;
675 if ((((tmp >> 11) & 0x1) == 0x1) &&
676 (((tmp >> 13) & 0x1) == 0x1) &&
677 (((tmp >> 5) & 0x1) == 0x0))
678 break;
679 else if (((tmp >> 5) & 0x1) == 0x1)
680 return -1;
682 /* clear interrupt,PI_175 PI_INT_ACK:WR:0:17 */
683 write32(&denali_pi[175], 0x00003f7c);
685 clrbits32(&denali_pi[100], 0x3 << 8);
687 return 0;
690 static int data_training_wl(u32 channel, const struct rk3399_sdram_params *params)
692 u32 *denali_pi = rk3399_ddr_pi[channel]->denali_pi;
693 u32 *denali_phy = rk3399_ddr_publ[channel]->denali_phy;
694 u32 obs_0, obs_1, obs_2, obs_3;
695 u32 rank = params->ch[channel].rank;
696 u32 i, tmp;
698 /* clear interrupt,PI_175 PI_INT_ACK:WR:0:17 */
699 write32(&denali_pi[175], 0x00003f7c);
701 for (i = 0; i < rank; i++) {
702 select_per_cs_training_index(channel, i);
703 /* PI_60 PI_WRLVL_EN:RW:8:2 */
704 clrsetbits32(&denali_pi[60], 0x3 << 8, 0x2 << 8);
705 /* PI_59 PI_WRLVL_REQ:WR:8:1,PI_WRLVL_CS:RW:16:2 */
706 clrsetbits32(&denali_pi[59], (0x1 << 8) | (0x3 << 16), (0x1 << 8) | (i << 16));
708 while (1) {
709 /* PI_174 PI_INT_STATUS:RD:8:18 */
710 tmp = read32(&denali_pi[174]) >> 8;
713 * check status obs, if error maybe can not
714 * get leveling done PHY_40/168/296/424
715 * phy_wrlvl_status_obs_X:0:13
717 obs_0 = read32(&denali_phy[40]);
718 obs_1 = read32(&denali_phy[168]);
719 obs_2 = read32(&denali_phy[296]);
720 obs_3 = read32(&denali_phy[424]);
721 if (((obs_0 >> 12) & 0x1) ||
722 ((obs_1 >> 12) & 0x1) ||
723 ((obs_2 >> 12) & 0x1) ||
724 ((obs_3 >> 12) & 0x1))
725 return -1;
726 if ((((tmp >> 10) & 0x1) == 0x1) &&
727 (((tmp >> 13) & 0x1) == 0x1) &&
728 (((tmp >> 4) & 0x1) == 0x0))
729 break;
730 else if (((tmp >> 4) & 0x1) == 0x1)
731 return -1;
733 /* clear interrupt,PI_175 PI_INT_ACK:WR:0:17 */
734 write32(&denali_pi[175], 0x00003f7c);
737 override_write_leveling_value(channel);
738 clrbits32(&denali_pi[60], 0x3 << 8);
740 return 0;
743 static int data_training_rg(u32 channel, const struct rk3399_sdram_params *params)
745 u32 *denali_pi = rk3399_ddr_pi[channel]->denali_pi;
746 u32 *denali_phy = rk3399_ddr_publ[channel]->denali_phy;
747 u32 rank = params->ch[channel].rank;
748 u32 obs_0, obs_1, obs_2, obs_3;
749 u32 reg_value = 0;
750 u32 i, tmp;
753 * The differential signal of DQS needs to keep low level
754 * before gate training. RPULL will connect 4Kn from PADP
755 * to VSS and a 4Kn from PADN to VDDQ to ensure it.
756 * But if it has PHY side ODT connect at this time,
757 * it will change the DQS signal level. So disable PHY
758 * side ODT before gate training and restore ODT state
759 * after gate training.
761 if (params->dramtype != LPDDR4) {
762 reg_value = (read32(&denali_phy[6]) >> 24) & 0x7;
765 * phy_dqs_tsel_enable_X 3bits
766 * DENALI_PHY_6/134/262/390 offset_24
768 clrbits32(&denali_phy[6], 0x7 << 24);
769 clrbits32(&denali_phy[134], 0x7 << 24);
770 clrbits32(&denali_phy[262], 0x7 << 24);
771 clrbits32(&denali_phy[390], 0x7 << 24);
774 /* clear interrupt,PI_175 PI_INT_ACK:WR:0:17 */
775 write32(&denali_pi[175], 0x00003f7c);
777 for (i = 0; i < rank; i++) {
778 select_per_cs_training_index(channel, i);
779 /* PI_80 PI_RDLVL_GATE_EN:RW:24:2 */
780 clrsetbits32(&denali_pi[80], 0x3 << 24, 0x2 << 24);
782 * PI_74 PI_RDLVL_GATE_REQ:WR:16:1
783 * PI_RDLVL_CS:RW:24:2
785 clrsetbits32(&denali_pi[74], (0x1 << 16) | (0x3 << 24),
786 (0x1 << 16) | (i << 24));
788 while (1) {
789 /* PI_174 PI_INT_STATUS:RD:8:18 */
790 tmp = read32(&denali_pi[174]) >> 8;
793 * check status obs
794 * PHY_43/171/299/427
795 * PHY_GTLVL_STATUS_OBS_x:16:8
797 obs_0 = read32(&denali_phy[43]);
798 obs_1 = read32(&denali_phy[171]);
799 obs_2 = read32(&denali_phy[299]);
800 obs_3 = read32(&denali_phy[427]);
801 if (((obs_0 >> (16 + 6)) & 0x3) ||
802 ((obs_1 >> (16 + 6)) & 0x3) ||
803 ((obs_2 >> (16 + 6)) & 0x3) ||
804 ((obs_3 >> (16 + 6)) & 0x3))
805 return -1;
806 if ((((tmp >> 9) & 0x1) == 0x1) &&
807 (((tmp >> 13) & 0x1) == 0x1) &&
808 (((tmp >> 3) & 0x1) == 0x0))
809 break;
810 else if (((tmp >> 3) & 0x1) == 0x1)
811 return -1;
813 /* clear interrupt,PI_175 PI_INT_ACK:WR:0:17 */
814 write32(&denali_pi[175], 0x00003f7c);
816 clrbits32(&denali_pi[80], 0x3 << 24);
818 if (params->dramtype != LPDDR4) {
820 * phy_dqs_tsel_enable_X 3bits
821 * DENALI_PHY_6/134/262/390 offset_24
823 tmp = reg_value << 24;
824 clrsetbits32(&denali_phy[6], 0x7 << 24, tmp);
825 clrsetbits32(&denali_phy[134], 0x7 << 24, tmp);
826 clrsetbits32(&denali_phy[262], 0x7 << 24, tmp);
827 clrsetbits32(&denali_phy[390], 0x7 << 24, tmp);
829 return 0;
832 static int data_training_rl(u32 channel, const struct rk3399_sdram_params *params)
834 u32 rank = params->ch[channel].rank;
835 u32 i, tmp;
837 u32 *denali_pi = rk3399_ddr_pi[channel]->denali_pi;
839 /* clear interrupt,PI_175 PI_INT_ACK:WR:0:17 */
840 write32(&denali_pi[175], 0x00003f7c);
842 for (i = 0; i < rank; i++) {
843 select_per_cs_training_index(channel, i);
844 /* PI_80 PI_RDLVL_EN:RW:16:2 */
845 clrsetbits32(&denali_pi[80], 0x3 << 16, 0x2 << 16);
846 /* PI_74 PI_RDLVL_REQ:WR:8:1,PI_RDLVL_CS:RW:24:2 */
847 clrsetbits32(&denali_pi[74], (0x1 << 8) | (0x3 << 24), (0x1 << 8) | (i << 24));
849 while (1) {
850 /* PI_174 PI_INT_STATUS:RD:8:18 */
851 tmp = read32(&denali_pi[174]) >> 8;
854 * make sure status obs not report error bit
855 * PHY_46/174/302/430
856 * phy_rdlvl_status_obs_X:16:8
858 if ((((tmp >> 8) & 0x1) == 0x1) && (((tmp >> 13) & 0x1) == 0x1)
859 && (((tmp >> 2) & 0x1) == 0x0))
860 break;
861 else if (((tmp >> 2) & 0x1) == 0x1)
862 return -1;
864 /* clear interrupt,PI_175 PI_INT_ACK:WR:0:17 */
865 write32(&denali_pi[175], 0x00003f7c);
867 clrbits32(&denali_pi[80], 0x3 << 16);
869 return 0;
872 static int data_training_wdql(u32 channel, const struct rk3399_sdram_params *params)
874 u32 *denali_pi = rk3399_ddr_pi[channel]->denali_pi;
875 const u32 rank_mask = get_rank_mask(channel, params);
876 u32 i, tmp;
878 /* clear interrupt,PI_175 PI_INT_ACK:WR:0:17 */
879 write32(&denali_pi[175], 0x00003f7c);
881 for (i = 0; i < MAX_RANKS_PER_CHANNEL; i++) {
882 if (!(rank_mask & (1 << i)))
883 continue;
885 select_per_cs_training_index(channel, i);
887 * disable PI_WDQLVL_VREF_EN before wdq leveling?
888 * PI_181 PI_WDQLVL_VREF_EN:RW:8:1
890 clrbits32(&denali_pi[181], 0x1 << 8);
891 /* PI_124 PI_WDQLVL_EN:RW:16:2 */
892 clrsetbits32(&denali_pi[124], 0x3 << 16, 0x2 << 16);
893 /* PI_121 PI_WDQLVL_REQ:WR:8:1,PI_WDQLVL_CS:RW:16:2 */
894 clrsetbits32(&denali_pi[121], (0x1 << 8) | (0x3 << 16), (0x1 << 8) | (i << 16));
896 while (1) {
897 /* PI_174 PI_INT_STATUS:RD:8:18 */
898 tmp = read32(&denali_pi[174]) >> 8;
899 if ((((tmp >> 12) & 0x1) == 0x1) && (((tmp >> 13) & 0x1) == 0x1)
900 && (((tmp >> 6) & 0x1) == 0x0))
901 break;
902 else if (((tmp >> 6) & 0x1) == 0x1)
903 return -1;
905 /* clear interrupt,PI_175 PI_INT_ACK:WR:0:17 */
906 write32(&denali_pi[175], 0x00003f7c);
908 clrbits32(&denali_pi[124], 0x3 << 16);
910 return 0;
914 static int data_training(u32 channel, const struct rk3399_sdram_params *params,
915 u32 training_flag)
917 u32 *denali_phy = rk3399_ddr_publ[channel]->denali_phy;
918 int ret;
920 /* PHY_927 PHY_PAD_DQS_DRIVE RPULL offset_22 */
921 setbits32(&denali_phy[927], (1 << 22));
923 if (training_flag == PI_FULL_TRAINING) {
924 if (params->dramtype == LPDDR4) {
925 training_flag = PI_CA_TRAINING | PI_WRITE_LEVELING |
926 PI_READ_GATE_TRAINING |
927 PI_READ_LEVELING | PI_WDQ_LEVELING;
928 } else if (params->dramtype == LPDDR3) {
929 training_flag = PI_CA_TRAINING | PI_WRITE_LEVELING |
930 PI_READ_GATE_TRAINING;
931 } else if (params->dramtype == DDR3) {
932 training_flag = PI_WRITE_LEVELING |
933 PI_READ_GATE_TRAINING |
934 PI_READ_LEVELING;
938 /* ca training(LPDDR4,LPDDR3 support) */
939 if ((training_flag & PI_CA_TRAINING) == PI_CA_TRAINING) {
940 ret = data_training_ca(channel, params);
941 if (ret) {
942 printk(BIOS_ERR, "Channel %d CA training failed\n", channel);
943 return ret;
947 /* write leveling(LPDDR4,LPDDR3,DDR3 support) */
948 if ((training_flag & PI_WRITE_LEVELING) == PI_WRITE_LEVELING) {
949 ret = data_training_wl(channel, params);
950 if (ret) {
951 printk(BIOS_ERR, "Channel %d WL training failed\n", channel);
952 return ret;
956 /* read gate training(LPDDR4,LPDDR3,DDR3 support) */
957 if ((training_flag & PI_READ_GATE_TRAINING) == PI_READ_GATE_TRAINING) {
958 ret = data_training_rg(channel, params);
959 if (ret) {
960 printk(BIOS_ERR, "Channel %d RG training failed\n", channel);
961 return ret;
965 /* read leveling(LPDDR4,LPDDR3,DDR3 support) */
966 if ((training_flag & PI_READ_LEVELING) == PI_READ_LEVELING) {
967 ret = data_training_rl(channel, params);
968 if (ret) {
969 printk(BIOS_ERR, "Channel %d RL training failed\n", channel);
970 return ret;
974 /* wdq leveling(LPDDR4 support) */
975 if ((training_flag & PI_WDQ_LEVELING) == PI_WDQ_LEVELING) {
976 ret = data_training_wdql(channel, params);
977 if (ret) {
978 printk(BIOS_ERR, "Channel %d WDQL training failed\n", channel);
979 return ret;
983 /* PHY_927 PHY_PAD_DQS_DRIVE RPULL offset_22 */
984 clrbits32(&denali_phy[927], (1 << 22));
986 return 0;
989 static void set_ddrconfig(const struct rk3399_sdram_params *params,
990 unsigned char channel, u32 ddrconfig)
992 /* only need to set ddrconfig */
993 struct rk3399_msch_regs *ddr_msch_regs = rk3399_msch[channel];
994 unsigned int cs0_cap = 0;
995 unsigned int cs1_cap = 0;
997 cs0_cap = (1 << (params->ch[channel].cs0_row
998 + params->ch[channel].col
999 + params->ch[channel].bk
1000 + params->ch[channel].bw - 20));
1001 if (params->ch[channel].rank > 1)
1002 cs1_cap = cs0_cap >> (params->ch[channel].cs0_row
1003 - params->ch[channel].cs1_row);
1004 if (params->ch[channel].row_3_4) {
1005 cs0_cap = cs0_cap * 3 / 4;
1006 cs1_cap = cs1_cap * 3 / 4;
1009 write32(&ddr_msch_regs->ddrconf, ddrconfig | (ddrconfig << 8));
1010 write32(&ddr_msch_regs->ddrsize, ((cs0_cap / 32) & 0xff) |
1011 (((cs1_cap / 32) & 0xff) << 8));
1014 static void dram_all_config(const struct rk3399_sdram_params *params)
1016 u32 sys_reg = 0;
1017 unsigned int channel;
1018 unsigned int use;
1020 sys_reg |= SYS_REG_ENC_DDRTYPE(params->dramtype);
1021 sys_reg |= SYS_REG_ENC_NUM_CH(params->num_channels);
1022 for (channel = 0, use = 0; (use < params->num_channels) && (channel < 2); channel++) {
1023 const struct rk3399_sdram_channel *info = &params->ch[channel];
1024 struct rk3399_msch_regs *ddr_msch_regs;
1025 const struct rk3399_msch_timings *noc_timing;
1027 if (params->ch[channel].col == 0)
1028 continue;
1029 use++;
1030 sys_reg |= SYS_REG_ENC_ROW_3_4(info->row_3_4, channel);
1031 sys_reg |= SYS_REG_ENC_CHINFO(channel);
1032 sys_reg |= SYS_REG_ENC_RANK(info->rank, channel);
1033 sys_reg |= SYS_REG_ENC_COL(info->col, channel);
1034 sys_reg |= SYS_REG_ENC_BK(info->bk, channel);
1035 sys_reg |= SYS_REG_ENC_CS0_ROW(info->cs0_row, channel);
1036 if (params->ch[channel].rank > 1)
1037 sys_reg |= SYS_REG_ENC_CS1_ROW(info->cs1_row, channel);
1038 sys_reg |= SYS_REG_ENC_BW(info->bw, channel);
1039 sys_reg |= SYS_REG_ENC_DBW(info->dbw, channel);
1041 ddr_msch_regs = rk3399_msch[channel];
1042 noc_timing = &params->ch[channel].noc_timings;
1043 write32(&ddr_msch_regs->ddrtiminga0.d32,
1044 noc_timing->ddrtiminga0.d32);
1045 write32(&ddr_msch_regs->ddrtimingb0.d32,
1046 noc_timing->ddrtimingb0.d32);
1047 write32(&ddr_msch_regs->ddrtimingc0.d32,
1048 noc_timing->ddrtimingc0.d32);
1049 write32(&ddr_msch_regs->devtodev0.d32,
1050 noc_timing->devtodev0.d32);
1051 write32(&ddr_msch_regs->ddrmode.d32,
1052 noc_timing->ddrmode.d32);
1054 /* rank 1 memory clock disable (dfi_dram_clk_disable = 1) */
1055 if (params->ch[channel].rank == 1)
1056 setbits32(&rk3399_ddr_pctl[channel]->denali_ctl[276],
1057 1 << 17);
1060 write32(&rk3399_pmugrf->os_reg2, sys_reg);
1061 DDR_STRIDE(params->stride);
1063 /* reboot hold register set */
1064 write32(&pmucru_ptr->pmucru_rstnhold_con[1],
1065 PRESET_SGRF_HOLD(0) | PRESET_GPIO0_HOLD(1) |
1066 PRESET_GPIO1_HOLD(1));
1067 clrsetbits32(&cru_ptr->glb_rst_con, 0x3, 0x3);
1070 static void switch_to_phy_index1(const struct rk3399_sdram_params *params)
1072 u32 channel;
1073 u32 *denali_phy;
1074 struct stopwatch sw;
1075 u32 ch_count = params->num_channels;
1077 stopwatch_init_msecs_expire(&sw, 100);
1078 write32(&rk3399_ddr_cic->cic_ctrl0,
1079 RK_CLRSETBITS(0x03 << 4 | 1 << 2 | 1,
1080 1 << 4 | 1 << 2 | 1));
1081 while (!(read32(&rk3399_ddr_cic->cic_status0) & (1 << 2))) {
1082 if (stopwatch_expired(&sw)) {
1083 printk(BIOS_ERR,
1084 "index1 frequency change overtime, reset\n");
1085 board_reset();
1089 stopwatch_init_msecs_expire(&sw, 100);
1090 write32(&rk3399_ddr_cic->cic_ctrl0, RK_CLRSETBITS(1 << 1, 1 << 1));
1091 while (!(read32(&rk3399_ddr_cic->cic_status0) & (1 << 0))) {
1092 if (stopwatch_expired(&sw)) {
1093 printk(BIOS_ERR,
1094 "index1 frequency done overtime, reset\n");
1095 board_reset();
1099 for (channel = 0; channel < ch_count; channel++) {
1100 denali_phy = rk3399_ddr_publ[channel]->denali_phy;
1101 clrsetbits32(&denali_phy[896], (0x3 << 8) | 1, 1 << 8);
1102 if (data_training(channel, params, PI_FULL_TRAINING)) {
1103 printk(BIOS_ERR, "index1 training failed, reset\n");
1104 board_reset();
1109 void sdram_init(const struct rk3399_sdram_params *params)
1111 unsigned char dramtype = params->dramtype;
1112 unsigned int ddr_freq = params->ddr_freq;
1113 int channel;
1115 printk(BIOS_INFO, "Starting SDRAM initialization...\n");
1117 if ((dramtype == DDR3 && ddr_freq > 800*MHz) ||
1118 (dramtype == LPDDR3 && ddr_freq > 933*MHz) ||
1119 (dramtype == LPDDR4 && ddr_freq > 800*MHz))
1120 die("SDRAM frequency is to high!");
1122 rkclk_configure_ddr(ddr_freq);
1124 for (channel = 0; channel < 2; channel++) {
1125 phy_pctrl_reset(channel);
1126 phy_dll_bypass_set(rk3399_ddr_publ[channel], ddr_freq);
1128 if (channel >= params->num_channels)
1129 continue;
1132 * TODO: we need to find the root cause why this
1133 * step may fail, before that, we just reset the
1134 * system, and start again.
1136 if (pctl_cfg(channel, params)) {
1137 printk(BIOS_ERR, "pctl_cfg fail, reset\n");
1138 board_reset();
1141 /* LPDDR2/LPDDR3 need to wait DAI complete, max 10us */
1142 if (dramtype == LPDDR3)
1143 udelay(10);
1145 if (data_training(channel, params, PI_FULL_TRAINING)) {
1146 printk(BIOS_ERR, "SDRAM initialization failed, reset\n");
1147 board_reset();
1150 set_ddrconfig(params, channel, params->ch[channel].ddrconfig);
1152 dram_all_config(params);
1153 switch_to_phy_index1(params);
1155 printk(BIOS_INFO, "Finish SDRAM initialization...\n");
1158 size_t sdram_size_mb(void)
1160 u32 rank, col, bk, cs0_row, cs1_row, bw, row_3_4;
1161 size_t chipsize_mb = 0;
1162 static size_t size_mb = 0;
1163 u32 ch;
1165 if (!size_mb) {
1166 u32 sys_reg = read32(&rk3399_pmugrf->os_reg2);
1167 u32 ch_num = SYS_REG_DEC_NUM_CH(sys_reg);
1169 for (ch = 0; ch < ch_num; ch++) {
1170 rank = SYS_REG_DEC_RANK(sys_reg, ch);
1171 col = SYS_REG_DEC_COL(sys_reg, ch);
1172 bk = SYS_REG_DEC_BK(sys_reg, ch);
1173 cs0_row = SYS_REG_DEC_CS0_ROW(sys_reg, ch);
1174 cs1_row = SYS_REG_DEC_CS1_ROW(sys_reg, ch);
1175 bw = SYS_REG_DEC_BW(sys_reg, ch);
1176 row_3_4 = SYS_REG_DEC_ROW_3_4(sys_reg, ch);
1178 chipsize_mb = (1 << (cs0_row + col + bk + bw - 20));
1180 if (rank > 1)
1181 chipsize_mb += chipsize_mb >>
1182 (cs0_row - cs1_row);
1183 if (row_3_4)
1184 chipsize_mb = chipsize_mb * 3 / 4;
1185 size_mb += chipsize_mb;
1189 * we use the 0x00000000~0xf7ffffff space
1190 * since 0xf8000000~0xffffffff is soc register space
1191 * so we reserve it
1193 size_mb = MIN(size_mb, 0xf8000000/MiB);
1196 return size_mb;