1 /* SPDX-License-Identifier: GPL-2.0-only */
3 #include <device/mmio.h>
4 #include <console/console.h>
7 #include <soc/addressmap.h>
15 #define DDR_PI_OFFSET 0x800
16 #define DDR_PHY_OFFSET 0x2000
17 #define DDRC0_PI_BASE_ADDR (DDRC0_BASE_ADDR + DDR_PI_OFFSET)
18 #define DDRC0_PHY_BASE_ADDR (DDRC0_BASE_ADDR + DDR_PHY_OFFSET)
19 #define DDRC1_PI_BASE_ADDR (DDRC1_BASE_ADDR + DDR_PI_OFFSET)
20 #define DDRC1_PHY_BASE_ADDR (DDRC1_BASE_ADDR + DDR_PHY_OFFSET)
22 static struct rk3399_ddr_pctl_regs
* const rk3399_ddr_pctl
[2] = {
23 (void *)DDRC0_BASE_ADDR
, (void *)DDRC1_BASE_ADDR
};
24 static struct rk3399_ddr_pi_regs
* const rk3399_ddr_pi
[2] = {
25 (void *)DDRC0_PI_BASE_ADDR
, (void *)DDRC1_PI_BASE_ADDR
};
26 static struct rk3399_ddr_publ_regs
* const rk3399_ddr_publ
[2] = {
27 (void *)DDRC0_PHY_BASE_ADDR
, (void *)DDRC1_PHY_BASE_ADDR
};
28 static struct rk3399_msch_regs
* const rk3399_msch
[2] = {
29 (void *)SERVER_MSCH0_BASE_ADDR
, (void *)SERVER_MSCH1_BASE_ADDR
};
30 static struct rk3399_ddr_cic_regs
*const rk3399_ddr_cic
= (void *)CIC_BASE_ADDR
;
33 * sys_reg bitfield struct
54 #define SYS_REG_ENC_ROW_3_4(n, ch) ((n) << (30 + (ch)))
55 #define SYS_REG_DEC_ROW_3_4(n, ch) ((n >> (30 + ch)) & 0x1)
56 #define SYS_REG_ENC_CHINFO(ch) (1 << (28 + (ch)))
57 #define SYS_REG_ENC_DDRTYPE(n) ((n) << 13)
58 #define SYS_REG_ENC_NUM_CH(n) (((n) - 1) << 12)
59 #define SYS_REG_DEC_NUM_CH(n) (1 + ((n >> 12) & 0x1))
60 #define SYS_REG_ENC_RANK(n, ch) (((n) - 1) << (11 + ((ch) * 16)))
61 #define SYS_REG_DEC_RANK(n, ch) (1 + ((n >> (11 + 16 * ch)) & 0x1))
62 #define SYS_REG_ENC_COL(n, ch) (((n) - 9) << (9 + ((ch) * 16)))
63 #define SYS_REG_DEC_COL(n, ch) (9 + ((n >> (9 + 16 * ch)) & 0x3))
64 #define SYS_REG_ENC_BK(n, ch) (((n) == 3 ? 0 : 1) \
66 #define SYS_REG_DEC_BK(n, ch) (3 - ((n >> (8 + 16 * ch)) & 0x1))
67 #define SYS_REG_ENC_CS0_ROW(n, ch) (((n) - 13) << (6 + ((ch) * 16)))
68 #define SYS_REG_DEC_CS0_ROW(n, ch) (13 + ((n >> (6 + 16 * ch)) & 0x3))
69 #define SYS_REG_ENC_CS1_ROW(n, ch) (((n) - 13) << (4 + ((ch) * 16)))
70 #define SYS_REG_DEC_CS1_ROW(n, ch) (13 + ((n >> (4 + 16 * ch)) & 0x3))
71 #define SYS_REG_ENC_BW(n, ch) ((2 >> (n)) << (2 + ((ch) * 16)))
72 #define SYS_REG_DEC_BW(n, ch) (2 >> ((n >> (2 + 16 * ch)) & 0x3))
73 #define SYS_REG_ENC_DBW(n, ch) ((2 >> (n)) << (0 + ((ch) * 16)))
74 #define SYS_REG_DEC_DBW(n, ch) (2 >> ((n >> (0 + 16 * ch)) & 0x3))
76 #define DDR_STRIDE(n) write32(&rk3399_pmusgrf->soc_con4,\
77 (0x1F << (10 + 16)) | (n << 10))
79 #define PRESET_SGRF_HOLD(n) ((0x1 << (6+16)) | ((n) << 6))
80 #define PRESET_GPIO0_HOLD(n) ((0x1 << (7+16)) | ((n) << 7))
81 #define PRESET_GPIO1_HOLD(n) ((0x1 << (8+16)) | ((n) << 8))
83 #define PHY_DRV_ODT_Hi_Z (0x0)
84 #define PHY_DRV_ODT_240 (0x1)
85 #define PHY_DRV_ODT_120 (0x8)
86 #define PHY_DRV_ODT_80 (0x9)
87 #define PHY_DRV_ODT_60 (0xc)
88 #define PHY_DRV_ODT_48 (0xd)
89 #define PHY_DRV_ODT_40 (0xe)
90 #define PHY_DRV_ODT_34_3 (0xf)
92 #define MAX_RANKS_PER_CHANNEL 4
94 static void copy_to_reg(u32
*dest
, const u32
*src
, u32 n
)
98 for (i
= 0; i
< n
/ sizeof(u32
); i
++) {
105 static void phy_pctrl_reset(u32 channel
)
107 rkclk_ddr_reset(channel
, 1, 1);
110 rkclk_ddr_reset(channel
, 1, 0);
113 rkclk_ddr_reset(channel
, 0, 0);
117 static void phy_dll_bypass_set(struct rk3399_ddr_publ_regs
*ddr_publ_regs
,
120 u32
*denali_phy
= ddr_publ_regs
->denali_phy
;
122 if (freq
<= 125*MHz
) {
123 /* phy_sw_master_mode_X PHY_86/214/342/470 4bits offset_8 */
124 setbits32(&denali_phy
[86], (0x3 << 2) << 8);
125 setbits32(&denali_phy
[214], (0x3 << 2) << 8);
126 setbits32(&denali_phy
[342], (0x3 << 2) << 8);
127 setbits32(&denali_phy
[470], (0x3 << 2) << 8);
129 /* phy_adrctl_sw_master_mode PHY_547/675/803 4bits offset_16 */
130 setbits32(&denali_phy
[547], (0x3 << 2) << 16);
131 setbits32(&denali_phy
[675], (0x3 << 2) << 16);
132 setbits32(&denali_phy
[803], (0x3 << 2) << 16);
134 /* phy_sw_master_mode_X PHY_86/214/342/470 4bits offset_8 */
135 clrbits32(&denali_phy
[86], (0x3 << 2) << 8);
136 clrbits32(&denali_phy
[214], (0x3 << 2) << 8);
137 clrbits32(&denali_phy
[342], (0x3 << 2) << 8);
138 clrbits32(&denali_phy
[470], (0x3 << 2) << 8);
140 /* phy_adrctl_sw_master_mode PHY_547/675/803 4bits offset_16 */
141 clrbits32(&denali_phy
[547], (0x3 << 2) << 16);
142 clrbits32(&denali_phy
[675], (0x3 << 2) << 16);
143 clrbits32(&denali_phy
[803], (0x3 << 2) << 16);
147 static void set_memory_map(u32 channel
, const struct rk3399_sdram_params
*params
)
149 const struct rk3399_sdram_channel
*sdram_ch
= ¶ms
->ch
[channel
];
150 u32
*denali_ctl
= rk3399_ddr_pctl
[channel
]->denali_ctl
;
151 u32
*denali_pi
= rk3399_ddr_pi
[channel
]->denali_pi
;
156 if ((sdram_ch
->ddrconfig
< 2) || (sdram_ch
->ddrconfig
== 4))
158 else if (sdram_ch
->ddrconfig
== 3)
163 cs_map
= (sdram_ch
->rank
> 1) ? 3 : 1;
164 reduc
= (sdram_ch
->bw
== 2) ? 0 : 1;
166 clrsetbits32(&denali_ctl
[191], 0xF, (12 - sdram_ch
->col
));
167 clrsetbits32(&denali_ctl
[190], (0x3 << 16) | (0x7 << 24),
168 ((3 - sdram_ch
->bk
) << 16) |
171 clrsetbits32(&denali_ctl
[196], 0x3 | (1 << 16),
172 cs_map
| (reduc
<< 16));
174 /* PI_199 PI_COL_DIFF:RW:0:4 */
175 clrsetbits32(&denali_pi
[199], 0xF, (12 - sdram_ch
->col
));
177 /* PI_155 PI_ROW_DIFF:RW:24:3 PI_BANK_DIFF:RW:16:2 */
178 clrsetbits32(&denali_pi
[155], (0x3 << 16) | (0x7 << 24),
179 ((3 - sdram_ch
->bk
) << 16) |
181 /* PI_41 PI_CS_MAP:RW:24:4 */
182 clrsetbits32(&denali_pi
[41], 0xf << 24, cs_map
<< 24);
183 if ((sdram_ch
->rank
== 1) && (params
->dramtype
== DDR3
))
184 write32(&denali_pi
[34], 0x2EC7FFFF);
187 static void set_ds_odt(u32 channel
, const struct rk3399_sdram_params
*params
)
189 u32
*denali_phy
= rk3399_ddr_publ
[channel
]->denali_phy
;
191 u32 tsel_idle_en
, tsel_wr_en
, tsel_rd_en
;
192 u32 tsel_idle_select_p
, tsel_wr_select_p
, tsel_rd_select_p
;
193 u32 ca_tsel_wr_select_p
, ca_tsel_wr_select_n
;
194 u32 tsel_idle_select_n
, tsel_wr_select_n
, tsel_rd_select_n
;
197 if (params
->dramtype
== LPDDR4
) {
198 tsel_rd_select_p
= PHY_DRV_ODT_Hi_Z
;
199 tsel_rd_select_n
= PHY_DRV_ODT_240
;
201 tsel_wr_select_p
= PHY_DRV_ODT_40
;
202 tsel_wr_select_n
= PHY_DRV_ODT_40
;
204 tsel_idle_select_p
= PHY_DRV_ODT_Hi_Z
;
205 tsel_idle_select_n
= PHY_DRV_ODT_240
;
207 ca_tsel_wr_select_p
= PHY_DRV_ODT_40
;
208 ca_tsel_wr_select_n
= PHY_DRV_ODT_40
;
209 } else if (params
->dramtype
== LPDDR3
) {
210 tsel_rd_select_p
= PHY_DRV_ODT_240
;
211 tsel_rd_select_n
= PHY_DRV_ODT_Hi_Z
;
213 tsel_wr_select_p
= PHY_DRV_ODT_34_3
;
214 tsel_wr_select_n
= PHY_DRV_ODT_34_3
;
216 tsel_idle_select_p
= PHY_DRV_ODT_240
;
217 tsel_idle_select_n
= PHY_DRV_ODT_Hi_Z
;
219 ca_tsel_wr_select_p
= PHY_DRV_ODT_48
;
220 ca_tsel_wr_select_n
= PHY_DRV_ODT_48
;
222 tsel_rd_select_p
= PHY_DRV_ODT_240
;
223 tsel_rd_select_n
= PHY_DRV_ODT_240
;
225 tsel_wr_select_p
= PHY_DRV_ODT_34_3
;
226 tsel_wr_select_n
= PHY_DRV_ODT_34_3
;
228 tsel_idle_select_p
= PHY_DRV_ODT_240
;
229 tsel_idle_select_n
= PHY_DRV_ODT_240
;
231 ca_tsel_wr_select_p
= PHY_DRV_ODT_34_3
;
232 ca_tsel_wr_select_n
= PHY_DRV_ODT_34_3
;
235 if (params
->odt
== 1)
244 * phy_dq_tsel_select_X 24bits DENALI_PHY_6/134/262/390 offset_0
245 * sets termination values for read/idle cycles and drive strength
246 * for write cycles for DQ/DM
248 reg_value
= tsel_rd_select_n
| (tsel_rd_select_p
<< 0x4) |
249 (tsel_wr_select_n
<< 8) | (tsel_wr_select_p
<< 12) |
250 (tsel_idle_select_n
<< 16) | (tsel_idle_select_p
<< 20);
251 clrsetbits32(&denali_phy
[6], 0xffffff, reg_value
);
252 clrsetbits32(&denali_phy
[134], 0xffffff, reg_value
);
253 clrsetbits32(&denali_phy
[262], 0xffffff, reg_value
);
254 clrsetbits32(&denali_phy
[390], 0xffffff, reg_value
);
257 * phy_dqs_tsel_select_X 24bits DENALI_PHY_7/135/263/391 offset_0
258 * sets termination values for read/idle cycles and drive strength
259 * for write cycles for DQS
261 clrsetbits32(&denali_phy
[7], 0xffffff, reg_value
);
262 clrsetbits32(&denali_phy
[135], 0xffffff, reg_value
);
263 clrsetbits32(&denali_phy
[263], 0xffffff, reg_value
);
264 clrsetbits32(&denali_phy
[391], 0xffffff, reg_value
);
266 /* phy_adr_tsel_select_ 8bits DENALI_PHY_544/672/800 offset_0 */
267 reg_value
= ca_tsel_wr_select_n
| (ca_tsel_wr_select_p
<< 0x4);
268 clrsetbits32(&denali_phy
[544], 0xff, reg_value
);
269 clrsetbits32(&denali_phy
[672], 0xff, reg_value
);
270 clrsetbits32(&denali_phy
[800], 0xff, reg_value
);
272 /* phy_pad_addr_drive 8bits DENALI_PHY_928 offset_0 */
273 clrsetbits32(&denali_phy
[928], 0xff, reg_value
);
275 /* phy_pad_rst_drive 8bits DENALI_PHY_937 offset_0 */
276 clrsetbits32(&denali_phy
[937], 0xff, reg_value
);
278 /* phy_pad_cke_drive 8bits DENALI_PHY_935 offset_0 */
279 clrsetbits32(&denali_phy
[935], 0xff, reg_value
);
281 /* phy_pad_cs_drive 8bits DENALI_PHY_939 offset_0 */
282 clrsetbits32(&denali_phy
[939], 0xff, reg_value
);
284 /* phy_pad_clk_drive 8bits DENALI_PHY_929 offset_0 */
285 clrsetbits32(&denali_phy
[929], 0xff, reg_value
);
287 /* phy_pad_fdbk_drive 23bit DENALI_PHY_924/925 */
288 clrsetbits32(&denali_phy
[924], 0xff,
289 tsel_wr_select_n
| (tsel_wr_select_p
<< 4));
290 clrsetbits32(&denali_phy
[925], 0xff,
291 tsel_rd_select_n
| (tsel_rd_select_p
<< 4));
293 /* phy_dq_tsel_enable_X 3bits DENALI_PHY_5/133/261/389 offset_16 */
294 reg_value
= (tsel_rd_en
| (tsel_wr_en
<< 1) | (tsel_idle_en
<< 2))
296 clrsetbits32(&denali_phy
[5], 0x7 << 16, reg_value
);
297 clrsetbits32(&denali_phy
[133], 0x7 << 16, reg_value
);
298 clrsetbits32(&denali_phy
[261], 0x7 << 16, reg_value
);
299 clrsetbits32(&denali_phy
[389], 0x7 << 16, reg_value
);
301 /* phy_dqs_tsel_enable_X 3bits DENALI_PHY_6/134/262/390 offset_24 */
302 reg_value
= (tsel_rd_en
| (tsel_wr_en
<< 1) | (tsel_idle_en
<< 2))
304 clrsetbits32(&denali_phy
[6], 0x7 << 24, reg_value
);
305 clrsetbits32(&denali_phy
[134], 0x7 << 24, reg_value
);
306 clrsetbits32(&denali_phy
[262], 0x7 << 24, reg_value
);
307 clrsetbits32(&denali_phy
[390], 0x7 << 24, reg_value
);
309 /* phy_adr_tsel_enable_ 1bit DENALI_PHY_518/646/774 offset_8 */
310 reg_value
= tsel_wr_en
<< 8;
311 clrsetbits32(&denali_phy
[518], 0x1 << 8, reg_value
);
312 clrsetbits32(&denali_phy
[646], 0x1 << 8, reg_value
);
313 clrsetbits32(&denali_phy
[774], 0x1 << 8, reg_value
);
315 /* phy_pad_addr_term tsel 1bit DENALI_PHY_933 offset_17 */
316 reg_value
= tsel_wr_en
<< 17;
317 clrsetbits32(&denali_phy
[933], 0x1 << 17, reg_value
);
319 * pad_rst/cke/cs/clk_term tsel 1bits
320 * DENALI_PHY_938/936/940/934 offset_17
322 clrsetbits32(&denali_phy
[938], 0x1 << 17, reg_value
);
323 clrsetbits32(&denali_phy
[936], 0x1 << 17, reg_value
);
324 clrsetbits32(&denali_phy
[940], 0x1 << 17, reg_value
);
325 clrsetbits32(&denali_phy
[934], 0x1 << 17, reg_value
);
327 /* phy_pad_fdbk_term 1bit DENALI_PHY_930 offset_17 */
328 clrsetbits32(&denali_phy
[930], 0x1 << 17, reg_value
);
331 static void phy_io_config(u32 channel
, const struct rk3399_sdram_params
*params
)
333 u32
*denali_phy
= rk3399_ddr_publ
[channel
]->denali_phy
;
334 u32 vref_mode_dq
, vref_value_dq
, vref_mode_ac
, vref_value_ac
;
337 u32 drv_value
, odt_value
;
341 if (params
->dramtype
== LPDDR4
) {
344 vref_value_dq
= 0x1f;
346 vref_value_ac
= 0x1f;
347 } else if (params
->dramtype
== LPDDR3
) {
348 if (params
->odt
== 1) {
349 vref_mode_dq
= 0x5; /* LPDDR3 ODT */
350 drv_value
= (read32(&denali_phy
[6]) >> 12) & 0xf;
351 odt_value
= (read32(&denali_phy
[6]) >> 4) & 0xf;
352 if (drv_value
== PHY_DRV_ODT_48
) {
354 case PHY_DRV_ODT_240
:
355 vref_value_dq
= 0x16;
357 case PHY_DRV_ODT_120
:
358 vref_value_dq
= 0x26;
361 vref_value_dq
= 0x36;
364 die("Halting: Invalid ODT value.\n");
366 } else if (drv_value
== PHY_DRV_ODT_40
) {
368 case PHY_DRV_ODT_240
:
369 vref_value_dq
= 0x19;
371 case PHY_DRV_ODT_120
:
372 vref_value_dq
= 0x23;
375 vref_value_dq
= 0x31;
378 die("Halting: Invalid ODT value.\n");
380 } else if (drv_value
== PHY_DRV_ODT_34_3
) {
382 case PHY_DRV_ODT_240
:
383 vref_value_dq
= 0x17;
385 case PHY_DRV_ODT_120
:
386 vref_value_dq
= 0x20;
389 vref_value_dq
= 0x2e;
392 die("Halting: Invalid ODT value.\n");
395 die("Halting: Invalid DRV value.\n");
398 vref_mode_dq
= 0x2; /* LPDDR3 */
399 vref_value_dq
= 0x1f;
402 vref_value_ac
= 0x1f;
403 } else if (params
->dramtype
== DDR3
) {
406 vref_value_dq
= 0x1f;
408 vref_value_ac
= 0x1f;
410 die("Halting: Unknown DRAM type.\n");
413 reg_value
= (vref_mode_dq
<< 9) | (0x1 << 8) | vref_value_dq
;
415 /* PHY_913 PHY_PAD_VREF_CTRL_DQ_0 12bits offset_8 */
416 clrsetbits32(&denali_phy
[913], 0xfff << 8, reg_value
<< 8);
417 /* PHY_914 PHY_PAD_VREF_CTRL_DQ_1 12bits offset_0 */
418 clrsetbits32(&denali_phy
[914], 0xfff, reg_value
);
419 /* PHY_914 PHY_PAD_VREF_CTRL_DQ_2 12bits offset_16 */
420 clrsetbits32(&denali_phy
[914], 0xfff << 16, reg_value
<< 16);
421 /* PHY_915 PHY_PAD_VREF_CTRL_DQ_3 12bits offset_0 */
422 clrsetbits32(&denali_phy
[915], 0xfff, reg_value
);
424 reg_value
= (vref_mode_ac
<< 9) | (0x1 << 8) | vref_value_ac
;
426 /* PHY_915 PHY_PAD_VREF_CTRL_AC 12bits offset_16 */
427 clrsetbits32(&denali_phy
[915], 0xfff << 16, reg_value
<< 16);
429 if (params
->dramtype
== LPDDR4
)
431 else if (params
->dramtype
== LPDDR3
)
433 else if (params
->dramtype
== DDR3
)
436 /* PHY_924 PHY_PAD_FDBK_DRIVE */
437 clrsetbits32(&denali_phy
[924], 0x7 << 15, mode_sel
<< 15);
438 /* PHY_926 PHY_PAD_DATA_DRIVE */
439 clrsetbits32(&denali_phy
[926], 0x7 << 6, mode_sel
<< 6);
440 /* PHY_927 PHY_PAD_DQS_DRIVE */
441 clrsetbits32(&denali_phy
[927], 0x7 << 6, mode_sel
<< 6);
442 /* PHY_928 PHY_PAD_ADDR_DRIVE */
443 clrsetbits32(&denali_phy
[928], 0x7 << 14, mode_sel
<< 14);
444 /* PHY_929 PHY_PAD_CLK_DRIVE */
445 clrsetbits32(&denali_phy
[929], 0x7 << 14, mode_sel
<< 14);
446 /* PHY_935 PHY_PAD_CKE_DRIVE */
447 clrsetbits32(&denali_phy
[935], 0x7 << 14, mode_sel
<< 14);
448 /* PHY_937 PHY_PAD_RST_DRIVE */
449 clrsetbits32(&denali_phy
[937], 0x7 << 14, mode_sel
<< 14);
450 /* PHY_939 PHY_PAD_CS_DRIVE */
451 clrsetbits32(&denali_phy
[939], 0x7 << 14, mode_sel
<< 14);
454 if (params
->ddr_freq
< 400 * MHz
)
456 else if (params
->ddr_freq
< 800 * MHz
)
458 else if (params
->ddr_freq
< 1200 * MHz
)
463 /* PHY_924 PHY_PAD_FDBK_DRIVE */
464 clrsetbits32(&denali_phy
[924], 0x3 << 21, speed
<< 21);
465 /* PHY_926 PHY_PAD_DATA_DRIVE */
466 clrsetbits32(&denali_phy
[926], 0x3 << 9, speed
<< 9);
467 /* PHY_927 PHY_PAD_DQS_DRIVE */
468 clrsetbits32(&denali_phy
[927], 0x3 << 9, speed
<< 9);
469 /* PHY_928 PHY_PAD_ADDR_DRIVE */
470 clrsetbits32(&denali_phy
[928], 0x3 << 17, speed
<< 17);
471 /* PHY_929 PHY_PAD_CLK_DRIVE */
472 clrsetbits32(&denali_phy
[929], 0x3 << 17, speed
<< 17);
473 /* PHY_935 PHY_PAD_CKE_DRIVE */
474 clrsetbits32(&denali_phy
[935], 0x3 << 17, speed
<< 17);
475 /* PHY_937 PHY_PAD_RST_DRIVE */
476 clrsetbits32(&denali_phy
[937], 0x3 << 17, speed
<< 17);
477 /* PHY_939 PHY_PAD_CS_DRIVE */
478 clrsetbits32(&denali_phy
[939], 0x3 << 17, speed
<< 17);
481 static int pctl_cfg(u32 channel
, const struct rk3399_sdram_params
*params
)
483 u32
*denali_ctl
= rk3399_ddr_pctl
[channel
]->denali_ctl
;
484 u32
*denali_pi
= rk3399_ddr_pi
[channel
]->denali_pi
;
485 u32
*denali_phy
= rk3399_ddr_publ
[channel
]->denali_phy
;
486 const u32
*params_ctl
= params
->pctl_regs
.denali_ctl
;
487 const u32
*params_phy
= params
->phy_regs
.denali_phy
;
489 u32 pwrup_srefresh_exit
;
493 * work around controller bug:
494 * Do not program DRAM_CLASS until NO_PHY_IND_TRAIN_INT is programmed
496 copy_to_reg(&denali_ctl
[1], ¶ms_ctl
[1],
497 sizeof(struct rk3399_ddr_pctl_regs
) - 4);
498 write32(&denali_ctl
[0], params_ctl
[0]);
499 copy_to_reg(denali_pi
, ¶ms
->pi_regs
.denali_pi
[0],
500 sizeof(struct rk3399_ddr_pi_regs
));
501 /* rank count need to set for init */
502 set_memory_map(channel
, params
);
504 write32(&denali_phy
[910], params
->phy_regs
.denali_phy
[910]);
505 write32(&denali_phy
[911], params
->phy_regs
.denali_phy
[911]);
506 write32(&denali_phy
[912], params
->phy_regs
.denali_phy
[912]);
508 pwrup_srefresh_exit
= read32(&denali_ctl
[68]) & PWRUP_SREFRESH_EXIT
;
509 clrbits32(&denali_ctl
[68], PWRUP_SREFRESH_EXIT
);
512 clrsetbits32(&denali_phy
[957], 0x3 << 24, 1 << 24);
514 setbits32(&denali_pi
[0], START
);
515 setbits32(&denali_ctl
[0], START
);
518 tmp
= read32(&denali_phy
[920]);
519 tmp1
= read32(&denali_phy
[921]);
520 tmp2
= read32(&denali_phy
[922]);
521 if ((((tmp
>> 16) & 0x1) == 0x1) &&
522 (((tmp1
>> 16) & 0x1) == 0x1) &&
523 (((tmp1
>> 0) & 0x1) == 0x1) &&
524 (((tmp2
>> 0) & 0x1) == 0x1))
528 copy_to_reg(&denali_phy
[896], ¶ms_phy
[896], (958 - 895) * 4);
529 copy_to_reg(&denali_phy
[0], ¶ms_phy
[0], (90 - 0 + 1) * 4);
530 copy_to_reg(&denali_phy
[128], ¶ms_phy
[128], (218 - 128 + 1) * 4);
531 copy_to_reg(&denali_phy
[256], ¶ms_phy
[256], (346 - 256 + 1) * 4);
532 copy_to_reg(&denali_phy
[384], ¶ms_phy
[384], (474 - 384 + 1) * 4);
533 copy_to_reg(&denali_phy
[512], ¶ms_phy
[512], (549 - 512 + 1) * 4);
534 copy_to_reg(&denali_phy
[640], ¶ms_phy
[640], (677 - 640 + 1) * 4);
535 copy_to_reg(&denali_phy
[768], ¶ms_phy
[768], (805 - 768 + 1) * 4);
536 set_ds_odt(channel
, params
);
539 * phy_dqs_tsel_wr_timing_X 8bits DENALI_PHY_84/212/340/468 offset_8
540 * dqs_tsel_wr_end[7:4] add Half cycle
542 tmp
= (read32(&denali_phy
[84]) >> 8) & 0xff;
543 clrsetbits32(&denali_phy
[84], 0xff << 8, (tmp
+ 0x10) << 8);
544 tmp
= (read32(&denali_phy
[212]) >> 8) & 0xff;
545 clrsetbits32(&denali_phy
[212], 0xff << 8, (tmp
+ 0x10) << 8);
546 tmp
= (read32(&denali_phy
[340]) >> 8) & 0xff;
547 clrsetbits32(&denali_phy
[340], 0xff << 8, (tmp
+ 0x10) << 8);
548 tmp
= (read32(&denali_phy
[468]) >> 8) & 0xff;
549 clrsetbits32(&denali_phy
[468], 0xff << 8, (tmp
+ 0x10) << 8);
552 * phy_dqs_tsel_wr_timing_X 8bits DENALI_PHY_83/211/339/467 offset_8
553 * dq_tsel_wr_end[7:4] add Half cycle
555 tmp
= (read32(&denali_phy
[83]) >> 16) & 0xff;
556 clrsetbits32(&denali_phy
[83], 0xff << 16, (tmp
+ 0x10) << 16);
557 tmp
= (read32(&denali_phy
[211]) >> 16) & 0xff;
558 clrsetbits32(&denali_phy
[211], 0xff << 16, (tmp
+ 0x10) << 16);
559 tmp
= (read32(&denali_phy
[339]) >> 16) & 0xff;
560 clrsetbits32(&denali_phy
[339], 0xff << 16, (tmp
+ 0x10) << 16);
561 tmp
= (read32(&denali_phy
[467]) >> 16) & 0xff;
562 clrsetbits32(&denali_phy
[467], 0xff << 16, (tmp
+ 0x10) << 16);
564 phy_io_config(channel
, params
);
567 clrsetbits32(&denali_phy
[957], 0x3 << 24, 0x2 << 24);
569 /* FIXME: need to care ERROR bit */
570 stopwatch_init_msecs_expire(&sw
, 100);
571 while (!(read32(&denali_ctl
[203]) & (1 << 3))) {
572 if (stopwatch_expired(&sw
))
576 clrsetbits32(&denali_ctl
[68], PWRUP_SREFRESH_EXIT
,
577 pwrup_srefresh_exit
);
581 static void select_per_cs_training_index(u32 channel
, u32 rank
)
583 u32
*denali_phy
= rk3399_ddr_publ
[channel
]->denali_phy
;
585 /* PHY_84 PHY_PER_CS_TRAINING_EN_0 1bit offset_16 */
586 if ((read32(&denali_phy
[84])>>16) & 1) {
589 * phy_per_cs_training_index_X 1bit offset_24
591 clrsetbits32(&denali_phy
[8], 0x1 << 24, rank
<< 24);
592 clrsetbits32(&denali_phy
[136], 0x1 << 24, rank
<< 24);
593 clrsetbits32(&denali_phy
[264], 0x1 << 24, rank
<< 24);
594 clrsetbits32(&denali_phy
[392], 0x1 << 24, rank
<< 24);
598 static void override_write_leveling_value(u32 channel
)
600 u32
*denali_ctl
= rk3399_ddr_pctl
[channel
]->denali_ctl
;
601 u32
*denali_phy
= rk3399_ddr_publ
[channel
]->denali_phy
;
604 /* PHY_896 PHY_FREQ_SEL_MULTICAST_EN 1bit offset_0 */
605 setbits32(&denali_phy
[896], 1);
609 * phy_per_cs_training_multicast_en_X 1bit offset_16
611 clrsetbits32(&denali_phy
[8], 0x1 << 16, 1 << 16);
612 clrsetbits32(&denali_phy
[136], 0x1 << 16, 1 << 16);
613 clrsetbits32(&denali_phy
[264], 0x1 << 16, 1 << 16);
614 clrsetbits32(&denali_phy
[392], 0x1 << 16, 1 << 16);
616 for (byte
= 0; byte
< 4; byte
++)
617 clrsetbits32(&denali_phy
[63 + (128 * byte
)], 0xffff << 16,
620 /* PHY_896 PHY_FREQ_SEL_MULTICAST_EN 1bit offset_0 */
621 clrbits32(&denali_phy
[896], 1);
623 /* CTL_200 ctrlupd_req 1bit offset_8 */
624 clrsetbits32(&denali_ctl
[200], 0x1 << 8, 0x1 << 8);
627 static u32
get_rank_mask(u32 channel
, const struct rk3399_sdram_params
*params
)
629 const u32 rank
= params
->ch
[channel
].rank
;
631 /* required rank mask is different for LPDDR4 */
632 if (params
->dramtype
== LPDDR4
)
633 return (rank
== 1) ? 0x5 : 0xf;
635 return (rank
== 1) ? 0x1 : 0x3;
638 static int data_training_ca(u32 channel
, const struct rk3399_sdram_params
*params
)
640 u32
*denali_pi
= rk3399_ddr_pi
[channel
]->denali_pi
;
641 u32
*denali_phy
= rk3399_ddr_publ
[channel
]->denali_phy
;
642 u32 obs_0
, obs_1
, obs_2
;
643 const u32 rank_mask
= get_rank_mask(channel
, params
);
646 /* clear interrupt,PI_175 PI_INT_ACK:WR:0:17 */
647 write32(&denali_pi
[175], 0x00003f7c);
649 for (i
= 0; i
< MAX_RANKS_PER_CHANNEL
; i
++) {
650 if (!(rank_mask
& (1 << i
)))
653 select_per_cs_training_index(channel
, i
);
654 /* PI_100 PI_CALVL_EN:RW:8:2 */
655 clrsetbits32(&denali_pi
[100], 0x3 << 8, 0x2 << 8);
656 /* PI_92 PI_CALVL_REQ:WR:16:1,PI_CALVL_CS:RW:24:2 */
657 clrsetbits32(&denali_pi
[92], (0x1 << 16) | (0x3 << 24),
658 (0x1 << 16) | (i
<< 24));
661 /* PI_174 PI_INT_STATUS:RD:8:18 */
662 tmp
= read32(&denali_pi
[174]) >> 8;
666 * PHY_532/660/789 phy_adr_calvl_obs1_:0:32
668 obs_0
= read32(&denali_phy
[532]);
669 obs_1
= read32(&denali_phy
[660]);
670 obs_2
= read32(&denali_phy
[788]);
671 if (((obs_0
>> 30) & 0x3) ||
672 ((obs_1
>> 30) & 0x3) ||
673 ((obs_2
>> 30) & 0x3))
675 if ((((tmp
>> 11) & 0x1) == 0x1) &&
676 (((tmp
>> 13) & 0x1) == 0x1) &&
677 (((tmp
>> 5) & 0x1) == 0x0))
679 else if (((tmp
>> 5) & 0x1) == 0x1)
682 /* clear interrupt,PI_175 PI_INT_ACK:WR:0:17 */
683 write32(&denali_pi
[175], 0x00003f7c);
685 clrbits32(&denali_pi
[100], 0x3 << 8);
690 static int data_training_wl(u32 channel
, const struct rk3399_sdram_params
*params
)
692 u32
*denali_pi
= rk3399_ddr_pi
[channel
]->denali_pi
;
693 u32
*denali_phy
= rk3399_ddr_publ
[channel
]->denali_phy
;
694 u32 obs_0
, obs_1
, obs_2
, obs_3
;
695 u32 rank
= params
->ch
[channel
].rank
;
698 /* clear interrupt,PI_175 PI_INT_ACK:WR:0:17 */
699 write32(&denali_pi
[175], 0x00003f7c);
701 for (i
= 0; i
< rank
; i
++) {
702 select_per_cs_training_index(channel
, i
);
703 /* PI_60 PI_WRLVL_EN:RW:8:2 */
704 clrsetbits32(&denali_pi
[60], 0x3 << 8, 0x2 << 8);
705 /* PI_59 PI_WRLVL_REQ:WR:8:1,PI_WRLVL_CS:RW:16:2 */
706 clrsetbits32(&denali_pi
[59], (0x1 << 8) | (0x3 << 16), (0x1 << 8) | (i
<< 16));
709 /* PI_174 PI_INT_STATUS:RD:8:18 */
710 tmp
= read32(&denali_pi
[174]) >> 8;
713 * check status obs, if error maybe can not
714 * get leveling done PHY_40/168/296/424
715 * phy_wrlvl_status_obs_X:0:13
717 obs_0
= read32(&denali_phy
[40]);
718 obs_1
= read32(&denali_phy
[168]);
719 obs_2
= read32(&denali_phy
[296]);
720 obs_3
= read32(&denali_phy
[424]);
721 if (((obs_0
>> 12) & 0x1) ||
722 ((obs_1
>> 12) & 0x1) ||
723 ((obs_2
>> 12) & 0x1) ||
724 ((obs_3
>> 12) & 0x1))
726 if ((((tmp
>> 10) & 0x1) == 0x1) &&
727 (((tmp
>> 13) & 0x1) == 0x1) &&
728 (((tmp
>> 4) & 0x1) == 0x0))
730 else if (((tmp
>> 4) & 0x1) == 0x1)
733 /* clear interrupt,PI_175 PI_INT_ACK:WR:0:17 */
734 write32(&denali_pi
[175], 0x00003f7c);
737 override_write_leveling_value(channel
);
738 clrbits32(&denali_pi
[60], 0x3 << 8);
743 static int data_training_rg(u32 channel
, const struct rk3399_sdram_params
*params
)
745 u32
*denali_pi
= rk3399_ddr_pi
[channel
]->denali_pi
;
746 u32
*denali_phy
= rk3399_ddr_publ
[channel
]->denali_phy
;
747 u32 rank
= params
->ch
[channel
].rank
;
748 u32 obs_0
, obs_1
, obs_2
, obs_3
;
753 * The differential signal of DQS needs to keep low level
754 * before gate training. RPULL will connect 4Kn from PADP
755 * to VSS and a 4Kn from PADN to VDDQ to ensure it.
756 * But if it has PHY side ODT connect at this time,
757 * it will change the DQS signal level. So disable PHY
758 * side ODT before gate training and restore ODT state
759 * after gate training.
761 if (params
->dramtype
!= LPDDR4
) {
762 reg_value
= (read32(&denali_phy
[6]) >> 24) & 0x7;
765 * phy_dqs_tsel_enable_X 3bits
766 * DENALI_PHY_6/134/262/390 offset_24
768 clrbits32(&denali_phy
[6], 0x7 << 24);
769 clrbits32(&denali_phy
[134], 0x7 << 24);
770 clrbits32(&denali_phy
[262], 0x7 << 24);
771 clrbits32(&denali_phy
[390], 0x7 << 24);
774 /* clear interrupt,PI_175 PI_INT_ACK:WR:0:17 */
775 write32(&denali_pi
[175], 0x00003f7c);
777 for (i
= 0; i
< rank
; i
++) {
778 select_per_cs_training_index(channel
, i
);
779 /* PI_80 PI_RDLVL_GATE_EN:RW:24:2 */
780 clrsetbits32(&denali_pi
[80], 0x3 << 24, 0x2 << 24);
782 * PI_74 PI_RDLVL_GATE_REQ:WR:16:1
783 * PI_RDLVL_CS:RW:24:2
785 clrsetbits32(&denali_pi
[74], (0x1 << 16) | (0x3 << 24),
786 (0x1 << 16) | (i
<< 24));
789 /* PI_174 PI_INT_STATUS:RD:8:18 */
790 tmp
= read32(&denali_pi
[174]) >> 8;
795 * PHY_GTLVL_STATUS_OBS_x:16:8
797 obs_0
= read32(&denali_phy
[43]);
798 obs_1
= read32(&denali_phy
[171]);
799 obs_2
= read32(&denali_phy
[299]);
800 obs_3
= read32(&denali_phy
[427]);
801 if (((obs_0
>> (16 + 6)) & 0x3) ||
802 ((obs_1
>> (16 + 6)) & 0x3) ||
803 ((obs_2
>> (16 + 6)) & 0x3) ||
804 ((obs_3
>> (16 + 6)) & 0x3))
806 if ((((tmp
>> 9) & 0x1) == 0x1) &&
807 (((tmp
>> 13) & 0x1) == 0x1) &&
808 (((tmp
>> 3) & 0x1) == 0x0))
810 else if (((tmp
>> 3) & 0x1) == 0x1)
813 /* clear interrupt,PI_175 PI_INT_ACK:WR:0:17 */
814 write32(&denali_pi
[175], 0x00003f7c);
816 clrbits32(&denali_pi
[80], 0x3 << 24);
818 if (params
->dramtype
!= LPDDR4
) {
820 * phy_dqs_tsel_enable_X 3bits
821 * DENALI_PHY_6/134/262/390 offset_24
823 tmp
= reg_value
<< 24;
824 clrsetbits32(&denali_phy
[6], 0x7 << 24, tmp
);
825 clrsetbits32(&denali_phy
[134], 0x7 << 24, tmp
);
826 clrsetbits32(&denali_phy
[262], 0x7 << 24, tmp
);
827 clrsetbits32(&denali_phy
[390], 0x7 << 24, tmp
);
832 static int data_training_rl(u32 channel
, const struct rk3399_sdram_params
*params
)
834 u32 rank
= params
->ch
[channel
].rank
;
837 u32
*denali_pi
= rk3399_ddr_pi
[channel
]->denali_pi
;
839 /* clear interrupt,PI_175 PI_INT_ACK:WR:0:17 */
840 write32(&denali_pi
[175], 0x00003f7c);
842 for (i
= 0; i
< rank
; i
++) {
843 select_per_cs_training_index(channel
, i
);
844 /* PI_80 PI_RDLVL_EN:RW:16:2 */
845 clrsetbits32(&denali_pi
[80], 0x3 << 16, 0x2 << 16);
846 /* PI_74 PI_RDLVL_REQ:WR:8:1,PI_RDLVL_CS:RW:24:2 */
847 clrsetbits32(&denali_pi
[74], (0x1 << 8) | (0x3 << 24), (0x1 << 8) | (i
<< 24));
850 /* PI_174 PI_INT_STATUS:RD:8:18 */
851 tmp
= read32(&denali_pi
[174]) >> 8;
854 * make sure status obs not report error bit
856 * phy_rdlvl_status_obs_X:16:8
858 if ((((tmp
>> 8) & 0x1) == 0x1) && (((tmp
>> 13) & 0x1) == 0x1)
859 && (((tmp
>> 2) & 0x1) == 0x0))
861 else if (((tmp
>> 2) & 0x1) == 0x1)
864 /* clear interrupt,PI_175 PI_INT_ACK:WR:0:17 */
865 write32(&denali_pi
[175], 0x00003f7c);
867 clrbits32(&denali_pi
[80], 0x3 << 16);
872 static int data_training_wdql(u32 channel
, const struct rk3399_sdram_params
*params
)
874 u32
*denali_pi
= rk3399_ddr_pi
[channel
]->denali_pi
;
875 const u32 rank_mask
= get_rank_mask(channel
, params
);
878 /* clear interrupt,PI_175 PI_INT_ACK:WR:0:17 */
879 write32(&denali_pi
[175], 0x00003f7c);
881 for (i
= 0; i
< MAX_RANKS_PER_CHANNEL
; i
++) {
882 if (!(rank_mask
& (1 << i
)))
885 select_per_cs_training_index(channel
, i
);
887 * disable PI_WDQLVL_VREF_EN before wdq leveling?
888 * PI_181 PI_WDQLVL_VREF_EN:RW:8:1
890 clrbits32(&denali_pi
[181], 0x1 << 8);
891 /* PI_124 PI_WDQLVL_EN:RW:16:2 */
892 clrsetbits32(&denali_pi
[124], 0x3 << 16, 0x2 << 16);
893 /* PI_121 PI_WDQLVL_REQ:WR:8:1,PI_WDQLVL_CS:RW:16:2 */
894 clrsetbits32(&denali_pi
[121], (0x1 << 8) | (0x3 << 16), (0x1 << 8) | (i
<< 16));
897 /* PI_174 PI_INT_STATUS:RD:8:18 */
898 tmp
= read32(&denali_pi
[174]) >> 8;
899 if ((((tmp
>> 12) & 0x1) == 0x1) && (((tmp
>> 13) & 0x1) == 0x1)
900 && (((tmp
>> 6) & 0x1) == 0x0))
902 else if (((tmp
>> 6) & 0x1) == 0x1)
905 /* clear interrupt,PI_175 PI_INT_ACK:WR:0:17 */
906 write32(&denali_pi
[175], 0x00003f7c);
908 clrbits32(&denali_pi
[124], 0x3 << 16);
914 static int data_training(u32 channel
, const struct rk3399_sdram_params
*params
,
917 u32
*denali_phy
= rk3399_ddr_publ
[channel
]->denali_phy
;
920 /* PHY_927 PHY_PAD_DQS_DRIVE RPULL offset_22 */
921 setbits32(&denali_phy
[927], (1 << 22));
923 if (training_flag
== PI_FULL_TRAINING
) {
924 if (params
->dramtype
== LPDDR4
) {
925 training_flag
= PI_CA_TRAINING
| PI_WRITE_LEVELING
|
926 PI_READ_GATE_TRAINING
|
927 PI_READ_LEVELING
| PI_WDQ_LEVELING
;
928 } else if (params
->dramtype
== LPDDR3
) {
929 training_flag
= PI_CA_TRAINING
| PI_WRITE_LEVELING
|
930 PI_READ_GATE_TRAINING
;
931 } else if (params
->dramtype
== DDR3
) {
932 training_flag
= PI_WRITE_LEVELING
|
933 PI_READ_GATE_TRAINING
|
938 /* ca training(LPDDR4,LPDDR3 support) */
939 if ((training_flag
& PI_CA_TRAINING
) == PI_CA_TRAINING
) {
940 ret
= data_training_ca(channel
, params
);
942 printk(BIOS_ERR
, "Channel %d CA training failed\n", channel
);
947 /* write leveling(LPDDR4,LPDDR3,DDR3 support) */
948 if ((training_flag
& PI_WRITE_LEVELING
) == PI_WRITE_LEVELING
) {
949 ret
= data_training_wl(channel
, params
);
951 printk(BIOS_ERR
, "Channel %d WL training failed\n", channel
);
956 /* read gate training(LPDDR4,LPDDR3,DDR3 support) */
957 if ((training_flag
& PI_READ_GATE_TRAINING
) == PI_READ_GATE_TRAINING
) {
958 ret
= data_training_rg(channel
, params
);
960 printk(BIOS_ERR
, "Channel %d RG training failed\n", channel
);
965 /* read leveling(LPDDR4,LPDDR3,DDR3 support) */
966 if ((training_flag
& PI_READ_LEVELING
) == PI_READ_LEVELING
) {
967 ret
= data_training_rl(channel
, params
);
969 printk(BIOS_ERR
, "Channel %d RL training failed\n", channel
);
974 /* wdq leveling(LPDDR4 support) */
975 if ((training_flag
& PI_WDQ_LEVELING
) == PI_WDQ_LEVELING
) {
976 ret
= data_training_wdql(channel
, params
);
978 printk(BIOS_ERR
, "Channel %d WDQL training failed\n", channel
);
983 /* PHY_927 PHY_PAD_DQS_DRIVE RPULL offset_22 */
984 clrbits32(&denali_phy
[927], (1 << 22));
989 static void set_ddrconfig(const struct rk3399_sdram_params
*params
,
990 unsigned char channel
, u32 ddrconfig
)
992 /* only need to set ddrconfig */
993 struct rk3399_msch_regs
*ddr_msch_regs
= rk3399_msch
[channel
];
994 unsigned int cs0_cap
= 0;
995 unsigned int cs1_cap
= 0;
997 cs0_cap
= (1 << (params
->ch
[channel
].cs0_row
998 + params
->ch
[channel
].col
999 + params
->ch
[channel
].bk
1000 + params
->ch
[channel
].bw
- 20));
1001 if (params
->ch
[channel
].rank
> 1)
1002 cs1_cap
= cs0_cap
>> (params
->ch
[channel
].cs0_row
1003 - params
->ch
[channel
].cs1_row
);
1004 if (params
->ch
[channel
].row_3_4
) {
1005 cs0_cap
= cs0_cap
* 3 / 4;
1006 cs1_cap
= cs1_cap
* 3 / 4;
1009 write32(&ddr_msch_regs
->ddrconf
, ddrconfig
| (ddrconfig
<< 8));
1010 write32(&ddr_msch_regs
->ddrsize
, ((cs0_cap
/ 32) & 0xff) |
1011 (((cs1_cap
/ 32) & 0xff) << 8));
1014 static void dram_all_config(const struct rk3399_sdram_params
*params
)
1017 unsigned int channel
;
1020 sys_reg
|= SYS_REG_ENC_DDRTYPE(params
->dramtype
);
1021 sys_reg
|= SYS_REG_ENC_NUM_CH(params
->num_channels
);
1022 for (channel
= 0, use
= 0; (use
< params
->num_channels
) && (channel
< 2); channel
++) {
1023 const struct rk3399_sdram_channel
*info
= ¶ms
->ch
[channel
];
1024 struct rk3399_msch_regs
*ddr_msch_regs
;
1025 const struct rk3399_msch_timings
*noc_timing
;
1027 if (params
->ch
[channel
].col
== 0)
1030 sys_reg
|= SYS_REG_ENC_ROW_3_4(info
->row_3_4
, channel
);
1031 sys_reg
|= SYS_REG_ENC_CHINFO(channel
);
1032 sys_reg
|= SYS_REG_ENC_RANK(info
->rank
, channel
);
1033 sys_reg
|= SYS_REG_ENC_COL(info
->col
, channel
);
1034 sys_reg
|= SYS_REG_ENC_BK(info
->bk
, channel
);
1035 sys_reg
|= SYS_REG_ENC_CS0_ROW(info
->cs0_row
, channel
);
1036 if (params
->ch
[channel
].rank
> 1)
1037 sys_reg
|= SYS_REG_ENC_CS1_ROW(info
->cs1_row
, channel
);
1038 sys_reg
|= SYS_REG_ENC_BW(info
->bw
, channel
);
1039 sys_reg
|= SYS_REG_ENC_DBW(info
->dbw
, channel
);
1041 ddr_msch_regs
= rk3399_msch
[channel
];
1042 noc_timing
= ¶ms
->ch
[channel
].noc_timings
;
1043 write32(&ddr_msch_regs
->ddrtiminga0
.d32
,
1044 noc_timing
->ddrtiminga0
.d32
);
1045 write32(&ddr_msch_regs
->ddrtimingb0
.d32
,
1046 noc_timing
->ddrtimingb0
.d32
);
1047 write32(&ddr_msch_regs
->ddrtimingc0
.d32
,
1048 noc_timing
->ddrtimingc0
.d32
);
1049 write32(&ddr_msch_regs
->devtodev0
.d32
,
1050 noc_timing
->devtodev0
.d32
);
1051 write32(&ddr_msch_regs
->ddrmode
.d32
,
1052 noc_timing
->ddrmode
.d32
);
1054 /* rank 1 memory clock disable (dfi_dram_clk_disable = 1) */
1055 if (params
->ch
[channel
].rank
== 1)
1056 setbits32(&rk3399_ddr_pctl
[channel
]->denali_ctl
[276],
1060 write32(&rk3399_pmugrf
->os_reg2
, sys_reg
);
1061 DDR_STRIDE(params
->stride
);
1063 /* reboot hold register set */
1064 write32(&pmucru_ptr
->pmucru_rstnhold_con
[1],
1065 PRESET_SGRF_HOLD(0) | PRESET_GPIO0_HOLD(1) |
1066 PRESET_GPIO1_HOLD(1));
1067 clrsetbits32(&cru_ptr
->glb_rst_con
, 0x3, 0x3);
1070 static void switch_to_phy_index1(const struct rk3399_sdram_params
*params
)
1074 struct stopwatch sw
;
1075 u32 ch_count
= params
->num_channels
;
1077 stopwatch_init_msecs_expire(&sw
, 100);
1078 write32(&rk3399_ddr_cic
->cic_ctrl0
,
1079 RK_CLRSETBITS(0x03 << 4 | 1 << 2 | 1,
1080 1 << 4 | 1 << 2 | 1));
1081 while (!(read32(&rk3399_ddr_cic
->cic_status0
) & (1 << 2))) {
1082 if (stopwatch_expired(&sw
)) {
1084 "index1 frequency change overtime, reset\n");
1089 stopwatch_init_msecs_expire(&sw
, 100);
1090 write32(&rk3399_ddr_cic
->cic_ctrl0
, RK_CLRSETBITS(1 << 1, 1 << 1));
1091 while (!(read32(&rk3399_ddr_cic
->cic_status0
) & (1 << 0))) {
1092 if (stopwatch_expired(&sw
)) {
1094 "index1 frequency done overtime, reset\n");
1099 for (channel
= 0; channel
< ch_count
; channel
++) {
1100 denali_phy
= rk3399_ddr_publ
[channel
]->denali_phy
;
1101 clrsetbits32(&denali_phy
[896], (0x3 << 8) | 1, 1 << 8);
1102 if (data_training(channel
, params
, PI_FULL_TRAINING
)) {
1103 printk(BIOS_ERR
, "index1 training failed, reset\n");
1109 void sdram_init(const struct rk3399_sdram_params
*params
)
1111 unsigned char dramtype
= params
->dramtype
;
1112 unsigned int ddr_freq
= params
->ddr_freq
;
1115 printk(BIOS_INFO
, "Starting SDRAM initialization...\n");
1117 if ((dramtype
== DDR3
&& ddr_freq
> 800*MHz
) ||
1118 (dramtype
== LPDDR3
&& ddr_freq
> 933*MHz
) ||
1119 (dramtype
== LPDDR4
&& ddr_freq
> 800*MHz
))
1120 die("SDRAM frequency is to high!");
1122 rkclk_configure_ddr(ddr_freq
);
1124 for (channel
= 0; channel
< 2; channel
++) {
1125 phy_pctrl_reset(channel
);
1126 phy_dll_bypass_set(rk3399_ddr_publ
[channel
], ddr_freq
);
1128 if (channel
>= params
->num_channels
)
1132 * TODO: we need to find the root cause why this
1133 * step may fail, before that, we just reset the
1134 * system, and start again.
1136 if (pctl_cfg(channel
, params
)) {
1137 printk(BIOS_ERR
, "pctl_cfg fail, reset\n");
1141 /* LPDDR2/LPDDR3 need to wait DAI complete, max 10us */
1142 if (dramtype
== LPDDR3
)
1145 if (data_training(channel
, params
, PI_FULL_TRAINING
)) {
1146 printk(BIOS_ERR
, "SDRAM initialization failed, reset\n");
1150 set_ddrconfig(params
, channel
, params
->ch
[channel
].ddrconfig
);
1152 dram_all_config(params
);
1153 switch_to_phy_index1(params
);
1155 printk(BIOS_INFO
, "Finish SDRAM initialization...\n");
1158 size_t sdram_size_mb(void)
1160 u32 rank
, col
, bk
, cs0_row
, cs1_row
, bw
, row_3_4
;
1161 size_t chipsize_mb
= 0;
1162 static size_t size_mb
= 0;
1166 u32 sys_reg
= read32(&rk3399_pmugrf
->os_reg2
);
1167 u32 ch_num
= SYS_REG_DEC_NUM_CH(sys_reg
);
1169 for (ch
= 0; ch
< ch_num
; ch
++) {
1170 rank
= SYS_REG_DEC_RANK(sys_reg
, ch
);
1171 col
= SYS_REG_DEC_COL(sys_reg
, ch
);
1172 bk
= SYS_REG_DEC_BK(sys_reg
, ch
);
1173 cs0_row
= SYS_REG_DEC_CS0_ROW(sys_reg
, ch
);
1174 cs1_row
= SYS_REG_DEC_CS1_ROW(sys_reg
, ch
);
1175 bw
= SYS_REG_DEC_BW(sys_reg
, ch
);
1176 row_3_4
= SYS_REG_DEC_ROW_3_4(sys_reg
, ch
);
1178 chipsize_mb
= (1 << (cs0_row
+ col
+ bk
+ bw
- 20));
1181 chipsize_mb
+= chipsize_mb
>>
1182 (cs0_row
- cs1_row
);
1184 chipsize_mb
= chipsize_mb
* 3 / 4;
1185 size_mb
+= chipsize_mb
;
1189 * we use the 0x00000000~0xf7ffffff space
1190 * since 0xf8000000~0xffffffff is soc register space
1193 size_mb
= MIN(size_mb
, 0xf8000000/MiB
);