2 * Copyright © 2018 Intel Corporation
4 * Permission is hereby granted, free of charge, to any person obtaining a
5 * copy of this software and associated documentation files (the "Software"),
6 * to deal in the Software without restriction, including without limitation
7 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
8 * and/or sell copies of the Software, and to permit persons to whom the
9 * Software is furnished to do so, subject to the following conditions:
11 * The above copyright notice and this permission notice (including the next
12 * paragraph) shall be included in all copies or substantial portions of the
15 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
16 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
17 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
18 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
19 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
20 * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
21 * DEALINGS IN THE SOFTWARE.
24 * Madhav Chauhan <madhav.chauhan@intel.com>
25 * Jani Nikula <jani.nikula@intel.com>
28 #include <drm/drm_atomic_helper.h>
29 #include <drm/drm_mipi_dsi.h>
31 #include "intel_atomic.h"
32 #include "intel_combo_phy.h"
33 #include "intel_connector.h"
34 #include "intel_ddi.h"
35 #include "intel_dsi.h"
36 #include "intel_panel.h"
37 #include "intel_vdsc.h"
39 static int header_credits_available(struct drm_i915_private
*dev_priv
,
40 enum transcoder dsi_trans
)
42 return (intel_de_read(dev_priv
, DSI_CMD_TXCTL(dsi_trans
)) & FREE_HEADER_CREDIT_MASK
)
43 >> FREE_HEADER_CREDIT_SHIFT
;
46 static int payload_credits_available(struct drm_i915_private
*dev_priv
,
47 enum transcoder dsi_trans
)
49 return (intel_de_read(dev_priv
, DSI_CMD_TXCTL(dsi_trans
)) & FREE_PLOAD_CREDIT_MASK
)
50 >> FREE_PLOAD_CREDIT_SHIFT
;
53 static void wait_for_header_credits(struct drm_i915_private
*dev_priv
,
54 enum transcoder dsi_trans
)
56 if (wait_for_us(header_credits_available(dev_priv
, dsi_trans
) >=
57 MAX_HEADER_CREDIT
, 100))
58 drm_err(&dev_priv
->drm
, "DSI header credits not released\n");
61 static void wait_for_payload_credits(struct drm_i915_private
*dev_priv
,
62 enum transcoder dsi_trans
)
64 if (wait_for_us(payload_credits_available(dev_priv
, dsi_trans
) >=
65 MAX_PLOAD_CREDIT
, 100))
66 drm_err(&dev_priv
->drm
, "DSI payload credits not released\n");
69 static enum transcoder
dsi_port_to_transcoder(enum port port
)
72 return TRANSCODER_DSI_0
;
74 return TRANSCODER_DSI_1
;
77 static void wait_for_cmds_dispatched_to_panel(struct intel_encoder
*encoder
)
79 struct drm_i915_private
*dev_priv
= to_i915(encoder
->base
.dev
);
80 struct intel_dsi
*intel_dsi
= enc_to_intel_dsi(encoder
);
81 struct mipi_dsi_device
*dsi
;
83 enum transcoder dsi_trans
;
86 /* wait for header/payload credits to be released */
87 for_each_dsi_port(port
, intel_dsi
->ports
) {
88 dsi_trans
= dsi_port_to_transcoder(port
);
89 wait_for_header_credits(dev_priv
, dsi_trans
);
90 wait_for_payload_credits(dev_priv
, dsi_trans
);
93 /* send nop DCS command */
94 for_each_dsi_port(port
, intel_dsi
->ports
) {
95 dsi
= intel_dsi
->dsi_hosts
[port
]->device
;
96 dsi
->mode_flags
|= MIPI_DSI_MODE_LPM
;
98 ret
= mipi_dsi_dcs_nop(dsi
);
100 drm_err(&dev_priv
->drm
,
101 "error sending DCS NOP command\n");
104 /* wait for header credits to be released */
105 for_each_dsi_port(port
, intel_dsi
->ports
) {
106 dsi_trans
= dsi_port_to_transcoder(port
);
107 wait_for_header_credits(dev_priv
, dsi_trans
);
110 /* wait for LP TX in progress bit to be cleared */
111 for_each_dsi_port(port
, intel_dsi
->ports
) {
112 dsi_trans
= dsi_port_to_transcoder(port
);
113 if (wait_for_us(!(intel_de_read(dev_priv
, DSI_LP_MSG(dsi_trans
)) &
114 LPTX_IN_PROGRESS
), 20))
115 drm_err(&dev_priv
->drm
, "LPTX bit not cleared\n");
119 static bool add_payld_to_queue(struct intel_dsi_host
*host
, const u8
*data
,
122 struct intel_dsi
*intel_dsi
= host
->intel_dsi
;
123 struct drm_i915_private
*dev_priv
= to_i915(intel_dsi
->base
.base
.dev
);
124 enum transcoder dsi_trans
= dsi_port_to_transcoder(host
->port
);
128 for (i
= 0; i
< len
; i
+= 4) {
131 free_credits
= payload_credits_available(dev_priv
, dsi_trans
);
132 if (free_credits
< 1) {
133 drm_err(&dev_priv
->drm
,
134 "Payload credit not available\n");
138 for (j
= 0; j
< min_t(u32
, len
- i
, 4); j
++)
139 tmp
|= *data
++ << 8 * j
;
141 intel_de_write(dev_priv
, DSI_CMD_TXPYLD(dsi_trans
), tmp
);
147 static int dsi_send_pkt_hdr(struct intel_dsi_host
*host
,
148 struct mipi_dsi_packet pkt
, bool enable_lpdt
)
150 struct intel_dsi
*intel_dsi
= host
->intel_dsi
;
151 struct drm_i915_private
*dev_priv
= to_i915(intel_dsi
->base
.base
.dev
);
152 enum transcoder dsi_trans
= dsi_port_to_transcoder(host
->port
);
156 /* check if header credit available */
157 free_credits
= header_credits_available(dev_priv
, dsi_trans
);
158 if (free_credits
< 1) {
159 drm_err(&dev_priv
->drm
,
160 "send pkt header failed, not enough hdr credits\n");
164 tmp
= intel_de_read(dev_priv
, DSI_CMD_TXHDR(dsi_trans
));
167 tmp
|= PAYLOAD_PRESENT
;
169 tmp
&= ~PAYLOAD_PRESENT
;
171 tmp
&= ~VBLANK_FENCE
;
174 tmp
|= LP_DATA_TRANSFER
;
176 tmp
&= ~(PARAM_WC_MASK
| VC_MASK
| DT_MASK
);
177 tmp
|= ((pkt
.header
[0] & VC_MASK
) << VC_SHIFT
);
178 tmp
|= ((pkt
.header
[0] & DT_MASK
) << DT_SHIFT
);
179 tmp
|= (pkt
.header
[1] << PARAM_WC_LOWER_SHIFT
);
180 tmp
|= (pkt
.header
[2] << PARAM_WC_UPPER_SHIFT
);
181 intel_de_write(dev_priv
, DSI_CMD_TXHDR(dsi_trans
), tmp
);
186 static int dsi_send_pkt_payld(struct intel_dsi_host
*host
,
187 struct mipi_dsi_packet pkt
)
189 struct intel_dsi
*intel_dsi
= host
->intel_dsi
;
190 struct drm_i915_private
*i915
= to_i915(intel_dsi
->base
.base
.dev
);
192 /* payload queue can accept *256 bytes*, check limit */
193 if (pkt
.payload_length
> MAX_PLOAD_CREDIT
* 4) {
194 drm_err(&i915
->drm
, "payload size exceeds max queue limit\n");
198 /* load data into command payload queue */
199 if (!add_payld_to_queue(host
, pkt
.payload
,
200 pkt
.payload_length
)) {
201 drm_err(&i915
->drm
, "adding payload to queue failed\n");
208 void icl_dsi_frame_update(struct intel_crtc_state
*crtc_state
)
210 struct intel_crtc
*crtc
= to_intel_crtc(crtc_state
->uapi
.crtc
);
211 struct drm_i915_private
*dev_priv
= to_i915(crtc
->base
.dev
);
215 mode_flags
= crtc_state
->mode_flags
;
218 * case 1 also covers dual link
219 * In case of dual link, frame update should be set on
222 if (mode_flags
& I915_MODE_FLAG_DSI_USE_TE0
)
224 else if (mode_flags
& I915_MODE_FLAG_DSI_USE_TE1
)
229 tmp
= intel_de_read(dev_priv
, DSI_CMD_FRMCTL(port
));
230 tmp
|= DSI_FRAME_UPDATE_REQUEST
;
231 intel_de_write(dev_priv
, DSI_CMD_FRMCTL(port
), tmp
);
234 static void dsi_program_swing_and_deemphasis(struct intel_encoder
*encoder
)
236 struct drm_i915_private
*dev_priv
= to_i915(encoder
->base
.dev
);
237 struct intel_dsi
*intel_dsi
= enc_to_intel_dsi(encoder
);
242 for_each_dsi_phy(phy
, intel_dsi
->phys
) {
244 * Program voltage swing and pre-emphasis level values as per
245 * table in BSPEC under DDI buffer programing
247 tmp
= intel_de_read(dev_priv
, ICL_PORT_TX_DW5_LN0(phy
));
248 tmp
&= ~(SCALING_MODE_SEL_MASK
| RTERM_SELECT_MASK
);
249 tmp
|= SCALING_MODE_SEL(0x2);
250 tmp
|= TAP2_DISABLE
| TAP3_DISABLE
;
251 tmp
|= RTERM_SELECT(0x6);
252 intel_de_write(dev_priv
, ICL_PORT_TX_DW5_GRP(phy
), tmp
);
254 tmp
= intel_de_read(dev_priv
, ICL_PORT_TX_DW5_AUX(phy
));
255 tmp
&= ~(SCALING_MODE_SEL_MASK
| RTERM_SELECT_MASK
);
256 tmp
|= SCALING_MODE_SEL(0x2);
257 tmp
|= TAP2_DISABLE
| TAP3_DISABLE
;
258 tmp
|= RTERM_SELECT(0x6);
259 intel_de_write(dev_priv
, ICL_PORT_TX_DW5_AUX(phy
), tmp
);
261 tmp
= intel_de_read(dev_priv
, ICL_PORT_TX_DW2_LN0(phy
));
262 tmp
&= ~(SWING_SEL_LOWER_MASK
| SWING_SEL_UPPER_MASK
|
264 tmp
|= SWING_SEL_UPPER(0x2);
265 tmp
|= SWING_SEL_LOWER(0x2);
266 tmp
|= RCOMP_SCALAR(0x98);
267 intel_de_write(dev_priv
, ICL_PORT_TX_DW2_GRP(phy
), tmp
);
269 tmp
= intel_de_read(dev_priv
, ICL_PORT_TX_DW2_AUX(phy
));
270 tmp
&= ~(SWING_SEL_LOWER_MASK
| SWING_SEL_UPPER_MASK
|
272 tmp
|= SWING_SEL_UPPER(0x2);
273 tmp
|= SWING_SEL_LOWER(0x2);
274 tmp
|= RCOMP_SCALAR(0x98);
275 intel_de_write(dev_priv
, ICL_PORT_TX_DW2_AUX(phy
), tmp
);
277 tmp
= intel_de_read(dev_priv
, ICL_PORT_TX_DW4_AUX(phy
));
278 tmp
&= ~(POST_CURSOR_1_MASK
| POST_CURSOR_2_MASK
|
280 tmp
|= POST_CURSOR_1(0x0);
281 tmp
|= POST_CURSOR_2(0x0);
282 tmp
|= CURSOR_COEFF(0x3f);
283 intel_de_write(dev_priv
, ICL_PORT_TX_DW4_AUX(phy
), tmp
);
285 for (lane
= 0; lane
<= 3; lane
++) {
286 /* Bspec: must not use GRP register for write */
287 tmp
= intel_de_read(dev_priv
,
288 ICL_PORT_TX_DW4_LN(lane
, phy
));
289 tmp
&= ~(POST_CURSOR_1_MASK
| POST_CURSOR_2_MASK
|
291 tmp
|= POST_CURSOR_1(0x0);
292 tmp
|= POST_CURSOR_2(0x0);
293 tmp
|= CURSOR_COEFF(0x3f);
294 intel_de_write(dev_priv
,
295 ICL_PORT_TX_DW4_LN(lane
, phy
), tmp
);
300 static void configure_dual_link_mode(struct intel_encoder
*encoder
,
301 const struct intel_crtc_state
*pipe_config
)
303 struct drm_i915_private
*dev_priv
= to_i915(encoder
->base
.dev
);
304 struct intel_dsi
*intel_dsi
= enc_to_intel_dsi(encoder
);
307 dss_ctl1
= intel_de_read(dev_priv
, DSS_CTL1
);
308 dss_ctl1
|= SPLITTER_ENABLE
;
309 dss_ctl1
&= ~OVERLAP_PIXELS_MASK
;
310 dss_ctl1
|= OVERLAP_PIXELS(intel_dsi
->pixel_overlap
);
312 if (intel_dsi
->dual_link
== DSI_DUAL_LINK_FRONT_BACK
) {
313 const struct drm_display_mode
*adjusted_mode
=
314 &pipe_config
->hw
.adjusted_mode
;
316 u16 hactive
= adjusted_mode
->crtc_hdisplay
;
319 dss_ctl1
&= ~DUAL_LINK_MODE_INTERLEAVE
;
320 dl_buffer_depth
= hactive
/ 2 + intel_dsi
->pixel_overlap
;
322 if (dl_buffer_depth
> MAX_DL_BUFFER_TARGET_DEPTH
)
323 drm_err(&dev_priv
->drm
,
324 "DL buffer depth exceed max value\n");
326 dss_ctl1
&= ~LEFT_DL_BUF_TARGET_DEPTH_MASK
;
327 dss_ctl1
|= LEFT_DL_BUF_TARGET_DEPTH(dl_buffer_depth
);
328 dss_ctl2
= intel_de_read(dev_priv
, DSS_CTL2
);
329 dss_ctl2
&= ~RIGHT_DL_BUF_TARGET_DEPTH_MASK
;
330 dss_ctl2
|= RIGHT_DL_BUF_TARGET_DEPTH(dl_buffer_depth
);
331 intel_de_write(dev_priv
, DSS_CTL2
, dss_ctl2
);
334 dss_ctl1
|= DUAL_LINK_MODE_INTERLEAVE
;
337 intel_de_write(dev_priv
, DSS_CTL1
, dss_ctl1
);
340 /* aka DSI 8X clock */
341 static int afe_clk(struct intel_encoder
*encoder
,
342 const struct intel_crtc_state
*crtc_state
)
344 struct intel_dsi
*intel_dsi
= enc_to_intel_dsi(encoder
);
347 if (crtc_state
->dsc
.compression_enable
)
348 bpp
= crtc_state
->dsc
.compressed_bpp
;
350 bpp
= mipi_dsi_pixel_format_to_bpp(intel_dsi
->pixel_format
);
352 return DIV_ROUND_CLOSEST(intel_dsi
->pclk
* bpp
, intel_dsi
->lane_count
);
355 static void gen11_dsi_program_esc_clk_div(struct intel_encoder
*encoder
,
356 const struct intel_crtc_state
*crtc_state
)
358 struct drm_i915_private
*dev_priv
= to_i915(encoder
->base
.dev
);
359 struct intel_dsi
*intel_dsi
= enc_to_intel_dsi(encoder
);
364 afe_clk_khz
= afe_clk(encoder
, crtc_state
);
365 esc_clk_div_m
= DIV_ROUND_UP(afe_clk_khz
, DSI_MAX_ESC_CLK
);
367 for_each_dsi_port(port
, intel_dsi
->ports
) {
368 intel_de_write(dev_priv
, ICL_DSI_ESC_CLK_DIV(port
),
369 esc_clk_div_m
& ICL_ESC_CLK_DIV_MASK
);
370 intel_de_posting_read(dev_priv
, ICL_DSI_ESC_CLK_DIV(port
));
373 for_each_dsi_port(port
, intel_dsi
->ports
) {
374 intel_de_write(dev_priv
, ICL_DPHY_ESC_CLK_DIV(port
),
375 esc_clk_div_m
& ICL_ESC_CLK_DIV_MASK
);
376 intel_de_posting_read(dev_priv
, ICL_DPHY_ESC_CLK_DIV(port
));
380 static void get_dsi_io_power_domains(struct drm_i915_private
*dev_priv
,
381 struct intel_dsi
*intel_dsi
)
385 for_each_dsi_port(port
, intel_dsi
->ports
) {
386 drm_WARN_ON(&dev_priv
->drm
, intel_dsi
->io_wakeref
[port
]);
387 intel_dsi
->io_wakeref
[port
] =
388 intel_display_power_get(dev_priv
,
390 POWER_DOMAIN_PORT_DDI_A_IO
:
391 POWER_DOMAIN_PORT_DDI_B_IO
);
395 static void gen11_dsi_enable_io_power(struct intel_encoder
*encoder
)
397 struct drm_i915_private
*dev_priv
= to_i915(encoder
->base
.dev
);
398 struct intel_dsi
*intel_dsi
= enc_to_intel_dsi(encoder
);
402 for_each_dsi_port(port
, intel_dsi
->ports
) {
403 tmp
= intel_de_read(dev_priv
, ICL_DSI_IO_MODECTL(port
));
404 tmp
|= COMBO_PHY_MODE_DSI
;
405 intel_de_write(dev_priv
, ICL_DSI_IO_MODECTL(port
), tmp
);
408 get_dsi_io_power_domains(dev_priv
, intel_dsi
);
411 static void gen11_dsi_power_up_lanes(struct intel_encoder
*encoder
)
413 struct drm_i915_private
*dev_priv
= to_i915(encoder
->base
.dev
);
414 struct intel_dsi
*intel_dsi
= enc_to_intel_dsi(encoder
);
417 for_each_dsi_phy(phy
, intel_dsi
->phys
)
418 intel_combo_phy_power_up_lanes(dev_priv
, phy
, true,
419 intel_dsi
->lane_count
, false);
422 static void gen11_dsi_config_phy_lanes_sequence(struct intel_encoder
*encoder
)
424 struct drm_i915_private
*dev_priv
= to_i915(encoder
->base
.dev
);
425 struct intel_dsi
*intel_dsi
= enc_to_intel_dsi(encoder
);
430 /* Step 4b(i) set loadgen select for transmit and aux lanes */
431 for_each_dsi_phy(phy
, intel_dsi
->phys
) {
432 tmp
= intel_de_read(dev_priv
, ICL_PORT_TX_DW4_AUX(phy
));
433 tmp
&= ~LOADGEN_SELECT
;
434 intel_de_write(dev_priv
, ICL_PORT_TX_DW4_AUX(phy
), tmp
);
435 for (lane
= 0; lane
<= 3; lane
++) {
436 tmp
= intel_de_read(dev_priv
,
437 ICL_PORT_TX_DW4_LN(lane
, phy
));
438 tmp
&= ~LOADGEN_SELECT
;
440 tmp
|= LOADGEN_SELECT
;
441 intel_de_write(dev_priv
,
442 ICL_PORT_TX_DW4_LN(lane
, phy
), tmp
);
446 /* Step 4b(ii) set latency optimization for transmit and aux lanes */
447 for_each_dsi_phy(phy
, intel_dsi
->phys
) {
448 tmp
= intel_de_read(dev_priv
, ICL_PORT_TX_DW2_AUX(phy
));
449 tmp
&= ~FRC_LATENCY_OPTIM_MASK
;
450 tmp
|= FRC_LATENCY_OPTIM_VAL(0x5);
451 intel_de_write(dev_priv
, ICL_PORT_TX_DW2_AUX(phy
), tmp
);
452 tmp
= intel_de_read(dev_priv
, ICL_PORT_TX_DW2_LN0(phy
));
453 tmp
&= ~FRC_LATENCY_OPTIM_MASK
;
454 tmp
|= FRC_LATENCY_OPTIM_VAL(0x5);
455 intel_de_write(dev_priv
, ICL_PORT_TX_DW2_GRP(phy
), tmp
);
457 /* For EHL, TGL, set latency optimization for PCS_DW1 lanes */
458 if (IS_JSL_EHL(dev_priv
) || (INTEL_GEN(dev_priv
) >= 12)) {
459 tmp
= intel_de_read(dev_priv
,
460 ICL_PORT_PCS_DW1_AUX(phy
));
461 tmp
&= ~LATENCY_OPTIM_MASK
;
462 tmp
|= LATENCY_OPTIM_VAL(0);
463 intel_de_write(dev_priv
, ICL_PORT_PCS_DW1_AUX(phy
),
466 tmp
= intel_de_read(dev_priv
,
467 ICL_PORT_PCS_DW1_LN0(phy
));
468 tmp
&= ~LATENCY_OPTIM_MASK
;
469 tmp
|= LATENCY_OPTIM_VAL(0x1);
470 intel_de_write(dev_priv
, ICL_PORT_PCS_DW1_GRP(phy
),
477 static void gen11_dsi_voltage_swing_program_seq(struct intel_encoder
*encoder
)
479 struct drm_i915_private
*dev_priv
= to_i915(encoder
->base
.dev
);
480 struct intel_dsi
*intel_dsi
= enc_to_intel_dsi(encoder
);
484 /* clear common keeper enable bit */
485 for_each_dsi_phy(phy
, intel_dsi
->phys
) {
486 tmp
= intel_de_read(dev_priv
, ICL_PORT_PCS_DW1_LN0(phy
));
487 tmp
&= ~COMMON_KEEPER_EN
;
488 intel_de_write(dev_priv
, ICL_PORT_PCS_DW1_GRP(phy
), tmp
);
489 tmp
= intel_de_read(dev_priv
, ICL_PORT_PCS_DW1_AUX(phy
));
490 tmp
&= ~COMMON_KEEPER_EN
;
491 intel_de_write(dev_priv
, ICL_PORT_PCS_DW1_AUX(phy
), tmp
);
495 * Set SUS Clock Config bitfield to 11b
496 * Note: loadgen select program is done
497 * as part of lane phy sequence configuration
499 for_each_dsi_phy(phy
, intel_dsi
->phys
) {
500 tmp
= intel_de_read(dev_priv
, ICL_PORT_CL_DW5(phy
));
501 tmp
|= SUS_CLOCK_CONFIG
;
502 intel_de_write(dev_priv
, ICL_PORT_CL_DW5(phy
), tmp
);
505 /* Clear training enable to change swing values */
506 for_each_dsi_phy(phy
, intel_dsi
->phys
) {
507 tmp
= intel_de_read(dev_priv
, ICL_PORT_TX_DW5_LN0(phy
));
508 tmp
&= ~TX_TRAINING_EN
;
509 intel_de_write(dev_priv
, ICL_PORT_TX_DW5_GRP(phy
), tmp
);
510 tmp
= intel_de_read(dev_priv
, ICL_PORT_TX_DW5_AUX(phy
));
511 tmp
&= ~TX_TRAINING_EN
;
512 intel_de_write(dev_priv
, ICL_PORT_TX_DW5_AUX(phy
), tmp
);
515 /* Program swing and de-emphasis */
516 dsi_program_swing_and_deemphasis(encoder
);
518 /* Set training enable to trigger update */
519 for_each_dsi_phy(phy
, intel_dsi
->phys
) {
520 tmp
= intel_de_read(dev_priv
, ICL_PORT_TX_DW5_LN0(phy
));
521 tmp
|= TX_TRAINING_EN
;
522 intel_de_write(dev_priv
, ICL_PORT_TX_DW5_GRP(phy
), tmp
);
523 tmp
= intel_de_read(dev_priv
, ICL_PORT_TX_DW5_AUX(phy
));
524 tmp
|= TX_TRAINING_EN
;
525 intel_de_write(dev_priv
, ICL_PORT_TX_DW5_AUX(phy
), tmp
);
529 static void gen11_dsi_enable_ddi_buffer(struct intel_encoder
*encoder
)
531 struct drm_i915_private
*dev_priv
= to_i915(encoder
->base
.dev
);
532 struct intel_dsi
*intel_dsi
= enc_to_intel_dsi(encoder
);
536 for_each_dsi_port(port
, intel_dsi
->ports
) {
537 tmp
= intel_de_read(dev_priv
, DDI_BUF_CTL(port
));
538 tmp
|= DDI_BUF_CTL_ENABLE
;
539 intel_de_write(dev_priv
, DDI_BUF_CTL(port
), tmp
);
541 if (wait_for_us(!(intel_de_read(dev_priv
, DDI_BUF_CTL(port
)) &
544 drm_err(&dev_priv
->drm
, "DDI port:%c buffer idle\n",
550 gen11_dsi_setup_dphy_timings(struct intel_encoder
*encoder
,
551 const struct intel_crtc_state
*crtc_state
)
553 struct drm_i915_private
*dev_priv
= to_i915(encoder
->base
.dev
);
554 struct intel_dsi
*intel_dsi
= enc_to_intel_dsi(encoder
);
559 /* Program T-INIT master registers */
560 for_each_dsi_port(port
, intel_dsi
->ports
) {
561 tmp
= intel_de_read(dev_priv
, ICL_DSI_T_INIT_MASTER(port
));
562 tmp
&= ~MASTER_INIT_TIMER_MASK
;
563 tmp
|= intel_dsi
->init_count
;
564 intel_de_write(dev_priv
, ICL_DSI_T_INIT_MASTER(port
), tmp
);
567 /* Program DPHY clock lanes timings */
568 for_each_dsi_port(port
, intel_dsi
->ports
) {
569 intel_de_write(dev_priv
, DPHY_CLK_TIMING_PARAM(port
),
570 intel_dsi
->dphy_reg
);
572 /* shadow register inside display core */
573 intel_de_write(dev_priv
, DSI_CLK_TIMING_PARAM(port
),
574 intel_dsi
->dphy_reg
);
577 /* Program DPHY data lanes timings */
578 for_each_dsi_port(port
, intel_dsi
->ports
) {
579 intel_de_write(dev_priv
, DPHY_DATA_TIMING_PARAM(port
),
580 intel_dsi
->dphy_data_lane_reg
);
582 /* shadow register inside display core */
583 intel_de_write(dev_priv
, DSI_DATA_TIMING_PARAM(port
),
584 intel_dsi
->dphy_data_lane_reg
);
588 * If DSI link operating at or below an 800 MHz,
589 * TA_SURE should be override and programmed to
590 * a value '0' inside TA_PARAM_REGISTERS otherwise
591 * leave all fields at HW default values.
593 if (IS_GEN(dev_priv
, 11)) {
594 if (afe_clk(encoder
, crtc_state
) <= 800000) {
595 for_each_dsi_port(port
, intel_dsi
->ports
) {
596 tmp
= intel_de_read(dev_priv
,
597 DPHY_TA_TIMING_PARAM(port
));
598 tmp
&= ~TA_SURE_MASK
;
599 tmp
|= TA_SURE_OVERRIDE
| TA_SURE(0);
600 intel_de_write(dev_priv
,
601 DPHY_TA_TIMING_PARAM(port
),
604 /* shadow register inside display core */
605 tmp
= intel_de_read(dev_priv
,
606 DSI_TA_TIMING_PARAM(port
));
607 tmp
&= ~TA_SURE_MASK
;
608 tmp
|= TA_SURE_OVERRIDE
| TA_SURE(0);
609 intel_de_write(dev_priv
,
610 DSI_TA_TIMING_PARAM(port
), tmp
);
615 if (IS_JSL_EHL(dev_priv
)) {
616 for_each_dsi_phy(phy
, intel_dsi
->phys
) {
617 tmp
= intel_de_read(dev_priv
, ICL_DPHY_CHKN(phy
));
618 tmp
|= ICL_DPHY_CHKN_AFE_OVER_PPI_STRAP
;
619 intel_de_write(dev_priv
, ICL_DPHY_CHKN(phy
), tmp
);
624 static void gen11_dsi_gate_clocks(struct intel_encoder
*encoder
)
626 struct drm_i915_private
*dev_priv
= to_i915(encoder
->base
.dev
);
627 struct intel_dsi
*intel_dsi
= enc_to_intel_dsi(encoder
);
631 mutex_lock(&dev_priv
->dpll
.lock
);
632 tmp
= intel_de_read(dev_priv
, ICL_DPCLKA_CFGCR0
);
633 for_each_dsi_phy(phy
, intel_dsi
->phys
)
634 tmp
|= ICL_DPCLKA_CFGCR0_DDI_CLK_OFF(phy
);
636 intel_de_write(dev_priv
, ICL_DPCLKA_CFGCR0
, tmp
);
637 mutex_unlock(&dev_priv
->dpll
.lock
);
640 static void gen11_dsi_ungate_clocks(struct intel_encoder
*encoder
)
642 struct drm_i915_private
*dev_priv
= to_i915(encoder
->base
.dev
);
643 struct intel_dsi
*intel_dsi
= enc_to_intel_dsi(encoder
);
647 mutex_lock(&dev_priv
->dpll
.lock
);
648 tmp
= intel_de_read(dev_priv
, ICL_DPCLKA_CFGCR0
);
649 for_each_dsi_phy(phy
, intel_dsi
->phys
)
650 tmp
&= ~ICL_DPCLKA_CFGCR0_DDI_CLK_OFF(phy
);
652 intel_de_write(dev_priv
, ICL_DPCLKA_CFGCR0
, tmp
);
653 mutex_unlock(&dev_priv
->dpll
.lock
);
656 static void gen11_dsi_map_pll(struct intel_encoder
*encoder
,
657 const struct intel_crtc_state
*crtc_state
)
659 struct drm_i915_private
*dev_priv
= to_i915(encoder
->base
.dev
);
660 struct intel_dsi
*intel_dsi
= enc_to_intel_dsi(encoder
);
661 struct intel_shared_dpll
*pll
= crtc_state
->shared_dpll
;
665 mutex_lock(&dev_priv
->dpll
.lock
);
667 val
= intel_de_read(dev_priv
, ICL_DPCLKA_CFGCR0
);
668 for_each_dsi_phy(phy
, intel_dsi
->phys
) {
669 val
&= ~ICL_DPCLKA_CFGCR0_DDI_CLK_SEL_MASK(phy
);
670 val
|= ICL_DPCLKA_CFGCR0_DDI_CLK_SEL(pll
->info
->id
, phy
);
672 intel_de_write(dev_priv
, ICL_DPCLKA_CFGCR0
, val
);
674 for_each_dsi_phy(phy
, intel_dsi
->phys
) {
675 if (INTEL_GEN(dev_priv
) >= 12)
676 val
|= ICL_DPCLKA_CFGCR0_DDI_CLK_OFF(phy
);
678 val
&= ~ICL_DPCLKA_CFGCR0_DDI_CLK_OFF(phy
);
680 intel_de_write(dev_priv
, ICL_DPCLKA_CFGCR0
, val
);
682 intel_de_posting_read(dev_priv
, ICL_DPCLKA_CFGCR0
);
684 mutex_unlock(&dev_priv
->dpll
.lock
);
688 gen11_dsi_configure_transcoder(struct intel_encoder
*encoder
,
689 const struct intel_crtc_state
*pipe_config
)
691 struct drm_i915_private
*dev_priv
= to_i915(encoder
->base
.dev
);
692 struct intel_dsi
*intel_dsi
= enc_to_intel_dsi(encoder
);
693 struct intel_crtc
*intel_crtc
= to_intel_crtc(pipe_config
->uapi
.crtc
);
694 enum pipe pipe
= intel_crtc
->pipe
;
697 enum transcoder dsi_trans
;
699 for_each_dsi_port(port
, intel_dsi
->ports
) {
700 dsi_trans
= dsi_port_to_transcoder(port
);
701 tmp
= intel_de_read(dev_priv
, DSI_TRANS_FUNC_CONF(dsi_trans
));
703 if (intel_dsi
->eotp_pkt
)
704 tmp
&= ~EOTP_DISABLED
;
706 tmp
|= EOTP_DISABLED
;
708 /* enable link calibration if freq > 1.5Gbps */
709 if (afe_clk(encoder
, pipe_config
) >= 1500 * 1000) {
710 tmp
&= ~LINK_CALIBRATION_MASK
;
711 tmp
|= CALIBRATION_ENABLED_INITIAL_ONLY
;
714 /* configure continuous clock */
715 tmp
&= ~CONTINUOUS_CLK_MASK
;
716 if (intel_dsi
->clock_stop
)
717 tmp
|= CLK_ENTER_LP_AFTER_DATA
;
719 tmp
|= CLK_HS_CONTINUOUS
;
721 /* configure buffer threshold limit to minimum */
722 tmp
&= ~PIX_BUF_THRESHOLD_MASK
;
723 tmp
|= PIX_BUF_THRESHOLD_1_4
;
725 /* set virtual channel to '0' */
726 tmp
&= ~PIX_VIRT_CHAN_MASK
;
727 tmp
|= PIX_VIRT_CHAN(0);
729 /* program BGR transmission */
730 if (intel_dsi
->bgr_enabled
)
731 tmp
|= BGR_TRANSMISSION
;
733 /* select pixel format */
734 tmp
&= ~PIX_FMT_MASK
;
735 if (pipe_config
->dsc
.compression_enable
) {
736 tmp
|= PIX_FMT_COMPRESSED
;
738 switch (intel_dsi
->pixel_format
) {
740 MISSING_CASE(intel_dsi
->pixel_format
);
742 case MIPI_DSI_FMT_RGB565
:
743 tmp
|= PIX_FMT_RGB565
;
745 case MIPI_DSI_FMT_RGB666_PACKED
:
746 tmp
|= PIX_FMT_RGB666_PACKED
;
748 case MIPI_DSI_FMT_RGB666
:
749 tmp
|= PIX_FMT_RGB666_LOOSE
;
751 case MIPI_DSI_FMT_RGB888
:
752 tmp
|= PIX_FMT_RGB888
;
757 if (INTEL_GEN(dev_priv
) >= 12) {
758 if (is_vid_mode(intel_dsi
))
759 tmp
|= BLANKING_PACKET_ENABLE
;
762 /* program DSI operation mode */
763 if (is_vid_mode(intel_dsi
)) {
764 tmp
&= ~OP_MODE_MASK
;
765 switch (intel_dsi
->video_mode_format
) {
767 MISSING_CASE(intel_dsi
->video_mode_format
);
769 case VIDEO_MODE_NON_BURST_WITH_SYNC_EVENTS
:
770 tmp
|= VIDEO_MODE_SYNC_EVENT
;
772 case VIDEO_MODE_NON_BURST_WITH_SYNC_PULSE
:
773 tmp
|= VIDEO_MODE_SYNC_PULSE
;
778 * FIXME: Retrieve this info from VBT.
779 * As per the spec when dsi transcoder is operating
780 * in TE GATE mode, TE comes from GPIO
781 * which is UTIL PIN for DSI 0.
782 * Also this GPIO would not be used for other
783 * purposes is an assumption.
785 tmp
&= ~OP_MODE_MASK
;
786 tmp
|= CMD_MODE_TE_GATE
;
787 tmp
|= TE_SOURCE_GPIO
;
790 intel_de_write(dev_priv
, DSI_TRANS_FUNC_CONF(dsi_trans
), tmp
);
793 /* enable port sync mode if dual link */
794 if (intel_dsi
->dual_link
) {
795 for_each_dsi_port(port
, intel_dsi
->ports
) {
796 dsi_trans
= dsi_port_to_transcoder(port
);
797 tmp
= intel_de_read(dev_priv
,
798 TRANS_DDI_FUNC_CTL2(dsi_trans
));
799 tmp
|= PORT_SYNC_MODE_ENABLE
;
800 intel_de_write(dev_priv
,
801 TRANS_DDI_FUNC_CTL2(dsi_trans
), tmp
);
804 /* configure stream splitting */
805 configure_dual_link_mode(encoder
, pipe_config
);
808 for_each_dsi_port(port
, intel_dsi
->ports
) {
809 dsi_trans
= dsi_port_to_transcoder(port
);
811 /* select data lane width */
812 tmp
= intel_de_read(dev_priv
, TRANS_DDI_FUNC_CTL(dsi_trans
));
813 tmp
&= ~DDI_PORT_WIDTH_MASK
;
814 tmp
|= DDI_PORT_WIDTH(intel_dsi
->lane_count
);
816 /* select input pipe */
817 tmp
&= ~TRANS_DDI_EDP_INPUT_MASK
;
823 tmp
|= TRANS_DDI_EDP_INPUT_A_ON
;
826 tmp
|= TRANS_DDI_EDP_INPUT_B_ONOFF
;
829 tmp
|= TRANS_DDI_EDP_INPUT_C_ONOFF
;
832 tmp
|= TRANS_DDI_EDP_INPUT_D_ONOFF
;
836 /* enable DDI buffer */
837 tmp
|= TRANS_DDI_FUNC_ENABLE
;
838 intel_de_write(dev_priv
, TRANS_DDI_FUNC_CTL(dsi_trans
), tmp
);
841 /* wait for link ready */
842 for_each_dsi_port(port
, intel_dsi
->ports
) {
843 dsi_trans
= dsi_port_to_transcoder(port
);
844 if (wait_for_us((intel_de_read(dev_priv
, DSI_TRANS_FUNC_CONF(dsi_trans
)) &
846 drm_err(&dev_priv
->drm
, "DSI link not ready\n");
851 gen11_dsi_set_transcoder_timings(struct intel_encoder
*encoder
,
852 const struct intel_crtc_state
*crtc_state
)
854 struct drm_i915_private
*dev_priv
= to_i915(encoder
->base
.dev
);
855 struct intel_dsi
*intel_dsi
= enc_to_intel_dsi(encoder
);
856 const struct drm_display_mode
*adjusted_mode
=
857 &crtc_state
->hw
.adjusted_mode
;
859 enum transcoder dsi_trans
;
860 /* horizontal timings */
861 u16 htotal
, hactive
, hsync_start
, hsync_end
, hsync_size
;
863 /* vertical timings */
864 u16 vtotal
, vactive
, vsync_start
, vsync_end
, vsync_shift
;
865 int mul
= 1, div
= 1;
868 * Adjust horizontal timings (htotal, hsync_start, hsync_end) to account
869 * for slower link speed if DSC is enabled.
871 * The compression frequency ratio is the ratio between compressed and
872 * non-compressed link speeds, and simplifies down to the ratio between
873 * compressed and non-compressed bpp.
875 if (crtc_state
->dsc
.compression_enable
) {
876 mul
= crtc_state
->dsc
.compressed_bpp
;
877 div
= mipi_dsi_pixel_format_to_bpp(intel_dsi
->pixel_format
);
880 hactive
= adjusted_mode
->crtc_hdisplay
;
882 if (is_vid_mode(intel_dsi
))
883 htotal
= DIV_ROUND_UP(adjusted_mode
->crtc_htotal
* mul
, div
);
885 htotal
= DIV_ROUND_UP((hactive
+ 160) * mul
, div
);
887 hsync_start
= DIV_ROUND_UP(adjusted_mode
->crtc_hsync_start
* mul
, div
);
888 hsync_end
= DIV_ROUND_UP(adjusted_mode
->crtc_hsync_end
* mul
, div
);
889 hsync_size
= hsync_end
- hsync_start
;
890 hback_porch
= (adjusted_mode
->crtc_htotal
-
891 adjusted_mode
->crtc_hsync_end
);
892 vactive
= adjusted_mode
->crtc_vdisplay
;
894 if (is_vid_mode(intel_dsi
)) {
895 vtotal
= adjusted_mode
->crtc_vtotal
;
897 int bpp
, line_time_us
, byte_clk_period_ns
;
899 if (crtc_state
->dsc
.compression_enable
)
900 bpp
= crtc_state
->dsc
.compressed_bpp
;
902 bpp
= mipi_dsi_pixel_format_to_bpp(intel_dsi
->pixel_format
);
904 byte_clk_period_ns
= 1000000 / afe_clk(encoder
, crtc_state
);
905 line_time_us
= (htotal
* (bpp
/ 8) * byte_clk_period_ns
) / (1000 * intel_dsi
->lane_count
);
906 vtotal
= vactive
+ DIV_ROUND_UP(400, line_time_us
);
908 vsync_start
= adjusted_mode
->crtc_vsync_start
;
909 vsync_end
= adjusted_mode
->crtc_vsync_end
;
910 vsync_shift
= hsync_start
- htotal
/ 2;
912 if (intel_dsi
->dual_link
) {
914 if (intel_dsi
->dual_link
== DSI_DUAL_LINK_FRONT_BACK
)
915 hactive
+= intel_dsi
->pixel_overlap
;
919 /* minimum hactive as per bspec: 256 pixels */
920 if (adjusted_mode
->crtc_hdisplay
< 256)
921 drm_err(&dev_priv
->drm
, "hactive is less then 256 pixels\n");
923 /* if RGB666 format, then hactive must be multiple of 4 pixels */
924 if (intel_dsi
->pixel_format
== MIPI_DSI_FMT_RGB666
&& hactive
% 4 != 0)
925 drm_err(&dev_priv
->drm
,
926 "hactive pixels are not multiple of 4\n");
928 /* program TRANS_HTOTAL register */
929 for_each_dsi_port(port
, intel_dsi
->ports
) {
930 dsi_trans
= dsi_port_to_transcoder(port
);
931 intel_de_write(dev_priv
, HTOTAL(dsi_trans
),
932 (hactive
- 1) | ((htotal
- 1) << 16));
935 /* TRANS_HSYNC register to be programmed only for video mode */
936 if (is_vid_mode(intel_dsi
)) {
937 if (intel_dsi
->video_mode_format
==
938 VIDEO_MODE_NON_BURST_WITH_SYNC_PULSE
) {
939 /* BSPEC: hsync size should be atleast 16 pixels */
941 drm_err(&dev_priv
->drm
,
942 "hsync size < 16 pixels\n");
945 if (hback_porch
< 16)
946 drm_err(&dev_priv
->drm
, "hback porch < 16 pixels\n");
948 if (intel_dsi
->dual_link
) {
953 for_each_dsi_port(port
, intel_dsi
->ports
) {
954 dsi_trans
= dsi_port_to_transcoder(port
);
955 intel_de_write(dev_priv
, HSYNC(dsi_trans
),
956 (hsync_start
- 1) | ((hsync_end
- 1) << 16));
960 /* program TRANS_VTOTAL register */
961 for_each_dsi_port(port
, intel_dsi
->ports
) {
962 dsi_trans
= dsi_port_to_transcoder(port
);
964 * FIXME: Programing this by assuming progressive mode, since
965 * non-interlaced info from VBT is not saved inside
966 * struct drm_display_mode.
967 * For interlace mode: program required pixel minus 2
969 intel_de_write(dev_priv
, VTOTAL(dsi_trans
),
970 (vactive
- 1) | ((vtotal
- 1) << 16));
973 if (vsync_end
< vsync_start
|| vsync_end
> vtotal
)
974 drm_err(&dev_priv
->drm
, "Invalid vsync_end value\n");
976 if (vsync_start
< vactive
)
977 drm_err(&dev_priv
->drm
, "vsync_start less than vactive\n");
979 /* program TRANS_VSYNC register for video mode only */
980 if (is_vid_mode(intel_dsi
)) {
981 for_each_dsi_port(port
, intel_dsi
->ports
) {
982 dsi_trans
= dsi_port_to_transcoder(port
);
983 intel_de_write(dev_priv
, VSYNC(dsi_trans
),
984 (vsync_start
- 1) | ((vsync_end
- 1) << 16));
989 * FIXME: It has to be programmed only for video modes and interlaced
990 * modes. Put the check condition here once interlaced
991 * info available as described above.
992 * program TRANS_VSYNCSHIFT register
994 if (is_vid_mode(intel_dsi
)) {
995 for_each_dsi_port(port
, intel_dsi
->ports
) {
996 dsi_trans
= dsi_port_to_transcoder(port
);
997 intel_de_write(dev_priv
, VSYNCSHIFT(dsi_trans
),
1002 /* program TRANS_VBLANK register, should be same as vtotal programmed */
1003 if (INTEL_GEN(dev_priv
) >= 12) {
1004 for_each_dsi_port(port
, intel_dsi
->ports
) {
1005 dsi_trans
= dsi_port_to_transcoder(port
);
1006 intel_de_write(dev_priv
, VBLANK(dsi_trans
),
1007 (vactive
- 1) | ((vtotal
- 1) << 16));
1012 static void gen11_dsi_enable_transcoder(struct intel_encoder
*encoder
)
1014 struct drm_i915_private
*dev_priv
= to_i915(encoder
->base
.dev
);
1015 struct intel_dsi
*intel_dsi
= enc_to_intel_dsi(encoder
);
1017 enum transcoder dsi_trans
;
1020 for_each_dsi_port(port
, intel_dsi
->ports
) {
1021 dsi_trans
= dsi_port_to_transcoder(port
);
1022 tmp
= intel_de_read(dev_priv
, PIPECONF(dsi_trans
));
1023 tmp
|= PIPECONF_ENABLE
;
1024 intel_de_write(dev_priv
, PIPECONF(dsi_trans
), tmp
);
1026 /* wait for transcoder to be enabled */
1027 if (intel_de_wait_for_set(dev_priv
, PIPECONF(dsi_trans
),
1028 I965_PIPECONF_ACTIVE
, 10))
1029 drm_err(&dev_priv
->drm
,
1030 "DSI transcoder not enabled\n");
1034 static void gen11_dsi_setup_timeouts(struct intel_encoder
*encoder
,
1035 const struct intel_crtc_state
*crtc_state
)
1037 struct drm_i915_private
*dev_priv
= to_i915(encoder
->base
.dev
);
1038 struct intel_dsi
*intel_dsi
= enc_to_intel_dsi(encoder
);
1040 enum transcoder dsi_trans
;
1041 u32 tmp
, hs_tx_timeout
, lp_rx_timeout
, ta_timeout
, divisor
, mul
;
1044 * escape clock count calculation:
1045 * BYTE_CLK_COUNT = TIME_NS/(8 * UI)
1046 * UI (nsec) = (10^6)/Bitrate
1047 * TIME_NS = (BYTE_CLK_COUNT * 8 * 10^6)/ Bitrate
1048 * ESCAPE_CLK_COUNT = TIME_NS/ESC_CLK_NS
1050 divisor
= intel_dsi_tlpx_ns(intel_dsi
) * afe_clk(encoder
, crtc_state
) * 1000;
1052 hs_tx_timeout
= DIV_ROUND_UP(intel_dsi
->hs_tx_timeout
* mul
,
1054 lp_rx_timeout
= DIV_ROUND_UP(intel_dsi
->lp_rx_timeout
* mul
, divisor
);
1055 ta_timeout
= DIV_ROUND_UP(intel_dsi
->turn_arnd_val
* mul
, divisor
);
1057 for_each_dsi_port(port
, intel_dsi
->ports
) {
1058 dsi_trans
= dsi_port_to_transcoder(port
);
1060 /* program hst_tx_timeout */
1061 tmp
= intel_de_read(dev_priv
, DSI_HSTX_TO(dsi_trans
));
1062 tmp
&= ~HSTX_TIMEOUT_VALUE_MASK
;
1063 tmp
|= HSTX_TIMEOUT_VALUE(hs_tx_timeout
);
1064 intel_de_write(dev_priv
, DSI_HSTX_TO(dsi_trans
), tmp
);
1066 /* FIXME: DSI_CALIB_TO */
1068 /* program lp_rx_host timeout */
1069 tmp
= intel_de_read(dev_priv
, DSI_LPRX_HOST_TO(dsi_trans
));
1070 tmp
&= ~LPRX_TIMEOUT_VALUE_MASK
;
1071 tmp
|= LPRX_TIMEOUT_VALUE(lp_rx_timeout
);
1072 intel_de_write(dev_priv
, DSI_LPRX_HOST_TO(dsi_trans
), tmp
);
1074 /* FIXME: DSI_PWAIT_TO */
1076 /* program turn around timeout */
1077 tmp
= intel_de_read(dev_priv
, DSI_TA_TO(dsi_trans
));
1078 tmp
&= ~TA_TIMEOUT_VALUE_MASK
;
1079 tmp
|= TA_TIMEOUT_VALUE(ta_timeout
);
1080 intel_de_write(dev_priv
, DSI_TA_TO(dsi_trans
), tmp
);
1084 static void gen11_dsi_config_util_pin(struct intel_encoder
*encoder
,
1087 struct drm_i915_private
*dev_priv
= to_i915(encoder
->base
.dev
);
1088 struct intel_dsi
*intel_dsi
= enc_to_intel_dsi(encoder
);
1092 * used as TE i/p for DSI0,
1093 * for dual link/DSI1 TE is from slave DSI1
1096 if (is_vid_mode(intel_dsi
) || (intel_dsi
->ports
& BIT(PORT_B
)))
1099 tmp
= intel_de_read(dev_priv
, UTIL_PIN_CTL
);
1102 tmp
|= UTIL_PIN_DIRECTION_INPUT
;
1103 tmp
|= UTIL_PIN_ENABLE
;
1105 tmp
&= ~UTIL_PIN_ENABLE
;
1107 intel_de_write(dev_priv
, UTIL_PIN_CTL
, tmp
);
1111 gen11_dsi_enable_port_and_phy(struct intel_encoder
*encoder
,
1112 const struct intel_crtc_state
*crtc_state
)
1114 struct drm_i915_private
*dev_priv
= to_i915(encoder
->base
.dev
);
1116 /* step 4a: power up all lanes of the DDI used by DSI */
1117 gen11_dsi_power_up_lanes(encoder
);
1119 /* step 4b: configure lane sequencing of the Combo-PHY transmitters */
1120 gen11_dsi_config_phy_lanes_sequence(encoder
);
1122 /* step 4c: configure voltage swing and skew */
1123 gen11_dsi_voltage_swing_program_seq(encoder
);
1125 /* enable DDI buffer */
1126 gen11_dsi_enable_ddi_buffer(encoder
);
1128 /* setup D-PHY timings */
1129 gen11_dsi_setup_dphy_timings(encoder
, crtc_state
);
1131 /* Since transcoder is configured to take events from GPIO */
1132 gen11_dsi_config_util_pin(encoder
, true);
1134 /* step 4h: setup DSI protocol timeouts */
1135 gen11_dsi_setup_timeouts(encoder
, crtc_state
);
1137 /* Step (4h, 4i, 4j, 4k): Configure transcoder */
1138 gen11_dsi_configure_transcoder(encoder
, crtc_state
);
1140 /* Step 4l: Gate DDI clocks */
1141 if (IS_GEN(dev_priv
, 11))
1142 gen11_dsi_gate_clocks(encoder
);
1145 static void gen11_dsi_powerup_panel(struct intel_encoder
*encoder
)
1147 struct drm_i915_private
*dev_priv
= to_i915(encoder
->base
.dev
);
1148 struct intel_dsi
*intel_dsi
= enc_to_intel_dsi(encoder
);
1149 struct mipi_dsi_device
*dsi
;
1151 enum transcoder dsi_trans
;
1155 /* set maximum return packet size */
1156 for_each_dsi_port(port
, intel_dsi
->ports
) {
1157 dsi_trans
= dsi_port_to_transcoder(port
);
1160 * FIXME: This uses the number of DW's currently in the payload
1161 * receive queue. This is probably not what we want here.
1163 tmp
= intel_de_read(dev_priv
, DSI_CMD_RXCTL(dsi_trans
));
1164 tmp
&= NUMBER_RX_PLOAD_DW_MASK
;
1165 /* multiply "Number Rx Payload DW" by 4 to get max value */
1167 dsi
= intel_dsi
->dsi_hosts
[port
]->device
;
1168 ret
= mipi_dsi_set_maximum_return_packet_size(dsi
, tmp
);
1170 drm_err(&dev_priv
->drm
,
1171 "error setting max return pkt size%d\n", tmp
);
1174 /* panel power on related mipi dsi vbt sequences */
1175 intel_dsi_vbt_exec_sequence(intel_dsi
, MIPI_SEQ_POWER_ON
);
1176 intel_dsi_msleep(intel_dsi
, intel_dsi
->panel_on_delay
);
1177 intel_dsi_vbt_exec_sequence(intel_dsi
, MIPI_SEQ_DEASSERT_RESET
);
1178 intel_dsi_vbt_exec_sequence(intel_dsi
, MIPI_SEQ_INIT_OTP
);
1179 intel_dsi_vbt_exec_sequence(intel_dsi
, MIPI_SEQ_DISPLAY_ON
);
1181 /* ensure all panel commands dispatched before enabling transcoder */
1182 wait_for_cmds_dispatched_to_panel(encoder
);
1185 static void gen11_dsi_pre_pll_enable(struct intel_atomic_state
*state
,
1186 struct intel_encoder
*encoder
,
1187 const struct intel_crtc_state
*crtc_state
,
1188 const struct drm_connector_state
*conn_state
)
1190 /* step2: enable IO power */
1191 gen11_dsi_enable_io_power(encoder
);
1193 /* step3: enable DSI PLL */
1194 gen11_dsi_program_esc_clk_div(encoder
, crtc_state
);
1197 static void gen11_dsi_pre_enable(struct intel_atomic_state
*state
,
1198 struct intel_encoder
*encoder
,
1199 const struct intel_crtc_state
*pipe_config
,
1200 const struct drm_connector_state
*conn_state
)
1203 gen11_dsi_map_pll(encoder
, pipe_config
);
1205 /* step4: enable DSI port and DPHY */
1206 gen11_dsi_enable_port_and_phy(encoder
, pipe_config
);
1208 /* step5: program and powerup panel */
1209 gen11_dsi_powerup_panel(encoder
);
1211 intel_dsc_enable(encoder
, pipe_config
);
1213 /* step6c: configure transcoder timings */
1214 gen11_dsi_set_transcoder_timings(encoder
, pipe_config
);
1217 static void gen11_dsi_enable(struct intel_atomic_state
*state
,
1218 struct intel_encoder
*encoder
,
1219 const struct intel_crtc_state
*crtc_state
,
1220 const struct drm_connector_state
*conn_state
)
1222 struct intel_dsi
*intel_dsi
= enc_to_intel_dsi(encoder
);
1224 drm_WARN_ON(state
->base
.dev
, crtc_state
->has_pch_encoder
);
1226 /* step6d: enable dsi transcoder */
1227 gen11_dsi_enable_transcoder(encoder
);
1229 /* step7: enable backlight */
1230 intel_panel_enable_backlight(crtc_state
, conn_state
);
1231 intel_dsi_vbt_exec_sequence(intel_dsi
, MIPI_SEQ_BACKLIGHT_ON
);
1233 intel_crtc_vblank_on(crtc_state
);
1236 static void gen11_dsi_disable_transcoder(struct intel_encoder
*encoder
)
1238 struct drm_i915_private
*dev_priv
= to_i915(encoder
->base
.dev
);
1239 struct intel_dsi
*intel_dsi
= enc_to_intel_dsi(encoder
);
1241 enum transcoder dsi_trans
;
1244 for_each_dsi_port(port
, intel_dsi
->ports
) {
1245 dsi_trans
= dsi_port_to_transcoder(port
);
1247 /* disable transcoder */
1248 tmp
= intel_de_read(dev_priv
, PIPECONF(dsi_trans
));
1249 tmp
&= ~PIPECONF_ENABLE
;
1250 intel_de_write(dev_priv
, PIPECONF(dsi_trans
), tmp
);
1252 /* wait for transcoder to be disabled */
1253 if (intel_de_wait_for_clear(dev_priv
, PIPECONF(dsi_trans
),
1254 I965_PIPECONF_ACTIVE
, 50))
1255 drm_err(&dev_priv
->drm
,
1256 "DSI trancoder not disabled\n");
1260 static void gen11_dsi_powerdown_panel(struct intel_encoder
*encoder
)
1262 struct intel_dsi
*intel_dsi
= enc_to_intel_dsi(encoder
);
1264 intel_dsi_vbt_exec_sequence(intel_dsi
, MIPI_SEQ_DISPLAY_OFF
);
1265 intel_dsi_vbt_exec_sequence(intel_dsi
, MIPI_SEQ_ASSERT_RESET
);
1266 intel_dsi_vbt_exec_sequence(intel_dsi
, MIPI_SEQ_POWER_OFF
);
1268 /* ensure cmds dispatched to panel */
1269 wait_for_cmds_dispatched_to_panel(encoder
);
1272 static void gen11_dsi_deconfigure_trancoder(struct intel_encoder
*encoder
)
1274 struct drm_i915_private
*dev_priv
= to_i915(encoder
->base
.dev
);
1275 struct intel_dsi
*intel_dsi
= enc_to_intel_dsi(encoder
);
1277 enum transcoder dsi_trans
;
1280 /* disable periodic update mode */
1281 if (is_cmd_mode(intel_dsi
)) {
1282 for_each_dsi_port(port
, intel_dsi
->ports
) {
1283 tmp
= intel_de_read(dev_priv
, DSI_CMD_FRMCTL(port
));
1284 tmp
&= ~DSI_PERIODIC_FRAME_UPDATE_ENABLE
;
1285 intel_de_write(dev_priv
, DSI_CMD_FRMCTL(port
), tmp
);
1289 /* put dsi link in ULPS */
1290 for_each_dsi_port(port
, intel_dsi
->ports
) {
1291 dsi_trans
= dsi_port_to_transcoder(port
);
1292 tmp
= intel_de_read(dev_priv
, DSI_LP_MSG(dsi_trans
));
1293 tmp
|= LINK_ENTER_ULPS
;
1294 tmp
&= ~LINK_ULPS_TYPE_LP11
;
1295 intel_de_write(dev_priv
, DSI_LP_MSG(dsi_trans
), tmp
);
1297 if (wait_for_us((intel_de_read(dev_priv
, DSI_LP_MSG(dsi_trans
)) &
1300 drm_err(&dev_priv
->drm
, "DSI link not in ULPS\n");
1303 /* disable ddi function */
1304 for_each_dsi_port(port
, intel_dsi
->ports
) {
1305 dsi_trans
= dsi_port_to_transcoder(port
);
1306 tmp
= intel_de_read(dev_priv
, TRANS_DDI_FUNC_CTL(dsi_trans
));
1307 tmp
&= ~TRANS_DDI_FUNC_ENABLE
;
1308 intel_de_write(dev_priv
, TRANS_DDI_FUNC_CTL(dsi_trans
), tmp
);
1311 /* disable port sync mode if dual link */
1312 if (intel_dsi
->dual_link
) {
1313 for_each_dsi_port(port
, intel_dsi
->ports
) {
1314 dsi_trans
= dsi_port_to_transcoder(port
);
1315 tmp
= intel_de_read(dev_priv
,
1316 TRANS_DDI_FUNC_CTL2(dsi_trans
));
1317 tmp
&= ~PORT_SYNC_MODE_ENABLE
;
1318 intel_de_write(dev_priv
,
1319 TRANS_DDI_FUNC_CTL2(dsi_trans
), tmp
);
1324 static void gen11_dsi_disable_port(struct intel_encoder
*encoder
)
1326 struct drm_i915_private
*dev_priv
= to_i915(encoder
->base
.dev
);
1327 struct intel_dsi
*intel_dsi
= enc_to_intel_dsi(encoder
);
1331 gen11_dsi_ungate_clocks(encoder
);
1332 for_each_dsi_port(port
, intel_dsi
->ports
) {
1333 tmp
= intel_de_read(dev_priv
, DDI_BUF_CTL(port
));
1334 tmp
&= ~DDI_BUF_CTL_ENABLE
;
1335 intel_de_write(dev_priv
, DDI_BUF_CTL(port
), tmp
);
1337 if (wait_for_us((intel_de_read(dev_priv
, DDI_BUF_CTL(port
)) &
1340 drm_err(&dev_priv
->drm
,
1341 "DDI port:%c buffer not idle\n",
1344 gen11_dsi_gate_clocks(encoder
);
1347 static void gen11_dsi_disable_io_power(struct intel_encoder
*encoder
)
1349 struct drm_i915_private
*dev_priv
= to_i915(encoder
->base
.dev
);
1350 struct intel_dsi
*intel_dsi
= enc_to_intel_dsi(encoder
);
1354 for_each_dsi_port(port
, intel_dsi
->ports
) {
1355 intel_wakeref_t wakeref
;
1357 wakeref
= fetch_and_zero(&intel_dsi
->io_wakeref
[port
]);
1358 intel_display_power_put(dev_priv
,
1360 POWER_DOMAIN_PORT_DDI_A_IO
:
1361 POWER_DOMAIN_PORT_DDI_B_IO
,
1365 /* set mode to DDI */
1366 for_each_dsi_port(port
, intel_dsi
->ports
) {
1367 tmp
= intel_de_read(dev_priv
, ICL_DSI_IO_MODECTL(port
));
1368 tmp
&= ~COMBO_PHY_MODE_DSI
;
1369 intel_de_write(dev_priv
, ICL_DSI_IO_MODECTL(port
), tmp
);
1373 static void gen11_dsi_disable(struct intel_atomic_state
*state
,
1374 struct intel_encoder
*encoder
,
1375 const struct intel_crtc_state
*old_crtc_state
,
1376 const struct drm_connector_state
*old_conn_state
)
1378 struct intel_dsi
*intel_dsi
= enc_to_intel_dsi(encoder
);
1380 /* step1: turn off backlight */
1381 intel_dsi_vbt_exec_sequence(intel_dsi
, MIPI_SEQ_BACKLIGHT_OFF
);
1382 intel_panel_disable_backlight(old_conn_state
);
1384 /* step2d,e: disable transcoder and wait */
1385 gen11_dsi_disable_transcoder(encoder
);
1387 /* step2f,g: powerdown panel */
1388 gen11_dsi_powerdown_panel(encoder
);
1390 /* step2h,i,j: deconfig trancoder */
1391 gen11_dsi_deconfigure_trancoder(encoder
);
1393 /* step3: disable port */
1394 gen11_dsi_disable_port(encoder
);
1396 gen11_dsi_config_util_pin(encoder
, false);
1398 /* step4: disable IO power */
1399 gen11_dsi_disable_io_power(encoder
);
1402 static void gen11_dsi_post_disable(struct intel_atomic_state
*state
,
1403 struct intel_encoder
*encoder
,
1404 const struct intel_crtc_state
*old_crtc_state
,
1405 const struct drm_connector_state
*old_conn_state
)
1407 intel_crtc_vblank_off(old_crtc_state
);
1409 intel_dsc_disable(old_crtc_state
);
1411 skl_scaler_disable(old_crtc_state
);
1414 static enum drm_mode_status
gen11_dsi_mode_valid(struct drm_connector
*connector
,
1415 struct drm_display_mode
*mode
)
1418 return intel_dsi_mode_valid(connector
, mode
);
1421 static void gen11_dsi_get_timings(struct intel_encoder
*encoder
,
1422 struct intel_crtc_state
*pipe_config
)
1424 struct intel_dsi
*intel_dsi
= enc_to_intel_dsi(encoder
);
1425 struct drm_display_mode
*adjusted_mode
=
1426 &pipe_config
->hw
.adjusted_mode
;
1428 if (pipe_config
->dsc
.compressed_bpp
) {
1429 int div
= pipe_config
->dsc
.compressed_bpp
;
1430 int mul
= mipi_dsi_pixel_format_to_bpp(intel_dsi
->pixel_format
);
1432 adjusted_mode
->crtc_htotal
=
1433 DIV_ROUND_UP(adjusted_mode
->crtc_htotal
* mul
, div
);
1434 adjusted_mode
->crtc_hsync_start
=
1435 DIV_ROUND_UP(adjusted_mode
->crtc_hsync_start
* mul
, div
);
1436 adjusted_mode
->crtc_hsync_end
=
1437 DIV_ROUND_UP(adjusted_mode
->crtc_hsync_end
* mul
, div
);
1440 if (intel_dsi
->dual_link
) {
1441 adjusted_mode
->crtc_hdisplay
*= 2;
1442 if (intel_dsi
->dual_link
== DSI_DUAL_LINK_FRONT_BACK
)
1443 adjusted_mode
->crtc_hdisplay
-=
1444 intel_dsi
->pixel_overlap
;
1445 adjusted_mode
->crtc_htotal
*= 2;
1447 adjusted_mode
->crtc_hblank_start
= adjusted_mode
->crtc_hdisplay
;
1448 adjusted_mode
->crtc_hblank_end
= adjusted_mode
->crtc_htotal
;
1450 if (intel_dsi
->operation_mode
== INTEL_DSI_VIDEO_MODE
) {
1451 if (intel_dsi
->dual_link
) {
1452 adjusted_mode
->crtc_hsync_start
*= 2;
1453 adjusted_mode
->crtc_hsync_end
*= 2;
1456 adjusted_mode
->crtc_vblank_start
= adjusted_mode
->crtc_vdisplay
;
1457 adjusted_mode
->crtc_vblank_end
= adjusted_mode
->crtc_vtotal
;
1460 static bool gen11_dsi_is_periodic_cmd_mode(struct intel_dsi
*intel_dsi
)
1462 struct drm_device
*dev
= intel_dsi
->base
.base
.dev
;
1463 struct drm_i915_private
*dev_priv
= to_i915(dev
);
1464 enum transcoder dsi_trans
;
1467 if (intel_dsi
->ports
== BIT(PORT_B
))
1468 dsi_trans
= TRANSCODER_DSI_1
;
1470 dsi_trans
= TRANSCODER_DSI_0
;
1472 val
= intel_de_read(dev_priv
, DSI_TRANS_FUNC_CONF(dsi_trans
));
1473 return (val
& DSI_PERIODIC_FRAME_UPDATE_ENABLE
);
1476 static void gen11_dsi_get_cmd_mode_config(struct intel_dsi
*intel_dsi
,
1477 struct intel_crtc_state
*pipe_config
)
1479 if (intel_dsi
->ports
== (BIT(PORT_B
) | BIT(PORT_A
)))
1480 pipe_config
->mode_flags
|= I915_MODE_FLAG_DSI_USE_TE1
|
1481 I915_MODE_FLAG_DSI_USE_TE0
;
1482 else if (intel_dsi
->ports
== BIT(PORT_B
))
1483 pipe_config
->mode_flags
|= I915_MODE_FLAG_DSI_USE_TE1
;
1485 pipe_config
->mode_flags
|= I915_MODE_FLAG_DSI_USE_TE0
;
1488 static void gen11_dsi_get_config(struct intel_encoder
*encoder
,
1489 struct intel_crtc_state
*pipe_config
)
1491 struct drm_i915_private
*i915
= to_i915(encoder
->base
.dev
);
1492 struct intel_crtc
*crtc
= to_intel_crtc(pipe_config
->uapi
.crtc
);
1493 struct intel_dsi
*intel_dsi
= enc_to_intel_dsi(encoder
);
1495 /* FIXME: adapt icl_ddi_clock_get() for DSI and use that? */
1496 pipe_config
->port_clock
= intel_dpll_get_freq(i915
,
1497 pipe_config
->shared_dpll
,
1498 &pipe_config
->dpll_hw_state
);
1500 pipe_config
->hw
.adjusted_mode
.crtc_clock
= intel_dsi
->pclk
;
1501 if (intel_dsi
->dual_link
)
1502 pipe_config
->hw
.adjusted_mode
.crtc_clock
*= 2;
1504 gen11_dsi_get_timings(encoder
, pipe_config
);
1505 pipe_config
->output_types
|= BIT(INTEL_OUTPUT_DSI
);
1506 pipe_config
->pipe_bpp
= bdw_get_pipemisc_bpp(crtc
);
1508 /* Get the details on which TE should be enabled */
1509 if (is_cmd_mode(intel_dsi
))
1510 gen11_dsi_get_cmd_mode_config(intel_dsi
, pipe_config
);
1512 if (gen11_dsi_is_periodic_cmd_mode(intel_dsi
))
1513 pipe_config
->mode_flags
|= I915_MODE_FLAG_DSI_PERIODIC_CMD_MODE
;
1516 static int gen11_dsi_dsc_compute_config(struct intel_encoder
*encoder
,
1517 struct intel_crtc_state
*crtc_state
)
1519 struct drm_i915_private
*dev_priv
= to_i915(encoder
->base
.dev
);
1520 struct drm_dsc_config
*vdsc_cfg
= &crtc_state
->dsc
.config
;
1521 int dsc_max_bpc
= INTEL_GEN(dev_priv
) >= 12 ? 12 : 10;
1525 use_dsc
= intel_bios_get_dsc_params(encoder
, crtc_state
, dsc_max_bpc
);
1529 if (crtc_state
->pipe_bpp
< 8 * 3)
1532 /* FIXME: split only when necessary */
1533 if (crtc_state
->dsc
.slice_count
> 1)
1534 crtc_state
->dsc
.dsc_split
= true;
1536 vdsc_cfg
->convert_rgb
= true;
1538 ret
= intel_dsc_compute_params(encoder
, crtc_state
);
1542 /* DSI specific sanity checks on the common code */
1543 drm_WARN_ON(&dev_priv
->drm
, vdsc_cfg
->vbr_enable
);
1544 drm_WARN_ON(&dev_priv
->drm
, vdsc_cfg
->simple_422
);
1545 drm_WARN_ON(&dev_priv
->drm
,
1546 vdsc_cfg
->pic_width
% vdsc_cfg
->slice_width
);
1547 drm_WARN_ON(&dev_priv
->drm
, vdsc_cfg
->slice_height
< 8);
1548 drm_WARN_ON(&dev_priv
->drm
,
1549 vdsc_cfg
->pic_height
% vdsc_cfg
->slice_height
);
1551 ret
= drm_dsc_compute_rc_parameters(vdsc_cfg
);
1555 crtc_state
->dsc
.compression_enable
= true;
1560 static int gen11_dsi_compute_config(struct intel_encoder
*encoder
,
1561 struct intel_crtc_state
*pipe_config
,
1562 struct drm_connector_state
*conn_state
)
1564 struct drm_i915_private
*i915
= to_i915(encoder
->base
.dev
);
1565 struct intel_dsi
*intel_dsi
= container_of(encoder
, struct intel_dsi
,
1567 struct intel_connector
*intel_connector
= intel_dsi
->attached_connector
;
1568 const struct drm_display_mode
*fixed_mode
=
1569 intel_connector
->panel
.fixed_mode
;
1570 struct drm_display_mode
*adjusted_mode
=
1571 &pipe_config
->hw
.adjusted_mode
;
1574 pipe_config
->output_format
= INTEL_OUTPUT_FORMAT_RGB
;
1575 intel_fixed_panel_mode(fixed_mode
, adjusted_mode
);
1577 ret
= intel_pch_panel_fitting(pipe_config
, conn_state
);
1581 adjusted_mode
->flags
= 0;
1583 /* Dual link goes to trancoder DSI'0' */
1584 if (intel_dsi
->ports
== BIT(PORT_B
))
1585 pipe_config
->cpu_transcoder
= TRANSCODER_DSI_1
;
1587 pipe_config
->cpu_transcoder
= TRANSCODER_DSI_0
;
1589 if (intel_dsi
->pixel_format
== MIPI_DSI_FMT_RGB888
)
1590 pipe_config
->pipe_bpp
= 24;
1592 pipe_config
->pipe_bpp
= 18;
1594 pipe_config
->clock_set
= true;
1596 if (gen11_dsi_dsc_compute_config(encoder
, pipe_config
))
1597 drm_dbg_kms(&i915
->drm
, "Attempting to use DSC failed\n");
1599 pipe_config
->port_clock
= afe_clk(encoder
, pipe_config
) / 5;
1602 * In case of TE GATE cmd mode, we
1603 * receive TE from the slave if
1604 * dual link is enabled
1606 if (is_cmd_mode(intel_dsi
))
1607 gen11_dsi_get_cmd_mode_config(intel_dsi
, pipe_config
);
1612 static void gen11_dsi_get_power_domains(struct intel_encoder
*encoder
,
1613 struct intel_crtc_state
*crtc_state
)
1615 struct drm_i915_private
*i915
= to_i915(encoder
->base
.dev
);
1617 get_dsi_io_power_domains(i915
,
1618 enc_to_intel_dsi(encoder
));
1621 static bool gen11_dsi_get_hw_state(struct intel_encoder
*encoder
,
1624 struct drm_i915_private
*dev_priv
= to_i915(encoder
->base
.dev
);
1625 struct intel_dsi
*intel_dsi
= enc_to_intel_dsi(encoder
);
1626 enum transcoder dsi_trans
;
1627 intel_wakeref_t wakeref
;
1632 wakeref
= intel_display_power_get_if_enabled(dev_priv
,
1633 encoder
->power_domain
);
1637 for_each_dsi_port(port
, intel_dsi
->ports
) {
1638 dsi_trans
= dsi_port_to_transcoder(port
);
1639 tmp
= intel_de_read(dev_priv
, TRANS_DDI_FUNC_CTL(dsi_trans
));
1640 switch (tmp
& TRANS_DDI_EDP_INPUT_MASK
) {
1641 case TRANS_DDI_EDP_INPUT_A_ON
:
1644 case TRANS_DDI_EDP_INPUT_B_ONOFF
:
1647 case TRANS_DDI_EDP_INPUT_C_ONOFF
:
1650 case TRANS_DDI_EDP_INPUT_D_ONOFF
:
1654 drm_err(&dev_priv
->drm
, "Invalid PIPE input\n");
1658 tmp
= intel_de_read(dev_priv
, PIPECONF(dsi_trans
));
1659 ret
= tmp
& PIPECONF_ENABLE
;
1662 intel_display_power_put(dev_priv
, encoder
->power_domain
, wakeref
);
1666 static bool gen11_dsi_initial_fastset_check(struct intel_encoder
*encoder
,
1667 struct intel_crtc_state
*crtc_state
)
1669 if (crtc_state
->dsc
.compression_enable
) {
1670 drm_dbg_kms(encoder
->base
.dev
, "Forcing full modeset due to DSC being enabled\n");
1671 crtc_state
->uapi
.mode_changed
= true;
1679 static void gen11_dsi_encoder_destroy(struct drm_encoder
*encoder
)
1681 intel_encoder_destroy(encoder
);
1684 static const struct drm_encoder_funcs gen11_dsi_encoder_funcs
= {
1685 .destroy
= gen11_dsi_encoder_destroy
,
1688 static const struct drm_connector_funcs gen11_dsi_connector_funcs
= {
1689 .detect
= intel_panel_detect
,
1690 .late_register
= intel_connector_register
,
1691 .early_unregister
= intel_connector_unregister
,
1692 .destroy
= intel_connector_destroy
,
1693 .fill_modes
= drm_helper_probe_single_connector_modes
,
1694 .atomic_get_property
= intel_digital_connector_atomic_get_property
,
1695 .atomic_set_property
= intel_digital_connector_atomic_set_property
,
1696 .atomic_destroy_state
= drm_atomic_helper_connector_destroy_state
,
1697 .atomic_duplicate_state
= intel_digital_connector_duplicate_state
,
1700 static const struct drm_connector_helper_funcs gen11_dsi_connector_helper_funcs
= {
1701 .get_modes
= intel_dsi_get_modes
,
1702 .mode_valid
= gen11_dsi_mode_valid
,
1703 .atomic_check
= intel_digital_connector_atomic_check
,
1706 static int gen11_dsi_host_attach(struct mipi_dsi_host
*host
,
1707 struct mipi_dsi_device
*dsi
)
1712 static int gen11_dsi_host_detach(struct mipi_dsi_host
*host
,
1713 struct mipi_dsi_device
*dsi
)
1718 static ssize_t
gen11_dsi_host_transfer(struct mipi_dsi_host
*host
,
1719 const struct mipi_dsi_msg
*msg
)
1721 struct intel_dsi_host
*intel_dsi_host
= to_intel_dsi_host(host
);
1722 struct mipi_dsi_packet dsi_pkt
;
1724 bool enable_lpdt
= false;
1726 ret
= mipi_dsi_create_packet(&dsi_pkt
, msg
);
1730 if (msg
->flags
& MIPI_DSI_MSG_USE_LPM
)
1733 /* send packet header */
1734 ret
= dsi_send_pkt_hdr(intel_dsi_host
, dsi_pkt
, enable_lpdt
);
1738 /* only long packet contains payload */
1739 if (mipi_dsi_packet_format_is_long(msg
->type
)) {
1740 ret
= dsi_send_pkt_payld(intel_dsi_host
, dsi_pkt
);
1745 //TODO: add payload receive code if needed
1747 ret
= sizeof(dsi_pkt
.header
) + dsi_pkt
.payload_length
;
1752 static const struct mipi_dsi_host_ops gen11_dsi_host_ops
= {
1753 .attach
= gen11_dsi_host_attach
,
1754 .detach
= gen11_dsi_host_detach
,
1755 .transfer
= gen11_dsi_host_transfer
,
1758 #define ICL_PREPARE_CNT_MAX 0x7
1759 #define ICL_CLK_ZERO_CNT_MAX 0xf
1760 #define ICL_TRAIL_CNT_MAX 0x7
1761 #define ICL_TCLK_PRE_CNT_MAX 0x3
1762 #define ICL_TCLK_POST_CNT_MAX 0x7
1763 #define ICL_HS_ZERO_CNT_MAX 0xf
1764 #define ICL_EXIT_ZERO_CNT_MAX 0x7
1766 static void icl_dphy_param_init(struct intel_dsi
*intel_dsi
)
1768 struct drm_device
*dev
= intel_dsi
->base
.base
.dev
;
1769 struct drm_i915_private
*dev_priv
= to_i915(dev
);
1770 struct mipi_config
*mipi_config
= dev_priv
->vbt
.dsi
.config
;
1772 u32 prepare_cnt
, exit_zero_cnt
, clk_zero_cnt
, trail_cnt
;
1773 u32 ths_prepare_ns
, tclk_trail_ns
;
1775 u32 tclk_pre_cnt
, tclk_post_cnt
;
1777 tlpx_ns
= intel_dsi_tlpx_ns(intel_dsi
);
1779 tclk_trail_ns
= max(mipi_config
->tclk_trail
, mipi_config
->ths_trail
);
1780 ths_prepare_ns
= max(mipi_config
->ths_prepare
,
1781 mipi_config
->tclk_prepare
);
1784 * prepare cnt in escape clocks
1785 * this field represents a hexadecimal value with a precision
1786 * of 1.2 – i.e. the most significant bit is the integer
1787 * and the least significant 2 bits are fraction bits.
1788 * so, the field can represent a range of 0.25 to 1.75
1790 prepare_cnt
= DIV_ROUND_UP(ths_prepare_ns
* 4, tlpx_ns
);
1791 if (prepare_cnt
> ICL_PREPARE_CNT_MAX
) {
1792 drm_dbg_kms(&dev_priv
->drm
, "prepare_cnt out of range (%d)\n",
1794 prepare_cnt
= ICL_PREPARE_CNT_MAX
;
1797 /* clk zero count in escape clocks */
1798 clk_zero_cnt
= DIV_ROUND_UP(mipi_config
->tclk_prepare_clkzero
-
1799 ths_prepare_ns
, tlpx_ns
);
1800 if (clk_zero_cnt
> ICL_CLK_ZERO_CNT_MAX
) {
1801 drm_dbg_kms(&dev_priv
->drm
,
1802 "clk_zero_cnt out of range (%d)\n", clk_zero_cnt
);
1803 clk_zero_cnt
= ICL_CLK_ZERO_CNT_MAX
;
1806 /* trail cnt in escape clocks*/
1807 trail_cnt
= DIV_ROUND_UP(tclk_trail_ns
, tlpx_ns
);
1808 if (trail_cnt
> ICL_TRAIL_CNT_MAX
) {
1809 drm_dbg_kms(&dev_priv
->drm
, "trail_cnt out of range (%d)\n",
1811 trail_cnt
= ICL_TRAIL_CNT_MAX
;
1814 /* tclk pre count in escape clocks */
1815 tclk_pre_cnt
= DIV_ROUND_UP(mipi_config
->tclk_pre
, tlpx_ns
);
1816 if (tclk_pre_cnt
> ICL_TCLK_PRE_CNT_MAX
) {
1817 drm_dbg_kms(&dev_priv
->drm
,
1818 "tclk_pre_cnt out of range (%d)\n", tclk_pre_cnt
);
1819 tclk_pre_cnt
= ICL_TCLK_PRE_CNT_MAX
;
1822 /* tclk post count in escape clocks */
1823 tclk_post_cnt
= DIV_ROUND_UP(mipi_config
->tclk_post
, tlpx_ns
);
1824 if (tclk_post_cnt
> ICL_TCLK_POST_CNT_MAX
) {
1825 drm_dbg_kms(&dev_priv
->drm
,
1826 "tclk_post_cnt out of range (%d)\n",
1828 tclk_post_cnt
= ICL_TCLK_POST_CNT_MAX
;
1831 /* hs zero cnt in escape clocks */
1832 hs_zero_cnt
= DIV_ROUND_UP(mipi_config
->ths_prepare_hszero
-
1833 ths_prepare_ns
, tlpx_ns
);
1834 if (hs_zero_cnt
> ICL_HS_ZERO_CNT_MAX
) {
1835 drm_dbg_kms(&dev_priv
->drm
, "hs_zero_cnt out of range (%d)\n",
1837 hs_zero_cnt
= ICL_HS_ZERO_CNT_MAX
;
1840 /* hs exit zero cnt in escape clocks */
1841 exit_zero_cnt
= DIV_ROUND_UP(mipi_config
->ths_exit
, tlpx_ns
);
1842 if (exit_zero_cnt
> ICL_EXIT_ZERO_CNT_MAX
) {
1843 drm_dbg_kms(&dev_priv
->drm
,
1844 "exit_zero_cnt out of range (%d)\n",
1846 exit_zero_cnt
= ICL_EXIT_ZERO_CNT_MAX
;
1849 /* clock lane dphy timings */
1850 intel_dsi
->dphy_reg
= (CLK_PREPARE_OVERRIDE
|
1851 CLK_PREPARE(prepare_cnt
) |
1853 CLK_ZERO(clk_zero_cnt
) |
1855 CLK_PRE(tclk_pre_cnt
) |
1857 CLK_POST(tclk_post_cnt
) |
1858 CLK_TRAIL_OVERRIDE
|
1859 CLK_TRAIL(trail_cnt
));
1861 /* data lanes dphy timings */
1862 intel_dsi
->dphy_data_lane_reg
= (HS_PREPARE_OVERRIDE
|
1863 HS_PREPARE(prepare_cnt
) |
1865 HS_ZERO(hs_zero_cnt
) |
1867 HS_TRAIL(trail_cnt
) |
1869 HS_EXIT(exit_zero_cnt
));
1871 intel_dsi_log_params(intel_dsi
);
1874 static void icl_dsi_add_properties(struct intel_connector
*connector
)
1876 u32 allowed_scalers
;
1878 allowed_scalers
= BIT(DRM_MODE_SCALE_ASPECT
) |
1879 BIT(DRM_MODE_SCALE_FULLSCREEN
) |
1880 BIT(DRM_MODE_SCALE_CENTER
);
1882 drm_connector_attach_scaling_mode_property(&connector
->base
,
1885 connector
->base
.state
->scaling_mode
= DRM_MODE_SCALE_ASPECT
;
1887 drm_connector_set_panel_orientation_with_quirk(&connector
->base
,
1888 intel_dsi_get_panel_orientation(connector
),
1889 connector
->panel
.fixed_mode
->hdisplay
,
1890 connector
->panel
.fixed_mode
->vdisplay
);
1893 void icl_dsi_init(struct drm_i915_private
*dev_priv
)
1895 struct drm_device
*dev
= &dev_priv
->drm
;
1896 struct intel_dsi
*intel_dsi
;
1897 struct intel_encoder
*encoder
;
1898 struct intel_connector
*intel_connector
;
1899 struct drm_connector
*connector
;
1900 struct drm_display_mode
*fixed_mode
;
1903 if (!intel_bios_is_dsi_present(dev_priv
, &port
))
1906 intel_dsi
= kzalloc(sizeof(*intel_dsi
), GFP_KERNEL
);
1910 intel_connector
= intel_connector_alloc();
1911 if (!intel_connector
) {
1916 encoder
= &intel_dsi
->base
;
1917 intel_dsi
->attached_connector
= intel_connector
;
1918 connector
= &intel_connector
->base
;
1920 /* register DSI encoder with DRM subsystem */
1921 drm_encoder_init(dev
, &encoder
->base
, &gen11_dsi_encoder_funcs
,
1922 DRM_MODE_ENCODER_DSI
, "DSI %c", port_name(port
));
1924 encoder
->pre_pll_enable
= gen11_dsi_pre_pll_enable
;
1925 encoder
->pre_enable
= gen11_dsi_pre_enable
;
1926 encoder
->enable
= gen11_dsi_enable
;
1927 encoder
->disable
= gen11_dsi_disable
;
1928 encoder
->post_disable
= gen11_dsi_post_disable
;
1929 encoder
->port
= port
;
1930 encoder
->get_config
= gen11_dsi_get_config
;
1931 encoder
->update_pipe
= intel_panel_update_backlight
;
1932 encoder
->compute_config
= gen11_dsi_compute_config
;
1933 encoder
->get_hw_state
= gen11_dsi_get_hw_state
;
1934 encoder
->initial_fastset_check
= gen11_dsi_initial_fastset_check
;
1935 encoder
->type
= INTEL_OUTPUT_DSI
;
1936 encoder
->cloneable
= 0;
1937 encoder
->pipe_mask
= ~0;
1938 encoder
->power_domain
= POWER_DOMAIN_PORT_DSI
;
1939 encoder
->get_power_domains
= gen11_dsi_get_power_domains
;
1941 /* register DSI connector with DRM subsystem */
1942 drm_connector_init(dev
, connector
, &gen11_dsi_connector_funcs
,
1943 DRM_MODE_CONNECTOR_DSI
);
1944 drm_connector_helper_add(connector
, &gen11_dsi_connector_helper_funcs
);
1945 connector
->display_info
.subpixel_order
= SubPixelHorizontalRGB
;
1946 connector
->interlace_allowed
= false;
1947 connector
->doublescan_allowed
= false;
1948 intel_connector
->get_hw_state
= intel_connector_get_hw_state
;
1950 /* attach connector to encoder */
1951 intel_connector_attach_encoder(intel_connector
, encoder
);
1953 mutex_lock(&dev
->mode_config
.mutex
);
1954 fixed_mode
= intel_panel_vbt_fixed_mode(intel_connector
);
1955 mutex_unlock(&dev
->mode_config
.mutex
);
1958 drm_err(&dev_priv
->drm
, "DSI fixed mode info missing\n");
1962 intel_panel_init(&intel_connector
->panel
, fixed_mode
, NULL
);
1963 intel_panel_setup_backlight(connector
, INVALID_PIPE
);
1965 if (dev_priv
->vbt
.dsi
.config
->dual_link
)
1966 intel_dsi
->ports
= BIT(PORT_A
) | BIT(PORT_B
);
1968 intel_dsi
->ports
= BIT(port
);
1970 intel_dsi
->dcs_backlight_ports
= dev_priv
->vbt
.dsi
.bl_ports
;
1971 intel_dsi
->dcs_cabc_ports
= dev_priv
->vbt
.dsi
.cabc_ports
;
1973 for_each_dsi_port(port
, intel_dsi
->ports
) {
1974 struct intel_dsi_host
*host
;
1976 host
= intel_dsi_host_init(intel_dsi
, &gen11_dsi_host_ops
, port
);
1980 intel_dsi
->dsi_hosts
[port
] = host
;
1983 if (!intel_dsi_vbt_init(intel_dsi
, MIPI_DSI_GENERIC_PANEL_ID
)) {
1984 drm_dbg_kms(&dev_priv
->drm
, "no device found\n");
1988 icl_dphy_param_init(intel_dsi
);
1990 icl_dsi_add_properties(intel_connector
);
1994 drm_connector_cleanup(connector
);
1995 drm_encoder_cleanup(&encoder
->base
);
1997 kfree(intel_connector
);