Merge tag 'v3.3.7' into 3.3/master
[zen-stable.git] / drivers / gpu / drm / nouveau / nv50_crtc.c
blob8f6c2ace3adf5c48b40863b34b240ceca88082ea
1 /*
2 * Copyright (C) 2008 Maarten Maathuis.
3 * All Rights Reserved.
5 * Permission is hereby granted, free of charge, to any person obtaining
6 * a copy of this software and associated documentation files (the
7 * "Software"), to deal in the Software without restriction, including
8 * without limitation the rights to use, copy, modify, merge, publish,
9 * distribute, sublicense, and/or sell copies of the Software, and to
10 * permit persons to whom the Software is furnished to do so, subject to
11 * the following conditions:
13 * The above copyright notice and this permission notice (including the
14 * next paragraph) shall be included in all copies or substantial
15 * portions of the Software.
17 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
18 * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
19 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.
20 * IN NO EVENT SHALL THE COPYRIGHT OWNER(S) AND/OR ITS SUPPLIERS BE
21 * LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION
22 * OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION
23 * WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
27 #include "drmP.h"
28 #include "drm_mode.h"
29 #include "drm_crtc_helper.h"
31 #define NOUVEAU_DMA_DEBUG (nouveau_reg_debug & NOUVEAU_REG_DEBUG_EVO)
32 #include "nouveau_reg.h"
33 #include "nouveau_drv.h"
34 #include "nouveau_hw.h"
35 #include "nouveau_encoder.h"
36 #include "nouveau_crtc.h"
37 #include "nouveau_fb.h"
38 #include "nouveau_connector.h"
39 #include "nv50_display.h"
41 static void
42 nv50_crtc_lut_load(struct drm_crtc *crtc)
44 struct nouveau_crtc *nv_crtc = nouveau_crtc(crtc);
45 void __iomem *lut = nvbo_kmap_obj_iovirtual(nv_crtc->lut.nvbo);
46 int i;
48 NV_DEBUG_KMS(crtc->dev, "\n");
50 for (i = 0; i < 256; i++) {
51 writew(nv_crtc->lut.r[i] >> 2, lut + 8*i + 0);
52 writew(nv_crtc->lut.g[i] >> 2, lut + 8*i + 2);
53 writew(nv_crtc->lut.b[i] >> 2, lut + 8*i + 4);
56 if (nv_crtc->lut.depth == 30) {
57 writew(nv_crtc->lut.r[i - 1] >> 2, lut + 8*i + 0);
58 writew(nv_crtc->lut.g[i - 1] >> 2, lut + 8*i + 2);
59 writew(nv_crtc->lut.b[i - 1] >> 2, lut + 8*i + 4);
63 int
64 nv50_crtc_blank(struct nouveau_crtc *nv_crtc, bool blanked)
66 struct drm_device *dev = nv_crtc->base.dev;
67 struct drm_nouveau_private *dev_priv = dev->dev_private;
68 struct nouveau_channel *evo = nv50_display(dev)->master;
69 int index = nv_crtc->index, ret;
71 NV_DEBUG_KMS(dev, "index %d\n", nv_crtc->index);
72 NV_DEBUG_KMS(dev, "%s\n", blanked ? "blanked" : "unblanked");
74 if (blanked) {
75 nv_crtc->cursor.hide(nv_crtc, false);
77 ret = RING_SPACE(evo, dev_priv->chipset != 0x50 ? 7 : 5);
78 if (ret) {
79 NV_ERROR(dev, "no space while blanking crtc\n");
80 return ret;
82 BEGIN_RING(evo, 0, NV50_EVO_CRTC(index, CLUT_MODE), 2);
83 OUT_RING(evo, NV50_EVO_CRTC_CLUT_MODE_BLANK);
84 OUT_RING(evo, 0);
85 if (dev_priv->chipset != 0x50) {
86 BEGIN_RING(evo, 0, NV84_EVO_CRTC(index, CLUT_DMA), 1);
87 OUT_RING(evo, NV84_EVO_CRTC_CLUT_DMA_HANDLE_NONE);
90 BEGIN_RING(evo, 0, NV50_EVO_CRTC(index, FB_DMA), 1);
91 OUT_RING(evo, NV50_EVO_CRTC_FB_DMA_HANDLE_NONE);
92 } else {
93 if (nv_crtc->cursor.visible)
94 nv_crtc->cursor.show(nv_crtc, false);
95 else
96 nv_crtc->cursor.hide(nv_crtc, false);
98 ret = RING_SPACE(evo, dev_priv->chipset != 0x50 ? 10 : 8);
99 if (ret) {
100 NV_ERROR(dev, "no space while unblanking crtc\n");
101 return ret;
103 BEGIN_RING(evo, 0, NV50_EVO_CRTC(index, CLUT_MODE), 2);
104 OUT_RING(evo, nv_crtc->lut.depth == 8 ?
105 NV50_EVO_CRTC_CLUT_MODE_OFF :
106 NV50_EVO_CRTC_CLUT_MODE_ON);
107 OUT_RING(evo, nv_crtc->lut.nvbo->bo.offset >> 8);
108 if (dev_priv->chipset != 0x50) {
109 BEGIN_RING(evo, 0, NV84_EVO_CRTC(index, CLUT_DMA), 1);
110 OUT_RING(evo, NvEvoVRAM);
113 BEGIN_RING(evo, 0, NV50_EVO_CRTC(index, FB_OFFSET), 2);
114 OUT_RING(evo, nv_crtc->fb.offset >> 8);
115 OUT_RING(evo, 0);
116 BEGIN_RING(evo, 0, NV50_EVO_CRTC(index, FB_DMA), 1);
117 if (dev_priv->chipset != 0x50)
118 if (nv_crtc->fb.tile_flags == 0x7a00 ||
119 nv_crtc->fb.tile_flags == 0xfe00)
120 OUT_RING(evo, NvEvoFB32);
121 else
122 if (nv_crtc->fb.tile_flags == 0x7000)
123 OUT_RING(evo, NvEvoFB16);
124 else
125 OUT_RING(evo, NvEvoVRAM_LP);
126 else
127 OUT_RING(evo, NvEvoVRAM_LP);
130 nv_crtc->fb.blanked = blanked;
131 return 0;
134 static int
135 nv50_crtc_set_dither(struct nouveau_crtc *nv_crtc, bool update)
137 struct nouveau_channel *evo = nv50_display(nv_crtc->base.dev)->master;
138 struct nouveau_connector *nv_connector;
139 struct drm_connector *connector;
140 int head = nv_crtc->index, ret;
141 u32 mode = 0x00;
143 nv_connector = nouveau_crtc_connector_get(nv_crtc);
144 connector = &nv_connector->base;
145 if (nv_connector->dithering_mode == DITHERING_MODE_AUTO) {
146 if (nv_crtc->base.fb->depth > connector->display_info.bpc * 3)
147 mode = DITHERING_MODE_DYNAMIC2X2;
148 } else {
149 mode = nv_connector->dithering_mode;
152 if (nv_connector->dithering_depth == DITHERING_DEPTH_AUTO) {
153 if (connector->display_info.bpc >= 8)
154 mode |= DITHERING_DEPTH_8BPC;
155 } else {
156 mode |= nv_connector->dithering_depth;
159 ret = RING_SPACE(evo, 2 + (update ? 2 : 0));
160 if (ret == 0) {
161 BEGIN_RING(evo, 0, NV50_EVO_CRTC(head, DITHER_CTRL), 1);
162 OUT_RING (evo, mode);
163 if (update) {
164 BEGIN_RING(evo, 0, NV50_EVO_UPDATE, 1);
165 OUT_RING (evo, 0);
166 FIRE_RING (evo);
170 return ret;
173 struct nouveau_connector *
174 nouveau_crtc_connector_get(struct nouveau_crtc *nv_crtc)
176 struct drm_device *dev = nv_crtc->base.dev;
177 struct drm_connector *connector;
178 struct drm_crtc *crtc = to_drm_crtc(nv_crtc);
180 /* The safest approach is to find an encoder with the right crtc, that
181 * is also linked to a connector. */
182 list_for_each_entry(connector, &dev->mode_config.connector_list, head) {
183 if (connector->encoder)
184 if (connector->encoder->crtc == crtc)
185 return nouveau_connector(connector);
188 return NULL;
191 static int
192 nv50_crtc_set_scale(struct nouveau_crtc *nv_crtc, bool update)
194 struct nouveau_connector *nv_connector;
195 struct drm_crtc *crtc = &nv_crtc->base;
196 struct drm_device *dev = crtc->dev;
197 struct nouveau_channel *evo = nv50_display(dev)->master;
198 struct drm_display_mode *umode = &crtc->mode;
199 struct drm_display_mode *omode;
200 int scaling_mode, ret;
201 u32 ctrl = 0, oX, oY;
203 NV_DEBUG_KMS(dev, "\n");
205 nv_connector = nouveau_crtc_connector_get(nv_crtc);
206 if (!nv_connector || !nv_connector->native_mode) {
207 NV_ERROR(dev, "no native mode, forcing panel scaling\n");
208 scaling_mode = DRM_MODE_SCALE_NONE;
209 } else {
210 scaling_mode = nv_connector->scaling_mode;
213 /* start off at the resolution we programmed the crtc for, this
214 * effectively handles NONE/FULL scaling
216 if (scaling_mode != DRM_MODE_SCALE_NONE)
217 omode = nv_connector->native_mode;
218 else
219 omode = umode;
221 oX = omode->hdisplay;
222 oY = omode->vdisplay;
223 if (omode->flags & DRM_MODE_FLAG_DBLSCAN)
224 oY *= 2;
226 /* add overscan compensation if necessary, will keep the aspect
227 * ratio the same as the backend mode unless overridden by the
228 * user setting both hborder and vborder properties.
230 if (nv_connector && ( nv_connector->underscan == UNDERSCAN_ON ||
231 (nv_connector->underscan == UNDERSCAN_AUTO &&
232 nv_connector->edid &&
233 drm_detect_hdmi_monitor(nv_connector->edid)))) {
234 u32 bX = nv_connector->underscan_hborder;
235 u32 bY = nv_connector->underscan_vborder;
236 u32 aspect = (oY << 19) / oX;
238 if (bX) {
239 oX -= (bX * 2);
240 if (bY) oY -= (bY * 2);
241 else oY = ((oX * aspect) + (aspect / 2)) >> 19;
242 } else {
243 oX -= (oX >> 4) + 32;
244 if (bY) oY -= (bY * 2);
245 else oY = ((oX * aspect) + (aspect / 2)) >> 19;
249 /* handle CENTER/ASPECT scaling, taking into account the areas
250 * removed already for overscan compensation
252 switch (scaling_mode) {
253 case DRM_MODE_SCALE_CENTER:
254 oX = min((u32)umode->hdisplay, oX);
255 oY = min((u32)umode->vdisplay, oY);
256 /* fall-through */
257 case DRM_MODE_SCALE_ASPECT:
258 if (oY < oX) {
259 u32 aspect = (umode->hdisplay << 19) / umode->vdisplay;
260 oX = ((oY * aspect) + (aspect / 2)) >> 19;
261 } else {
262 u32 aspect = (umode->vdisplay << 19) / umode->hdisplay;
263 oY = ((oX * aspect) + (aspect / 2)) >> 19;
265 break;
266 default:
267 break;
270 if (umode->hdisplay != oX || umode->vdisplay != oY ||
271 umode->flags & DRM_MODE_FLAG_INTERLACE ||
272 umode->flags & DRM_MODE_FLAG_DBLSCAN)
273 ctrl |= NV50_EVO_CRTC_SCALE_CTRL_ACTIVE;
275 ret = RING_SPACE(evo, 5);
276 if (ret)
277 return ret;
279 BEGIN_RING(evo, 0, NV50_EVO_CRTC(nv_crtc->index, SCALE_CTRL), 1);
280 OUT_RING (evo, ctrl);
281 BEGIN_RING(evo, 0, NV50_EVO_CRTC(nv_crtc->index, SCALE_RES1), 2);
282 OUT_RING (evo, oY << 16 | oX);
283 OUT_RING (evo, oY << 16 | oX);
285 if (update) {
286 nv50_display_flip_stop(crtc);
287 nv50_display_sync(dev);
288 nv50_display_flip_next(crtc, crtc->fb, NULL);
291 return 0;
295 nv50_crtc_set_clock(struct drm_device *dev, int head, int pclk)
297 struct drm_nouveau_private *dev_priv = dev->dev_private;
298 struct pll_lims pll;
299 uint32_t reg1, reg2;
300 int ret, N1, M1, N2, M2, P;
302 ret = get_pll_limits(dev, PLL_VPLL0 + head, &pll);
303 if (ret)
304 return ret;
306 if (pll.vco2.maxfreq) {
307 ret = nv50_calc_pll(dev, &pll, pclk, &N1, &M1, &N2, &M2, &P);
308 if (ret <= 0)
309 return 0;
311 NV_DEBUG(dev, "pclk %d out %d NM1 %d %d NM2 %d %d P %d\n",
312 pclk, ret, N1, M1, N2, M2, P);
314 reg1 = nv_rd32(dev, pll.reg + 4) & 0xff00ff00;
315 reg2 = nv_rd32(dev, pll.reg + 8) & 0x8000ff00;
316 nv_wr32(dev, pll.reg + 0, 0x10000611);
317 nv_wr32(dev, pll.reg + 4, reg1 | (M1 << 16) | N1);
318 nv_wr32(dev, pll.reg + 8, reg2 | (P << 28) | (M2 << 16) | N2);
319 } else
320 if (dev_priv->chipset < NV_C0) {
321 ret = nva3_calc_pll(dev, &pll, pclk, &N1, &N2, &M1, &P);
322 if (ret <= 0)
323 return 0;
325 NV_DEBUG(dev, "pclk %d out %d N %d fN 0x%04x M %d P %d\n",
326 pclk, ret, N1, N2, M1, P);
328 reg1 = nv_rd32(dev, pll.reg + 4) & 0xffc00000;
329 nv_wr32(dev, pll.reg + 0, 0x50000610);
330 nv_wr32(dev, pll.reg + 4, reg1 | (P << 16) | (M1 << 8) | N1);
331 nv_wr32(dev, pll.reg + 8, N2);
332 } else {
333 ret = nva3_calc_pll(dev, &pll, pclk, &N1, &N2, &M1, &P);
334 if (ret <= 0)
335 return 0;
337 NV_DEBUG(dev, "pclk %d out %d N %d fN 0x%04x M %d P %d\n",
338 pclk, ret, N1, N2, M1, P);
340 nv_mask(dev, pll.reg + 0x0c, 0x00000000, 0x00000100);
341 nv_wr32(dev, pll.reg + 0x04, (P << 16) | (N1 << 8) | M1);
342 nv_wr32(dev, pll.reg + 0x10, N2 << 16);
345 return 0;
348 static void
349 nv50_crtc_destroy(struct drm_crtc *crtc)
351 struct drm_device *dev;
352 struct nouveau_crtc *nv_crtc;
354 if (!crtc)
355 return;
357 dev = crtc->dev;
358 nv_crtc = nouveau_crtc(crtc);
360 NV_DEBUG_KMS(dev, "\n");
362 drm_crtc_cleanup(&nv_crtc->base);
364 nouveau_bo_unmap(nv_crtc->lut.nvbo);
365 nouveau_bo_ref(NULL, &nv_crtc->lut.nvbo);
366 nouveau_bo_unmap(nv_crtc->cursor.nvbo);
367 nouveau_bo_ref(NULL, &nv_crtc->cursor.nvbo);
368 kfree(nv_crtc);
372 nv50_crtc_cursor_set(struct drm_crtc *crtc, struct drm_file *file_priv,
373 uint32_t buffer_handle, uint32_t width, uint32_t height)
375 struct drm_device *dev = crtc->dev;
376 struct nouveau_crtc *nv_crtc = nouveau_crtc(crtc);
377 struct nouveau_bo *cursor = NULL;
378 struct drm_gem_object *gem;
379 int ret = 0, i;
381 if (!buffer_handle) {
382 nv_crtc->cursor.hide(nv_crtc, true);
383 return 0;
386 if (width != 64 || height != 64)
387 return -EINVAL;
389 gem = drm_gem_object_lookup(dev, file_priv, buffer_handle);
390 if (!gem)
391 return -ENOENT;
392 cursor = nouveau_gem_object(gem);
394 ret = nouveau_bo_map(cursor);
395 if (ret)
396 goto out;
398 /* The simple will do for now. */
399 for (i = 0; i < 64 * 64; i++)
400 nouveau_bo_wr32(nv_crtc->cursor.nvbo, i, nouveau_bo_rd32(cursor, i));
402 nouveau_bo_unmap(cursor);
404 nv_crtc->cursor.set_offset(nv_crtc, nv_crtc->cursor.nvbo->bo.offset);
405 nv_crtc->cursor.show(nv_crtc, true);
407 out:
408 drm_gem_object_unreference_unlocked(gem);
409 return ret;
413 nv50_crtc_cursor_move(struct drm_crtc *crtc, int x, int y)
415 struct nouveau_crtc *nv_crtc = nouveau_crtc(crtc);
417 nv_crtc->cursor.set_pos(nv_crtc, x, y);
418 return 0;
421 static void
422 nv50_crtc_gamma_set(struct drm_crtc *crtc, u16 *r, u16 *g, u16 *b,
423 uint32_t start, uint32_t size)
425 int end = (start + size > 256) ? 256 : start + size, i;
426 struct nouveau_crtc *nv_crtc = nouveau_crtc(crtc);
428 for (i = start; i < end; i++) {
429 nv_crtc->lut.r[i] = r[i];
430 nv_crtc->lut.g[i] = g[i];
431 nv_crtc->lut.b[i] = b[i];
434 /* We need to know the depth before we upload, but it's possible to
435 * get called before a framebuffer is bound. If this is the case,
436 * mark the lut values as dirty by setting depth==0, and it'll be
437 * uploaded on the first mode_set_base()
439 if (!nv_crtc->base.fb) {
440 nv_crtc->lut.depth = 0;
441 return;
444 nv50_crtc_lut_load(crtc);
447 static void
448 nv50_crtc_save(struct drm_crtc *crtc)
450 NV_ERROR(crtc->dev, "!!\n");
453 static void
454 nv50_crtc_restore(struct drm_crtc *crtc)
456 NV_ERROR(crtc->dev, "!!\n");
459 static const struct drm_crtc_funcs nv50_crtc_funcs = {
460 .save = nv50_crtc_save,
461 .restore = nv50_crtc_restore,
462 .cursor_set = nv50_crtc_cursor_set,
463 .cursor_move = nv50_crtc_cursor_move,
464 .gamma_set = nv50_crtc_gamma_set,
465 .set_config = drm_crtc_helper_set_config,
466 .page_flip = nouveau_crtc_page_flip,
467 .destroy = nv50_crtc_destroy,
470 static void
471 nv50_crtc_dpms(struct drm_crtc *crtc, int mode)
475 static void
476 nv50_crtc_prepare(struct drm_crtc *crtc)
478 struct nouveau_crtc *nv_crtc = nouveau_crtc(crtc);
479 struct drm_device *dev = crtc->dev;
481 NV_DEBUG_KMS(dev, "index %d\n", nv_crtc->index);
483 nv50_display_flip_stop(crtc);
484 drm_vblank_pre_modeset(dev, nv_crtc->index);
485 nv50_crtc_blank(nv_crtc, true);
488 static void
489 nv50_crtc_commit(struct drm_crtc *crtc)
491 struct drm_device *dev = crtc->dev;
492 struct nouveau_crtc *nv_crtc = nouveau_crtc(crtc);
494 NV_DEBUG_KMS(dev, "index %d\n", nv_crtc->index);
496 nv50_crtc_blank(nv_crtc, false);
497 drm_vblank_post_modeset(dev, nv_crtc->index);
498 nv50_display_sync(dev);
499 nv50_display_flip_next(crtc, crtc->fb, NULL);
502 static bool
503 nv50_crtc_mode_fixup(struct drm_crtc *crtc, struct drm_display_mode *mode,
504 struct drm_display_mode *adjusted_mode)
506 return true;
509 static int
510 nv50_crtc_do_mode_set_base(struct drm_crtc *crtc,
511 struct drm_framebuffer *passed_fb,
512 int x, int y, bool atomic)
514 struct nouveau_crtc *nv_crtc = nouveau_crtc(crtc);
515 struct drm_device *dev = nv_crtc->base.dev;
516 struct drm_nouveau_private *dev_priv = dev->dev_private;
517 struct nouveau_channel *evo = nv50_display(dev)->master;
518 struct drm_framebuffer *drm_fb;
519 struct nouveau_framebuffer *fb;
520 int ret;
522 NV_DEBUG_KMS(dev, "index %d\n", nv_crtc->index);
524 /* no fb bound */
525 if (!atomic && !crtc->fb) {
526 NV_DEBUG_KMS(dev, "No FB bound\n");
527 return 0;
530 /* If atomic, we want to switch to the fb we were passed, so
531 * now we update pointers to do that. (We don't pin; just
532 * assume we're already pinned and update the base address.)
534 if (atomic) {
535 drm_fb = passed_fb;
536 fb = nouveau_framebuffer(passed_fb);
537 } else {
538 drm_fb = crtc->fb;
539 fb = nouveau_framebuffer(crtc->fb);
540 /* If not atomic, we can go ahead and pin, and unpin the
541 * old fb we were passed.
543 ret = nouveau_bo_pin(fb->nvbo, TTM_PL_FLAG_VRAM);
544 if (ret)
545 return ret;
547 if (passed_fb) {
548 struct nouveau_framebuffer *ofb = nouveau_framebuffer(passed_fb);
549 nouveau_bo_unpin(ofb->nvbo);
553 nv_crtc->fb.offset = fb->nvbo->bo.offset;
554 nv_crtc->fb.tile_flags = nouveau_bo_tile_layout(fb->nvbo);
555 nv_crtc->fb.cpp = drm_fb->bits_per_pixel / 8;
556 if (!nv_crtc->fb.blanked && dev_priv->chipset != 0x50) {
557 ret = RING_SPACE(evo, 2);
558 if (ret)
559 return ret;
561 BEGIN_RING(evo, 0, NV50_EVO_CRTC(nv_crtc->index, FB_DMA), 1);
562 OUT_RING (evo, fb->r_dma);
565 ret = RING_SPACE(evo, 12);
566 if (ret)
567 return ret;
569 BEGIN_RING(evo, 0, NV50_EVO_CRTC(nv_crtc->index, FB_OFFSET), 5);
570 OUT_RING (evo, nv_crtc->fb.offset >> 8);
571 OUT_RING (evo, 0);
572 OUT_RING (evo, (drm_fb->height << 16) | drm_fb->width);
573 OUT_RING (evo, fb->r_pitch);
574 OUT_RING (evo, fb->r_format);
576 BEGIN_RING(evo, 0, NV50_EVO_CRTC(nv_crtc->index, CLUT_MODE), 1);
577 OUT_RING (evo, fb->base.depth == 8 ?
578 NV50_EVO_CRTC_CLUT_MODE_OFF : NV50_EVO_CRTC_CLUT_MODE_ON);
580 BEGIN_RING(evo, 0, NV50_EVO_CRTC(nv_crtc->index, COLOR_CTRL), 1);
581 OUT_RING (evo, NV50_EVO_CRTC_COLOR_CTRL_COLOR);
582 BEGIN_RING(evo, 0, NV50_EVO_CRTC(nv_crtc->index, FB_POS), 1);
583 OUT_RING (evo, (y << 16) | x);
585 if (nv_crtc->lut.depth != fb->base.depth) {
586 nv_crtc->lut.depth = fb->base.depth;
587 nv50_crtc_lut_load(crtc);
590 return 0;
593 static int
594 nv50_crtc_mode_set(struct drm_crtc *crtc, struct drm_display_mode *umode,
595 struct drm_display_mode *mode, int x, int y,
596 struct drm_framebuffer *old_fb)
598 struct drm_device *dev = crtc->dev;
599 struct nouveau_channel *evo = nv50_display(dev)->master;
600 struct nouveau_crtc *nv_crtc = nouveau_crtc(crtc);
601 u32 head = nv_crtc->index * 0x400;
602 u32 ilace = (mode->flags & DRM_MODE_FLAG_INTERLACE) ? 2 : 1;
603 u32 vscan = (mode->flags & DRM_MODE_FLAG_DBLSCAN) ? 2 : 1;
604 u32 hactive, hsynce, hbackp, hfrontp, hblanke, hblanks;
605 u32 vactive, vsynce, vbackp, vfrontp, vblanke, vblanks;
606 u32 vblan2e = 0, vblan2s = 1;
607 int ret;
609 /* hw timing description looks like this:
611 * <sync> <back porch> <---------display---------> <front porch>
612 * ______
613 * |____________|---------------------------|____________|
615 * ^ synce ^ blanke ^ blanks ^ active
617 * interlaced modes also have 2 additional values pointing at the end
618 * and start of the next field's blanking period.
621 hactive = mode->htotal;
622 hsynce = mode->hsync_end - mode->hsync_start - 1;
623 hbackp = mode->htotal - mode->hsync_end;
624 hblanke = hsynce + hbackp;
625 hfrontp = mode->hsync_start - mode->hdisplay;
626 hblanks = mode->htotal - hfrontp - 1;
628 vactive = mode->vtotal * vscan / ilace;
629 vsynce = ((mode->vsync_end - mode->vsync_start) * vscan / ilace) - 1;
630 vbackp = (mode->vtotal - mode->vsync_end) * vscan / ilace;
631 vblanke = vsynce + vbackp;
632 vfrontp = (mode->vsync_start - mode->vdisplay) * vscan / ilace;
633 vblanks = vactive - vfrontp - 1;
634 if (mode->flags & DRM_MODE_FLAG_INTERLACE) {
635 vblan2e = vactive + vsynce + vbackp;
636 vblan2s = vblan2e + (mode->vdisplay * vscan / ilace);
637 vactive = (vactive * 2) + 1;
640 ret = RING_SPACE(evo, 18);
641 if (ret == 0) {
642 BEGIN_RING(evo, 0, 0x0804 + head, 2);
643 OUT_RING (evo, 0x00800000 | mode->clock);
644 OUT_RING (evo, (ilace == 2) ? 2 : 0);
645 BEGIN_RING(evo, 0, 0x0810 + head, 6);
646 OUT_RING (evo, 0x00000000); /* border colour */
647 OUT_RING (evo, (vactive << 16) | hactive);
648 OUT_RING (evo, ( vsynce << 16) | hsynce);
649 OUT_RING (evo, (vblanke << 16) | hblanke);
650 OUT_RING (evo, (vblanks << 16) | hblanks);
651 OUT_RING (evo, (vblan2e << 16) | vblan2s);
652 BEGIN_RING(evo, 0, 0x082c + head, 1);
653 OUT_RING (evo, 0x00000000);
654 BEGIN_RING(evo, 0, 0x0900 + head, 1);
655 OUT_RING (evo, 0x00000311); /* makes sync channel work */
656 BEGIN_RING(evo, 0, 0x08c8 + head, 1);
657 OUT_RING (evo, (umode->vdisplay << 16) | umode->hdisplay);
658 BEGIN_RING(evo, 0, 0x08d4 + head, 1);
659 OUT_RING (evo, 0x00000000); /* screen position */
662 nv_crtc->set_dither(nv_crtc, false);
663 nv_crtc->set_scale(nv_crtc, false);
665 return nv50_crtc_do_mode_set_base(crtc, old_fb, x, y, false);
668 static int
669 nv50_crtc_mode_set_base(struct drm_crtc *crtc, int x, int y,
670 struct drm_framebuffer *old_fb)
672 int ret;
674 nv50_display_flip_stop(crtc);
675 ret = nv50_crtc_do_mode_set_base(crtc, old_fb, x, y, false);
676 if (ret)
677 return ret;
679 ret = nv50_display_sync(crtc->dev);
680 if (ret)
681 return ret;
683 return nv50_display_flip_next(crtc, crtc->fb, NULL);
686 static int
687 nv50_crtc_mode_set_base_atomic(struct drm_crtc *crtc,
688 struct drm_framebuffer *fb,
689 int x, int y, enum mode_set_atomic state)
691 int ret;
693 nv50_display_flip_stop(crtc);
694 ret = nv50_crtc_do_mode_set_base(crtc, fb, x, y, true);
695 if (ret)
696 return ret;
698 return nv50_display_sync(crtc->dev);
701 static const struct drm_crtc_helper_funcs nv50_crtc_helper_funcs = {
702 .dpms = nv50_crtc_dpms,
703 .prepare = nv50_crtc_prepare,
704 .commit = nv50_crtc_commit,
705 .mode_fixup = nv50_crtc_mode_fixup,
706 .mode_set = nv50_crtc_mode_set,
707 .mode_set_base = nv50_crtc_mode_set_base,
708 .mode_set_base_atomic = nv50_crtc_mode_set_base_atomic,
709 .load_lut = nv50_crtc_lut_load,
713 nv50_crtc_create(struct drm_device *dev, int index)
715 struct nouveau_crtc *nv_crtc = NULL;
716 int ret, i;
718 NV_DEBUG_KMS(dev, "\n");
720 nv_crtc = kzalloc(sizeof(*nv_crtc), GFP_KERNEL);
721 if (!nv_crtc)
722 return -ENOMEM;
724 /* Default CLUT parameters, will be activated on the hw upon
725 * first mode set.
727 for (i = 0; i < 256; i++) {
728 nv_crtc->lut.r[i] = i << 8;
729 nv_crtc->lut.g[i] = i << 8;
730 nv_crtc->lut.b[i] = i << 8;
732 nv_crtc->lut.depth = 0;
734 ret = nouveau_bo_new(dev, 4096, 0x100, TTM_PL_FLAG_VRAM,
735 0, 0x0000, &nv_crtc->lut.nvbo);
736 if (!ret) {
737 ret = nouveau_bo_pin(nv_crtc->lut.nvbo, TTM_PL_FLAG_VRAM);
738 if (!ret)
739 ret = nouveau_bo_map(nv_crtc->lut.nvbo);
740 if (ret)
741 nouveau_bo_ref(NULL, &nv_crtc->lut.nvbo);
744 if (ret) {
745 kfree(nv_crtc);
746 return ret;
749 nv_crtc->index = index;
751 /* set function pointers */
752 nv_crtc->set_dither = nv50_crtc_set_dither;
753 nv_crtc->set_scale = nv50_crtc_set_scale;
755 drm_crtc_init(dev, &nv_crtc->base, &nv50_crtc_funcs);
756 drm_crtc_helper_add(&nv_crtc->base, &nv50_crtc_helper_funcs);
757 drm_mode_crtc_set_gamma_size(&nv_crtc->base, 256);
759 ret = nouveau_bo_new(dev, 64*64*4, 0x100, TTM_PL_FLAG_VRAM,
760 0, 0x0000, &nv_crtc->cursor.nvbo);
761 if (!ret) {
762 ret = nouveau_bo_pin(nv_crtc->cursor.nvbo, TTM_PL_FLAG_VRAM);
763 if (!ret)
764 ret = nouveau_bo_map(nv_crtc->cursor.nvbo);
765 if (ret)
766 nouveau_bo_ref(NULL, &nv_crtc->cursor.nvbo);
769 nv50_cursor_init(nv_crtc);
770 return 0;