vt: vt_ioctl: fix VT_DISALLOCATE freeing in-use virtual console
[linux/fpc-iii.git] / drivers / gpu / drm / xen / xen_drm_front_kms.c
bloba3479eb72d79446c098cf7ebf7d3fb792a669f7e
1 // SPDX-License-Identifier: GPL-2.0 OR MIT
3 /*
4 * Xen para-virtual DRM device
6 * Copyright (C) 2016-2018 EPAM Systems Inc.
8 * Author: Oleksandr Andrushchenko <oleksandr_andrushchenko@epam.com>
9 */
11 #include "xen_drm_front_kms.h"
13 #include <drm/drmP.h>
14 #include <drm/drm_atomic.h>
15 #include <drm/drm_atomic_helper.h>
16 #include <drm/drm_crtc_helper.h>
17 #include <drm/drm_gem.h>
18 #include <drm/drm_gem_framebuffer_helper.h>
20 #include "xen_drm_front.h"
21 #include "xen_drm_front_conn.h"
24 * Timeout in ms to wait for frame done event from the backend:
25 * must be a bit more than IO time-out
27 #define FRAME_DONE_TO_MS (XEN_DRM_FRONT_WAIT_BACK_MS + 100)
29 static struct xen_drm_front_drm_pipeline *
30 to_xen_drm_pipeline(struct drm_simple_display_pipe *pipe)
32 return container_of(pipe, struct xen_drm_front_drm_pipeline, pipe);
35 static void fb_destroy(struct drm_framebuffer *fb)
37 struct xen_drm_front_drm_info *drm_info = fb->dev->dev_private;
38 int idx;
40 if (drm_dev_enter(fb->dev, &idx)) {
41 xen_drm_front_fb_detach(drm_info->front_info,
42 xen_drm_front_fb_to_cookie(fb));
43 drm_dev_exit(idx);
45 drm_gem_fb_destroy(fb);
48 static struct drm_framebuffer_funcs fb_funcs = {
49 .destroy = fb_destroy,
52 static struct drm_framebuffer *
53 fb_create(struct drm_device *dev, struct drm_file *filp,
54 const struct drm_mode_fb_cmd2 *mode_cmd)
56 struct xen_drm_front_drm_info *drm_info = dev->dev_private;
57 static struct drm_framebuffer *fb;
58 struct drm_gem_object *gem_obj;
59 int ret;
61 fb = drm_gem_fb_create_with_funcs(dev, filp, mode_cmd, &fb_funcs);
62 if (IS_ERR_OR_NULL(fb))
63 return fb;
65 gem_obj = drm_gem_object_lookup(filp, mode_cmd->handles[0]);
66 if (!gem_obj) {
67 DRM_ERROR("Failed to lookup GEM object\n");
68 ret = -ENOENT;
69 goto fail;
72 drm_gem_object_put_unlocked(gem_obj);
74 ret = xen_drm_front_fb_attach(drm_info->front_info,
75 xen_drm_front_dbuf_to_cookie(gem_obj),
76 xen_drm_front_fb_to_cookie(fb),
77 fb->width, fb->height,
78 fb->format->format);
79 if (ret < 0) {
80 DRM_ERROR("Back failed to attach FB %p: %d\n", fb, ret);
81 goto fail;
84 return fb;
86 fail:
87 drm_gem_fb_destroy(fb);
88 return ERR_PTR(ret);
91 static const struct drm_mode_config_funcs mode_config_funcs = {
92 .fb_create = fb_create,
93 .atomic_check = drm_atomic_helper_check,
94 .atomic_commit = drm_atomic_helper_commit,
97 static void send_pending_event(struct xen_drm_front_drm_pipeline *pipeline)
99 struct drm_crtc *crtc = &pipeline->pipe.crtc;
100 struct drm_device *dev = crtc->dev;
101 unsigned long flags;
103 spin_lock_irqsave(&dev->event_lock, flags);
104 if (pipeline->pending_event)
105 drm_crtc_send_vblank_event(crtc, pipeline->pending_event);
106 pipeline->pending_event = NULL;
107 spin_unlock_irqrestore(&dev->event_lock, flags);
110 static void display_enable(struct drm_simple_display_pipe *pipe,
111 struct drm_crtc_state *crtc_state,
112 struct drm_plane_state *plane_state)
114 struct xen_drm_front_drm_pipeline *pipeline =
115 to_xen_drm_pipeline(pipe);
116 struct drm_crtc *crtc = &pipe->crtc;
117 struct drm_framebuffer *fb = plane_state->fb;
118 int ret, idx;
120 if (!drm_dev_enter(pipe->crtc.dev, &idx))
121 return;
123 ret = xen_drm_front_mode_set(pipeline, crtc->x, crtc->y,
124 fb->width, fb->height,
125 fb->format->cpp[0] * 8,
126 xen_drm_front_fb_to_cookie(fb));
128 if (ret) {
129 DRM_ERROR("Failed to enable display: %d\n", ret);
130 pipeline->conn_connected = false;
133 drm_dev_exit(idx);
136 static void display_disable(struct drm_simple_display_pipe *pipe)
138 struct xen_drm_front_drm_pipeline *pipeline =
139 to_xen_drm_pipeline(pipe);
140 int ret = 0, idx;
142 if (drm_dev_enter(pipe->crtc.dev, &idx)) {
143 ret = xen_drm_front_mode_set(pipeline, 0, 0, 0, 0, 0,
144 xen_drm_front_fb_to_cookie(NULL));
145 drm_dev_exit(idx);
147 if (ret)
148 DRM_ERROR("Failed to disable display: %d\n", ret);
150 /* Make sure we can restart with enabled connector next time */
151 pipeline->conn_connected = true;
153 /* release stalled event if any */
154 send_pending_event(pipeline);
157 void xen_drm_front_kms_on_frame_done(struct xen_drm_front_drm_pipeline *pipeline,
158 u64 fb_cookie)
161 * This runs in interrupt context, e.g. under
162 * drm_info->front_info->io_lock, so we cannot call _sync version
163 * to cancel the work
165 cancel_delayed_work(&pipeline->pflip_to_worker);
167 send_pending_event(pipeline);
170 static void pflip_to_worker(struct work_struct *work)
172 struct delayed_work *delayed_work = to_delayed_work(work);
173 struct xen_drm_front_drm_pipeline *pipeline =
174 container_of(delayed_work,
175 struct xen_drm_front_drm_pipeline,
176 pflip_to_worker);
178 DRM_ERROR("Frame done timed-out, releasing");
179 send_pending_event(pipeline);
182 static bool display_send_page_flip(struct drm_simple_display_pipe *pipe,
183 struct drm_plane_state *old_plane_state)
185 struct drm_plane_state *plane_state =
186 drm_atomic_get_new_plane_state(old_plane_state->state,
187 &pipe->plane);
190 * If old_plane_state->fb is NULL and plane_state->fb is not,
191 * then this is an atomic commit which will enable display.
192 * If old_plane_state->fb is not NULL and plane_state->fb is,
193 * then this is an atomic commit which will disable display.
194 * Ignore these and do not send page flip as this framebuffer will be
195 * sent to the backend as a part of display_set_config call.
197 if (old_plane_state->fb && plane_state->fb) {
198 struct xen_drm_front_drm_pipeline *pipeline =
199 to_xen_drm_pipeline(pipe);
200 struct xen_drm_front_drm_info *drm_info = pipeline->drm_info;
201 int ret;
203 schedule_delayed_work(&pipeline->pflip_to_worker,
204 msecs_to_jiffies(FRAME_DONE_TO_MS));
206 ret = xen_drm_front_page_flip(drm_info->front_info,
207 pipeline->index,
208 xen_drm_front_fb_to_cookie(plane_state->fb));
209 if (ret) {
210 DRM_ERROR("Failed to send page flip request to backend: %d\n", ret);
212 pipeline->conn_connected = false;
214 * Report the flip not handled, so pending event is
215 * sent, unblocking user-space.
217 return false;
220 * Signal that page flip was handled, pending event will be sent
221 * on frame done event from the backend.
223 return true;
226 return false;
229 static void display_update(struct drm_simple_display_pipe *pipe,
230 struct drm_plane_state *old_plane_state)
232 struct xen_drm_front_drm_pipeline *pipeline =
233 to_xen_drm_pipeline(pipe);
234 struct drm_crtc *crtc = &pipe->crtc;
235 struct drm_pending_vblank_event *event;
236 int idx;
238 event = crtc->state->event;
239 if (event) {
240 struct drm_device *dev = crtc->dev;
241 unsigned long flags;
243 WARN_ON(pipeline->pending_event);
245 spin_lock_irqsave(&dev->event_lock, flags);
246 crtc->state->event = NULL;
248 pipeline->pending_event = event;
249 spin_unlock_irqrestore(&dev->event_lock, flags);
252 if (!drm_dev_enter(pipe->crtc.dev, &idx)) {
253 send_pending_event(pipeline);
254 return;
258 * Send page flip request to the backend *after* we have event cached
259 * above, so on page flip done event from the backend we can
260 * deliver it and there is no race condition between this code and
261 * event from the backend.
262 * If this is not a page flip, e.g. no flip done event from the backend
263 * is expected, then send now.
265 if (!display_send_page_flip(pipe, old_plane_state))
266 send_pending_event(pipeline);
268 drm_dev_exit(idx);
271 static enum drm_mode_status
272 display_mode_valid(struct drm_crtc *crtc, const struct drm_display_mode *mode)
274 struct xen_drm_front_drm_pipeline *pipeline =
275 container_of(crtc, struct xen_drm_front_drm_pipeline,
276 pipe.crtc);
278 if (mode->hdisplay != pipeline->width)
279 return MODE_ERROR;
281 if (mode->vdisplay != pipeline->height)
282 return MODE_ERROR;
284 return MODE_OK;
287 static const struct drm_simple_display_pipe_funcs display_funcs = {
288 .mode_valid = display_mode_valid,
289 .enable = display_enable,
290 .disable = display_disable,
291 .prepare_fb = drm_gem_fb_simple_display_pipe_prepare_fb,
292 .update = display_update,
295 static int display_pipe_init(struct xen_drm_front_drm_info *drm_info,
296 int index, struct xen_drm_front_cfg_connector *cfg,
297 struct xen_drm_front_drm_pipeline *pipeline)
299 struct drm_device *dev = drm_info->drm_dev;
300 const u32 *formats;
301 int format_count;
302 int ret;
304 pipeline->drm_info = drm_info;
305 pipeline->index = index;
306 pipeline->height = cfg->height;
307 pipeline->width = cfg->width;
309 INIT_DELAYED_WORK(&pipeline->pflip_to_worker, pflip_to_worker);
311 ret = xen_drm_front_conn_init(drm_info, &pipeline->conn);
312 if (ret)
313 return ret;
315 formats = xen_drm_front_conn_get_formats(&format_count);
317 return drm_simple_display_pipe_init(dev, &pipeline->pipe,
318 &display_funcs, formats,
319 format_count, NULL,
320 &pipeline->conn);
323 int xen_drm_front_kms_init(struct xen_drm_front_drm_info *drm_info)
325 struct drm_device *dev = drm_info->drm_dev;
326 int i, ret;
328 drm_mode_config_init(dev);
330 dev->mode_config.min_width = 0;
331 dev->mode_config.min_height = 0;
332 dev->mode_config.max_width = 4095;
333 dev->mode_config.max_height = 2047;
334 dev->mode_config.funcs = &mode_config_funcs;
336 for (i = 0; i < drm_info->front_info->cfg.num_connectors; i++) {
337 struct xen_drm_front_cfg_connector *cfg =
338 &drm_info->front_info->cfg.connectors[i];
339 struct xen_drm_front_drm_pipeline *pipeline =
340 &drm_info->pipeline[i];
342 ret = display_pipe_init(drm_info, i, cfg, pipeline);
343 if (ret) {
344 drm_mode_config_cleanup(dev);
345 return ret;
349 drm_mode_config_reset(dev);
350 drm_kms_helper_poll_init(dev);
351 return 0;
354 void xen_drm_front_kms_fini(struct xen_drm_front_drm_info *drm_info)
356 int i;
358 for (i = 0; i < drm_info->front_info->cfg.num_connectors; i++) {
359 struct xen_drm_front_drm_pipeline *pipeline =
360 &drm_info->pipeline[i];
362 cancel_delayed_work_sync(&pipeline->pflip_to_worker);
364 send_pending_event(pipeline);