5 * \author Rickard E. (Rik) Faith <faith@valinux.com>
6 * \author Gareth Hughes <gareth@valinux.com>
10 * Created: Fri Mar 19 14:30:16 1999 by faith@valinux.com
12 * Copyright 1999, 2000 Precision Insight, Inc., Cedar Park, Texas.
13 * Copyright 2000 VA Linux Systems, Inc., Sunnyvale, California.
14 * All Rights Reserved.
16 * Permission is hereby granted, free of charge, to any person obtaining a
17 * copy of this software and associated documentation files (the "Software"),
18 * to deal in the Software without restriction, including without limitation
19 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
20 * and/or sell copies of the Software, and to permit persons to whom the
21 * Software is furnished to do so, subject to the following conditions:
23 * The above copyright notice and this permission notice (including the next
24 * paragraph) shall be included in all copies or substantial portions of the
27 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
28 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
29 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
30 * VA LINUX SYSTEMS AND/OR ITS SUPPLIERS BE LIABLE FOR ANY CLAIM, DAMAGES OR
31 * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
32 * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
33 * OTHER DEALINGS IN THE SOFTWARE.
38 #include <linux/interrupt.h> /* For task queue support */
41 * Get interrupt from bus id.
43 * \param inode device inode.
44 * \param file_priv DRM file private.
46 * \param arg user argument, pointing to a drm_irq_busid structure.
47 * \return zero on success or a negative number on failure.
49 * Finds the PCI device with the specified bus id and gets its IRQ number.
50 * This IOCTL is deprecated, and will now return EINVAL for any busid not equal
51 * to that of the device that this DRM instance attached to.
53 int drm_irq_by_busid(struct drm_device
*dev
, void *data
,
54 struct drm_file
*file_priv
)
56 struct drm_irq_busid
*p
= data
;
58 if (!drm_core_check_feature(dev
, DRIVER_HAVE_IRQ
))
61 if ((p
->busnum
>> 8) != drm_get_pci_domain(dev
) ||
62 (p
->busnum
& 0xff) != dev
->pdev
->bus
->number
||
63 p
->devnum
!= PCI_SLOT(dev
->pdev
->devfn
) || p
->funcnum
!= PCI_FUNC(dev
->pdev
->devfn
))
68 DRM_DEBUG("%d:%d:%d => IRQ %d\n", p
->busnum
, p
->devnum
, p
->funcnum
,
75 * Install IRQ handler.
77 * \param dev DRM device.
78 * \param irq IRQ number.
80 * Initializes the IRQ related data, and setups drm_device::vbl_queue. Installs the handler, calling the driver
81 * \c drm_driver_irq_preinstall() and \c drm_driver_irq_postinstall() functions
82 * before and after the installation.
84 static int drm_irq_install(struct drm_device
* dev
)
87 unsigned long sh_flags
= 0;
89 if (!drm_core_check_feature(dev
, DRIVER_HAVE_IRQ
))
95 mutex_lock(&dev
->struct_mutex
);
97 /* Driver must have been initialized */
98 if (!dev
->dev_private
) {
99 mutex_unlock(&dev
->struct_mutex
);
103 if (dev
->irq_enabled
) {
104 mutex_unlock(&dev
->struct_mutex
);
107 dev
->irq_enabled
= 1;
108 mutex_unlock(&dev
->struct_mutex
);
110 DRM_DEBUG("%s: irq=%d\n", __FUNCTION__
, dev
->irq
);
112 if (drm_core_check_feature(dev
, DRIVER_IRQ_VBL
)) {
113 init_waitqueue_head(&dev
->vbl_queue
);
115 spin_lock_init(&dev
->vbl_lock
);
117 INIT_LIST_HEAD(&dev
->vbl_sigs
);
118 INIT_LIST_HEAD(&dev
->vbl_sigs2
);
120 dev
->vbl_pending
= 0;
123 /* Before installing handler */
124 dev
->driver
->irq_preinstall(dev
);
126 /* Install handler */
127 if (drm_core_check_feature(dev
, DRIVER_IRQ_SHARED
))
128 sh_flags
= IRQF_SHARED
;
130 ret
= request_irq(dev
->irq
, dev
->driver
->irq_handler
,
131 sh_flags
, dev
->devname
, dev
);
133 mutex_lock(&dev
->struct_mutex
);
134 dev
->irq_enabled
= 0;
135 mutex_unlock(&dev
->struct_mutex
);
139 /* After installing handler */
140 dev
->driver
->irq_postinstall(dev
);
146 * Uninstall the IRQ handler.
148 * \param dev DRM device.
150 * Calls the driver's \c drm_driver_irq_uninstall() function, and stops the irq.
152 int drm_irq_uninstall(struct drm_device
* dev
)
156 if (!drm_core_check_feature(dev
, DRIVER_HAVE_IRQ
))
159 mutex_lock(&dev
->struct_mutex
);
160 irq_enabled
= dev
->irq_enabled
;
161 dev
->irq_enabled
= 0;
162 mutex_unlock(&dev
->struct_mutex
);
167 DRM_DEBUG("%s: irq=%d\n", __FUNCTION__
, dev
->irq
);
169 dev
->driver
->irq_uninstall(dev
);
171 free_irq(dev
->irq
, dev
);
173 dev
->locked_tasklet_func
= NULL
;
178 EXPORT_SYMBOL(drm_irq_uninstall
);
183 * \param inode device inode.
184 * \param file_priv DRM file private.
185 * \param cmd command.
186 * \param arg user argument, pointing to a drm_control structure.
187 * \return zero on success or a negative number on failure.
189 * Calls irq_install() or irq_uninstall() according to \p arg.
191 int drm_control(struct drm_device
*dev
, void *data
,
192 struct drm_file
*file_priv
)
194 struct drm_control
*ctl
= data
;
196 /* if we haven't irq we fallback for compatibility reasons - this used to be a separate function in drm_dma.h */
200 case DRM_INST_HANDLER
:
201 if (!drm_core_check_feature(dev
, DRIVER_HAVE_IRQ
))
203 if (dev
->if_version
< DRM_IF_VERSION(1, 2) &&
204 ctl
->irq
!= dev
->irq
)
206 return drm_irq_install(dev
);
207 case DRM_UNINST_HANDLER
:
208 if (!drm_core_check_feature(dev
, DRIVER_HAVE_IRQ
))
210 return drm_irq_uninstall(dev
);
219 * \param inode device inode.
220 * \param file_priv DRM file private.
221 * \param cmd command.
222 * \param data user argument, pointing to a drm_wait_vblank structure.
223 * \return zero on success or a negative number on failure.
225 * Verifies the IRQ is installed.
227 * If a signal is requested checks if this task has already scheduled the same signal
228 * for the same vblank sequence number - nothing to be done in
229 * that case. If the number of tasks waiting for the interrupt exceeds 100 the
230 * function fails. Otherwise adds a new entry to drm_device::vbl_sigs for this
233 * If a signal is not requested, then calls vblank_wait().
235 int drm_wait_vblank(struct drm_device
*dev
, void *data
, struct drm_file
*file_priv
)
237 union drm_wait_vblank
*vblwait
= data
;
240 unsigned int flags
, seq
;
242 if ((!dev
->irq
) || (!dev
->irq_enabled
))
245 if (vblwait
->request
.type
&
246 ~(_DRM_VBLANK_TYPES_MASK
| _DRM_VBLANK_FLAGS_MASK
)) {
247 DRM_ERROR("Unsupported type value 0x%x, supported mask 0x%x\n",
248 vblwait
->request
.type
,
249 (_DRM_VBLANK_TYPES_MASK
| _DRM_VBLANK_FLAGS_MASK
));
253 flags
= vblwait
->request
.type
& _DRM_VBLANK_FLAGS_MASK
;
255 if (!drm_core_check_feature(dev
, (flags
& _DRM_VBLANK_SECONDARY
) ?
256 DRIVER_IRQ_VBL2
: DRIVER_IRQ_VBL
))
259 seq
= atomic_read((flags
& _DRM_VBLANK_SECONDARY
) ? &dev
->vbl_received2
260 : &dev
->vbl_received
);
262 switch (vblwait
->request
.type
& _DRM_VBLANK_TYPES_MASK
) {
263 case _DRM_VBLANK_RELATIVE
:
264 vblwait
->request
.sequence
+= seq
;
265 vblwait
->request
.type
&= ~_DRM_VBLANK_RELATIVE
;
266 case _DRM_VBLANK_ABSOLUTE
:
272 if ((flags
& _DRM_VBLANK_NEXTONMISS
) &&
273 (seq
- vblwait
->request
.sequence
) <= (1<<23)) {
274 vblwait
->request
.sequence
= seq
+ 1;
277 if (flags
& _DRM_VBLANK_SIGNAL
) {
278 unsigned long irqflags
;
279 struct list_head
*vbl_sigs
= (flags
& _DRM_VBLANK_SECONDARY
)
280 ? &dev
->vbl_sigs2
: &dev
->vbl_sigs
;
281 struct drm_vbl_sig
*vbl_sig
;
283 spin_lock_irqsave(&dev
->vbl_lock
, irqflags
);
285 /* Check if this task has already scheduled the same signal
286 * for the same vblank sequence number; nothing to be done in
289 list_for_each_entry(vbl_sig
, vbl_sigs
, head
) {
290 if (vbl_sig
->sequence
== vblwait
->request
.sequence
291 && vbl_sig
->info
.si_signo
==
292 vblwait
->request
.signal
293 && vbl_sig
->task
== current
) {
294 spin_unlock_irqrestore(&dev
->vbl_lock
,
296 vblwait
->reply
.sequence
= seq
;
301 if (dev
->vbl_pending
>= 100) {
302 spin_unlock_irqrestore(&dev
->vbl_lock
, irqflags
);
308 spin_unlock_irqrestore(&dev
->vbl_lock
, irqflags
);
312 drm_alloc(sizeof(struct drm_vbl_sig
), DRM_MEM_DRIVER
))) {
316 memset((void *)vbl_sig
, 0, sizeof(*vbl_sig
));
318 vbl_sig
->sequence
= vblwait
->request
.sequence
;
319 vbl_sig
->info
.si_signo
= vblwait
->request
.signal
;
320 vbl_sig
->task
= current
;
322 spin_lock_irqsave(&dev
->vbl_lock
, irqflags
);
324 list_add_tail(&vbl_sig
->head
, vbl_sigs
);
326 spin_unlock_irqrestore(&dev
->vbl_lock
, irqflags
);
328 vblwait
->reply
.sequence
= seq
;
330 if (flags
& _DRM_VBLANK_SECONDARY
) {
331 if (dev
->driver
->vblank_wait2
)
332 ret
= dev
->driver
->vblank_wait2(dev
, &vblwait
->request
.sequence
);
333 } else if (dev
->driver
->vblank_wait
)
335 dev
->driver
->vblank_wait(dev
,
336 &vblwait
->request
.sequence
);
338 do_gettimeofday(&now
);
339 vblwait
->reply
.tval_sec
= now
.tv_sec
;
340 vblwait
->reply
.tval_usec
= now
.tv_usec
;
348 * Send the VBLANK signals.
350 * \param dev DRM device.
352 * Sends a signal for each task in drm_device::vbl_sigs and empties the list.
354 * If a signal is not requested, then calls vblank_wait().
356 void drm_vbl_send_signals(struct drm_device
* dev
)
361 spin_lock_irqsave(&dev
->vbl_lock
, flags
);
363 for (i
= 0; i
< 2; i
++) {
364 struct drm_vbl_sig
*vbl_sig
, *tmp
;
365 struct list_head
*vbl_sigs
= i
? &dev
->vbl_sigs2
: &dev
->vbl_sigs
;
366 unsigned int vbl_seq
= atomic_read(i
? &dev
->vbl_received2
:
369 list_for_each_entry_safe(vbl_sig
, tmp
, vbl_sigs
, head
) {
370 if ((vbl_seq
- vbl_sig
->sequence
) <= (1 << 23)) {
371 vbl_sig
->info
.si_code
= vbl_seq
;
372 send_sig_info(vbl_sig
->info
.si_signo
,
373 &vbl_sig
->info
, vbl_sig
->task
);
375 list_del(&vbl_sig
->head
);
377 drm_free(vbl_sig
, sizeof(*vbl_sig
),
385 spin_unlock_irqrestore(&dev
->vbl_lock
, flags
);
388 EXPORT_SYMBOL(drm_vbl_send_signals
);
391 * Tasklet wrapper function.
393 * \param data DRM device in disguise.
395 * Attempts to grab the HW lock and calls the driver callback on success. On
396 * failure, leave the lock marked as contended so the callback can be called
399 static void drm_locked_tasklet_func(unsigned long data
)
401 struct drm_device
*dev
= (struct drm_device
*)data
;
402 unsigned long irqflags
;
404 spin_lock_irqsave(&dev
->tasklet_lock
, irqflags
);
406 if (!dev
->locked_tasklet_func
||
407 !drm_lock_take(&dev
->lock
,
408 DRM_KERNEL_CONTEXT
)) {
409 spin_unlock_irqrestore(&dev
->tasklet_lock
, irqflags
);
413 dev
->lock
.lock_time
= jiffies
;
414 atomic_inc(&dev
->counts
[_DRM_STAT_LOCKS
]);
416 dev
->locked_tasklet_func(dev
);
418 drm_lock_free(&dev
->lock
,
421 dev
->locked_tasklet_func
= NULL
;
423 spin_unlock_irqrestore(&dev
->tasklet_lock
, irqflags
);
427 * Schedule a tasklet to call back a driver hook with the HW lock held.
429 * \param dev DRM device.
430 * \param func Driver callback.
432 * This is intended for triggering actions that require the HW lock from an
433 * interrupt handler. The lock will be grabbed ASAP after the interrupt handler
434 * completes. Note that the callback may be called from interrupt or process
435 * context, it must not make any assumptions about this. Also, the HW lock will
436 * be held with the kernel context or any client context.
438 void drm_locked_tasklet(struct drm_device
*dev
, void (*func
)(struct drm_device
*))
440 unsigned long irqflags
;
441 static DECLARE_TASKLET(drm_tasklet
, drm_locked_tasklet_func
, 0);
443 if (!drm_core_check_feature(dev
, DRIVER_HAVE_IRQ
) ||
444 test_bit(TASKLET_STATE_SCHED
, &drm_tasklet
.state
))
447 spin_lock_irqsave(&dev
->tasklet_lock
, irqflags
);
449 if (dev
->locked_tasklet_func
) {
450 spin_unlock_irqrestore(&dev
->tasklet_lock
, irqflags
);
454 dev
->locked_tasklet_func
= func
;
456 spin_unlock_irqrestore(&dev
->tasklet_lock
, irqflags
);
458 drm_tasklet
.data
= (unsigned long)dev
;
460 tasklet_hi_schedule(&drm_tasklet
);
462 EXPORT_SYMBOL(drm_locked_tasklet
);