Merge branch 'for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tiwai/sound-2.6
[deliverable/linux.git] / drivers / gpu / drm / drm_irq.c
1 /**
2 * \file drm_irq.c
3 * IRQ support
4 *
5 * \author Rickard E. (Rik) Faith <faith@valinux.com>
6 * \author Gareth Hughes <gareth@valinux.com>
7 */
8
9 /*
10 * Created: Fri Mar 19 14:30:16 1999 by faith@valinux.com
11 *
12 * Copyright 1999, 2000 Precision Insight, Inc., Cedar Park, Texas.
13 * Copyright 2000 VA Linux Systems, Inc., Sunnyvale, California.
14 * All Rights Reserved.
15 *
16 * Permission is hereby granted, free of charge, to any person obtaining a
17 * copy of this software and associated documentation files (the "Software"),
18 * to deal in the Software without restriction, including without limitation
19 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
20 * and/or sell copies of the Software, and to permit persons to whom the
21 * Software is furnished to do so, subject to the following conditions:
22 *
23 * The above copyright notice and this permission notice (including the next
24 * paragraph) shall be included in all copies or substantial portions of the
25 * Software.
26 *
27 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
28 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
29 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
30 * VA LINUX SYSTEMS AND/OR ITS SUPPLIERS BE LIABLE FOR ANY CLAIM, DAMAGES OR
31 * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
32 * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
33 * OTHER DEALINGS IN THE SOFTWARE.
34 */
35
36 #include "drmP.h"
37
38 #include <linux/interrupt.h> /* For task queue support */
39
40 /**
41 * Get interrupt from bus id.
42 *
43 * \param inode device inode.
44 * \param file_priv DRM file private.
45 * \param cmd command.
46 * \param arg user argument, pointing to a drm_irq_busid structure.
47 * \return zero on success or a negative number on failure.
48 *
49 * Finds the PCI device with the specified bus id and gets its IRQ number.
50 * This IOCTL is deprecated, and will now return EINVAL for any busid not equal
51 * to that of the device that this DRM instance attached to.
52 */
53 int drm_irq_by_busid(struct drm_device *dev, void *data,
54 struct drm_file *file_priv)
55 {
56 struct drm_irq_busid *p = data;
57
58 if (!drm_core_check_feature(dev, DRIVER_HAVE_IRQ))
59 return -EINVAL;
60
61 if ((p->busnum >> 8) != drm_get_pci_domain(dev) ||
62 (p->busnum & 0xff) != dev->pdev->bus->number ||
63 p->devnum != PCI_SLOT(dev->pdev->devfn) || p->funcnum != PCI_FUNC(dev->pdev->devfn))
64 return -EINVAL;
65
66 p->irq = dev->pdev->irq;
67
68 DRM_DEBUG("%d:%d:%d => IRQ %d\n", p->busnum, p->devnum, p->funcnum,
69 p->irq);
70
71 return 0;
72 }
73
74 static void vblank_disable_fn(unsigned long arg)
75 {
76 struct drm_device *dev = (struct drm_device *)arg;
77 unsigned long irqflags;
78 int i;
79
80 if (!dev->vblank_disable_allowed)
81 return;
82
83 for (i = 0; i < dev->num_crtcs; i++) {
84 spin_lock_irqsave(&dev->vbl_lock, irqflags);
85 if (atomic_read(&dev->vblank_refcount[i]) == 0 &&
86 dev->vblank_enabled[i]) {
87 DRM_DEBUG("disabling vblank on crtc %d\n", i);
88 dev->last_vblank[i] =
89 dev->driver->get_vblank_counter(dev, i);
90 dev->driver->disable_vblank(dev, i);
91 dev->vblank_enabled[i] = 0;
92 }
93 spin_unlock_irqrestore(&dev->vbl_lock, irqflags);
94 }
95 }
96
97 static void drm_vblank_cleanup(struct drm_device *dev)
98 {
99 /* Bail if the driver didn't call drm_vblank_init() */
100 if (dev->num_crtcs == 0)
101 return;
102
103 del_timer(&dev->vblank_disable_timer);
104
105 vblank_disable_fn((unsigned long)dev);
106
107 drm_free(dev->vbl_queue, sizeof(*dev->vbl_queue) * dev->num_crtcs,
108 DRM_MEM_DRIVER);
109 drm_free(dev->vbl_sigs, sizeof(*dev->vbl_sigs) * dev->num_crtcs,
110 DRM_MEM_DRIVER);
111 drm_free(dev->_vblank_count, sizeof(*dev->_vblank_count) *
112 dev->num_crtcs, DRM_MEM_DRIVER);
113 drm_free(dev->vblank_refcount, sizeof(*dev->vblank_refcount) *
114 dev->num_crtcs, DRM_MEM_DRIVER);
115 drm_free(dev->vblank_enabled, sizeof(*dev->vblank_enabled) *
116 dev->num_crtcs, DRM_MEM_DRIVER);
117 drm_free(dev->last_vblank, sizeof(*dev->last_vblank) * dev->num_crtcs,
118 DRM_MEM_DRIVER);
119 drm_free(dev->vblank_inmodeset, sizeof(*dev->vblank_inmodeset) *
120 dev->num_crtcs, DRM_MEM_DRIVER);
121
122 dev->num_crtcs = 0;
123 }
124
125 int drm_vblank_init(struct drm_device *dev, int num_crtcs)
126 {
127 int i, ret = -ENOMEM;
128
129 setup_timer(&dev->vblank_disable_timer, vblank_disable_fn,
130 (unsigned long)dev);
131 spin_lock_init(&dev->vbl_lock);
132 atomic_set(&dev->vbl_signal_pending, 0);
133 dev->num_crtcs = num_crtcs;
134
135 dev->vbl_queue = drm_alloc(sizeof(wait_queue_head_t) * num_crtcs,
136 DRM_MEM_DRIVER);
137 if (!dev->vbl_queue)
138 goto err;
139
140 dev->vbl_sigs = drm_alloc(sizeof(struct list_head) * num_crtcs,
141 DRM_MEM_DRIVER);
142 if (!dev->vbl_sigs)
143 goto err;
144
145 dev->_vblank_count = drm_alloc(sizeof(atomic_t) * num_crtcs,
146 DRM_MEM_DRIVER);
147 if (!dev->_vblank_count)
148 goto err;
149
150 dev->vblank_refcount = drm_alloc(sizeof(atomic_t) * num_crtcs,
151 DRM_MEM_DRIVER);
152 if (!dev->vblank_refcount)
153 goto err;
154
155 dev->vblank_enabled = drm_calloc(num_crtcs, sizeof(int),
156 DRM_MEM_DRIVER);
157 if (!dev->vblank_enabled)
158 goto err;
159
160 dev->last_vblank = drm_calloc(num_crtcs, sizeof(u32), DRM_MEM_DRIVER);
161 if (!dev->last_vblank)
162 goto err;
163
164 dev->vblank_inmodeset = drm_calloc(num_crtcs, sizeof(int),
165 DRM_MEM_DRIVER);
166 if (!dev->vblank_inmodeset)
167 goto err;
168
169 /* Zero per-crtc vblank stuff */
170 for (i = 0; i < num_crtcs; i++) {
171 init_waitqueue_head(&dev->vbl_queue[i]);
172 INIT_LIST_HEAD(&dev->vbl_sigs[i]);
173 atomic_set(&dev->_vblank_count[i], 0);
174 atomic_set(&dev->vblank_refcount[i], 0);
175 }
176
177 dev->vblank_disable_allowed = 0;
178
179 return 0;
180
181 err:
182 drm_vblank_cleanup(dev);
183 return ret;
184 }
185 EXPORT_SYMBOL(drm_vblank_init);
186
187 /**
188 * Install IRQ handler.
189 *
190 * \param dev DRM device.
191 *
192 * Initializes the IRQ related data. Installs the handler, calling the driver
193 * \c drm_driver_irq_preinstall() and \c drm_driver_irq_postinstall() functions
194 * before and after the installation.
195 */
196 int drm_irq_install(struct drm_device *dev)
197 {
198 int ret = 0;
199 unsigned long sh_flags = 0;
200
201 if (!drm_core_check_feature(dev, DRIVER_HAVE_IRQ))
202 return -EINVAL;
203
204 if (dev->pdev->irq == 0)
205 return -EINVAL;
206
207 mutex_lock(&dev->struct_mutex);
208
209 /* Driver must have been initialized */
210 if (!dev->dev_private) {
211 mutex_unlock(&dev->struct_mutex);
212 return -EINVAL;
213 }
214
215 if (dev->irq_enabled) {
216 mutex_unlock(&dev->struct_mutex);
217 return -EBUSY;
218 }
219 dev->irq_enabled = 1;
220 mutex_unlock(&dev->struct_mutex);
221
222 DRM_DEBUG("irq=%d\n", dev->pdev->irq);
223
224 /* Before installing handler */
225 dev->driver->irq_preinstall(dev);
226
227 /* Install handler */
228 if (drm_core_check_feature(dev, DRIVER_IRQ_SHARED))
229 sh_flags = IRQF_SHARED;
230
231 ret = request_irq(drm_dev_to_irq(dev), dev->driver->irq_handler,
232 sh_flags, dev->devname, dev);
233
234 if (ret < 0) {
235 mutex_lock(&dev->struct_mutex);
236 dev->irq_enabled = 0;
237 mutex_unlock(&dev->struct_mutex);
238 return ret;
239 }
240
241 /* After installing handler */
242 ret = dev->driver->irq_postinstall(dev);
243 if (ret < 0) {
244 mutex_lock(&dev->struct_mutex);
245 dev->irq_enabled = 0;
246 mutex_unlock(&dev->struct_mutex);
247 }
248
249 return ret;
250 }
251 EXPORT_SYMBOL(drm_irq_install);
252
253 /**
254 * Uninstall the IRQ handler.
255 *
256 * \param dev DRM device.
257 *
258 * Calls the driver's \c drm_driver_irq_uninstall() function, and stops the irq.
259 */
260 int drm_irq_uninstall(struct drm_device * dev)
261 {
262 int irq_enabled;
263
264 if (!drm_core_check_feature(dev, DRIVER_HAVE_IRQ))
265 return -EINVAL;
266
267 mutex_lock(&dev->struct_mutex);
268 irq_enabled = dev->irq_enabled;
269 dev->irq_enabled = 0;
270 mutex_unlock(&dev->struct_mutex);
271
272 if (!irq_enabled)
273 return -EINVAL;
274
275 DRM_DEBUG("irq=%d\n", dev->pdev->irq);
276
277 dev->driver->irq_uninstall(dev);
278
279 free_irq(dev->pdev->irq, dev);
280
281 drm_vblank_cleanup(dev);
282
283 dev->locked_tasklet_func = NULL;
284
285 return 0;
286 }
287 EXPORT_SYMBOL(drm_irq_uninstall);
288
289 /**
290 * IRQ control ioctl.
291 *
292 * \param inode device inode.
293 * \param file_priv DRM file private.
294 * \param cmd command.
295 * \param arg user argument, pointing to a drm_control structure.
296 * \return zero on success or a negative number on failure.
297 *
298 * Calls irq_install() or irq_uninstall() according to \p arg.
299 */
300 int drm_control(struct drm_device *dev, void *data,
301 struct drm_file *file_priv)
302 {
303 struct drm_control *ctl = data;
304
305 /* if we haven't irq we fallback for compatibility reasons - this used to be a separate function in drm_dma.h */
306
307
308 switch (ctl->func) {
309 case DRM_INST_HANDLER:
310 if (!drm_core_check_feature(dev, DRIVER_HAVE_IRQ))
311 return 0;
312 if (dev->if_version < DRM_IF_VERSION(1, 2) &&
313 ctl->irq != dev->pdev->irq)
314 return -EINVAL;
315 return drm_irq_install(dev);
316 case DRM_UNINST_HANDLER:
317 if (!drm_core_check_feature(dev, DRIVER_HAVE_IRQ))
318 return 0;
319 return drm_irq_uninstall(dev);
320 default:
321 return -EINVAL;
322 }
323 }
324
325 /**
326 * drm_vblank_count - retrieve "cooked" vblank counter value
327 * @dev: DRM device
328 * @crtc: which counter to retrieve
329 *
330 * Fetches the "cooked" vblank count value that represents the number of
331 * vblank events since the system was booted, including lost events due to
332 * modesetting activity.
333 */
334 u32 drm_vblank_count(struct drm_device *dev, int crtc)
335 {
336 return atomic_read(&dev->_vblank_count[crtc]);
337 }
338 EXPORT_SYMBOL(drm_vblank_count);
339
340 /**
341 * drm_update_vblank_count - update the master vblank counter
342 * @dev: DRM device
343 * @crtc: counter to update
344 *
345 * Call back into the driver to update the appropriate vblank counter
346 * (specified by @crtc). Deal with wraparound, if it occurred, and
347 * update the last read value so we can deal with wraparound on the next
348 * call if necessary.
349 *
350 * Only necessary when going from off->on, to account for frames we
351 * didn't get an interrupt for.
352 *
353 * Note: caller must hold dev->vbl_lock since this reads & writes
354 * device vblank fields.
355 */
356 static void drm_update_vblank_count(struct drm_device *dev, int crtc)
357 {
358 u32 cur_vblank, diff;
359
360 /*
361 * Interrupts were disabled prior to this call, so deal with counter
362 * wrap if needed.
363 * NOTE! It's possible we lost a full dev->max_vblank_count events
364 * here if the register is small or we had vblank interrupts off for
365 * a long time.
366 */
367 cur_vblank = dev->driver->get_vblank_counter(dev, crtc);
368 diff = cur_vblank - dev->last_vblank[crtc];
369 if (cur_vblank < dev->last_vblank[crtc]) {
370 diff += dev->max_vblank_count;
371
372 DRM_DEBUG("last_vblank[%d]=0x%x, cur_vblank=0x%x => diff=0x%x\n",
373 crtc, dev->last_vblank[crtc], cur_vblank, diff);
374 }
375
376 DRM_DEBUG("enabling vblank interrupts on crtc %d, missed %d\n",
377 crtc, diff);
378
379 atomic_add(diff, &dev->_vblank_count[crtc]);
380 }
381
382 /**
383 * drm_vblank_get - get a reference count on vblank events
384 * @dev: DRM device
385 * @crtc: which CRTC to own
386 *
387 * Acquire a reference count on vblank events to avoid having them disabled
388 * while in use.
389 *
390 * RETURNS
391 * Zero on success, nonzero on failure.
392 */
393 int drm_vblank_get(struct drm_device *dev, int crtc)
394 {
395 unsigned long irqflags;
396 int ret = 0;
397
398 spin_lock_irqsave(&dev->vbl_lock, irqflags);
399 /* Going from 0->1 means we have to enable interrupts again */
400 if (atomic_add_return(1, &dev->vblank_refcount[crtc]) == 1 &&
401 !dev->vblank_enabled[crtc]) {
402 ret = dev->driver->enable_vblank(dev, crtc);
403 DRM_DEBUG("enabling vblank on crtc %d, ret: %d\n", crtc, ret);
404 if (ret)
405 atomic_dec(&dev->vblank_refcount[crtc]);
406 else {
407 dev->vblank_enabled[crtc] = 1;
408 drm_update_vblank_count(dev, crtc);
409 }
410 }
411 spin_unlock_irqrestore(&dev->vbl_lock, irqflags);
412
413 return ret;
414 }
415 EXPORT_SYMBOL(drm_vblank_get);
416
417 /**
418 * drm_vblank_put - give up ownership of vblank events
419 * @dev: DRM device
420 * @crtc: which counter to give up
421 *
422 * Release ownership of a given vblank counter, turning off interrupts
423 * if possible.
424 */
425 void drm_vblank_put(struct drm_device *dev, int crtc)
426 {
427 /* Last user schedules interrupt disable */
428 if (atomic_dec_and_test(&dev->vblank_refcount[crtc]))
429 mod_timer(&dev->vblank_disable_timer, jiffies + 5*DRM_HZ);
430 }
431 EXPORT_SYMBOL(drm_vblank_put);
432
433 /**
434 * drm_modeset_ctl - handle vblank event counter changes across mode switch
435 * @DRM_IOCTL_ARGS: standard ioctl arguments
436 *
437 * Applications should call the %_DRM_PRE_MODESET and %_DRM_POST_MODESET
438 * ioctls around modesetting so that any lost vblank events are accounted for.
439 *
440 * Generally the counter will reset across mode sets. If interrupts are
441 * enabled around this call, we don't have to do anything since the counter
442 * will have already been incremented.
443 */
444 int drm_modeset_ctl(struct drm_device *dev, void *data,
445 struct drm_file *file_priv)
446 {
447 struct drm_modeset_ctl *modeset = data;
448 unsigned long irqflags;
449 int crtc, ret = 0;
450
451 /* If drm_vblank_init() hasn't been called yet, just no-op */
452 if (!dev->num_crtcs)
453 goto out;
454
455 crtc = modeset->crtc;
456 if (crtc >= dev->num_crtcs) {
457 ret = -EINVAL;
458 goto out;
459 }
460
461 /*
462 * To avoid all the problems that might happen if interrupts
463 * were enabled/disabled around or between these calls, we just
464 * have the kernel take a reference on the CRTC (just once though
465 * to avoid corrupting the count if multiple, mismatch calls occur),
466 * so that interrupts remain enabled in the interim.
467 */
468 switch (modeset->cmd) {
469 case _DRM_PRE_MODESET:
470 if (!dev->vblank_inmodeset[crtc]) {
471 dev->vblank_inmodeset[crtc] = 1;
472 drm_vblank_get(dev, crtc);
473 }
474 break;
475 case _DRM_POST_MODESET:
476 if (dev->vblank_inmodeset[crtc]) {
477 spin_lock_irqsave(&dev->vbl_lock, irqflags);
478 dev->vblank_disable_allowed = 1;
479 dev->vblank_inmodeset[crtc] = 0;
480 spin_unlock_irqrestore(&dev->vbl_lock, irqflags);
481 drm_vblank_put(dev, crtc);
482 }
483 break;
484 default:
485 ret = -EINVAL;
486 break;
487 }
488
489 out:
490 return ret;
491 }
492
493 /**
494 * Wait for VBLANK.
495 *
496 * \param inode device inode.
497 * \param file_priv DRM file private.
498 * \param cmd command.
499 * \param data user argument, pointing to a drm_wait_vblank structure.
500 * \return zero on success or a negative number on failure.
501 *
502 * Verifies the IRQ is installed.
503 *
504 * If a signal is requested checks if this task has already scheduled the same signal
505 * for the same vblank sequence number - nothing to be done in
506 * that case. If the number of tasks waiting for the interrupt exceeds 100 the
507 * function fails. Otherwise adds a new entry to drm_device::vbl_sigs for this
508 * task.
509 *
510 * If a signal is not requested, then calls vblank_wait().
511 */
512 int drm_wait_vblank(struct drm_device *dev, void *data,
513 struct drm_file *file_priv)
514 {
515 union drm_wait_vblank *vblwait = data;
516 int ret = 0;
517 unsigned int flags, seq, crtc;
518
519 if ((!dev->pdev->irq) || (!dev->irq_enabled))
520 return -EINVAL;
521
522 if (vblwait->request.type &
523 ~(_DRM_VBLANK_TYPES_MASK | _DRM_VBLANK_FLAGS_MASK)) {
524 DRM_ERROR("Unsupported type value 0x%x, supported mask 0x%x\n",
525 vblwait->request.type,
526 (_DRM_VBLANK_TYPES_MASK | _DRM_VBLANK_FLAGS_MASK));
527 return -EINVAL;
528 }
529
530 flags = vblwait->request.type & _DRM_VBLANK_FLAGS_MASK;
531 crtc = flags & _DRM_VBLANK_SECONDARY ? 1 : 0;
532
533 if (crtc >= dev->num_crtcs)
534 return -EINVAL;
535
536 ret = drm_vblank_get(dev, crtc);
537 if (ret) {
538 DRM_ERROR("failed to acquire vblank counter, %d\n", ret);
539 return ret;
540 }
541 seq = drm_vblank_count(dev, crtc);
542
543 switch (vblwait->request.type & _DRM_VBLANK_TYPES_MASK) {
544 case _DRM_VBLANK_RELATIVE:
545 vblwait->request.sequence += seq;
546 vblwait->request.type &= ~_DRM_VBLANK_RELATIVE;
547 case _DRM_VBLANK_ABSOLUTE:
548 break;
549 default:
550 ret = -EINVAL;
551 goto done;
552 }
553
554 if ((flags & _DRM_VBLANK_NEXTONMISS) &&
555 (seq - vblwait->request.sequence) <= (1<<23)) {
556 vblwait->request.sequence = seq + 1;
557 }
558
559 if (flags & _DRM_VBLANK_SIGNAL) {
560 unsigned long irqflags;
561 struct list_head *vbl_sigs = &dev->vbl_sigs[crtc];
562 struct drm_vbl_sig *vbl_sig;
563
564 spin_lock_irqsave(&dev->vbl_lock, irqflags);
565
566 /* Check if this task has already scheduled the same signal
567 * for the same vblank sequence number; nothing to be done in
568 * that case
569 */
570 list_for_each_entry(vbl_sig, vbl_sigs, head) {
571 if (vbl_sig->sequence == vblwait->request.sequence
572 && vbl_sig->info.si_signo ==
573 vblwait->request.signal
574 && vbl_sig->task == current) {
575 spin_unlock_irqrestore(&dev->vbl_lock,
576 irqflags);
577 vblwait->reply.sequence = seq;
578 goto done;
579 }
580 }
581
582 if (atomic_read(&dev->vbl_signal_pending) >= 100) {
583 spin_unlock_irqrestore(&dev->vbl_lock, irqflags);
584 ret = -EBUSY;
585 goto done;
586 }
587
588 spin_unlock_irqrestore(&dev->vbl_lock, irqflags);
589
590 vbl_sig = drm_calloc(1, sizeof(struct drm_vbl_sig),
591 DRM_MEM_DRIVER);
592 if (!vbl_sig) {
593 ret = -ENOMEM;
594 goto done;
595 }
596
597 /* Get a refcount on the vblank, which will be released by
598 * drm_vbl_send_signals().
599 */
600 ret = drm_vblank_get(dev, crtc);
601 if (ret) {
602 drm_free(vbl_sig, sizeof(struct drm_vbl_sig),
603 DRM_MEM_DRIVER);
604 goto done;
605 }
606
607 atomic_inc(&dev->vbl_signal_pending);
608
609 vbl_sig->sequence = vblwait->request.sequence;
610 vbl_sig->info.si_signo = vblwait->request.signal;
611 vbl_sig->task = current;
612
613 spin_lock_irqsave(&dev->vbl_lock, irqflags);
614
615 list_add_tail(&vbl_sig->head, vbl_sigs);
616
617 spin_unlock_irqrestore(&dev->vbl_lock, irqflags);
618
619 vblwait->reply.sequence = seq;
620 } else {
621 DRM_DEBUG("waiting on vblank count %d, crtc %d\n",
622 vblwait->request.sequence, crtc);
623 DRM_WAIT_ON(ret, dev->vbl_queue[crtc], 3 * DRM_HZ,
624 ((drm_vblank_count(dev, crtc)
625 - vblwait->request.sequence) <= (1 << 23)));
626
627 if (ret != -EINTR) {
628 struct timeval now;
629
630 do_gettimeofday(&now);
631
632 vblwait->reply.tval_sec = now.tv_sec;
633 vblwait->reply.tval_usec = now.tv_usec;
634 vblwait->reply.sequence = drm_vblank_count(dev, crtc);
635 DRM_DEBUG("returning %d to client\n",
636 vblwait->reply.sequence);
637 } else {
638 DRM_DEBUG("vblank wait interrupted by signal\n");
639 }
640 }
641
642 done:
643 drm_vblank_put(dev, crtc);
644 return ret;
645 }
646
647 /**
648 * Send the VBLANK signals.
649 *
650 * \param dev DRM device.
651 * \param crtc CRTC where the vblank event occurred
652 *
653 * Sends a signal for each task in drm_device::vbl_sigs and empties the list.
654 *
655 * If a signal is not requested, then calls vblank_wait().
656 */
657 static void drm_vbl_send_signals(struct drm_device *dev, int crtc)
658 {
659 struct drm_vbl_sig *vbl_sig, *tmp;
660 struct list_head *vbl_sigs;
661 unsigned int vbl_seq;
662 unsigned long flags;
663
664 spin_lock_irqsave(&dev->vbl_lock, flags);
665
666 vbl_sigs = &dev->vbl_sigs[crtc];
667 vbl_seq = drm_vblank_count(dev, crtc);
668
669 list_for_each_entry_safe(vbl_sig, tmp, vbl_sigs, head) {
670 if ((vbl_seq - vbl_sig->sequence) <= (1 << 23)) {
671 vbl_sig->info.si_code = vbl_seq;
672 send_sig_info(vbl_sig->info.si_signo,
673 &vbl_sig->info, vbl_sig->task);
674
675 list_del(&vbl_sig->head);
676
677 drm_free(vbl_sig, sizeof(*vbl_sig),
678 DRM_MEM_DRIVER);
679 atomic_dec(&dev->vbl_signal_pending);
680 drm_vblank_put(dev, crtc);
681 }
682 }
683
684 spin_unlock_irqrestore(&dev->vbl_lock, flags);
685 }
686
687 /**
688 * drm_handle_vblank - handle a vblank event
689 * @dev: DRM device
690 * @crtc: where this event occurred
691 *
692 * Drivers should call this routine in their vblank interrupt handlers to
693 * update the vblank counter and send any signals that may be pending.
694 */
695 void drm_handle_vblank(struct drm_device *dev, int crtc)
696 {
697 atomic_inc(&dev->_vblank_count[crtc]);
698 DRM_WAKEUP(&dev->vbl_queue[crtc]);
699 drm_vbl_send_signals(dev, crtc);
700 }
701 EXPORT_SYMBOL(drm_handle_vblank);
702
703 /**
704 * Tasklet wrapper function.
705 *
706 * \param data DRM device in disguise.
707 *
708 * Attempts to grab the HW lock and calls the driver callback on success. On
709 * failure, leave the lock marked as contended so the callback can be called
710 * from drm_unlock().
711 */
712 static void drm_locked_tasklet_func(unsigned long data)
713 {
714 struct drm_device *dev = (struct drm_device *)data;
715 unsigned long irqflags;
716 void (*tasklet_func)(struct drm_device *);
717
718 spin_lock_irqsave(&dev->tasklet_lock, irqflags);
719 tasklet_func = dev->locked_tasklet_func;
720 spin_unlock_irqrestore(&dev->tasklet_lock, irqflags);
721
722 if (!tasklet_func ||
723 !drm_lock_take(&dev->lock,
724 DRM_KERNEL_CONTEXT)) {
725 return;
726 }
727
728 dev->lock.lock_time = jiffies;
729 atomic_inc(&dev->counts[_DRM_STAT_LOCKS]);
730
731 spin_lock_irqsave(&dev->tasklet_lock, irqflags);
732 tasklet_func = dev->locked_tasklet_func;
733 dev->locked_tasklet_func = NULL;
734 spin_unlock_irqrestore(&dev->tasklet_lock, irqflags);
735
736 if (tasklet_func != NULL)
737 tasklet_func(dev);
738
739 drm_lock_free(&dev->lock,
740 DRM_KERNEL_CONTEXT);
741 }
742
743 /**
744 * Schedule a tasklet to call back a driver hook with the HW lock held.
745 *
746 * \param dev DRM device.
747 * \param func Driver callback.
748 *
749 * This is intended for triggering actions that require the HW lock from an
750 * interrupt handler. The lock will be grabbed ASAP after the interrupt handler
751 * completes. Note that the callback may be called from interrupt or process
752 * context, it must not make any assumptions about this. Also, the HW lock will
753 * be held with the kernel context or any client context.
754 */
755 void drm_locked_tasklet(struct drm_device *dev, void (*func)(struct drm_device *))
756 {
757 unsigned long irqflags;
758 static DECLARE_TASKLET(drm_tasklet, drm_locked_tasklet_func, 0);
759
760 if (!drm_core_check_feature(dev, DRIVER_HAVE_IRQ) ||
761 test_bit(TASKLET_STATE_SCHED, &drm_tasklet.state))
762 return;
763
764 spin_lock_irqsave(&dev->tasklet_lock, irqflags);
765
766 if (dev->locked_tasklet_func) {
767 spin_unlock_irqrestore(&dev->tasklet_lock, irqflags);
768 return;
769 }
770
771 dev->locked_tasklet_func = func;
772
773 spin_unlock_irqrestore(&dev->tasklet_lock, irqflags);
774
775 drm_tasklet.data = (unsigned long)dev;
776
777 tasklet_hi_schedule(&drm_tasklet);
778 }
779 EXPORT_SYMBOL(drm_locked_tasklet);
This page took 0.046559 seconds and 6 git commands to generate.