[ARM] pxa: Gumstix Verdex PCMCIA support
[linux-2.6/verdex.git] / drivers / media / video / ivtv / ivtv-irq.c
blobcd9db0bf33bf078946d08307a2414b901c9c1cde
1 /* interrupt handling
2 Copyright (C) 2003-2004 Kevin Thayer <nufan_wfk at yahoo.com>
3 Copyright (C) 2004 Chris Kennedy <c@groovy.org>
4 Copyright (C) 2005-2007 Hans Verkuil <hverkuil@xs4all.nl>
6 This program is free software; you can redistribute it and/or modify
7 it under the terms of the GNU General Public License as published by
8 the Free Software Foundation; either version 2 of the License, or
9 (at your option) any later version.
11 This program is distributed in the hope that it will be useful,
12 but WITHOUT ANY WARRANTY; without even the implied warranty of
13 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
14 GNU General Public License for more details.
16 You should have received a copy of the GNU General Public License
17 along with this program; if not, write to the Free Software
18 Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
21 #include "ivtv-driver.h"
22 #include "ivtv-queue.h"
23 #include "ivtv-udma.h"
24 #include "ivtv-irq.h"
25 #include "ivtv-mailbox.h"
26 #include "ivtv-vbi.h"
27 #include "ivtv-yuv.h"
29 #define DMA_MAGIC_COOKIE 0x000001fe
31 static void ivtv_dma_dec_start(struct ivtv_stream *s);
33 static const int ivtv_stream_map[] = {
34 IVTV_ENC_STREAM_TYPE_MPG,
35 IVTV_ENC_STREAM_TYPE_YUV,
36 IVTV_ENC_STREAM_TYPE_PCM,
37 IVTV_ENC_STREAM_TYPE_VBI,
41 static void ivtv_pio_work_handler(struct ivtv *itv)
43 struct ivtv_stream *s = &itv->streams[itv->cur_pio_stream];
44 struct ivtv_buffer *buf;
45 int i = 0;
47 IVTV_DEBUG_HI_DMA("ivtv_pio_work_handler\n");
48 if (itv->cur_pio_stream < 0 || itv->cur_pio_stream >= IVTV_MAX_STREAMS ||
49 s->vdev == NULL || !ivtv_use_pio(s)) {
50 itv->cur_pio_stream = -1;
51 /* trigger PIO complete user interrupt */
52 write_reg(IVTV_IRQ_ENC_PIO_COMPLETE, 0x44);
53 return;
55 IVTV_DEBUG_HI_DMA("Process PIO %s\n", s->name);
56 list_for_each_entry(buf, &s->q_dma.list, list) {
57 u32 size = s->sg_processing[i].size & 0x3ffff;
59 /* Copy the data from the card to the buffer */
60 if (s->type == IVTV_DEC_STREAM_TYPE_VBI) {
61 memcpy_fromio(buf->buf, itv->dec_mem + s->sg_processing[i].src - IVTV_DECODER_OFFSET, size);
63 else {
64 memcpy_fromio(buf->buf, itv->enc_mem + s->sg_processing[i].src, size);
66 i++;
67 if (i == s->sg_processing_size)
68 break;
70 write_reg(IVTV_IRQ_ENC_PIO_COMPLETE, 0x44);
73 void ivtv_irq_work_handler(struct work_struct *work)
75 struct ivtv *itv = container_of(work, struct ivtv, irq_work_queue);
77 DEFINE_WAIT(wait);
79 if (test_and_clear_bit(IVTV_F_I_WORK_INITED, &itv->i_flags)) {
80 struct sched_param param = { .sched_priority = 99 };
82 /* This thread must use the FIFO scheduler as it
83 is realtime sensitive. */
84 sched_setscheduler(current, SCHED_FIFO, &param);
86 if (test_and_clear_bit(IVTV_F_I_WORK_HANDLER_PIO, &itv->i_flags))
87 ivtv_pio_work_handler(itv);
89 if (test_and_clear_bit(IVTV_F_I_WORK_HANDLER_VBI, &itv->i_flags))
90 ivtv_vbi_work_handler(itv);
92 if (test_and_clear_bit(IVTV_F_I_WORK_HANDLER_YUV, &itv->i_flags))
93 ivtv_yuv_work_handler(itv);
96 /* Determine the required DMA size, setup enough buffers in the predma queue and
97 actually copy the data from the card to the buffers in case a PIO transfer is
98 required for this stream.
100 static int stream_enc_dma_append(struct ivtv_stream *s, u32 data[CX2341X_MBOX_MAX_DATA])
102 struct ivtv *itv = s->itv;
103 struct ivtv_buffer *buf;
104 u32 bytes_needed = 0;
105 u32 offset, size;
106 u32 UVoffset = 0, UVsize = 0;
107 int skip_bufs = s->q_predma.buffers;
108 int idx = s->sg_pending_size;
109 int rc;
111 /* sanity checks */
112 if (s->vdev == NULL) {
113 IVTV_DEBUG_WARN("Stream %s not started\n", s->name);
114 return -1;
116 if (!test_bit(IVTV_F_S_CLAIMED, &s->s_flags)) {
117 IVTV_DEBUG_WARN("Stream %s not open\n", s->name);
118 return -1;
121 /* determine offset, size and PTS for the various streams */
122 switch (s->type) {
123 case IVTV_ENC_STREAM_TYPE_MPG:
124 offset = data[1];
125 size = data[2];
126 s->pending_pts = 0;
127 break;
129 case IVTV_ENC_STREAM_TYPE_YUV:
130 offset = data[1];
131 size = data[2];
132 UVoffset = data[3];
133 UVsize = data[4];
134 s->pending_pts = ((u64) data[5] << 32) | data[6];
135 break;
137 case IVTV_ENC_STREAM_TYPE_PCM:
138 offset = data[1] + 12;
139 size = data[2] - 12;
140 s->pending_pts = read_dec(offset - 8) |
141 ((u64)(read_dec(offset - 12)) << 32);
142 if (itv->has_cx23415)
143 offset += IVTV_DECODER_OFFSET;
144 break;
146 case IVTV_ENC_STREAM_TYPE_VBI:
147 size = itv->vbi.enc_size * itv->vbi.fpi;
148 offset = read_enc(itv->vbi.enc_start - 4) + 12;
149 if (offset == 12) {
150 IVTV_DEBUG_INFO("VBI offset == 0\n");
151 return -1;
153 s->pending_pts = read_enc(offset - 4) | ((u64)read_enc(offset - 8) << 32);
154 break;
156 case IVTV_DEC_STREAM_TYPE_VBI:
157 size = read_dec(itv->vbi.dec_start + 4) + 8;
158 offset = read_dec(itv->vbi.dec_start) + itv->vbi.dec_start;
159 s->pending_pts = 0;
160 offset += IVTV_DECODER_OFFSET;
161 break;
162 default:
163 /* shouldn't happen */
164 return -1;
167 /* if this is the start of the DMA then fill in the magic cookie */
168 if (s->sg_pending_size == 0 && ivtv_use_dma(s)) {
169 if (itv->has_cx23415 && (s->type == IVTV_ENC_STREAM_TYPE_PCM ||
170 s->type == IVTV_DEC_STREAM_TYPE_VBI)) {
171 s->pending_backup = read_dec(offset - IVTV_DECODER_OFFSET);
172 write_dec_sync(cpu_to_le32(DMA_MAGIC_COOKIE), offset - IVTV_DECODER_OFFSET);
174 else {
175 s->pending_backup = read_enc(offset);
176 write_enc_sync(cpu_to_le32(DMA_MAGIC_COOKIE), offset);
178 s->pending_offset = offset;
181 bytes_needed = size;
182 if (s->type == IVTV_ENC_STREAM_TYPE_YUV) {
183 /* The size for the Y samples needs to be rounded upwards to a
184 multiple of the buf_size. The UV samples then start in the
185 next buffer. */
186 bytes_needed = s->buf_size * ((bytes_needed + s->buf_size - 1) / s->buf_size);
187 bytes_needed += UVsize;
190 IVTV_DEBUG_HI_DMA("%s %s: 0x%08x bytes at 0x%08x\n",
191 ivtv_use_pio(s) ? "PIO" : "DMA", s->name, bytes_needed, offset);
193 rc = ivtv_queue_move(s, &s->q_free, &s->q_full, &s->q_predma, bytes_needed);
194 if (rc < 0) { /* Insufficient buffers */
195 IVTV_DEBUG_WARN("Cannot obtain %d bytes for %s data transfer\n",
196 bytes_needed, s->name);
197 return -1;
199 if (rc && !s->buffers_stolen && test_bit(IVTV_F_S_APPL_IO, &s->s_flags)) {
200 IVTV_WARN("All %s stream buffers are full. Dropping data.\n", s->name);
201 IVTV_WARN("Cause: the application is not reading fast enough.\n");
203 s->buffers_stolen = rc;
205 /* got the buffers, now fill in sg_pending */
206 buf = list_entry(s->q_predma.list.next, struct ivtv_buffer, list);
207 memset(buf->buf, 0, 128);
208 list_for_each_entry(buf, &s->q_predma.list, list) {
209 if (skip_bufs-- > 0)
210 continue;
211 s->sg_pending[idx].dst = buf->dma_handle;
212 s->sg_pending[idx].src = offset;
213 s->sg_pending[idx].size = s->buf_size;
214 buf->bytesused = min(size, s->buf_size);
215 buf->dma_xfer_cnt = s->dma_xfer_cnt;
217 s->q_predma.bytesused += buf->bytesused;
218 size -= buf->bytesused;
219 offset += s->buf_size;
221 /* Sync SG buffers */
222 ivtv_buf_sync_for_device(s, buf);
224 if (size == 0) { /* YUV */
225 /* process the UV section */
226 offset = UVoffset;
227 size = UVsize;
229 idx++;
231 s->sg_pending_size = idx;
232 return 0;
235 static void dma_post(struct ivtv_stream *s)
237 struct ivtv *itv = s->itv;
238 struct ivtv_buffer *buf = NULL;
239 struct list_head *p;
240 u32 offset;
241 __le32 *u32buf;
242 int x = 0;
244 IVTV_DEBUG_HI_DMA("%s %s completed (%x)\n", ivtv_use_pio(s) ? "PIO" : "DMA",
245 s->name, s->dma_offset);
246 list_for_each(p, &s->q_dma.list) {
247 buf = list_entry(p, struct ivtv_buffer, list);
248 u32buf = (__le32 *)buf->buf;
250 /* Sync Buffer */
251 ivtv_buf_sync_for_cpu(s, buf);
253 if (x == 0 && ivtv_use_dma(s)) {
254 offset = s->dma_last_offset;
255 if (u32buf[offset / 4] != DMA_MAGIC_COOKIE)
257 for (offset = 0; offset < 64; offset++) {
258 if (u32buf[offset] == DMA_MAGIC_COOKIE) {
259 break;
262 offset *= 4;
263 if (offset == 256) {
264 IVTV_DEBUG_WARN("%s: Couldn't find start of buffer within the first 256 bytes\n", s->name);
265 offset = s->dma_last_offset;
267 if (s->dma_last_offset != offset)
268 IVTV_DEBUG_WARN("%s: offset %d -> %d\n", s->name, s->dma_last_offset, offset);
269 s->dma_last_offset = offset;
271 if (itv->has_cx23415 && (s->type == IVTV_ENC_STREAM_TYPE_PCM ||
272 s->type == IVTV_DEC_STREAM_TYPE_VBI)) {
273 write_dec_sync(0, s->dma_offset - IVTV_DECODER_OFFSET);
275 else {
276 write_enc_sync(0, s->dma_offset);
278 if (offset) {
279 buf->bytesused -= offset;
280 memcpy(buf->buf, buf->buf + offset, buf->bytesused + offset);
282 *u32buf = cpu_to_le32(s->dma_backup);
284 x++;
285 /* flag byteswap ABCD -> DCBA for MPG & VBI data outside irq */
286 if (s->type == IVTV_ENC_STREAM_TYPE_MPG ||
287 s->type == IVTV_ENC_STREAM_TYPE_VBI)
288 buf->b_flags |= IVTV_F_B_NEED_BUF_SWAP;
290 if (buf)
291 buf->bytesused += s->dma_last_offset;
292 if (buf && s->type == IVTV_DEC_STREAM_TYPE_VBI) {
293 list_for_each_entry(buf, &s->q_dma.list, list) {
294 /* Parse and Groom VBI Data */
295 s->q_dma.bytesused -= buf->bytesused;
296 ivtv_process_vbi_data(itv, buf, 0, s->type);
297 s->q_dma.bytesused += buf->bytesused;
299 if (s->id == -1) {
300 ivtv_queue_move(s, &s->q_dma, NULL, &s->q_free, 0);
301 return;
304 ivtv_queue_move(s, &s->q_dma, NULL, &s->q_full, s->q_dma.bytesused);
305 if (s->id != -1)
306 wake_up(&s->waitq);
309 void ivtv_dma_stream_dec_prepare(struct ivtv_stream *s, u32 offset, int lock)
311 struct ivtv *itv = s->itv;
312 struct yuv_playback_info *yi = &itv->yuv_info;
313 u8 frame = yi->draw_frame;
314 struct yuv_frame_info *f = &yi->new_frame_info[frame];
315 struct ivtv_buffer *buf;
316 u32 y_size = 720 * ((f->src_h + 31) & ~31);
317 u32 uv_offset = offset + IVTV_YUV_BUFFER_UV_OFFSET;
318 int y_done = 0;
319 int bytes_written = 0;
320 unsigned long flags = 0;
321 int idx = 0;
323 IVTV_DEBUG_HI_DMA("DEC PREPARE DMA %s: %08x %08x\n", s->name, s->q_predma.bytesused, offset);
325 /* Insert buffer block for YUV if needed */
326 if (s->type == IVTV_DEC_STREAM_TYPE_YUV && f->offset_y) {
327 if (yi->blanking_dmaptr) {
328 s->sg_pending[idx].src = yi->blanking_dmaptr;
329 s->sg_pending[idx].dst = offset;
330 s->sg_pending[idx].size = 720 * 16;
332 offset += 720 * 16;
333 idx++;
336 list_for_each_entry(buf, &s->q_predma.list, list) {
337 /* YUV UV Offset from Y Buffer */
338 if (s->type == IVTV_DEC_STREAM_TYPE_YUV && !y_done &&
339 (bytes_written + buf->bytesused) >= y_size) {
340 s->sg_pending[idx].src = buf->dma_handle;
341 s->sg_pending[idx].dst = offset;
342 s->sg_pending[idx].size = y_size - bytes_written;
343 offset = uv_offset;
344 if (s->sg_pending[idx].size != buf->bytesused) {
345 idx++;
346 s->sg_pending[idx].src =
347 buf->dma_handle + s->sg_pending[idx - 1].size;
348 s->sg_pending[idx].dst = offset;
349 s->sg_pending[idx].size =
350 buf->bytesused - s->sg_pending[idx - 1].size;
351 offset += s->sg_pending[idx].size;
353 y_done = 1;
354 } else {
355 s->sg_pending[idx].src = buf->dma_handle;
356 s->sg_pending[idx].dst = offset;
357 s->sg_pending[idx].size = buf->bytesused;
358 offset += buf->bytesused;
360 bytes_written += buf->bytesused;
362 /* Sync SG buffers */
363 ivtv_buf_sync_for_device(s, buf);
364 idx++;
366 s->sg_pending_size = idx;
368 /* Sync Hardware SG List of buffers */
369 ivtv_stream_sync_for_device(s);
370 if (lock)
371 spin_lock_irqsave(&itv->dma_reg_lock, flags);
372 if (!test_bit(IVTV_F_I_DMA, &itv->i_flags)) {
373 ivtv_dma_dec_start(s);
375 else {
376 set_bit(IVTV_F_S_DMA_PENDING, &s->s_flags);
378 if (lock)
379 spin_unlock_irqrestore(&itv->dma_reg_lock, flags);
382 static void ivtv_dma_enc_start_xfer(struct ivtv_stream *s)
384 struct ivtv *itv = s->itv;
386 s->sg_dma->src = cpu_to_le32(s->sg_processing[s->sg_processed].src);
387 s->sg_dma->dst = cpu_to_le32(s->sg_processing[s->sg_processed].dst);
388 s->sg_dma->size = cpu_to_le32(s->sg_processing[s->sg_processed].size | 0x80000000);
389 s->sg_processed++;
390 /* Sync Hardware SG List of buffers */
391 ivtv_stream_sync_for_device(s);
392 write_reg(s->sg_handle, IVTV_REG_ENCDMAADDR);
393 write_reg_sync(read_reg(IVTV_REG_DMAXFER) | 0x02, IVTV_REG_DMAXFER);
394 itv->dma_timer.expires = jiffies + msecs_to_jiffies(300);
395 add_timer(&itv->dma_timer);
398 static void ivtv_dma_dec_start_xfer(struct ivtv_stream *s)
400 struct ivtv *itv = s->itv;
402 s->sg_dma->src = cpu_to_le32(s->sg_processing[s->sg_processed].src);
403 s->sg_dma->dst = cpu_to_le32(s->sg_processing[s->sg_processed].dst);
404 s->sg_dma->size = cpu_to_le32(s->sg_processing[s->sg_processed].size | 0x80000000);
405 s->sg_processed++;
406 /* Sync Hardware SG List of buffers */
407 ivtv_stream_sync_for_device(s);
408 write_reg(s->sg_handle, IVTV_REG_DECDMAADDR);
409 write_reg_sync(read_reg(IVTV_REG_DMAXFER) | 0x01, IVTV_REG_DMAXFER);
410 itv->dma_timer.expires = jiffies + msecs_to_jiffies(300);
411 add_timer(&itv->dma_timer);
414 /* start the encoder DMA */
415 static void ivtv_dma_enc_start(struct ivtv_stream *s)
417 struct ivtv *itv = s->itv;
418 struct ivtv_stream *s_vbi = &itv->streams[IVTV_ENC_STREAM_TYPE_VBI];
419 int i;
421 IVTV_DEBUG_HI_DMA("start %s for %s\n", ivtv_use_dma(s) ? "DMA" : "PIO", s->name);
423 if (s->q_predma.bytesused)
424 ivtv_queue_move(s, &s->q_predma, NULL, &s->q_dma, s->q_predma.bytesused);
426 if (ivtv_use_dma(s))
427 s->sg_pending[s->sg_pending_size - 1].size += 256;
429 /* If this is an MPEG stream, and VBI data is also pending, then append the
430 VBI DMA to the MPEG DMA and transfer both sets of data at once.
432 VBI DMA is a second class citizen compared to MPEG and mixing them together
433 will confuse the firmware (the end of a VBI DMA is seen as the end of a
434 MPEG DMA, thus effectively dropping an MPEG frame). So instead we make
435 sure we only use the MPEG DMA to transfer the VBI DMA if both are in
436 use. This way no conflicts occur. */
437 clear_bit(IVTV_F_S_DMA_HAS_VBI, &s->s_flags);
438 if (s->type == IVTV_ENC_STREAM_TYPE_MPG && s_vbi->sg_pending_size &&
439 s->sg_pending_size + s_vbi->sg_pending_size <= s->buffers) {
440 ivtv_queue_move(s_vbi, &s_vbi->q_predma, NULL, &s_vbi->q_dma, s_vbi->q_predma.bytesused);
441 if (ivtv_use_dma(s_vbi))
442 s_vbi->sg_pending[s_vbi->sg_pending_size - 1].size += 256;
443 for (i = 0; i < s_vbi->sg_pending_size; i++) {
444 s->sg_pending[s->sg_pending_size++] = s_vbi->sg_pending[i];
446 s_vbi->dma_offset = s_vbi->pending_offset;
447 s_vbi->sg_pending_size = 0;
448 s_vbi->dma_xfer_cnt++;
449 set_bit(IVTV_F_S_DMA_HAS_VBI, &s->s_flags);
450 IVTV_DEBUG_HI_DMA("include DMA for %s\n", s_vbi->name);
453 s->dma_xfer_cnt++;
454 memcpy(s->sg_processing, s->sg_pending, sizeof(struct ivtv_sg_host_element) * s->sg_pending_size);
455 s->sg_processing_size = s->sg_pending_size;
456 s->sg_pending_size = 0;
457 s->sg_processed = 0;
458 s->dma_offset = s->pending_offset;
459 s->dma_backup = s->pending_backup;
460 s->dma_pts = s->pending_pts;
462 if (ivtv_use_pio(s)) {
463 set_bit(IVTV_F_I_WORK_HANDLER_PIO, &itv->i_flags);
464 set_bit(IVTV_F_I_HAVE_WORK, &itv->i_flags);
465 set_bit(IVTV_F_I_PIO, &itv->i_flags);
466 itv->cur_pio_stream = s->type;
468 else {
469 itv->dma_retries = 0;
470 ivtv_dma_enc_start_xfer(s);
471 set_bit(IVTV_F_I_DMA, &itv->i_flags);
472 itv->cur_dma_stream = s->type;
476 static void ivtv_dma_dec_start(struct ivtv_stream *s)
478 struct ivtv *itv = s->itv;
480 if (s->q_predma.bytesused)
481 ivtv_queue_move(s, &s->q_predma, NULL, &s->q_dma, s->q_predma.bytesused);
482 s->dma_xfer_cnt++;
483 memcpy(s->sg_processing, s->sg_pending, sizeof(struct ivtv_sg_host_element) * s->sg_pending_size);
484 s->sg_processing_size = s->sg_pending_size;
485 s->sg_pending_size = 0;
486 s->sg_processed = 0;
488 IVTV_DEBUG_HI_DMA("start DMA for %s\n", s->name);
489 itv->dma_retries = 0;
490 ivtv_dma_dec_start_xfer(s);
491 set_bit(IVTV_F_I_DMA, &itv->i_flags);
492 itv->cur_dma_stream = s->type;
495 static void ivtv_irq_dma_read(struct ivtv *itv)
497 struct ivtv_stream *s = NULL;
498 struct ivtv_buffer *buf;
499 int hw_stream_type = 0;
501 IVTV_DEBUG_HI_IRQ("DEC DMA READ\n");
503 del_timer(&itv->dma_timer);
505 if (!test_bit(IVTV_F_I_UDMA, &itv->i_flags) && itv->cur_dma_stream < 0)
506 return;
508 if (!test_bit(IVTV_F_I_UDMA, &itv->i_flags)) {
509 s = &itv->streams[itv->cur_dma_stream];
510 ivtv_stream_sync_for_cpu(s);
512 if (read_reg(IVTV_REG_DMASTATUS) & 0x14) {
513 IVTV_DEBUG_WARN("DEC DMA ERROR %x (xfer %d of %d, retry %d)\n",
514 read_reg(IVTV_REG_DMASTATUS),
515 s->sg_processed, s->sg_processing_size, itv->dma_retries);
516 write_reg(read_reg(IVTV_REG_DMASTATUS) & 3, IVTV_REG_DMASTATUS);
517 if (itv->dma_retries == 3) {
518 /* Too many retries, give up on this frame */
519 itv->dma_retries = 0;
520 s->sg_processed = s->sg_processing_size;
522 else {
523 /* Retry, starting with the first xfer segment.
524 Just retrying the current segment is not sufficient. */
525 s->sg_processed = 0;
526 itv->dma_retries++;
529 if (s->sg_processed < s->sg_processing_size) {
530 /* DMA next buffer */
531 ivtv_dma_dec_start_xfer(s);
532 return;
534 if (s->type == IVTV_DEC_STREAM_TYPE_YUV)
535 hw_stream_type = 2;
536 IVTV_DEBUG_HI_DMA("DEC DATA READ %s: %d\n", s->name, s->q_dma.bytesused);
538 /* For some reason must kick the firmware, like PIO mode,
539 I think this tells the firmware we are done and the size
540 of the xfer so it can calculate what we need next.
541 I think we can do this part ourselves but would have to
542 fully calculate xfer info ourselves and not use interrupts
544 ivtv_vapi(itv, CX2341X_DEC_SCHED_DMA_FROM_HOST, 3, 0, s->q_dma.bytesused,
545 hw_stream_type);
547 /* Free last DMA call */
548 while ((buf = ivtv_dequeue(s, &s->q_dma)) != NULL) {
549 ivtv_buf_sync_for_cpu(s, buf);
550 ivtv_enqueue(s, buf, &s->q_free);
552 wake_up(&s->waitq);
554 clear_bit(IVTV_F_I_UDMA, &itv->i_flags);
555 clear_bit(IVTV_F_I_DMA, &itv->i_flags);
556 itv->cur_dma_stream = -1;
557 wake_up(&itv->dma_waitq);
560 static void ivtv_irq_enc_dma_complete(struct ivtv *itv)
562 u32 data[CX2341X_MBOX_MAX_DATA];
563 struct ivtv_stream *s;
565 ivtv_api_get_data(&itv->enc_mbox, IVTV_MBOX_DMA_END, data);
566 IVTV_DEBUG_HI_IRQ("ENC DMA COMPLETE %x %d (%d)\n", data[0], data[1], itv->cur_dma_stream);
568 del_timer(&itv->dma_timer);
570 if (itv->cur_dma_stream < 0)
571 return;
573 s = &itv->streams[itv->cur_dma_stream];
574 ivtv_stream_sync_for_cpu(s);
576 if (data[0] & 0x18) {
577 IVTV_DEBUG_WARN("ENC DMA ERROR %x (offset %08x, xfer %d of %d, retry %d)\n", data[0],
578 s->dma_offset, s->sg_processed, s->sg_processing_size, itv->dma_retries);
579 write_reg(read_reg(IVTV_REG_DMASTATUS) & 3, IVTV_REG_DMASTATUS);
580 if (itv->dma_retries == 3) {
581 /* Too many retries, give up on this frame */
582 itv->dma_retries = 0;
583 s->sg_processed = s->sg_processing_size;
585 else {
586 /* Retry, starting with the first xfer segment.
587 Just retrying the current segment is not sufficient. */
588 s->sg_processed = 0;
589 itv->dma_retries++;
592 if (s->sg_processed < s->sg_processing_size) {
593 /* DMA next buffer */
594 ivtv_dma_enc_start_xfer(s);
595 return;
597 clear_bit(IVTV_F_I_DMA, &itv->i_flags);
598 itv->cur_dma_stream = -1;
599 dma_post(s);
600 if (test_and_clear_bit(IVTV_F_S_DMA_HAS_VBI, &s->s_flags)) {
601 s = &itv->streams[IVTV_ENC_STREAM_TYPE_VBI];
602 dma_post(s);
604 s->sg_processing_size = 0;
605 s->sg_processed = 0;
606 wake_up(&itv->dma_waitq);
609 static void ivtv_irq_enc_pio_complete(struct ivtv *itv)
611 struct ivtv_stream *s;
613 if (itv->cur_pio_stream < 0 || itv->cur_pio_stream >= IVTV_MAX_STREAMS) {
614 itv->cur_pio_stream = -1;
615 return;
617 s = &itv->streams[itv->cur_pio_stream];
618 IVTV_DEBUG_HI_IRQ("ENC PIO COMPLETE %s\n", s->name);
619 clear_bit(IVTV_F_I_PIO, &itv->i_flags);
620 itv->cur_pio_stream = -1;
621 dma_post(s);
622 if (s->type == IVTV_ENC_STREAM_TYPE_MPG)
623 ivtv_vapi(itv, CX2341X_ENC_SCHED_DMA_TO_HOST, 3, 0, 0, 0);
624 else if (s->type == IVTV_ENC_STREAM_TYPE_YUV)
625 ivtv_vapi(itv, CX2341X_ENC_SCHED_DMA_TO_HOST, 3, 0, 0, 1);
626 else if (s->type == IVTV_ENC_STREAM_TYPE_PCM)
627 ivtv_vapi(itv, CX2341X_ENC_SCHED_DMA_TO_HOST, 3, 0, 0, 2);
628 clear_bit(IVTV_F_I_PIO, &itv->i_flags);
629 if (test_and_clear_bit(IVTV_F_S_DMA_HAS_VBI, &s->s_flags)) {
630 s = &itv->streams[IVTV_ENC_STREAM_TYPE_VBI];
631 dma_post(s);
633 wake_up(&itv->dma_waitq);
636 static void ivtv_irq_dma_err(struct ivtv *itv)
638 u32 data[CX2341X_MBOX_MAX_DATA];
640 del_timer(&itv->dma_timer);
641 ivtv_api_get_data(&itv->enc_mbox, IVTV_MBOX_DMA_END, data);
642 IVTV_DEBUG_WARN("DMA ERROR %08x %08x %08x %d\n", data[0], data[1],
643 read_reg(IVTV_REG_DMASTATUS), itv->cur_dma_stream);
644 write_reg(read_reg(IVTV_REG_DMASTATUS) & 3, IVTV_REG_DMASTATUS);
645 if (!test_bit(IVTV_F_I_UDMA, &itv->i_flags) &&
646 itv->cur_dma_stream >= 0 && itv->cur_dma_stream < IVTV_MAX_STREAMS) {
647 struct ivtv_stream *s = &itv->streams[itv->cur_dma_stream];
649 /* retry */
650 if (s->type >= IVTV_DEC_STREAM_TYPE_MPG)
651 ivtv_dma_dec_start(s);
652 else
653 ivtv_dma_enc_start(s);
654 return;
656 if (test_bit(IVTV_F_I_UDMA, &itv->i_flags)) {
657 ivtv_udma_start(itv);
658 return;
660 clear_bit(IVTV_F_I_UDMA, &itv->i_flags);
661 clear_bit(IVTV_F_I_DMA, &itv->i_flags);
662 itv->cur_dma_stream = -1;
663 wake_up(&itv->dma_waitq);
666 static void ivtv_irq_enc_start_cap(struct ivtv *itv)
668 u32 data[CX2341X_MBOX_MAX_DATA];
669 struct ivtv_stream *s;
671 /* Get DMA destination and size arguments from card */
672 ivtv_api_get_data(&itv->enc_mbox, IVTV_MBOX_DMA, data);
673 IVTV_DEBUG_HI_IRQ("ENC START CAP %d: %08x %08x\n", data[0], data[1], data[2]);
675 if (data[0] > 2 || data[1] == 0 || data[2] == 0) {
676 IVTV_DEBUG_WARN("Unknown input: %08x %08x %08x\n",
677 data[0], data[1], data[2]);
678 return;
680 s = &itv->streams[ivtv_stream_map[data[0]]];
681 if (!stream_enc_dma_append(s, data)) {
682 set_bit(ivtv_use_pio(s) ? IVTV_F_S_PIO_PENDING : IVTV_F_S_DMA_PENDING, &s->s_flags);
686 static void ivtv_irq_enc_vbi_cap(struct ivtv *itv)
688 u32 data[CX2341X_MBOX_MAX_DATA];
689 struct ivtv_stream *s;
691 IVTV_DEBUG_HI_IRQ("ENC START VBI CAP\n");
692 s = &itv->streams[IVTV_ENC_STREAM_TYPE_VBI];
694 if (!stream_enc_dma_append(s, data))
695 set_bit(ivtv_use_pio(s) ? IVTV_F_S_PIO_PENDING : IVTV_F_S_DMA_PENDING, &s->s_flags);
698 static void ivtv_irq_dec_vbi_reinsert(struct ivtv *itv)
700 u32 data[CX2341X_MBOX_MAX_DATA];
701 struct ivtv_stream *s = &itv->streams[IVTV_DEC_STREAM_TYPE_VBI];
703 IVTV_DEBUG_HI_IRQ("DEC VBI REINSERT\n");
704 if (test_bit(IVTV_F_S_CLAIMED, &s->s_flags) &&
705 !stream_enc_dma_append(s, data)) {
706 set_bit(IVTV_F_S_PIO_PENDING, &s->s_flags);
710 static void ivtv_irq_dec_data_req(struct ivtv *itv)
712 u32 data[CX2341X_MBOX_MAX_DATA];
713 struct ivtv_stream *s;
715 /* YUV or MPG */
716 ivtv_api_get_data(&itv->dec_mbox, IVTV_MBOX_DMA, data);
718 if (test_bit(IVTV_F_I_DEC_YUV, &itv->i_flags)) {
719 itv->dma_data_req_size =
720 1080 * ((itv->yuv_info.v4l2_src_h + 31) & ~31);
721 itv->dma_data_req_offset = data[1];
722 if (atomic_read(&itv->yuv_info.next_dma_frame) >= 0)
723 ivtv_yuv_frame_complete(itv);
724 s = &itv->streams[IVTV_DEC_STREAM_TYPE_YUV];
726 else {
727 itv->dma_data_req_size = min_t(u32, data[2], 0x10000);
728 itv->dma_data_req_offset = data[1];
729 s = &itv->streams[IVTV_DEC_STREAM_TYPE_MPG];
731 IVTV_DEBUG_HI_IRQ("DEC DATA REQ %s: %d %08x %u\n", s->name, s->q_full.bytesused,
732 itv->dma_data_req_offset, itv->dma_data_req_size);
733 if (itv->dma_data_req_size == 0 || s->q_full.bytesused < itv->dma_data_req_size) {
734 set_bit(IVTV_F_S_NEEDS_DATA, &s->s_flags);
736 else {
737 if (test_bit(IVTV_F_I_DEC_YUV, &itv->i_flags))
738 ivtv_yuv_setup_stream_frame(itv);
739 clear_bit(IVTV_F_S_NEEDS_DATA, &s->s_flags);
740 ivtv_queue_move(s, &s->q_full, NULL, &s->q_predma, itv->dma_data_req_size);
741 ivtv_dma_stream_dec_prepare(s, itv->dma_data_req_offset + IVTV_DECODER_OFFSET, 0);
745 static void ivtv_irq_vsync(struct ivtv *itv)
747 /* The vsync interrupt is unusual in that it won't clear until
748 * the end of the first line for the current field, at which
749 * point it clears itself. This can result in repeated vsync
750 * interrupts, or a missed vsync. Read some of the registers
751 * to determine the line being displayed and ensure we handle
752 * one vsync per frame.
754 unsigned int frame = read_reg(0x28c0) & 1;
755 struct yuv_playback_info *yi = &itv->yuv_info;
756 int last_dma_frame = atomic_read(&yi->next_dma_frame);
757 struct yuv_frame_info *f = &yi->new_frame_info[last_dma_frame];
759 if (0) IVTV_DEBUG_IRQ("DEC VSYNC\n");
761 if (((frame ^ f->sync_field) == 0 &&
762 ((itv->last_vsync_field & 1) ^ f->sync_field)) ||
763 (frame != (itv->last_vsync_field & 1) && !f->interlaced)) {
764 int next_dma_frame = last_dma_frame;
766 if (!(f->interlaced && f->delay && yi->fields_lapsed < 1)) {
767 if (next_dma_frame >= 0 && next_dma_frame != atomic_read(&yi->next_fill_frame)) {
768 write_reg(yuv_offset[next_dma_frame] >> 4, 0x82c);
769 write_reg((yuv_offset[next_dma_frame] + IVTV_YUV_BUFFER_UV_OFFSET) >> 4, 0x830);
770 write_reg(yuv_offset[next_dma_frame] >> 4, 0x834);
771 write_reg((yuv_offset[next_dma_frame] + IVTV_YUV_BUFFER_UV_OFFSET) >> 4, 0x838);
772 next_dma_frame = (next_dma_frame + 1) % IVTV_YUV_BUFFERS;
773 atomic_set(&yi->next_dma_frame, next_dma_frame);
774 yi->fields_lapsed = -1;
775 yi->running = 1;
779 if (frame != (itv->last_vsync_field & 1)) {
780 struct ivtv_stream *s = ivtv_get_output_stream(itv);
782 itv->last_vsync_field += 1;
783 if (frame == 0) {
784 clear_bit(IVTV_F_I_VALID_DEC_TIMINGS, &itv->i_flags);
785 clear_bit(IVTV_F_I_EV_VSYNC_FIELD, &itv->i_flags);
787 else {
788 set_bit(IVTV_F_I_EV_VSYNC_FIELD, &itv->i_flags);
790 if (test_bit(IVTV_F_I_EV_VSYNC_ENABLED, &itv->i_flags)) {
791 set_bit(IVTV_F_I_EV_VSYNC, &itv->i_flags);
792 wake_up(&itv->event_waitq);
794 wake_up(&itv->vsync_waitq);
795 if (s)
796 wake_up(&s->waitq);
798 /* Send VBI to saa7127 */
799 if (frame && (itv->output_mode == OUT_PASSTHROUGH ||
800 test_bit(IVTV_F_I_UPDATE_WSS, &itv->i_flags) ||
801 test_bit(IVTV_F_I_UPDATE_VPS, &itv->i_flags) ||
802 test_bit(IVTV_F_I_UPDATE_CC, &itv->i_flags))) {
803 set_bit(IVTV_F_I_WORK_HANDLER_VBI, &itv->i_flags);
804 set_bit(IVTV_F_I_HAVE_WORK, &itv->i_flags);
807 /* Check if we need to update the yuv registers */
808 if (yi->running && (yi->yuv_forced_update || f->update)) {
809 if (!f->update) {
810 last_dma_frame =
811 (u8)(atomic_read(&yi->next_dma_frame) -
812 1) % IVTV_YUV_BUFFERS;
813 f = &yi->new_frame_info[last_dma_frame];
816 if (f->src_w) {
817 yi->update_frame = last_dma_frame;
818 f->update = 0;
819 yi->yuv_forced_update = 0;
820 set_bit(IVTV_F_I_WORK_HANDLER_YUV, &itv->i_flags);
821 set_bit(IVTV_F_I_HAVE_WORK, &itv->i_flags);
825 yi->fields_lapsed++;
829 #define IVTV_IRQ_DMA (IVTV_IRQ_DMA_READ | IVTV_IRQ_ENC_DMA_COMPLETE | IVTV_IRQ_DMA_ERR | IVTV_IRQ_ENC_START_CAP | IVTV_IRQ_ENC_VBI_CAP | IVTV_IRQ_DEC_DATA_REQ | IVTV_IRQ_DEC_VBI_RE_INSERT)
831 irqreturn_t ivtv_irq_handler(int irq, void *dev_id)
833 struct ivtv *itv = (struct ivtv *)dev_id;
834 u32 combo;
835 u32 stat;
836 int i;
837 u8 vsync_force = 0;
839 spin_lock(&itv->dma_reg_lock);
840 /* get contents of irq status register */
841 stat = read_reg(IVTV_REG_IRQSTATUS);
843 combo = ~itv->irqmask & stat;
845 /* Clear out IRQ */
846 if (combo) write_reg(combo, IVTV_REG_IRQSTATUS);
848 if (0 == combo) {
849 /* The vsync interrupt is unusual and clears itself. If we
850 * took too long, we may have missed it. Do some checks
852 if (~itv->irqmask & IVTV_IRQ_DEC_VSYNC) {
853 /* vsync is enabled, see if we're in a new field */
854 if ((itv->last_vsync_field & 1) != (read_reg(0x28c0) & 1)) {
855 /* New field, looks like we missed it */
856 IVTV_DEBUG_YUV("VSync interrupt missed %d\n",read_reg(0x28c0)>>16);
857 vsync_force = 1;
861 if (!vsync_force) {
862 /* No Vsync expected, wasn't for us */
863 spin_unlock(&itv->dma_reg_lock);
864 return IRQ_NONE;
868 /* Exclude interrupts noted below from the output, otherwise the log is flooded with
869 these messages */
870 if (combo & ~0xff6d0400)
871 IVTV_DEBUG_HI_IRQ("======= valid IRQ bits: 0x%08x ======\n", combo);
873 if (combo & IVTV_IRQ_DEC_DMA_COMPLETE) {
874 IVTV_DEBUG_HI_IRQ("DEC DMA COMPLETE\n");
877 if (combo & IVTV_IRQ_DMA_READ) {
878 ivtv_irq_dma_read(itv);
881 if (combo & IVTV_IRQ_ENC_DMA_COMPLETE) {
882 ivtv_irq_enc_dma_complete(itv);
885 if (combo & IVTV_IRQ_ENC_PIO_COMPLETE) {
886 ivtv_irq_enc_pio_complete(itv);
889 if (combo & IVTV_IRQ_DMA_ERR) {
890 ivtv_irq_dma_err(itv);
893 if (combo & IVTV_IRQ_ENC_START_CAP) {
894 ivtv_irq_enc_start_cap(itv);
897 if (combo & IVTV_IRQ_ENC_VBI_CAP) {
898 ivtv_irq_enc_vbi_cap(itv);
901 if (combo & IVTV_IRQ_DEC_VBI_RE_INSERT) {
902 ivtv_irq_dec_vbi_reinsert(itv);
905 if (combo & IVTV_IRQ_ENC_EOS) {
906 IVTV_DEBUG_IRQ("ENC EOS\n");
907 set_bit(IVTV_F_I_EOS, &itv->i_flags);
908 wake_up(&itv->eos_waitq);
911 if (combo & IVTV_IRQ_DEC_DATA_REQ) {
912 ivtv_irq_dec_data_req(itv);
915 /* Decoder Vertical Sync - We can't rely on 'combo', so check if vsync enabled */
916 if (~itv->irqmask & IVTV_IRQ_DEC_VSYNC) {
917 ivtv_irq_vsync(itv);
920 if (combo & IVTV_IRQ_ENC_VIM_RST) {
921 IVTV_DEBUG_IRQ("VIM RST\n");
922 /*ivtv_vapi(itv, CX2341X_ENC_REFRESH_INPUT, 0); */
925 if (combo & IVTV_IRQ_DEC_AUD_MODE_CHG) {
926 IVTV_DEBUG_INFO("Stereo mode changed\n");
929 if ((combo & IVTV_IRQ_DMA) && !test_bit(IVTV_F_I_DMA, &itv->i_flags)) {
930 itv->irq_rr_idx++;
931 for (i = 0; i < IVTV_MAX_STREAMS; i++) {
932 int idx = (i + itv->irq_rr_idx) % IVTV_MAX_STREAMS;
933 struct ivtv_stream *s = &itv->streams[idx];
935 if (!test_and_clear_bit(IVTV_F_S_DMA_PENDING, &s->s_flags))
936 continue;
937 if (s->type >= IVTV_DEC_STREAM_TYPE_MPG)
938 ivtv_dma_dec_start(s);
939 else
940 ivtv_dma_enc_start(s);
941 break;
943 if (i == IVTV_MAX_STREAMS && test_and_clear_bit(IVTV_F_I_UDMA_PENDING, &itv->i_flags)) {
944 ivtv_udma_start(itv);
948 if ((combo & IVTV_IRQ_DMA) && !test_bit(IVTV_F_I_PIO, &itv->i_flags)) {
949 itv->irq_rr_idx++;
950 for (i = 0; i < IVTV_MAX_STREAMS; i++) {
951 int idx = (i + itv->irq_rr_idx) % IVTV_MAX_STREAMS;
952 struct ivtv_stream *s = &itv->streams[idx];
954 if (!test_and_clear_bit(IVTV_F_S_PIO_PENDING, &s->s_flags))
955 continue;
956 if (s->type == IVTV_DEC_STREAM_TYPE_VBI || s->type < IVTV_DEC_STREAM_TYPE_MPG)
957 ivtv_dma_enc_start(s);
958 break;
962 if (test_and_clear_bit(IVTV_F_I_HAVE_WORK, &itv->i_flags)) {
963 queue_work(itv->irq_work_queues, &itv->irq_work_queue);
966 spin_unlock(&itv->dma_reg_lock);
968 /* If we've just handled a 'forced' vsync, it's safest to say it
969 * wasn't ours. Another device may have triggered it at just
970 * the right time.
972 return vsync_force ? IRQ_NONE : IRQ_HANDLED;
975 void ivtv_unfinished_dma(unsigned long arg)
977 struct ivtv *itv = (struct ivtv *)arg;
979 if (!test_bit(IVTV_F_I_DMA, &itv->i_flags))
980 return;
981 IVTV_ERR("DMA TIMEOUT %08x %d\n", read_reg(IVTV_REG_DMASTATUS), itv->cur_dma_stream);
983 write_reg(read_reg(IVTV_REG_DMASTATUS) & 3, IVTV_REG_DMASTATUS);
984 clear_bit(IVTV_F_I_UDMA, &itv->i_flags);
985 clear_bit(IVTV_F_I_DMA, &itv->i_flags);
986 itv->cur_dma_stream = -1;
987 wake_up(&itv->dma_waitq);