2 * c8sectpfe-core.c - C8SECTPFE STi DVB driver
4 * Copyright (c) STMicroelectronics 2015
6 * Author:Peter Bennett <peter.bennett@st.com>
7 * Peter Griffin <peter.griffin@linaro.org>
9 * This program is free software; you can redistribute it and/or
10 * modify it under the terms of the GNU General Public License as
11 * published by the Free Software Foundation; either version 2 of
12 * the License, or (at your option) any later version.
14 #include <linux/atomic.h>
15 #include <linux/clk.h>
16 #include <linux/completion.h>
17 #include <linux/delay.h>
18 #include <linux/device.h>
19 #include <linux/dma-mapping.h>
20 #include <linux/dvb/dmx.h>
21 #include <linux/dvb/frontend.h>
22 #include <linux/errno.h>
23 #include <linux/firmware.h>
24 #include <linux/init.h>
25 #include <linux/interrupt.h>
27 #include <linux/module.h>
28 #include <linux/of_gpio.h>
29 #include <linux/of_platform.h>
30 #include <linux/platform_device.h>
31 #include <linux/usb.h>
32 #include <linux/slab.h>
33 #include <linux/time.h>
34 #include <linux/version.h>
35 #include <linux/wait.h>
36 #include <linux/pinctrl/pinctrl.h>
38 #include "c8sectpfe-core.h"
39 #include "c8sectpfe-common.h"
40 #include "c8sectpfe-debugfs.h"
42 #include "dvb_demux.h"
43 #include "dvb_frontend.h"
46 #define FIRMWARE_MEMDMA "pti_memdma_h407.elf"
47 MODULE_FIRMWARE(FIRMWARE_MEMDMA
);
49 #define PID_TABLE_SIZE 1024
52 static int load_c8sectpfe_fw_step1(struct c8sectpfei
*fei
);
54 #define TS_PKT_SIZE 188
55 #define HEADER_SIZE (4)
56 #define PACKET_SIZE (TS_PKT_SIZE+HEADER_SIZE)
58 #define FEI_ALIGNMENT (32)
59 /* hw requires minimum of 8*PACKET_SIZE and padded to 8byte boundary */
60 #define FEI_BUFFER_SIZE (8*PACKET_SIZE*340)
64 static void c8sectpfe_timer_interrupt(unsigned long ac8sectpfei
)
66 struct c8sectpfei
*fei
= (struct c8sectpfei
*)ac8sectpfei
;
67 struct channel_info
*channel
;
70 /* iterate through input block channels */
71 for (chan_num
= 0; chan_num
< fei
->tsin_count
; chan_num
++) {
72 channel
= fei
->channel_data
[chan_num
];
74 /* is this descriptor initialised and TP enabled */
75 if (channel
->irec
&& readl(channel
->irec
+ DMA_PRDS_TPENABLE
))
76 tasklet_schedule(&channel
->tsklet
);
79 fei
->timer
.expires
= jiffies
+ msecs_to_jiffies(POLL_MSECS
);
80 add_timer(&fei
->timer
);
83 static void channel_swdemux_tsklet(unsigned long data
)
85 struct channel_info
*channel
= (struct channel_info
*)data
;
86 struct c8sectpfei
*fei
= channel
->fei
;
88 int pos
, num_packets
, n
, size
;
91 if (unlikely(!channel
|| !channel
->irec
))
94 wp
= readl(channel
->irec
+ DMA_PRDS_BUSWP_TP(0));
95 rp
= readl(channel
->irec
+ DMA_PRDS_BUSRP_TP(0));
97 pos
= rp
- channel
->back_buffer_busaddr
;
101 wp
= channel
->back_buffer_busaddr
+ FEI_BUFFER_SIZE
;
104 num_packets
= size
/ PACKET_SIZE
;
106 /* manage cache so data is visible to CPU */
107 dma_sync_single_for_cpu(fei
->dev
,
112 buf
= (u8
*) channel
->back_buffer_aligned
;
115 "chan=%d channel=%p num_packets = %d, buf = %p, pos = 0x%x\n\t"
116 "rp=0x%lx, wp=0x%lx\n",
117 channel
->tsin_id
, channel
, num_packets
, buf
, pos
, rp
, wp
);
119 for (n
= 0; n
< num_packets
; n
++) {
120 dvb_dmx_swfilter_packets(
122 demux
[channel
->demux_mapping
].dvb_demux
,
128 /* advance the read pointer */
129 if (wp
== (channel
->back_buffer_busaddr
+ FEI_BUFFER_SIZE
))
130 writel(channel
->back_buffer_busaddr
, channel
->irec
+
131 DMA_PRDS_BUSRP_TP(0));
133 writel(wp
, channel
->irec
+ DMA_PRDS_BUSWP_TP(0));
136 static int c8sectpfe_start_feed(struct dvb_demux_feed
*dvbdmxfeed
)
138 struct dvb_demux
*demux
= dvbdmxfeed
->demux
;
139 struct stdemux
*stdemux
= (struct stdemux
*)demux
->priv
;
140 struct c8sectpfei
*fei
= stdemux
->c8sectpfei
;
141 struct channel_info
*channel
;
143 unsigned long *bitmap
;
145 switch (dvbdmxfeed
->type
) {
151 dev_err(fei
->dev
, "%s:%d Error bailing\n"
152 , __func__
, __LINE__
);
156 if (dvbdmxfeed
->type
== DMX_TYPE_TS
) {
157 switch (dvbdmxfeed
->pes_type
) {
160 case DMX_PES_TELETEXT
:
165 dev_err(fei
->dev
, "%s:%d Error bailing\n"
166 , __func__
, __LINE__
);
171 if (!atomic_read(&fei
->fw_loaded
)) {
172 dev_err(fei
->dev
, "%s: c8sectpfe fw not loaded\n", __func__
);
176 mutex_lock(&fei
->lock
);
178 channel
= fei
->channel_data
[stdemux
->tsin_index
];
180 bitmap
= (unsigned long *) channel
->pid_buffer_aligned
;
182 /* 8192 is a special PID */
183 if (dvbdmxfeed
->pid
== 8192) {
184 tmp
= readl(fei
->io
+ C8SECTPFE_IB_PID_SET(channel
->tsin_id
));
185 tmp
&= ~C8SECTPFE_PID_ENABLE
;
186 writel(tmp
, fei
->io
+ C8SECTPFE_IB_PID_SET(channel
->tsin_id
));
189 bitmap_set(bitmap
, dvbdmxfeed
->pid
, 1);
192 /* manage cache so PID bitmap is visible to HW */
193 dma_sync_single_for_device(fei
->dev
,
194 channel
->pid_buffer_busaddr
,
200 if (fei
->global_feed_count
== 0) {
201 fei
->timer
.expires
= jiffies
+
202 msecs_to_jiffies(msecs_to_jiffies(POLL_MSECS
));
204 add_timer(&fei
->timer
);
207 if (stdemux
->running_feed_count
== 0) {
209 dev_dbg(fei
->dev
, "Starting channel=%p\n", channel
);
211 tasklet_init(&channel
->tsklet
, channel_swdemux_tsklet
,
212 (unsigned long) channel
);
214 /* Reset the internal inputblock sram pointers */
215 writel(channel
->fifo
,
216 fei
->io
+ C8SECTPFE_IB_BUFF_STRT(channel
->tsin_id
));
217 writel(channel
->fifo
+ FIFO_LEN
- 1,
218 fei
->io
+ C8SECTPFE_IB_BUFF_END(channel
->tsin_id
));
220 writel(channel
->fifo
,
221 fei
->io
+ C8SECTPFE_IB_READ_PNT(channel
->tsin_id
));
222 writel(channel
->fifo
,
223 fei
->io
+ C8SECTPFE_IB_WRT_PNT(channel
->tsin_id
));
226 /* reset read / write memdma ptrs for this channel */
227 writel(channel
->back_buffer_busaddr
, channel
->irec
+
228 DMA_PRDS_BUSBASE_TP(0));
230 tmp
= channel
->back_buffer_busaddr
+ FEI_BUFFER_SIZE
- 1;
231 writel(tmp
, channel
->irec
+ DMA_PRDS_BUSTOP_TP(0));
233 writel(channel
->back_buffer_busaddr
, channel
->irec
+
234 DMA_PRDS_BUSWP_TP(0));
236 /* Issue a reset and enable InputBlock */
237 writel(C8SECTPFE_SYS_ENABLE
| C8SECTPFE_SYS_RESET
238 , fei
->io
+ C8SECTPFE_IB_SYS(channel
->tsin_id
));
240 /* and enable the tp */
241 writel(0x1, channel
->irec
+ DMA_PRDS_TPENABLE
);
243 dev_dbg(fei
->dev
, "%s:%d Starting DMA feed on stdemux=%p\n"
244 , __func__
, __LINE__
, stdemux
);
247 stdemux
->running_feed_count
++;
248 fei
->global_feed_count
++;
250 mutex_unlock(&fei
->lock
);
255 static int c8sectpfe_stop_feed(struct dvb_demux_feed
*dvbdmxfeed
)
258 struct dvb_demux
*demux
= dvbdmxfeed
->demux
;
259 struct stdemux
*stdemux
= (struct stdemux
*)demux
->priv
;
260 struct c8sectpfei
*fei
= stdemux
->c8sectpfei
;
261 struct channel_info
*channel
;
265 unsigned long *bitmap
;
267 if (!atomic_read(&fei
->fw_loaded
)) {
268 dev_err(fei
->dev
, "%s: c8sectpfe fw not loaded\n", __func__
);
272 mutex_lock(&fei
->lock
);
274 channel
= fei
->channel_data
[stdemux
->tsin_index
];
276 bitmap
= (unsigned long *) channel
->pid_buffer_aligned
;
278 if (dvbdmxfeed
->pid
== 8192) {
279 tmp
= readl(fei
->io
+ C8SECTPFE_IB_PID_SET(channel
->tsin_id
));
280 tmp
|= C8SECTPFE_PID_ENABLE
;
281 writel(tmp
, fei
->io
+ C8SECTPFE_IB_PID_SET(channel
->tsin_id
));
283 bitmap_clear(bitmap
, dvbdmxfeed
->pid
, 1);
286 /* manage cache so data is visible to HW */
287 dma_sync_single_for_device(fei
->dev
,
288 channel
->pid_buffer_busaddr
,
292 if (--stdemux
->running_feed_count
== 0) {
294 channel
= fei
->channel_data
[stdemux
->tsin_index
];
296 /* TP re-configuration on page 168 of functional spec */
298 /* disable IB (prevents more TS data going to memdma) */
299 writel(0, fei
->io
+ C8SECTPFE_IB_SYS(channel
->tsin_id
));
301 /* disable this channels descriptor */
302 writel(0, channel
->irec
+ DMA_PRDS_TPENABLE
);
304 tasklet_disable(&channel
->tsklet
);
306 /* now request memdma channel goes idle */
307 idlereq
= (1 << channel
->tsin_id
) | IDLEREQ
;
308 writel(idlereq
, fei
->io
+ DMA_IDLE_REQ
);
310 /* wait for idle irq handler to signal completion */
311 ret
= wait_for_completion_timeout(&channel
->idle_completion
,
312 msecs_to_jiffies(100));
316 "Timeout waiting for idle irq on tsin%d\n",
319 reinit_completion(&channel
->idle_completion
);
321 /* reset read / write ptrs for this channel */
323 writel(channel
->back_buffer_busaddr
,
324 channel
->irec
+ DMA_PRDS_BUSBASE_TP(0));
326 tmp
= channel
->back_buffer_busaddr
+ FEI_BUFFER_SIZE
- 1;
327 writel(tmp
, channel
->irec
+ DMA_PRDS_BUSTOP_TP(0));
329 writel(channel
->back_buffer_busaddr
,
330 channel
->irec
+ DMA_PRDS_BUSWP_TP(0));
333 "%s:%d stopping DMA feed on stdemux=%p channel=%d\n",
334 __func__
, __LINE__
, stdemux
, channel
->tsin_id
);
336 /* turn off all PIDS in the bitmap */
337 memset((void *)channel
->pid_buffer_aligned
338 , 0x00, PID_TABLE_SIZE
);
340 /* manage cache so data is visible to HW */
341 dma_sync_single_for_device(fei
->dev
,
342 channel
->pid_buffer_busaddr
,
349 if (--fei
->global_feed_count
== 0) {
350 dev_dbg(fei
->dev
, "%s:%d global_feed_count=%d\n"
351 , __func__
, __LINE__
, fei
->global_feed_count
);
353 del_timer(&fei
->timer
);
356 mutex_unlock(&fei
->lock
);
361 static struct channel_info
*find_channel(struct c8sectpfei
*fei
, int tsin_num
)
365 for (i
= 0; i
< C8SECTPFE_MAX_TSIN_CHAN
; i
++) {
366 if (!fei
->channel_data
[i
])
369 if (fei
->channel_data
[i
]->tsin_id
== tsin_num
)
370 return fei
->channel_data
[i
];
376 static void c8sectpfe_getconfig(struct c8sectpfei
*fei
)
378 struct c8sectpfe_hw
*hw
= &fei
->hw_stats
;
380 hw
->num_ib
= readl(fei
->io
+ SYS_CFG_NUM_IB
);
381 hw
->num_mib
= readl(fei
->io
+ SYS_CFG_NUM_MIB
);
382 hw
->num_swts
= readl(fei
->io
+ SYS_CFG_NUM_SWTS
);
383 hw
->num_tsout
= readl(fei
->io
+ SYS_CFG_NUM_TSOUT
);
384 hw
->num_ccsc
= readl(fei
->io
+ SYS_CFG_NUM_CCSC
);
385 hw
->num_ram
= readl(fei
->io
+ SYS_CFG_NUM_RAM
);
386 hw
->num_tp
= readl(fei
->io
+ SYS_CFG_NUM_TP
);
388 dev_info(fei
->dev
, "C8SECTPFE hw supports the following:\n");
389 dev_info(fei
->dev
, "Input Blocks: %d\n", hw
->num_ib
);
390 dev_info(fei
->dev
, "Merged Input Blocks: %d\n", hw
->num_mib
);
391 dev_info(fei
->dev
, "Software Transport Stream Inputs: %d\n"
393 dev_info(fei
->dev
, "Transport Stream Output: %d\n", hw
->num_tsout
);
394 dev_info(fei
->dev
, "Cable Card Converter: %d\n", hw
->num_ccsc
);
395 dev_info(fei
->dev
, "RAMs supported by C8SECTPFE: %d\n", hw
->num_ram
);
396 dev_info(fei
->dev
, "Tango TPs supported by C8SECTPFE: %d\n"
400 static irqreturn_t
c8sectpfe_idle_irq_handler(int irq
, void *priv
)
402 struct c8sectpfei
*fei
= priv
;
403 struct channel_info
*chan
;
405 unsigned long tmp
= readl(fei
->io
+ DMA_IDLE_REQ
);
407 /* page 168 of functional spec: Clear the idle request
408 by writing 0 to the C8SECTPFE_DMA_IDLE_REQ register. */
410 /* signal idle completion */
411 for_each_set_bit(bit
, &tmp
, fei
->hw_stats
.num_ib
) {
413 chan
= find_channel(fei
, bit
);
416 complete(&chan
->idle_completion
);
419 writel(0, fei
->io
+ DMA_IDLE_REQ
);
425 static void free_input_block(struct c8sectpfei
*fei
, struct channel_info
*tsin
)
430 if (tsin
->back_buffer_busaddr
)
431 if (!dma_mapping_error(fei
->dev
, tsin
->back_buffer_busaddr
))
432 dma_unmap_single(fei
->dev
, tsin
->back_buffer_busaddr
,
433 FEI_BUFFER_SIZE
, DMA_BIDIRECTIONAL
);
435 kfree(tsin
->back_buffer_start
);
437 if (tsin
->pid_buffer_busaddr
)
438 if (!dma_mapping_error(fei
->dev
, tsin
->pid_buffer_busaddr
))
439 dma_unmap_single(fei
->dev
, tsin
->pid_buffer_busaddr
,
440 PID_TABLE_SIZE
, DMA_BIDIRECTIONAL
);
442 kfree(tsin
->pid_buffer_start
);
447 static int configure_memdma_and_inputblock(struct c8sectpfei
*fei
,
448 struct channel_info
*tsin
)
452 char tsin_pin_name
[MAX_NAME
];
457 dev_dbg(fei
->dev
, "%s:%d Configuring channel=%p tsin=%d\n"
458 , __func__
, __LINE__
, tsin
, tsin
->tsin_id
);
460 init_completion(&tsin
->idle_completion
);
462 tsin
->back_buffer_start
= kzalloc(FEI_BUFFER_SIZE
+
463 FEI_ALIGNMENT
, GFP_KERNEL
);
465 if (!tsin
->back_buffer_start
) {
470 /* Ensure backbuffer is 32byte aligned */
471 tsin
->back_buffer_aligned
= tsin
->back_buffer_start
474 tsin
->back_buffer_aligned
= (void *)
475 (((uintptr_t) tsin
->back_buffer_aligned
) & ~0x1F);
477 tsin
->back_buffer_busaddr
= dma_map_single(fei
->dev
,
478 (void *)tsin
->back_buffer_aligned
,
482 if (dma_mapping_error(fei
->dev
, tsin
->back_buffer_busaddr
)) {
483 dev_err(fei
->dev
, "failed to map back_buffer\n");
489 * The pid buffer can be configured (in hw) for byte or bit
490 * per pid. By powers of deduction we conclude stih407 family
491 * is configured (at SoC design stage) for bit per pid.
493 tsin
->pid_buffer_start
= kzalloc(2048, GFP_KERNEL
);
495 if (!tsin
->pid_buffer_start
) {
501 * PID buffer needs to be aligned to size of the pid table
502 * which at bit per pid is 1024 bytes (8192 pids / 8).
503 * PIDF_BASE register enforces this alignment when writing
507 tsin
->pid_buffer_aligned
= tsin
->pid_buffer_start
+
510 tsin
->pid_buffer_aligned
= (void *)
511 (((uintptr_t) tsin
->pid_buffer_aligned
) & ~0x3ff);
513 tsin
->pid_buffer_busaddr
= dma_map_single(fei
->dev
,
514 tsin
->pid_buffer_aligned
,
518 if (dma_mapping_error(fei
->dev
, tsin
->pid_buffer_busaddr
)) {
519 dev_err(fei
->dev
, "failed to map pid_bitmap\n");
524 /* manage cache so pid bitmap is visible to HW */
525 dma_sync_single_for_device(fei
->dev
,
526 tsin
->pid_buffer_busaddr
,
530 snprintf(tsin_pin_name
, MAX_NAME
, "tsin%d-%s", tsin
->tsin_id
,
531 (tsin
->serial_not_parallel
? "serial" : "parallel"));
533 tsin
->pstate
= pinctrl_lookup_state(fei
->pinctrl
, tsin_pin_name
);
534 if (IS_ERR(tsin
->pstate
)) {
535 dev_err(fei
->dev
, "%s: pinctrl_lookup_state couldn't find %s state\n"
536 , __func__
, tsin_pin_name
);
537 ret
= PTR_ERR(tsin
->pstate
);
541 ret
= pinctrl_select_state(fei
->pinctrl
, tsin
->pstate
);
544 dev_err(fei
->dev
, "%s: pinctrl_select_state failed\n"
549 /* Enable this input block */
550 tmp
= readl(fei
->io
+ SYS_INPUT_CLKEN
);
551 tmp
|= BIT(tsin
->tsin_id
);
552 writel(tmp
, fei
->io
+ SYS_INPUT_CLKEN
);
554 if (tsin
->serial_not_parallel
)
555 tmp
|= C8SECTPFE_SERIAL_NOT_PARALLEL
;
557 if (tsin
->invert_ts_clk
)
558 tmp
|= C8SECTPFE_INVERT_TSCLK
;
560 if (tsin
->async_not_sync
)
561 tmp
|= C8SECTPFE_ASYNC_NOT_SYNC
;
563 tmp
|= C8SECTPFE_ALIGN_BYTE_SOP
| C8SECTPFE_BYTE_ENDIANNESS_MSB
;
565 writel(tmp
, fei
->io
+ C8SECTPFE_IB_IP_FMT_CFG(tsin
->tsin_id
));
567 writel(C8SECTPFE_SYNC(0x9) |
568 C8SECTPFE_DROP(0x9) |
569 C8SECTPFE_TOKEN(0x47),
570 fei
->io
+ C8SECTPFE_IB_SYNCLCKDRP_CFG(tsin
->tsin_id
));
572 writel(TS_PKT_SIZE
, fei
->io
+ C8SECTPFE_IB_PKT_LEN(tsin
->tsin_id
));
574 /* Place the FIFO's at the end of the irec descriptors */
576 tsin
->fifo
= (tsin
->tsin_id
* FIFO_LEN
);
578 writel(tsin
->fifo
, fei
->io
+ C8SECTPFE_IB_BUFF_STRT(tsin
->tsin_id
));
579 writel(tsin
->fifo
+ FIFO_LEN
- 1,
580 fei
->io
+ C8SECTPFE_IB_BUFF_END(tsin
->tsin_id
));
582 writel(tsin
->fifo
, fei
->io
+ C8SECTPFE_IB_READ_PNT(tsin
->tsin_id
));
583 writel(tsin
->fifo
, fei
->io
+ C8SECTPFE_IB_WRT_PNT(tsin
->tsin_id
));
585 writel(tsin
->pid_buffer_busaddr
,
586 fei
->io
+ PIDF_BASE(tsin
->tsin_id
));
588 dev_info(fei
->dev
, "chan=%d PIDF_BASE=0x%x pid_bus_addr=%pad\n",
589 tsin
->tsin_id
, readl(fei
->io
+ PIDF_BASE(tsin
->tsin_id
)),
590 &tsin
->pid_buffer_busaddr
);
592 /* Configure and enable HW PID filtering */
595 * The PID value is created by assembling the first 8 bytes of
596 * the TS packet into a 64-bit word in big-endian format. A
597 * slice of that 64-bit word is taken from
598 * (PID_OFFSET+PID_NUM_BITS-1) to PID_OFFSET.
600 tmp
= (C8SECTPFE_PID_ENABLE
| C8SECTPFE_PID_NUMBITS(13)
601 | C8SECTPFE_PID_OFFSET(40));
603 writel(tmp
, fei
->io
+ C8SECTPFE_IB_PID_SET(tsin
->tsin_id
));
605 dev_dbg(fei
->dev
, "chan=%d setting wp: %d, rp: %d, buf: %d-%d\n",
607 readl(fei
->io
+ C8SECTPFE_IB_WRT_PNT(tsin
->tsin_id
)),
608 readl(fei
->io
+ C8SECTPFE_IB_READ_PNT(tsin
->tsin_id
)),
609 readl(fei
->io
+ C8SECTPFE_IB_BUFF_STRT(tsin
->tsin_id
)),
610 readl(fei
->io
+ C8SECTPFE_IB_BUFF_END(tsin
->tsin_id
)));
612 /* Get base addpress of pointer record block from DMEM */
613 tsin
->irec
= fei
->io
+ DMA_MEMDMA_OFFSET
+ DMA_DMEM_OFFSET
+
614 readl(fei
->io
+ DMA_PTRREC_BASE
);
616 /* fill out pointer record data structure */
618 /* advance pointer record block to our channel */
619 tsin
->irec
+= (tsin
->tsin_id
* DMA_PRDS_SIZE
);
621 writel(tsin
->fifo
, tsin
->irec
+ DMA_PRDS_MEMBASE
);
623 writel(tsin
->fifo
+ FIFO_LEN
- 1, tsin
->irec
+ DMA_PRDS_MEMTOP
);
625 writel((188 + 7)&~7, tsin
->irec
+ DMA_PRDS_PKTSIZE
);
627 writel(0x1, tsin
->irec
+ DMA_PRDS_TPENABLE
);
629 /* read/write pointers with physical bus address */
631 writel(tsin
->back_buffer_busaddr
, tsin
->irec
+ DMA_PRDS_BUSBASE_TP(0));
633 tmp
= tsin
->back_buffer_busaddr
+ FEI_BUFFER_SIZE
- 1;
634 writel(tmp
, tsin
->irec
+ DMA_PRDS_BUSTOP_TP(0));
636 writel(tsin
->back_buffer_busaddr
, tsin
->irec
+ DMA_PRDS_BUSWP_TP(0));
637 writel(tsin
->back_buffer_busaddr
, tsin
->irec
+ DMA_PRDS_BUSRP_TP(0));
639 /* initialize tasklet */
640 tasklet_init(&tsin
->tsklet
, channel_swdemux_tsklet
,
641 (unsigned long) tsin
);
646 free_input_block(fei
, tsin
);
650 static irqreturn_t
c8sectpfe_error_irq_handler(int irq
, void *priv
)
652 struct c8sectpfei
*fei
= priv
;
654 dev_err(fei
->dev
, "%s: error handling not yet implemented\n"
658 * TODO FIXME we should detect some error conditions here
659 * and ideally so something about them!
665 static int c8sectpfe_probe(struct platform_device
*pdev
)
667 struct device
*dev
= &pdev
->dev
;
668 struct device_node
*child
, *np
= dev
->of_node
;
669 struct c8sectpfei
*fei
;
670 struct resource
*res
;
672 struct channel_info
*tsin
;
674 /* Allocate the c8sectpfei structure */
675 fei
= devm_kzalloc(dev
, sizeof(struct c8sectpfei
), GFP_KERNEL
);
681 res
= platform_get_resource_byname(pdev
, IORESOURCE_MEM
, "c8sectpfe");
682 fei
->io
= devm_ioremap_resource(dev
, res
);
684 return PTR_ERR(fei
->io
);
686 res
= platform_get_resource_byname(pdev
, IORESOURCE_MEM
,
688 fei
->sram
= devm_ioremap_resource(dev
, res
);
689 if (IS_ERR(fei
->sram
))
690 return PTR_ERR(fei
->sram
);
692 fei
->sram_size
= res
->end
- res
->start
;
694 fei
->idle_irq
= platform_get_irq_byname(pdev
, "c8sectpfe-idle-irq");
695 if (fei
->idle_irq
< 0) {
696 dev_err(dev
, "Can't get c8sectpfe-idle-irq\n");
697 return fei
->idle_irq
;
700 fei
->error_irq
= platform_get_irq_byname(pdev
, "c8sectpfe-error-irq");
701 if (fei
->error_irq
< 0) {
702 dev_err(dev
, "Can't get c8sectpfe-error-irq\n");
703 return fei
->error_irq
;
706 platform_set_drvdata(pdev
, fei
);
708 fei
->c8sectpfeclk
= devm_clk_get(dev
, "c8sectpfe");
709 if (IS_ERR(fei
->c8sectpfeclk
)) {
710 dev_err(dev
, "c8sectpfe clk not found\n");
711 return PTR_ERR(fei
->c8sectpfeclk
);
714 ret
= clk_prepare_enable(fei
->c8sectpfeclk
);
716 dev_err(dev
, "Failed to enable c8sectpfe clock\n");
720 /* to save power disable all IP's (on by default) */
721 writel(0, fei
->io
+ SYS_INPUT_CLKEN
);
723 /* Enable memdma clock */
724 writel(MEMDMAENABLE
, fei
->io
+ SYS_OTHER_CLKEN
);
726 /* clear internal sram */
727 memset_io(fei
->sram
, 0x0, fei
->sram_size
);
729 c8sectpfe_getconfig(fei
);
731 ret
= devm_request_irq(dev
, fei
->idle_irq
, c8sectpfe_idle_irq_handler
,
732 0, "c8sectpfe-idle-irq", fei
);
734 dev_err(dev
, "Can't register c8sectpfe-idle-irq IRQ.\n");
735 goto err_clk_disable
;
738 ret
= devm_request_irq(dev
, fei
->error_irq
,
739 c8sectpfe_error_irq_handler
, 0,
740 "c8sectpfe-error-irq", fei
);
742 dev_err(dev
, "Can't register c8sectpfe-error-irq IRQ.\n");
743 goto err_clk_disable
;
746 fei
->tsin_count
= of_get_child_count(np
);
748 if (fei
->tsin_count
> C8SECTPFE_MAX_TSIN_CHAN
||
749 fei
->tsin_count
> fei
->hw_stats
.num_ib
) {
751 dev_err(dev
, "More tsin declared than exist on SoC!\n");
753 goto err_clk_disable
;
756 fei
->pinctrl
= devm_pinctrl_get(dev
);
758 if (IS_ERR(fei
->pinctrl
)) {
759 dev_err(dev
, "Error getting tsin pins\n");
760 ret
= PTR_ERR(fei
->pinctrl
);
761 goto err_clk_disable
;
764 for_each_child_of_node(np
, child
) {
765 struct device_node
*i2c_bus
;
767 fei
->channel_data
[index
] = devm_kzalloc(dev
,
768 sizeof(struct channel_info
),
771 if (!fei
->channel_data
[index
]) {
773 goto err_clk_disable
;
776 tsin
= fei
->channel_data
[index
];
780 ret
= of_property_read_u32(child
, "tsin-num", &tsin
->tsin_id
);
782 dev_err(&pdev
->dev
, "No tsin_num found\n");
783 goto err_clk_disable
;
786 /* sanity check value */
787 if (tsin
->tsin_id
> fei
->hw_stats
.num_ib
) {
789 "tsin-num %d specified greater than number\n\t"
790 "of input block hw in SoC! (%d)",
791 tsin
->tsin_id
, fei
->hw_stats
.num_ib
);
793 goto err_clk_disable
;
796 tsin
->invert_ts_clk
= of_property_read_bool(child
,
799 tsin
->serial_not_parallel
= of_property_read_bool(child
,
800 "serial-not-parallel");
802 tsin
->async_not_sync
= of_property_read_bool(child
,
805 ret
= of_property_read_u32(child
, "dvb-card",
808 dev_err(&pdev
->dev
, "No dvb-card found\n");
809 goto err_clk_disable
;
812 i2c_bus
= of_parse_phandle(child
, "i2c-bus", 0);
814 dev_err(&pdev
->dev
, "No i2c-bus found\n");
815 goto err_clk_disable
;
818 of_find_i2c_adapter_by_node(i2c_bus
);
819 if (!tsin
->i2c_adapter
) {
820 dev_err(&pdev
->dev
, "No i2c adapter found\n");
821 of_node_put(i2c_bus
);
822 goto err_clk_disable
;
824 of_node_put(i2c_bus
);
826 tsin
->rst_gpio
= of_get_named_gpio(child
, "rst-gpio", 0);
828 ret
= gpio_is_valid(tsin
->rst_gpio
);
831 "reset gpio for tsin%d not valid (gpio=%d)\n",
832 tsin
->tsin_id
, tsin
->rst_gpio
);
833 goto err_clk_disable
;
836 ret
= devm_gpio_request_one(dev
, tsin
->rst_gpio
,
837 GPIOF_OUT_INIT_LOW
, "NIM reset");
838 if (ret
&& ret
!= -EBUSY
) {
839 dev_err(dev
, "Can't request tsin%d reset gpio\n"
840 , fei
->channel_data
[index
]->tsin_id
);
841 goto err_clk_disable
;
845 /* toggle reset lines */
846 gpio_direction_output(tsin
->rst_gpio
, 0);
847 usleep_range(3500, 5000);
848 gpio_direction_output(tsin
->rst_gpio
, 1);
849 usleep_range(3000, 5000);
852 tsin
->demux_mapping
= index
;
855 "channel=%p n=%d tsin_num=%d, invert-ts-clk=%d\n\t"
856 "serial-not-parallel=%d pkt-clk-valid=%d dvb-card=%d\n",
857 fei
->channel_data
[index
], index
,
858 tsin
->tsin_id
, tsin
->invert_ts_clk
,
859 tsin
->serial_not_parallel
, tsin
->async_not_sync
,
865 /* Setup timer interrupt */
866 init_timer(&fei
->timer
);
867 fei
->timer
.function
= c8sectpfe_timer_interrupt
;
868 fei
->timer
.data
= (unsigned long)fei
;
870 mutex_init(&fei
->lock
);
872 /* Get the configuration information about the tuners */
873 ret
= c8sectpfe_tuner_register_frontend(&fei
->c8sectpfe
[0],
875 c8sectpfe_start_feed
,
876 c8sectpfe_stop_feed
);
878 dev_err(dev
, "c8sectpfe_tuner_register_frontend failed (%d)\n",
880 goto err_clk_disable
;
883 /* ensure all other init has been done before requesting firmware */
884 ret
= load_c8sectpfe_fw_step1(fei
);
886 dev_err(dev
, "Couldn't load slim core firmware\n");
887 goto err_clk_disable
;
890 c8sectpfe_debugfs_init(fei
);
895 /* TODO uncomment when upstream has taken a reference on this clk */
896 /*clk_disable_unprepare(fei->c8sectpfeclk);*/
900 static int c8sectpfe_remove(struct platform_device
*pdev
)
902 struct c8sectpfei
*fei
= platform_get_drvdata(pdev
);
903 struct channel_info
*channel
;
906 wait_for_completion(&fei
->fw_ack
);
908 c8sectpfe_tuner_unregister_frontend(fei
->c8sectpfe
[0], fei
);
911 * Now loop through and un-configure each of the InputBlock resources
913 for (i
= 0; i
< fei
->tsin_count
; i
++) {
914 channel
= fei
->channel_data
[i
];
915 free_input_block(fei
, channel
);
918 c8sectpfe_debugfs_exit(fei
);
920 dev_info(fei
->dev
, "Stopping memdma SLIM core\n");
921 if (readl(fei
->io
+ DMA_CPU_RUN
))
922 writel(0x0, fei
->io
+ DMA_CPU_RUN
);
924 /* unclock all internal IP's */
925 if (readl(fei
->io
+ SYS_INPUT_CLKEN
))
926 writel(0, fei
->io
+ SYS_INPUT_CLKEN
);
928 if (readl(fei
->io
+ SYS_OTHER_CLKEN
))
929 writel(0, fei
->io
+ SYS_OTHER_CLKEN
);
931 /* TODO uncomment when upstream has taken a reference on this clk */
933 if (fei->c8sectpfeclk)
934 clk_disable_unprepare(fei->c8sectpfeclk);
941 static int configure_channels(struct c8sectpfei
*fei
)
944 struct channel_info
*tsin
;
945 struct device_node
*child
, *np
= fei
->dev
->of_node
;
947 /* iterate round each tsin and configure memdma descriptor and IB hw */
948 for_each_child_of_node(np
, child
) {
950 tsin
= fei
->channel_data
[index
];
952 ret
= configure_memdma_and_inputblock(fei
,
953 fei
->channel_data
[index
]);
957 "configure_memdma_and_inputblock failed\n");
966 for (index
= 0; index
< fei
->tsin_count
; index
++) {
967 tsin
= fei
->channel_data
[index
];
968 free_input_block(fei
, tsin
);
974 c8sectpfe_elf_sanity_check(struct c8sectpfei
*fei
, const struct firmware
*fw
)
976 struct elf32_hdr
*ehdr
;
980 dev_err(fei
->dev
, "failed to load %s\n", FIRMWARE_MEMDMA
);
984 if (fw
->size
< sizeof(struct elf32_hdr
)) {
985 dev_err(fei
->dev
, "Image is too small\n");
989 ehdr
= (struct elf32_hdr
*)fw
->data
;
991 /* We only support ELF32 at this point */
992 class = ehdr
->e_ident
[EI_CLASS
];
993 if (class != ELFCLASS32
) {
994 dev_err(fei
->dev
, "Unsupported class: %d\n", class);
998 if (ehdr
->e_ident
[EI_DATA
] != ELFDATA2LSB
) {
999 dev_err(fei
->dev
, "Unsupported firmware endianness\n");
1003 if (fw
->size
< ehdr
->e_shoff
+ sizeof(struct elf32_shdr
)) {
1004 dev_err(fei
->dev
, "Image is too small\n");
1008 if (memcmp(ehdr
->e_ident
, ELFMAG
, SELFMAG
)) {
1009 dev_err(fei
->dev
, "Image is corrupted (bad magic)\n");
1013 /* Check ELF magic */
1014 ehdr
= (Elf32_Ehdr
*)fw
->data
;
1015 if (ehdr
->e_ident
[EI_MAG0
] != ELFMAG0
||
1016 ehdr
->e_ident
[EI_MAG1
] != ELFMAG1
||
1017 ehdr
->e_ident
[EI_MAG2
] != ELFMAG2
||
1018 ehdr
->e_ident
[EI_MAG3
] != ELFMAG3
) {
1019 dev_err(fei
->dev
, "Invalid ELF magic\n");
1023 if (ehdr
->e_type
!= ET_EXEC
) {
1024 dev_err(fei
->dev
, "Unsupported ELF header type\n");
1028 if (ehdr
->e_phoff
> fw
->size
) {
1029 dev_err(fei
->dev
, "Firmware size is too small\n");
1037 static void load_imem_segment(struct c8sectpfei
*fei
, Elf32_Phdr
*phdr
,
1038 const struct firmware
*fw
, u8 __iomem
*dest
,
1041 const u8
*imem_src
= fw
->data
+ phdr
->p_offset
;
1045 * For IMEM segments, the segment contains 24-bit
1046 * instructions which must be padded to 32-bit
1047 * instructions before being written. The written
1048 * segment is padded with NOP instructions.
1052 "Loading IMEM segment %d 0x%08x\n\t"
1053 " (0x%x bytes) -> 0x%p (0x%x bytes)\n", seg_num
,
1054 phdr
->p_paddr
, phdr
->p_filesz
,
1055 dest
, phdr
->p_memsz
+ phdr
->p_memsz
/ 3);
1057 for (i
= 0; i
< phdr
->p_filesz
; i
++) {
1059 writeb(readb((void __iomem
*)imem_src
), (void __iomem
*)dest
);
1061 /* Every 3 bytes, add an additional
1062 * padding zero in destination */
1065 writeb(0x00, (void __iomem
*)dest
);
1073 static void load_dmem_segment(struct c8sectpfei
*fei
, Elf32_Phdr
*phdr
,
1074 const struct firmware
*fw
, u8 __iomem
*dst
, int seg_num
)
1077 * For DMEM segments copy the segment data from the ELF
1078 * file and pad segment with zeroes
1082 "Loading DMEM segment %d 0x%08x\n\t"
1083 "(0x%x bytes) -> 0x%p (0x%x bytes)\n",
1084 seg_num
, phdr
->p_paddr
, phdr
->p_filesz
,
1085 dst
, phdr
->p_memsz
);
1087 memcpy((void __force
*)dst
, (void *)fw
->data
+ phdr
->p_offset
,
1090 memset((void __force
*)dst
+ phdr
->p_filesz
, 0,
1091 phdr
->p_memsz
- phdr
->p_filesz
);
1094 static int load_slim_core_fw(const struct firmware
*fw
, void *context
)
1096 struct c8sectpfei
*fei
= context
;
1102 if (!fw
|| !context
)
1105 ehdr
= (Elf32_Ehdr
*)fw
->data
;
1106 phdr
= (Elf32_Phdr
*)(fw
->data
+ ehdr
->e_phoff
);
1108 /* go through the available ELF segments */
1109 for (i
= 0; i
< ehdr
->e_phnum
; i
++, phdr
++) {
1111 /* Only consider LOAD segments */
1112 if (phdr
->p_type
!= PT_LOAD
)
1116 * Check segment is contained within the fw->data buffer
1118 if (phdr
->p_offset
+ phdr
->p_filesz
> fw
->size
) {
1120 "Segment %d is outside of firmware file\n", i
);
1126 * MEMDMA IMEM has executable flag set, otherwise load
1127 * this segment into DMEM.
1131 if (phdr
->p_flags
& PF_X
) {
1132 dst
= (u8 __iomem
*) fei
->io
+ DMA_MEMDMA_IMEM
;
1134 * The Slim ELF file uses 32-bit word addressing for
1137 dst
+= (phdr
->p_paddr
& 0xFFFFF) * sizeof(unsigned int);
1138 load_imem_segment(fei
, phdr
, fw
, dst
, i
);
1140 dst
= (u8 __iomem
*) fei
->io
+ DMA_MEMDMA_DMEM
;
1142 * The Slim ELF file uses 32-bit word addressing for
1145 dst
+= (phdr
->p_paddr
& 0xFFFFF) * sizeof(unsigned int);
1146 load_dmem_segment(fei
, phdr
, fw
, dst
, i
);
1150 release_firmware(fw
);
1154 static void load_c8sectpfe_fw_cb(const struct firmware
*fw
, void *context
)
1156 struct c8sectpfei
*fei
= context
;
1159 err
= c8sectpfe_elf_sanity_check(fei
, fw
);
1161 dev_err(fei
->dev
, "c8sectpfe_elf_sanity_check failed err=(%d)\n"
1166 err
= load_slim_core_fw(fw
, context
);
1168 dev_err(fei
->dev
, "load_slim_core_fw failed err=(%d)\n", err
);
1172 /* now the firmware is loaded configure the input blocks */
1173 err
= configure_channels(fei
);
1175 dev_err(fei
->dev
, "configure_channels failed err=(%d)\n", err
);
1180 * STBus target port can access IMEM and DMEM ports
1181 * without waiting for CPU
1183 writel(0x1, fei
->io
+ DMA_PER_STBUS_SYNC
);
1185 dev_info(fei
->dev
, "Boot the memdma SLIM core\n");
1186 writel(0x1, fei
->io
+ DMA_CPU_RUN
);
1188 atomic_set(&fei
->fw_loaded
, 1);
1190 complete_all(&fei
->fw_ack
);
1193 static int load_c8sectpfe_fw_step1(struct c8sectpfei
*fei
)
1197 dev_info(fei
->dev
, "Loading firmware: %s\n", FIRMWARE_MEMDMA
);
1199 init_completion(&fei
->fw_ack
);
1200 atomic_set(&fei
->fw_loaded
, 0);
1202 err
= request_firmware_nowait(THIS_MODULE
, FW_ACTION_HOTPLUG
,
1203 FIRMWARE_MEMDMA
, fei
->dev
, GFP_KERNEL
, fei
,
1204 load_c8sectpfe_fw_cb
);
1207 dev_err(fei
->dev
, "request_firmware_nowait err: %d.\n", err
);
1208 complete_all(&fei
->fw_ack
);
1215 static const struct of_device_id c8sectpfe_match
[] = {
1216 { .compatible
= "st,stih407-c8sectpfe" },
1219 MODULE_DEVICE_TABLE(of
, c8sectpfe_match
);
1221 static struct platform_driver c8sectpfe_driver
= {
1223 .name
= "c8sectpfe",
1224 .of_match_table
= of_match_ptr(c8sectpfe_match
),
1226 .probe
= c8sectpfe_probe
,
1227 .remove
= c8sectpfe_remove
,
1230 module_platform_driver(c8sectpfe_driver
);
1232 MODULE_AUTHOR("Peter Bennett <peter.bennett@st.com>");
1233 MODULE_AUTHOR("Peter Griffin <peter.griffin@linaro.org>");
1234 MODULE_DESCRIPTION("C8SECTPFE STi DVB Driver");
1235 MODULE_LICENSE("GPL");