1 // SPDX-License-Identifier: GPL-2.0-or-later
5 * Derived from ivtv-queue.c
7 * Copyright (C) 2007 Hans Verkuil <hverkuil@xs4all.nl>
8 * Copyright (C) 2008 Andy Walls <awalls@md.metrocast.net>
11 #include "cx18-driver.h"
12 #include "cx18-queue.h"
13 #include "cx18-streams.h"
17 void cx18_buf_swap(struct cx18_buffer
*buf
)
21 for (i
= 0; i
< buf
->bytesused
; i
+= 4)
22 swab32s((u32
*)(buf
->buf
+ i
));
25 void _cx18_mdl_swap(struct cx18_mdl
*mdl
)
27 struct cx18_buffer
*buf
;
29 list_for_each_entry(buf
, &mdl
->buf_list
, list
) {
30 if (buf
->bytesused
== 0)
36 void cx18_queue_init(struct cx18_queue
*q
)
38 INIT_LIST_HEAD(&q
->list
);
39 atomic_set(&q
->depth
, 0);
43 struct cx18_queue
*_cx18_enqueue(struct cx18_stream
*s
, struct cx18_mdl
*mdl
,
44 struct cx18_queue
*q
, int to_front
)
46 /* clear the mdl if it is not to be enqueued to the full queue */
47 if (q
!= &s
->q_full
) {
55 /* q_busy is restricted to a max buffer count imposed by firmware */
56 if (q
== &s
->q_busy
&&
57 atomic_read(&q
->depth
) >= CX18_MAX_FW_MDLS_PER_STREAM
)
63 list_add(&mdl
->list
, &q
->list
); /* LIFO */
65 list_add_tail(&mdl
->list
, &q
->list
); /* FIFO */
66 q
->bytesused
+= mdl
->bytesused
- mdl
->readpos
;
67 atomic_inc(&q
->depth
);
69 spin_unlock(&q
->lock
);
73 struct cx18_mdl
*cx18_dequeue(struct cx18_stream
*s
, struct cx18_queue
*q
)
75 struct cx18_mdl
*mdl
= NULL
;
78 if (!list_empty(&q
->list
)) {
79 mdl
= list_first_entry(&q
->list
, struct cx18_mdl
, list
);
80 list_del_init(&mdl
->list
);
81 q
->bytesused
-= mdl
->bytesused
- mdl
->readpos
;
83 atomic_dec(&q
->depth
);
85 spin_unlock(&q
->lock
);
89 static void _cx18_mdl_update_bufs_for_cpu(struct cx18_stream
*s
,
92 struct cx18_buffer
*buf
;
93 u32 buf_size
= s
->buf_size
;
94 u32 bytesused
= mdl
->bytesused
;
96 list_for_each_entry(buf
, &mdl
->buf_list
, list
) {
98 if (bytesused
>= buf_size
) {
99 buf
->bytesused
= buf_size
;
100 bytesused
-= buf_size
;
102 buf
->bytesused
= bytesused
;
105 cx18_buf_sync_for_cpu(s
, buf
);
109 static inline void cx18_mdl_update_bufs_for_cpu(struct cx18_stream
*s
,
110 struct cx18_mdl
*mdl
)
112 struct cx18_buffer
*buf
;
114 if (list_is_singular(&mdl
->buf_list
)) {
115 buf
= list_first_entry(&mdl
->buf_list
, struct cx18_buffer
,
117 buf
->bytesused
= mdl
->bytesused
;
119 cx18_buf_sync_for_cpu(s
, buf
);
121 _cx18_mdl_update_bufs_for_cpu(s
, mdl
);
125 struct cx18_mdl
*cx18_queue_get_mdl(struct cx18_stream
*s
, u32 id
,
128 struct cx18
*cx
= s
->cx
;
129 struct cx18_mdl
*mdl
;
130 struct cx18_mdl
*tmp
;
131 struct cx18_mdl
*ret
= NULL
;
135 * We don't have to acquire multiple q locks here, because we are
136 * serialized by the single threaded work handler.
137 * MDLs from the firmware will thus remain in order as
138 * they are moved from q_busy to q_full or to the dvb ring buffer.
140 spin_lock(&s
->q_busy
.lock
);
141 list_for_each_entry_safe(mdl
, tmp
, &s
->q_busy
.list
, list
) {
143 * We should find what the firmware told us is done,
144 * right at the front of the queue. If we don't, we likely have
145 * missed an mdl done message from the firmware.
146 * Once we skip an mdl repeatedly, relative to the size of
147 * q_busy, we have high confidence we've missed it.
151 if (mdl
->skipped
>= atomic_read(&s
->q_busy
.depth
)-1) {
152 /* mdl must have fallen out of rotation */
153 CX18_WARN("Skipped %s, MDL %d, %d times - it must have dropped out of rotation\n",
156 /* Sweep it up to put it back into rotation */
157 list_move_tail(&mdl
->list
, &sweep_up
);
158 atomic_dec(&s
->q_busy
.depth
);
163 * We pull the desired mdl off of the queue here. Something
164 * will have to put it back on a queue later.
166 list_del_init(&mdl
->list
);
167 atomic_dec(&s
->q_busy
.depth
);
171 spin_unlock(&s
->q_busy
.lock
);
174 * We found the mdl for which we were looking. Get it ready for
175 * the caller to put on q_full or in the dvb ring buffer.
178 ret
->bytesused
= bytesused
;
180 /* 0'ed readpos, m_flags & curr_buf when mdl went on q_busy */
181 cx18_mdl_update_bufs_for_cpu(s
, ret
);
182 if (s
->type
!= CX18_ENC_STREAM_TYPE_TS
)
183 set_bit(CX18_F_M_NEED_SWAP
, &ret
->m_flags
);
186 /* Put any mdls the firmware is ignoring back into normal rotation */
187 list_for_each_entry_safe(mdl
, tmp
, &sweep_up
, list
) {
188 list_del_init(&mdl
->list
);
189 cx18_enqueue(s
, mdl
, &s
->q_free
);
194 /* Move all mdls of a queue, while flushing the mdl */
195 static void cx18_queue_flush(struct cx18_stream
*s
,
196 struct cx18_queue
*q_src
, struct cx18_queue
*q_dst
)
198 struct cx18_mdl
*mdl
;
200 /* It only makes sense to flush to q_free or q_idle */
201 if (q_src
== q_dst
|| q_dst
== &s
->q_full
|| q_dst
== &s
->q_busy
)
204 spin_lock(&q_src
->lock
);
205 spin_lock(&q_dst
->lock
);
206 while (!list_empty(&q_src
->list
)) {
207 mdl
= list_first_entry(&q_src
->list
, struct cx18_mdl
, list
);
208 list_move_tail(&mdl
->list
, &q_dst
->list
);
213 mdl
->curr_buf
= NULL
;
214 atomic_inc(&q_dst
->depth
);
216 cx18_queue_init(q_src
);
217 spin_unlock(&q_src
->lock
);
218 spin_unlock(&q_dst
->lock
);
221 void cx18_flush_queues(struct cx18_stream
*s
)
223 cx18_queue_flush(s
, &s
->q_busy
, &s
->q_free
);
224 cx18_queue_flush(s
, &s
->q_full
, &s
->q_free
);
228 * Note, s->buf_pool is not protected by a lock,
229 * the stream better not have *anything* going on when calling this
231 void cx18_unload_queues(struct cx18_stream
*s
)
233 struct cx18_queue
*q_idle
= &s
->q_idle
;
234 struct cx18_mdl
*mdl
;
235 struct cx18_buffer
*buf
;
237 /* Move all MDLS to q_idle */
238 cx18_queue_flush(s
, &s
->q_busy
, q_idle
);
239 cx18_queue_flush(s
, &s
->q_full
, q_idle
);
240 cx18_queue_flush(s
, &s
->q_free
, q_idle
);
242 /* Reset MDL id's and move all buffers back to the stream's buf_pool */
243 spin_lock(&q_idle
->lock
);
244 list_for_each_entry(mdl
, &q_idle
->list
, list
) {
245 while (!list_empty(&mdl
->buf_list
)) {
246 buf
= list_first_entry(&mdl
->buf_list
,
247 struct cx18_buffer
, list
);
248 list_move_tail(&buf
->list
, &s
->buf_pool
);
252 mdl
->id
= s
->mdl_base_idx
; /* reset id to a "safe" value */
253 /* all other mdl fields were cleared by cx18_queue_flush() */
255 spin_unlock(&q_idle
->lock
);
259 * Note, s->buf_pool is not protected by a lock,
260 * the stream better not have *anything* going on when calling this
262 void cx18_load_queues(struct cx18_stream
*s
)
264 struct cx18
*cx
= s
->cx
;
265 struct cx18_mdl
*mdl
;
266 struct cx18_buffer
*buf
;
269 u32 partial_buf_size
;
272 * Attach buffers to MDLs, give the MDLs ids, and add MDLs to q_free
273 * Excess MDLs are left on q_idle
274 * Excess buffers are left in buf_pool and/or on an MDL in q_idle
276 mdl_id
= s
->mdl_base_idx
;
277 for (mdl
= cx18_dequeue(s
, &s
->q_idle
), i
= s
->bufs_per_mdl
;
278 mdl
!= NULL
&& i
== s
->bufs_per_mdl
;
279 mdl
= cx18_dequeue(s
, &s
->q_idle
)) {
283 for (i
= 0; i
< s
->bufs_per_mdl
; i
++) {
284 if (list_empty(&s
->buf_pool
))
287 buf
= list_first_entry(&s
->buf_pool
, struct cx18_buffer
,
289 list_move_tail(&buf
->list
, &mdl
->buf_list
);
291 /* update the firmware's MDL array with this buffer */
292 cx18_writel(cx
, buf
->dma_handle
,
293 &cx
->scb
->cpu_mdl
[mdl_id
+ i
].paddr
);
294 cx18_writel(cx
, s
->buf_size
,
295 &cx
->scb
->cpu_mdl
[mdl_id
+ i
].length
);
298 if (i
== s
->bufs_per_mdl
) {
300 * The encoder doesn't honor s->mdl_size. So in the
301 * case of a non-integral number of buffers to meet
302 * mdl_size, we lie about the size of the last buffer
303 * in the MDL to get the encoder to really only send
304 * us mdl_size bytes per MDL transfer.
306 partial_buf_size
= s
->mdl_size
% s
->buf_size
;
307 if (partial_buf_size
) {
308 cx18_writel(cx
, partial_buf_size
,
309 &cx
->scb
->cpu_mdl
[mdl_id
+ i
- 1].length
);
311 cx18_enqueue(s
, mdl
, &s
->q_free
);
313 /* Not enough buffers for this MDL; we won't use it */
314 cx18_push(s
, mdl
, &s
->q_idle
);
320 void _cx18_mdl_sync_for_device(struct cx18_stream
*s
, struct cx18_mdl
*mdl
)
323 u32 buf_size
= s
->buf_size
;
324 struct pci_dev
*pci_dev
= s
->cx
->pci_dev
;
325 struct cx18_buffer
*buf
;
327 list_for_each_entry(buf
, &mdl
->buf_list
, list
)
328 pci_dma_sync_single_for_device(pci_dev
, buf
->dma_handle
,
332 int cx18_stream_alloc(struct cx18_stream
*s
)
334 struct cx18
*cx
= s
->cx
;
340 CX18_DEBUG_INFO("Allocate %s stream: %d x %d buffers (%d.%02d kB total)\n",
341 s
->name
, s
->buffers
, s
->buf_size
,
342 s
->buffers
* s
->buf_size
/ 1024,
343 (s
->buffers
* s
->buf_size
* 100 / 1024) % 100);
345 if (((char __iomem
*)&cx
->scb
->cpu_mdl
[cx
->free_mdl_idx
+ s
->buffers
] -
346 (char __iomem
*)cx
->scb
) > SCB_RESERVED_SIZE
) {
347 unsigned bufsz
= (((char __iomem
*)cx
->scb
) + SCB_RESERVED_SIZE
-
348 ((char __iomem
*)cx
->scb
->cpu_mdl
));
350 CX18_ERR("Too many buffers, cannot fit in SCB area\n");
351 CX18_ERR("Max buffers = %zu\n",
352 bufsz
/ sizeof(struct cx18_mdl_ent
));
356 s
->mdl_base_idx
= cx
->free_mdl_idx
;
358 /* allocate stream buffers and MDLs */
359 for (i
= 0; i
< s
->buffers
; i
++) {
360 struct cx18_mdl
*mdl
;
361 struct cx18_buffer
*buf
;
363 /* 1 MDL per buffer to handle the worst & also default case */
364 mdl
= kzalloc(sizeof(struct cx18_mdl
), GFP_KERNEL
|__GFP_NOWARN
);
368 buf
= kzalloc(sizeof(struct cx18_buffer
),
369 GFP_KERNEL
|__GFP_NOWARN
);
375 buf
->buf
= kmalloc(s
->buf_size
, GFP_KERNEL
|__GFP_NOWARN
);
376 if (buf
->buf
== NULL
) {
382 INIT_LIST_HEAD(&mdl
->list
);
383 INIT_LIST_HEAD(&mdl
->buf_list
);
384 mdl
->id
= s
->mdl_base_idx
; /* a somewhat safe value */
385 cx18_enqueue(s
, mdl
, &s
->q_idle
);
387 INIT_LIST_HEAD(&buf
->list
);
388 buf
->dma_handle
= pci_map_single(s
->cx
->pci_dev
,
389 buf
->buf
, s
->buf_size
, s
->dma
);
390 cx18_buf_sync_for_cpu(s
, buf
);
391 list_add_tail(&buf
->list
, &s
->buf_pool
);
393 if (i
== s
->buffers
) {
394 cx
->free_mdl_idx
+= s
->buffers
;
397 CX18_ERR("Couldn't allocate buffers for %s stream\n", s
->name
);
402 void cx18_stream_free(struct cx18_stream
*s
)
404 struct cx18_mdl
*mdl
;
405 struct cx18_buffer
*buf
;
406 struct cx18
*cx
= s
->cx
;
408 CX18_DEBUG_INFO("Deallocating buffers for %s stream\n", s
->name
);
410 /* move all buffers to buf_pool and all MDLs to q_idle */
411 cx18_unload_queues(s
);
414 while ((mdl
= cx18_dequeue(s
, &s
->q_idle
)))
418 while (!list_empty(&s
->buf_pool
)) {
419 buf
= list_first_entry(&s
->buf_pool
, struct cx18_buffer
, list
);
420 list_del_init(&buf
->list
);
422 pci_unmap_single(s
->cx
->pci_dev
, buf
->dma_handle
,
423 s
->buf_size
, s
->dma
);