1 // SPDX-License-Identifier: GPL-2.0-only
3 * Keystone accumulator queue manager
5 * Copyright (C) 2014 Texas Instruments Incorporated - http://www.ti.com
6 * Author: Sandeep Nair <sandeep_n@ti.com>
7 * Cyril Chemparathy <cyril@ti.com>
8 * Santosh Shilimkar <santosh.shilimkar@ti.com>
11 #include <linux/dma-mapping.h>
13 #include <linux/interrupt.h>
14 #include <linux/module.h>
15 #include <linux/of_address.h>
16 #include <linux/soc/ti/knav_qmss.h>
18 #include "knav_qmss.h"
20 #define knav_range_offset_to_inst(kdev, range, q) \
21 (range->queue_base_inst + (q << kdev->inst_shift))
23 static void __knav_acc_notify(struct knav_range_info
*range
,
24 struct knav_acc_channel
*acc
)
26 struct knav_device
*kdev
= range
->kdev
;
27 struct knav_queue_inst
*inst
;
28 int range_base
, queue
;
30 range_base
= kdev
->base_id
+ range
->queue_base
;
32 if (range
->flags
& RANGE_MULTI_QUEUE
) {
33 for (queue
= 0; queue
< range
->num_queues
; queue
++) {
34 inst
= knav_range_offset_to_inst(kdev
, range
,
36 if (inst
->notify_needed
) {
37 inst
->notify_needed
= 0;
38 dev_dbg(kdev
->dev
, "acc-irq: notifying %d\n",
40 knav_queue_notify(inst
);
44 queue
= acc
->channel
- range
->acc_info
.start_channel
;
45 inst
= knav_range_offset_to_inst(kdev
, range
, queue
);
46 dev_dbg(kdev
->dev
, "acc-irq: notifying %d\n",
48 knav_queue_notify(inst
);
52 static int knav_acc_set_notify(struct knav_range_info
*range
,
53 struct knav_queue_inst
*kq
,
56 struct knav_pdsp_info
*pdsp
= range
->acc_info
.pdsp
;
57 struct knav_device
*kdev
= range
->kdev
;
61 * when enabling, we need to re-trigger an interrupt if we
62 * have descriptors pending
64 if (!enabled
|| atomic_read(&kq
->desc_count
) <= 0)
67 kq
->notify_needed
= 1;
68 atomic_inc(&kq
->acc
->retrigger_count
);
69 mask
= BIT(kq
->acc
->channel
% 32);
70 offset
= ACC_INTD_OFFSET_STATUS(kq
->acc
->channel
);
71 dev_dbg(kdev
->dev
, "setup-notify: re-triggering irq for %s\n",
73 writel_relaxed(mask
, pdsp
->intd
+ offset
);
77 static irqreturn_t
knav_acc_int_handler(int irq
, void *_instdata
)
79 struct knav_acc_channel
*acc
;
80 struct knav_queue_inst
*kq
= NULL
;
81 struct knav_range_info
*range
;
82 struct knav_pdsp_info
*pdsp
;
83 struct knav_acc_info
*info
;
84 struct knav_device
*kdev
;
86 u32
*list
, *list_cpu
, val
, idx
, notifies
;
87 int range_base
, channel
, queue
= 0;
91 info
= &range
->acc_info
;
93 pdsp
= range
->acc_info
.pdsp
;
96 range_base
= kdev
->base_id
+ range
->queue_base
;
97 if ((range
->flags
& RANGE_MULTI_QUEUE
) == 0) {
98 for (queue
= 0; queue
< range
->num_irqs
; queue
++)
99 if (range
->irqs
[queue
].irq
== irq
)
101 kq
= knav_range_offset_to_inst(kdev
, range
, queue
);
105 channel
= acc
->channel
;
106 list_dma
= acc
->list_dma
[acc
->list_index
];
107 list_cpu
= acc
->list_cpu
[acc
->list_index
];
108 dev_dbg(kdev
->dev
, "acc-irq: channel %d, list %d, virt %p, dma %pad\n",
109 channel
, acc
->list_index
, list_cpu
, &list_dma
);
110 if (atomic_read(&acc
->retrigger_count
)) {
111 atomic_dec(&acc
->retrigger_count
);
112 __knav_acc_notify(range
, acc
);
113 writel_relaxed(1, pdsp
->intd
+ ACC_INTD_OFFSET_COUNT(channel
));
114 /* ack the interrupt */
115 writel_relaxed(ACC_CHANNEL_INT_BASE
+ channel
,
116 pdsp
->intd
+ ACC_INTD_OFFSET_EOI
);
121 notifies
= readl_relaxed(pdsp
->intd
+ ACC_INTD_OFFSET_COUNT(channel
));
123 dma_sync_single_for_cpu(kdev
->dev
, list_dma
, info
->list_size
,
126 for (list
= list_cpu
; list
< list_cpu
+ (info
->list_size
/ sizeof(u32
));
127 list
+= ACC_LIST_ENTRY_WORDS
) {
128 if (ACC_LIST_ENTRY_WORDS
== 1) {
130 "acc-irq: list %d, entry @%p, %08x\n",
131 acc
->list_index
, list
, list
[0]);
132 } else if (ACC_LIST_ENTRY_WORDS
== 2) {
134 "acc-irq: list %d, entry @%p, %08x %08x\n",
135 acc
->list_index
, list
, list
[0], list
[1]);
136 } else if (ACC_LIST_ENTRY_WORDS
== 4) {
138 "acc-irq: list %d, entry @%p, %08x %08x %08x %08x\n",
139 acc
->list_index
, list
, list
[0], list
[1],
143 val
= list
[ACC_LIST_ENTRY_DESC_IDX
];
147 if (range
->flags
& RANGE_MULTI_QUEUE
) {
148 queue
= list
[ACC_LIST_ENTRY_QUEUE_IDX
] >> 16;
149 if (queue
< range_base
||
150 queue
>= range_base
+ range
->num_queues
) {
152 "bad queue %d, expecting %d-%d\n",
154 range_base
+ range
->num_queues
);
158 kq
= knav_range_offset_to_inst(kdev
, range
,
162 if (atomic_inc_return(&kq
->desc_count
) >= ACC_DESCS_MAX
) {
163 atomic_dec(&kq
->desc_count
);
165 "acc-irq: queue %d full, entry dropped\n",
170 idx
= atomic_inc_return(&kq
->desc_tail
) & ACC_DESCS_MASK
;
171 kq
->descs
[idx
] = val
;
172 kq
->notify_needed
= 1;
173 dev_dbg(kdev
->dev
, "acc-irq: enqueue %08x at %d, queue %d\n",
174 val
, idx
, queue
+ range_base
);
177 __knav_acc_notify(range
, acc
);
178 memset(list_cpu
, 0, info
->list_size
);
179 dma_sync_single_for_device(kdev
->dev
, list_dma
, info
->list_size
,
182 /* flip to the other list */
183 acc
->list_index
^= 1;
185 /* reset the interrupt counter */
186 writel_relaxed(1, pdsp
->intd
+ ACC_INTD_OFFSET_COUNT(channel
));
188 /* ack the interrupt */
189 writel_relaxed(ACC_CHANNEL_INT_BASE
+ channel
,
190 pdsp
->intd
+ ACC_INTD_OFFSET_EOI
);
195 static int knav_range_setup_acc_irq(struct knav_range_info
*range
,
196 int queue
, bool enabled
)
198 struct knav_device
*kdev
= range
->kdev
;
199 struct knav_acc_channel
*acc
;
200 struct cpumask
*cpu_mask
;
204 if (range
->flags
& RANGE_MULTI_QUEUE
) {
206 irq
= range
->irqs
[0].irq
;
207 cpu_mask
= range
->irqs
[0].cpu_mask
;
209 acc
= range
->acc
+ queue
;
210 irq
= range
->irqs
[queue
].irq
;
211 cpu_mask
= range
->irqs
[queue
].cpu_mask
;
214 old
= acc
->open_mask
;
216 new = old
| BIT(queue
);
218 new = old
& ~BIT(queue
);
219 acc
->open_mask
= new;
222 "setup-acc-irq: open mask old %08x, new %08x, channel %s\n",
223 old
, new, acc
->name
);
225 if (likely(new == old
))
230 "setup-acc-irq: requesting %s for channel %s\n",
231 acc
->name
, acc
->name
);
232 ret
= request_irq(irq
, knav_acc_int_handler
, 0, acc
->name
,
234 if (!ret
&& cpu_mask
) {
235 ret
= irq_set_affinity_hint(irq
, cpu_mask
);
237 dev_warn(range
->kdev
->dev
,
238 "Failed to set IRQ affinity\n");
245 dev_dbg(kdev
->dev
, "setup-acc-irq: freeing %s for channel %s\n",
246 acc
->name
, acc
->name
);
247 ret
= irq_set_affinity_hint(irq
, NULL
);
249 dev_warn(range
->kdev
->dev
,
250 "Failed to set IRQ affinity\n");
251 free_irq(irq
, range
);
257 static const char *knav_acc_result_str(enum knav_acc_result result
)
259 static const char * const result_str
[] = {
260 [ACC_RET_IDLE
] = "idle",
261 [ACC_RET_SUCCESS
] = "success",
262 [ACC_RET_INVALID_COMMAND
] = "invalid command",
263 [ACC_RET_INVALID_CHANNEL
] = "invalid channel",
264 [ACC_RET_INACTIVE_CHANNEL
] = "inactive channel",
265 [ACC_RET_ACTIVE_CHANNEL
] = "active channel",
266 [ACC_RET_INVALID_QUEUE
] = "invalid queue",
267 [ACC_RET_INVALID_RET
] = "invalid return code",
270 if (result
>= ARRAY_SIZE(result_str
))
271 return result_str
[ACC_RET_INVALID_RET
];
273 return result_str
[result
];
276 static enum knav_acc_result
277 knav_acc_write(struct knav_device
*kdev
, struct knav_pdsp_info
*pdsp
,
278 struct knav_reg_acc_command
*cmd
)
282 dev_dbg(kdev
->dev
, "acc command %08x %08x %08x %08x %08x\n",
283 cmd
->command
, cmd
->queue_mask
, cmd
->list_dma
,
284 cmd
->queue_num
, cmd
->timer_config
);
286 writel_relaxed(cmd
->timer_config
, &pdsp
->acc_command
->timer_config
);
287 writel_relaxed(cmd
->queue_num
, &pdsp
->acc_command
->queue_num
);
288 writel_relaxed(cmd
->list_dma
, &pdsp
->acc_command
->list_dma
);
289 writel_relaxed(cmd
->queue_mask
, &pdsp
->acc_command
->queue_mask
);
290 writel_relaxed(cmd
->command
, &pdsp
->acc_command
->command
);
292 /* wait for the command to clear */
294 result
= readl_relaxed(&pdsp
->acc_command
->command
);
295 } while ((result
>> 8) & 0xff);
297 return (result
>> 24) & 0xff;
300 static void knav_acc_setup_cmd(struct knav_device
*kdev
,
301 struct knav_range_info
*range
,
302 struct knav_reg_acc_command
*cmd
,
305 struct knav_acc_info
*info
= &range
->acc_info
;
306 struct knav_acc_channel
*acc
;
310 if (range
->flags
& RANGE_MULTI_QUEUE
) {
312 queue_base
= range
->queue_base
;
313 queue_mask
= BIT(range
->num_queues
) - 1;
315 acc
= range
->acc
+ queue
;
316 queue_base
= range
->queue_base
+ queue
;
320 memset(cmd
, 0, sizeof(*cmd
));
321 cmd
->command
= acc
->channel
;
322 cmd
->queue_mask
= queue_mask
;
323 cmd
->list_dma
= (u32
)acc
->list_dma
[0];
324 cmd
->queue_num
= info
->list_entries
<< 16;
325 cmd
->queue_num
|= queue_base
;
327 cmd
->timer_config
= ACC_LIST_ENTRY_TYPE
<< 18;
328 if (range
->flags
& RANGE_MULTI_QUEUE
)
329 cmd
->timer_config
|= ACC_CFG_MULTI_QUEUE
;
330 cmd
->timer_config
|= info
->pacing_mode
<< 16;
331 cmd
->timer_config
|= info
->timer_count
;
334 static void knav_acc_stop(struct knav_device
*kdev
,
335 struct knav_range_info
*range
,
338 struct knav_reg_acc_command cmd
;
339 struct knav_acc_channel
*acc
;
340 enum knav_acc_result result
;
342 acc
= range
->acc
+ queue
;
344 knav_acc_setup_cmd(kdev
, range
, &cmd
, queue
);
345 cmd
.command
|= ACC_CMD_DISABLE_CHANNEL
<< 8;
346 result
= knav_acc_write(kdev
, range
->acc_info
.pdsp
, &cmd
);
348 dev_dbg(kdev
->dev
, "stopped acc channel %s, result %s\n",
349 acc
->name
, knav_acc_result_str(result
));
352 static enum knav_acc_result
knav_acc_start(struct knav_device
*kdev
,
353 struct knav_range_info
*range
,
356 struct knav_reg_acc_command cmd
;
357 struct knav_acc_channel
*acc
;
358 enum knav_acc_result result
;
360 acc
= range
->acc
+ queue
;
362 knav_acc_setup_cmd(kdev
, range
, &cmd
, queue
);
363 cmd
.command
|= ACC_CMD_ENABLE_CHANNEL
<< 8;
364 result
= knav_acc_write(kdev
, range
->acc_info
.pdsp
, &cmd
);
366 dev_dbg(kdev
->dev
, "started acc channel %s, result %s\n",
367 acc
->name
, knav_acc_result_str(result
));
372 static int knav_acc_init_range(struct knav_range_info
*range
)
374 struct knav_device
*kdev
= range
->kdev
;
375 struct knav_acc_channel
*acc
;
376 enum knav_acc_result result
;
379 for (queue
= 0; queue
< range
->num_queues
; queue
++) {
380 acc
= range
->acc
+ queue
;
382 knav_acc_stop(kdev
, range
, queue
);
384 result
= knav_acc_start(kdev
, range
, queue
);
386 if (result
!= ACC_RET_SUCCESS
)
389 if (range
->flags
& RANGE_MULTI_QUEUE
)
395 static int knav_acc_init_queue(struct knav_range_info
*range
,
396 struct knav_queue_inst
*kq
)
398 unsigned id
= kq
->id
- range
->queue_base
;
400 kq
->descs
= devm_kcalloc(range
->kdev
->dev
,
401 ACC_DESCS_MAX
, sizeof(u32
), GFP_KERNEL
);
405 kq
->acc
= range
->acc
;
406 if ((range
->flags
& RANGE_MULTI_QUEUE
) == 0)
411 static int knav_acc_open_queue(struct knav_range_info
*range
,
412 struct knav_queue_inst
*inst
, unsigned flags
)
414 unsigned id
= inst
->id
- range
->queue_base
;
416 return knav_range_setup_acc_irq(range
, id
, true);
419 static int knav_acc_close_queue(struct knav_range_info
*range
,
420 struct knav_queue_inst
*inst
)
422 unsigned id
= inst
->id
- range
->queue_base
;
424 return knav_range_setup_acc_irq(range
, id
, false);
427 static int knav_acc_free_range(struct knav_range_info
*range
)
429 struct knav_device
*kdev
= range
->kdev
;
430 struct knav_acc_channel
*acc
;
431 struct knav_acc_info
*info
;
432 int channel
, channels
;
434 info
= &range
->acc_info
;
436 if (range
->flags
& RANGE_MULTI_QUEUE
)
439 channels
= range
->num_queues
;
441 for (channel
= 0; channel
< channels
; channel
++) {
442 acc
= range
->acc
+ channel
;
443 if (!acc
->list_cpu
[0])
445 dma_unmap_single(kdev
->dev
, acc
->list_dma
[0],
446 info
->mem_size
, DMA_BIDIRECTIONAL
);
447 free_pages_exact(acc
->list_cpu
[0], info
->mem_size
);
449 devm_kfree(range
->kdev
->dev
, range
->acc
);
453 struct knav_range_ops knav_acc_range_ops
= {
454 .set_notify
= knav_acc_set_notify
,
455 .init_queue
= knav_acc_init_queue
,
456 .open_queue
= knav_acc_open_queue
,
457 .close_queue
= knav_acc_close_queue
,
458 .init_range
= knav_acc_init_range
,
459 .free_range
= knav_acc_free_range
,
463 * knav_init_acc_range: Initialise accumulator ranges
467 * @range: qmms range information
469 * Return 0 on success or error
471 int knav_init_acc_range(struct knav_device
*kdev
,
472 struct device_node
*node
,
473 struct knav_range_info
*range
)
475 struct knav_acc_channel
*acc
;
476 struct knav_pdsp_info
*pdsp
;
477 struct knav_acc_info
*info
;
478 int ret
, channel
, channels
;
479 int list_size
, mem_size
;
484 range
->flags
|= RANGE_HAS_ACCUMULATOR
;
485 info
= &range
->acc_info
;
487 ret
= of_property_read_u32_array(node
, "accumulator", config
, 5);
491 info
->pdsp_id
= config
[0];
492 info
->start_channel
= config
[1];
493 info
->list_entries
= config
[2];
494 info
->pacing_mode
= config
[3];
495 info
->timer_count
= config
[4] / ACC_DEFAULT_PERIOD
;
497 if (info
->start_channel
> ACC_MAX_CHANNEL
) {
498 dev_err(kdev
->dev
, "channel %d invalid for range %s\n",
499 info
->start_channel
, range
->name
);
503 if (info
->pacing_mode
> 3) {
504 dev_err(kdev
->dev
, "pacing mode %d invalid for range %s\n",
505 info
->pacing_mode
, range
->name
);
509 pdsp
= knav_find_pdsp(kdev
, info
->pdsp_id
);
511 dev_err(kdev
->dev
, "pdsp id %d not found for range %s\n",
512 info
->pdsp_id
, range
->name
);
516 if (!pdsp
->started
) {
517 dev_err(kdev
->dev
, "pdsp id %d not started for range %s\n",
518 info
->pdsp_id
, range
->name
);
523 channels
= range
->num_queues
;
524 if (of_get_property(node
, "multi-queue", NULL
)) {
525 range
->flags
|= RANGE_MULTI_QUEUE
;
527 if (range
->queue_base
& (32 - 1)) {
529 "misaligned multi-queue accumulator range %s\n",
533 if (range
->num_queues
> 32) {
535 "too many queues in accumulator range %s\n",
541 /* figure out list size */
542 list_size
= info
->list_entries
;
543 list_size
*= ACC_LIST_ENTRY_WORDS
* sizeof(u32
);
544 info
->list_size
= list_size
;
545 mem_size
= PAGE_ALIGN(list_size
* 2);
546 info
->mem_size
= mem_size
;
547 range
->acc
= devm_kcalloc(kdev
->dev
, channels
, sizeof(*range
->acc
),
552 for (channel
= 0; channel
< channels
; channel
++) {
553 acc
= range
->acc
+ channel
;
554 acc
->channel
= info
->start_channel
+ channel
;
556 /* allocate memory for the two lists */
557 list_mem
= alloc_pages_exact(mem_size
, GFP_KERNEL
| GFP_DMA
);
561 list_dma
= dma_map_single(kdev
->dev
, list_mem
, mem_size
,
563 if (dma_mapping_error(kdev
->dev
, list_dma
)) {
564 free_pages_exact(list_mem
, mem_size
);
568 memset(list_mem
, 0, mem_size
);
569 dma_sync_single_for_device(kdev
->dev
, list_dma
, mem_size
,
571 scnprintf(acc
->name
, sizeof(acc
->name
), "hwqueue-acc-%d",
573 acc
->list_cpu
[0] = list_mem
;
574 acc
->list_cpu
[1] = list_mem
+ list_size
;
575 acc
->list_dma
[0] = list_dma
;
576 acc
->list_dma
[1] = list_dma
+ list_size
;
577 dev_dbg(kdev
->dev
, "%s: channel %d, dma %pad, virt %8p\n",
578 acc
->name
, acc
->channel
, &list_dma
, list_mem
);
581 range
->ops
= &knav_acc_range_ops
;
584 EXPORT_SYMBOL_GPL(knav_init_acc_range
);