1 // SPDX-License-Identifier: GPL-2.0
3 * core.c - Implementation of core module of MOST Linux driver stack
5 * Copyright (C) 2013-2020 Microchip Technology Germany II GmbH & Co. KG
8 #include <linux/module.h>
10 #include <linux/slab.h>
11 #include <linux/init.h>
12 #include <linux/device.h>
13 #include <linux/list.h>
14 #include <linux/poll.h>
15 #include <linux/wait.h>
16 #include <linux/kobject.h>
17 #include <linux/mutex.h>
18 #include <linux/completion.h>
19 #include <linux/sysfs.h>
20 #include <linux/kthread.h>
21 #include <linux/dma-mapping.h>
22 #include <linux/idr.h>
23 #include <linux/most.h>
25 #define MAX_CHANNELS 64
26 #define STRING_SIZE 80
28 static struct ida mdev_id
;
29 static int dummy_num_buffers
;
30 static struct list_head comp_list
;
33 struct most_component
*comp
;
40 struct completion cleanup
;
42 atomic_t mbo_nq_level
;
44 char name
[STRING_SIZE
];
46 struct mutex start_mutex
; /* channel activation synchronization */
47 struct mutex nq_mutex
; /* nq thread synchronization */
49 struct most_interface
*iface
;
50 struct most_channel_config cfg
;
53 struct list_head fifo
;
54 spinlock_t fifo_lock
; /* fifo access synchronization */
55 struct list_head halt_fifo
;
56 struct list_head list
;
59 struct list_head trash_fifo
;
60 struct task_struct
*hdm_enqueue_task
;
61 wait_queue_head_t hdm_fifo_wq
;
65 #define to_channel(d) container_of(d, struct most_channel, dev)
67 struct interface_private
{
69 char name
[STRING_SIZE
];
70 struct most_channel
*channel
[MAX_CHANNELS
];
71 struct list_head channel_list
;
75 int most_ch_data_type
;
78 { MOST_CH_CONTROL
, "control" },
79 { MOST_CH_ASYNC
, "async" },
80 { MOST_CH_SYNC
, "sync" },
81 { MOST_CH_ISOC
, "isoc"},
82 { MOST_CH_ISOC
, "isoc_avp"},
86 * list_pop_mbo - retrieves the first MBO of the list and removes it
87 * @ptr: the list head to grab the MBO from.
89 #define list_pop_mbo(ptr) \
91 struct mbo *_mbo = list_first_entry(ptr, struct mbo, list); \
92 list_del(&_mbo->list); \
97 * most_free_mbo_coherent - free an MBO and its coherent buffer
100 static void most_free_mbo_coherent(struct mbo
*mbo
)
102 struct most_channel
*c
= mbo
->context
;
103 u16
const coherent_buf_size
= c
->cfg
.buffer_size
+ c
->cfg
.extra_len
;
105 if (c
->iface
->dma_free
)
106 c
->iface
->dma_free(mbo
, coherent_buf_size
);
108 kfree(mbo
->virt_address
);
110 if (atomic_sub_and_test(1, &c
->mbo_ref
))
111 complete(&c
->cleanup
);
115 * flush_channel_fifos - clear the channel fifos
116 * @c: pointer to channel object
118 static void flush_channel_fifos(struct most_channel
*c
)
120 unsigned long flags
, hf_flags
;
121 struct mbo
*mbo
, *tmp
;
123 if (list_empty(&c
->fifo
) && list_empty(&c
->halt_fifo
))
126 spin_lock_irqsave(&c
->fifo_lock
, flags
);
127 list_for_each_entry_safe(mbo
, tmp
, &c
->fifo
, list
) {
128 list_del(&mbo
->list
);
129 spin_unlock_irqrestore(&c
->fifo_lock
, flags
);
130 most_free_mbo_coherent(mbo
);
131 spin_lock_irqsave(&c
->fifo_lock
, flags
);
133 spin_unlock_irqrestore(&c
->fifo_lock
, flags
);
135 spin_lock_irqsave(&c
->fifo_lock
, hf_flags
);
136 list_for_each_entry_safe(mbo
, tmp
, &c
->halt_fifo
, list
) {
137 list_del(&mbo
->list
);
138 spin_unlock_irqrestore(&c
->fifo_lock
, hf_flags
);
139 most_free_mbo_coherent(mbo
);
140 spin_lock_irqsave(&c
->fifo_lock
, hf_flags
);
142 spin_unlock_irqrestore(&c
->fifo_lock
, hf_flags
);
144 if (unlikely((!list_empty(&c
->fifo
) || !list_empty(&c
->halt_fifo
))))
145 dev_warn(&c
->dev
, "Channel or trash fifo not empty\n");
149 * flush_trash_fifo - clear the trash fifo
150 * @c: pointer to channel object
152 static int flush_trash_fifo(struct most_channel
*c
)
154 struct mbo
*mbo
, *tmp
;
157 spin_lock_irqsave(&c
->fifo_lock
, flags
);
158 list_for_each_entry_safe(mbo
, tmp
, &c
->trash_fifo
, list
) {
159 list_del(&mbo
->list
);
160 spin_unlock_irqrestore(&c
->fifo_lock
, flags
);
161 most_free_mbo_coherent(mbo
);
162 spin_lock_irqsave(&c
->fifo_lock
, flags
);
164 spin_unlock_irqrestore(&c
->fifo_lock
, flags
);
168 static ssize_t
available_directions_show(struct device
*dev
,
169 struct device_attribute
*attr
,
172 struct most_channel
*c
= to_channel(dev
);
173 unsigned int i
= c
->channel_id
;
176 if (c
->iface
->channel_vector
[i
].direction
& MOST_CH_RX
)
178 if (c
->iface
->channel_vector
[i
].direction
& MOST_CH_TX
)
184 static ssize_t
available_datatypes_show(struct device
*dev
,
185 struct device_attribute
*attr
,
188 struct most_channel
*c
= to_channel(dev
);
189 unsigned int i
= c
->channel_id
;
192 if (c
->iface
->channel_vector
[i
].data_type
& MOST_CH_CONTROL
)
193 strcat(buf
, "control ");
194 if (c
->iface
->channel_vector
[i
].data_type
& MOST_CH_ASYNC
)
195 strcat(buf
, "async ");
196 if (c
->iface
->channel_vector
[i
].data_type
& MOST_CH_SYNC
)
197 strcat(buf
, "sync ");
198 if (c
->iface
->channel_vector
[i
].data_type
& MOST_CH_ISOC
)
199 strcat(buf
, "isoc ");
204 static ssize_t
number_of_packet_buffers_show(struct device
*dev
,
205 struct device_attribute
*attr
,
208 struct most_channel
*c
= to_channel(dev
);
209 unsigned int i
= c
->channel_id
;
211 return snprintf(buf
, PAGE_SIZE
, "%d\n",
212 c
->iface
->channel_vector
[i
].num_buffers_packet
);
215 static ssize_t
number_of_stream_buffers_show(struct device
*dev
,
216 struct device_attribute
*attr
,
219 struct most_channel
*c
= to_channel(dev
);
220 unsigned int i
= c
->channel_id
;
222 return snprintf(buf
, PAGE_SIZE
, "%d\n",
223 c
->iface
->channel_vector
[i
].num_buffers_streaming
);
226 static ssize_t
size_of_packet_buffer_show(struct device
*dev
,
227 struct device_attribute
*attr
,
230 struct most_channel
*c
= to_channel(dev
);
231 unsigned int i
= c
->channel_id
;
233 return snprintf(buf
, PAGE_SIZE
, "%d\n",
234 c
->iface
->channel_vector
[i
].buffer_size_packet
);
237 static ssize_t
size_of_stream_buffer_show(struct device
*dev
,
238 struct device_attribute
*attr
,
241 struct most_channel
*c
= to_channel(dev
);
242 unsigned int i
= c
->channel_id
;
244 return snprintf(buf
, PAGE_SIZE
, "%d\n",
245 c
->iface
->channel_vector
[i
].buffer_size_streaming
);
248 static ssize_t
channel_starving_show(struct device
*dev
,
249 struct device_attribute
*attr
,
252 struct most_channel
*c
= to_channel(dev
);
254 return snprintf(buf
, PAGE_SIZE
, "%d\n", c
->is_starving
);
257 static ssize_t
set_number_of_buffers_show(struct device
*dev
,
258 struct device_attribute
*attr
,
261 struct most_channel
*c
= to_channel(dev
);
263 return snprintf(buf
, PAGE_SIZE
, "%d\n", c
->cfg
.num_buffers
);
266 static ssize_t
set_buffer_size_show(struct device
*dev
,
267 struct device_attribute
*attr
,
270 struct most_channel
*c
= to_channel(dev
);
272 return snprintf(buf
, PAGE_SIZE
, "%d\n", c
->cfg
.buffer_size
);
275 static ssize_t
set_direction_show(struct device
*dev
,
276 struct device_attribute
*attr
,
279 struct most_channel
*c
= to_channel(dev
);
281 if (c
->cfg
.direction
& MOST_CH_TX
)
282 return snprintf(buf
, PAGE_SIZE
, "tx\n");
283 else if (c
->cfg
.direction
& MOST_CH_RX
)
284 return snprintf(buf
, PAGE_SIZE
, "rx\n");
285 return snprintf(buf
, PAGE_SIZE
, "unconfigured\n");
288 static ssize_t
set_datatype_show(struct device
*dev
,
289 struct device_attribute
*attr
,
293 struct most_channel
*c
= to_channel(dev
);
295 for (i
= 0; i
< ARRAY_SIZE(ch_data_type
); i
++) {
296 if (c
->cfg
.data_type
& ch_data_type
[i
].most_ch_data_type
)
297 return snprintf(buf
, PAGE_SIZE
, "%s",
298 ch_data_type
[i
].name
);
300 return snprintf(buf
, PAGE_SIZE
, "unconfigured\n");
303 static ssize_t
set_subbuffer_size_show(struct device
*dev
,
304 struct device_attribute
*attr
,
307 struct most_channel
*c
= to_channel(dev
);
309 return snprintf(buf
, PAGE_SIZE
, "%d\n", c
->cfg
.subbuffer_size
);
312 static ssize_t
set_packets_per_xact_show(struct device
*dev
,
313 struct device_attribute
*attr
,
316 struct most_channel
*c
= to_channel(dev
);
318 return snprintf(buf
, PAGE_SIZE
, "%d\n", c
->cfg
.packets_per_xact
);
321 static ssize_t
set_dbr_size_show(struct device
*dev
,
322 struct device_attribute
*attr
, char *buf
)
324 struct most_channel
*c
= to_channel(dev
);
326 return snprintf(buf
, PAGE_SIZE
, "%d\n", c
->cfg
.dbr_size
);
329 #define to_dev_attr(a) container_of(a, struct device_attribute, attr)
330 static umode_t
channel_attr_is_visible(struct kobject
*kobj
,
331 struct attribute
*attr
, int index
)
333 struct device_attribute
*dev_attr
= to_dev_attr(attr
);
334 struct device
*dev
= kobj_to_dev(kobj
);
335 struct most_channel
*c
= to_channel(dev
);
337 if (!strcmp(dev_attr
->attr
.name
, "set_dbr_size") &&
338 (c
->iface
->interface
!= ITYPE_MEDIALB_DIM2
))
340 if (!strcmp(dev_attr
->attr
.name
, "set_packets_per_xact") &&
341 (c
->iface
->interface
!= ITYPE_USB
))
347 #define DEV_ATTR(_name) (&dev_attr_##_name.attr)
349 static DEVICE_ATTR_RO(available_directions
);
350 static DEVICE_ATTR_RO(available_datatypes
);
351 static DEVICE_ATTR_RO(number_of_packet_buffers
);
352 static DEVICE_ATTR_RO(number_of_stream_buffers
);
353 static DEVICE_ATTR_RO(size_of_stream_buffer
);
354 static DEVICE_ATTR_RO(size_of_packet_buffer
);
355 static DEVICE_ATTR_RO(channel_starving
);
356 static DEVICE_ATTR_RO(set_buffer_size
);
357 static DEVICE_ATTR_RO(set_number_of_buffers
);
358 static DEVICE_ATTR_RO(set_direction
);
359 static DEVICE_ATTR_RO(set_datatype
);
360 static DEVICE_ATTR_RO(set_subbuffer_size
);
361 static DEVICE_ATTR_RO(set_packets_per_xact
);
362 static DEVICE_ATTR_RO(set_dbr_size
);
364 static struct attribute
*channel_attrs
[] = {
365 DEV_ATTR(available_directions
),
366 DEV_ATTR(available_datatypes
),
367 DEV_ATTR(number_of_packet_buffers
),
368 DEV_ATTR(number_of_stream_buffers
),
369 DEV_ATTR(size_of_stream_buffer
),
370 DEV_ATTR(size_of_packet_buffer
),
371 DEV_ATTR(channel_starving
),
372 DEV_ATTR(set_buffer_size
),
373 DEV_ATTR(set_number_of_buffers
),
374 DEV_ATTR(set_direction
),
375 DEV_ATTR(set_datatype
),
376 DEV_ATTR(set_subbuffer_size
),
377 DEV_ATTR(set_packets_per_xact
),
378 DEV_ATTR(set_dbr_size
),
382 static struct attribute_group channel_attr_group
= {
383 .attrs
= channel_attrs
,
384 .is_visible
= channel_attr_is_visible
,
387 static const struct attribute_group
*channel_attr_groups
[] = {
392 static ssize_t
description_show(struct device
*dev
,
393 struct device_attribute
*attr
,
396 struct most_interface
*iface
= dev_get_drvdata(dev
);
398 return snprintf(buf
, PAGE_SIZE
, "%s\n", iface
->description
);
401 static ssize_t
interface_show(struct device
*dev
,
402 struct device_attribute
*attr
,
405 struct most_interface
*iface
= dev_get_drvdata(dev
);
407 switch (iface
->interface
) {
409 return snprintf(buf
, PAGE_SIZE
, "loopback\n");
411 return snprintf(buf
, PAGE_SIZE
, "i2c\n");
413 return snprintf(buf
, PAGE_SIZE
, "i2s\n");
415 return snprintf(buf
, PAGE_SIZE
, "tsi\n");
417 return snprintf(buf
, PAGE_SIZE
, "hbi\n");
418 case ITYPE_MEDIALB_DIM
:
419 return snprintf(buf
, PAGE_SIZE
, "mlb_dim\n");
420 case ITYPE_MEDIALB_DIM2
:
421 return snprintf(buf
, PAGE_SIZE
, "mlb_dim2\n");
423 return snprintf(buf
, PAGE_SIZE
, "usb\n");
425 return snprintf(buf
, PAGE_SIZE
, "pcie\n");
427 return snprintf(buf
, PAGE_SIZE
, "unknown\n");
430 static DEVICE_ATTR_RO(description
);
431 static DEVICE_ATTR_RO(interface
);
433 static struct attribute
*interface_attrs
[] = {
434 DEV_ATTR(description
),
439 static struct attribute_group interface_attr_group
= {
440 .attrs
= interface_attrs
,
443 static const struct attribute_group
*interface_attr_groups
[] = {
444 &interface_attr_group
,
448 static struct most_component
*match_component(char *name
)
450 struct most_component
*comp
;
452 list_for_each_entry(comp
, &comp_list
, list
) {
453 if (!strcmp(comp
->name
, name
))
459 struct show_links_data
{
464 static int print_links(struct device
*dev
, void *data
)
466 struct show_links_data
*d
= data
;
469 struct most_channel
*c
;
470 struct most_interface
*iface
= dev_get_drvdata(dev
);
472 list_for_each_entry(c
, &iface
->p
->channel_list
, list
) {
474 offs
+= scnprintf(buf
+ offs
,
478 dev_name(iface
->dev
),
482 offs
+= scnprintf(buf
+ offs
,
486 dev_name(iface
->dev
),
494 static int most_match(struct device
*dev
, struct device_driver
*drv
)
496 if (!strcmp(dev_name(dev
), "most"))
502 static struct bus_type mostbus
= {
507 static ssize_t
links_show(struct device_driver
*drv
, char *buf
)
509 struct show_links_data d
= { .buf
= buf
};
511 bus_for_each_dev(&mostbus
, NULL
, &d
, print_links
);
515 static ssize_t
components_show(struct device_driver
*drv
, char *buf
)
517 struct most_component
*comp
;
520 list_for_each_entry(comp
, &comp_list
, list
) {
521 offs
+= scnprintf(buf
+ offs
, PAGE_SIZE
- offs
, "%s\n",
528 * get_channel - get pointer to channel
529 * @mdev: name of the device interface
530 * @mdev_ch: name of channel
532 static struct most_channel
*get_channel(char *mdev
, char *mdev_ch
)
534 struct device
*dev
= NULL
;
535 struct most_interface
*iface
;
536 struct most_channel
*c
, *tmp
;
538 dev
= bus_find_device_by_name(&mostbus
, NULL
, mdev
);
542 iface
= dev_get_drvdata(dev
);
543 list_for_each_entry_safe(c
, tmp
, &iface
->p
->channel_list
, list
) {
544 if (!strcmp(dev_name(&c
->dev
), mdev_ch
))
551 inline int link_channel_to_component(struct most_channel
*c
,
552 struct most_component
*comp
,
557 struct most_component
**comp_ptr
;
560 comp_ptr
= &c
->pipe0
.comp
;
561 else if (!c
->pipe1
.comp
)
562 comp_ptr
= &c
->pipe1
.comp
;
567 ret
= comp
->probe_channel(c
->iface
, c
->channel_id
, &c
->cfg
, name
,
576 int most_set_cfg_buffer_size(char *mdev
, char *mdev_ch
, u16 val
)
578 struct most_channel
*c
= get_channel(mdev
, mdev_ch
);
582 c
->cfg
.buffer_size
= val
;
586 int most_set_cfg_subbuffer_size(char *mdev
, char *mdev_ch
, u16 val
)
588 struct most_channel
*c
= get_channel(mdev
, mdev_ch
);
592 c
->cfg
.subbuffer_size
= val
;
596 int most_set_cfg_dbr_size(char *mdev
, char *mdev_ch
, u16 val
)
598 struct most_channel
*c
= get_channel(mdev
, mdev_ch
);
602 c
->cfg
.dbr_size
= val
;
606 int most_set_cfg_num_buffers(char *mdev
, char *mdev_ch
, u16 val
)
608 struct most_channel
*c
= get_channel(mdev
, mdev_ch
);
612 c
->cfg
.num_buffers
= val
;
616 int most_set_cfg_datatype(char *mdev
, char *mdev_ch
, char *buf
)
619 struct most_channel
*c
= get_channel(mdev
, mdev_ch
);
623 for (i
= 0; i
< ARRAY_SIZE(ch_data_type
); i
++) {
624 if (!strcmp(buf
, ch_data_type
[i
].name
)) {
625 c
->cfg
.data_type
= ch_data_type
[i
].most_ch_data_type
;
630 if (i
== ARRAY_SIZE(ch_data_type
))
631 dev_warn(&c
->dev
, "Invalid attribute settings\n");
635 int most_set_cfg_direction(char *mdev
, char *mdev_ch
, char *buf
)
637 struct most_channel
*c
= get_channel(mdev
, mdev_ch
);
641 if (!strcmp(buf
, "dir_rx")) {
642 c
->cfg
.direction
= MOST_CH_RX
;
643 } else if (!strcmp(buf
, "rx")) {
644 c
->cfg
.direction
= MOST_CH_RX
;
645 } else if (!strcmp(buf
, "dir_tx")) {
646 c
->cfg
.direction
= MOST_CH_TX
;
647 } else if (!strcmp(buf
, "tx")) {
648 c
->cfg
.direction
= MOST_CH_TX
;
650 dev_err(&c
->dev
, "Invalid direction\n");
656 int most_set_cfg_packets_xact(char *mdev
, char *mdev_ch
, u16 val
)
658 struct most_channel
*c
= get_channel(mdev
, mdev_ch
);
662 c
->cfg
.packets_per_xact
= val
;
666 int most_cfg_complete(char *comp_name
)
668 struct most_component
*comp
;
670 comp
= match_component(comp_name
);
674 return comp
->cfg_complete();
677 int most_add_link(char *mdev
, char *mdev_ch
, char *comp_name
, char *link_name
,
680 struct most_channel
*c
= get_channel(mdev
, mdev_ch
);
681 struct most_component
*comp
= match_component(comp_name
);
686 return link_channel_to_component(c
, comp
, link_name
, comp_param
);
689 int most_remove_link(char *mdev
, char *mdev_ch
, char *comp_name
)
691 struct most_channel
*c
;
692 struct most_component
*comp
;
694 comp
= match_component(comp_name
);
697 c
= get_channel(mdev
, mdev_ch
);
701 if (comp
->disconnect_channel(c
->iface
, c
->channel_id
))
703 if (c
->pipe0
.comp
== comp
)
704 c
->pipe0
.comp
= NULL
;
705 if (c
->pipe1
.comp
== comp
)
706 c
->pipe1
.comp
= NULL
;
710 #define DRV_ATTR(_name) (&driver_attr_##_name.attr)
712 static DRIVER_ATTR_RO(links
);
713 static DRIVER_ATTR_RO(components
);
715 static struct attribute
*mc_attrs
[] = {
717 DRV_ATTR(components
),
721 static struct attribute_group mc_attr_group
= {
725 static const struct attribute_group
*mc_attr_groups
[] = {
730 static struct device_driver mostbus_driver
= {
733 .groups
= mc_attr_groups
,
736 static inline void trash_mbo(struct mbo
*mbo
)
739 struct most_channel
*c
= mbo
->context
;
741 spin_lock_irqsave(&c
->fifo_lock
, flags
);
742 list_add(&mbo
->list
, &c
->trash_fifo
);
743 spin_unlock_irqrestore(&c
->fifo_lock
, flags
);
746 static bool hdm_mbo_ready(struct most_channel
*c
)
753 spin_lock_irq(&c
->fifo_lock
);
754 empty
= list_empty(&c
->halt_fifo
);
755 spin_unlock_irq(&c
->fifo_lock
);
760 static void nq_hdm_mbo(struct mbo
*mbo
)
763 struct most_channel
*c
= mbo
->context
;
765 spin_lock_irqsave(&c
->fifo_lock
, flags
);
766 list_add_tail(&mbo
->list
, &c
->halt_fifo
);
767 spin_unlock_irqrestore(&c
->fifo_lock
, flags
);
768 wake_up_interruptible(&c
->hdm_fifo_wq
);
771 static int hdm_enqueue_thread(void *data
)
773 struct most_channel
*c
= data
;
776 typeof(c
->iface
->enqueue
) enqueue
= c
->iface
->enqueue
;
778 while (likely(!kthread_should_stop())) {
779 wait_event_interruptible(c
->hdm_fifo_wq
,
781 kthread_should_stop());
783 mutex_lock(&c
->nq_mutex
);
784 spin_lock_irq(&c
->fifo_lock
);
785 if (unlikely(c
->enqueue_halt
|| list_empty(&c
->halt_fifo
))) {
786 spin_unlock_irq(&c
->fifo_lock
);
787 mutex_unlock(&c
->nq_mutex
);
791 mbo
= list_pop_mbo(&c
->halt_fifo
);
792 spin_unlock_irq(&c
->fifo_lock
);
794 if (c
->cfg
.direction
== MOST_CH_RX
)
795 mbo
->buffer_length
= c
->cfg
.buffer_size
;
797 ret
= enqueue(mbo
->ifp
, mbo
->hdm_channel_id
, mbo
);
798 mutex_unlock(&c
->nq_mutex
);
801 dev_err(&c
->dev
, "Buffer enqueue failed\n");
803 c
->hdm_enqueue_task
= NULL
;
811 static int run_enqueue_thread(struct most_channel
*c
, int channel_id
)
813 struct task_struct
*task
=
814 kthread_run(hdm_enqueue_thread
, c
, "hdm_fifo_%d",
818 return PTR_ERR(task
);
820 c
->hdm_enqueue_task
= task
;
825 * arm_mbo - recycle MBO for further usage
828 * This puts an MBO back to the list to have it ready for up coming
831 * In case the MBO belongs to a channel that recently has been
832 * poisoned, the MBO is scheduled to be trashed.
833 * Calls the completion handler of an attached component.
835 static void arm_mbo(struct mbo
*mbo
)
838 struct most_channel
*c
;
842 if (c
->is_poisoned
) {
847 spin_lock_irqsave(&c
->fifo_lock
, flags
);
848 ++*mbo
->num_buffers_ptr
;
849 list_add_tail(&mbo
->list
, &c
->fifo
);
850 spin_unlock_irqrestore(&c
->fifo_lock
, flags
);
852 if (c
->pipe0
.refs
&& c
->pipe0
.comp
->tx_completion
)
853 c
->pipe0
.comp
->tx_completion(c
->iface
, c
->channel_id
);
855 if (c
->pipe1
.refs
&& c
->pipe1
.comp
->tx_completion
)
856 c
->pipe1
.comp
->tx_completion(c
->iface
, c
->channel_id
);
860 * arm_mbo_chain - helper function that arms an MBO chain for the HDM
861 * @c: pointer to interface channel
862 * @dir: direction of the channel
863 * @compl: pointer to completion function
865 * This allocates buffer objects including the containing DMA coherent
866 * buffer and puts them in the fifo.
867 * Buffers of Rx channels are put in the kthread fifo, hence immediately
868 * submitted to the HDM.
870 * Returns the number of allocated and enqueued MBOs.
872 static int arm_mbo_chain(struct most_channel
*c
, int dir
,
873 void (*compl)(struct mbo
*))
878 u32 coherent_buf_size
= c
->cfg
.buffer_size
+ c
->cfg
.extra_len
;
880 atomic_set(&c
->mbo_nq_level
, 0);
882 for (i
= 0; i
< c
->cfg
.num_buffers
; i
++) {
883 mbo
= kzalloc(sizeof(*mbo
), GFP_KERNEL
);
889 mbo
->hdm_channel_id
= c
->channel_id
;
890 if (c
->iface
->dma_alloc
) {
892 c
->iface
->dma_alloc(mbo
, coherent_buf_size
);
895 kzalloc(coherent_buf_size
, GFP_KERNEL
);
897 if (!mbo
->virt_address
)
900 mbo
->complete
= compl;
901 mbo
->num_buffers_ptr
= &dummy_num_buffers
;
902 if (dir
== MOST_CH_RX
) {
904 atomic_inc(&c
->mbo_nq_level
);
906 spin_lock_irqsave(&c
->fifo_lock
, flags
);
907 list_add_tail(&mbo
->list
, &c
->fifo
);
908 spin_unlock_irqrestore(&c
->fifo_lock
, flags
);
911 return c
->cfg
.num_buffers
;
917 flush_channel_fifos(c
);
922 * most_submit_mbo - submits an MBO to fifo
925 void most_submit_mbo(struct mbo
*mbo
)
927 if (WARN_ONCE(!mbo
|| !mbo
->context
,
928 "Bad buffer or missing channel reference\n"))
933 EXPORT_SYMBOL_GPL(most_submit_mbo
);
936 * most_write_completion - write completion handler
939 * This recycles the MBO for further usage. In case the channel has been
940 * poisoned, the MBO is scheduled to be trashed.
942 static void most_write_completion(struct mbo
*mbo
)
944 struct most_channel
*c
;
947 if (unlikely(c
->is_poisoned
|| (mbo
->status
== MBO_E_CLOSE
)))
953 int channel_has_mbo(struct most_interface
*iface
, int id
,
954 struct most_component
*comp
)
956 struct most_channel
*c
= iface
->p
->channel
[id
];
963 if (c
->pipe0
.refs
&& c
->pipe1
.refs
&&
964 ((comp
== c
->pipe0
.comp
&& c
->pipe0
.num_buffers
<= 0) ||
965 (comp
== c
->pipe1
.comp
&& c
->pipe1
.num_buffers
<= 0)))
968 spin_lock_irqsave(&c
->fifo_lock
, flags
);
969 empty
= list_empty(&c
->fifo
);
970 spin_unlock_irqrestore(&c
->fifo_lock
, flags
);
973 EXPORT_SYMBOL_GPL(channel_has_mbo
);
976 * most_get_mbo - get pointer to an MBO of pool
977 * @iface: pointer to interface instance
979 * @comp: driver component
981 * This attempts to get a free buffer out of the channel fifo.
982 * Returns a pointer to MBO on success or NULL otherwise.
984 struct mbo
*most_get_mbo(struct most_interface
*iface
, int id
,
985 struct most_component
*comp
)
988 struct most_channel
*c
;
990 int *num_buffers_ptr
;
992 c
= iface
->p
->channel
[id
];
996 if (c
->pipe0
.refs
&& c
->pipe1
.refs
&&
997 ((comp
== c
->pipe0
.comp
&& c
->pipe0
.num_buffers
<= 0) ||
998 (comp
== c
->pipe1
.comp
&& c
->pipe1
.num_buffers
<= 0)))
1001 if (comp
== c
->pipe0
.comp
)
1002 num_buffers_ptr
= &c
->pipe0
.num_buffers
;
1003 else if (comp
== c
->pipe1
.comp
)
1004 num_buffers_ptr
= &c
->pipe1
.num_buffers
;
1006 num_buffers_ptr
= &dummy_num_buffers
;
1008 spin_lock_irqsave(&c
->fifo_lock
, flags
);
1009 if (list_empty(&c
->fifo
)) {
1010 spin_unlock_irqrestore(&c
->fifo_lock
, flags
);
1013 mbo
= list_pop_mbo(&c
->fifo
);
1015 spin_unlock_irqrestore(&c
->fifo_lock
, flags
);
1017 mbo
->num_buffers_ptr
= num_buffers_ptr
;
1018 mbo
->buffer_length
= c
->cfg
.buffer_size
;
1021 EXPORT_SYMBOL_GPL(most_get_mbo
);
1024 * most_put_mbo - return buffer to pool
1027 void most_put_mbo(struct mbo
*mbo
)
1029 struct most_channel
*c
= mbo
->context
;
1031 if (c
->cfg
.direction
== MOST_CH_TX
) {
1036 atomic_inc(&c
->mbo_nq_level
);
1038 EXPORT_SYMBOL_GPL(most_put_mbo
);
1041 * most_read_completion - read completion handler
1044 * This function is called by the HDM when data has been received from the
1045 * hardware and copied to the buffer of the MBO.
1047 * In case the channel has been poisoned it puts the buffer in the trash queue.
1048 * Otherwise, it passes the buffer to an component for further processing.
1050 static void most_read_completion(struct mbo
*mbo
)
1052 struct most_channel
*c
= mbo
->context
;
1054 if (unlikely(c
->is_poisoned
|| (mbo
->status
== MBO_E_CLOSE
))) {
1059 if (mbo
->status
== MBO_E_INVAL
) {
1061 atomic_inc(&c
->mbo_nq_level
);
1065 if (atomic_sub_and_test(1, &c
->mbo_nq_level
))
1068 if (c
->pipe0
.refs
&& c
->pipe0
.comp
->rx_completion
&&
1069 c
->pipe0
.comp
->rx_completion(mbo
) == 0)
1072 if (c
->pipe1
.refs
&& c
->pipe1
.comp
->rx_completion
&&
1073 c
->pipe1
.comp
->rx_completion(mbo
) == 0)
1080 * most_start_channel - prepares a channel for communication
1081 * @iface: pointer to interface instance
1083 * @comp: driver component
1085 * This prepares the channel for usage. Cross-checks whether the
1086 * channel's been properly configured.
1088 * Returns 0 on success or error code otherwise.
1090 int most_start_channel(struct most_interface
*iface
, int id
,
1091 struct most_component
*comp
)
1095 struct most_channel
*c
= iface
->p
->channel
[id
];
1100 mutex_lock(&c
->start_mutex
);
1101 if (c
->pipe0
.refs
+ c
->pipe1
.refs
> 0)
1102 goto out
; /* already started by another component */
1104 if (!try_module_get(iface
->mod
)) {
1105 dev_err(&c
->dev
, "Failed to acquire HDM lock\n");
1106 mutex_unlock(&c
->start_mutex
);
1110 c
->cfg
.extra_len
= 0;
1111 if (c
->iface
->configure(c
->iface
, c
->channel_id
, &c
->cfg
)) {
1112 dev_err(&c
->dev
, "Channel configuration failed. Go check settings...\n");
1114 goto err_put_module
;
1117 init_waitqueue_head(&c
->hdm_fifo_wq
);
1119 if (c
->cfg
.direction
== MOST_CH_RX
)
1120 num_buffer
= arm_mbo_chain(c
, c
->cfg
.direction
,
1121 most_read_completion
);
1123 num_buffer
= arm_mbo_chain(c
, c
->cfg
.direction
,
1124 most_write_completion
);
1125 if (unlikely(!num_buffer
)) {
1127 goto err_put_module
;
1130 ret
= run_enqueue_thread(c
, id
);
1132 goto err_put_module
;
1135 c
->pipe0
.num_buffers
= c
->cfg
.num_buffers
/ 2;
1136 c
->pipe1
.num_buffers
= c
->cfg
.num_buffers
- c
->pipe0
.num_buffers
;
1137 atomic_set(&c
->mbo_ref
, num_buffer
);
1140 if (comp
== c
->pipe0
.comp
)
1142 if (comp
== c
->pipe1
.comp
)
1144 mutex_unlock(&c
->start_mutex
);
1148 module_put(iface
->mod
);
1149 mutex_unlock(&c
->start_mutex
);
1152 EXPORT_SYMBOL_GPL(most_start_channel
);
1155 * most_stop_channel - stops a running channel
1156 * @iface: pointer to interface instance
1158 * @comp: driver component
1160 int most_stop_channel(struct most_interface
*iface
, int id
,
1161 struct most_component
*comp
)
1163 struct most_channel
*c
;
1165 if (unlikely((!iface
) || (id
>= iface
->num_channels
) || (id
< 0))) {
1166 pr_err("Bad interface or index out of range\n");
1169 c
= iface
->p
->channel
[id
];
1173 mutex_lock(&c
->start_mutex
);
1174 if (c
->pipe0
.refs
+ c
->pipe1
.refs
>= 2)
1177 if (c
->hdm_enqueue_task
)
1178 kthread_stop(c
->hdm_enqueue_task
);
1179 c
->hdm_enqueue_task
= NULL
;
1182 module_put(iface
->mod
);
1184 c
->is_poisoned
= true;
1185 if (c
->iface
->poison_channel(c
->iface
, c
->channel_id
)) {
1186 dev_err(&c
->dev
, "Failed to stop channel %d of interface %s\n", c
->channel_id
,
1187 c
->iface
->description
);
1188 mutex_unlock(&c
->start_mutex
);
1191 flush_trash_fifo(c
);
1192 flush_channel_fifos(c
);
1194 #ifdef CMPL_INTERRUPTIBLE
1195 if (wait_for_completion_interruptible(&c
->cleanup
)) {
1196 dev_err(&c
->dev
, "Interrupted while cleaning up channel %d\n", c
->channel_id
);
1197 mutex_unlock(&c
->start_mutex
);
1201 wait_for_completion(&c
->cleanup
);
1203 c
->is_poisoned
= false;
1206 if (comp
== c
->pipe0
.comp
)
1208 if (comp
== c
->pipe1
.comp
)
1210 mutex_unlock(&c
->start_mutex
);
1213 EXPORT_SYMBOL_GPL(most_stop_channel
);
1216 * most_register_component - registers a driver component with the core
1217 * @comp: driver component
1219 int most_register_component(struct most_component
*comp
)
1222 pr_err("Bad component\n");
1225 list_add_tail(&comp
->list
, &comp_list
);
1228 EXPORT_SYMBOL_GPL(most_register_component
);
1230 static int disconnect_channels(struct device
*dev
, void *data
)
1232 struct most_interface
*iface
;
1233 struct most_channel
*c
, *tmp
;
1234 struct most_component
*comp
= data
;
1236 iface
= dev_get_drvdata(dev
);
1237 list_for_each_entry_safe(c
, tmp
, &iface
->p
->channel_list
, list
) {
1238 if (c
->pipe0
.comp
== comp
|| c
->pipe1
.comp
== comp
)
1239 comp
->disconnect_channel(c
->iface
, c
->channel_id
);
1240 if (c
->pipe0
.comp
== comp
)
1241 c
->pipe0
.comp
= NULL
;
1242 if (c
->pipe1
.comp
== comp
)
1243 c
->pipe1
.comp
= NULL
;
1249 * most_deregister_component - deregisters a driver component with the core
1250 * @comp: driver component
1252 int most_deregister_component(struct most_component
*comp
)
1255 pr_err("Bad component\n");
1259 bus_for_each_dev(&mostbus
, NULL
, comp
, disconnect_channels
);
1260 list_del(&comp
->list
);
1263 EXPORT_SYMBOL_GPL(most_deregister_component
);
1265 static void release_channel(struct device
*dev
)
1267 struct most_channel
*c
= to_channel(dev
);
1273 * most_register_interface - registers an interface with core
1274 * @iface: device interface
1276 * Allocates and initializes a new interface instance and all of its channels.
1277 * Returns a pointer to kobject or an error pointer.
1279 int most_register_interface(struct most_interface
*iface
)
1283 struct most_channel
*c
;
1285 if (!iface
|| !iface
->enqueue
|| !iface
->configure
||
1286 !iface
->poison_channel
|| (iface
->num_channels
> MAX_CHANNELS
))
1289 id
= ida_simple_get(&mdev_id
, 0, 0, GFP_KERNEL
);
1291 dev_err(iface
->dev
, "Failed to allocate device ID\n");
1295 iface
->p
= kzalloc(sizeof(*iface
->p
), GFP_KERNEL
);
1297 ida_simple_remove(&mdev_id
, id
);
1301 INIT_LIST_HEAD(&iface
->p
->channel_list
);
1302 iface
->p
->dev_id
= id
;
1303 strscpy(iface
->p
->name
, iface
->description
, sizeof(iface
->p
->name
));
1304 iface
->dev
->bus
= &mostbus
;
1305 iface
->dev
->groups
= interface_attr_groups
;
1306 dev_set_drvdata(iface
->dev
, iface
);
1307 if (device_register(iface
->dev
)) {
1308 dev_err(iface
->dev
, "Failed to register interface device\n");
1310 put_device(iface
->dev
);
1311 ida_simple_remove(&mdev_id
, id
);
1315 for (i
= 0; i
< iface
->num_channels
; i
++) {
1316 const char *name_suffix
= iface
->channel_vector
[i
].name_suffix
;
1318 c
= kzalloc(sizeof(*c
), GFP_KERNEL
);
1320 goto err_free_resources
;
1322 snprintf(c
->name
, STRING_SIZE
, "ch%d", i
);
1324 snprintf(c
->name
, STRING_SIZE
, "%s", name_suffix
);
1325 c
->dev
.init_name
= c
->name
;
1326 c
->dev
.parent
= iface
->dev
;
1327 c
->dev
.groups
= channel_attr_groups
;
1328 c
->dev
.release
= release_channel
;
1329 iface
->p
->channel
[i
] = c
;
1333 c
->keep_mbo
= false;
1334 c
->enqueue_halt
= false;
1335 c
->is_poisoned
= false;
1336 c
->cfg
.direction
= 0;
1337 c
->cfg
.data_type
= 0;
1338 c
->cfg
.num_buffers
= 0;
1339 c
->cfg
.buffer_size
= 0;
1340 c
->cfg
.subbuffer_size
= 0;
1341 c
->cfg
.packets_per_xact
= 0;
1342 spin_lock_init(&c
->fifo_lock
);
1343 INIT_LIST_HEAD(&c
->fifo
);
1344 INIT_LIST_HEAD(&c
->trash_fifo
);
1345 INIT_LIST_HEAD(&c
->halt_fifo
);
1346 init_completion(&c
->cleanup
);
1347 atomic_set(&c
->mbo_ref
, 0);
1348 mutex_init(&c
->start_mutex
);
1349 mutex_init(&c
->nq_mutex
);
1350 list_add_tail(&c
->list
, &iface
->p
->channel_list
);
1351 if (device_register(&c
->dev
)) {
1352 dev_err(&c
->dev
, "Failed to register channel device\n");
1353 goto err_free_most_channel
;
1356 most_interface_register_notify(iface
->description
);
1359 err_free_most_channel
:
1360 put_device(&c
->dev
);
1364 c
= iface
->p
->channel
[--i
];
1365 device_unregister(&c
->dev
);
1368 device_unregister(iface
->dev
);
1369 ida_simple_remove(&mdev_id
, id
);
1372 EXPORT_SYMBOL_GPL(most_register_interface
);
1375 * most_deregister_interface - deregisters an interface with core
1376 * @iface: device interface
1378 * Before removing an interface instance from the list, all running
1379 * channels are stopped and poisoned.
1381 void most_deregister_interface(struct most_interface
*iface
)
1384 struct most_channel
*c
;
1386 for (i
= 0; i
< iface
->num_channels
; i
++) {
1387 c
= iface
->p
->channel
[i
];
1389 c
->pipe0
.comp
->disconnect_channel(c
->iface
,
1392 c
->pipe1
.comp
->disconnect_channel(c
->iface
,
1394 c
->pipe0
.comp
= NULL
;
1395 c
->pipe1
.comp
= NULL
;
1397 device_unregister(&c
->dev
);
1400 ida_simple_remove(&mdev_id
, iface
->p
->dev_id
);
1402 device_unregister(iface
->dev
);
1404 EXPORT_SYMBOL_GPL(most_deregister_interface
);
1407 * most_stop_enqueue - prevents core from enqueueing MBOs
1408 * @iface: pointer to interface
1411 * This is called by an HDM that _cannot_ attend to its duties and
1412 * is imminent to get run over by the core. The core is not going to
1413 * enqueue any further packets unless the flagging HDM calls
1414 * most_resume enqueue().
1416 void most_stop_enqueue(struct most_interface
*iface
, int id
)
1418 struct most_channel
*c
= iface
->p
->channel
[id
];
1423 mutex_lock(&c
->nq_mutex
);
1424 c
->enqueue_halt
= true;
1425 mutex_unlock(&c
->nq_mutex
);
1427 EXPORT_SYMBOL_GPL(most_stop_enqueue
);
1430 * most_resume_enqueue - allow core to enqueue MBOs again
1431 * @iface: pointer to interface
1434 * This clears the enqueue halt flag and enqueues all MBOs currently
1435 * sitting in the wait fifo.
1437 void most_resume_enqueue(struct most_interface
*iface
, int id
)
1439 struct most_channel
*c
= iface
->p
->channel
[id
];
1444 mutex_lock(&c
->nq_mutex
);
1445 c
->enqueue_halt
= false;
1446 mutex_unlock(&c
->nq_mutex
);
1448 wake_up_interruptible(&c
->hdm_fifo_wq
);
1450 EXPORT_SYMBOL_GPL(most_resume_enqueue
);
1452 static int __init
most_init(void)
1456 INIT_LIST_HEAD(&comp_list
);
1459 err
= bus_register(&mostbus
);
1461 pr_err("Failed to register most bus\n");
1464 err
= driver_register(&mostbus_driver
);
1466 pr_err("Failed to register core driver\n");
1467 goto err_unregister_bus
;
1473 bus_unregister(&mostbus
);
1477 static void __exit
most_exit(void)
1479 driver_unregister(&mostbus_driver
);
1480 bus_unregister(&mostbus
);
1481 ida_destroy(&mdev_id
);
1484 subsys_initcall(most_init
);
1485 module_exit(most_exit
);
1486 MODULE_LICENSE("GPL");
1487 MODULE_AUTHOR("Christian Gromm <christian.gromm@microchip.com>");
1488 MODULE_DESCRIPTION("Core module of stacked MOST Linux driver");