4 * Author: Martyn Welch <martyn.welch@ge.com>
5 * Copyright 2008 GE Intelligent Platforms Embedded Systems, Inc.
7 * Based on work by Tom Armistead and Ajit Prem
8 * Copyright 2004 Motorola Inc.
10 * This program is free software; you can redistribute it and/or modify it
11 * under the terms of the GNU General Public License as published by the
12 * Free Software Foundation; either version 2 of the License, or (at your
13 * option) any later version.
16 #include <linux/module.h>
17 #include <linux/moduleparam.h>
19 #include <linux/types.h>
20 #include <linux/kernel.h>
21 #include <linux/errno.h>
22 #include <linux/pci.h>
23 #include <linux/poll.h>
24 #include <linux/highmem.h>
25 #include <linux/interrupt.h>
26 #include <linux/pagemap.h>
27 #include <linux/device.h>
28 #include <linux/dma-mapping.h>
29 #include <linux/syscalls.h>
30 #include <linux/mutex.h>
31 #include <linux/spinlock.h>
32 #include <linux/slab.h>
35 #include "vme_bridge.h"
37 /* Bitmask and mutex to keep track of bridge numbers */
38 static unsigned int vme_bus_numbers
;
39 static DEFINE_MUTEX(vme_bus_num_mtx
);
41 static void __exit
vme_exit(void);
42 static int __init
vme_init(void);
46 * Find the bridge resource associated with a specific device resource
48 static struct vme_bridge
*dev_to_bridge(struct device
*dev
)
50 return dev
->platform_data
;
54 * Find the bridge that the resource is associated with.
56 static struct vme_bridge
*find_bridge(struct vme_resource
*resource
)
58 /* Get list to search */
59 switch (resource
->type
) {
61 return list_entry(resource
->entry
, struct vme_master_resource
,
65 return list_entry(resource
->entry
, struct vme_slave_resource
,
69 return list_entry(resource
->entry
, struct vme_dma_resource
,
73 return list_entry(resource
->entry
, struct vme_lm_resource
,
77 printk(KERN_ERR
"Unknown resource type\n");
84 * Allocate a contiguous block of memory for use by the driver. This is used to
85 * create the buffers for the slave windows.
87 * XXX VME bridges could be available on buses other than PCI. At the momment
88 * this framework only supports PCI devices.
90 void *vme_alloc_consistent(struct vme_resource
*resource
, size_t size
,
93 struct vme_bridge
*bridge
;
96 if (resource
== NULL
) {
97 printk(KERN_ERR
"No resource\n");
101 bridge
= find_bridge(resource
);
102 if (bridge
== NULL
) {
103 printk(KERN_ERR
"Can't find bridge\n");
107 /* Find pci_dev container of dev */
108 if (bridge
->parent
== NULL
) {
109 printk(KERN_ERR
"Dev entry NULL\n");
112 pdev
= container_of(bridge
->parent
, struct pci_dev
, dev
);
114 return pci_alloc_consistent(pdev
, size
, dma
);
116 EXPORT_SYMBOL(vme_alloc_consistent
);
119 * Free previously allocated contiguous block of memory.
121 * XXX VME bridges could be available on buses other than PCI. At the momment
122 * this framework only supports PCI devices.
124 void vme_free_consistent(struct vme_resource
*resource
, size_t size
,
125 void *vaddr
, dma_addr_t dma
)
127 struct vme_bridge
*bridge
;
128 struct pci_dev
*pdev
;
130 if (resource
== NULL
) {
131 printk(KERN_ERR
"No resource\n");
135 bridge
= find_bridge(resource
);
136 if (bridge
== NULL
) {
137 printk(KERN_ERR
"Can't find bridge\n");
141 /* Find pci_dev container of dev */
142 pdev
= container_of(bridge
->parent
, struct pci_dev
, dev
);
144 pci_free_consistent(pdev
, size
, vaddr
, dma
);
146 EXPORT_SYMBOL(vme_free_consistent
);
148 size_t vme_get_size(struct vme_resource
*resource
)
151 unsigned long long base
, size
;
153 vme_address_t aspace
;
157 switch (resource
->type
) {
159 retval
= vme_master_get(resource
, &enabled
, &base
, &size
,
160 &aspace
, &cycle
, &dwidth
);
165 retval
= vme_slave_get(resource
, &enabled
, &base
, &size
,
166 &buf_base
, &aspace
, &cycle
);
174 printk(KERN_ERR
"Unknown resource type\n");
179 EXPORT_SYMBOL(vme_get_size
);
181 static int vme_check_window(vme_address_t aspace
, unsigned long long vme_base
,
182 unsigned long long size
)
188 if (((vme_base
+ size
) > VME_A16_MAX
) ||
189 (vme_base
> VME_A16_MAX
))
193 if (((vme_base
+ size
) > VME_A24_MAX
) ||
194 (vme_base
> VME_A24_MAX
))
198 if (((vme_base
+ size
) > VME_A32_MAX
) ||
199 (vme_base
> VME_A32_MAX
))
204 * Any value held in an unsigned long long can be used as the
209 if (((vme_base
+ size
) > VME_CRCSR_MAX
) ||
210 (vme_base
> VME_CRCSR_MAX
))
220 printk(KERN_ERR
"Invalid address space\n");
229 * Request a slave image with specific attributes, return some unique
232 struct vme_resource
*vme_slave_request(struct device
*dev
,
233 vme_address_t address
, vme_cycle_t cycle
)
235 struct vme_bridge
*bridge
;
236 struct list_head
*slave_pos
= NULL
;
237 struct vme_slave_resource
*allocated_image
= NULL
;
238 struct vme_slave_resource
*slave_image
= NULL
;
239 struct vme_resource
*resource
= NULL
;
241 bridge
= dev_to_bridge(dev
);
242 if (bridge
== NULL
) {
243 printk(KERN_ERR
"Can't find VME bus\n");
247 /* Loop through slave resources */
248 list_for_each(slave_pos
, &bridge
->slave_resources
) {
249 slave_image
= list_entry(slave_pos
,
250 struct vme_slave_resource
, list
);
252 if (slave_image
== NULL
) {
253 printk(KERN_ERR
"Registered NULL Slave resource\n");
257 /* Find an unlocked and compatible image */
258 mutex_lock(&slave_image
->mtx
);
259 if (((slave_image
->address_attr
& address
) == address
) &&
260 ((slave_image
->cycle_attr
& cycle
) == cycle
) &&
261 (slave_image
->locked
== 0)) {
263 slave_image
->locked
= 1;
264 mutex_unlock(&slave_image
->mtx
);
265 allocated_image
= slave_image
;
268 mutex_unlock(&slave_image
->mtx
);
272 if (allocated_image
== NULL
)
275 resource
= kmalloc(sizeof(struct vme_resource
), GFP_KERNEL
);
276 if (resource
== NULL
) {
277 printk(KERN_WARNING
"Unable to allocate resource structure\n");
280 resource
->type
= VME_SLAVE
;
281 resource
->entry
= &allocated_image
->list
;
287 mutex_lock(&slave_image
->mtx
);
288 slave_image
->locked
= 0;
289 mutex_unlock(&slave_image
->mtx
);
294 EXPORT_SYMBOL(vme_slave_request
);
296 int vme_slave_set(struct vme_resource
*resource
, int enabled
,
297 unsigned long long vme_base
, unsigned long long size
,
298 dma_addr_t buf_base
, vme_address_t aspace
, vme_cycle_t cycle
)
300 struct vme_bridge
*bridge
= find_bridge(resource
);
301 struct vme_slave_resource
*image
;
304 if (resource
->type
!= VME_SLAVE
) {
305 printk(KERN_ERR
"Not a slave resource\n");
309 image
= list_entry(resource
->entry
, struct vme_slave_resource
, list
);
311 if (bridge
->slave_set
== NULL
) {
312 printk(KERN_ERR
"Function not supported\n");
316 if (!(((image
->address_attr
& aspace
) == aspace
) &&
317 ((image
->cycle_attr
& cycle
) == cycle
))) {
318 printk(KERN_ERR
"Invalid attributes\n");
322 retval
= vme_check_window(aspace
, vme_base
, size
);
326 return bridge
->slave_set(image
, enabled
, vme_base
, size
, buf_base
,
329 EXPORT_SYMBOL(vme_slave_set
);
331 int vme_slave_get(struct vme_resource
*resource
, int *enabled
,
332 unsigned long long *vme_base
, unsigned long long *size
,
333 dma_addr_t
*buf_base
, vme_address_t
*aspace
, vme_cycle_t
*cycle
)
335 struct vme_bridge
*bridge
= find_bridge(resource
);
336 struct vme_slave_resource
*image
;
338 if (resource
->type
!= VME_SLAVE
) {
339 printk(KERN_ERR
"Not a slave resource\n");
343 image
= list_entry(resource
->entry
, struct vme_slave_resource
, list
);
345 if (bridge
->slave_get
== NULL
) {
346 printk(KERN_ERR
"vme_slave_get not supported\n");
350 return bridge
->slave_get(image
, enabled
, vme_base
, size
, buf_base
,
353 EXPORT_SYMBOL(vme_slave_get
);
355 void vme_slave_free(struct vme_resource
*resource
)
357 struct vme_slave_resource
*slave_image
;
359 if (resource
->type
!= VME_SLAVE
) {
360 printk(KERN_ERR
"Not a slave resource\n");
364 slave_image
= list_entry(resource
->entry
, struct vme_slave_resource
,
366 if (slave_image
== NULL
) {
367 printk(KERN_ERR
"Can't find slave resource\n");
372 mutex_lock(&slave_image
->mtx
);
373 if (slave_image
->locked
== 0)
374 printk(KERN_ERR
"Image is already free\n");
376 slave_image
->locked
= 0;
377 mutex_unlock(&slave_image
->mtx
);
379 /* Free up resource memory */
382 EXPORT_SYMBOL(vme_slave_free
);
385 * Request a master image with specific attributes, return some unique
388 struct vme_resource
*vme_master_request(struct device
*dev
,
389 vme_address_t address
, vme_cycle_t cycle
, vme_width_t dwidth
)
391 struct vme_bridge
*bridge
;
392 struct list_head
*master_pos
= NULL
;
393 struct vme_master_resource
*allocated_image
= NULL
;
394 struct vme_master_resource
*master_image
= NULL
;
395 struct vme_resource
*resource
= NULL
;
397 bridge
= dev_to_bridge(dev
);
398 if (bridge
== NULL
) {
399 printk(KERN_ERR
"Can't find VME bus\n");
403 /* Loop through master resources */
404 list_for_each(master_pos
, &bridge
->master_resources
) {
405 master_image
= list_entry(master_pos
,
406 struct vme_master_resource
, list
);
408 if (master_image
== NULL
) {
409 printk(KERN_WARNING
"Registered NULL master resource\n");
413 /* Find an unlocked and compatible image */
414 spin_lock(&master_image
->lock
);
415 if (((master_image
->address_attr
& address
) == address
) &&
416 ((master_image
->cycle_attr
& cycle
) == cycle
) &&
417 ((master_image
->width_attr
& dwidth
) == dwidth
) &&
418 (master_image
->locked
== 0)) {
420 master_image
->locked
= 1;
421 spin_unlock(&master_image
->lock
);
422 allocated_image
= master_image
;
425 spin_unlock(&master_image
->lock
);
428 /* Check to see if we found a resource */
429 if (allocated_image
== NULL
) {
430 printk(KERN_ERR
"Can't find a suitable resource\n");
434 resource
= kmalloc(sizeof(struct vme_resource
), GFP_KERNEL
);
435 if (resource
== NULL
) {
436 printk(KERN_ERR
"Unable to allocate resource structure\n");
439 resource
->type
= VME_MASTER
;
440 resource
->entry
= &allocated_image
->list
;
446 spin_lock(&master_image
->lock
);
447 master_image
->locked
= 0;
448 spin_unlock(&master_image
->lock
);
453 EXPORT_SYMBOL(vme_master_request
);
455 int vme_master_set(struct vme_resource
*resource
, int enabled
,
456 unsigned long long vme_base
, unsigned long long size
,
457 vme_address_t aspace
, vme_cycle_t cycle
, vme_width_t dwidth
)
459 struct vme_bridge
*bridge
= find_bridge(resource
);
460 struct vme_master_resource
*image
;
463 if (resource
->type
!= VME_MASTER
) {
464 printk(KERN_ERR
"Not a master resource\n");
468 image
= list_entry(resource
->entry
, struct vme_master_resource
, list
);
470 if (bridge
->master_set
== NULL
) {
471 printk(KERN_WARNING
"vme_master_set not supported\n");
475 if (!(((image
->address_attr
& aspace
) == aspace
) &&
476 ((image
->cycle_attr
& cycle
) == cycle
) &&
477 ((image
->width_attr
& dwidth
) == dwidth
))) {
478 printk(KERN_WARNING
"Invalid attributes\n");
482 retval
= vme_check_window(aspace
, vme_base
, size
);
486 return bridge
->master_set(image
, enabled
, vme_base
, size
, aspace
,
489 EXPORT_SYMBOL(vme_master_set
);
491 int vme_master_get(struct vme_resource
*resource
, int *enabled
,
492 unsigned long long *vme_base
, unsigned long long *size
,
493 vme_address_t
*aspace
, vme_cycle_t
*cycle
, vme_width_t
*dwidth
)
495 struct vme_bridge
*bridge
= find_bridge(resource
);
496 struct vme_master_resource
*image
;
498 if (resource
->type
!= VME_MASTER
) {
499 printk(KERN_ERR
"Not a master resource\n");
503 image
= list_entry(resource
->entry
, struct vme_master_resource
, list
);
505 if (bridge
->master_get
== NULL
) {
506 printk(KERN_WARNING
"vme_master_set not supported\n");
510 return bridge
->master_get(image
, enabled
, vme_base
, size
, aspace
,
513 EXPORT_SYMBOL(vme_master_get
);
516 * Read data out of VME space into a buffer.
518 ssize_t
vme_master_read(struct vme_resource
*resource
, void *buf
, size_t count
,
521 struct vme_bridge
*bridge
= find_bridge(resource
);
522 struct vme_master_resource
*image
;
525 if (bridge
->master_read
== NULL
) {
526 printk(KERN_WARNING
"Reading from resource not supported\n");
530 if (resource
->type
!= VME_MASTER
) {
531 printk(KERN_ERR
"Not a master resource\n");
535 image
= list_entry(resource
->entry
, struct vme_master_resource
, list
);
537 length
= vme_get_size(resource
);
539 if (offset
> length
) {
540 printk(KERN_WARNING
"Invalid Offset\n");
544 if ((offset
+ count
) > length
)
545 count
= length
- offset
;
547 return bridge
->master_read(image
, buf
, count
, offset
);
550 EXPORT_SYMBOL(vme_master_read
);
553 * Write data out to VME space from a buffer.
555 ssize_t
vme_master_write(struct vme_resource
*resource
, void *buf
,
556 size_t count
, loff_t offset
)
558 struct vme_bridge
*bridge
= find_bridge(resource
);
559 struct vme_master_resource
*image
;
562 if (bridge
->master_write
== NULL
) {
563 printk(KERN_WARNING
"Writing to resource not supported\n");
567 if (resource
->type
!= VME_MASTER
) {
568 printk(KERN_ERR
"Not a master resource\n");
572 image
= list_entry(resource
->entry
, struct vme_master_resource
, list
);
574 length
= vme_get_size(resource
);
576 if (offset
> length
) {
577 printk(KERN_WARNING
"Invalid Offset\n");
581 if ((offset
+ count
) > length
)
582 count
= length
- offset
;
584 return bridge
->master_write(image
, buf
, count
, offset
);
586 EXPORT_SYMBOL(vme_master_write
);
589 * Perform RMW cycle to provided location.
591 unsigned int vme_master_rmw(struct vme_resource
*resource
, unsigned int mask
,
592 unsigned int compare
, unsigned int swap
, loff_t offset
)
594 struct vme_bridge
*bridge
= find_bridge(resource
);
595 struct vme_master_resource
*image
;
597 if (bridge
->master_rmw
== NULL
) {
598 printk(KERN_WARNING
"Writing to resource not supported\n");
602 if (resource
->type
!= VME_MASTER
) {
603 printk(KERN_ERR
"Not a master resource\n");
607 image
= list_entry(resource
->entry
, struct vme_master_resource
, list
);
609 return bridge
->master_rmw(image
, mask
, compare
, swap
, offset
);
611 EXPORT_SYMBOL(vme_master_rmw
);
613 void vme_master_free(struct vme_resource
*resource
)
615 struct vme_master_resource
*master_image
;
617 if (resource
->type
!= VME_MASTER
) {
618 printk(KERN_ERR
"Not a master resource\n");
622 master_image
= list_entry(resource
->entry
, struct vme_master_resource
,
624 if (master_image
== NULL
) {
625 printk(KERN_ERR
"Can't find master resource\n");
630 spin_lock(&master_image
->lock
);
631 if (master_image
->locked
== 0)
632 printk(KERN_ERR
"Image is already free\n");
634 master_image
->locked
= 0;
635 spin_unlock(&master_image
->lock
);
637 /* Free up resource memory */
640 EXPORT_SYMBOL(vme_master_free
);
643 * Request a DMA controller with specific attributes, return some unique
646 struct vme_resource
*vme_dma_request(struct device
*dev
, vme_dma_route_t route
)
648 struct vme_bridge
*bridge
;
649 struct list_head
*dma_pos
= NULL
;
650 struct vme_dma_resource
*allocated_ctrlr
= NULL
;
651 struct vme_dma_resource
*dma_ctrlr
= NULL
;
652 struct vme_resource
*resource
= NULL
;
654 /* XXX Not checking resource attributes */
655 printk(KERN_ERR
"No VME resource Attribute tests done\n");
657 bridge
= dev_to_bridge(dev
);
658 if (bridge
== NULL
) {
659 printk(KERN_ERR
"Can't find VME bus\n");
663 /* Loop through DMA resources */
664 list_for_each(dma_pos
, &bridge
->dma_resources
) {
665 dma_ctrlr
= list_entry(dma_pos
,
666 struct vme_dma_resource
, list
);
668 if (dma_ctrlr
== NULL
) {
669 printk(KERN_ERR
"Registered NULL DMA resource\n");
673 /* Find an unlocked and compatible controller */
674 mutex_lock(&dma_ctrlr
->mtx
);
675 if (((dma_ctrlr
->route_attr
& route
) == route
) &&
676 (dma_ctrlr
->locked
== 0)) {
678 dma_ctrlr
->locked
= 1;
679 mutex_unlock(&dma_ctrlr
->mtx
);
680 allocated_ctrlr
= dma_ctrlr
;
683 mutex_unlock(&dma_ctrlr
->mtx
);
686 /* Check to see if we found a resource */
687 if (allocated_ctrlr
== NULL
)
690 resource
= kmalloc(sizeof(struct vme_resource
), GFP_KERNEL
);
691 if (resource
== NULL
) {
692 printk(KERN_WARNING
"Unable to allocate resource structure\n");
695 resource
->type
= VME_DMA
;
696 resource
->entry
= &allocated_ctrlr
->list
;
702 mutex_lock(&dma_ctrlr
->mtx
);
703 dma_ctrlr
->locked
= 0;
704 mutex_unlock(&dma_ctrlr
->mtx
);
709 EXPORT_SYMBOL(vme_dma_request
);
714 struct vme_dma_list
*vme_new_dma_list(struct vme_resource
*resource
)
716 struct vme_dma_resource
*ctrlr
;
717 struct vme_dma_list
*dma_list
;
719 if (resource
->type
!= VME_DMA
) {
720 printk(KERN_ERR
"Not a DMA resource\n");
724 ctrlr
= list_entry(resource
->entry
, struct vme_dma_resource
, list
);
726 dma_list
= kmalloc(sizeof(struct vme_dma_list
), GFP_KERNEL
);
727 if (dma_list
== NULL
) {
728 printk(KERN_ERR
"Unable to allocate memory for new dma list\n");
731 INIT_LIST_HEAD(&dma_list
->entries
);
732 dma_list
->parent
= ctrlr
;
733 mutex_init(&dma_list
->mtx
);
737 EXPORT_SYMBOL(vme_new_dma_list
);
740 * Create "Pattern" type attributes
742 struct vme_dma_attr
*vme_dma_pattern_attribute(u32 pattern
,
745 struct vme_dma_attr
*attributes
;
746 struct vme_dma_pattern
*pattern_attr
;
748 attributes
= kmalloc(sizeof(struct vme_dma_attr
), GFP_KERNEL
);
749 if (attributes
== NULL
) {
750 printk(KERN_ERR
"Unable to allocate memory for attributes "
755 pattern_attr
= kmalloc(sizeof(struct vme_dma_pattern
), GFP_KERNEL
);
756 if (pattern_attr
== NULL
) {
757 printk(KERN_ERR
"Unable to allocate memory for pattern "
762 attributes
->type
= VME_DMA_PATTERN
;
763 attributes
->private = (void *)pattern_attr
;
765 pattern_attr
->pattern
= pattern
;
766 pattern_attr
->type
= type
;
775 EXPORT_SYMBOL(vme_dma_pattern_attribute
);
778 * Create "PCI" type attributes
780 struct vme_dma_attr
*vme_dma_pci_attribute(dma_addr_t address
)
782 struct vme_dma_attr
*attributes
;
783 struct vme_dma_pci
*pci_attr
;
785 /* XXX Run some sanity checks here */
787 attributes
= kmalloc(sizeof(struct vme_dma_attr
), GFP_KERNEL
);
788 if (attributes
== NULL
) {
789 printk(KERN_ERR
"Unable to allocate memory for attributes "
794 pci_attr
= kmalloc(sizeof(struct vme_dma_pci
), GFP_KERNEL
);
795 if (pci_attr
== NULL
) {
796 printk(KERN_ERR
"Unable to allocate memory for pci "
803 attributes
->type
= VME_DMA_PCI
;
804 attributes
->private = (void *)pci_attr
;
806 pci_attr
->address
= address
;
815 EXPORT_SYMBOL(vme_dma_pci_attribute
);
818 * Create "VME" type attributes
820 struct vme_dma_attr
*vme_dma_vme_attribute(unsigned long long address
,
821 vme_address_t aspace
, vme_cycle_t cycle
, vme_width_t dwidth
)
823 struct vme_dma_attr
*attributes
;
824 struct vme_dma_vme
*vme_attr
;
826 attributes
= kmalloc(
827 sizeof(struct vme_dma_attr
), GFP_KERNEL
);
828 if (attributes
== NULL
) {
829 printk(KERN_ERR
"Unable to allocate memory for attributes "
834 vme_attr
= kmalloc(sizeof(struct vme_dma_vme
), GFP_KERNEL
);
835 if (vme_attr
== NULL
) {
836 printk(KERN_ERR
"Unable to allocate memory for vme "
841 attributes
->type
= VME_DMA_VME
;
842 attributes
->private = (void *)vme_attr
;
844 vme_attr
->address
= address
;
845 vme_attr
->aspace
= aspace
;
846 vme_attr
->cycle
= cycle
;
847 vme_attr
->dwidth
= dwidth
;
856 EXPORT_SYMBOL(vme_dma_vme_attribute
);
861 void vme_dma_free_attribute(struct vme_dma_attr
*attributes
)
863 kfree(attributes
->private);
866 EXPORT_SYMBOL(vme_dma_free_attribute
);
868 int vme_dma_list_add(struct vme_dma_list
*list
, struct vme_dma_attr
*src
,
869 struct vme_dma_attr
*dest
, size_t count
)
871 struct vme_bridge
*bridge
= list
->parent
->parent
;
874 if (bridge
->dma_list_add
== NULL
) {
875 printk(KERN_WARNING
"Link List DMA generation not supported\n");
879 if (!mutex_trylock(&list
->mtx
)) {
880 printk(KERN_ERR
"Link List already submitted\n");
884 retval
= bridge
->dma_list_add(list
, src
, dest
, count
);
886 mutex_unlock(&list
->mtx
);
890 EXPORT_SYMBOL(vme_dma_list_add
);
892 int vme_dma_list_exec(struct vme_dma_list
*list
)
894 struct vme_bridge
*bridge
= list
->parent
->parent
;
897 if (bridge
->dma_list_exec
== NULL
) {
898 printk(KERN_ERR
"Link List DMA execution not supported\n");
902 mutex_lock(&list
->mtx
);
904 retval
= bridge
->dma_list_exec(list
);
906 mutex_unlock(&list
->mtx
);
910 EXPORT_SYMBOL(vme_dma_list_exec
);
912 int vme_dma_list_free(struct vme_dma_list
*list
)
914 struct vme_bridge
*bridge
= list
->parent
->parent
;
917 if (bridge
->dma_list_empty
== NULL
) {
918 printk(KERN_WARNING
"Emptying of Link Lists not supported\n");
922 if (!mutex_trylock(&list
->mtx
)) {
923 printk(KERN_ERR
"Link List in use\n");
928 * Empty out all of the entries from the dma list. We need to go to the
929 * low level driver as dma entries are driver specific.
931 retval
= bridge
->dma_list_empty(list
);
933 printk(KERN_ERR
"Unable to empty link-list entries\n");
934 mutex_unlock(&list
->mtx
);
937 mutex_unlock(&list
->mtx
);
942 EXPORT_SYMBOL(vme_dma_list_free
);
944 int vme_dma_free(struct vme_resource
*resource
)
946 struct vme_dma_resource
*ctrlr
;
948 if (resource
->type
!= VME_DMA
) {
949 printk(KERN_ERR
"Not a DMA resource\n");
953 ctrlr
= list_entry(resource
->entry
, struct vme_dma_resource
, list
);
955 if (!mutex_trylock(&ctrlr
->mtx
)) {
956 printk(KERN_ERR
"Resource busy, can't free\n");
960 if (!(list_empty(&ctrlr
->pending
) && list_empty(&ctrlr
->running
))) {
961 printk(KERN_WARNING
"Resource still processing transfers\n");
962 mutex_unlock(&ctrlr
->mtx
);
968 mutex_unlock(&ctrlr
->mtx
);
972 EXPORT_SYMBOL(vme_dma_free
);
974 void vme_irq_handler(struct vme_bridge
*bridge
, int level
, int statid
)
976 void (*call
)(int, int, void *);
979 call
= bridge
->irq
[level
- 1].callback
[statid
].func
;
980 priv_data
= bridge
->irq
[level
- 1].callback
[statid
].priv_data
;
983 call(level
, statid
, priv_data
);
985 printk(KERN_WARNING
"Spurilous VME interrupt, level:%x, "
986 "vector:%x\n", level
, statid
);
988 EXPORT_SYMBOL(vme_irq_handler
);
990 int vme_irq_request(struct device
*dev
, int level
, int statid
,
991 void (*callback
)(int, int, void *),
994 struct vme_bridge
*bridge
;
996 bridge
= dev_to_bridge(dev
);
997 if (bridge
== NULL
) {
998 printk(KERN_ERR
"Can't find VME bus\n");
1002 if ((level
< 1) || (level
> 7)) {
1003 printk(KERN_ERR
"Invalid interrupt level\n");
1007 if (bridge
->irq_set
== NULL
) {
1008 printk(KERN_ERR
"Configuring interrupts not supported\n");
1012 mutex_lock(&bridge
->irq_mtx
);
1014 if (bridge
->irq
[level
- 1].callback
[statid
].func
) {
1015 mutex_unlock(&bridge
->irq_mtx
);
1016 printk(KERN_WARNING
"VME Interrupt already taken\n");
1020 bridge
->irq
[level
- 1].count
++;
1021 bridge
->irq
[level
- 1].callback
[statid
].priv_data
= priv_data
;
1022 bridge
->irq
[level
- 1].callback
[statid
].func
= callback
;
1024 /* Enable IRQ level */
1025 bridge
->irq_set(bridge
, level
, 1, 1);
1027 mutex_unlock(&bridge
->irq_mtx
);
1031 EXPORT_SYMBOL(vme_irq_request
);
1033 void vme_irq_free(struct device
*dev
, int level
, int statid
)
1035 struct vme_bridge
*bridge
;
1037 bridge
= dev_to_bridge(dev
);
1038 if (bridge
== NULL
) {
1039 printk(KERN_ERR
"Can't find VME bus\n");
1043 if ((level
< 1) || (level
> 7)) {
1044 printk(KERN_ERR
"Invalid interrupt level\n");
1048 if (bridge
->irq_set
== NULL
) {
1049 printk(KERN_ERR
"Configuring interrupts not supported\n");
1053 mutex_lock(&bridge
->irq_mtx
);
1055 bridge
->irq
[level
- 1].count
--;
1057 /* Disable IRQ level if no more interrupts attached at this level*/
1058 if (bridge
->irq
[level
- 1].count
== 0)
1059 bridge
->irq_set(bridge
, level
, 0, 1);
1061 bridge
->irq
[level
- 1].callback
[statid
].func
= NULL
;
1062 bridge
->irq
[level
- 1].callback
[statid
].priv_data
= NULL
;
1064 mutex_unlock(&bridge
->irq_mtx
);
1066 EXPORT_SYMBOL(vme_irq_free
);
1068 int vme_irq_generate(struct device
*dev
, int level
, int statid
)
1070 struct vme_bridge
*bridge
;
1072 bridge
= dev_to_bridge(dev
);
1073 if (bridge
== NULL
) {
1074 printk(KERN_ERR
"Can't find VME bus\n");
1078 if ((level
< 1) || (level
> 7)) {
1079 printk(KERN_WARNING
"Invalid interrupt level\n");
1083 if (bridge
->irq_generate
== NULL
) {
1084 printk(KERN_WARNING
"Interrupt generation not supported\n");
1088 return bridge
->irq_generate(bridge
, level
, statid
);
1090 EXPORT_SYMBOL(vme_irq_generate
);
1093 * Request the location monitor, return resource or NULL
1095 struct vme_resource
*vme_lm_request(struct device
*dev
)
1097 struct vme_bridge
*bridge
;
1098 struct list_head
*lm_pos
= NULL
;
1099 struct vme_lm_resource
*allocated_lm
= NULL
;
1100 struct vme_lm_resource
*lm
= NULL
;
1101 struct vme_resource
*resource
= NULL
;
1103 bridge
= dev_to_bridge(dev
);
1104 if (bridge
== NULL
) {
1105 printk(KERN_ERR
"Can't find VME bus\n");
1109 /* Loop through DMA resources */
1110 list_for_each(lm_pos
, &bridge
->lm_resources
) {
1111 lm
= list_entry(lm_pos
,
1112 struct vme_lm_resource
, list
);
1115 printk(KERN_ERR
"Registered NULL Location Monitor "
1120 /* Find an unlocked controller */
1121 mutex_lock(&lm
->mtx
);
1122 if (lm
->locked
== 0) {
1124 mutex_unlock(&lm
->mtx
);
1128 mutex_unlock(&lm
->mtx
);
1131 /* Check to see if we found a resource */
1132 if (allocated_lm
== NULL
)
1135 resource
= kmalloc(sizeof(struct vme_resource
), GFP_KERNEL
);
1136 if (resource
== NULL
) {
1137 printk(KERN_ERR
"Unable to allocate resource structure\n");
1140 resource
->type
= VME_LM
;
1141 resource
->entry
= &allocated_lm
->list
;
1147 mutex_lock(&lm
->mtx
);
1149 mutex_unlock(&lm
->mtx
);
1154 EXPORT_SYMBOL(vme_lm_request
);
1156 int vme_lm_count(struct vme_resource
*resource
)
1158 struct vme_lm_resource
*lm
;
1160 if (resource
->type
!= VME_LM
) {
1161 printk(KERN_ERR
"Not a Location Monitor resource\n");
1165 lm
= list_entry(resource
->entry
, struct vme_lm_resource
, list
);
1167 return lm
->monitors
;
1169 EXPORT_SYMBOL(vme_lm_count
);
1171 int vme_lm_set(struct vme_resource
*resource
, unsigned long long lm_base
,
1172 vme_address_t aspace
, vme_cycle_t cycle
)
1174 struct vme_bridge
*bridge
= find_bridge(resource
);
1175 struct vme_lm_resource
*lm
;
1177 if (resource
->type
!= VME_LM
) {
1178 printk(KERN_ERR
"Not a Location Monitor resource\n");
1182 lm
= list_entry(resource
->entry
, struct vme_lm_resource
, list
);
1184 if (bridge
->lm_set
== NULL
) {
1185 printk(KERN_ERR
"vme_lm_set not supported\n");
1189 return bridge
->lm_set(lm
, lm_base
, aspace
, cycle
);
1191 EXPORT_SYMBOL(vme_lm_set
);
1193 int vme_lm_get(struct vme_resource
*resource
, unsigned long long *lm_base
,
1194 vme_address_t
*aspace
, vme_cycle_t
*cycle
)
1196 struct vme_bridge
*bridge
= find_bridge(resource
);
1197 struct vme_lm_resource
*lm
;
1199 if (resource
->type
!= VME_LM
) {
1200 printk(KERN_ERR
"Not a Location Monitor resource\n");
1204 lm
= list_entry(resource
->entry
, struct vme_lm_resource
, list
);
1206 if (bridge
->lm_get
== NULL
) {
1207 printk(KERN_ERR
"vme_lm_get not supported\n");
1211 return bridge
->lm_get(lm
, lm_base
, aspace
, cycle
);
1213 EXPORT_SYMBOL(vme_lm_get
);
1215 int vme_lm_attach(struct vme_resource
*resource
, int monitor
,
1216 void (*callback
)(int))
1218 struct vme_bridge
*bridge
= find_bridge(resource
);
1219 struct vme_lm_resource
*lm
;
1221 if (resource
->type
!= VME_LM
) {
1222 printk(KERN_ERR
"Not a Location Monitor resource\n");
1226 lm
= list_entry(resource
->entry
, struct vme_lm_resource
, list
);
1228 if (bridge
->lm_attach
== NULL
) {
1229 printk(KERN_ERR
"vme_lm_attach not supported\n");
1233 return bridge
->lm_attach(lm
, monitor
, callback
);
1235 EXPORT_SYMBOL(vme_lm_attach
);
1237 int vme_lm_detach(struct vme_resource
*resource
, int monitor
)
1239 struct vme_bridge
*bridge
= find_bridge(resource
);
1240 struct vme_lm_resource
*lm
;
1242 if (resource
->type
!= VME_LM
) {
1243 printk(KERN_ERR
"Not a Location Monitor resource\n");
1247 lm
= list_entry(resource
->entry
, struct vme_lm_resource
, list
);
1249 if (bridge
->lm_detach
== NULL
) {
1250 printk(KERN_ERR
"vme_lm_detach not supported\n");
1254 return bridge
->lm_detach(lm
, monitor
);
1256 EXPORT_SYMBOL(vme_lm_detach
);
1258 void vme_lm_free(struct vme_resource
*resource
)
1260 struct vme_lm_resource
*lm
;
1262 if (resource
->type
!= VME_LM
) {
1263 printk(KERN_ERR
"Not a Location Monitor resource\n");
1267 lm
= list_entry(resource
->entry
, struct vme_lm_resource
, list
);
1269 mutex_lock(&lm
->mtx
);
1272 * Check to see that there aren't any callbacks still attached, if
1273 * there are we should probably be detaching them!
1278 mutex_unlock(&lm
->mtx
);
1282 EXPORT_SYMBOL(vme_lm_free
);
1284 int vme_slot_get(struct device
*bus
)
1286 struct vme_bridge
*bridge
;
1288 bridge
= dev_to_bridge(bus
);
1289 if (bridge
== NULL
) {
1290 printk(KERN_ERR
"Can't find VME bus\n");
1294 if (bridge
->slot_get
== NULL
) {
1295 printk(KERN_WARNING
"vme_slot_get not supported\n");
1299 return bridge
->slot_get(bridge
);
1301 EXPORT_SYMBOL(vme_slot_get
);
1304 /* - Bridge Registration --------------------------------------------------- */
1306 static int vme_alloc_bus_num(void)
1310 mutex_lock(&vme_bus_num_mtx
);
1311 for (i
= 0; i
< sizeof(vme_bus_numbers
) * 8; i
++) {
1312 if (((vme_bus_numbers
>> i
) & 0x1) == 0) {
1313 vme_bus_numbers
|= (0x1 << i
);
1317 mutex_unlock(&vme_bus_num_mtx
);
1322 static void vme_free_bus_num(int bus
)
1324 mutex_lock(&vme_bus_num_mtx
);
1325 vme_bus_numbers
&= ~(0x1 << bus
);
1326 mutex_unlock(&vme_bus_num_mtx
);
1329 int vme_register_bridge(struct vme_bridge
*bridge
)
1335 bridge
->num
= vme_alloc_bus_num();
1337 /* This creates 32 vme "slot" devices. This equates to a slot for each
1338 * ID available in a system conforming to the ANSI/VITA 1-1994
1341 for (i
= 0; i
< VME_SLOTS_MAX
; i
++) {
1342 dev
= &bridge
->dev
[i
];
1343 memset(dev
, 0, sizeof(struct device
));
1345 dev
->parent
= bridge
->parent
;
1346 dev
->bus
= &vme_bus_type
;
1348 * We save a pointer to the bridge in platform_data so that we
1349 * can get to it later. We keep driver_data for use by the
1350 * driver that binds against the slot
1352 dev
->platform_data
= bridge
;
1353 dev_set_name(dev
, "vme-%x.%x", bridge
->num
, i
+ 1);
1355 retval
= device_register(dev
);
1364 dev
= &bridge
->dev
[i
];
1365 device_unregister(dev
);
1367 vme_free_bus_num(bridge
->num
);
1370 EXPORT_SYMBOL(vme_register_bridge
);
1372 void vme_unregister_bridge(struct vme_bridge
*bridge
)
1378 for (i
= 0; i
< VME_SLOTS_MAX
; i
++) {
1379 dev
= &bridge
->dev
[i
];
1380 device_unregister(dev
);
1382 vme_free_bus_num(bridge
->num
);
1384 EXPORT_SYMBOL(vme_unregister_bridge
);
1387 /* - Driver Registration --------------------------------------------------- */
1389 int vme_register_driver(struct vme_driver
*drv
)
1391 drv
->driver
.name
= drv
->name
;
1392 drv
->driver
.bus
= &vme_bus_type
;
1394 return driver_register(&drv
->driver
);
1396 EXPORT_SYMBOL(vme_register_driver
);
1398 void vme_unregister_driver(struct vme_driver
*drv
)
1400 driver_unregister(&drv
->driver
);
1402 EXPORT_SYMBOL(vme_unregister_driver
);
1404 /* - Bus Registration ------------------------------------------------------ */
1406 static int vme_calc_slot(struct device
*dev
)
1408 struct vme_bridge
*bridge
;
1411 bridge
= dev_to_bridge(dev
);
1413 /* Determine slot number */
1415 while (num
< VME_SLOTS_MAX
) {
1416 if (&bridge
->dev
[num
] == dev
)
1421 if (num
== VME_SLOTS_MAX
) {
1422 dev_err(dev
, "Failed to identify slot\n");
1432 static struct vme_driver
*dev_to_vme_driver(struct device
*dev
)
1434 if (dev
->driver
== NULL
)
1435 printk(KERN_ERR
"Bugger dev->driver is NULL\n");
1437 return container_of(dev
->driver
, struct vme_driver
, driver
);
1440 static int vme_bus_match(struct device
*dev
, struct device_driver
*drv
)
1442 struct vme_bridge
*bridge
;
1443 struct vme_driver
*driver
;
1446 bridge
= dev_to_bridge(dev
);
1447 driver
= container_of(drv
, struct vme_driver
, driver
);
1449 num
= vme_calc_slot(dev
);
1453 if (driver
->bind_table
== NULL
) {
1454 dev_err(dev
, "Bind table NULL\n");
1459 while ((driver
->bind_table
[i
].bus
!= 0) ||
1460 (driver
->bind_table
[i
].slot
!= 0)) {
1462 if (bridge
->num
== driver
->bind_table
[i
].bus
) {
1463 if (num
== driver
->bind_table
[i
].slot
)
1466 if (driver
->bind_table
[i
].slot
== VME_SLOT_ALL
)
1469 if ((driver
->bind_table
[i
].slot
== VME_SLOT_CURRENT
) &&
1470 (num
== vme_slot_get(dev
)))
1481 static int vme_bus_probe(struct device
*dev
)
1483 struct vme_bridge
*bridge
;
1484 struct vme_driver
*driver
;
1485 int retval
= -ENODEV
;
1487 driver
= dev_to_vme_driver(dev
);
1488 bridge
= dev_to_bridge(dev
);
1490 if (driver
->probe
!= NULL
)
1491 retval
= driver
->probe(dev
, bridge
->num
, vme_calc_slot(dev
));
1496 static int vme_bus_remove(struct device
*dev
)
1498 struct vme_bridge
*bridge
;
1499 struct vme_driver
*driver
;
1500 int retval
= -ENODEV
;
1502 driver
= dev_to_vme_driver(dev
);
1503 bridge
= dev_to_bridge(dev
);
1505 if (driver
->remove
!= NULL
)
1506 retval
= driver
->remove(dev
, bridge
->num
, vme_calc_slot(dev
));
1511 struct bus_type vme_bus_type
= {
1513 .match
= vme_bus_match
,
1514 .probe
= vme_bus_probe
,
1515 .remove
= vme_bus_remove
,
1517 EXPORT_SYMBOL(vme_bus_type
);
1519 static int __init
vme_init(void)
1521 return bus_register(&vme_bus_type
);
1524 static void __exit
vme_exit(void)
1526 bus_unregister(&vme_bus_type
);
1529 MODULE_DESCRIPTION("VME bridge driver framework");
1530 MODULE_AUTHOR("Martyn Welch <martyn.welch@ge.com");
1531 MODULE_LICENSE("GPL");
1533 module_init(vme_init
);
1534 module_exit(vme_exit
);