1 // SPDX-License-Identifier: GPL-2.0
2 #include <linux/device.h>
7 * On the state of PCI's devres implementation:
9 * The older devres API for PCI has two significant problems:
11 * 1. It is very strongly tied to the statically allocated mapping table in
12 * struct pcim_iomap_devres below. This is mostly solved in the sense of the
13 * pcim_ functions in this file providing things like ranged mapping by
14 * bypassing this table, whereas the functions that were present in the old
15 * API still enter the mapping addresses into the table for users of the old
18 * 2. The region-request-functions in pci.c do become managed IF the device has
19 * been enabled with pcim_enable_device() instead of pci_enable_device().
20 * This resulted in the API becoming inconsistent: Some functions have an
21 * obviously managed counter-part (e.g., pci_iomap() <-> pcim_iomap()),
22 * whereas some don't and are never managed, while others don't and are
23 * _sometimes_ managed (e.g. pci_request_region()).
25 * Consequently, in the new API, region requests performed by the pcim_
26 * functions are automatically cleaned up through the devres callback
27 * pcim_addr_resource_release().
29 * Users of pcim_enable_device() + pci_*region*() are redirected in
30 * pci.c to the managed functions here in this file. This isn't exactly
31 * perfect, but the only alternative way would be to port ALL drivers
32 * using said combination to pcim_ functions.
35 * Remove the legacy table entirely once all calls to pcim_iomap_table() in
36 * the kernel have been removed.
40 * Legacy struct storing addresses to whole mapped BARs.
42 struct pcim_iomap_devres
{
43 void __iomem
*table
[PCI_STD_NUM_BARS
];
46 /* Used to restore the old INTx state on driver detach. */
47 struct pcim_intx_devres
{
51 enum pcim_addr_devres_type
{
52 /* Default initializer. */
53 PCIM_ADDR_DEVRES_TYPE_INVALID
,
55 /* A requested region spanning an entire BAR. */
56 PCIM_ADDR_DEVRES_TYPE_REGION
,
59 * A requested region spanning an entire BAR, and a mapping for
62 PCIM_ADDR_DEVRES_TYPE_REGION_MAPPING
,
65 * A mapping within a BAR, either spanning the whole BAR or just a
66 * range. Without a requested region.
68 PCIM_ADDR_DEVRES_TYPE_MAPPING
,
72 * This struct envelops IO or MEM addresses, i.e., mappings and region
73 * requests, because those are very frequently requested and released
76 struct pcim_addr_devres
{
77 enum pcim_addr_devres_type type
;
78 void __iomem
*baseaddr
;
84 static inline void pcim_addr_devres_clear(struct pcim_addr_devres
*res
)
86 memset(res
, 0, sizeof(*res
));
91 * The following functions, __pcim_*_region*, exist as counterparts to the
92 * versions from pci.c - which, unfortunately, can be in "hybrid mode", i.e.,
93 * sometimes managed, sometimes not.
95 * To separate the APIs cleanly, we define our own, simplified versions here.
99 * __pcim_request_region_range - Request a ranged region
100 * @pdev: PCI device the region belongs to
101 * @bar: BAR the range is within
102 * @offset: offset from the BAR's start address
103 * @maxlen: length in bytes, beginning at @offset
104 * @name: name associated with the request
105 * @req_flags: flags for the request, e.g., for kernel-exclusive requests
107 * Returns: 0 on success, a negative error code on failure.
109 * Request a range within a device's PCI BAR. Sanity check the input.
111 static int __pcim_request_region_range(struct pci_dev
*pdev
, int bar
,
112 unsigned long offset
,
113 unsigned long maxlen
,
114 const char *name
, int req_flags
)
116 resource_size_t start
= pci_resource_start(pdev
, bar
);
117 resource_size_t len
= pci_resource_len(pdev
, bar
);
118 unsigned long dev_flags
= pci_resource_flags(pdev
, bar
);
120 if (start
== 0 || len
== 0) /* Unused BAR. */
128 if (len
> maxlen
&& maxlen
!= 0)
131 if (dev_flags
& IORESOURCE_IO
) {
132 if (!request_region(start
, len
, name
))
134 } else if (dev_flags
& IORESOURCE_MEM
) {
135 if (!__request_mem_region(start
, len
, name
, req_flags
))
138 /* That's not a device we can request anything on. */
145 static void __pcim_release_region_range(struct pci_dev
*pdev
, int bar
,
146 unsigned long offset
,
147 unsigned long maxlen
)
149 resource_size_t start
= pci_resource_start(pdev
, bar
);
150 resource_size_t len
= pci_resource_len(pdev
, bar
);
151 unsigned long flags
= pci_resource_flags(pdev
, bar
);
153 if (len
<= offset
|| start
== 0)
156 if (len
== 0 || maxlen
== 0) /* This an unused BAR. Do nothing. */
165 if (flags
& IORESOURCE_IO
)
166 release_region(start
, len
);
167 else if (flags
& IORESOURCE_MEM
)
168 release_mem_region(start
, len
);
171 static int __pcim_request_region(struct pci_dev
*pdev
, int bar
,
172 const char *name
, int flags
)
174 unsigned long offset
= 0;
175 unsigned long len
= pci_resource_len(pdev
, bar
);
177 return __pcim_request_region_range(pdev
, bar
, offset
, len
, name
, flags
);
180 static void __pcim_release_region(struct pci_dev
*pdev
, int bar
)
182 unsigned long offset
= 0;
183 unsigned long len
= pci_resource_len(pdev
, bar
);
185 __pcim_release_region_range(pdev
, bar
, offset
, len
);
188 static void pcim_addr_resource_release(struct device
*dev
, void *resource_raw
)
190 struct pci_dev
*pdev
= to_pci_dev(dev
);
191 struct pcim_addr_devres
*res
= resource_raw
;
194 case PCIM_ADDR_DEVRES_TYPE_REGION
:
195 __pcim_release_region(pdev
, res
->bar
);
197 case PCIM_ADDR_DEVRES_TYPE_REGION_MAPPING
:
198 pci_iounmap(pdev
, res
->baseaddr
);
199 __pcim_release_region(pdev
, res
->bar
);
201 case PCIM_ADDR_DEVRES_TYPE_MAPPING
:
202 pci_iounmap(pdev
, res
->baseaddr
);
209 static struct pcim_addr_devres
*pcim_addr_devres_alloc(struct pci_dev
*pdev
)
211 struct pcim_addr_devres
*res
;
213 res
= devres_alloc_node(pcim_addr_resource_release
, sizeof(*res
),
214 GFP_KERNEL
, dev_to_node(&pdev
->dev
));
216 pcim_addr_devres_clear(res
);
220 /* Just for consistency and readability. */
221 static inline void pcim_addr_devres_free(struct pcim_addr_devres
*res
)
227 * Used by devres to identify a pcim_addr_devres.
229 static int pcim_addr_resources_match(struct device
*dev
,
230 void *a_raw
, void *b_raw
)
232 struct pcim_addr_devres
*a
, *b
;
237 if (a
->type
!= b
->type
)
241 case PCIM_ADDR_DEVRES_TYPE_REGION
:
242 case PCIM_ADDR_DEVRES_TYPE_REGION_MAPPING
:
243 return a
->bar
== b
->bar
;
244 case PCIM_ADDR_DEVRES_TYPE_MAPPING
:
245 return a
->baseaddr
== b
->baseaddr
;
251 static void devm_pci_unmap_iospace(struct device
*dev
, void *ptr
)
253 struct resource
**res
= ptr
;
255 pci_unmap_iospace(*res
);
259 * devm_pci_remap_iospace - Managed pci_remap_iospace()
260 * @dev: Generic device to remap IO address for
261 * @res: Resource describing the I/O space
262 * @phys_addr: physical address of range to be mapped
264 * Managed pci_remap_iospace(). Map is automatically unmapped on driver
267 int devm_pci_remap_iospace(struct device
*dev
, const struct resource
*res
,
268 phys_addr_t phys_addr
)
270 const struct resource
**ptr
;
273 ptr
= devres_alloc(devm_pci_unmap_iospace
, sizeof(*ptr
), GFP_KERNEL
);
277 error
= pci_remap_iospace(res
, phys_addr
);
282 devres_add(dev
, ptr
);
287 EXPORT_SYMBOL(devm_pci_remap_iospace
);
290 * devm_pci_remap_cfgspace - Managed pci_remap_cfgspace()
291 * @dev: Generic device to remap IO address for
292 * @offset: Resource address to map
295 * Managed pci_remap_cfgspace(). Map is automatically unmapped on driver
298 void __iomem
*devm_pci_remap_cfgspace(struct device
*dev
,
299 resource_size_t offset
,
300 resource_size_t size
)
302 void __iomem
**ptr
, *addr
;
304 ptr
= devres_alloc(devm_ioremap_release
, sizeof(*ptr
), GFP_KERNEL
);
308 addr
= pci_remap_cfgspace(offset
, size
);
311 devres_add(dev
, ptr
);
317 EXPORT_SYMBOL(devm_pci_remap_cfgspace
);
320 * devm_pci_remap_cfg_resource - check, request region and ioremap cfg resource
321 * @dev: generic device to handle the resource for
322 * @res: configuration space resource to be handled
324 * Checks that a resource is a valid memory region, requests the memory
325 * region and ioremaps with pci_remap_cfgspace() API that ensures the
326 * proper PCI configuration space memory attributes are guaranteed.
328 * All operations are managed and will be undone on driver detach.
330 * Returns a pointer to the remapped memory or an IOMEM_ERR_PTR() encoded error
331 * code on failure. Usage example::
333 * res = platform_get_resource(pdev, IORESOURCE_MEM, 0);
334 * base = devm_pci_remap_cfg_resource(&pdev->dev, res);
336 * return PTR_ERR(base);
338 void __iomem
*devm_pci_remap_cfg_resource(struct device
*dev
,
339 struct resource
*res
)
341 resource_size_t size
;
343 void __iomem
*dest_ptr
;
347 if (!res
|| resource_type(res
) != IORESOURCE_MEM
) {
348 dev_err(dev
, "invalid resource\n");
349 return IOMEM_ERR_PTR(-EINVAL
);
352 size
= resource_size(res
);
355 name
= devm_kasprintf(dev
, GFP_KERNEL
, "%s %s", dev_name(dev
),
358 name
= devm_kstrdup(dev
, dev_name(dev
), GFP_KERNEL
);
360 return IOMEM_ERR_PTR(-ENOMEM
);
362 if (!devm_request_mem_region(dev
, res
->start
, size
, name
)) {
363 dev_err(dev
, "can't request region for resource %pR\n", res
);
364 return IOMEM_ERR_PTR(-EBUSY
);
367 dest_ptr
= devm_pci_remap_cfgspace(dev
, res
->start
, size
);
369 dev_err(dev
, "ioremap failed for resource %pR\n", res
);
370 devm_release_mem_region(dev
, res
->start
, size
);
371 dest_ptr
= IOMEM_ERR_PTR(-ENOMEM
);
376 EXPORT_SYMBOL(devm_pci_remap_cfg_resource
);
378 static void __pcim_clear_mwi(void *pdev_raw
)
380 struct pci_dev
*pdev
= pdev_raw
;
386 * pcim_set_mwi - a device-managed pci_set_mwi()
387 * @pdev: the PCI device for which MWI is enabled
389 * Managed pci_set_mwi().
391 * RETURNS: An appropriate -ERRNO error value on error, or zero for success.
393 int pcim_set_mwi(struct pci_dev
*pdev
)
397 ret
= devm_add_action(&pdev
->dev
, __pcim_clear_mwi
, pdev
);
401 ret
= pci_set_mwi(pdev
);
403 devm_remove_action(&pdev
->dev
, __pcim_clear_mwi
, pdev
);
407 EXPORT_SYMBOL(pcim_set_mwi
);
409 static inline bool mask_contains_bar(int mask
, int bar
)
411 return mask
& BIT(bar
);
415 * This is a copy of pci_intx() used to bypass the problem of recursive
416 * function calls due to the hybrid nature of pci_intx().
418 static void __pcim_intx(struct pci_dev
*pdev
, int enable
)
420 u16 pci_command
, new;
422 pci_read_config_word(pdev
, PCI_COMMAND
, &pci_command
);
425 new = pci_command
& ~PCI_COMMAND_INTX_DISABLE
;
427 new = pci_command
| PCI_COMMAND_INTX_DISABLE
;
429 if (new != pci_command
)
430 pci_write_config_word(pdev
, PCI_COMMAND
, new);
433 static void pcim_intx_restore(struct device
*dev
, void *data
)
435 struct pci_dev
*pdev
= to_pci_dev(dev
);
436 struct pcim_intx_devres
*res
= data
;
438 __pcim_intx(pdev
, res
->orig_intx
);
441 static struct pcim_intx_devres
*get_or_create_intx_devres(struct device
*dev
)
443 struct pcim_intx_devres
*res
;
445 res
= devres_find(dev
, pcim_intx_restore
, NULL
, NULL
);
449 res
= devres_alloc(pcim_intx_restore
, sizeof(*res
), GFP_KERNEL
);
451 devres_add(dev
, res
);
457 * pcim_intx - managed pci_intx()
458 * @pdev: the PCI device to operate on
459 * @enable: boolean: whether to enable or disable PCI INTx
461 * Returns: 0 on success, -ENOMEM on error.
463 * Enable/disable PCI INTx for device @pdev.
464 * Restore the original state on driver detach.
466 int pcim_intx(struct pci_dev
*pdev
, int enable
)
468 struct pcim_intx_devres
*res
;
470 res
= get_or_create_intx_devres(&pdev
->dev
);
474 res
->orig_intx
= !enable
;
475 __pcim_intx(pdev
, enable
);
480 static void pcim_disable_device(void *pdev_raw
)
482 struct pci_dev
*pdev
= pdev_raw
;
485 pci_disable_device(pdev
);
487 pdev
->is_managed
= false;
491 * pcim_enable_device - Managed pci_enable_device()
492 * @pdev: PCI device to be initialized
494 * Returns: 0 on success, negative error code on failure.
496 * Managed pci_enable_device(). Device will automatically be disabled on
499 int pcim_enable_device(struct pci_dev
*pdev
)
503 ret
= devm_add_action(&pdev
->dev
, pcim_disable_device
, pdev
);
508 * We prefer removing the action in case of an error over
509 * devm_add_action_or_reset() because the latter could theoretically be
510 * disturbed by users having pinned the device too soon.
512 ret
= pci_enable_device(pdev
);
514 devm_remove_action(&pdev
->dev
, pcim_disable_device
, pdev
);
518 pdev
->is_managed
= true;
522 EXPORT_SYMBOL(pcim_enable_device
);
525 * pcim_pin_device - Pin managed PCI device
526 * @pdev: PCI device to pin
528 * Pin managed PCI device @pdev. Pinned device won't be disabled on driver
529 * detach. @pdev must have been enabled with pcim_enable_device().
531 void pcim_pin_device(struct pci_dev
*pdev
)
535 EXPORT_SYMBOL(pcim_pin_device
);
537 static void pcim_iomap_release(struct device
*gendev
, void *res
)
540 * Do nothing. This is legacy code.
542 * Cleanup of the mappings is now done directly through the callbacks
543 * registered when creating them.
548 * pcim_iomap_table - access iomap allocation table (DEPRECATED)
549 * @pdev: PCI device to access iomap table for
552 * Const pointer to array of __iomem pointers on success, NULL on failure.
554 * Access iomap allocation table for @dev. If iomap table doesn't
555 * exist and @pdev is managed, it will be allocated. All iomaps
556 * recorded in the iomap table are automatically unmapped on driver
559 * This function might sleep when the table is first allocated but can
560 * be safely called without context and guaranteed to succeed once
563 * This function is DEPRECATED. Do not use it in new code. Instead, obtain a
564 * mapping's address directly from one of the pcim_* mapping functions. For
566 * void __iomem \*mappy = pcim_iomap(pdev, bar, length);
568 void __iomem
* const *pcim_iomap_table(struct pci_dev
*pdev
)
570 struct pcim_iomap_devres
*dr
, *new_dr
;
572 dr
= devres_find(&pdev
->dev
, pcim_iomap_release
, NULL
, NULL
);
576 new_dr
= devres_alloc_node(pcim_iomap_release
, sizeof(*new_dr
), GFP_KERNEL
,
577 dev_to_node(&pdev
->dev
));
580 dr
= devres_get(&pdev
->dev
, new_dr
, NULL
, NULL
);
583 EXPORT_SYMBOL(pcim_iomap_table
);
586 * Fill the legacy mapping-table, so that drivers using the old API can
587 * still get a BAR's mapping address through pcim_iomap_table().
589 static int pcim_add_mapping_to_legacy_table(struct pci_dev
*pdev
,
590 void __iomem
*mapping
, int bar
)
592 void __iomem
**legacy_iomap_table
;
594 if (bar
>= PCI_STD_NUM_BARS
)
597 legacy_iomap_table
= (void __iomem
**)pcim_iomap_table(pdev
);
598 if (!legacy_iomap_table
)
601 /* The legacy mechanism doesn't allow for duplicate mappings. */
602 WARN_ON(legacy_iomap_table
[bar
]);
604 legacy_iomap_table
[bar
] = mapping
;
610 * Remove a mapping. The table only contains whole-BAR mappings, so this will
611 * never interfere with ranged mappings.
613 static void pcim_remove_mapping_from_legacy_table(struct pci_dev
*pdev
,
617 void __iomem
**legacy_iomap_table
;
619 legacy_iomap_table
= (void __iomem
**)pcim_iomap_table(pdev
);
620 if (!legacy_iomap_table
)
623 for (bar
= 0; bar
< PCI_STD_NUM_BARS
; bar
++) {
624 if (legacy_iomap_table
[bar
] == addr
) {
625 legacy_iomap_table
[bar
] = NULL
;
632 * The same as pcim_remove_mapping_from_legacy_table(), but identifies the
633 * mapping by its BAR index.
635 static void pcim_remove_bar_from_legacy_table(struct pci_dev
*pdev
, int bar
)
637 void __iomem
**legacy_iomap_table
;
639 if (bar
>= PCI_STD_NUM_BARS
)
642 legacy_iomap_table
= (void __iomem
**)pcim_iomap_table(pdev
);
643 if (!legacy_iomap_table
)
646 legacy_iomap_table
[bar
] = NULL
;
650 * pcim_iomap - Managed pcim_iomap()
651 * @pdev: PCI device to iomap for
653 * @maxlen: Maximum length of iomap
655 * Returns: __iomem pointer on success, NULL on failure.
657 * Managed pci_iomap(). Map is automatically unmapped on driver detach. If
658 * desired, unmap manually only with pcim_iounmap().
660 * This SHOULD only be used once per BAR.
663 * Contrary to the other pcim_* functions, this function does not return an
664 * IOMEM_ERR_PTR() on failure, but a simple NULL. This is done for backwards
667 void __iomem
*pcim_iomap(struct pci_dev
*pdev
, int bar
, unsigned long maxlen
)
669 void __iomem
*mapping
;
670 struct pcim_addr_devres
*res
;
672 res
= pcim_addr_devres_alloc(pdev
);
675 res
->type
= PCIM_ADDR_DEVRES_TYPE_MAPPING
;
677 mapping
= pci_iomap(pdev
, bar
, maxlen
);
680 res
->baseaddr
= mapping
;
682 if (pcim_add_mapping_to_legacy_table(pdev
, mapping
, bar
) != 0)
685 devres_add(&pdev
->dev
, res
);
689 pci_iounmap(pdev
, mapping
);
691 pcim_addr_devres_free(res
);
694 EXPORT_SYMBOL(pcim_iomap
);
697 * pcim_iounmap - Managed pci_iounmap()
698 * @pdev: PCI device to iounmap for
699 * @addr: Address to unmap
701 * Managed pci_iounmap(). @addr must have been mapped using a pcim_* mapping
704 void pcim_iounmap(struct pci_dev
*pdev
, void __iomem
*addr
)
706 struct pcim_addr_devres res_searched
;
708 pcim_addr_devres_clear(&res_searched
);
709 res_searched
.type
= PCIM_ADDR_DEVRES_TYPE_MAPPING
;
710 res_searched
.baseaddr
= addr
;
712 if (devres_release(&pdev
->dev
, pcim_addr_resource_release
,
713 pcim_addr_resources_match
, &res_searched
) != 0) {
714 /* Doesn't exist. User passed nonsense. */
718 pcim_remove_mapping_from_legacy_table(pdev
, addr
);
720 EXPORT_SYMBOL(pcim_iounmap
);
723 * pcim_iomap_region - Request and iomap a PCI BAR
724 * @pdev: PCI device to map IO resources for
725 * @bar: Index of a BAR to map
726 * @name: Name associated with the request
728 * Returns: __iomem pointer on success, an IOMEM_ERR_PTR on failure.
730 * Mapping and region will get automatically released on driver detach. If
731 * desired, release manually only with pcim_iounmap_region().
733 void __iomem
*pcim_iomap_region(struct pci_dev
*pdev
, int bar
,
737 struct pcim_addr_devres
*res
;
739 res
= pcim_addr_devres_alloc(pdev
);
741 return IOMEM_ERR_PTR(-ENOMEM
);
743 res
->type
= PCIM_ADDR_DEVRES_TYPE_REGION_MAPPING
;
746 ret
= __pcim_request_region(pdev
, bar
, name
, 0);
750 res
->baseaddr
= pci_iomap(pdev
, bar
, 0);
751 if (!res
->baseaddr
) {
756 devres_add(&pdev
->dev
, res
);
757 return res
->baseaddr
;
760 __pcim_release_region(pdev
, bar
);
762 pcim_addr_devres_free(res
);
764 return IOMEM_ERR_PTR(ret
);
766 EXPORT_SYMBOL(pcim_iomap_region
);
769 * pcim_iounmap_region - Unmap and release a PCI BAR
770 * @pdev: PCI device to operate on
771 * @bar: Index of BAR to unmap and release
773 * Unmap a BAR and release its region manually. Only pass BARs that were
774 * previously mapped by pcim_iomap_region().
776 void pcim_iounmap_region(struct pci_dev
*pdev
, int bar
)
778 struct pcim_addr_devres res_searched
;
780 pcim_addr_devres_clear(&res_searched
);
781 res_searched
.type
= PCIM_ADDR_DEVRES_TYPE_REGION_MAPPING
;
782 res_searched
.bar
= bar
;
784 devres_release(&pdev
->dev
, pcim_addr_resource_release
,
785 pcim_addr_resources_match
, &res_searched
);
787 EXPORT_SYMBOL(pcim_iounmap_region
);
790 * pcim_iomap_regions - Request and iomap PCI BARs (DEPRECATED)
791 * @pdev: PCI device to map IO resources for
792 * @mask: Mask of BARs to request and iomap
793 * @name: Name associated with the requests
795 * Returns: 0 on success, negative error code on failure.
797 * Request and iomap regions specified by @mask.
799 * This function is DEPRECATED. Do not use it in new code.
800 * Use pcim_iomap_region() instead.
802 int pcim_iomap_regions(struct pci_dev
*pdev
, int mask
, const char *name
)
806 void __iomem
*mapping
;
808 for (bar
= 0; bar
< DEVICE_COUNT_RESOURCE
; bar
++) {
809 if (!mask_contains_bar(mask
, bar
))
812 mapping
= pcim_iomap_region(pdev
, bar
, name
);
813 if (IS_ERR(mapping
)) {
814 ret
= PTR_ERR(mapping
);
817 ret
= pcim_add_mapping_to_legacy_table(pdev
, mapping
, bar
);
826 pcim_iounmap_region(pdev
, bar
);
827 pcim_remove_bar_from_legacy_table(pdev
, bar
);
832 EXPORT_SYMBOL(pcim_iomap_regions
);
834 static int _pcim_request_region(struct pci_dev
*pdev
, int bar
, const char *name
,
838 struct pcim_addr_devres
*res
;
840 res
= pcim_addr_devres_alloc(pdev
);
843 res
->type
= PCIM_ADDR_DEVRES_TYPE_REGION
;
846 ret
= __pcim_request_region(pdev
, bar
, name
, request_flags
);
848 pcim_addr_devres_free(res
);
852 devres_add(&pdev
->dev
, res
);
857 * pcim_request_region - Request a PCI BAR
858 * @pdev: PCI device to requestion region for
859 * @bar: Index of BAR to request
860 * @name: Name associated with the request
862 * Returns: 0 on success, a negative error code on failure.
864 * Request region specified by @bar.
866 * The region will automatically be released on driver detach. If desired,
867 * release manually only with pcim_release_region().
869 int pcim_request_region(struct pci_dev
*pdev
, int bar
, const char *name
)
871 return _pcim_request_region(pdev
, bar
, name
, 0);
873 EXPORT_SYMBOL(pcim_request_region
);
876 * pcim_request_region_exclusive - Request a PCI BAR exclusively
877 * @pdev: PCI device to requestion region for
878 * @bar: Index of BAR to request
879 * @name: Name associated with the request
881 * Returns: 0 on success, a negative error code on failure.
883 * Request region specified by @bar exclusively.
885 * The region will automatically be released on driver detach. If desired,
886 * release manually only with pcim_release_region().
888 int pcim_request_region_exclusive(struct pci_dev
*pdev
, int bar
, const char *name
)
890 return _pcim_request_region(pdev
, bar
, name
, IORESOURCE_EXCLUSIVE
);
894 * pcim_release_region - Release a PCI BAR
895 * @pdev: PCI device to operate on
896 * @bar: Index of BAR to release
898 * Release a region manually that was previously requested by
899 * pcim_request_region().
901 void pcim_release_region(struct pci_dev
*pdev
, int bar
)
903 struct pcim_addr_devres res_searched
;
905 pcim_addr_devres_clear(&res_searched
);
906 res_searched
.type
= PCIM_ADDR_DEVRES_TYPE_REGION
;
907 res_searched
.bar
= bar
;
909 devres_release(&pdev
->dev
, pcim_addr_resource_release
,
910 pcim_addr_resources_match
, &res_searched
);
915 * pcim_release_all_regions - Release all regions of a PCI-device
916 * @pdev: the PCI device
918 * Release all regions previously requested through pcim_request_region()
919 * or pcim_request_all_regions().
921 * Can be called from any context, i.e., not necessarily as a counterpart to
922 * pcim_request_all_regions().
924 static void pcim_release_all_regions(struct pci_dev
*pdev
)
928 for (bar
= 0; bar
< PCI_STD_NUM_BARS
; bar
++)
929 pcim_release_region(pdev
, bar
);
933 * pcim_request_all_regions - Request all regions
934 * @pdev: PCI device to map IO resources for
935 * @name: name associated with the request
937 * Returns: 0 on success, negative error code on failure.
939 * Requested regions will automatically be released at driver detach. If
940 * desired, release individual regions with pcim_release_region() or all of
941 * them at once with pcim_release_all_regions().
943 int pcim_request_all_regions(struct pci_dev
*pdev
, const char *name
)
948 for (bar
= 0; bar
< PCI_STD_NUM_BARS
; bar
++) {
949 ret
= pcim_request_region(pdev
, bar
, name
);
957 pcim_release_all_regions(pdev
);
961 EXPORT_SYMBOL(pcim_request_all_regions
);
964 * pcim_iounmap_regions - Unmap and release PCI BARs (DEPRECATED)
965 * @pdev: PCI device to map IO resources for
966 * @mask: Mask of BARs to unmap and release
968 * Unmap and release regions specified by @mask.
970 * This function is DEPRECATED. Do not use it in new code.
971 * Use pcim_iounmap_region() instead.
973 void pcim_iounmap_regions(struct pci_dev
*pdev
, int mask
)
977 for (i
= 0; i
< PCI_STD_NUM_BARS
; i
++) {
978 if (!mask_contains_bar(mask
, i
))
981 pcim_iounmap_region(pdev
, i
);
982 pcim_remove_bar_from_legacy_table(pdev
, i
);
985 EXPORT_SYMBOL(pcim_iounmap_regions
);
988 * pcim_iomap_range - Create a ranged __iomap mapping within a PCI BAR
989 * @pdev: PCI device to map IO resources for
990 * @bar: Index of the BAR
991 * @offset: Offset from the begin of the BAR
992 * @len: Length in bytes for the mapping
994 * Returns: __iomem pointer on success, an IOMEM_ERR_PTR on failure.
996 * Creates a new IO-Mapping within the specified @bar, ranging from @offset to
999 * The mapping will automatically get unmapped on driver detach. If desired,
1000 * release manually only with pcim_iounmap().
1002 void __iomem
*pcim_iomap_range(struct pci_dev
*pdev
, int bar
,
1003 unsigned long offset
, unsigned long len
)
1005 void __iomem
*mapping
;
1006 struct pcim_addr_devres
*res
;
1008 res
= pcim_addr_devres_alloc(pdev
);
1010 return IOMEM_ERR_PTR(-ENOMEM
);
1012 mapping
= pci_iomap_range(pdev
, bar
, offset
, len
);
1014 pcim_addr_devres_free(res
);
1015 return IOMEM_ERR_PTR(-EINVAL
);
1018 res
->type
= PCIM_ADDR_DEVRES_TYPE_MAPPING
;
1019 res
->baseaddr
= mapping
;
1022 * Ranged mappings don't get added to the legacy-table, since the table
1023 * only ever keeps track of whole BARs.
1026 devres_add(&pdev
->dev
, res
);
1029 EXPORT_SYMBOL(pcim_iomap_range
);