2 * drivers/pci/setup-bus.c
4 * Extruded from code written by
5 * Dave Rusling (david.rusling@reo.mts.dec.com)
6 * David Mosberger (davidm@cs.arizona.edu)
7 * David Miller (davem@redhat.com)
9 * Support routines for initializing a PCI subsystem.
13 * Nov 2000, Ivan Kokshaysky <ink@jurassic.park.msu.ru>
14 * PCI-PCI bridges cleanup, sorted resource allocation.
15 * Feb 2002, Ivan Kokshaysky <ink@jurassic.park.msu.ru>
16 * Converted to allocation in 3 passes, which gives
17 * tighter packing. Prefetchable range support.
20 #include <linux/init.h>
21 #include <linux/kernel.h>
22 #include <linux/module.h>
23 #include <linux/pci.h>
24 #include <linux/errno.h>
25 #include <linux/ioport.h>
26 #include <linux/cache.h>
27 #include <linux/slab.h>
30 struct resource_list_x
{
31 struct resource_list_x
*next
;
34 resource_size_t start
;
39 static void add_to_failed_list(struct resource_list_x
*head
,
40 struct pci_dev
*dev
, struct resource
*res
)
42 struct resource_list_x
*list
= head
;
43 struct resource_list_x
*ln
= list
->next
;
44 struct resource_list_x
*tmp
;
46 tmp
= kmalloc(sizeof(*tmp
), GFP_KERNEL
);
48 pr_warning("add_to_failed_list: kmalloc() failed!\n");
55 tmp
->start
= res
->start
;
57 tmp
->flags
= res
->flags
;
61 static void free_failed_list(struct resource_list_x
*head
)
63 struct resource_list_x
*list
, *tmp
;
65 for (list
= head
->next
; list
;) {
74 static void __dev_sort_resources(struct pci_dev
*dev
,
75 struct resource_list
*head
)
77 u16
class = dev
->class >> 8;
79 /* Don't touch classless devices or host bridges or ioapics. */
80 if (class == PCI_CLASS_NOT_DEFINED
|| class == PCI_CLASS_BRIDGE_HOST
)
83 /* Don't touch ioapic devices already enabled by firmware */
84 if (class == PCI_CLASS_SYSTEM_PIC
) {
86 pci_read_config_word(dev
, PCI_COMMAND
, &command
);
87 if (command
& (PCI_COMMAND_IO
| PCI_COMMAND_MEMORY
))
91 pdev_sort_resources(dev
, head
);
94 static void __assign_resources_sorted(struct resource_list
*head
,
95 struct resource_list_x
*fail_head
)
98 struct resource_list
*list
, *tmp
;
101 for (list
= head
->next
; list
;) {
103 idx
= res
- &list
->dev
->resource
[0];
105 if (pci_assign_resource(list
->dev
, idx
)) {
106 if (fail_head
&& !pci_is_root_bus(list
->dev
->bus
)) {
108 * if the failed res is for ROM BAR, and it will
109 * be enabled later, don't add it to the list
111 if (!((idx
== PCI_ROM_RESOURCE
) &&
112 (!(res
->flags
& IORESOURCE_ROM_ENABLE
))))
113 add_to_failed_list(fail_head
, list
->dev
, res
);
125 static void pdev_assign_resources_sorted(struct pci_dev
*dev
,
126 struct resource_list_x
*fail_head
)
128 struct resource_list head
;
131 __dev_sort_resources(dev
, &head
);
132 __assign_resources_sorted(&head
, fail_head
);
136 static void pbus_assign_resources_sorted(const struct pci_bus
*bus
,
137 struct resource_list_x
*fail_head
)
140 struct resource_list head
;
143 list_for_each_entry(dev
, &bus
->devices
, bus_list
)
144 __dev_sort_resources(dev
, &head
);
146 __assign_resources_sorted(&head
, fail_head
);
149 void pci_setup_cardbus(struct pci_bus
*bus
)
151 struct pci_dev
*bridge
= bus
->self
;
152 struct resource
*res
;
153 struct pci_bus_region region
;
155 dev_info(&bridge
->dev
, "CardBus bridge to [bus %02x-%02x]\n",
156 bus
->secondary
, bus
->subordinate
);
158 res
= bus
->resource
[0];
159 pcibios_resource_to_bus(bridge
, ®ion
, res
);
160 if (res
->flags
& IORESOURCE_IO
) {
162 * The IO resource is allocated a range twice as large as it
163 * would normally need. This allows us to set both IO regs.
165 dev_info(&bridge
->dev
, " bridge window %pR\n", res
);
166 pci_write_config_dword(bridge
, PCI_CB_IO_BASE_0
,
168 pci_write_config_dword(bridge
, PCI_CB_IO_LIMIT_0
,
172 res
= bus
->resource
[1];
173 pcibios_resource_to_bus(bridge
, ®ion
, res
);
174 if (res
->flags
& IORESOURCE_IO
) {
175 dev_info(&bridge
->dev
, " bridge window %pR\n", res
);
176 pci_write_config_dword(bridge
, PCI_CB_IO_BASE_1
,
178 pci_write_config_dword(bridge
, PCI_CB_IO_LIMIT_1
,
182 res
= bus
->resource
[2];
183 pcibios_resource_to_bus(bridge
, ®ion
, res
);
184 if (res
->flags
& IORESOURCE_MEM
) {
185 dev_info(&bridge
->dev
, " bridge window %pR\n", res
);
186 pci_write_config_dword(bridge
, PCI_CB_MEMORY_BASE_0
,
188 pci_write_config_dword(bridge
, PCI_CB_MEMORY_LIMIT_0
,
192 res
= bus
->resource
[3];
193 pcibios_resource_to_bus(bridge
, ®ion
, res
);
194 if (res
->flags
& IORESOURCE_MEM
) {
195 dev_info(&bridge
->dev
, " bridge window %pR\n", res
);
196 pci_write_config_dword(bridge
, PCI_CB_MEMORY_BASE_1
,
198 pci_write_config_dword(bridge
, PCI_CB_MEMORY_LIMIT_1
,
202 EXPORT_SYMBOL(pci_setup_cardbus
);
204 /* Initialize bridges with base/limit values we have collected.
205 PCI-to-PCI Bridge Architecture Specification rev. 1.1 (1998)
206 requires that if there is no I/O ports or memory behind the
207 bridge, corresponding range must be turned off by writing base
208 value greater than limit to the bridge's base/limit registers.
210 Note: care must be taken when updating I/O base/limit registers
211 of bridges which support 32-bit I/O. This update requires two
212 config space writes, so it's quite possible that an I/O window of
213 the bridge will have some undesirable address (e.g. 0) after the
214 first write. Ditto 64-bit prefetchable MMIO. */
215 static void pci_setup_bridge_io(struct pci_bus
*bus
)
217 struct pci_dev
*bridge
= bus
->self
;
218 struct resource
*res
;
219 struct pci_bus_region region
;
222 /* Set up the top and bottom of the PCI I/O segment for this bus. */
223 res
= bus
->resource
[0];
224 pcibios_resource_to_bus(bridge
, ®ion
, res
);
225 if (res
->flags
& IORESOURCE_IO
) {
226 pci_read_config_dword(bridge
, PCI_IO_BASE
, &l
);
228 l
|= (region
.start
>> 8) & 0x00f0;
229 l
|= region
.end
& 0xf000;
230 /* Set up upper 16 bits of I/O base/limit. */
231 io_upper16
= (region
.end
& 0xffff0000) | (region
.start
>> 16);
232 dev_info(&bridge
->dev
, " bridge window %pR\n", res
);
234 /* Clear upper 16 bits of I/O base/limit. */
237 dev_info(&bridge
->dev
, " bridge window [io disabled]\n");
239 /* Temporarily disable the I/O range before updating PCI_IO_BASE. */
240 pci_write_config_dword(bridge
, PCI_IO_BASE_UPPER16
, 0x0000ffff);
241 /* Update lower 16 bits of I/O base/limit. */
242 pci_write_config_dword(bridge
, PCI_IO_BASE
, l
);
243 /* Update upper 16 bits of I/O base/limit. */
244 pci_write_config_dword(bridge
, PCI_IO_BASE_UPPER16
, io_upper16
);
247 static void pci_setup_bridge_mmio(struct pci_bus
*bus
)
249 struct pci_dev
*bridge
= bus
->self
;
250 struct resource
*res
;
251 struct pci_bus_region region
;
254 /* Set up the top and bottom of the PCI Memory segment for this bus. */
255 res
= bus
->resource
[1];
256 pcibios_resource_to_bus(bridge
, ®ion
, res
);
257 if (res
->flags
& IORESOURCE_MEM
) {
258 l
= (region
.start
>> 16) & 0xfff0;
259 l
|= region
.end
& 0xfff00000;
260 dev_info(&bridge
->dev
, " bridge window %pR\n", res
);
263 dev_info(&bridge
->dev
, " bridge window [mem disabled]\n");
265 pci_write_config_dword(bridge
, PCI_MEMORY_BASE
, l
);
268 static void pci_setup_bridge_mmio_pref(struct pci_bus
*bus
)
270 struct pci_dev
*bridge
= bus
->self
;
271 struct resource
*res
;
272 struct pci_bus_region region
;
275 /* Clear out the upper 32 bits of PREF limit.
276 If PCI_PREF_BASE_UPPER32 was non-zero, this temporarily
277 disables PREF range, which is ok. */
278 pci_write_config_dword(bridge
, PCI_PREF_LIMIT_UPPER32
, 0);
280 /* Set up PREF base/limit. */
282 res
= bus
->resource
[2];
283 pcibios_resource_to_bus(bridge
, ®ion
, res
);
284 if (res
->flags
& IORESOURCE_PREFETCH
) {
285 l
= (region
.start
>> 16) & 0xfff0;
286 l
|= region
.end
& 0xfff00000;
287 if (res
->flags
& IORESOURCE_MEM_64
) {
288 bu
= upper_32_bits(region
.start
);
289 lu
= upper_32_bits(region
.end
);
291 dev_info(&bridge
->dev
, " bridge window %pR\n", res
);
294 dev_info(&bridge
->dev
, " bridge window [mem pref disabled]\n");
296 pci_write_config_dword(bridge
, PCI_PREF_MEMORY_BASE
, l
);
298 /* Set the upper 32 bits of PREF base & limit. */
299 pci_write_config_dword(bridge
, PCI_PREF_BASE_UPPER32
, bu
);
300 pci_write_config_dword(bridge
, PCI_PREF_LIMIT_UPPER32
, lu
);
303 static void __pci_setup_bridge(struct pci_bus
*bus
, unsigned long type
)
305 struct pci_dev
*bridge
= bus
->self
;
307 dev_info(&bridge
->dev
, "PCI bridge to [bus %02x-%02x]\n",
308 bus
->secondary
, bus
->subordinate
);
310 if (type
& IORESOURCE_IO
)
311 pci_setup_bridge_io(bus
);
313 if (type
& IORESOURCE_MEM
)
314 pci_setup_bridge_mmio(bus
);
316 if (type
& IORESOURCE_PREFETCH
)
317 pci_setup_bridge_mmio_pref(bus
);
319 pci_write_config_word(bridge
, PCI_BRIDGE_CONTROL
, bus
->bridge_ctl
);
322 static void pci_setup_bridge(struct pci_bus
*bus
)
324 unsigned long type
= IORESOURCE_IO
| IORESOURCE_MEM
|
327 __pci_setup_bridge(bus
, type
);
330 /* Check whether the bridge supports optional I/O and
331 prefetchable memory ranges. If not, the respective
332 base/limit registers must be read-only and read as 0. */
333 static void pci_bridge_check_ranges(struct pci_bus
*bus
)
337 struct pci_dev
*bridge
= bus
->self
;
338 struct resource
*b_res
;
340 b_res
= &bridge
->resource
[PCI_BRIDGE_RESOURCES
];
341 b_res
[1].flags
|= IORESOURCE_MEM
;
343 pci_read_config_word(bridge
, PCI_IO_BASE
, &io
);
345 pci_write_config_word(bridge
, PCI_IO_BASE
, 0xf0f0);
346 pci_read_config_word(bridge
, PCI_IO_BASE
, &io
);
347 pci_write_config_word(bridge
, PCI_IO_BASE
, 0x0);
350 b_res
[0].flags
|= IORESOURCE_IO
;
351 /* DECchip 21050 pass 2 errata: the bridge may miss an address
352 disconnect boundary by one PCI data phase.
353 Workaround: do not use prefetching on this device. */
354 if (bridge
->vendor
== PCI_VENDOR_ID_DEC
&& bridge
->device
== 0x0001)
356 pci_read_config_dword(bridge
, PCI_PREF_MEMORY_BASE
, &pmem
);
358 pci_write_config_dword(bridge
, PCI_PREF_MEMORY_BASE
,
360 pci_read_config_dword(bridge
, PCI_PREF_MEMORY_BASE
, &pmem
);
361 pci_write_config_dword(bridge
, PCI_PREF_MEMORY_BASE
, 0x0);
364 b_res
[2].flags
|= IORESOURCE_MEM
| IORESOURCE_PREFETCH
;
365 if ((pmem
& PCI_PREF_RANGE_TYPE_MASK
) ==
366 PCI_PREF_RANGE_TYPE_64
) {
367 b_res
[2].flags
|= IORESOURCE_MEM_64
;
368 b_res
[2].flags
|= PCI_PREF_RANGE_TYPE_64
;
372 /* double check if bridge does support 64 bit pref */
373 if (b_res
[2].flags
& IORESOURCE_MEM_64
) {
374 u32 mem_base_hi
, tmp
;
375 pci_read_config_dword(bridge
, PCI_PREF_BASE_UPPER32
,
377 pci_write_config_dword(bridge
, PCI_PREF_BASE_UPPER32
,
379 pci_read_config_dword(bridge
, PCI_PREF_BASE_UPPER32
, &tmp
);
381 b_res
[2].flags
&= ~IORESOURCE_MEM_64
;
382 pci_write_config_dword(bridge
, PCI_PREF_BASE_UPPER32
,
387 /* Helper function for sizing routines: find first available
388 bus resource of a given type. Note: we intentionally skip
389 the bus resources which have already been assigned (that is,
390 have non-NULL parent resource). */
391 static struct resource
*find_free_bus_resource(struct pci_bus
*bus
, unsigned long type
)
395 unsigned long type_mask
= IORESOURCE_IO
| IORESOURCE_MEM
|
398 pci_bus_for_each_resource(bus
, r
, i
) {
399 if (r
== &ioport_resource
|| r
== &iomem_resource
)
401 if (r
&& (r
->flags
& type_mask
) == type
&& !r
->parent
)
407 /* Sizing the IO windows of the PCI-PCI bridge is trivial,
408 since these windows have 4K granularity and the IO ranges
409 of non-bridge PCI devices are limited to 256 bytes.
410 We must be careful with the ISA aliasing though. */
411 static void pbus_size_io(struct pci_bus
*bus
, resource_size_t min_size
)
414 struct resource
*b_res
= find_free_bus_resource(bus
, IORESOURCE_IO
);
415 unsigned long size
= 0, size1
= 0, old_size
;
420 list_for_each_entry(dev
, &bus
->devices
, bus_list
) {
423 for (i
= 0; i
< PCI_NUM_RESOURCES
; i
++) {
424 struct resource
*r
= &dev
->resource
[i
];
425 unsigned long r_size
;
427 if (r
->parent
|| !(r
->flags
& IORESOURCE_IO
))
429 r_size
= resource_size(r
);
432 /* Might be re-aligned for ISA */
440 old_size
= resource_size(b_res
);
443 /* To be fixed in 2.5: we should have sort of HAVE_ISA
444 flag in the struct pci_bus. */
445 #if defined(CONFIG_ISA) || defined(CONFIG_EISA)
446 size
= (size
& 0xff) + ((size
& ~0xffUL
) << 2);
448 size
= ALIGN(size
+ size1
, 4096);
452 if (b_res
->start
|| b_res
->end
)
453 dev_info(&bus
->self
->dev
, "disabling bridge window "
454 "%pR to [bus %02x-%02x] (unused)\n", b_res
,
455 bus
->secondary
, bus
->subordinate
);
459 /* Alignment of the IO window is always 4K */
461 b_res
->end
= b_res
->start
+ size
- 1;
462 b_res
->flags
|= IORESOURCE_STARTALIGN
;
465 /* Calculate the size of the bus and minimal alignment which
466 guarantees that all child resources fit in this size. */
467 static int pbus_size_mem(struct pci_bus
*bus
, unsigned long mask
,
468 unsigned long type
, resource_size_t min_size
)
471 resource_size_t min_align
, align
, size
, old_size
;
472 resource_size_t aligns
[12]; /* Alignments from 1Mb to 2Gb */
473 int order
, max_order
;
474 struct resource
*b_res
= find_free_bus_resource(bus
, type
);
475 unsigned int mem64_mask
= 0;
480 memset(aligns
, 0, sizeof(aligns
));
484 mem64_mask
= b_res
->flags
& IORESOURCE_MEM_64
;
485 b_res
->flags
&= ~IORESOURCE_MEM_64
;
487 list_for_each_entry(dev
, &bus
->devices
, bus_list
) {
490 for (i
= 0; i
< PCI_NUM_RESOURCES
; i
++) {
491 struct resource
*r
= &dev
->resource
[i
];
492 resource_size_t r_size
;
494 if (r
->parent
|| (r
->flags
& mask
) != type
)
496 r_size
= resource_size(r
);
497 /* For bridges size != alignment */
498 align
= pci_resource_alignment(dev
, r
);
499 order
= __ffs(align
) - 20;
501 dev_warn(&dev
->dev
, "disabling BAR %d: %pR "
502 "(bad alignment %#llx)\n", i
, r
,
503 (unsigned long long) align
);
510 /* Exclude ranges with size > align from
511 calculation of the alignment. */
513 aligns
[order
] += align
;
514 if (order
> max_order
)
516 mem64_mask
&= r
->flags
& IORESOURCE_MEM_64
;
521 old_size
= resource_size(b_res
);
529 for (order
= 0; order
<= max_order
; order
++) {
530 resource_size_t align1
= 1;
532 align1
<<= (order
+ 20);
536 else if (ALIGN(align
+ min_align
, min_align
) < align1
)
537 min_align
= align1
>> 1;
538 align
+= aligns
[order
];
540 size
= ALIGN(size
, min_align
);
542 if (b_res
->start
|| b_res
->end
)
543 dev_info(&bus
->self
->dev
, "disabling bridge window "
544 "%pR to [bus %02x-%02x] (unused)\n", b_res
,
545 bus
->secondary
, bus
->subordinate
);
549 b_res
->start
= min_align
;
550 b_res
->end
= size
+ min_align
- 1;
551 b_res
->flags
|= IORESOURCE_STARTALIGN
;
552 b_res
->flags
|= mem64_mask
;
556 static void pci_bus_size_cardbus(struct pci_bus
*bus
)
558 struct pci_dev
*bridge
= bus
->self
;
559 struct resource
*b_res
= &bridge
->resource
[PCI_BRIDGE_RESOURCES
];
563 * Reserve some resources for CardBus. We reserve
564 * a fixed amount of bus space for CardBus bridges.
567 b_res
[0].end
= pci_cardbus_io_size
- 1;
568 b_res
[0].flags
|= IORESOURCE_IO
| IORESOURCE_SIZEALIGN
;
571 b_res
[1].end
= pci_cardbus_io_size
- 1;
572 b_res
[1].flags
|= IORESOURCE_IO
| IORESOURCE_SIZEALIGN
;
575 * Check whether prefetchable memory is supported
578 pci_read_config_word(bridge
, PCI_CB_BRIDGE_CONTROL
, &ctrl
);
579 if (!(ctrl
& PCI_CB_BRIDGE_CTL_PREFETCH_MEM0
)) {
580 ctrl
|= PCI_CB_BRIDGE_CTL_PREFETCH_MEM0
;
581 pci_write_config_word(bridge
, PCI_CB_BRIDGE_CONTROL
, ctrl
);
582 pci_read_config_word(bridge
, PCI_CB_BRIDGE_CONTROL
, &ctrl
);
586 * If we have prefetchable memory support, allocate
587 * two regions. Otherwise, allocate one region of
590 if (ctrl
& PCI_CB_BRIDGE_CTL_PREFETCH_MEM0
) {
592 b_res
[2].end
= pci_cardbus_mem_size
- 1;
593 b_res
[2].flags
|= IORESOURCE_MEM
| IORESOURCE_PREFETCH
| IORESOURCE_SIZEALIGN
;
596 b_res
[3].end
= pci_cardbus_mem_size
- 1;
597 b_res
[3].flags
|= IORESOURCE_MEM
| IORESOURCE_SIZEALIGN
;
600 b_res
[3].end
= pci_cardbus_mem_size
* 2 - 1;
601 b_res
[3].flags
|= IORESOURCE_MEM
| IORESOURCE_SIZEALIGN
;
605 void __ref
pci_bus_size_bridges(struct pci_bus
*bus
)
608 unsigned long mask
, prefmask
;
609 resource_size_t min_mem_size
= 0, min_io_size
= 0;
611 list_for_each_entry(dev
, &bus
->devices
, bus_list
) {
612 struct pci_bus
*b
= dev
->subordinate
;
616 switch (dev
->class >> 8) {
617 case PCI_CLASS_BRIDGE_CARDBUS
:
618 pci_bus_size_cardbus(b
);
621 case PCI_CLASS_BRIDGE_PCI
:
623 pci_bus_size_bridges(b
);
632 switch (bus
->self
->class >> 8) {
633 case PCI_CLASS_BRIDGE_CARDBUS
:
634 /* don't size cardbuses yet. */
637 case PCI_CLASS_BRIDGE_PCI
:
638 pci_bridge_check_ranges(bus
);
639 if (bus
->self
->is_hotplug_bridge
) {
640 min_io_size
= pci_hotplug_io_size
;
641 min_mem_size
= pci_hotplug_mem_size
;
644 pbus_size_io(bus
, min_io_size
);
645 /* If the bridge supports prefetchable range, size it
646 separately. If it doesn't, or its prefetchable window
647 has already been allocated by arch code, try
648 non-prefetchable range for both types of PCI memory
650 mask
= IORESOURCE_MEM
;
651 prefmask
= IORESOURCE_MEM
| IORESOURCE_PREFETCH
;
652 if (pbus_size_mem(bus
, prefmask
, prefmask
, min_mem_size
))
653 mask
= prefmask
; /* Success, size non-prefetch only. */
655 min_mem_size
+= min_mem_size
;
656 pbus_size_mem(bus
, mask
, IORESOURCE_MEM
, min_mem_size
);
660 EXPORT_SYMBOL(pci_bus_size_bridges
);
662 static void __ref
__pci_bus_assign_resources(const struct pci_bus
*bus
,
663 struct resource_list_x
*fail_head
)
668 pbus_assign_resources_sorted(bus
, fail_head
);
670 list_for_each_entry(dev
, &bus
->devices
, bus_list
) {
671 b
= dev
->subordinate
;
675 __pci_bus_assign_resources(b
, fail_head
);
677 switch (dev
->class >> 8) {
678 case PCI_CLASS_BRIDGE_PCI
:
679 if (!pci_is_enabled(dev
))
683 case PCI_CLASS_BRIDGE_CARDBUS
:
684 pci_setup_cardbus(b
);
688 dev_info(&dev
->dev
, "not setting up bridge for bus "
689 "%04x:%02x\n", pci_domain_nr(b
), b
->number
);
695 void __ref
pci_bus_assign_resources(const struct pci_bus
*bus
)
697 __pci_bus_assign_resources(bus
, NULL
);
699 EXPORT_SYMBOL(pci_bus_assign_resources
);
701 static void __ref
__pci_bridge_assign_resources(const struct pci_dev
*bridge
,
702 struct resource_list_x
*fail_head
)
706 pdev_assign_resources_sorted((struct pci_dev
*)bridge
, fail_head
);
708 b
= bridge
->subordinate
;
712 __pci_bus_assign_resources(b
, fail_head
);
714 switch (bridge
->class >> 8) {
715 case PCI_CLASS_BRIDGE_PCI
:
719 case PCI_CLASS_BRIDGE_CARDBUS
:
720 pci_setup_cardbus(b
);
724 dev_info(&bridge
->dev
, "not setting up bridge for bus "
725 "%04x:%02x\n", pci_domain_nr(b
), b
->number
);
729 static void pci_bridge_release_resources(struct pci_bus
*bus
,
733 bool changed
= false;
736 unsigned long type_mask
= IORESOURCE_IO
| IORESOURCE_MEM
|
740 for (idx
= PCI_BRIDGE_RESOURCES
; idx
<= PCI_BRIDGE_RESOURCE_END
;
742 r
= &dev
->resource
[idx
];
743 if ((r
->flags
& type_mask
) != type
)
748 * if there are children under that, we should release them
751 release_child_resources(r
);
752 if (!release_resource(r
)) {
753 dev_printk(KERN_DEBUG
, &dev
->dev
,
754 "resource %d %pR released\n", idx
, r
);
755 /* keep the old size */
756 r
->end
= resource_size(r
) - 1;
764 /* avoiding touch the one without PREF */
765 if (type
& IORESOURCE_PREFETCH
)
766 type
= IORESOURCE_PREFETCH
;
767 __pci_setup_bridge(bus
, type
);
776 * try to release pci bridge resources that is from leaf bridge,
777 * so we can allocate big new one later
779 static void __ref
pci_bus_release_bridge_resources(struct pci_bus
*bus
,
781 enum release_type rel_type
)
784 bool is_leaf_bridge
= true;
786 list_for_each_entry(dev
, &bus
->devices
, bus_list
) {
787 struct pci_bus
*b
= dev
->subordinate
;
791 is_leaf_bridge
= false;
793 if ((dev
->class >> 8) != PCI_CLASS_BRIDGE_PCI
)
796 if (rel_type
== whole_subtree
)
797 pci_bus_release_bridge_resources(b
, type
,
801 if (pci_is_root_bus(bus
))
804 if ((bus
->self
->class >> 8) != PCI_CLASS_BRIDGE_PCI
)
807 if ((rel_type
== whole_subtree
) || is_leaf_bridge
)
808 pci_bridge_release_resources(bus
, type
);
811 static void pci_bus_dump_res(struct pci_bus
*bus
)
813 struct resource
*res
;
816 pci_bus_for_each_resource(bus
, res
, i
) {
817 if (!res
|| !res
->end
|| !res
->flags
)
820 dev_printk(KERN_DEBUG
, &bus
->dev
, "resource %d %pR\n", i
, res
);
824 static void pci_bus_dump_resources(struct pci_bus
*bus
)
830 pci_bus_dump_res(bus
);
832 list_for_each_entry(dev
, &bus
->devices
, bus_list
) {
833 b
= dev
->subordinate
;
837 pci_bus_dump_resources(b
);
841 static int __init
pci_bus_get_depth(struct pci_bus
*bus
)
846 list_for_each_entry(dev
, &bus
->devices
, bus_list
) {
848 struct pci_bus
*b
= dev
->subordinate
;
852 ret
= pci_bus_get_depth(b
);
859 static int __init
pci_get_max_depth(void)
864 list_for_each_entry(bus
, &pci_root_buses
, node
) {
867 ret
= pci_bus_get_depth(bus
);
876 * first try will not touch pci bridge res
877 * second and later try will clear small leaf bridge res
878 * will stop till to the max deepth if can not find good one
881 pci_assign_unassigned_resources(void)
885 enum release_type rel_type
= leaf_only
;
886 struct resource_list_x head
, *list
;
887 unsigned long type_mask
= IORESOURCE_IO
| IORESOURCE_MEM
|
889 unsigned long failed_type
;
890 int max_depth
= pci_get_max_depth();
895 pci_try_num
= max_depth
+ 1;
896 printk(KERN_DEBUG
"PCI: max bus depth: %d pci_try_num: %d\n",
897 max_depth
, pci_try_num
);
900 /* Depth first, calculate sizes and alignments of all
901 subordinate buses. */
902 list_for_each_entry(bus
, &pci_root_buses
, node
) {
903 pci_bus_size_bridges(bus
);
905 /* Depth last, allocate resources and update the hardware. */
906 list_for_each_entry(bus
, &pci_root_buses
, node
) {
907 __pci_bus_assign_resources(bus
, &head
);
911 /* any device complain? */
913 goto enable_and_dump
;
915 for (list
= head
.next
; list
;) {
916 failed_type
|= list
->flags
;
920 * io port are tight, don't try extra
921 * or if reach the limit, don't want to try more
923 failed_type
&= type_mask
;
924 if ((failed_type
== IORESOURCE_IO
) || (tried_times
>= pci_try_num
)) {
925 free_failed_list(&head
);
926 goto enable_and_dump
;
929 printk(KERN_DEBUG
"PCI: No. %d try to assign unassigned res\n",
932 /* third times and later will not check if it is leaf */
933 if ((tried_times
+ 1) > 2)
934 rel_type
= whole_subtree
;
937 * Try to release leaf bridge's resources that doesn't fit resource of
938 * child device under that bridge
940 for (list
= head
.next
; list
;) {
941 bus
= list
->dev
->bus
;
942 pci_bus_release_bridge_resources(bus
, list
->flags
& type_mask
,
946 /* restore size and flags */
947 for (list
= head
.next
; list
;) {
948 struct resource
*res
= list
->res
;
950 res
->start
= list
->start
;
951 res
->end
= list
->end
;
952 res
->flags
= list
->flags
;
953 if (list
->dev
->subordinate
)
958 free_failed_list(&head
);
963 /* Depth last, update the hardware. */
964 list_for_each_entry(bus
, &pci_root_buses
, node
)
965 pci_enable_bridges(bus
);
967 /* dump the resource on buses */
968 list_for_each_entry(bus
, &pci_root_buses
, node
) {
969 pci_bus_dump_resources(bus
);
973 void pci_assign_unassigned_bridge_resources(struct pci_dev
*bridge
)
975 struct pci_bus
*parent
= bridge
->subordinate
;
977 struct resource_list_x head
, *list
;
979 unsigned long type_mask
= IORESOURCE_IO
| IORESOURCE_MEM
|
985 pci_bus_size_bridges(parent
);
986 __pci_bridge_assign_resources(bridge
, &head
);
987 retval
= pci_reenable_device(bridge
);
988 pci_set_master(bridge
);
989 pci_enable_bridges(parent
);
996 if (tried_times
>= 2) {
997 /* still fail, don't need to try more */
998 free_failed_list(&head
);
1002 printk(KERN_DEBUG
"PCI: No. %d try to assign unassigned res\n",
1006 * Try to release leaf bridge's resources that doesn't fit resource of
1007 * child device under that bridge
1009 for (list
= head
.next
; list
;) {
1010 struct pci_bus
*bus
= list
->dev
->bus
;
1011 unsigned long flags
= list
->flags
;
1013 pci_bus_release_bridge_resources(bus
, flags
& type_mask
,
1017 /* restore size and flags */
1018 for (list
= head
.next
; list
;) {
1019 struct resource
*res
= list
->res
;
1021 res
->start
= list
->start
;
1022 res
->end
= list
->end
;
1023 res
->flags
= list
->flags
;
1024 if (list
->dev
->subordinate
)
1029 free_failed_list(&head
);
1033 EXPORT_SYMBOL_GPL(pci_assign_unassigned_bridge_resources
);