2 * PCI Error Recovery Driver for RPA-compliant PPC64 platform.
3 * Copyright IBM Corp. 2004 2005
4 * Copyright Linas Vepstas <linas@linas.org> 2004, 2005
8 * This program is free software; you can redistribute it and/or modify
9 * it under the terms of the GNU General Public License as published by
10 * the Free Software Foundation; either version 2 of the License, or (at
11 * your option) any later version.
13 * This program is distributed in the hope that it will be useful, but
14 * WITHOUT ANY WARRANTY; without even the implied warranty of
15 * MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE, GOOD TITLE or
16 * NON INFRINGEMENT. See the GNU General Public License for more
19 * You should have received a copy of the GNU General Public License
20 * along with this program; if not, write to the Free Software
21 * Foundation, Inc., 675 Mass Ave, Cambridge, MA 02139, USA.
23 * Send comments and feedback to Linas Vepstas <linas@austin.ibm.com>
25 #include <linux/delay.h>
26 #include <linux/interrupt.h>
27 #include <linux/irq.h>
28 #include <linux/module.h>
29 #include <linux/pci.h>
31 #include <asm/eeh_event.h>
32 #include <asm/ppc-pci.h>
33 #include <asm/pci-bridge.h>
38 * eeh_pcid_name - Retrieve name of PCI device driver
41 * This routine is used to retrieve the name of PCI device driver
44 static inline const char *eeh_pcid_name(struct pci_dev
*pdev
)
46 if (pdev
&& pdev
->dev
.driver
)
47 return pdev
->dev
.driver
->name
;
52 * eeh_pcid_get - Get the PCI device driver
55 * The function is used to retrieve the PCI device driver for
56 * the indicated PCI device. Besides, we will increase the reference
57 * of the PCI device driver to prevent that being unloaded on
58 * the fly. Otherwise, kernel crash would be seen.
60 static inline struct pci_driver
*eeh_pcid_get(struct pci_dev
*pdev
)
62 if (!pdev
|| !pdev
->driver
)
65 if (!try_module_get(pdev
->driver
->driver
.owner
))
72 * eeh_pcid_put - Dereference on the PCI device driver
75 * The function is called to do dereference on the PCI device
76 * driver of the indicated PCI device.
78 static inline void eeh_pcid_put(struct pci_dev
*pdev
)
80 if (!pdev
|| !pdev
->driver
)
83 module_put(pdev
->driver
->driver
.owner
);
87 static void print_device_node_tree(struct pci_dn
*pdn
, int dent
)
90 struct device_node
*pc
;
94 for (i
= 0; i
< dent
; i
++)
96 printk("dn=%s mode=%x \tcfg_addr=%x pe_addr=%x \tfull=%s\n",
97 pdn
->node
->name
, pdn
->eeh_mode
, pdn
->eeh_config_addr
,
98 pdn
->eeh_pe_config_addr
, pdn
->node
->full_name
);
100 pc
= pdn
->node
->child
;
102 print_device_node_tree(PCI_DN(pc
), dent
);
109 * eeh_disable_irq - Disable interrupt for the recovering device
112 * This routine must be called when reporting temporary or permanent
113 * error to the particular PCI device to disable interrupt of that
114 * device. If the device has enabled MSI or MSI-X interrupt, we needn't
115 * do real work because EEH should freeze DMA transfers for those PCI
116 * devices encountering EEH errors, which includes MSI or MSI-X.
118 static void eeh_disable_irq(struct pci_dev
*dev
)
120 struct eeh_dev
*edev
= pci_dev_to_eeh_dev(dev
);
122 /* Don't disable MSI and MSI-X interrupts. They are
123 * effectively disabled by the DMA Stopped state
124 * when an EEH error occurs.
126 if (dev
->msi_enabled
|| dev
->msix_enabled
)
129 if (!irq_has_action(dev
->irq
))
132 edev
->mode
|= EEH_DEV_IRQ_DISABLED
;
133 disable_irq_nosync(dev
->irq
);
137 * eeh_enable_irq - Enable interrupt for the recovering device
140 * This routine must be called to enable interrupt while failed
141 * device could be resumed.
143 static void eeh_enable_irq(struct pci_dev
*dev
)
145 struct eeh_dev
*edev
= pci_dev_to_eeh_dev(dev
);
147 if ((edev
->mode
) & EEH_DEV_IRQ_DISABLED
) {
148 edev
->mode
&= ~EEH_DEV_IRQ_DISABLED
;
152 * This is just ass backwards. This maze has
153 * unbalanced irq_enable/disable calls. So instead of
154 * finding the root cause it works around the warning
155 * in the irq_enable code by conditionally calling
158 * That's just wrong.The warning in the core code is
159 * there to tell people to fix their assymetries in
160 * their own code, not by abusing the core information
163 * I so wish that the assymetry would be the other way
164 * round and a few more irq_disable calls render that
165 * shit unusable forever.
169 if (irqd_irq_disabled(irq_get_irq_data(dev
->irq
)))
170 enable_irq(dev
->irq
);
174 static bool eeh_dev_removed(struct eeh_dev
*edev
)
176 /* EEH device removed ? */
177 if (!edev
|| (edev
->mode
& EEH_DEV_REMOVED
))
183 static void *eeh_dev_save_state(void *data
, void *userdata
)
185 struct eeh_dev
*edev
= data
;
186 struct pci_dev
*pdev
;
191 pdev
= eeh_dev_to_pci_dev(edev
);
195 pci_save_state(pdev
);
200 * eeh_report_error - Report pci error to each device driver
202 * @userdata: return value
204 * Report an EEH error to each device driver, collect up and
205 * merge the device driver responses. Cumulative response
206 * passed back in "userdata".
208 static void *eeh_report_error(void *data
, void *userdata
)
210 struct eeh_dev
*edev
= (struct eeh_dev
*)data
;
211 struct pci_dev
*dev
= eeh_dev_to_pci_dev(edev
);
212 enum pci_ers_result rc
, *res
= userdata
;
213 struct pci_driver
*driver
;
215 if (!dev
|| eeh_dev_removed(edev
))
217 dev
->error_state
= pci_channel_io_frozen
;
219 driver
= eeh_pcid_get(dev
);
220 if (!driver
) return NULL
;
222 eeh_disable_irq(dev
);
224 if (!driver
->err_handler
||
225 !driver
->err_handler
->error_detected
) {
230 rc
= driver
->err_handler
->error_detected(dev
, pci_channel_io_frozen
);
232 /* A driver that needs a reset trumps all others */
233 if (rc
== PCI_ERS_RESULT_NEED_RESET
) *res
= rc
;
234 if (*res
== PCI_ERS_RESULT_NONE
) *res
= rc
;
241 * eeh_report_mmio_enabled - Tell drivers that MMIO has been enabled
243 * @userdata: return value
245 * Tells each device driver that IO ports, MMIO and config space I/O
246 * are now enabled. Collects up and merges the device driver responses.
247 * Cumulative response passed back in "userdata".
249 static void *eeh_report_mmio_enabled(void *data
, void *userdata
)
251 struct eeh_dev
*edev
= (struct eeh_dev
*)data
;
252 struct pci_dev
*dev
= eeh_dev_to_pci_dev(edev
);
253 enum pci_ers_result rc
, *res
= userdata
;
254 struct pci_driver
*driver
;
256 if (!dev
|| eeh_dev_removed(edev
))
259 driver
= eeh_pcid_get(dev
);
260 if (!driver
) return NULL
;
262 if (!driver
->err_handler
||
263 !driver
->err_handler
->mmio_enabled
||
264 (edev
->mode
& EEH_DEV_NO_HANDLER
)) {
269 rc
= driver
->err_handler
->mmio_enabled(dev
);
271 /* A driver that needs a reset trumps all others */
272 if (rc
== PCI_ERS_RESULT_NEED_RESET
) *res
= rc
;
273 if (*res
== PCI_ERS_RESULT_NONE
) *res
= rc
;
280 * eeh_report_reset - Tell device that slot has been reset
282 * @userdata: return value
284 * This routine must be called while EEH tries to reset particular
285 * PCI device so that the associated PCI device driver could take
286 * some actions, usually to save data the driver needs so that the
287 * driver can work again while the device is recovered.
289 static void *eeh_report_reset(void *data
, void *userdata
)
291 struct eeh_dev
*edev
= (struct eeh_dev
*)data
;
292 struct pci_dev
*dev
= eeh_dev_to_pci_dev(edev
);
293 enum pci_ers_result rc
, *res
= userdata
;
294 struct pci_driver
*driver
;
296 if (!dev
|| eeh_dev_removed(edev
))
298 dev
->error_state
= pci_channel_io_normal
;
300 driver
= eeh_pcid_get(dev
);
301 if (!driver
) return NULL
;
305 if (!driver
->err_handler
||
306 !driver
->err_handler
->slot_reset
||
307 (edev
->mode
& EEH_DEV_NO_HANDLER
)) {
312 rc
= driver
->err_handler
->slot_reset(dev
);
313 if ((*res
== PCI_ERS_RESULT_NONE
) ||
314 (*res
== PCI_ERS_RESULT_RECOVERED
)) *res
= rc
;
315 if (*res
== PCI_ERS_RESULT_DISCONNECT
&&
316 rc
== PCI_ERS_RESULT_NEED_RESET
) *res
= rc
;
322 static void *eeh_dev_restore_state(void *data
, void *userdata
)
324 struct eeh_dev
*edev
= data
;
325 struct pci_dev
*pdev
;
330 pdev
= eeh_dev_to_pci_dev(edev
);
334 pci_restore_state(pdev
);
339 * eeh_report_resume - Tell device to resume normal operations
341 * @userdata: return value
343 * This routine must be called to notify the device driver that it
344 * could resume so that the device driver can do some initialization
345 * to make the recovered device work again.
347 static void *eeh_report_resume(void *data
, void *userdata
)
349 struct eeh_dev
*edev
= (struct eeh_dev
*)data
;
350 struct pci_dev
*dev
= eeh_dev_to_pci_dev(edev
);
351 struct pci_driver
*driver
;
353 if (!dev
|| eeh_dev_removed(edev
))
355 dev
->error_state
= pci_channel_io_normal
;
357 driver
= eeh_pcid_get(dev
);
358 if (!driver
) return NULL
;
362 if (!driver
->err_handler
||
363 !driver
->err_handler
->resume
||
364 (edev
->mode
& EEH_DEV_NO_HANDLER
)) {
365 edev
->mode
&= ~EEH_DEV_NO_HANDLER
;
370 driver
->err_handler
->resume(dev
);
377 * eeh_report_failure - Tell device driver that device is dead.
379 * @userdata: return value
381 * This informs the device driver that the device is permanently
382 * dead, and that no further recovery attempts will be made on it.
384 static void *eeh_report_failure(void *data
, void *userdata
)
386 struct eeh_dev
*edev
= (struct eeh_dev
*)data
;
387 struct pci_dev
*dev
= eeh_dev_to_pci_dev(edev
);
388 struct pci_driver
*driver
;
390 if (!dev
|| eeh_dev_removed(edev
))
392 dev
->error_state
= pci_channel_io_perm_failure
;
394 driver
= eeh_pcid_get(dev
);
395 if (!driver
) return NULL
;
397 eeh_disable_irq(dev
);
399 if (!driver
->err_handler
||
400 !driver
->err_handler
->error_detected
) {
405 driver
->err_handler
->error_detected(dev
, pci_channel_io_perm_failure
);
411 static void *eeh_rmv_device(void *data
, void *userdata
)
413 struct pci_driver
*driver
;
414 struct eeh_dev
*edev
= (struct eeh_dev
*)data
;
415 struct pci_dev
*dev
= eeh_dev_to_pci_dev(edev
);
416 int *removed
= (int *)userdata
;
419 * Actually, we should remove the PCI bridges as well.
420 * However, that's lots of complexity to do that,
421 * particularly some of devices under the bridge might
422 * support EEH. So we just care about PCI devices for
425 if (!dev
|| (dev
->hdr_type
& PCI_HEADER_TYPE_BRIDGE
))
429 * We rely on count-based pcibios_release_device() to
430 * detach permanently offlined PEs. Unfortunately, that's
431 * not reliable enough. We might have the permanently
432 * offlined PEs attached, but we needn't take care of
433 * them and their child devices.
435 if (eeh_dev_removed(edev
))
438 driver
= eeh_pcid_get(dev
);
441 if (driver
->err_handler
)
445 /* Remove it from PCI subsystem */
446 pr_debug("EEH: Removing %s without EEH sensitive driver\n",
448 edev
->bus
= dev
->bus
;
449 edev
->mode
|= EEH_DEV_DISCONNECTED
;
452 pci_lock_rescan_remove();
453 pci_stop_and_remove_bus_device(dev
);
454 pci_unlock_rescan_remove();
459 static void *eeh_pe_detach_dev(void *data
, void *userdata
)
461 struct eeh_pe
*pe
= (struct eeh_pe
*)data
;
462 struct eeh_dev
*edev
, *tmp
;
464 eeh_pe_for_each_dev(pe
, edev
, tmp
) {
465 if (!(edev
->mode
& EEH_DEV_DISCONNECTED
))
468 edev
->mode
&= ~(EEH_DEV_DISCONNECTED
| EEH_DEV_IRQ_DISABLED
);
469 eeh_rmv_from_parent_pe(edev
);
476 * Explicitly clear PE's frozen state for PowerNV where
477 * we have frozen PE until BAR restore is completed. It's
478 * harmless to clear it for pSeries. To be consistent with
479 * PE reset (for 3 times), we try to clear the frozen state
480 * for 3 times as well.
482 static void *__eeh_clear_pe_frozen_state(void *data
, void *flag
)
484 struct eeh_pe
*pe
= (struct eeh_pe
*)data
;
485 bool *clear_sw_state
= flag
;
488 for (i
= 0; rc
&& i
< 3; i
++)
489 rc
= eeh_unfreeze_pe(pe
, clear_sw_state
);
491 /* Stop immediately on any errors */
493 pr_warn("%s: Failure %d unfreezing PHB#%x-PE#%x\n",
494 __func__
, rc
, pe
->phb
->global_number
, pe
->addr
);
501 static int eeh_clear_pe_frozen_state(struct eeh_pe
*pe
,
506 rc
= eeh_pe_traverse(pe
, __eeh_clear_pe_frozen_state
, &clear_sw_state
);
508 eeh_pe_state_clear(pe
, EEH_PE_ISOLATED
);
510 return rc
? -EIO
: 0;
513 int eeh_pe_reset_and_recover(struct eeh_pe
*pe
)
517 /* Bail if the PE is being recovered */
518 if (pe
->state
& EEH_PE_RECOVERING
)
521 /* Put the PE into recovery mode */
522 eeh_pe_state_mark(pe
, EEH_PE_RECOVERING
);
525 eeh_pe_dev_traverse(pe
, eeh_dev_save_state
, NULL
);
528 eeh_pe_dev_traverse(pe
, eeh_report_error
, &result
);
531 eeh_pe_state_mark(pe
, EEH_PE_RESET
);
532 ret
= eeh_reset_pe(pe
);
534 eeh_pe_state_clear(pe
, EEH_PE_RECOVERING
| EEH_PE_RESET
);
537 eeh_pe_state_clear(pe
, EEH_PE_RESET
);
539 /* Unfreeze the PE */
540 ret
= eeh_clear_pe_frozen_state(pe
, true);
542 eeh_pe_state_clear(pe
, EEH_PE_RECOVERING
);
546 /* Notify completion of reset */
547 eeh_pe_dev_traverse(pe
, eeh_report_reset
, &result
);
549 /* Restore device state */
550 eeh_pe_dev_traverse(pe
, eeh_dev_restore_state
, NULL
);
553 eeh_pe_dev_traverse(pe
, eeh_report_resume
, NULL
);
555 /* Clear recovery mode */
556 eeh_pe_state_clear(pe
, EEH_PE_RECOVERING
);
562 * eeh_reset_device - Perform actual reset of a pci slot
564 * @bus: PCI bus corresponding to the isolcated slot
566 * This routine must be called to do reset on the indicated PE.
567 * During the reset, udev might be invoked because those affected
568 * PCI devices will be removed and then added.
570 static int eeh_reset_device(struct eeh_pe
*pe
, struct pci_bus
*bus
)
572 struct pci_bus
*frozen_bus
= eeh_pe_bus_get(pe
);
573 struct timeval tstamp
;
574 int cnt
, rc
, removed
= 0;
576 /* pcibios will clear the counter; save the value */
577 cnt
= pe
->freeze_count
;
581 * We don't remove the corresponding PE instances because
582 * we need the information afterwords. The attached EEH
583 * devices are expected to be attached soon when calling
584 * into pcibios_add_pci_devices().
586 eeh_pe_state_mark(pe
, EEH_PE_KEEP
);
588 pci_lock_rescan_remove();
589 pcibios_remove_pci_devices(bus
);
590 pci_unlock_rescan_remove();
591 } else if (frozen_bus
) {
592 eeh_pe_dev_traverse(pe
, eeh_rmv_device
, &removed
);
596 * Reset the pci controller. (Asserts RST#; resets config space).
597 * Reconfigure bridges and devices. Don't try to bring the system
598 * up if the reset failed for some reason.
600 * During the reset, it's very dangerous to have uncontrolled PCI
601 * config accesses. So we prefer to block them. However, controlled
602 * PCI config accesses initiated from EEH itself are allowed.
604 eeh_pe_state_mark(pe
, EEH_PE_RESET
);
605 rc
= eeh_reset_pe(pe
);
607 eeh_pe_state_clear(pe
, EEH_PE_RESET
);
611 pci_lock_rescan_remove();
614 eeh_ops
->configure_bridge(pe
);
615 eeh_pe_restore_bars(pe
);
616 eeh_pe_state_clear(pe
, EEH_PE_RESET
);
618 /* Clear frozen state */
619 rc
= eeh_clear_pe_frozen_state(pe
, false);
623 /* Give the system 5 seconds to finish running the user-space
624 * hotplug shutdown scripts, e.g. ifdown for ethernet. Yes,
625 * this is a hack, but if we don't do this, and try to bring
626 * the device up before the scripts have taken it down,
627 * potentially weird things happen.
630 pr_info("EEH: Sleep 5s ahead of complete hotplug\n");
634 * The EEH device is still connected with its parent
635 * PE. We should disconnect it so the binding can be
636 * rebuilt when adding PCI devices.
638 eeh_pe_traverse(pe
, eeh_pe_detach_dev
, NULL
);
639 pcibios_add_pci_devices(bus
);
640 } else if (frozen_bus
&& removed
) {
641 pr_info("EEH: Sleep 5s ahead of partial hotplug\n");
644 eeh_pe_traverse(pe
, eeh_pe_detach_dev
, NULL
);
645 pcibios_add_pci_devices(frozen_bus
);
647 eeh_pe_state_clear(pe
, EEH_PE_KEEP
);
650 pe
->freeze_count
= cnt
;
652 pci_unlock_rescan_remove();
656 /* The longest amount of time to wait for a pci device
657 * to come back on line, in seconds.
659 #define MAX_WAIT_FOR_RECOVERY 300
661 static void eeh_handle_normal_event(struct eeh_pe
*pe
)
663 struct pci_bus
*frozen_bus
;
665 enum pci_ers_result result
= PCI_ERS_RESULT_NONE
;
667 frozen_bus
= eeh_pe_bus_get(pe
);
669 pr_err("%s: Cannot find PCI bus for PHB#%d-PE#%x\n",
670 __func__
, pe
->phb
->global_number
, pe
->addr
);
674 eeh_pe_update_time_stamp(pe
);
676 if (pe
->freeze_count
> EEH_MAX_ALLOWED_FREEZES
)
677 goto excess_failures
;
678 pr_warn("EEH: This PCI device has failed %d times in the last hour\n",
681 /* Walk the various device drivers attached to this slot through
682 * a reset sequence, giving each an opportunity to do what it needs
683 * to accomplish the reset. Each child gets a report of the
684 * status ... if any child can't handle the reset, then the entire
685 * slot is dlpar removed and added.
687 pr_info("EEH: Notify device drivers to shutdown\n");
688 eeh_pe_dev_traverse(pe
, eeh_report_error
, &result
);
690 /* Get the current PCI slot state. This can take a long time,
691 * sometimes over 3 seconds for certain systems.
693 rc
= eeh_ops
->wait_state(pe
, MAX_WAIT_FOR_RECOVERY
*1000);
694 if (rc
< 0 || rc
== EEH_STATE_NOT_SUPPORT
) {
695 pr_warn("EEH: Permanent failure\n");
699 /* Since rtas may enable MMIO when posting the error log,
700 * don't post the error log until after all dev drivers
701 * have been informed.
703 pr_info("EEH: Collect temporary log\n");
704 eeh_slot_error_detail(pe
, EEH_LOG_TEMP
);
706 /* If all device drivers were EEH-unaware, then shut
707 * down all of the device drivers, and hope they
708 * go down willingly, without panicing the system.
710 if (result
== PCI_ERS_RESULT_NONE
) {
711 pr_info("EEH: Reset with hotplug activity\n");
712 rc
= eeh_reset_device(pe
, frozen_bus
);
714 pr_warn("%s: Unable to reset, err=%d\n",
720 /* If all devices reported they can proceed, then re-enable MMIO */
721 if (result
== PCI_ERS_RESULT_CAN_RECOVER
) {
722 pr_info("EEH: Enable I/O for affected devices\n");
723 rc
= eeh_pci_enable(pe
, EEH_OPT_THAW_MMIO
);
728 result
= PCI_ERS_RESULT_NEED_RESET
;
730 pr_info("EEH: Notify device drivers to resume I/O\n");
731 eeh_pe_dev_traverse(pe
, eeh_report_mmio_enabled
, &result
);
735 /* If all devices reported they can proceed, then re-enable DMA */
736 if (result
== PCI_ERS_RESULT_CAN_RECOVER
) {
737 pr_info("EEH: Enabled DMA for affected devices\n");
738 rc
= eeh_pci_enable(pe
, EEH_OPT_THAW_DMA
);
743 result
= PCI_ERS_RESULT_NEED_RESET
;
746 * We didn't do PE reset for the case. The PE
747 * is still in frozen state. Clear it before
750 eeh_pe_state_clear(pe
, EEH_PE_ISOLATED
);
751 result
= PCI_ERS_RESULT_RECOVERED
;
755 /* If any device has a hard failure, then shut off everything. */
756 if (result
== PCI_ERS_RESULT_DISCONNECT
) {
757 pr_warn("EEH: Device driver gave up\n");
761 /* If any device called out for a reset, then reset the slot */
762 if (result
== PCI_ERS_RESULT_NEED_RESET
) {
763 pr_info("EEH: Reset without hotplug activity\n");
764 rc
= eeh_reset_device(pe
, NULL
);
766 pr_warn("%s: Cannot reset, err=%d\n",
771 pr_info("EEH: Notify device drivers "
772 "the completion of reset\n");
773 result
= PCI_ERS_RESULT_NONE
;
774 eeh_pe_dev_traverse(pe
, eeh_report_reset
, &result
);
777 /* All devices should claim they have recovered by now. */
778 if ((result
!= PCI_ERS_RESULT_RECOVERED
) &&
779 (result
!= PCI_ERS_RESULT_NONE
)) {
780 pr_warn("EEH: Not recovered\n");
784 /* Tell all device drivers that they can resume operations */
785 pr_info("EEH: Notify device driver to resume\n");
786 eeh_pe_dev_traverse(pe
, eeh_report_resume
, NULL
);
792 * About 90% of all real-life EEH failures in the field
793 * are due to poorly seated PCI cards. Only 10% or so are
794 * due to actual, failed cards.
796 pr_err("EEH: PHB#%d-PE#%x has failed %d times in the\n"
797 "last hour and has been permanently disabled.\n"
798 "Please try reseating or replacing it.\n",
799 pe
->phb
->global_number
, pe
->addr
,
804 pr_err("EEH: Unable to recover from failure from PHB#%d-PE#%x.\n"
805 "Please try reseating or replacing it\n",
806 pe
->phb
->global_number
, pe
->addr
);
809 eeh_slot_error_detail(pe
, EEH_LOG_PERM
);
811 /* Notify all devices that they're about to go down. */
812 eeh_pe_dev_traverse(pe
, eeh_report_failure
, NULL
);
814 /* Mark the PE to be removed permanently */
815 pe
->freeze_count
= EEH_MAX_ALLOWED_FREEZES
+ 1;
818 * Shut down the device drivers for good. We mark
819 * all removed devices correctly to avoid access
820 * the their PCI config any more.
823 eeh_pe_dev_mode_mark(pe
, EEH_DEV_REMOVED
);
825 pci_lock_rescan_remove();
826 pcibios_remove_pci_devices(frozen_bus
);
827 pci_unlock_rescan_remove();
831 static void eeh_handle_special_event(void)
833 struct eeh_pe
*pe
, *phb_pe
;
835 struct pci_controller
*hose
;
841 rc
= eeh_ops
->next_error(&pe
);
844 case EEH_NEXT_ERR_DEAD_IOC
:
845 /* Mark all PHBs in dead state */
846 eeh_serialize_lock(&flags
);
848 /* Purge all events */
849 eeh_remove_event(NULL
, true);
851 list_for_each_entry(hose
, &hose_list
, list_node
) {
852 phb_pe
= eeh_phb_pe_get(hose
);
853 if (!phb_pe
) continue;
855 eeh_pe_state_mark(phb_pe
, EEH_PE_ISOLATED
);
858 eeh_serialize_unlock(flags
);
861 case EEH_NEXT_ERR_FROZEN_PE
:
862 case EEH_NEXT_ERR_FENCED_PHB
:
863 case EEH_NEXT_ERR_DEAD_PHB
:
864 /* Mark the PE in fenced state */
865 eeh_serialize_lock(&flags
);
867 /* Purge all events of the PHB */
868 eeh_remove_event(pe
, true);
870 if (rc
== EEH_NEXT_ERR_DEAD_PHB
)
871 eeh_pe_state_mark(pe
, EEH_PE_ISOLATED
);
873 eeh_pe_state_mark(pe
,
874 EEH_PE_ISOLATED
| EEH_PE_RECOVERING
);
876 eeh_serialize_unlock(flags
);
879 case EEH_NEXT_ERR_NONE
:
882 pr_warn("%s: Invalid value %d from next_error()\n",
888 * For fenced PHB and frozen PE, it's handled as normal
889 * event. We have to remove the affected PHBs for dead
892 if (rc
== EEH_NEXT_ERR_FROZEN_PE
||
893 rc
== EEH_NEXT_ERR_FENCED_PHB
) {
894 eeh_handle_normal_event(pe
);
895 eeh_pe_state_clear(pe
, EEH_PE_RECOVERING
);
897 pci_lock_rescan_remove();
898 list_for_each_entry(hose
, &hose_list
, list_node
) {
899 phb_pe
= eeh_phb_pe_get(hose
);
901 !(phb_pe
->state
& EEH_PE_ISOLATED
) ||
902 (phb_pe
->state
& EEH_PE_RECOVERING
))
905 /* Notify all devices to be down */
906 bus
= eeh_pe_bus_get(phb_pe
);
907 eeh_pe_dev_traverse(pe
,
908 eeh_report_failure
, NULL
);
909 pcibios_remove_pci_devices(bus
);
911 pci_unlock_rescan_remove();
915 * If we have detected dead IOC, we needn't proceed
916 * any more since all PHBs would have been removed
918 if (rc
== EEH_NEXT_ERR_DEAD_IOC
)
920 } while (rc
!= EEH_NEXT_ERR_NONE
);
924 * eeh_handle_event - Reset a PCI device after hard lockup.
927 * While PHB detects address or data parity errors on particular PCI
928 * slot, the associated PE will be frozen. Besides, DMA's occurring
929 * to wild addresses (which usually happen due to bugs in device
930 * drivers or in PCI adapter firmware) can cause EEH error. #SERR,
931 * #PERR or other misc PCI-related errors also can trigger EEH errors.
933 * Recovery process consists of unplugging the device driver (which
934 * generated hotplug events to userspace), then issuing a PCI #RST to
935 * the device, then reconfiguring the PCI config space for all bridges
936 * & devices under this slot, and then finally restarting the device
937 * drivers (which cause a second set of hotplug events to go out to
940 void eeh_handle_event(struct eeh_pe
*pe
)
943 eeh_handle_normal_event(pe
);
945 eeh_handle_special_event();