1 /* SPDX-License-Identifier: GPL-2.0-or-later */
2 #ifndef _ASM_X86_ACPI_H
3 #define _ASM_X86_ACPI_H
6 * Copyright (C) 2001 Paul Diefenbaugh <paul.s.diefenbaugh@intel.com>
7 * Copyright (C) 2001 Patrick Mochel <mochel@osdl.org>
9 #include <acpi/pdc_intel.h>
12 #include <asm/fixmap.h>
13 #include <asm/processor.h>
15 #include <asm/mpspec.h>
16 #include <asm/x86_init.h>
18 #ifdef CONFIG_ACPI_APEI
19 # include <asm/pgtable_types.h>
23 extern int acpi_lapic
;
24 extern int acpi_ioapic
;
25 extern int acpi_noirq
;
26 extern int acpi_strict
;
27 extern int acpi_disabled
;
28 extern int acpi_pci_disabled
;
29 extern int acpi_skip_timer_override
;
30 extern int acpi_use_timer_override
;
31 extern int acpi_fix_pin2_polarity
;
32 extern int acpi_disable_cmcff
;
34 extern u8 acpi_sci_flags
;
35 extern u32 acpi_sci_override_gsi
;
36 void acpi_pic_sci_set_trigger(unsigned int, u16
);
40 extern int (*__acpi_register_gsi
)(struct device
*dev
, u32 gsi
,
41 int trigger
, int polarity
);
42 extern void (*__acpi_unregister_gsi
)(u32 gsi
);
44 static inline void disable_acpi(void)
47 acpi_pci_disabled
= 1;
51 extern int acpi_gsi_to_irq(u32 gsi
, unsigned int *irq
);
53 static inline void acpi_noirq_set(void) { acpi_noirq
= 1; }
54 static inline void acpi_disable_pci(void)
56 acpi_pci_disabled
= 1;
60 /* Low-level suspend routine. */
61 extern int (*acpi_suspend_lowlevel
)(void);
63 /* Physical address to resume after wakeup */
64 unsigned long acpi_get_wakeup_address(void);
67 * Check if the CPU can handle C2 and deeper
69 static inline unsigned int acpi_processor_cstate_check(unsigned int max_cstate
)
72 * Early models (<=5) of AMD Opterons are not supposed to go into
75 * Steppings 0x0A and later are good
77 if (boot_cpu_data
.x86
== 0x0F &&
78 boot_cpu_data
.x86_vendor
== X86_VENDOR_AMD
&&
79 boot_cpu_data
.x86_model
<= 0x05 &&
80 boot_cpu_data
.x86_stepping
< 0x0A)
82 else if (boot_cpu_has(X86_BUG_AMD_APIC_C1E
))
88 static inline bool arch_has_acpi_pdc(void)
90 struct cpuinfo_x86
*c
= &cpu_data(0);
91 return (c
->x86_vendor
== X86_VENDOR_INTEL
||
92 c
->x86_vendor
== X86_VENDOR_CENTAUR
);
95 static inline void arch_acpi_set_pdc_bits(u32
*buf
)
97 struct cpuinfo_x86
*c
= &cpu_data(0);
99 buf
[2] |= ACPI_PDC_C_CAPABILITY_SMP
;
101 if (cpu_has(c
, X86_FEATURE_EST
))
102 buf
[2] |= ACPI_PDC_EST_CAPABILITY_SWSMP
;
104 if (cpu_has(c
, X86_FEATURE_ACPI
))
105 buf
[2] |= ACPI_PDC_T_FFH
;
108 * If mwait/monitor is unsupported, C2/C3_FFH will be disabled
110 if (!cpu_has(c
, X86_FEATURE_MWAIT
))
111 buf
[2] &= ~(ACPI_PDC_C_C2C3_FFH
);
114 static inline bool acpi_has_cpu_in_madt(void)
119 #define ACPI_HAVE_ARCH_SET_ROOT_POINTER
120 static inline void acpi_arch_set_root_pointer(u64 addr
)
122 x86_init
.acpi
.set_root_pointer(addr
);
125 #define ACPI_HAVE_ARCH_GET_ROOT_POINTER
126 static inline u64
acpi_arch_get_root_pointer(void)
128 return x86_init
.acpi
.get_root_pointer();
131 void acpi_generic_reduced_hw_init(void);
133 void x86_default_set_root_pointer(u64 addr
);
134 u64
x86_default_get_root_pointer(void);
136 #else /* !CONFIG_ACPI */
139 #define acpi_ioapic 0
140 #define acpi_disable_cmcff 0
141 static inline void acpi_noirq_set(void) { }
142 static inline void acpi_disable_pci(void) { }
143 static inline void disable_acpi(void) { }
145 static inline void acpi_generic_reduced_hw_init(void) { }
147 static inline void x86_default_set_root_pointer(u64 addr
) { }
149 static inline u64
x86_default_get_root_pointer(void)
154 #endif /* !CONFIG_ACPI */
156 #define ARCH_HAS_POWER_INIT 1
158 #ifdef CONFIG_ACPI_NUMA
159 extern int x86_acpi_numa_init(void);
160 #endif /* CONFIG_ACPI_NUMA */
162 #define acpi_unlazy_tlb(x) leave_mm(x)
164 #ifdef CONFIG_ACPI_APEI
165 static inline pgprot_t
arch_apei_get_mem_attribute(phys_addr_t addr
)
168 * We currently have no way to look up the EFI memory map
169 * attributes for a region in a consistent way, because the
170 * memmap is discarded after efi_free_boot_services(). So if
171 * you call efi_mem_attributes() during boot and at runtime,
172 * you could theoretically see different attributes.
174 * We are yet to see any x86 platforms that require anything
175 * other than PAGE_KERNEL (some ARM64 platforms require the
176 * equivalent of PAGE_KERNEL_NOCACHE). Additionally, if SME
177 * is active, the ACPI information will not be encrypted,
178 * so return PAGE_KERNEL_NOENC until we know differently.
180 return PAGE_KERNEL_NOENC
;
184 #define ACPI_TABLE_UPGRADE_MAX_PHYS (max_low_pfn_mapped << PAGE_SHIFT)
186 #endif /* _ASM_X86_ACPI_H */