| /* SPDX-License-Identifier: GPL-2.0-or-later */ |
| #ifndef _ASM_X86_ACPI_H |
| #define _ASM_X86_ACPI_H |
| |
| /* |
| * Copyright (C) 2001 Paul Diefenbaugh <paul.s.diefenbaugh@intel.com> |
| * Copyright (C) 2001 Patrick Mochel <mochel@osdl.org> |
| */ |
| #include <acpi/pdc_intel.h> |
| |
| #include <asm/numa.h> |
| #include <asm/fixmap.h> |
| #include <asm/processor.h> |
| #include <asm/mmu.h> |
| #include <asm/mpspec.h> |
| #include <asm/x86_init.h> |
| |
| #ifdef CONFIG_ACPI_APEI |
| # include <asm/pgtable_types.h> |
| #endif |
| |
| #ifdef CONFIG_ACPI |
| extern int acpi_lapic; |
| extern int acpi_ioapic; |
| extern int acpi_noirq; |
| extern int acpi_strict; |
| extern int acpi_disabled; |
| extern int acpi_pci_disabled; |
| extern int acpi_skip_timer_override; |
| extern int acpi_use_timer_override; |
| extern int acpi_fix_pin2_polarity; |
| extern int acpi_disable_cmcff; |
| |
| extern u8 acpi_sci_flags; |
| extern u32 acpi_sci_override_gsi; |
| void acpi_pic_sci_set_trigger(unsigned int, u16); |
| |
| struct device; |
| |
| extern int (*__acpi_register_gsi)(struct device *dev, u32 gsi, |
| int trigger, int polarity); |
| extern void (*__acpi_unregister_gsi)(u32 gsi); |
| |
| static inline void disable_acpi(void) |
| { |
| acpi_disabled = 1; |
| acpi_pci_disabled = 1; |
| acpi_noirq = 1; |
| } |
| |
| extern int acpi_gsi_to_irq(u32 gsi, unsigned int *irq); |
| |
| static inline void acpi_noirq_set(void) { acpi_noirq = 1; } |
| static inline void acpi_disable_pci(void) |
| { |
| acpi_pci_disabled = 1; |
| acpi_noirq_set(); |
| } |
| |
| /* Low-level suspend routine. */ |
| extern int (*acpi_suspend_lowlevel)(void); |
| |
| /* Physical address to resume after wakeup */ |
| unsigned long acpi_get_wakeup_address(void); |
| |
| /* |
| * Check if the CPU can handle C2 and deeper |
| */ |
| static inline unsigned int acpi_processor_cstate_check(unsigned int max_cstate) |
| { |
| /* |
| * Early models (<=5) of AMD Opterons are not supposed to go into |
| * C2 state. |
| * |
| * Steppings 0x0A and later are good |
| */ |
| if (boot_cpu_data.x86 == 0x0F && |
| boot_cpu_data.x86_vendor == X86_VENDOR_AMD && |
| boot_cpu_data.x86_model <= 0x05 && |
| boot_cpu_data.x86_stepping < 0x0A) |
| return 1; |
| else if (boot_cpu_has(X86_BUG_AMD_APIC_C1E)) |
| return 1; |
| else |
| return max_cstate; |
| } |
| |
| static inline bool arch_has_acpi_pdc(void) |
| { |
| struct cpuinfo_x86 *c = &cpu_data(0); |
| return (c->x86_vendor == X86_VENDOR_INTEL || |
| c->x86_vendor == X86_VENDOR_CENTAUR); |
| } |
| |
| static inline void arch_acpi_set_pdc_bits(u32 *buf) |
| { |
| struct cpuinfo_x86 *c = &cpu_data(0); |
| |
| buf[2] |= ACPI_PDC_C_CAPABILITY_SMP; |
| |
| if (cpu_has(c, X86_FEATURE_EST)) |
| buf[2] |= ACPI_PDC_EST_CAPABILITY_SWSMP; |
| |
| if (cpu_has(c, X86_FEATURE_ACPI)) |
| buf[2] |= ACPI_PDC_T_FFH; |
| |
| /* |
| * If mwait/monitor is unsupported, C2/C3_FFH will be disabled |
| */ |
| if (!cpu_has(c, X86_FEATURE_MWAIT)) |
| buf[2] &= ~(ACPI_PDC_C_C2C3_FFH); |
| } |
| |
| static inline bool acpi_has_cpu_in_madt(void) |
| { |
| return !!acpi_lapic; |
| } |
| |
| #define ACPI_HAVE_ARCH_SET_ROOT_POINTER |
| static inline void acpi_arch_set_root_pointer(u64 addr) |
| { |
| x86_init.acpi.set_root_pointer(addr); |
| } |
| |
| #define ACPI_HAVE_ARCH_GET_ROOT_POINTER |
| static inline u64 acpi_arch_get_root_pointer(void) |
| { |
| return x86_init.acpi.get_root_pointer(); |
| } |
| |
| void acpi_generic_reduced_hw_init(void); |
| |
| void x86_default_set_root_pointer(u64 addr); |
| u64 x86_default_get_root_pointer(void); |
| |
| #else /* !CONFIG_ACPI */ |
| |
| #define acpi_lapic 0 |
| #define acpi_ioapic 0 |
| #define acpi_disable_cmcff 0 |
| static inline void acpi_noirq_set(void) { } |
| static inline void acpi_disable_pci(void) { } |
| static inline void disable_acpi(void) { } |
| |
| static inline void acpi_generic_reduced_hw_init(void) { } |
| |
| static inline void x86_default_set_root_pointer(u64 addr) { } |
| |
| static inline u64 x86_default_get_root_pointer(void) |
| { |
| return 0; |
| } |
| |
| #endif /* !CONFIG_ACPI */ |
| |
| #define ARCH_HAS_POWER_INIT 1 |
| |
| #ifdef CONFIG_ACPI_NUMA |
| extern int x86_acpi_numa_init(void); |
| #endif /* CONFIG_ACPI_NUMA */ |
| |
| struct cper_ia_proc_ctx; |
| |
| #ifdef CONFIG_ACPI_APEI |
| static inline pgprot_t arch_apei_get_mem_attribute(phys_addr_t addr) |
| { |
| /* |
| * We currently have no way to look up the EFI memory map |
| * attributes for a region in a consistent way, because the |
| * memmap is discarded after efi_free_boot_services(). So if |
| * you call efi_mem_attributes() during boot and at runtime, |
| * you could theoretically see different attributes. |
| * |
| * We are yet to see any x86 platforms that require anything |
| * other than PAGE_KERNEL (some ARM64 platforms require the |
| * equivalent of PAGE_KERNEL_NOCACHE). Additionally, if SME |
| * is active, the ACPI information will not be encrypted, |
| * so return PAGE_KERNEL_NOENC until we know differently. |
| */ |
| return PAGE_KERNEL_NOENC; |
| } |
| |
| int arch_apei_report_x86_error(struct cper_ia_proc_ctx *ctx_info, |
| u64 lapic_id); |
| #else |
| static inline int arch_apei_report_x86_error(struct cper_ia_proc_ctx *ctx_info, |
| u64 lapic_id) |
| { |
| return -EINVAL; |
| } |
| #endif |
| |
| #define ACPI_TABLE_UPGRADE_MAX_PHYS (max_low_pfn_mapped << PAGE_SHIFT) |
| |
| #endif /* _ASM_X86_ACPI_H */ |