| // SPDX-License-Identifier: GPL-2.0-only |
| /* |
| * OMAP WakeupGen Source file |
| * |
| * OMAP WakeupGen is the interrupt controller extension used along |
| * with ARM GIC to wake the CPU out from low power states on |
| * external interrupts. It is responsible for generating wakeup |
| * event from the incoming interrupts and enable bits. It is |
| * implemented in MPU always ON power domain. During normal operation, |
| * WakeupGen delivers external interrupts directly to the GIC. |
| * |
| * Copyright (C) 2011 Texas Instruments, Inc. |
| * Santosh Shilimkar <santosh.shilimkar@ti.com> |
| */ |
| |
| #include <linux/kernel.h> |
| #include <linux/init.h> |
| #include <linux/io.h> |
| #include <linux/irq.h> |
| #include <linux/irqchip.h> |
| #include <linux/irqdomain.h> |
| #include <linux/of_address.h> |
| #include <linux/platform_device.h> |
| #include <linux/cpu.h> |
| #include <linux/notifier.h> |
| #include <linux/cpu_pm.h> |
| |
| #include "omap-wakeupgen.h" |
| #include "omap-secure.h" |
| |
| #include "soc.h" |
| #include "omap4-sar-layout.h" |
| #include "common.h" |
| #include "pm.h" |
| |
| #define AM43XX_NR_REG_BANKS 7 |
| #define AM43XX_IRQS 224 |
| #define MAX_NR_REG_BANKS AM43XX_NR_REG_BANKS |
| #define MAX_IRQS AM43XX_IRQS |
| #define DEFAULT_NR_REG_BANKS 5 |
| #define DEFAULT_IRQS 160 |
| #define WKG_MASK_ALL 0x00000000 |
| #define WKG_UNMASK_ALL 0xffffffff |
| #define CPU_ENA_OFFSET 0x400 |
| #define CPU0_ID 0x0 |
| #define CPU1_ID 0x1 |
| #define OMAP4_NR_BANKS 4 |
| #define OMAP4_NR_IRQS 128 |
| |
| #define SYS_NIRQ1_EXT_SYS_IRQ_1 7 |
| #define SYS_NIRQ2_EXT_SYS_IRQ_2 119 |
| |
| static void __iomem *wakeupgen_base; |
| static void __iomem *sar_base; |
| static DEFINE_RAW_SPINLOCK(wakeupgen_lock); |
| static unsigned int irq_target_cpu[MAX_IRQS]; |
| static unsigned int irq_banks = DEFAULT_NR_REG_BANKS; |
| static unsigned int max_irqs = DEFAULT_IRQS; |
| static unsigned int omap_secure_apis; |
| |
| #ifdef CONFIG_CPU_PM |
| static unsigned int wakeupgen_context[MAX_NR_REG_BANKS]; |
| #endif |
| |
| struct omap_wakeupgen_ops { |
| void (*save_context)(void); |
| void (*restore_context)(void); |
| }; |
| |
| static struct omap_wakeupgen_ops *wakeupgen_ops; |
| |
| /* |
| * Static helper functions. |
| */ |
| static inline u32 wakeupgen_readl(u8 idx, u32 cpu) |
| { |
| return readl_relaxed(wakeupgen_base + OMAP_WKG_ENB_A_0 + |
| (cpu * CPU_ENA_OFFSET) + (idx * 4)); |
| } |
| |
| static inline void wakeupgen_writel(u32 val, u8 idx, u32 cpu) |
| { |
| writel_relaxed(val, wakeupgen_base + OMAP_WKG_ENB_A_0 + |
| (cpu * CPU_ENA_OFFSET) + (idx * 4)); |
| } |
| |
| static inline void sar_writel(u32 val, u32 offset, u8 idx) |
| { |
| writel_relaxed(val, sar_base + offset + (idx * 4)); |
| } |
| |
| static inline int _wakeupgen_get_irq_info(u32 irq, u32 *bit_posn, u8 *reg_index) |
| { |
| /* |
| * Each WakeupGen register controls 32 interrupt. |
| * i.e. 1 bit per SPI IRQ |
| */ |
| *reg_index = irq >> 5; |
| *bit_posn = irq %= 32; |
| |
| return 0; |
| } |
| |
| static void _wakeupgen_clear(unsigned int irq, unsigned int cpu) |
| { |
| u32 val, bit_number; |
| u8 i; |
| |
| if (_wakeupgen_get_irq_info(irq, &bit_number, &i)) |
| return; |
| |
| val = wakeupgen_readl(i, cpu); |
| val &= ~BIT(bit_number); |
| wakeupgen_writel(val, i, cpu); |
| } |
| |
| static void _wakeupgen_set(unsigned int irq, unsigned int cpu) |
| { |
| u32 val, bit_number; |
| u8 i; |
| |
| if (_wakeupgen_get_irq_info(irq, &bit_number, &i)) |
| return; |
| |
| val = wakeupgen_readl(i, cpu); |
| val |= BIT(bit_number); |
| wakeupgen_writel(val, i, cpu); |
| } |
| |
| /* |
| * Architecture specific Mask extension |
| */ |
| static void wakeupgen_mask(struct irq_data *d) |
| { |
| unsigned long flags; |
| |
| raw_spin_lock_irqsave(&wakeupgen_lock, flags); |
| _wakeupgen_clear(d->hwirq, irq_target_cpu[d->hwirq]); |
| raw_spin_unlock_irqrestore(&wakeupgen_lock, flags); |
| irq_chip_mask_parent(d); |
| } |
| |
| /* |
| * Architecture specific Unmask extension |
| */ |
| static void wakeupgen_unmask(struct irq_data *d) |
| { |
| unsigned long flags; |
| |
| raw_spin_lock_irqsave(&wakeupgen_lock, flags); |
| _wakeupgen_set(d->hwirq, irq_target_cpu[d->hwirq]); |
| raw_spin_unlock_irqrestore(&wakeupgen_lock, flags); |
| irq_chip_unmask_parent(d); |
| } |
| |
| /* |
| * The sys_nirq pins bypass peripheral modules and are wired directly |
| * to MPUSS wakeupgen. They get automatically inverted for GIC. |
| */ |
| static int wakeupgen_irq_set_type(struct irq_data *d, unsigned int type) |
| { |
| bool inverted = false; |
| |
| switch (type) { |
| case IRQ_TYPE_LEVEL_LOW: |
| type &= ~IRQ_TYPE_LEVEL_MASK; |
| type |= IRQ_TYPE_LEVEL_HIGH; |
| inverted = true; |
| break; |
| case IRQ_TYPE_EDGE_FALLING: |
| type &= ~IRQ_TYPE_EDGE_BOTH; |
| type |= IRQ_TYPE_EDGE_RISING; |
| inverted = true; |
| break; |
| default: |
| break; |
| } |
| |
| if (inverted && d->hwirq != SYS_NIRQ1_EXT_SYS_IRQ_1 && |
| d->hwirq != SYS_NIRQ2_EXT_SYS_IRQ_2) |
| pr_warn("wakeupgen: irq%li polarity inverted in dts\n", |
| d->hwirq); |
| |
| return irq_chip_set_type_parent(d, type); |
| } |
| |
| #ifdef CONFIG_HOTPLUG_CPU |
| static DEFINE_PER_CPU(u32 [MAX_NR_REG_BANKS], irqmasks); |
| |
| static void _wakeupgen_save_masks(unsigned int cpu) |
| { |
| u8 i; |
| |
| for (i = 0; i < irq_banks; i++) |
| per_cpu(irqmasks, cpu)[i] = wakeupgen_readl(i, cpu); |
| } |
| |
| static void _wakeupgen_restore_masks(unsigned int cpu) |
| { |
| u8 i; |
| |
| for (i = 0; i < irq_banks; i++) |
| wakeupgen_writel(per_cpu(irqmasks, cpu)[i], i, cpu); |
| } |
| |
| static void _wakeupgen_set_all(unsigned int cpu, unsigned int reg) |
| { |
| u8 i; |
| |
| for (i = 0; i < irq_banks; i++) |
| wakeupgen_writel(reg, i, cpu); |
| } |
| |
| /* |
| * Mask or unmask all interrupts on given CPU. |
| * 0 = Mask all interrupts on the 'cpu' |
| * 1 = Unmask all interrupts on the 'cpu' |
| * Ensure that the initial mask is maintained. This is faster than |
| * iterating through GIC registers to arrive at the correct masks. |
| */ |
| static void wakeupgen_irqmask_all(unsigned int cpu, unsigned int set) |
| { |
| unsigned long flags; |
| |
| raw_spin_lock_irqsave(&wakeupgen_lock, flags); |
| if (set) { |
| _wakeupgen_save_masks(cpu); |
| _wakeupgen_set_all(cpu, WKG_MASK_ALL); |
| } else { |
| _wakeupgen_set_all(cpu, WKG_UNMASK_ALL); |
| _wakeupgen_restore_masks(cpu); |
| } |
| raw_spin_unlock_irqrestore(&wakeupgen_lock, flags); |
| } |
| #endif |
| |
| #ifdef CONFIG_CPU_PM |
| static inline void omap4_irq_save_context(void) |
| { |
| u32 i, val; |
| |
| if (omap_rev() == OMAP4430_REV_ES1_0) |
| return; |
| |
| for (i = 0; i < irq_banks; i++) { |
| /* Save the CPUx interrupt mask for IRQ 0 to 127 */ |
| val = wakeupgen_readl(i, 0); |
| sar_writel(val, WAKEUPGENENB_OFFSET_CPU0, i); |
| val = wakeupgen_readl(i, 1); |
| sar_writel(val, WAKEUPGENENB_OFFSET_CPU1, i); |
| |
| /* |
| * Disable the secure interrupts for CPUx. The restore |
| * code blindly restores secure and non-secure interrupt |
| * masks from SAR RAM. Secure interrupts are not suppose |
| * to be enabled from HLOS. So overwrite the SAR location |
| * so that the secure interrupt remains disabled. |
| */ |
| sar_writel(0x0, WAKEUPGENENB_SECURE_OFFSET_CPU0, i); |
| sar_writel(0x0, WAKEUPGENENB_SECURE_OFFSET_CPU1, i); |
| } |
| |
| /* Save AuxBoot* registers */ |
| val = readl_relaxed(wakeupgen_base + OMAP_AUX_CORE_BOOT_0); |
| writel_relaxed(val, sar_base + AUXCOREBOOT0_OFFSET); |
| val = readl_relaxed(wakeupgen_base + OMAP_AUX_CORE_BOOT_1); |
| writel_relaxed(val, sar_base + AUXCOREBOOT1_OFFSET); |
| |
| /* Save SyncReq generation logic */ |
| val = readl_relaxed(wakeupgen_base + OMAP_PTMSYNCREQ_MASK); |
| writel_relaxed(val, sar_base + PTMSYNCREQ_MASK_OFFSET); |
| val = readl_relaxed(wakeupgen_base + OMAP_PTMSYNCREQ_EN); |
| writel_relaxed(val, sar_base + PTMSYNCREQ_EN_OFFSET); |
| |
| /* Set the Backup Bit Mask status */ |
| val = readl_relaxed(sar_base + SAR_BACKUP_STATUS_OFFSET); |
| val |= SAR_BACKUP_STATUS_WAKEUPGEN; |
| writel_relaxed(val, sar_base + SAR_BACKUP_STATUS_OFFSET); |
| |
| } |
| |
| static inline void omap5_irq_save_context(void) |
| { |
| u32 i, val; |
| |
| for (i = 0; i < irq_banks; i++) { |
| /* Save the CPUx interrupt mask for IRQ 0 to 159 */ |
| val = wakeupgen_readl(i, 0); |
| sar_writel(val, OMAP5_WAKEUPGENENB_OFFSET_CPU0, i); |
| val = wakeupgen_readl(i, 1); |
| sar_writel(val, OMAP5_WAKEUPGENENB_OFFSET_CPU1, i); |
| sar_writel(0x0, OMAP5_WAKEUPGENENB_SECURE_OFFSET_CPU0, i); |
| sar_writel(0x0, OMAP5_WAKEUPGENENB_SECURE_OFFSET_CPU1, i); |
| } |
| |
| /* Save AuxBoot* registers */ |
| val = readl_relaxed(wakeupgen_base + OMAP_AUX_CORE_BOOT_0); |
| writel_relaxed(val, sar_base + OMAP5_AUXCOREBOOT0_OFFSET); |
| val = readl_relaxed(wakeupgen_base + OMAP_AUX_CORE_BOOT_0); |
| writel_relaxed(val, sar_base + OMAP5_AUXCOREBOOT1_OFFSET); |
| |
| /* Set the Backup Bit Mask status */ |
| val = readl_relaxed(sar_base + OMAP5_SAR_BACKUP_STATUS_OFFSET); |
| val |= SAR_BACKUP_STATUS_WAKEUPGEN; |
| writel_relaxed(val, sar_base + OMAP5_SAR_BACKUP_STATUS_OFFSET); |
| |
| } |
| |
| static inline void am43xx_irq_save_context(void) |
| { |
| u32 i; |
| |
| for (i = 0; i < irq_banks; i++) { |
| wakeupgen_context[i] = wakeupgen_readl(i, 0); |
| wakeupgen_writel(0, i, CPU0_ID); |
| } |
| } |
| |
| /* |
| * Save WakeupGen interrupt context in SAR BANK3. Restore is done by |
| * ROM code. WakeupGen IP is integrated along with GIC to manage the |
| * interrupt wakeups from CPU low power states. It manages |
| * masking/unmasking of Shared peripheral interrupts(SPI). So the |
| * interrupt enable/disable control should be in sync and consistent |
| * at WakeupGen and GIC so that interrupts are not lost. |
| */ |
| static void irq_save_context(void) |
| { |
| /* DRA7 has no SAR to save */ |
| if (soc_is_dra7xx()) |
| return; |
| |
| if (wakeupgen_ops && wakeupgen_ops->save_context) |
| wakeupgen_ops->save_context(); |
| } |
| |
| /* |
| * Clear WakeupGen SAR backup status. |
| */ |
| static void irq_sar_clear(void) |
| { |
| u32 val; |
| u32 offset = SAR_BACKUP_STATUS_OFFSET; |
| /* DRA7 has no SAR to save */ |
| if (soc_is_dra7xx()) |
| return; |
| |
| if (soc_is_omap54xx()) |
| offset = OMAP5_SAR_BACKUP_STATUS_OFFSET; |
| |
| val = readl_relaxed(sar_base + offset); |
| val &= ~SAR_BACKUP_STATUS_WAKEUPGEN; |
| writel_relaxed(val, sar_base + offset); |
| } |
| |
| static void am43xx_irq_restore_context(void) |
| { |
| u32 i; |
| |
| for (i = 0; i < irq_banks; i++) |
| wakeupgen_writel(wakeupgen_context[i], i, CPU0_ID); |
| } |
| |
| static void irq_restore_context(void) |
| { |
| if (wakeupgen_ops && wakeupgen_ops->restore_context) |
| wakeupgen_ops->restore_context(); |
| } |
| |
| /* |
| * Save GIC and Wakeupgen interrupt context using secure API |
| * for HS/EMU devices. |
| */ |
| static void irq_save_secure_context(void) |
| { |
| u32 ret; |
| ret = omap_secure_dispatcher(OMAP4_HAL_SAVEGIC_INDEX, |
| FLAG_START_CRITICAL, |
| 0, 0, 0, 0, 0); |
| if (ret != API_HAL_RET_VALUE_OK) |
| pr_err("GIC and Wakeupgen context save failed\n"); |
| } |
| |
| /* Define ops for context save and restore for each SoC */ |
| static struct omap_wakeupgen_ops omap4_wakeupgen_ops = { |
| .save_context = omap4_irq_save_context, |
| .restore_context = irq_sar_clear, |
| }; |
| |
| static struct omap_wakeupgen_ops omap5_wakeupgen_ops = { |
| .save_context = omap5_irq_save_context, |
| .restore_context = irq_sar_clear, |
| }; |
| |
| static struct omap_wakeupgen_ops am43xx_wakeupgen_ops = { |
| .save_context = am43xx_irq_save_context, |
| .restore_context = am43xx_irq_restore_context, |
| }; |
| #else |
| static struct omap_wakeupgen_ops omap4_wakeupgen_ops = {}; |
| static struct omap_wakeupgen_ops omap5_wakeupgen_ops = {}; |
| static struct omap_wakeupgen_ops am43xx_wakeupgen_ops = {}; |
| #endif |
| |
| #ifdef CONFIG_HOTPLUG_CPU |
| static int omap_wakeupgen_cpu_online(unsigned int cpu) |
| { |
| wakeupgen_irqmask_all(cpu, 0); |
| return 0; |
| } |
| |
| static int omap_wakeupgen_cpu_dead(unsigned int cpu) |
| { |
| wakeupgen_irqmask_all(cpu, 1); |
| return 0; |
| } |
| |
| static void __init irq_hotplug_init(void) |
| { |
| cpuhp_setup_state_nocalls(CPUHP_AP_ONLINE_DYN, "arm/omap-wake:online", |
| omap_wakeupgen_cpu_online, NULL); |
| cpuhp_setup_state_nocalls(CPUHP_ARM_OMAP_WAKE_DEAD, |
| "arm/omap-wake:dead", NULL, |
| omap_wakeupgen_cpu_dead); |
| } |
| #else |
| static void __init irq_hotplug_init(void) |
| {} |
| #endif |
| |
| #ifdef CONFIG_CPU_PM |
| static int irq_notifier(struct notifier_block *self, unsigned long cmd, void *v) |
| { |
| switch (cmd) { |
| case CPU_CLUSTER_PM_ENTER: |
| if (omap_type() == OMAP2_DEVICE_TYPE_GP || soc_is_am43xx()) |
| irq_save_context(); |
| else |
| irq_save_secure_context(); |
| break; |
| case CPU_CLUSTER_PM_EXIT: |
| if (omap_type() == OMAP2_DEVICE_TYPE_GP || soc_is_am43xx()) |
| irq_restore_context(); |
| break; |
| } |
| return NOTIFY_OK; |
| } |
| |
| static struct notifier_block irq_notifier_block = { |
| .notifier_call = irq_notifier, |
| }; |
| |
| static void __init irq_pm_init(void) |
| { |
| /* FIXME: Remove this when MPU OSWR support is added */ |
| if (!IS_PM44XX_ERRATUM(PM_OMAP4_CPU_OSWR_DISABLE)) |
| cpu_pm_register_notifier(&irq_notifier_block); |
| } |
| #else |
| static void __init irq_pm_init(void) |
| {} |
| #endif |
| |
| void __iomem *omap_get_wakeupgen_base(void) |
| { |
| return wakeupgen_base; |
| } |
| |
| int omap_secure_apis_support(void) |
| { |
| return omap_secure_apis; |
| } |
| |
| static struct irq_chip wakeupgen_chip = { |
| .name = "WUGEN", |
| .irq_eoi = irq_chip_eoi_parent, |
| .irq_mask = wakeupgen_mask, |
| .irq_unmask = wakeupgen_unmask, |
| .irq_retrigger = irq_chip_retrigger_hierarchy, |
| .irq_set_type = wakeupgen_irq_set_type, |
| .flags = IRQCHIP_SKIP_SET_WAKE | IRQCHIP_MASK_ON_SUSPEND, |
| #ifdef CONFIG_SMP |
| .irq_set_affinity = irq_chip_set_affinity_parent, |
| #endif |
| }; |
| |
| static int wakeupgen_domain_translate(struct irq_domain *d, |
| struct irq_fwspec *fwspec, |
| unsigned long *hwirq, |
| unsigned int *type) |
| { |
| if (is_of_node(fwspec->fwnode)) { |
| if (fwspec->param_count != 3) |
| return -EINVAL; |
| |
| /* No PPI should point to this domain */ |
| if (fwspec->param[0] != 0) |
| return -EINVAL; |
| |
| *hwirq = fwspec->param[1]; |
| *type = fwspec->param[2]; |
| return 0; |
| } |
| |
| return -EINVAL; |
| } |
| |
| static int wakeupgen_domain_alloc(struct irq_domain *domain, |
| unsigned int virq, |
| unsigned int nr_irqs, void *data) |
| { |
| struct irq_fwspec *fwspec = data; |
| struct irq_fwspec parent_fwspec; |
| irq_hw_number_t hwirq; |
| int i; |
| |
| if (fwspec->param_count != 3) |
| return -EINVAL; /* Not GIC compliant */ |
| if (fwspec->param[0] != 0) |
| return -EINVAL; /* No PPI should point to this domain */ |
| |
| hwirq = fwspec->param[1]; |
| if (hwirq >= MAX_IRQS) |
| return -EINVAL; /* Can't deal with this */ |
| |
| for (i = 0; i < nr_irqs; i++) |
| irq_domain_set_hwirq_and_chip(domain, virq + i, hwirq + i, |
| &wakeupgen_chip, NULL); |
| |
| parent_fwspec = *fwspec; |
| parent_fwspec.fwnode = domain->parent->fwnode; |
| return irq_domain_alloc_irqs_parent(domain, virq, nr_irqs, |
| &parent_fwspec); |
| } |
| |
| static const struct irq_domain_ops wakeupgen_domain_ops = { |
| .translate = wakeupgen_domain_translate, |
| .alloc = wakeupgen_domain_alloc, |
| .free = irq_domain_free_irqs_common, |
| }; |
| |
| /* |
| * Initialise the wakeupgen module. |
| */ |
| static int __init wakeupgen_init(struct device_node *node, |
| struct device_node *parent) |
| { |
| struct irq_domain *parent_domain, *domain; |
| int i; |
| unsigned int boot_cpu = smp_processor_id(); |
| u32 val; |
| |
| if (!parent) { |
| pr_err("%pOF: no parent, giving up\n", node); |
| return -ENODEV; |
| } |
| |
| parent_domain = irq_find_host(parent); |
| if (!parent_domain) { |
| pr_err("%pOF: unable to obtain parent domain\n", node); |
| return -ENXIO; |
| } |
| /* Not supported on OMAP4 ES1.0 silicon */ |
| if (omap_rev() == OMAP4430_REV_ES1_0) { |
| WARN(1, "WakeupGen: Not supported on OMAP4430 ES1.0\n"); |
| return -EPERM; |
| } |
| |
| /* Static mapping, never released */ |
| wakeupgen_base = of_iomap(node, 0); |
| if (WARN_ON(!wakeupgen_base)) |
| return -ENOMEM; |
| |
| if (cpu_is_omap44xx()) { |
| irq_banks = OMAP4_NR_BANKS; |
| max_irqs = OMAP4_NR_IRQS; |
| omap_secure_apis = 1; |
| wakeupgen_ops = &omap4_wakeupgen_ops; |
| } else if (soc_is_omap54xx()) { |
| wakeupgen_ops = &omap5_wakeupgen_ops; |
| } else if (soc_is_am43xx()) { |
| irq_banks = AM43XX_NR_REG_BANKS; |
| max_irqs = AM43XX_IRQS; |
| wakeupgen_ops = &am43xx_wakeupgen_ops; |
| } |
| |
| domain = irq_domain_add_hierarchy(parent_domain, 0, max_irqs, |
| node, &wakeupgen_domain_ops, |
| NULL); |
| if (!domain) { |
| iounmap(wakeupgen_base); |
| return -ENOMEM; |
| } |
| |
| /* Clear all IRQ bitmasks at wakeupGen level */ |
| for (i = 0; i < irq_banks; i++) { |
| wakeupgen_writel(0, i, CPU0_ID); |
| if (!soc_is_am43xx()) |
| wakeupgen_writel(0, i, CPU1_ID); |
| } |
| |
| /* |
| * FIXME: Add support to set_smp_affinity() once the core |
| * GIC code has necessary hooks in place. |
| */ |
| |
| /* Associate all the IRQs to boot CPU like GIC init does. */ |
| for (i = 0; i < max_irqs; i++) |
| irq_target_cpu[i] = boot_cpu; |
| |
| /* |
| * Enables OMAP5 ES2 PM Mode using ES2_PM_MODE in AMBA_IF_MODE |
| * 0x0: ES1 behavior, CPU cores would enter and exit OFF mode together. |
| * 0x1: ES2 behavior, CPU cores are allowed to enter/exit OFF mode |
| * independently. |
| * This needs to be set one time thanks to always ON domain. |
| * |
| * We do not support ES1 behavior anymore. OMAP5 is assumed to be |
| * ES2.0, and the same is applicable for DRA7. |
| */ |
| if (soc_is_omap54xx() || soc_is_dra7xx()) { |
| val = __raw_readl(wakeupgen_base + OMAP_AMBA_IF_MODE); |
| val |= BIT(5); |
| omap_smc1(OMAP5_MON_AMBA_IF_INDEX, val); |
| } |
| |
| irq_hotplug_init(); |
| irq_pm_init(); |
| |
| sar_base = omap4_get_sar_ram_base(); |
| |
| return 0; |
| } |
| IRQCHIP_DECLARE(ti_wakeupgen, "ti,omap4-wugen-mpu", wakeupgen_init); |