| /* |
| * linux/arch/arm/mach-omap2/irq.c |
| * |
| * Interrupt handler for OMAP2 boards. |
| * |
| * Copyright (C) 2005 Nokia Corporation |
| * Author: Paul Mundt <paul.mundt@nokia.com> |
| * |
| * This file is subject to the terms and conditions of the GNU General Public |
| * License. See the file "COPYING" in the main directory of this archive |
| * for more details. |
| */ |
| #include <linux/kernel.h> |
| #include <linux/module.h> |
| #include <linux/init.h> |
| #include <linux/interrupt.h> |
| #include <linux/io.h> |
| |
| #include <asm/exception.h> |
| #include <linux/irqchip.h> |
| #include <linux/irqdomain.h> |
| #include <linux/of.h> |
| #include <linux/of_address.h> |
| #include <linux/of_irq.h> |
| |
| #include <linux/irqchip/irq-omap-intc.h> |
| |
| /* selected INTC register offsets */ |
| |
| #define INTC_REVISION 0x0000 |
| #define INTC_SYSCONFIG 0x0010 |
| #define INTC_SYSSTATUS 0x0014 |
| #define INTC_SIR 0x0040 |
| #define INTC_CONTROL 0x0048 |
| #define INTC_PROTECTION 0x004C |
| #define INTC_IDLE 0x0050 |
| #define INTC_THRESHOLD 0x0068 |
| #define INTC_MIR0 0x0084 |
| #define INTC_MIR_CLEAR0 0x0088 |
| #define INTC_MIR_SET0 0x008c |
| #define INTC_PENDING_IRQ0 0x0098 |
| #define INTC_PENDING_IRQ1 0x00b8 |
| #define INTC_PENDING_IRQ2 0x00d8 |
| #define INTC_PENDING_IRQ3 0x00f8 |
| #define INTC_ILR0 0x0100 |
| |
| #define ACTIVEIRQ_MASK 0x7f /* omap2/3 active interrupt bits */ |
| #define SPURIOUSIRQ_MASK (0x1ffffff << 7) |
| #define INTCPS_NR_ILR_REGS 128 |
| #define INTCPS_NR_MIR_REGS 4 |
| |
| #define INTC_IDLE_FUNCIDLE (1 << 0) |
| #define INTC_IDLE_TURBO (1 << 1) |
| |
| #define INTC_PROTECTION_ENABLE (1 << 0) |
| |
| struct omap_intc_regs { |
| u32 sysconfig; |
| u32 protection; |
| u32 idle; |
| u32 threshold; |
| u32 ilr[INTCPS_NR_ILR_REGS]; |
| u32 mir[INTCPS_NR_MIR_REGS]; |
| }; |
| static struct omap_intc_regs intc_context; |
| |
| static struct irq_domain *domain; |
| static void __iomem *omap_irq_base; |
| static int omap_nr_pending; |
| static int omap_nr_irqs; |
| |
| static void intc_writel(u32 reg, u32 val) |
| { |
| writel_relaxed(val, omap_irq_base + reg); |
| } |
| |
| static u32 intc_readl(u32 reg) |
| { |
| return readl_relaxed(omap_irq_base + reg); |
| } |
| |
| void omap_intc_save_context(void) |
| { |
| int i; |
| |
| intc_context.sysconfig = |
| intc_readl(INTC_SYSCONFIG); |
| intc_context.protection = |
| intc_readl(INTC_PROTECTION); |
| intc_context.idle = |
| intc_readl(INTC_IDLE); |
| intc_context.threshold = |
| intc_readl(INTC_THRESHOLD); |
| |
| for (i = 0; i < omap_nr_irqs; i++) |
| intc_context.ilr[i] = |
| intc_readl((INTC_ILR0 + 0x4 * i)); |
| for (i = 0; i < INTCPS_NR_MIR_REGS; i++) |
| intc_context.mir[i] = |
| intc_readl(INTC_MIR0 + (0x20 * i)); |
| } |
| |
| void omap_intc_restore_context(void) |
| { |
| int i; |
| |
| intc_writel(INTC_SYSCONFIG, intc_context.sysconfig); |
| intc_writel(INTC_PROTECTION, intc_context.protection); |
| intc_writel(INTC_IDLE, intc_context.idle); |
| intc_writel(INTC_THRESHOLD, intc_context.threshold); |
| |
| for (i = 0; i < omap_nr_irqs; i++) |
| intc_writel(INTC_ILR0 + 0x4 * i, |
| intc_context.ilr[i]); |
| |
| for (i = 0; i < INTCPS_NR_MIR_REGS; i++) |
| intc_writel(INTC_MIR0 + 0x20 * i, |
| intc_context.mir[i]); |
| /* MIRs are saved and restore with other PRCM registers */ |
| } |
| |
| void omap3_intc_prepare_idle(void) |
| { |
| /* |
| * Disable autoidle as it can stall interrupt controller, |
| * cf. errata ID i540 for 3430 (all revisions up to 3.1.x) |
| */ |
| intc_writel(INTC_SYSCONFIG, 0); |
| intc_writel(INTC_IDLE, INTC_IDLE_TURBO); |
| } |
| |
| void omap3_intc_resume_idle(void) |
| { |
| /* Re-enable autoidle */ |
| intc_writel(INTC_SYSCONFIG, 1); |
| intc_writel(INTC_IDLE, 0); |
| } |
| |
| /* XXX: FIQ and additional INTC support (only MPU at the moment) */ |
| static void omap_ack_irq(struct irq_data *d) |
| { |
| intc_writel(INTC_CONTROL, 0x1); |
| } |
| |
| static void omap_mask_ack_irq(struct irq_data *d) |
| { |
| irq_gc_mask_disable_reg(d); |
| omap_ack_irq(d); |
| } |
| |
| static void __init omap_irq_soft_reset(void) |
| { |
| unsigned long tmp; |
| |
| tmp = intc_readl(INTC_REVISION) & 0xff; |
| |
| pr_info("IRQ: Found an INTC at 0x%p (revision %ld.%ld) with %d interrupts\n", |
| omap_irq_base, tmp >> 4, tmp & 0xf, omap_nr_irqs); |
| |
| tmp = intc_readl(INTC_SYSCONFIG); |
| tmp |= 1 << 1; /* soft reset */ |
| intc_writel(INTC_SYSCONFIG, tmp); |
| |
| while (!(intc_readl(INTC_SYSSTATUS) & 0x1)) |
| /* Wait for reset to complete */; |
| |
| /* Enable autoidle */ |
| intc_writel(INTC_SYSCONFIG, 1 << 0); |
| } |
| |
| int omap_irq_pending(void) |
| { |
| int i; |
| |
| for (i = 0; i < omap_nr_pending; i++) |
| if (intc_readl(INTC_PENDING_IRQ0 + (0x20 * i))) |
| return 1; |
| return 0; |
| } |
| |
| void omap3_intc_suspend(void) |
| { |
| /* A pending interrupt would prevent OMAP from entering suspend */ |
| omap_ack_irq(NULL); |
| } |
| |
| static int __init omap_alloc_gc_of(struct irq_domain *d, void __iomem *base) |
| { |
| int ret; |
| int i; |
| |
| ret = irq_alloc_domain_generic_chips(d, 32, 1, "INTC", |
| handle_level_irq, IRQ_NOREQUEST | IRQ_NOPROBE, |
| IRQ_LEVEL, 0); |
| if (ret) { |
| pr_warn("Failed to allocate irq chips\n"); |
| return ret; |
| } |
| |
| for (i = 0; i < omap_nr_pending; i++) { |
| struct irq_chip_generic *gc; |
| struct irq_chip_type *ct; |
| |
| gc = irq_get_domain_generic_chip(d, 32 * i); |
| gc->reg_base = base; |
| ct = gc->chip_types; |
| |
| ct->type = IRQ_TYPE_LEVEL_MASK; |
| |
| ct->chip.irq_ack = omap_mask_ack_irq; |
| ct->chip.irq_mask = irq_gc_mask_disable_reg; |
| ct->chip.irq_unmask = irq_gc_unmask_enable_reg; |
| |
| ct->chip.flags |= IRQCHIP_SKIP_SET_WAKE; |
| |
| ct->regs.enable = INTC_MIR_CLEAR0 + 32 * i; |
| ct->regs.disable = INTC_MIR_SET0 + 32 * i; |
| } |
| |
| return 0; |
| } |
| |
| static void __init omap_alloc_gc_legacy(void __iomem *base, |
| unsigned int irq_start, unsigned int num) |
| { |
| struct irq_chip_generic *gc; |
| struct irq_chip_type *ct; |
| |
| gc = irq_alloc_generic_chip("INTC", 1, irq_start, base, |
| handle_level_irq); |
| ct = gc->chip_types; |
| ct->chip.irq_ack = omap_mask_ack_irq; |
| ct->chip.irq_mask = irq_gc_mask_disable_reg; |
| ct->chip.irq_unmask = irq_gc_unmask_enable_reg; |
| ct->chip.flags |= IRQCHIP_SKIP_SET_WAKE; |
| |
| ct->regs.enable = INTC_MIR_CLEAR0; |
| ct->regs.disable = INTC_MIR_SET0; |
| irq_setup_generic_chip(gc, IRQ_MSK(num), IRQ_GC_INIT_MASK_CACHE, |
| IRQ_NOREQUEST | IRQ_NOPROBE, 0); |
| } |
| |
| static int __init omap_init_irq_of(struct device_node *node) |
| { |
| int ret; |
| |
| omap_irq_base = of_iomap(node, 0); |
| if (WARN_ON(!omap_irq_base)) |
| return -ENOMEM; |
| |
| domain = irq_domain_add_linear(node, omap_nr_irqs, |
| &irq_generic_chip_ops, NULL); |
| |
| omap_irq_soft_reset(); |
| |
| ret = omap_alloc_gc_of(domain, omap_irq_base); |
| if (ret < 0) |
| irq_domain_remove(domain); |
| |
| return ret; |
| } |
| |
| static int __init omap_init_irq_legacy(u32 base, struct device_node *node) |
| { |
| int j, irq_base; |
| |
| omap_irq_base = ioremap(base, SZ_4K); |
| if (WARN_ON(!omap_irq_base)) |
| return -ENOMEM; |
| |
| irq_base = irq_alloc_descs(-1, 0, omap_nr_irqs, 0); |
| if (irq_base < 0) { |
| pr_warn("Couldn't allocate IRQ numbers\n"); |
| irq_base = 0; |
| } |
| |
| domain = irq_domain_add_legacy(node, omap_nr_irqs, irq_base, 0, |
| &irq_domain_simple_ops, NULL); |
| |
| omap_irq_soft_reset(); |
| |
| for (j = 0; j < omap_nr_irqs; j += 32) |
| omap_alloc_gc_legacy(omap_irq_base + j, j + irq_base, 32); |
| |
| return 0; |
| } |
| |
| static void __init omap_irq_enable_protection(void) |
| { |
| u32 reg; |
| |
| reg = intc_readl(INTC_PROTECTION); |
| reg |= INTC_PROTECTION_ENABLE; |
| intc_writel(INTC_PROTECTION, reg); |
| } |
| |
| static int __init omap_init_irq(u32 base, struct device_node *node) |
| { |
| int ret; |
| |
| /* |
| * FIXME legacy OMAP DMA driver sitting under arch/arm/plat-omap/dma.c |
| * depends is still not ready for linear IRQ domains; because of that |
| * we need to temporarily "blacklist" OMAP2 and OMAP3 devices from using |
| * linear IRQ Domain until that driver is finally fixed. |
| */ |
| if (of_device_is_compatible(node, "ti,omap2-intc") || |
| of_device_is_compatible(node, "ti,omap3-intc")) { |
| struct resource res; |
| |
| if (of_address_to_resource(node, 0, &res)) |
| return -ENOMEM; |
| |
| base = res.start; |
| ret = omap_init_irq_legacy(base, node); |
| } else if (node) { |
| ret = omap_init_irq_of(node); |
| } else { |
| ret = omap_init_irq_legacy(base, NULL); |
| } |
| |
| if (ret == 0) |
| omap_irq_enable_protection(); |
| |
| return ret; |
| } |
| |
| static void __exception_irq_entry omap_intc_handle_irq(struct pt_regs *regs) |
| { |
| extern unsigned long irq_err_count; |
| u32 irqnr; |
| |
| irqnr = intc_readl(INTC_SIR); |
| |
| /* |
| * A spurious IRQ can result if interrupt that triggered the |
| * sorting is no longer active during the sorting (10 INTC |
| * functional clock cycles after interrupt assertion). Or a |
| * change in interrupt mask affected the result during sorting |
| * time. There is no special handling required except ignoring |
| * the SIR register value just read and retrying. |
| * See section 6.2.5 of AM335x TRM Literature Number: SPRUH73K |
| * |
| * Many a times, a spurious interrupt situation has been fixed |
| * by adding a flush for the posted write acking the IRQ in |
| * the device driver. Typically, this is going be the device |
| * driver whose interrupt was handled just before the spurious |
| * IRQ occurred. Pay attention to those device drivers if you |
| * run into hitting the spurious IRQ condition below. |
| */ |
| if (unlikely((irqnr & SPURIOUSIRQ_MASK) == SPURIOUSIRQ_MASK)) { |
| pr_err_once("%s: spurious irq!\n", __func__); |
| irq_err_count++; |
| omap_ack_irq(NULL); |
| return; |
| } |
| |
| irqnr &= ACTIVEIRQ_MASK; |
| generic_handle_domain_irq(domain, irqnr); |
| } |
| |
| static int __init intc_of_init(struct device_node *node, |
| struct device_node *parent) |
| { |
| int ret; |
| |
| omap_nr_pending = 3; |
| omap_nr_irqs = 96; |
| |
| if (WARN_ON(!node)) |
| return -ENODEV; |
| |
| if (of_device_is_compatible(node, "ti,dm814-intc") || |
| of_device_is_compatible(node, "ti,dm816-intc") || |
| of_device_is_compatible(node, "ti,am33xx-intc")) { |
| omap_nr_irqs = 128; |
| omap_nr_pending = 4; |
| } |
| |
| ret = omap_init_irq(-1, of_node_get(node)); |
| if (ret < 0) |
| return ret; |
| |
| set_handle_irq(omap_intc_handle_irq); |
| |
| return 0; |
| } |
| |
| IRQCHIP_DECLARE(omap2_intc, "ti,omap2-intc", intc_of_init); |
| IRQCHIP_DECLARE(omap3_intc, "ti,omap3-intc", intc_of_init); |
| IRQCHIP_DECLARE(dm814x_intc, "ti,dm814-intc", intc_of_init); |
| IRQCHIP_DECLARE(dm816x_intc, "ti,dm816-intc", intc_of_init); |
| IRQCHIP_DECLARE(am33xx_intc, "ti,am33xx-intc", intc_of_init); |