| // SPDX-License-Identifier: GPL-2.0-only |
| /* |
| * Copyright (C) 2011-12 Synopsys, Inc. (www.synopsys.com) |
| */ |
| |
| #include <linux/interrupt.h> |
| #include <linux/module.h> |
| #include <linux/of.h> |
| #include <linux/irqdomain.h> |
| #include <linux/irqchip.h> |
| #include <asm/irq.h> |
| |
| #define NR_CPU_IRQS 32 /* number of irq lines coming in */ |
| #define TIMER0_IRQ 3 /* Fixed by ISA */ |
| |
| /* |
| * Early Hardware specific Interrupt setup |
| * -Platform independent, needed for each CPU (not foldable into init_IRQ) |
| * -Called very early (start_kernel -> setup_arch -> setup_processor) |
| * |
| * what it does ? |
| * -Optionally, setup the High priority Interrupts as Level 2 IRQs |
| */ |
| void arc_init_IRQ(void) |
| { |
| unsigned int level_mask = 0, i; |
| |
| /* Is timer high priority Interrupt (Level2 in ARCompact jargon) */ |
| level_mask |= IS_ENABLED(CONFIG_ARC_COMPACT_IRQ_LEVELS) << TIMER0_IRQ; |
| |
| /* |
| * Write to register, even if no LV2 IRQs configured to reset it |
| * in case bootloader had mucked with it |
| */ |
| write_aux_reg(AUX_IRQ_LEV, level_mask); |
| |
| if (level_mask) |
| pr_info("Level-2 interrupts bitset %x\n", level_mask); |
| |
| /* |
| * Disable all IRQ lines so faulty external hardware won't |
| * trigger interrupt that kernel is not ready to handle. |
| */ |
| for (i = TIMER0_IRQ; i < NR_CPU_IRQS; i++) { |
| unsigned int ienb; |
| |
| ienb = read_aux_reg(AUX_IENABLE); |
| ienb &= ~(1 << i); |
| write_aux_reg(AUX_IENABLE, ienb); |
| } |
| } |
| |
| /* |
| * ARC700 core includes a simple on-chip intc supporting |
| * -per IRQ enable/disable |
| * -2 levels of interrupts (high/low) |
| * -all interrupts being level triggered |
| * |
| * To reduce platform code, we assume all IRQs directly hooked-up into intc. |
| * Platforms with external intc, hence cascaded IRQs, are free to over-ride |
| * below, per IRQ. |
| */ |
| |
| static void arc_irq_mask(struct irq_data *data) |
| { |
| unsigned int ienb; |
| |
| ienb = read_aux_reg(AUX_IENABLE); |
| ienb &= ~(1 << data->hwirq); |
| write_aux_reg(AUX_IENABLE, ienb); |
| } |
| |
| static void arc_irq_unmask(struct irq_data *data) |
| { |
| unsigned int ienb; |
| |
| ienb = read_aux_reg(AUX_IENABLE); |
| ienb |= (1 << data->hwirq); |
| write_aux_reg(AUX_IENABLE, ienb); |
| } |
| |
| static struct irq_chip onchip_intc = { |
| .name = "ARC In-core Intc", |
| .irq_mask = arc_irq_mask, |
| .irq_unmask = arc_irq_unmask, |
| }; |
| |
| static int arc_intc_domain_map(struct irq_domain *d, unsigned int irq, |
| irq_hw_number_t hw) |
| { |
| switch (hw) { |
| case TIMER0_IRQ: |
| irq_set_percpu_devid(irq); |
| irq_set_chip_and_handler(irq, &onchip_intc, handle_percpu_irq); |
| break; |
| default: |
| irq_set_chip_and_handler(irq, &onchip_intc, handle_level_irq); |
| } |
| return 0; |
| } |
| |
| static const struct irq_domain_ops arc_intc_domain_ops = { |
| .xlate = irq_domain_xlate_onecell, |
| .map = arc_intc_domain_map, |
| }; |
| |
| static int __init |
| init_onchip_IRQ(struct device_node *intc, struct device_node *parent) |
| { |
| struct irq_domain *root_domain; |
| |
| if (parent) |
| panic("DeviceTree incore intc not a root irq controller\n"); |
| |
| root_domain = irq_domain_add_linear(intc, NR_CPU_IRQS, |
| &arc_intc_domain_ops, NULL); |
| if (!root_domain) |
| panic("root irq domain not avail\n"); |
| |
| /* |
| * Needed for primary domain lookup to succeed |
| * This is a primary irqchip, and can never have a parent |
| */ |
| irq_set_default_host(root_domain); |
| |
| return 0; |
| } |
| |
| IRQCHIP_DECLARE(arc_intc, "snps,arc700-intc", init_onchip_IRQ); |
| |
| /* |
| * arch_local_irq_enable - Enable interrupts. |
| * |
| * 1. Explicitly called to re-enable interrupts |
| * 2. Implicitly called from spin_unlock_irq, write_unlock_irq etc |
| * which maybe in hard ISR itself |
| * |
| * Semantics of this function change depending on where it is called from: |
| * |
| * -If called from hard-ISR, it must not invert interrupt priorities |
| * e.g. suppose TIMER is high priority (Level 2) IRQ |
| * Time hard-ISR, timer_interrupt( ) calls spin_unlock_irq several times. |
| * Here local_irq_enable( ) shd not re-enable lower priority interrupts |
| * -If called from soft-ISR, it must re-enable all interrupts |
| * soft ISR are low priority jobs which can be very slow, thus all IRQs |
| * must be enabled while they run. |
| * Now hardware context wise we may still be in L2 ISR (not done rtie) |
| * still we must re-enable both L1 and L2 IRQs |
| * Another twist is prev scenario with flow being |
| * L1 ISR ==> interrupted by L2 ISR ==> L2 soft ISR |
| * here we must not re-enable Ll as prev Ll Interrupt's h/w context will get |
| * over-written (this is deficiency in ARC700 Interrupt mechanism) |
| */ |
| |
| #ifdef CONFIG_ARC_COMPACT_IRQ_LEVELS /* Complex version for 2 IRQ levels */ |
| |
| void arch_local_irq_enable(void) |
| { |
| unsigned long flags = arch_local_save_flags(); |
| |
| if (flags & STATUS_A2_MASK) |
| flags |= STATUS_E2_MASK; |
| else if (flags & STATUS_A1_MASK) |
| flags |= STATUS_E1_MASK; |
| |
| arch_local_irq_restore(flags); |
| } |
| |
| EXPORT_SYMBOL(arch_local_irq_enable); |
| #endif |