mirror of
https://github.com/edk2-porting/linux-next.git
synced 2024-12-25 21:54:06 +08:00
a8db8cf0d8
Each revmap type has different arguments for setting up the revmap. This patch splits up the generator functions so that each revmap type can do its own setup and the user doesn't need to keep track of how each revmap type handles the arguments. This patch also adds a host_data argument to the generators. There are cases where the host_data pointer will be needed before the function returns. ie. the legacy map calls the .map callback for each irq before returning. v2: - Add void *host_data argument to irq_domain_add_*() functions - fixed failure to compile - Moved IRQ_DOMAIN_MAP_* defines into irqdomain.c Signed-off-by: Grant Likely <grant.likely@secretlab.ca> Cc: Rob Herring <rob.herring@calxeda.com> Cc: Benjamin Herrenschmidt <benh@kernel.crashing.org> Cc: Thomas Gleixner <tglx@linutronix.de> Cc: Milton Miller <miltonm@bga.com> Tested-by: Olof Johansson <olof@lixom.net>
298 lines
8.1 KiB
C
298 lines
8.1 KiB
C
/*
|
|
* Interrupt handling for Marvell mv64360/mv64460 host bridges (Discovery)
|
|
*
|
|
* Author: Dale Farnsworth <dale@farnsworth.org>
|
|
*
|
|
* 2007 (c) MontaVista, Software, Inc. This file is licensed under
|
|
* the terms of the GNU General Public License version 2. This program
|
|
* is licensed "as is" without any warranty of any kind, whether express
|
|
* or implied.
|
|
*/
|
|
|
|
#include <linux/stddef.h>
|
|
#include <linux/kernel.h>
|
|
#include <linux/init.h>
|
|
#include <linux/irq.h>
|
|
#include <linux/interrupt.h>
|
|
#include <linux/spinlock.h>
|
|
|
|
#include <asm/byteorder.h>
|
|
#include <asm/io.h>
|
|
#include <asm/prom.h>
|
|
#include <asm/irq.h>
|
|
|
|
#include "mv64x60.h"
|
|
|
|
/* Interrupt Controller Interface Registers */
|
|
#define MV64X60_IC_MAIN_CAUSE_LO 0x0004
|
|
#define MV64X60_IC_MAIN_CAUSE_HI 0x000c
|
|
#define MV64X60_IC_CPU0_INTR_MASK_LO 0x0014
|
|
#define MV64X60_IC_CPU0_INTR_MASK_HI 0x001c
|
|
#define MV64X60_IC_CPU0_SELECT_CAUSE 0x0024
|
|
|
|
#define MV64X60_HIGH_GPP_GROUPS 0x0f000000
|
|
#define MV64X60_SELECT_CAUSE_HIGH 0x40000000
|
|
|
|
/* General Purpose Pins Controller Interface Registers */
|
|
#define MV64x60_GPP_INTR_CAUSE 0x0008
|
|
#define MV64x60_GPP_INTR_MASK 0x000c
|
|
|
|
#define MV64x60_LEVEL1_LOW 0
|
|
#define MV64x60_LEVEL1_HIGH 1
|
|
#define MV64x60_LEVEL1_GPP 2
|
|
|
|
#define MV64x60_LEVEL1_MASK 0x00000060
|
|
#define MV64x60_LEVEL1_OFFSET 5
|
|
|
|
#define MV64x60_LEVEL2_MASK 0x0000001f
|
|
|
|
#define MV64x60_NUM_IRQS 96
|
|
|
|
static DEFINE_SPINLOCK(mv64x60_lock);
|
|
|
|
static void __iomem *mv64x60_irq_reg_base;
|
|
static void __iomem *mv64x60_gpp_reg_base;
|
|
|
|
/*
|
|
* Interrupt Controller Handling
|
|
*
|
|
* The interrupt controller handles three groups of interrupts:
|
|
* main low: IRQ0-IRQ31
|
|
* main high: IRQ32-IRQ63
|
|
* gpp: IRQ64-IRQ95
|
|
*
|
|
* This code handles interrupts in two levels. Level 1 selects the
|
|
* interrupt group, and level 2 selects an IRQ within that group.
|
|
* Each group has its own irq_chip structure.
|
|
*/
|
|
|
|
static u32 mv64x60_cached_low_mask;
|
|
static u32 mv64x60_cached_high_mask = MV64X60_HIGH_GPP_GROUPS;
|
|
static u32 mv64x60_cached_gpp_mask;
|
|
|
|
static struct irq_domain *mv64x60_irq_host;
|
|
|
|
/*
|
|
* mv64x60_chip_low functions
|
|
*/
|
|
|
|
static void mv64x60_mask_low(struct irq_data *d)
|
|
{
|
|
int level2 = irqd_to_hwirq(d) & MV64x60_LEVEL2_MASK;
|
|
unsigned long flags;
|
|
|
|
spin_lock_irqsave(&mv64x60_lock, flags);
|
|
mv64x60_cached_low_mask &= ~(1 << level2);
|
|
out_le32(mv64x60_irq_reg_base + MV64X60_IC_CPU0_INTR_MASK_LO,
|
|
mv64x60_cached_low_mask);
|
|
spin_unlock_irqrestore(&mv64x60_lock, flags);
|
|
(void)in_le32(mv64x60_irq_reg_base + MV64X60_IC_CPU0_INTR_MASK_LO);
|
|
}
|
|
|
|
static void mv64x60_unmask_low(struct irq_data *d)
|
|
{
|
|
int level2 = irqd_to_hwirq(d) & MV64x60_LEVEL2_MASK;
|
|
unsigned long flags;
|
|
|
|
spin_lock_irqsave(&mv64x60_lock, flags);
|
|
mv64x60_cached_low_mask |= 1 << level2;
|
|
out_le32(mv64x60_irq_reg_base + MV64X60_IC_CPU0_INTR_MASK_LO,
|
|
mv64x60_cached_low_mask);
|
|
spin_unlock_irqrestore(&mv64x60_lock, flags);
|
|
(void)in_le32(mv64x60_irq_reg_base + MV64X60_IC_CPU0_INTR_MASK_LO);
|
|
}
|
|
|
|
static struct irq_chip mv64x60_chip_low = {
|
|
.name = "mv64x60_low",
|
|
.irq_mask = mv64x60_mask_low,
|
|
.irq_mask_ack = mv64x60_mask_low,
|
|
.irq_unmask = mv64x60_unmask_low,
|
|
};
|
|
|
|
/*
|
|
* mv64x60_chip_high functions
|
|
*/
|
|
|
|
static void mv64x60_mask_high(struct irq_data *d)
|
|
{
|
|
int level2 = irqd_to_hwirq(d) & MV64x60_LEVEL2_MASK;
|
|
unsigned long flags;
|
|
|
|
spin_lock_irqsave(&mv64x60_lock, flags);
|
|
mv64x60_cached_high_mask &= ~(1 << level2);
|
|
out_le32(mv64x60_irq_reg_base + MV64X60_IC_CPU0_INTR_MASK_HI,
|
|
mv64x60_cached_high_mask);
|
|
spin_unlock_irqrestore(&mv64x60_lock, flags);
|
|
(void)in_le32(mv64x60_irq_reg_base + MV64X60_IC_CPU0_INTR_MASK_HI);
|
|
}
|
|
|
|
static void mv64x60_unmask_high(struct irq_data *d)
|
|
{
|
|
int level2 = irqd_to_hwirq(d) & MV64x60_LEVEL2_MASK;
|
|
unsigned long flags;
|
|
|
|
spin_lock_irqsave(&mv64x60_lock, flags);
|
|
mv64x60_cached_high_mask |= 1 << level2;
|
|
out_le32(mv64x60_irq_reg_base + MV64X60_IC_CPU0_INTR_MASK_HI,
|
|
mv64x60_cached_high_mask);
|
|
spin_unlock_irqrestore(&mv64x60_lock, flags);
|
|
(void)in_le32(mv64x60_irq_reg_base + MV64X60_IC_CPU0_INTR_MASK_HI);
|
|
}
|
|
|
|
static struct irq_chip mv64x60_chip_high = {
|
|
.name = "mv64x60_high",
|
|
.irq_mask = mv64x60_mask_high,
|
|
.irq_mask_ack = mv64x60_mask_high,
|
|
.irq_unmask = mv64x60_unmask_high,
|
|
};
|
|
|
|
/*
|
|
* mv64x60_chip_gpp functions
|
|
*/
|
|
|
|
static void mv64x60_mask_gpp(struct irq_data *d)
|
|
{
|
|
int level2 = irqd_to_hwirq(d) & MV64x60_LEVEL2_MASK;
|
|
unsigned long flags;
|
|
|
|
spin_lock_irqsave(&mv64x60_lock, flags);
|
|
mv64x60_cached_gpp_mask &= ~(1 << level2);
|
|
out_le32(mv64x60_gpp_reg_base + MV64x60_GPP_INTR_MASK,
|
|
mv64x60_cached_gpp_mask);
|
|
spin_unlock_irqrestore(&mv64x60_lock, flags);
|
|
(void)in_le32(mv64x60_gpp_reg_base + MV64x60_GPP_INTR_MASK);
|
|
}
|
|
|
|
static void mv64x60_mask_ack_gpp(struct irq_data *d)
|
|
{
|
|
int level2 = irqd_to_hwirq(d) & MV64x60_LEVEL2_MASK;
|
|
unsigned long flags;
|
|
|
|
spin_lock_irqsave(&mv64x60_lock, flags);
|
|
mv64x60_cached_gpp_mask &= ~(1 << level2);
|
|
out_le32(mv64x60_gpp_reg_base + MV64x60_GPP_INTR_MASK,
|
|
mv64x60_cached_gpp_mask);
|
|
out_le32(mv64x60_gpp_reg_base + MV64x60_GPP_INTR_CAUSE,
|
|
~(1 << level2));
|
|
spin_unlock_irqrestore(&mv64x60_lock, flags);
|
|
(void)in_le32(mv64x60_gpp_reg_base + MV64x60_GPP_INTR_CAUSE);
|
|
}
|
|
|
|
static void mv64x60_unmask_gpp(struct irq_data *d)
|
|
{
|
|
int level2 = irqd_to_hwirq(d) & MV64x60_LEVEL2_MASK;
|
|
unsigned long flags;
|
|
|
|
spin_lock_irqsave(&mv64x60_lock, flags);
|
|
mv64x60_cached_gpp_mask |= 1 << level2;
|
|
out_le32(mv64x60_gpp_reg_base + MV64x60_GPP_INTR_MASK,
|
|
mv64x60_cached_gpp_mask);
|
|
spin_unlock_irqrestore(&mv64x60_lock, flags);
|
|
(void)in_le32(mv64x60_gpp_reg_base + MV64x60_GPP_INTR_MASK);
|
|
}
|
|
|
|
static struct irq_chip mv64x60_chip_gpp = {
|
|
.name = "mv64x60_gpp",
|
|
.irq_mask = mv64x60_mask_gpp,
|
|
.irq_mask_ack = mv64x60_mask_ack_gpp,
|
|
.irq_unmask = mv64x60_unmask_gpp,
|
|
};
|
|
|
|
/*
|
|
* mv64x60_host_ops functions
|
|
*/
|
|
|
|
static struct irq_chip *mv64x60_chips[] = {
|
|
[MV64x60_LEVEL1_LOW] = &mv64x60_chip_low,
|
|
[MV64x60_LEVEL1_HIGH] = &mv64x60_chip_high,
|
|
[MV64x60_LEVEL1_GPP] = &mv64x60_chip_gpp,
|
|
};
|
|
|
|
static int mv64x60_host_map(struct irq_domain *h, unsigned int virq,
|
|
irq_hw_number_t hwirq)
|
|
{
|
|
int level1;
|
|
|
|
irq_set_status_flags(virq, IRQ_LEVEL);
|
|
|
|
level1 = (hwirq & MV64x60_LEVEL1_MASK) >> MV64x60_LEVEL1_OFFSET;
|
|
BUG_ON(level1 > MV64x60_LEVEL1_GPP);
|
|
irq_set_chip_and_handler(virq, mv64x60_chips[level1],
|
|
handle_level_irq);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static struct irq_domain_ops mv64x60_host_ops = {
|
|
.map = mv64x60_host_map,
|
|
};
|
|
|
|
/*
|
|
* Global functions
|
|
*/
|
|
|
|
void __init mv64x60_init_irq(void)
|
|
{
|
|
struct device_node *np;
|
|
phys_addr_t paddr;
|
|
unsigned int size;
|
|
const unsigned int *reg;
|
|
unsigned long flags;
|
|
|
|
np = of_find_compatible_node(NULL, NULL, "marvell,mv64360-gpp");
|
|
reg = of_get_property(np, "reg", &size);
|
|
paddr = of_translate_address(np, reg);
|
|
mv64x60_gpp_reg_base = ioremap(paddr, reg[1]);
|
|
of_node_put(np);
|
|
|
|
np = of_find_compatible_node(NULL, NULL, "marvell,mv64360-pic");
|
|
reg = of_get_property(np, "reg", &size);
|
|
paddr = of_translate_address(np, reg);
|
|
mv64x60_irq_reg_base = ioremap(paddr, reg[1]);
|
|
|
|
mv64x60_irq_host = irq_domain_add_linear(np, MV64x60_NUM_IRQS,
|
|
&mv64x60_host_ops, NULL);
|
|
|
|
spin_lock_irqsave(&mv64x60_lock, flags);
|
|
out_le32(mv64x60_gpp_reg_base + MV64x60_GPP_INTR_MASK,
|
|
mv64x60_cached_gpp_mask);
|
|
out_le32(mv64x60_irq_reg_base + MV64X60_IC_CPU0_INTR_MASK_LO,
|
|
mv64x60_cached_low_mask);
|
|
out_le32(mv64x60_irq_reg_base + MV64X60_IC_CPU0_INTR_MASK_HI,
|
|
mv64x60_cached_high_mask);
|
|
|
|
out_le32(mv64x60_gpp_reg_base + MV64x60_GPP_INTR_CAUSE, 0);
|
|
out_le32(mv64x60_irq_reg_base + MV64X60_IC_MAIN_CAUSE_LO, 0);
|
|
out_le32(mv64x60_irq_reg_base + MV64X60_IC_MAIN_CAUSE_HI, 0);
|
|
spin_unlock_irqrestore(&mv64x60_lock, flags);
|
|
}
|
|
|
|
unsigned int mv64x60_get_irq(void)
|
|
{
|
|
u32 cause;
|
|
int level1;
|
|
irq_hw_number_t hwirq;
|
|
int virq = NO_IRQ;
|
|
|
|
cause = in_le32(mv64x60_irq_reg_base + MV64X60_IC_CPU0_SELECT_CAUSE);
|
|
if (cause & MV64X60_SELECT_CAUSE_HIGH) {
|
|
cause &= mv64x60_cached_high_mask;
|
|
level1 = MV64x60_LEVEL1_HIGH;
|
|
if (cause & MV64X60_HIGH_GPP_GROUPS) {
|
|
cause = in_le32(mv64x60_gpp_reg_base +
|
|
MV64x60_GPP_INTR_CAUSE);
|
|
cause &= mv64x60_cached_gpp_mask;
|
|
level1 = MV64x60_LEVEL1_GPP;
|
|
}
|
|
} else {
|
|
cause &= mv64x60_cached_low_mask;
|
|
level1 = MV64x60_LEVEL1_LOW;
|
|
}
|
|
if (cause) {
|
|
hwirq = (level1 << MV64x60_LEVEL1_OFFSET) | __ilog2(cause);
|
|
virq = irq_linear_revmap(mv64x60_irq_host, hwirq);
|
|
}
|
|
|
|
return virq;
|
|
}
|