MIPS: msp71xx: Convert to new irq_chip functions
Signed-off-by: Thomas Gleixner <tglx@linutronix.de> To: linux-mips@linux-mips.org Patchwork: https://patchwork.linux-mips.org/patch/2197/ Signed-off-by: Ralf Baechle <ralf@linux-mips.org>
This commit is contained in:
parent
7ec8af9e34
commit
d7881fbdf8
|
@ -77,7 +77,7 @@ static inline void cic_wmb(void)
|
||||||
dummy_read++;
|
dummy_read++;
|
||||||
}
|
}
|
||||||
|
|
||||||
static inline void unmask_cic_irq(unsigned int irq)
|
static void unmask_cic_irq(struct irq_data *d)
|
||||||
{
|
{
|
||||||
volatile u32 *cic_msk_reg = CIC_VPE0_MSK_REG;
|
volatile u32 *cic_msk_reg = CIC_VPE0_MSK_REG;
|
||||||
int vpe;
|
int vpe;
|
||||||
|
@ -89,18 +89,18 @@ static inline void unmask_cic_irq(unsigned int irq)
|
||||||
* Make sure we have IRQ affinity. It may have changed while
|
* Make sure we have IRQ affinity. It may have changed while
|
||||||
* we were processing the IRQ.
|
* we were processing the IRQ.
|
||||||
*/
|
*/
|
||||||
if (!cpumask_test_cpu(smp_processor_id(), irq_desc[irq].affinity))
|
if (!cpumask_test_cpu(smp_processor_id(), d->affinity))
|
||||||
return;
|
return;
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
vpe = get_current_vpe();
|
vpe = get_current_vpe();
|
||||||
LOCK_VPE(flags, mtflags);
|
LOCK_VPE(flags, mtflags);
|
||||||
cic_msk_reg[vpe] |= (1 << (irq - MSP_CIC_INTBASE));
|
cic_msk_reg[vpe] |= (1 << (d->irq - MSP_CIC_INTBASE));
|
||||||
UNLOCK_VPE(flags, mtflags);
|
UNLOCK_VPE(flags, mtflags);
|
||||||
cic_wmb();
|
cic_wmb();
|
||||||
}
|
}
|
||||||
|
|
||||||
static inline void mask_cic_irq(unsigned int irq)
|
static void mask_cic_irq(struct irq_data *d)
|
||||||
{
|
{
|
||||||
volatile u32 *cic_msk_reg = CIC_VPE0_MSK_REG;
|
volatile u32 *cic_msk_reg = CIC_VPE0_MSK_REG;
|
||||||
int vpe = get_current_vpe();
|
int vpe = get_current_vpe();
|
||||||
|
@ -108,33 +108,27 @@ static inline void mask_cic_irq(unsigned int irq)
|
||||||
unsigned long flags, mtflags;
|
unsigned long flags, mtflags;
|
||||||
#endif
|
#endif
|
||||||
LOCK_VPE(flags, mtflags);
|
LOCK_VPE(flags, mtflags);
|
||||||
cic_msk_reg[vpe] &= ~(1 << (irq - MSP_CIC_INTBASE));
|
cic_msk_reg[vpe] &= ~(1 << (d->irq - MSP_CIC_INTBASE));
|
||||||
UNLOCK_VPE(flags, mtflags);
|
UNLOCK_VPE(flags, mtflags);
|
||||||
cic_wmb();
|
cic_wmb();
|
||||||
}
|
}
|
||||||
static inline void msp_cic_irq_ack(unsigned int irq)
|
static void msp_cic_irq_ack(struct irq_data *d)
|
||||||
{
|
{
|
||||||
mask_cic_irq(irq);
|
mask_cic_irq(d);
|
||||||
/*
|
/*
|
||||||
* Only really necessary for 18, 16-14 and sometimes 3:0
|
* Only really necessary for 18, 16-14 and sometimes 3:0
|
||||||
* (since these can be edge sensitive) but it doesn't
|
* (since these can be edge sensitive) but it doesn't
|
||||||
* hurt for the others
|
* hurt for the others
|
||||||
*/
|
*/
|
||||||
*CIC_STS_REG = (1 << (irq - MSP_CIC_INTBASE));
|
*CIC_STS_REG = (1 << (d->irq - MSP_CIC_INTBASE));
|
||||||
smtc_im_ack_irq(irq);
|
smtc_im_ack_irq(d->irq);
|
||||||
}
|
|
||||||
|
|
||||||
static void msp_cic_irq_end(unsigned int irq)
|
|
||||||
{
|
|
||||||
if (!(irq_desc[irq].status & (IRQ_DISABLED | IRQ_INPROGRESS)))
|
|
||||||
unmask_cic_irq(irq);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/*Note: Limiting to VSMP . Not tested in SMTC */
|
/*Note: Limiting to VSMP . Not tested in SMTC */
|
||||||
|
|
||||||
#ifdef CONFIG_MIPS_MT_SMP
|
#ifdef CONFIG_MIPS_MT_SMP
|
||||||
static inline int msp_cic_irq_set_affinity(unsigned int irq,
|
static int msp_cic_irq_set_affinity(struct irq_data *d,
|
||||||
const struct cpumask *cpumask)
|
const struct cpumask *cpumask, bool force)
|
||||||
{
|
{
|
||||||
int cpu;
|
int cpu;
|
||||||
unsigned long flags;
|
unsigned long flags;
|
||||||
|
@ -163,13 +157,12 @@ static inline int msp_cic_irq_set_affinity(unsigned int irq,
|
||||||
|
|
||||||
static struct irq_chip msp_cic_irq_controller = {
|
static struct irq_chip msp_cic_irq_controller = {
|
||||||
.name = "MSP_CIC",
|
.name = "MSP_CIC",
|
||||||
.mask = mask_cic_irq,
|
.irq_mask = mask_cic_irq,
|
||||||
.mask_ack = msp_cic_irq_ack,
|
.irq_mask_ack = msp_cic_irq_ack,
|
||||||
.unmask = unmask_cic_irq,
|
.irq_unmask = unmask_cic_irq,
|
||||||
.ack = msp_cic_irq_ack,
|
.irq_ack = msp_cic_irq_ack,
|
||||||
.end = msp_cic_irq_end,
|
|
||||||
#ifdef CONFIG_MIPS_MT_SMP
|
#ifdef CONFIG_MIPS_MT_SMP
|
||||||
.set_affinity = msp_cic_irq_set_affinity,
|
.irq_set_affinity = msp_cic_irq_set_affinity,
|
||||||
#endif
|
#endif
|
||||||
};
|
};
|
||||||
|
|
||||||
|
@ -220,7 +213,5 @@ void msp_cic_irq_dispatch(void)
|
||||||
do_IRQ(ffs(pending) + MSP_CIC_INTBASE - 1);
|
do_IRQ(ffs(pending) + MSP_CIC_INTBASE - 1);
|
||||||
} else{
|
} else{
|
||||||
spurious_interrupt();
|
spurious_interrupt();
|
||||||
/* Re-enable the CIC cascaded interrupt. */
|
|
||||||
irq_desc[MSP_INT_CIC].chip->end(MSP_INT_CIC);
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -48,100 +48,61 @@ static inline void per_wmb(void)
|
||||||
dummy_read++;
|
dummy_read++;
|
||||||
}
|
}
|
||||||
|
|
||||||
static inline void unmask_per_irq(unsigned int irq)
|
static inline void unmask_per_irq(struct irq_data *d)
|
||||||
{
|
{
|
||||||
#ifdef CONFIG_SMP
|
#ifdef CONFIG_SMP
|
||||||
unsigned long flags;
|
unsigned long flags;
|
||||||
spin_lock_irqsave(&per_lock, flags);
|
spin_lock_irqsave(&per_lock, flags);
|
||||||
*PER_INT_MSK_REG |= (1 << (irq - MSP_PER_INTBASE));
|
*PER_INT_MSK_REG |= (1 << (d->irq - MSP_PER_INTBASE));
|
||||||
spin_unlock_irqrestore(&per_lock, flags);
|
spin_unlock_irqrestore(&per_lock, flags);
|
||||||
#else
|
#else
|
||||||
*PER_INT_MSK_REG |= (1 << (irq - MSP_PER_INTBASE));
|
*PER_INT_MSK_REG |= (1 << (d->irq - MSP_PER_INTBASE));
|
||||||
#endif
|
#endif
|
||||||
per_wmb();
|
per_wmb();
|
||||||
}
|
}
|
||||||
|
|
||||||
static inline void mask_per_irq(unsigned int irq)
|
static inline void mask_per_irq(struct irq_data *d)
|
||||||
{
|
{
|
||||||
#ifdef CONFIG_SMP
|
#ifdef CONFIG_SMP
|
||||||
unsigned long flags;
|
unsigned long flags;
|
||||||
spin_lock_irqsave(&per_lock, flags);
|
spin_lock_irqsave(&per_lock, flags);
|
||||||
*PER_INT_MSK_REG &= ~(1 << (irq - MSP_PER_INTBASE));
|
*PER_INT_MSK_REG &= ~(1 << (d->irq - MSP_PER_INTBASE));
|
||||||
spin_unlock_irqrestore(&per_lock, flags);
|
spin_unlock_irqrestore(&per_lock, flags);
|
||||||
#else
|
#else
|
||||||
*PER_INT_MSK_REG &= ~(1 << (irq - MSP_PER_INTBASE));
|
*PER_INT_MSK_REG &= ~(1 << (d->irq - MSP_PER_INTBASE));
|
||||||
#endif
|
#endif
|
||||||
per_wmb();
|
per_wmb();
|
||||||
}
|
}
|
||||||
|
|
||||||
static inline void msp_per_irq_enable(unsigned int irq)
|
static inline void msp_per_irq_ack(struct irq_data *d)
|
||||||
{
|
{
|
||||||
unmask_per_irq(irq);
|
mask_per_irq(d);
|
||||||
}
|
|
||||||
|
|
||||||
static inline void msp_per_irq_disable(unsigned int irq)
|
|
||||||
{
|
|
||||||
mask_per_irq(irq);
|
|
||||||
}
|
|
||||||
|
|
||||||
static unsigned int msp_per_irq_startup(unsigned int irq)
|
|
||||||
{
|
|
||||||
msp_per_irq_enable(irq);
|
|
||||||
return 0;
|
|
||||||
}
|
|
||||||
|
|
||||||
#define msp_per_irq_shutdown msp_per_irq_disable
|
|
||||||
|
|
||||||
static inline void msp_per_irq_ack(unsigned int irq)
|
|
||||||
{
|
|
||||||
mask_per_irq(irq);
|
|
||||||
/*
|
/*
|
||||||
* In the PER interrupt controller, only bits 11 and 10
|
* In the PER interrupt controller, only bits 11 and 10
|
||||||
* are write-to-clear, (SPI TX complete, SPI RX complete).
|
* are write-to-clear, (SPI TX complete, SPI RX complete).
|
||||||
* It does nothing for any others.
|
* It does nothing for any others.
|
||||||
*/
|
*/
|
||||||
|
*PER_INT_STS_REG = (1 << (d->irq - MSP_PER_INTBASE));
|
||||||
*PER_INT_STS_REG = (1 << (irq - MSP_PER_INTBASE));
|
|
||||||
|
|
||||||
/* Re-enable the CIC cascaded interrupt and return */
|
|
||||||
irq_desc[MSP_INT_CIC].chip->end(MSP_INT_CIC);
|
|
||||||
}
|
|
||||||
|
|
||||||
static void msp_per_irq_end(unsigned int irq)
|
|
||||||
{
|
|
||||||
if (!(irq_desc[irq].status & (IRQ_DISABLED | IRQ_INPROGRESS)))
|
|
||||||
unmask_per_irq(irq);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
#ifdef CONFIG_SMP
|
#ifdef CONFIG_SMP
|
||||||
static inline int msp_per_irq_set_affinity(unsigned int irq,
|
static int msp_per_irq_set_affinity(struct irq_data *d,
|
||||||
const struct cpumask *affinity)
|
const struct cpumask *affinity, bool force)
|
||||||
{
|
{
|
||||||
unsigned long flags;
|
/* WTF is this doing ????? */
|
||||||
/*
|
unmask_per_irq(d);
|
||||||
* Calls to ack, end, startup, enable are spinlocked in setup_irq and
|
|
||||||
* __do_IRQ.Callers of this function do not spinlock,so we need to
|
|
||||||
* do so ourselves.
|
|
||||||
*/
|
|
||||||
raw_spin_lock_irqsave(&irq_desc[irq].lock, flags);
|
|
||||||
msp_per_irq_enable(irq);
|
|
||||||
raw_spin_unlock_irqrestore(&irq_desc[irq].lock, flags);
|
|
||||||
return 0;
|
return 0;
|
||||||
|
|
||||||
}
|
}
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
static struct irq_chip msp_per_irq_controller = {
|
static struct irq_chip msp_per_irq_controller = {
|
||||||
.name = "MSP_PER",
|
.name = "MSP_PER",
|
||||||
.startup = msp_per_irq_startup,
|
.irq_enable = unmask_per_irq.
|
||||||
.shutdown = msp_per_irq_shutdown,
|
.irq_disable = mask_per_irq,
|
||||||
.enable = msp_per_irq_enable,
|
.irq_ack = msp_per_irq_ack,
|
||||||
.disable = msp_per_irq_disable,
|
|
||||||
#ifdef CONFIG_SMP
|
#ifdef CONFIG_SMP
|
||||||
.set_affinity = msp_per_irq_set_affinity,
|
.irq_set_affinity = msp_per_irq_set_affinity,
|
||||||
#endif
|
#endif
|
||||||
.ack = msp_per_irq_ack,
|
|
||||||
.end = msp_per_irq_end,
|
|
||||||
};
|
};
|
||||||
|
|
||||||
void __init msp_per_irq_init(void)
|
void __init msp_per_irq_init(void)
|
||||||
|
@ -152,10 +113,7 @@ void __init msp_per_irq_init(void)
|
||||||
*PER_INT_STS_REG = 0xFFFFFFFF;
|
*PER_INT_STS_REG = 0xFFFFFFFF;
|
||||||
/* initialize all the IRQ descriptors */
|
/* initialize all the IRQ descriptors */
|
||||||
for (i = MSP_PER_INTBASE; i < MSP_PER_INTBASE + 32; i++) {
|
for (i = MSP_PER_INTBASE; i < MSP_PER_INTBASE + 32; i++) {
|
||||||
irq_desc[i].status = IRQ_DISABLED;
|
irq_set_chip(i, &msp_per_irq_controller);
|
||||||
irq_desc[i].action = NULL;
|
|
||||||
irq_desc[i].depth = 1;
|
|
||||||
irq_desc[i].chip = &msp_per_irq_controller;
|
|
||||||
#ifdef CONFIG_MIPS_MT_SMTC
|
#ifdef CONFIG_MIPS_MT_SMTC
|
||||||
irq_hwmask[i] = C_IRQ4;
|
irq_hwmask[i] = C_IRQ4;
|
||||||
#endif
|
#endif
|
||||||
|
@ -173,7 +131,5 @@ void msp_per_irq_dispatch(void)
|
||||||
do_IRQ(ffs(pending) + MSP_PER_INTBASE - 1);
|
do_IRQ(ffs(pending) + MSP_PER_INTBASE - 1);
|
||||||
} else {
|
} else {
|
||||||
spurious_interrupt();
|
spurious_interrupt();
|
||||||
/* Re-enable the CIC cascaded interrupt and return */
|
|
||||||
irq_desc[MSP_INT_CIC].chip->end(MSP_INT_CIC);
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -21,8 +21,10 @@
|
||||||
#include <msp_slp_int.h>
|
#include <msp_slp_int.h>
|
||||||
#include <msp_regs.h>
|
#include <msp_regs.h>
|
||||||
|
|
||||||
static inline void unmask_msp_slp_irq(unsigned int irq)
|
static inline void unmask_msp_slp_irq(struct irq_data *d)
|
||||||
{
|
{
|
||||||
|
unsigned int irq = d->irq;
|
||||||
|
|
||||||
/* check for PER interrupt range */
|
/* check for PER interrupt range */
|
||||||
if (irq < MSP_PER_INTBASE)
|
if (irq < MSP_PER_INTBASE)
|
||||||
*SLP_INT_MSK_REG |= (1 << (irq - MSP_SLP_INTBASE));
|
*SLP_INT_MSK_REG |= (1 << (irq - MSP_SLP_INTBASE));
|
||||||
|
@ -30,8 +32,10 @@ static inline void unmask_msp_slp_irq(unsigned int irq)
|
||||||
*PER_INT_MSK_REG |= (1 << (irq - MSP_PER_INTBASE));
|
*PER_INT_MSK_REG |= (1 << (irq - MSP_PER_INTBASE));
|
||||||
}
|
}
|
||||||
|
|
||||||
static inline void mask_msp_slp_irq(unsigned int irq)
|
static inline void mask_msp_slp_irq(struct irq_data *d)
|
||||||
{
|
{
|
||||||
|
unsigned int irq = d->irq;
|
||||||
|
|
||||||
/* check for PER interrupt range */
|
/* check for PER interrupt range */
|
||||||
if (irq < MSP_PER_INTBASE)
|
if (irq < MSP_PER_INTBASE)
|
||||||
*SLP_INT_MSK_REG &= ~(1 << (irq - MSP_SLP_INTBASE));
|
*SLP_INT_MSK_REG &= ~(1 << (irq - MSP_SLP_INTBASE));
|
||||||
|
@ -43,8 +47,10 @@ static inline void mask_msp_slp_irq(unsigned int irq)
|
||||||
* While we ack the interrupt interrupts are disabled and thus we don't need
|
* While we ack the interrupt interrupts are disabled and thus we don't need
|
||||||
* to deal with concurrency issues. Same for msp_slp_irq_end.
|
* to deal with concurrency issues. Same for msp_slp_irq_end.
|
||||||
*/
|
*/
|
||||||
static inline void ack_msp_slp_irq(unsigned int irq)
|
static inline void ack_msp_slp_irq(struct irq_data *d)
|
||||||
{
|
{
|
||||||
|
unsigned int irq = d->irq;
|
||||||
|
|
||||||
/* check for PER interrupt range */
|
/* check for PER interrupt range */
|
||||||
if (irq < MSP_PER_INTBASE)
|
if (irq < MSP_PER_INTBASE)
|
||||||
*SLP_INT_STS_REG = (1 << (irq - MSP_SLP_INTBASE));
|
*SLP_INT_STS_REG = (1 << (irq - MSP_SLP_INTBASE));
|
||||||
|
@ -54,9 +60,9 @@ static inline void ack_msp_slp_irq(unsigned int irq)
|
||||||
|
|
||||||
static struct irq_chip msp_slp_irq_controller = {
|
static struct irq_chip msp_slp_irq_controller = {
|
||||||
.name = "MSP_SLP",
|
.name = "MSP_SLP",
|
||||||
.ack = ack_msp_slp_irq,
|
.irq_ack = ack_msp_slp_irq,
|
||||||
.mask = mask_msp_slp_irq,
|
.irq_mask = mask_msp_slp_irq,
|
||||||
.unmask = unmask_msp_slp_irq,
|
.irq_unmask = unmask_msp_slp_irq,
|
||||||
};
|
};
|
||||||
|
|
||||||
void __init msp_slp_irq_init(void)
|
void __init msp_slp_irq_init(void)
|
||||||
|
|
Loading…
Reference in New Issue