2012-06-14 01:01:28 +08:00
|
|
|
/*
|
|
|
|
* Marvell Armada 370 and Armada XP SoC IRQ handling
|
|
|
|
*
|
|
|
|
* Copyright (C) 2012 Marvell
|
|
|
|
*
|
|
|
|
* Lior Amsalem <alior@marvell.com>
|
|
|
|
* Gregory CLEMENT <gregory.clement@free-electrons.com>
|
|
|
|
* Thomas Petazzoni <thomas.petazzoni@free-electrons.com>
|
|
|
|
* Ben Dooks <ben.dooks@codethink.co.uk>
|
|
|
|
*
|
|
|
|
* This file is licensed under the terms of the GNU General Public
|
|
|
|
* License version 2. This program is licensed "as is" without any
|
|
|
|
* warranty of any kind, whether express or implied.
|
|
|
|
*/
|
|
|
|
|
|
|
|
#include <linux/kernel.h>
|
|
|
|
#include <linux/module.h>
|
|
|
|
#include <linux/init.h>
|
|
|
|
#include <linux/irq.h>
|
|
|
|
#include <linux/interrupt.h>
|
2015-07-08 05:11:46 +08:00
|
|
|
#include <linux/irqchip.h>
|
2014-02-11 04:00:02 +08:00
|
|
|
#include <linux/irqchip/chained_irq.h>
|
2014-04-14 21:54:02 +08:00
|
|
|
#include <linux/cpu.h>
|
2012-06-14 01:01:28 +08:00
|
|
|
#include <linux/io.h>
|
|
|
|
#include <linux/of_address.h>
|
|
|
|
#include <linux/of_irq.h>
|
2013-08-10 04:27:11 +08:00
|
|
|
#include <linux/of_pci.h>
|
2012-06-14 01:01:28 +08:00
|
|
|
#include <linux/irqdomain.h>
|
2013-08-10 04:27:11 +08:00
|
|
|
#include <linux/slab.h>
|
2014-11-22 00:00:00 +08:00
|
|
|
#include <linux/syscore_ops.h>
|
2013-08-10 04:27:11 +08:00
|
|
|
#include <linux/msi.h>
|
2012-06-14 01:01:28 +08:00
|
|
|
#include <asm/mach/arch.h>
|
|
|
|
#include <asm/exception.h>
|
2012-08-02 16:19:12 +08:00
|
|
|
#include <asm/smp_plat.h>
|
2013-04-10 05:26:15 +08:00
|
|
|
#include <asm/mach/irq.h>
|
|
|
|
|
2017-05-18 16:07:38 +08:00
|
|
|
/*
|
|
|
|
* Overall diagram of the Armada XP interrupt controller:
|
|
|
|
*
|
|
|
|
* To CPU 0 To CPU 1
|
|
|
|
*
|
|
|
|
* /\ /\
|
|
|
|
* || ||
|
|
|
|
* +---------------+ +---------------+
|
|
|
|
* | | | |
|
|
|
|
* | per-CPU | | per-CPU |
|
|
|
|
* | mask/unmask | | mask/unmask |
|
|
|
|
* | CPU0 | | CPU1 |
|
|
|
|
* | | | |
|
|
|
|
* +---------------+ +---------------+
|
|
|
|
* /\ /\
|
|
|
|
* || ||
|
|
|
|
* \\_______________________//
|
|
|
|
* ||
|
|
|
|
* +-------------------+
|
|
|
|
* | |
|
|
|
|
* | Global interrupt |
|
|
|
|
* | mask/unmask |
|
|
|
|
* | |
|
|
|
|
* +-------------------+
|
|
|
|
* /\
|
|
|
|
* ||
|
|
|
|
* interrupt from
|
|
|
|
* device
|
|
|
|
*
|
|
|
|
* The "global interrupt mask/unmask" is modified using the
|
|
|
|
* ARMADA_370_XP_INT_SET_ENABLE_OFFS and
|
|
|
|
* ARMADA_370_XP_INT_CLEAR_ENABLE_OFFS registers, which are relative
|
|
|
|
* to "main_int_base".
|
|
|
|
*
|
|
|
|
* The "per-CPU mask/unmask" is modified using the
|
|
|
|
* ARMADA_370_XP_INT_SET_MASK_OFFS and
|
|
|
|
* ARMADA_370_XP_INT_CLEAR_MASK_OFFS registers, which are relative to
|
|
|
|
* "per_cpu_int_base". This base address points to a special address,
|
|
|
|
* which automatically accesses the registers of the current CPU.
|
|
|
|
*
|
|
|
|
* The per-CPU mask/unmask can also be adjusted using the global
|
|
|
|
* per-interrupt ARMADA_370_XP_INT_SOURCE_CTL register, which we use
|
|
|
|
* to configure interrupt affinity.
|
|
|
|
*
|
|
|
|
* Due to this model, all interrupts need to be mask/unmasked at two
|
|
|
|
* different levels: at the global level and at the per-CPU level.
|
|
|
|
*
|
|
|
|
* This driver takes the following approach to deal with this:
|
|
|
|
*
|
|
|
|
* - For global interrupts:
|
|
|
|
*
|
|
|
|
* At ->map() time, a global interrupt is unmasked at the per-CPU
|
|
|
|
* mask/unmask level. It is therefore unmasked at this level for
|
|
|
|
* the current CPU, running the ->map() code. This allows to have
|
|
|
|
* the interrupt unmasked at this level in non-SMP
|
|
|
|
* configurations. In SMP configurations, the ->set_affinity()
|
|
|
|
* callback is called, which using the
|
|
|
|
* ARMADA_370_XP_INT_SOURCE_CTL() readjusts the per-CPU mask/unmask
|
|
|
|
* for the interrupt.
|
|
|
|
*
|
|
|
|
* The ->mask() and ->unmask() operations only mask/unmask the
|
|
|
|
* interrupt at the "global" level.
|
|
|
|
*
|
|
|
|
* So, a global interrupt is enabled at the per-CPU level as soon
|
|
|
|
* as it is mapped. At run time, the masking/unmasking takes place
|
|
|
|
* at the global level.
|
|
|
|
*
|
|
|
|
* - For per-CPU interrupts
|
|
|
|
*
|
|
|
|
* At ->map() time, a per-CPU interrupt is unmasked at the global
|
|
|
|
* mask/unmask level.
|
|
|
|
*
|
|
|
|
* The ->mask() and ->unmask() operations mask/unmask the interrupt
|
|
|
|
* at the per-CPU level.
|
|
|
|
*
|
|
|
|
* So, a per-CPU interrupt is enabled at the global level as soon
|
|
|
|
* as it is mapped. At run time, the masking/unmasking takes place
|
|
|
|
* at the per-CPU level.
|
|
|
|
*/
|
2012-06-14 01:01:28 +08:00
|
|
|
|
2017-05-18 16:07:37 +08:00
|
|
|
/* Registers relative to main_int_base */
|
2012-06-05 00:50:12 +08:00
|
|
|
#define ARMADA_370_XP_INT_CONTROL (0x00)
|
2017-05-18 16:07:37 +08:00
|
|
|
#define ARMADA_370_XP_SW_TRIG_INT_OFFS (0x04)
|
2012-06-14 01:01:28 +08:00
|
|
|
#define ARMADA_370_XP_INT_SET_ENABLE_OFFS (0x30)
|
|
|
|
#define ARMADA_370_XP_INT_CLEAR_ENABLE_OFFS (0x34)
|
2012-12-06 04:43:23 +08:00
|
|
|
#define ARMADA_370_XP_INT_SOURCE_CTL(irq) (0x100 + irq*4)
|
2014-03-05 04:43:41 +08:00
|
|
|
#define ARMADA_370_XP_INT_SOURCE_CPU_MASK 0xF
|
2014-09-25 19:17:19 +08:00
|
|
|
#define ARMADA_370_XP_INT_IRQ_FIQ_MASK(cpuid) ((BIT(0) | BIT(8)) << cpuid)
|
2012-06-14 01:01:28 +08:00
|
|
|
|
2017-05-18 16:07:37 +08:00
|
|
|
/* Registers relative to per_cpu_int_base */
|
|
|
|
#define ARMADA_370_XP_IN_DRBEL_CAUSE_OFFS (0x08)
|
|
|
|
#define ARMADA_370_XP_IN_DRBEL_MSK_OFFS (0x0c)
|
2014-02-11 04:00:02 +08:00
|
|
|
#define ARMADA_375_PPI_CAUSE (0x10)
|
2017-05-18 16:07:37 +08:00
|
|
|
#define ARMADA_370_XP_CPU_INTACK_OFFS (0x44)
|
|
|
|
#define ARMADA_370_XP_INT_SET_MASK_OFFS (0x48)
|
|
|
|
#define ARMADA_370_XP_INT_CLEAR_MASK_OFFS (0x4C)
|
|
|
|
#define ARMADA_370_XP_INT_FABRIC_MASK_OFFS (0x54)
|
|
|
|
#define ARMADA_370_XP_INT_CAUSE_PERF(cpu) (1 << cpu)
|
2012-08-02 16:19:12 +08:00
|
|
|
|
2012-12-06 04:43:23 +08:00
|
|
|
#define ARMADA_370_XP_MAX_PER_CPU_IRQS (28)
|
|
|
|
|
2013-04-10 05:26:17 +08:00
|
|
|
#define IPI_DOORBELL_START (0)
|
|
|
|
#define IPI_DOORBELL_END (8)
|
|
|
|
#define IPI_DOORBELL_MASK 0xFF
|
2013-08-10 04:27:11 +08:00
|
|
|
#define PCI_MSI_DOORBELL_START (16)
|
|
|
|
#define PCI_MSI_DOORBELL_NR (16)
|
|
|
|
#define PCI_MSI_DOORBELL_END (32)
|
|
|
|
#define PCI_MSI_DOORBELL_MASK 0xFFFF0000
|
2012-08-02 16:19:12 +08:00
|
|
|
|
2012-06-14 01:01:28 +08:00
|
|
|
static void __iomem *per_cpu_int_base;
|
|
|
|
static void __iomem *main_int_base;
|
|
|
|
static struct irq_domain *armada_370_xp_mpic_domain;
|
2014-11-22 00:00:00 +08:00
|
|
|
static u32 doorbell_mask_reg;
|
irqchip: armada-370-xp: Fix chained per-cpu interrupts
On the Cortex-A9-based Armada SoCs, the MPIC is not the primary interrupt
controller. Yet, it still has to handle some per-cpu interrupt.
To do so, it is chained with the GIC using a per-cpu interrupt. However, the
current code only call irq_set_chained_handler, which is called and enable that
interrupt only on the boot CPU, which means that the parent per-CPU interrupt
is never unmasked on the secondary CPUs, preventing the per-CPU interrupt to
actually work as expected.
This was not seen until now since the only MPIC PPI users were the Marvell
timers that were not working, but not used either since the system use the ARM
TWD by default, and the ethernet controllers, that are faking there interrupts
as SPI, and don't really expect to have interrupts on the secondary cores
anyway.
Add a CPU notifier that will enable the PPI on the secondary cores when they
are brought up.
Cc: <stable@vger.kernel.org> # 3.15+
Signed-off-by: Maxime Ripard <maxime.ripard@free-electrons.com>
Acked-by: Gregory CLEMENT <gregory.clement@free-electrons.com>
Link: https://lkml.kernel.org/r/1425378443-28822-1-git-send-email-maxime.ripard@free-electrons.com
Signed-off-by: Jason Cooper <jason@lakedaemon.net>
2015-03-03 18:27:23 +08:00
|
|
|
static int parent_irq;
|
2013-08-10 04:27:11 +08:00
|
|
|
#ifdef CONFIG_PCI_MSI
|
|
|
|
static struct irq_domain *armada_370_xp_msi_domain;
|
2016-02-10 22:46:57 +08:00
|
|
|
static struct irq_domain *armada_370_xp_msi_inner_domain;
|
2013-08-10 04:27:11 +08:00
|
|
|
static DECLARE_BITMAP(msi_used, PCI_MSI_DOORBELL_NR);
|
|
|
|
static DEFINE_MUTEX(msi_used_lock);
|
|
|
|
static phys_addr_t msi_doorbell_addr;
|
|
|
|
#endif
|
2012-06-14 01:01:28 +08:00
|
|
|
|
2015-03-03 18:43:15 +08:00
|
|
|
static inline bool is_percpu_irq(irq_hw_number_t irq)
|
|
|
|
{
|
2015-09-26 00:09:34 +08:00
|
|
|
if (irq <= ARMADA_370_XP_MAX_PER_CPU_IRQS)
|
2015-03-03 18:43:15 +08:00
|
|
|
return true;
|
2015-09-26 00:09:34 +08:00
|
|
|
|
|
|
|
return false;
|
2015-03-03 18:43:15 +08:00
|
|
|
}
|
|
|
|
|
2012-12-06 04:43:23 +08:00
|
|
|
/*
|
|
|
|
* In SMP mode:
|
|
|
|
* For shared global interrupts, mask/unmask global enable bit
|
2013-03-16 06:34:04 +08:00
|
|
|
* For CPU interrupts, mask/unmask the calling CPU's bit
|
2012-12-06 04:43:23 +08:00
|
|
|
*/
|
2012-06-14 01:01:28 +08:00
|
|
|
static void armada_370_xp_irq_mask(struct irq_data *d)
|
|
|
|
{
|
2012-12-06 04:43:23 +08:00
|
|
|
irq_hw_number_t hwirq = irqd_to_hwirq(d);
|
|
|
|
|
2015-03-03 18:43:15 +08:00
|
|
|
if (!is_percpu_irq(hwirq))
|
2012-12-06 04:43:23 +08:00
|
|
|
writel(hwirq, main_int_base +
|
|
|
|
ARMADA_370_XP_INT_CLEAR_ENABLE_OFFS);
|
|
|
|
else
|
|
|
|
writel(hwirq, per_cpu_int_base +
|
|
|
|
ARMADA_370_XP_INT_SET_MASK_OFFS);
|
2012-06-14 01:01:28 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
static void armada_370_xp_irq_unmask(struct irq_data *d)
|
|
|
|
{
|
2012-12-06 04:43:23 +08:00
|
|
|
irq_hw_number_t hwirq = irqd_to_hwirq(d);
|
|
|
|
|
2015-03-03 18:43:15 +08:00
|
|
|
if (!is_percpu_irq(hwirq))
|
2012-12-06 04:43:23 +08:00
|
|
|
writel(hwirq, main_int_base +
|
|
|
|
ARMADA_370_XP_INT_SET_ENABLE_OFFS);
|
|
|
|
else
|
|
|
|
writel(hwirq, per_cpu_int_base +
|
|
|
|
ARMADA_370_XP_INT_CLEAR_MASK_OFFS);
|
2012-06-14 01:01:28 +08:00
|
|
|
}
|
|
|
|
|
2013-08-10 04:27:11 +08:00
|
|
|
#ifdef CONFIG_PCI_MSI
|
|
|
|
|
2016-02-10 22:46:57 +08:00
|
|
|
static struct irq_chip armada_370_xp_msi_irq_chip = {
|
2016-02-10 22:46:59 +08:00
|
|
|
.name = "MPIC MSI",
|
2016-02-10 22:46:57 +08:00
|
|
|
.irq_mask = pci_msi_mask_irq,
|
|
|
|
.irq_unmask = pci_msi_unmask_irq,
|
|
|
|
};
|
2013-08-10 04:27:11 +08:00
|
|
|
|
2016-02-10 22:46:57 +08:00
|
|
|
static struct msi_domain_info armada_370_xp_msi_domain_info = {
|
2016-02-10 22:47:00 +08:00
|
|
|
.flags = (MSI_FLAG_USE_DEF_DOM_OPS | MSI_FLAG_USE_DEF_CHIP_OPS |
|
2017-08-18 20:59:26 +08:00
|
|
|
MSI_FLAG_MULTI_PCI_MSI | MSI_FLAG_PCI_MSIX),
|
2016-02-10 22:46:57 +08:00
|
|
|
.chip = &armada_370_xp_msi_irq_chip,
|
|
|
|
};
|
2013-08-10 04:27:11 +08:00
|
|
|
|
2016-02-10 22:46:57 +08:00
|
|
|
static void armada_370_xp_compose_msi_msg(struct irq_data *data, struct msi_msg *msg)
|
2013-08-10 04:27:11 +08:00
|
|
|
{
|
2022-04-22 12:35:32 +08:00
|
|
|
unsigned int cpu = cpumask_first(irq_data_get_effective_affinity_mask(data));
|
|
|
|
|
2016-02-10 22:46:57 +08:00
|
|
|
msg->address_lo = lower_32_bits(msi_doorbell_addr);
|
|
|
|
msg->address_hi = upper_32_bits(msi_doorbell_addr);
|
2022-04-22 12:35:32 +08:00
|
|
|
msg->data = BIT(cpu + 8) | (data->hwirq + PCI_MSI_DOORBELL_START);
|
2013-08-10 04:27:11 +08:00
|
|
|
}
|
|
|
|
|
2016-02-10 22:46:57 +08:00
|
|
|
static int armada_370_xp_msi_set_affinity(struct irq_data *irq_data,
|
|
|
|
const struct cpumask *mask, bool force)
|
2013-08-10 04:27:11 +08:00
|
|
|
{
|
2022-04-22 12:35:32 +08:00
|
|
|
unsigned int cpu;
|
|
|
|
|
|
|
|
if (!force)
|
|
|
|
cpu = cpumask_any_and(mask, cpu_online_mask);
|
|
|
|
else
|
|
|
|
cpu = cpumask_first(mask);
|
|
|
|
|
|
|
|
if (cpu >= nr_cpu_ids)
|
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
irq_data_update_effective_affinity(irq_data, cpumask_of(cpu));
|
|
|
|
|
|
|
|
return IRQ_SET_MASK_OK;
|
2016-02-10 22:46:57 +08:00
|
|
|
}
|
2013-08-10 04:27:11 +08:00
|
|
|
|
2016-02-10 22:46:57 +08:00
|
|
|
static struct irq_chip armada_370_xp_msi_bottom_irq_chip = {
|
2016-02-10 22:46:59 +08:00
|
|
|
.name = "MPIC MSI",
|
2016-02-10 22:46:57 +08:00
|
|
|
.irq_compose_msi_msg = armada_370_xp_compose_msi_msg,
|
|
|
|
.irq_set_affinity = armada_370_xp_msi_set_affinity,
|
|
|
|
};
|
2014-09-08 02:57:54 +08:00
|
|
|
|
2016-02-10 22:46:57 +08:00
|
|
|
static int armada_370_xp_msi_alloc(struct irq_domain *domain, unsigned int virq,
|
|
|
|
unsigned int nr_irqs, void *args)
|
|
|
|
{
|
2016-02-10 22:47:00 +08:00
|
|
|
int hwirq, i;
|
2013-08-10 04:27:11 +08:00
|
|
|
|
2016-02-10 22:46:57 +08:00
|
|
|
mutex_lock(&msi_used_lock);
|
2021-11-25 21:00:57 +08:00
|
|
|
hwirq = bitmap_find_free_region(msi_used, PCI_MSI_DOORBELL_NR,
|
|
|
|
order_base_2(nr_irqs));
|
|
|
|
mutex_unlock(&msi_used_lock);
|
2016-02-10 22:47:00 +08:00
|
|
|
|
2021-11-25 21:00:57 +08:00
|
|
|
if (hwirq < 0)
|
2016-02-10 22:46:57 +08:00
|
|
|
return -ENOSPC;
|
2013-08-10 04:27:11 +08:00
|
|
|
|
2016-02-10 22:47:00 +08:00
|
|
|
for (i = 0; i < nr_irqs; i++) {
|
|
|
|
irq_domain_set_info(domain, virq + i, hwirq + i,
|
|
|
|
&armada_370_xp_msi_bottom_irq_chip,
|
|
|
|
domain->host_data, handle_simple_irq,
|
|
|
|
NULL, NULL);
|
|
|
|
}
|
2013-08-10 04:27:11 +08:00
|
|
|
|
2021-11-25 21:00:56 +08:00
|
|
|
return 0;
|
2013-08-10 04:27:11 +08:00
|
|
|
}
|
|
|
|
|
2016-02-10 22:46:57 +08:00
|
|
|
static void armada_370_xp_msi_free(struct irq_domain *domain,
|
|
|
|
unsigned int virq, unsigned int nr_irqs)
|
2013-08-10 04:27:11 +08:00
|
|
|
{
|
2016-02-10 22:46:57 +08:00
|
|
|
struct irq_data *d = irq_domain_get_irq_data(domain, virq);
|
2013-08-10 04:27:11 +08:00
|
|
|
|
2016-02-10 22:46:57 +08:00
|
|
|
mutex_lock(&msi_used_lock);
|
2021-11-25 21:00:57 +08:00
|
|
|
bitmap_release_region(msi_used, d->hwirq, order_base_2(nr_irqs));
|
2016-02-10 22:46:57 +08:00
|
|
|
mutex_unlock(&msi_used_lock);
|
2013-08-10 04:27:11 +08:00
|
|
|
}
|
|
|
|
|
2016-02-10 22:46:57 +08:00
|
|
|
static const struct irq_domain_ops armada_370_xp_msi_domain_ops = {
|
|
|
|
.alloc = armada_370_xp_msi_alloc,
|
|
|
|
.free = armada_370_xp_msi_free,
|
2013-08-10 04:27:11 +08:00
|
|
|
};
|
|
|
|
|
2022-04-22 12:35:32 +08:00
|
|
|
static void armada_370_xp_msi_reenable_percpu(void)
|
2013-08-10 04:27:11 +08:00
|
|
|
{
|
|
|
|
u32 reg;
|
|
|
|
|
2022-04-22 12:35:32 +08:00
|
|
|
/* Enable MSI doorbell mask and combined cpu local interrupt */
|
|
|
|
reg = readl(per_cpu_int_base + ARMADA_370_XP_IN_DRBEL_MSK_OFFS)
|
|
|
|
| PCI_MSI_DOORBELL_MASK;
|
|
|
|
writel(reg, per_cpu_int_base + ARMADA_370_XP_IN_DRBEL_MSK_OFFS);
|
|
|
|
/* Unmask local doorbell interrupt */
|
|
|
|
writel(1, per_cpu_int_base + ARMADA_370_XP_INT_CLEAR_MASK_OFFS);
|
|
|
|
}
|
|
|
|
|
|
|
|
static int armada_370_xp_msi_init(struct device_node *node,
|
|
|
|
phys_addr_t main_int_phys_base)
|
|
|
|
{
|
2013-08-10 04:27:11 +08:00
|
|
|
msi_doorbell_addr = main_int_phys_base +
|
|
|
|
ARMADA_370_XP_SW_TRIG_INT_OFFS;
|
|
|
|
|
2016-02-10 22:46:57 +08:00
|
|
|
armada_370_xp_msi_inner_domain =
|
|
|
|
irq_domain_add_linear(NULL, PCI_MSI_DOORBELL_NR,
|
|
|
|
&armada_370_xp_msi_domain_ops, NULL);
|
|
|
|
if (!armada_370_xp_msi_inner_domain)
|
2013-08-10 04:27:11 +08:00
|
|
|
return -ENOMEM;
|
|
|
|
|
|
|
|
armada_370_xp_msi_domain =
|
2016-02-10 22:46:57 +08:00
|
|
|
pci_msi_create_irq_domain(of_node_to_fwnode(node),
|
|
|
|
&armada_370_xp_msi_domain_info,
|
|
|
|
armada_370_xp_msi_inner_domain);
|
2013-08-10 04:27:11 +08:00
|
|
|
if (!armada_370_xp_msi_domain) {
|
2016-02-10 22:46:57 +08:00
|
|
|
irq_domain_remove(armada_370_xp_msi_inner_domain);
|
2013-08-10 04:27:11 +08:00
|
|
|
return -ENOMEM;
|
|
|
|
}
|
|
|
|
|
2022-04-22 12:35:32 +08:00
|
|
|
armada_370_xp_msi_reenable_percpu();
|
2013-08-10 04:27:11 +08:00
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
#else
|
2022-04-22 12:35:32 +08:00
|
|
|
static void armada_370_xp_msi_reenable_percpu(void) {}
|
|
|
|
|
2013-08-10 04:27:11 +08:00
|
|
|
static inline int armada_370_xp_msi_init(struct device_node *node,
|
|
|
|
phys_addr_t main_int_phys_base)
|
|
|
|
{
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
2020-06-23 04:23:36 +08:00
|
|
|
static void armada_xp_mpic_perf_init(void)
|
|
|
|
{
|
2022-04-25 19:37:05 +08:00
|
|
|
unsigned long cpuid;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* This Performance Counter Overflow interrupt is specific for
|
|
|
|
* Armada 370 and XP. It is not available on Armada 375, 38x and 39x.
|
|
|
|
*/
|
|
|
|
if (!of_machine_is_compatible("marvell,armada-370-xp"))
|
|
|
|
return;
|
|
|
|
|
|
|
|
cpuid = cpu_logical_map(smp_processor_id());
|
2020-06-23 04:23:36 +08:00
|
|
|
|
|
|
|
/* Enable Performance Counter Overflow interrupts */
|
|
|
|
writel(ARMADA_370_XP_INT_CAUSE_PERF(cpuid),
|
|
|
|
per_cpu_int_base + ARMADA_370_XP_INT_FABRIC_MASK_OFFS);
|
|
|
|
}
|
|
|
|
|
2012-08-02 16:19:12 +08:00
|
|
|
#ifdef CONFIG_SMP
|
2020-06-23 04:23:36 +08:00
|
|
|
static struct irq_domain *ipi_domain;
|
|
|
|
|
|
|
|
static void armada_370_xp_ipi_mask(struct irq_data *d)
|
|
|
|
{
|
|
|
|
u32 reg;
|
|
|
|
reg = readl(per_cpu_int_base + ARMADA_370_XP_IN_DRBEL_MSK_OFFS);
|
|
|
|
reg &= ~BIT(d->hwirq);
|
|
|
|
writel(reg, per_cpu_int_base + ARMADA_370_XP_IN_DRBEL_MSK_OFFS);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void armada_370_xp_ipi_unmask(struct irq_data *d)
|
|
|
|
{
|
|
|
|
u32 reg;
|
|
|
|
reg = readl(per_cpu_int_base + ARMADA_370_XP_IN_DRBEL_MSK_OFFS);
|
|
|
|
reg |= BIT(d->hwirq);
|
|
|
|
writel(reg, per_cpu_int_base + ARMADA_370_XP_IN_DRBEL_MSK_OFFS);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void armada_370_xp_ipi_send_mask(struct irq_data *d,
|
|
|
|
const struct cpumask *mask)
|
|
|
|
{
|
|
|
|
unsigned long map = 0;
|
|
|
|
int cpu;
|
|
|
|
|
|
|
|
/* Convert our logical CPU mask into a physical one. */
|
|
|
|
for_each_cpu(cpu, mask)
|
|
|
|
map |= 1 << cpu_logical_map(cpu);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Ensure that stores to Normal memory are visible to the
|
|
|
|
* other CPUs before issuing the IPI.
|
|
|
|
*/
|
|
|
|
dsb();
|
|
|
|
|
|
|
|
/* submit softirq */
|
|
|
|
writel((map << 8) | d->hwirq, main_int_base +
|
|
|
|
ARMADA_370_XP_SW_TRIG_INT_OFFS);
|
|
|
|
}
|
|
|
|
|
2021-09-22 21:19:41 +08:00
|
|
|
static void armada_370_xp_ipi_ack(struct irq_data *d)
|
2020-06-23 04:23:36 +08:00
|
|
|
{
|
|
|
|
writel(~BIT(d->hwirq), per_cpu_int_base + ARMADA_370_XP_IN_DRBEL_CAUSE_OFFS);
|
|
|
|
}
|
|
|
|
|
|
|
|
static struct irq_chip ipi_irqchip = {
|
|
|
|
.name = "IPI",
|
2021-09-22 21:19:41 +08:00
|
|
|
.irq_ack = armada_370_xp_ipi_ack,
|
2020-06-23 04:23:36 +08:00
|
|
|
.irq_mask = armada_370_xp_ipi_mask,
|
|
|
|
.irq_unmask = armada_370_xp_ipi_unmask,
|
|
|
|
.ipi_send_mask = armada_370_xp_ipi_send_mask,
|
|
|
|
};
|
|
|
|
|
|
|
|
static int armada_370_xp_ipi_alloc(struct irq_domain *d,
|
|
|
|
unsigned int virq,
|
|
|
|
unsigned int nr_irqs, void *args)
|
|
|
|
{
|
|
|
|
int i;
|
|
|
|
|
|
|
|
for (i = 0; i < nr_irqs; i++) {
|
|
|
|
irq_set_percpu_devid(virq + i);
|
|
|
|
irq_domain_set_info(d, virq + i, i, &ipi_irqchip,
|
|
|
|
d->host_data,
|
2020-11-09 17:41:18 +08:00
|
|
|
handle_percpu_devid_irq,
|
2020-06-23 04:23:36 +08:00
|
|
|
NULL, NULL);
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void armada_370_xp_ipi_free(struct irq_domain *d,
|
|
|
|
unsigned int virq,
|
|
|
|
unsigned int nr_irqs)
|
|
|
|
{
|
|
|
|
/* Not freeing IPIs */
|
|
|
|
}
|
|
|
|
|
|
|
|
static const struct irq_domain_ops ipi_domain_ops = {
|
|
|
|
.alloc = armada_370_xp_ipi_alloc,
|
|
|
|
.free = armada_370_xp_ipi_free,
|
|
|
|
};
|
|
|
|
|
|
|
|
static void ipi_resume(void)
|
|
|
|
{
|
|
|
|
int i;
|
|
|
|
|
|
|
|
for (i = 0; i < IPI_DOORBELL_END; i++) {
|
|
|
|
int irq;
|
|
|
|
|
|
|
|
irq = irq_find_mapping(ipi_domain, i);
|
|
|
|
if (irq <= 0)
|
|
|
|
continue;
|
|
|
|
if (irq_percpu_is_enabled(irq)) {
|
|
|
|
struct irq_data *d;
|
|
|
|
d = irq_domain_get_irq_data(ipi_domain, irq);
|
|
|
|
armada_370_xp_ipi_unmask(d);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static __init void armada_xp_ipi_init(struct device_node *node)
|
|
|
|
{
|
|
|
|
int base_ipi;
|
|
|
|
|
|
|
|
ipi_domain = irq_domain_create_linear(of_node_to_fwnode(node),
|
|
|
|
IPI_DOORBELL_END,
|
|
|
|
&ipi_domain_ops, NULL);
|
|
|
|
if (WARN_ON(!ipi_domain))
|
|
|
|
return;
|
|
|
|
|
|
|
|
irq_domain_update_bus_token(ipi_domain, DOMAIN_BUS_IPI);
|
2022-12-13 22:08:43 +08:00
|
|
|
base_ipi = irq_domain_alloc_irqs(ipi_domain, IPI_DOORBELL_END, NUMA_NO_NODE, NULL);
|
2020-06-23 04:23:36 +08:00
|
|
|
if (WARN_ON(!base_ipi))
|
|
|
|
return;
|
|
|
|
|
|
|
|
set_smp_ipi_range(base_ipi, IPI_DOORBELL_END);
|
|
|
|
}
|
|
|
|
|
2014-01-21 05:52:05 +08:00
|
|
|
static DEFINE_RAW_SPINLOCK(irq_controller_lock);
|
|
|
|
|
2012-08-02 16:19:12 +08:00
|
|
|
static int armada_xp_set_affinity(struct irq_data *d,
|
|
|
|
const struct cpumask *mask_val, bool force)
|
|
|
|
{
|
2012-12-06 04:43:23 +08:00
|
|
|
irq_hw_number_t hwirq = irqd_to_hwirq(d);
|
2014-03-05 04:43:41 +08:00
|
|
|
unsigned long reg, mask;
|
2012-12-06 04:43:23 +08:00
|
|
|
int cpu;
|
|
|
|
|
2014-03-05 04:43:41 +08:00
|
|
|
/* Select a single core from the affinity mask which is online */
|
|
|
|
cpu = cpumask_any_and(mask_val, cpu_online_mask);
|
|
|
|
mask = 1UL << cpu_logical_map(cpu);
|
2012-12-06 04:43:23 +08:00
|
|
|
|
|
|
|
raw_spin_lock(&irq_controller_lock);
|
|
|
|
reg = readl(main_int_base + ARMADA_370_XP_INT_SOURCE_CTL(hwirq));
|
2014-03-05 04:43:41 +08:00
|
|
|
reg = (reg & (~ARMADA_370_XP_INT_SOURCE_CPU_MASK)) | mask;
|
2012-12-06 04:43:23 +08:00
|
|
|
writel(reg, main_int_base + ARMADA_370_XP_INT_SOURCE_CTL(hwirq));
|
|
|
|
raw_spin_unlock(&irq_controller_lock);
|
|
|
|
|
2017-08-18 16:39:19 +08:00
|
|
|
irq_data_update_effective_affinity(d, cpumask_of(cpu));
|
|
|
|
|
2014-10-24 19:59:16 +08:00
|
|
|
return IRQ_SET_MASK_OK;
|
2012-08-02 16:19:12 +08:00
|
|
|
}
|
2012-06-14 01:01:28 +08:00
|
|
|
|
2014-04-14 21:54:02 +08:00
|
|
|
static void armada_xp_mpic_smp_cpu_init(void)
|
2012-08-02 16:19:12 +08:00
|
|
|
{
|
2014-05-31 04:18:18 +08:00
|
|
|
u32 control;
|
|
|
|
int nr_irqs, i;
|
|
|
|
|
|
|
|
control = readl(main_int_base + ARMADA_370_XP_INT_CONTROL);
|
|
|
|
nr_irqs = (control >> 2) & 0x3ff;
|
|
|
|
|
|
|
|
for (i = 0; i < nr_irqs; i++)
|
|
|
|
writel(i, per_cpu_int_base + ARMADA_370_XP_INT_SET_MASK_OFFS);
|
|
|
|
|
2020-06-23 04:23:36 +08:00
|
|
|
/* Disable all IPIs */
|
|
|
|
writel(0, per_cpu_int_base + ARMADA_370_XP_IN_DRBEL_MSK_OFFS);
|
|
|
|
|
2012-08-02 16:19:12 +08:00
|
|
|
/* Clear pending IPIs */
|
|
|
|
writel(0, per_cpu_int_base + ARMADA_370_XP_IN_DRBEL_CAUSE_OFFS);
|
|
|
|
|
|
|
|
/* Unmask IPI interrupt */
|
|
|
|
writel(0, per_cpu_int_base + ARMADA_370_XP_INT_CLEAR_MASK_OFFS);
|
|
|
|
}
|
2014-04-14 21:54:02 +08:00
|
|
|
|
irqchip/armada-370-xp: Re-enable per-CPU interrupts at resume time
Commit d17cab4451df1 ("irqchip: Kill off set_irq_flags usage") changed
the code of armada_370_xp_mpic_irq_map() from using set_irq_flags() to
irq_set_probe().
While the commit log seems to imply that there are no functional
changes, there are indeed functional changes introduced by this commit:
the IRQ_NOAUTOEN flag is no longer cleared. This functional change
caused a regression on Armada XP, which no longer works properly after
suspend/resume because per-CPU interrupts remain disabled. This
regression was temporarly worked around in commit
353d6d6c82e5d ("irqchip/armada-370-xp: Fix regression by clearing
IRQ_NOAUTOEN"), but it is not the most satisfying solution. This commit
implements the solution that was initially discussed with Thomas
Gleixner.
Due to how the hardware registers work, the irq-armada-370-xp cannot
simply save/restore a bunch of registers at suspend/resume to make sure
that the interrupts remain in the same state after resuming. Therefore,
it relies on the kernel to say whether the interrupt is disabled or not,
using the irqd_irq_disabled() function. This was all working fine while
the IRQ_NOAUTOEN flag was cleared.
With the change introduced by Rob Herring in d17cab4451df1, the
IRQ_NOAUTOEN flag is now set for all interrupts. irqd_irq_disabled()
returns false for per-CPU interrupts, and therefore our per-CPU
interrupts are no longer re-enabled after resume.
This commit fixes that by using irqd_irq_disabled() only for global
interrupts, and using the newly introduced irq_percpu_is_enabled() for
per-CPU interrupts.
Also, it fixes a related problems that per-CPU interrupts were only
re-enabled on the boot CPU and not other CPUs. Until now this wasn't a
problem since on this platform, only the local timers are using per-CPU
interrupts and the local timers of secondary CPUs are turned off/on
during CPU hotplug before suspend, after after resume. However, since
Linux 4.4, we are also be using per-CPU interrupts for the network
controller, so we need to properly restore the per-CPU interrupts on
secondary CPUs as well.
Signed-off-by: Thomas Petazzoni <thomas.petazzoni@free-electrons.com>
Signed-off-by: Marc Zyngier <marc.zyngier@arm.com>
2017-05-18 16:07:39 +08:00
|
|
|
static void armada_xp_mpic_reenable_percpu(void)
|
|
|
|
{
|
|
|
|
unsigned int irq;
|
|
|
|
|
|
|
|
/* Re-enable per-CPU interrupts that were enabled before suspend */
|
|
|
|
for (irq = 0; irq < ARMADA_370_XP_MAX_PER_CPU_IRQS; irq++) {
|
|
|
|
struct irq_data *data;
|
|
|
|
int virq;
|
|
|
|
|
|
|
|
virq = irq_linear_revmap(armada_370_xp_mpic_domain, irq);
|
|
|
|
if (virq == 0)
|
|
|
|
continue;
|
|
|
|
|
|
|
|
data = irq_get_irq_data(virq);
|
|
|
|
|
|
|
|
if (!irq_percpu_is_enabled(virq))
|
|
|
|
continue;
|
|
|
|
|
|
|
|
armada_370_xp_irq_unmask(data);
|
|
|
|
}
|
2020-06-23 04:23:36 +08:00
|
|
|
|
|
|
|
ipi_resume();
|
2022-04-22 12:35:32 +08:00
|
|
|
|
|
|
|
armada_370_xp_msi_reenable_percpu();
|
irqchip/armada-370-xp: Re-enable per-CPU interrupts at resume time
Commit d17cab4451df1 ("irqchip: Kill off set_irq_flags usage") changed
the code of armada_370_xp_mpic_irq_map() from using set_irq_flags() to
irq_set_probe().
While the commit log seems to imply that there are no functional
changes, there are indeed functional changes introduced by this commit:
the IRQ_NOAUTOEN flag is no longer cleared. This functional change
caused a regression on Armada XP, which no longer works properly after
suspend/resume because per-CPU interrupts remain disabled. This
regression was temporarly worked around in commit
353d6d6c82e5d ("irqchip/armada-370-xp: Fix regression by clearing
IRQ_NOAUTOEN"), but it is not the most satisfying solution. This commit
implements the solution that was initially discussed with Thomas
Gleixner.
Due to how the hardware registers work, the irq-armada-370-xp cannot
simply save/restore a bunch of registers at suspend/resume to make sure
that the interrupts remain in the same state after resuming. Therefore,
it relies on the kernel to say whether the interrupt is disabled or not,
using the irqd_irq_disabled() function. This was all working fine while
the IRQ_NOAUTOEN flag was cleared.
With the change introduced by Rob Herring in d17cab4451df1, the
IRQ_NOAUTOEN flag is now set for all interrupts. irqd_irq_disabled()
returns false for per-CPU interrupts, and therefore our per-CPU
interrupts are no longer re-enabled after resume.
This commit fixes that by using irqd_irq_disabled() only for global
interrupts, and using the newly introduced irq_percpu_is_enabled() for
per-CPU interrupts.
Also, it fixes a related problems that per-CPU interrupts were only
re-enabled on the boot CPU and not other CPUs. Until now this wasn't a
problem since on this platform, only the local timers are using per-CPU
interrupts and the local timers of secondary CPUs are turned off/on
during CPU hotplug before suspend, after after resume. However, since
Linux 4.4, we are also be using per-CPU interrupts for the network
controller, so we need to properly restore the per-CPU interrupts on
secondary CPUs as well.
Signed-off-by: Thomas Petazzoni <thomas.petazzoni@free-electrons.com>
Signed-off-by: Marc Zyngier <marc.zyngier@arm.com>
2017-05-18 16:07:39 +08:00
|
|
|
}
|
|
|
|
|
2016-07-14 01:16:07 +08:00
|
|
|
static int armada_xp_mpic_starting_cpu(unsigned int cpu)
|
2014-04-14 21:54:02 +08:00
|
|
|
{
|
2016-07-14 01:16:07 +08:00
|
|
|
armada_xp_mpic_perf_init();
|
|
|
|
armada_xp_mpic_smp_cpu_init();
|
irqchip/armada-370-xp: Re-enable per-CPU interrupts at resume time
Commit d17cab4451df1 ("irqchip: Kill off set_irq_flags usage") changed
the code of armada_370_xp_mpic_irq_map() from using set_irq_flags() to
irq_set_probe().
While the commit log seems to imply that there are no functional
changes, there are indeed functional changes introduced by this commit:
the IRQ_NOAUTOEN flag is no longer cleared. This functional change
caused a regression on Armada XP, which no longer works properly after
suspend/resume because per-CPU interrupts remain disabled. This
regression was temporarly worked around in commit
353d6d6c82e5d ("irqchip/armada-370-xp: Fix regression by clearing
IRQ_NOAUTOEN"), but it is not the most satisfying solution. This commit
implements the solution that was initially discussed with Thomas
Gleixner.
Due to how the hardware registers work, the irq-armada-370-xp cannot
simply save/restore a bunch of registers at suspend/resume to make sure
that the interrupts remain in the same state after resuming. Therefore,
it relies on the kernel to say whether the interrupt is disabled or not,
using the irqd_irq_disabled() function. This was all working fine while
the IRQ_NOAUTOEN flag was cleared.
With the change introduced by Rob Herring in d17cab4451df1, the
IRQ_NOAUTOEN flag is now set for all interrupts. irqd_irq_disabled()
returns false for per-CPU interrupts, and therefore our per-CPU
interrupts are no longer re-enabled after resume.
This commit fixes that by using irqd_irq_disabled() only for global
interrupts, and using the newly introduced irq_percpu_is_enabled() for
per-CPU interrupts.
Also, it fixes a related problems that per-CPU interrupts were only
re-enabled on the boot CPU and not other CPUs. Until now this wasn't a
problem since on this platform, only the local timers are using per-CPU
interrupts and the local timers of secondary CPUs are turned off/on
during CPU hotplug before suspend, after after resume. However, since
Linux 4.4, we are also be using per-CPU interrupts for the network
controller, so we need to properly restore the per-CPU interrupts on
secondary CPUs as well.
Signed-off-by: Thomas Petazzoni <thomas.petazzoni@free-electrons.com>
Signed-off-by: Marc Zyngier <marc.zyngier@arm.com>
2017-05-18 16:07:39 +08:00
|
|
|
armada_xp_mpic_reenable_percpu();
|
2016-07-14 01:16:07 +08:00
|
|
|
return 0;
|
2014-04-14 21:54:02 +08:00
|
|
|
}
|
|
|
|
|
2016-07-14 01:16:07 +08:00
|
|
|
static int mpic_cascaded_starting_cpu(unsigned int cpu)
|
irqchip: armada-370-xp: Fix chained per-cpu interrupts
On the Cortex-A9-based Armada SoCs, the MPIC is not the primary interrupt
controller. Yet, it still has to handle some per-cpu interrupt.
To do so, it is chained with the GIC using a per-cpu interrupt. However, the
current code only call irq_set_chained_handler, which is called and enable that
interrupt only on the boot CPU, which means that the parent per-CPU interrupt
is never unmasked on the secondary CPUs, preventing the per-CPU interrupt to
actually work as expected.
This was not seen until now since the only MPIC PPI users were the Marvell
timers that were not working, but not used either since the system use the ARM
TWD by default, and the ethernet controllers, that are faking there interrupts
as SPI, and don't really expect to have interrupts on the secondary cores
anyway.
Add a CPU notifier that will enable the PPI on the secondary cores when they
are brought up.
Cc: <stable@vger.kernel.org> # 3.15+
Signed-off-by: Maxime Ripard <maxime.ripard@free-electrons.com>
Acked-by: Gregory CLEMENT <gregory.clement@free-electrons.com>
Link: https://lkml.kernel.org/r/1425378443-28822-1-git-send-email-maxime.ripard@free-electrons.com
Signed-off-by: Jason Cooper <jason@lakedaemon.net>
2015-03-03 18:27:23 +08:00
|
|
|
{
|
2016-07-14 01:16:07 +08:00
|
|
|
armada_xp_mpic_perf_init();
|
irqchip/armada-370-xp: Re-enable per-CPU interrupts at resume time
Commit d17cab4451df1 ("irqchip: Kill off set_irq_flags usage") changed
the code of armada_370_xp_mpic_irq_map() from using set_irq_flags() to
irq_set_probe().
While the commit log seems to imply that there are no functional
changes, there are indeed functional changes introduced by this commit:
the IRQ_NOAUTOEN flag is no longer cleared. This functional change
caused a regression on Armada XP, which no longer works properly after
suspend/resume because per-CPU interrupts remain disabled. This
regression was temporarly worked around in commit
353d6d6c82e5d ("irqchip/armada-370-xp: Fix regression by clearing
IRQ_NOAUTOEN"), but it is not the most satisfying solution. This commit
implements the solution that was initially discussed with Thomas
Gleixner.
Due to how the hardware registers work, the irq-armada-370-xp cannot
simply save/restore a bunch of registers at suspend/resume to make sure
that the interrupts remain in the same state after resuming. Therefore,
it relies on the kernel to say whether the interrupt is disabled or not,
using the irqd_irq_disabled() function. This was all working fine while
the IRQ_NOAUTOEN flag was cleared.
With the change introduced by Rob Herring in d17cab4451df1, the
IRQ_NOAUTOEN flag is now set for all interrupts. irqd_irq_disabled()
returns false for per-CPU interrupts, and therefore our per-CPU
interrupts are no longer re-enabled after resume.
This commit fixes that by using irqd_irq_disabled() only for global
interrupts, and using the newly introduced irq_percpu_is_enabled() for
per-CPU interrupts.
Also, it fixes a related problems that per-CPU interrupts were only
re-enabled on the boot CPU and not other CPUs. Until now this wasn't a
problem since on this platform, only the local timers are using per-CPU
interrupts and the local timers of secondary CPUs are turned off/on
during CPU hotplug before suspend, after after resume. However, since
Linux 4.4, we are also be using per-CPU interrupts for the network
controller, so we need to properly restore the per-CPU interrupts on
secondary CPUs as well.
Signed-off-by: Thomas Petazzoni <thomas.petazzoni@free-electrons.com>
Signed-off-by: Marc Zyngier <marc.zyngier@arm.com>
2017-05-18 16:07:39 +08:00
|
|
|
armada_xp_mpic_reenable_percpu();
|
2016-07-14 01:16:07 +08:00
|
|
|
enable_percpu_irq(parent_irq, IRQ_TYPE_NONE);
|
|
|
|
return 0;
|
irqchip: armada-370-xp: Fix chained per-cpu interrupts
On the Cortex-A9-based Armada SoCs, the MPIC is not the primary interrupt
controller. Yet, it still has to handle some per-cpu interrupt.
To do so, it is chained with the GIC using a per-cpu interrupt. However, the
current code only call irq_set_chained_handler, which is called and enable that
interrupt only on the boot CPU, which means that the parent per-CPU interrupt
is never unmasked on the secondary CPUs, preventing the per-CPU interrupt to
actually work as expected.
This was not seen until now since the only MPIC PPI users were the Marvell
timers that were not working, but not used either since the system use the ARM
TWD by default, and the ethernet controllers, that are faking there interrupts
as SPI, and don't really expect to have interrupts on the secondary cores
anyway.
Add a CPU notifier that will enable the PPI on the secondary cores when they
are brought up.
Cc: <stable@vger.kernel.org> # 3.15+
Signed-off-by: Maxime Ripard <maxime.ripard@free-electrons.com>
Acked-by: Gregory CLEMENT <gregory.clement@free-electrons.com>
Link: https://lkml.kernel.org/r/1425378443-28822-1-git-send-email-maxime.ripard@free-electrons.com
Signed-off-by: Jason Cooper <jason@lakedaemon.net>
2015-03-03 18:27:23 +08:00
|
|
|
}
|
2020-06-23 04:23:36 +08:00
|
|
|
#else
|
|
|
|
static void armada_xp_mpic_smp_cpu_init(void) {}
|
|
|
|
static void ipi_resume(void) {}
|
|
|
|
#endif
|
|
|
|
|
|
|
|
static struct irq_chip armada_370_xp_irq_chip = {
|
|
|
|
.name = "MPIC",
|
|
|
|
.irq_mask = armada_370_xp_irq_mask,
|
|
|
|
.irq_mask_ack = armada_370_xp_irq_mask,
|
|
|
|
.irq_unmask = armada_370_xp_irq_unmask,
|
|
|
|
#ifdef CONFIG_SMP
|
|
|
|
.irq_set_affinity = armada_xp_set_affinity,
|
2016-07-19 00:03:21 +08:00
|
|
|
#endif
|
2020-06-23 04:23:36 +08:00
|
|
|
.flags = IRQCHIP_SKIP_SET_WAKE | IRQCHIP_MASK_ON_SUSPEND,
|
|
|
|
};
|
|
|
|
|
|
|
|
static int armada_370_xp_mpic_irq_map(struct irq_domain *h,
|
|
|
|
unsigned int virq, irq_hw_number_t hw)
|
|
|
|
{
|
|
|
|
armada_370_xp_irq_mask(irq_get_irq_data(virq));
|
|
|
|
if (!is_percpu_irq(hw))
|
|
|
|
writel(hw, per_cpu_int_base +
|
|
|
|
ARMADA_370_XP_INT_CLEAR_MASK_OFFS);
|
|
|
|
else
|
|
|
|
writel(hw, main_int_base + ARMADA_370_XP_INT_SET_ENABLE_OFFS);
|
|
|
|
irq_set_status_flags(virq, IRQ_LEVEL);
|
|
|
|
|
|
|
|
if (is_percpu_irq(hw)) {
|
|
|
|
irq_set_percpu_devid(virq);
|
|
|
|
irq_set_chip_and_handler(virq, &armada_370_xp_irq_chip,
|
|
|
|
handle_percpu_devid_irq);
|
|
|
|
} else {
|
|
|
|
irq_set_chip_and_handler(virq, &armada_370_xp_irq_chip,
|
|
|
|
handle_level_irq);
|
|
|
|
irqd_set_single_target(irq_desc_get_irq_data(irq_to_desc(virq)));
|
|
|
|
}
|
|
|
|
irq_set_probe(virq);
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
2012-08-02 16:19:12 +08:00
|
|
|
|
2015-04-27 20:54:24 +08:00
|
|
|
static const struct irq_domain_ops armada_370_xp_mpic_irq_ops = {
|
2012-06-14 01:01:28 +08:00
|
|
|
.map = armada_370_xp_mpic_irq_map,
|
|
|
|
.xlate = irq_domain_xlate_onecell,
|
|
|
|
};
|
|
|
|
|
2014-02-11 04:00:01 +08:00
|
|
|
#ifdef CONFIG_PCI_MSI
|
2014-02-11 04:00:02 +08:00
|
|
|
static void armada_370_xp_handle_msi_irq(struct pt_regs *regs, bool is_chained)
|
2014-02-11 04:00:01 +08:00
|
|
|
{
|
|
|
|
u32 msimask, msinr;
|
|
|
|
|
|
|
|
msimask = readl_relaxed(per_cpu_int_base +
|
|
|
|
ARMADA_370_XP_IN_DRBEL_CAUSE_OFFS)
|
|
|
|
& PCI_MSI_DOORBELL_MASK;
|
|
|
|
|
|
|
|
writel(~msimask, per_cpu_int_base +
|
|
|
|
ARMADA_370_XP_IN_DRBEL_CAUSE_OFFS);
|
|
|
|
|
|
|
|
for (msinr = PCI_MSI_DOORBELL_START;
|
|
|
|
msinr < PCI_MSI_DOORBELL_END; msinr++) {
|
2021-05-05 00:42:18 +08:00
|
|
|
unsigned int irq;
|
2014-02-11 04:00:01 +08:00
|
|
|
|
|
|
|
if (!(msimask & BIT(msinr)))
|
|
|
|
continue;
|
|
|
|
|
2021-05-05 00:42:18 +08:00
|
|
|
irq = msinr - PCI_MSI_DOORBELL_START;
|
|
|
|
|
2021-10-21 03:23:09 +08:00
|
|
|
generic_handle_domain_irq(armada_370_xp_msi_inner_domain, irq);
|
2014-02-11 04:00:01 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
#else
|
2014-02-11 04:00:02 +08:00
|
|
|
static void armada_370_xp_handle_msi_irq(struct pt_regs *r, bool b) {}
|
2014-02-11 04:00:01 +08:00
|
|
|
#endif
|
|
|
|
|
2015-09-14 16:42:37 +08:00
|
|
|
static void armada_370_xp_mpic_handle_cascade_irq(struct irq_desc *desc)
|
2014-02-11 04:00:02 +08:00
|
|
|
{
|
2015-06-04 12:13:20 +08:00
|
|
|
struct irq_chip *chip = irq_desc_get_chip(desc);
|
2014-09-25 19:17:19 +08:00
|
|
|
unsigned long irqmap, irqn, irqsrc, cpuid;
|
2014-02-11 04:00:02 +08:00
|
|
|
|
|
|
|
chained_irq_enter(chip, desc);
|
|
|
|
|
|
|
|
irqmap = readl_relaxed(per_cpu_int_base + ARMADA_375_PPI_CAUSE);
|
2014-09-25 19:17:19 +08:00
|
|
|
cpuid = cpu_logical_map(smp_processor_id());
|
2014-02-11 04:00:02 +08:00
|
|
|
|
|
|
|
for_each_set_bit(irqn, &irqmap, BITS_PER_LONG) {
|
2014-09-25 19:17:19 +08:00
|
|
|
irqsrc = readl_relaxed(main_int_base +
|
|
|
|
ARMADA_370_XP_INT_SOURCE_CTL(irqn));
|
|
|
|
|
|
|
|
/* Check if the interrupt is not masked on current CPU.
|
|
|
|
* Test IRQ (0-1) and FIQ (8-9) mask bits.
|
|
|
|
*/
|
|
|
|
if (!(irqsrc & ARMADA_370_XP_INT_IRQ_FIQ_MASK(cpuid)))
|
|
|
|
continue;
|
|
|
|
|
|
|
|
if (irqn == 1) {
|
|
|
|
armada_370_xp_handle_msi_irq(NULL, true);
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
2021-05-05 00:42:18 +08:00
|
|
|
generic_handle_domain_irq(armada_370_xp_mpic_domain, irqn);
|
2014-02-11 04:00:02 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
chained_irq_exit(chip, desc);
|
|
|
|
}
|
|
|
|
|
2014-03-05 08:40:30 +08:00
|
|
|
static void __exception_irq_entry
|
2013-04-10 05:26:15 +08:00
|
|
|
armada_370_xp_handle_irq(struct pt_regs *regs)
|
2012-06-14 01:01:28 +08:00
|
|
|
{
|
|
|
|
u32 irqstat, irqnr;
|
|
|
|
|
|
|
|
do {
|
|
|
|
irqstat = readl_relaxed(per_cpu_int_base +
|
|
|
|
ARMADA_370_XP_CPU_INTACK_OFFS);
|
|
|
|
irqnr = irqstat & 0x3FF;
|
|
|
|
|
2012-08-02 16:19:12 +08:00
|
|
|
if (irqnr > 1022)
|
|
|
|
break;
|
|
|
|
|
2013-08-10 04:27:11 +08:00
|
|
|
if (irqnr > 1) {
|
2021-10-21 03:23:09 +08:00
|
|
|
generic_handle_domain_irq(armada_370_xp_mpic_domain,
|
|
|
|
irqnr);
|
2012-06-14 01:01:28 +08:00
|
|
|
continue;
|
|
|
|
}
|
2013-08-10 04:27:11 +08:00
|
|
|
|
|
|
|
/* MSI handling */
|
2014-02-11 04:00:01 +08:00
|
|
|
if (irqnr == 1)
|
2014-02-11 04:00:02 +08:00
|
|
|
armada_370_xp_handle_msi_irq(regs, false);
|
2013-08-10 04:27:11 +08:00
|
|
|
|
2012-08-02 16:19:12 +08:00
|
|
|
#ifdef CONFIG_SMP
|
|
|
|
/* IPI Handling */
|
|
|
|
if (irqnr == 0) {
|
2020-06-23 04:23:36 +08:00
|
|
|
unsigned long ipimask;
|
|
|
|
int ipi;
|
2012-08-02 16:19:12 +08:00
|
|
|
|
|
|
|
ipimask = readl_relaxed(per_cpu_int_base +
|
|
|
|
ARMADA_370_XP_IN_DRBEL_CAUSE_OFFS)
|
2013-04-10 05:26:17 +08:00
|
|
|
& IPI_DOORBELL_MASK;
|
2012-08-02 16:19:12 +08:00
|
|
|
|
2020-06-23 04:23:36 +08:00
|
|
|
for_each_set_bit(ipi, &ipimask, IPI_DOORBELL_END)
|
2021-10-21 03:23:09 +08:00
|
|
|
generic_handle_domain_irq(ipi_domain, ipi);
|
2012-08-02 16:19:12 +08:00
|
|
|
}
|
|
|
|
#endif
|
2012-06-14 01:01:28 +08:00
|
|
|
|
|
|
|
} while (1);
|
|
|
|
}
|
|
|
|
|
2014-11-22 00:00:00 +08:00
|
|
|
static int armada_370_xp_mpic_suspend(void)
|
|
|
|
{
|
|
|
|
doorbell_mask_reg = readl(per_cpu_int_base +
|
|
|
|
ARMADA_370_XP_IN_DRBEL_MSK_OFFS);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void armada_370_xp_mpic_resume(void)
|
|
|
|
{
|
|
|
|
int nirqs;
|
|
|
|
irq_hw_number_t irq;
|
|
|
|
|
|
|
|
/* Re-enable interrupts */
|
|
|
|
nirqs = (readl(main_int_base + ARMADA_370_XP_INT_CONTROL) >> 2) & 0x3ff;
|
|
|
|
for (irq = 0; irq < nirqs; irq++) {
|
|
|
|
struct irq_data *data;
|
|
|
|
int virq;
|
|
|
|
|
|
|
|
virq = irq_linear_revmap(armada_370_xp_mpic_domain, irq);
|
|
|
|
if (virq == 0)
|
|
|
|
continue;
|
|
|
|
|
irqchip/armada-370-xp: Re-enable per-CPU interrupts at resume time
Commit d17cab4451df1 ("irqchip: Kill off set_irq_flags usage") changed
the code of armada_370_xp_mpic_irq_map() from using set_irq_flags() to
irq_set_probe().
While the commit log seems to imply that there are no functional
changes, there are indeed functional changes introduced by this commit:
the IRQ_NOAUTOEN flag is no longer cleared. This functional change
caused a regression on Armada XP, which no longer works properly after
suspend/resume because per-CPU interrupts remain disabled. This
regression was temporarly worked around in commit
353d6d6c82e5d ("irqchip/armada-370-xp: Fix regression by clearing
IRQ_NOAUTOEN"), but it is not the most satisfying solution. This commit
implements the solution that was initially discussed with Thomas
Gleixner.
Due to how the hardware registers work, the irq-armada-370-xp cannot
simply save/restore a bunch of registers at suspend/resume to make sure
that the interrupts remain in the same state after resuming. Therefore,
it relies on the kernel to say whether the interrupt is disabled or not,
using the irqd_irq_disabled() function. This was all working fine while
the IRQ_NOAUTOEN flag was cleared.
With the change introduced by Rob Herring in d17cab4451df1, the
IRQ_NOAUTOEN flag is now set for all interrupts. irqd_irq_disabled()
returns false for per-CPU interrupts, and therefore our per-CPU
interrupts are no longer re-enabled after resume.
This commit fixes that by using irqd_irq_disabled() only for global
interrupts, and using the newly introduced irq_percpu_is_enabled() for
per-CPU interrupts.
Also, it fixes a related problems that per-CPU interrupts were only
re-enabled on the boot CPU and not other CPUs. Until now this wasn't a
problem since on this platform, only the local timers are using per-CPU
interrupts and the local timers of secondary CPUs are turned off/on
during CPU hotplug before suspend, after after resume. However, since
Linux 4.4, we are also be using per-CPU interrupts for the network
controller, so we need to properly restore the per-CPU interrupts on
secondary CPUs as well.
Signed-off-by: Thomas Petazzoni <thomas.petazzoni@free-electrons.com>
Signed-off-by: Marc Zyngier <marc.zyngier@arm.com>
2017-05-18 16:07:39 +08:00
|
|
|
data = irq_get_irq_data(virq);
|
|
|
|
|
|
|
|
if (!is_percpu_irq(irq)) {
|
|
|
|
/* Non per-CPU interrupts */
|
2014-11-22 00:00:00 +08:00
|
|
|
writel(irq, per_cpu_int_base +
|
|
|
|
ARMADA_370_XP_INT_CLEAR_MASK_OFFS);
|
irqchip/armada-370-xp: Re-enable per-CPU interrupts at resume time
Commit d17cab4451df1 ("irqchip: Kill off set_irq_flags usage") changed
the code of armada_370_xp_mpic_irq_map() from using set_irq_flags() to
irq_set_probe().
While the commit log seems to imply that there are no functional
changes, there are indeed functional changes introduced by this commit:
the IRQ_NOAUTOEN flag is no longer cleared. This functional change
caused a regression on Armada XP, which no longer works properly after
suspend/resume because per-CPU interrupts remain disabled. This
regression was temporarly worked around in commit
353d6d6c82e5d ("irqchip/armada-370-xp: Fix regression by clearing
IRQ_NOAUTOEN"), but it is not the most satisfying solution. This commit
implements the solution that was initially discussed with Thomas
Gleixner.
Due to how the hardware registers work, the irq-armada-370-xp cannot
simply save/restore a bunch of registers at suspend/resume to make sure
that the interrupts remain in the same state after resuming. Therefore,
it relies on the kernel to say whether the interrupt is disabled or not,
using the irqd_irq_disabled() function. This was all working fine while
the IRQ_NOAUTOEN flag was cleared.
With the change introduced by Rob Herring in d17cab4451df1, the
IRQ_NOAUTOEN flag is now set for all interrupts. irqd_irq_disabled()
returns false for per-CPU interrupts, and therefore our per-CPU
interrupts are no longer re-enabled after resume.
This commit fixes that by using irqd_irq_disabled() only for global
interrupts, and using the newly introduced irq_percpu_is_enabled() for
per-CPU interrupts.
Also, it fixes a related problems that per-CPU interrupts were only
re-enabled on the boot CPU and not other CPUs. Until now this wasn't a
problem since on this platform, only the local timers are using per-CPU
interrupts and the local timers of secondary CPUs are turned off/on
during CPU hotplug before suspend, after after resume. However, since
Linux 4.4, we are also be using per-CPU interrupts for the network
controller, so we need to properly restore the per-CPU interrupts on
secondary CPUs as well.
Signed-off-by: Thomas Petazzoni <thomas.petazzoni@free-electrons.com>
Signed-off-by: Marc Zyngier <marc.zyngier@arm.com>
2017-05-18 16:07:39 +08:00
|
|
|
if (!irqd_irq_disabled(data))
|
|
|
|
armada_370_xp_irq_unmask(data);
|
|
|
|
} else {
|
|
|
|
/* Per-CPU interrupts */
|
2014-11-22 00:00:00 +08:00
|
|
|
writel(irq, main_int_base +
|
|
|
|
ARMADA_370_XP_INT_SET_ENABLE_OFFS);
|
|
|
|
|
irqchip/armada-370-xp: Re-enable per-CPU interrupts at resume time
Commit d17cab4451df1 ("irqchip: Kill off set_irq_flags usage") changed
the code of armada_370_xp_mpic_irq_map() from using set_irq_flags() to
irq_set_probe().
While the commit log seems to imply that there are no functional
changes, there are indeed functional changes introduced by this commit:
the IRQ_NOAUTOEN flag is no longer cleared. This functional change
caused a regression on Armada XP, which no longer works properly after
suspend/resume because per-CPU interrupts remain disabled. This
regression was temporarly worked around in commit
353d6d6c82e5d ("irqchip/armada-370-xp: Fix regression by clearing
IRQ_NOAUTOEN"), but it is not the most satisfying solution. This commit
implements the solution that was initially discussed with Thomas
Gleixner.
Due to how the hardware registers work, the irq-armada-370-xp cannot
simply save/restore a bunch of registers at suspend/resume to make sure
that the interrupts remain in the same state after resuming. Therefore,
it relies on the kernel to say whether the interrupt is disabled or not,
using the irqd_irq_disabled() function. This was all working fine while
the IRQ_NOAUTOEN flag was cleared.
With the change introduced by Rob Herring in d17cab4451df1, the
IRQ_NOAUTOEN flag is now set for all interrupts. irqd_irq_disabled()
returns false for per-CPU interrupts, and therefore our per-CPU
interrupts are no longer re-enabled after resume.
This commit fixes that by using irqd_irq_disabled() only for global
interrupts, and using the newly introduced irq_percpu_is_enabled() for
per-CPU interrupts.
Also, it fixes a related problems that per-CPU interrupts were only
re-enabled on the boot CPU and not other CPUs. Until now this wasn't a
problem since on this platform, only the local timers are using per-CPU
interrupts and the local timers of secondary CPUs are turned off/on
during CPU hotplug before suspend, after after resume. However, since
Linux 4.4, we are also be using per-CPU interrupts for the network
controller, so we need to properly restore the per-CPU interrupts on
secondary CPUs as well.
Signed-off-by: Thomas Petazzoni <thomas.petazzoni@free-electrons.com>
Signed-off-by: Marc Zyngier <marc.zyngier@arm.com>
2017-05-18 16:07:39 +08:00
|
|
|
/*
|
|
|
|
* Re-enable on the current CPU,
|
|
|
|
* armada_xp_mpic_reenable_percpu() will take
|
|
|
|
* care of secondary CPUs when they come up.
|
|
|
|
*/
|
|
|
|
if (irq_percpu_is_enabled(virq))
|
|
|
|
armada_370_xp_irq_unmask(data);
|
|
|
|
}
|
2014-11-22 00:00:00 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
/* Reconfigure doorbells for IPIs and MSIs */
|
|
|
|
writel(doorbell_mask_reg,
|
|
|
|
per_cpu_int_base + ARMADA_370_XP_IN_DRBEL_MSK_OFFS);
|
|
|
|
if (doorbell_mask_reg & IPI_DOORBELL_MASK)
|
|
|
|
writel(0, per_cpu_int_base + ARMADA_370_XP_INT_CLEAR_MASK_OFFS);
|
|
|
|
if (doorbell_mask_reg & PCI_MSI_DOORBELL_MASK)
|
|
|
|
writel(1, per_cpu_int_base + ARMADA_370_XP_INT_CLEAR_MASK_OFFS);
|
2020-06-23 04:23:36 +08:00
|
|
|
|
|
|
|
ipi_resume();
|
2014-11-22 00:00:00 +08:00
|
|
|
}
|
|
|
|
|
2016-06-09 01:55:33 +08:00
|
|
|
static struct syscore_ops armada_370_xp_mpic_syscore_ops = {
|
2014-11-22 00:00:00 +08:00
|
|
|
.suspend = armada_370_xp_mpic_suspend,
|
|
|
|
.resume = armada_370_xp_mpic_resume,
|
|
|
|
};
|
|
|
|
|
2013-04-10 05:26:16 +08:00
|
|
|
static int __init armada_370_xp_mpic_of_init(struct device_node *node,
|
|
|
|
struct device_node *parent)
|
2012-06-14 01:01:28 +08:00
|
|
|
{
|
2013-08-10 04:27:10 +08:00
|
|
|
struct resource main_int_res, per_cpu_int_res;
|
irqchip: armada-370-xp: Fix chained per-cpu interrupts
On the Cortex-A9-based Armada SoCs, the MPIC is not the primary interrupt
controller. Yet, it still has to handle some per-cpu interrupt.
To do so, it is chained with the GIC using a per-cpu interrupt. However, the
current code only call irq_set_chained_handler, which is called and enable that
interrupt only on the boot CPU, which means that the parent per-CPU interrupt
is never unmasked on the secondary CPUs, preventing the per-CPU interrupt to
actually work as expected.
This was not seen until now since the only MPIC PPI users were the Marvell
timers that were not working, but not used either since the system use the ARM
TWD by default, and the ethernet controllers, that are faking there interrupts
as SPI, and don't really expect to have interrupts on the secondary cores
anyway.
Add a CPU notifier that will enable the PPI on the secondary cores when they
are brought up.
Cc: <stable@vger.kernel.org> # 3.15+
Signed-off-by: Maxime Ripard <maxime.ripard@free-electrons.com>
Acked-by: Gregory CLEMENT <gregory.clement@free-electrons.com>
Link: https://lkml.kernel.org/r/1425378443-28822-1-git-send-email-maxime.ripard@free-electrons.com
Signed-off-by: Jason Cooper <jason@lakedaemon.net>
2015-03-03 18:27:23 +08:00
|
|
|
int nr_irqs, i;
|
2013-04-10 05:26:16 +08:00
|
|
|
u32 control;
|
|
|
|
|
2013-08-10 04:27:10 +08:00
|
|
|
BUG_ON(of_address_to_resource(node, 0, &main_int_res));
|
|
|
|
BUG_ON(of_address_to_resource(node, 1, &per_cpu_int_res));
|
2013-04-10 05:26:16 +08:00
|
|
|
|
2013-08-10 04:27:10 +08:00
|
|
|
BUG_ON(!request_mem_region(main_int_res.start,
|
|
|
|
resource_size(&main_int_res),
|
|
|
|
node->full_name));
|
|
|
|
BUG_ON(!request_mem_region(per_cpu_int_res.start,
|
|
|
|
resource_size(&per_cpu_int_res),
|
|
|
|
node->full_name));
|
|
|
|
|
|
|
|
main_int_base = ioremap(main_int_res.start,
|
|
|
|
resource_size(&main_int_res));
|
2013-04-10 05:26:16 +08:00
|
|
|
BUG_ON(!main_int_base);
|
2013-08-10 04:27:10 +08:00
|
|
|
|
|
|
|
per_cpu_int_base = ioremap(per_cpu_int_res.start,
|
|
|
|
resource_size(&per_cpu_int_res));
|
2013-04-10 05:26:16 +08:00
|
|
|
BUG_ON(!per_cpu_int_base);
|
|
|
|
|
|
|
|
control = readl(main_int_base + ARMADA_370_XP_INT_CONTROL);
|
2014-05-31 04:18:18 +08:00
|
|
|
nr_irqs = (control >> 2) & 0x3ff;
|
|
|
|
|
|
|
|
for (i = 0; i < nr_irqs; i++)
|
|
|
|
writel(i, main_int_base + ARMADA_370_XP_INT_CLEAR_ENABLE_OFFS);
|
2013-04-10 05:26:16 +08:00
|
|
|
|
|
|
|
armada_370_xp_mpic_domain =
|
2014-05-31 04:18:18 +08:00
|
|
|
irq_domain_add_linear(node, nr_irqs,
|
2013-04-10 05:26:16 +08:00
|
|
|
&armada_370_xp_mpic_irq_ops, NULL);
|
2013-08-10 04:27:10 +08:00
|
|
|
BUG_ON(!armada_370_xp_mpic_domain);
|
2017-06-22 18:42:50 +08:00
|
|
|
irq_domain_update_bus_token(armada_370_xp_mpic_domain, DOMAIN_BUS_WIRED);
|
2013-04-10 05:26:16 +08:00
|
|
|
|
2015-03-03 18:43:14 +08:00
|
|
|
/* Setup for the boot CPU */
|
2015-03-03 18:43:16 +08:00
|
|
|
armada_xp_mpic_perf_init();
|
2013-04-10 05:26:16 +08:00
|
|
|
armada_xp_mpic_smp_cpu_init();
|
|
|
|
|
2013-08-10 04:27:11 +08:00
|
|
|
armada_370_xp_msi_init(node, main_int_res.start);
|
|
|
|
|
2014-02-11 04:00:02 +08:00
|
|
|
parent_irq = irq_of_parse_and_map(node, 0);
|
|
|
|
if (parent_irq <= 0) {
|
|
|
|
irq_set_default_host(armada_370_xp_mpic_domain);
|
|
|
|
set_handle_irq(armada_370_xp_handle_irq);
|
2014-04-14 21:54:01 +08:00
|
|
|
#ifdef CONFIG_SMP
|
2020-06-23 04:23:36 +08:00
|
|
|
armada_xp_ipi_init(node);
|
2016-07-14 01:16:07 +08:00
|
|
|
cpuhp_setup_state_nocalls(CPUHP_AP_IRQ_ARMADA_XP_STARTING,
|
2016-12-22 03:19:54 +08:00
|
|
|
"irqchip/armada/ipi:starting",
|
2016-07-14 01:16:07 +08:00
|
|
|
armada_xp_mpic_starting_cpu, NULL);
|
2014-04-14 21:54:01 +08:00
|
|
|
#endif
|
2014-02-11 04:00:02 +08:00
|
|
|
} else {
|
irqchip: armada-370-xp: Fix chained per-cpu interrupts
On the Cortex-A9-based Armada SoCs, the MPIC is not the primary interrupt
controller. Yet, it still has to handle some per-cpu interrupt.
To do so, it is chained with the GIC using a per-cpu interrupt. However, the
current code only call irq_set_chained_handler, which is called and enable that
interrupt only on the boot CPU, which means that the parent per-CPU interrupt
is never unmasked on the secondary CPUs, preventing the per-CPU interrupt to
actually work as expected.
This was not seen until now since the only MPIC PPI users were the Marvell
timers that were not working, but not used either since the system use the ARM
TWD by default, and the ethernet controllers, that are faking there interrupts
as SPI, and don't really expect to have interrupts on the secondary cores
anyway.
Add a CPU notifier that will enable the PPI on the secondary cores when they
are brought up.
Cc: <stable@vger.kernel.org> # 3.15+
Signed-off-by: Maxime Ripard <maxime.ripard@free-electrons.com>
Acked-by: Gregory CLEMENT <gregory.clement@free-electrons.com>
Link: https://lkml.kernel.org/r/1425378443-28822-1-git-send-email-maxime.ripard@free-electrons.com
Signed-off-by: Jason Cooper <jason@lakedaemon.net>
2015-03-03 18:27:23 +08:00
|
|
|
#ifdef CONFIG_SMP
|
2016-12-22 03:19:57 +08:00
|
|
|
cpuhp_setup_state_nocalls(CPUHP_AP_IRQ_ARMADA_XP_STARTING,
|
2016-12-22 03:19:54 +08:00
|
|
|
"irqchip/armada/cascade:starting",
|
2016-07-14 01:16:07 +08:00
|
|
|
mpic_cascaded_starting_cpu, NULL);
|
irqchip: armada-370-xp: Fix chained per-cpu interrupts
On the Cortex-A9-based Armada SoCs, the MPIC is not the primary interrupt
controller. Yet, it still has to handle some per-cpu interrupt.
To do so, it is chained with the GIC using a per-cpu interrupt. However, the
current code only call irq_set_chained_handler, which is called and enable that
interrupt only on the boot CPU, which means that the parent per-CPU interrupt
is never unmasked on the secondary CPUs, preventing the per-CPU interrupt to
actually work as expected.
This was not seen until now since the only MPIC PPI users were the Marvell
timers that were not working, but not used either since the system use the ARM
TWD by default, and the ethernet controllers, that are faking there interrupts
as SPI, and don't really expect to have interrupts on the secondary cores
anyway.
Add a CPU notifier that will enable the PPI on the secondary cores when they
are brought up.
Cc: <stable@vger.kernel.org> # 3.15+
Signed-off-by: Maxime Ripard <maxime.ripard@free-electrons.com>
Acked-by: Gregory CLEMENT <gregory.clement@free-electrons.com>
Link: https://lkml.kernel.org/r/1425378443-28822-1-git-send-email-maxime.ripard@free-electrons.com
Signed-off-by: Jason Cooper <jason@lakedaemon.net>
2015-03-03 18:27:23 +08:00
|
|
|
#endif
|
2014-02-11 04:00:02 +08:00
|
|
|
irq_set_chained_handler(parent_irq,
|
|
|
|
armada_370_xp_mpic_handle_cascade_irq);
|
|
|
|
}
|
2013-04-10 05:26:16 +08:00
|
|
|
|
2014-11-22 00:00:00 +08:00
|
|
|
register_syscore_ops(&armada_370_xp_mpic_syscore_ops);
|
|
|
|
|
2013-04-10 05:26:16 +08:00
|
|
|
return 0;
|
2012-06-14 01:01:28 +08:00
|
|
|
}
|
2013-04-10 05:26:16 +08:00
|
|
|
|
2013-04-10 05:26:15 +08:00
|
|
|
IRQCHIP_DECLARE(armada_370_xp_mpic, "marvell,mpic", armada_370_xp_mpic_of_init);
|