2019-06-03 05:44:50 +00:00
|
|
|
// SPDX-License-Identifier: GPL-2.0-only
|
2014-06-30 15:01:31 +00:00
|
|
|
/*
|
2016-12-19 17:01:52 +00:00
|
|
|
* Copyright (C) 2013-2017 ARM Limited, All Rights Reserved.
|
2014-06-30 15:01:31 +00:00
|
|
|
* Author: Marc Zyngier <marc.zyngier@arm.com>
|
|
|
|
*/
|
|
|
|
|
2016-04-11 15:32:55 +00:00
|
|
|
#define pr_fmt(fmt) "GICv3: " fmt
|
|
|
|
|
2016-01-19 13:11:15 +00:00
|
|
|
#include <linux/acpi.h>
|
2014-06-30 15:01:31 +00:00
|
|
|
#include <linux/cpu.h>
|
2014-08-26 15:03:35 +00:00
|
|
|
#include <linux/cpu_pm.h>
|
2014-06-30 15:01:31 +00:00
|
|
|
#include <linux/delay.h>
|
|
|
|
#include <linux/interrupt.h>
|
2016-01-19 13:11:15 +00:00
|
|
|
#include <linux/irqdomain.h>
|
2014-06-30 15:01:31 +00:00
|
|
|
#include <linux/of.h>
|
|
|
|
#include <linux/of_address.h>
|
|
|
|
#include <linux/of_irq.h>
|
|
|
|
#include <linux/percpu.h>
|
2019-01-31 14:58:59 +00:00
|
|
|
#include <linux/refcount.h>
|
2014-06-30 15:01:31 +00:00
|
|
|
#include <linux/slab.h>
|
|
|
|
|
2015-07-07 21:11:46 +00:00
|
|
|
#include <linux/irqchip.h>
|
2016-04-11 15:32:57 +00:00
|
|
|
#include <linux/irqchip/arm-gic-common.h>
|
2014-06-30 15:01:31 +00:00
|
|
|
#include <linux/irqchip/arm-gic-v3.h>
|
2016-04-11 08:57:54 +00:00
|
|
|
#include <linux/irqchip/irq-partition-percpu.h>
|
2014-06-30 15:01:31 +00:00
|
|
|
|
|
|
|
#include <asm/cputype.h>
|
|
|
|
#include <asm/exception.h>
|
|
|
|
#include <asm/smp_plat.h>
|
2015-08-26 16:00:42 +00:00
|
|
|
#include <asm/virt.h>
|
2014-06-30 15:01:31 +00:00
|
|
|
|
|
|
|
#include "irq-gic-common.h"
|
|
|
|
|
2019-01-31 14:58:58 +00:00
|
|
|
#define GICD_INT_NMI_PRI (GICD_INT_DEF_PRI & ~0x80)
|
|
|
|
|
2018-12-10 13:56:32 +00:00
|
|
|
#define FLAGS_WORKAROUND_GICR_WAKER_MSM8996 (1ULL << 0)
|
2020-03-11 11:56:49 +00:00
|
|
|
#define FLAGS_WORKAROUND_CAVIUM_ERRATUM_38539 (1ULL << 1)
|
2018-12-10 13:56:32 +00:00
|
|
|
|
2020-04-25 14:24:01 +00:00
|
|
|
#define GIC_IRQ_TYPE_PARTITION (GIC_IRQ_TYPE_LPI + 1)
|
|
|
|
|
2014-11-24 14:35:10 +00:00
|
|
|
struct redist_region {
|
|
|
|
void __iomem *redist_base;
|
|
|
|
phys_addr_t phys_base;
|
2016-01-19 13:11:16 +00:00
|
|
|
bool single_redist;
|
2014-11-24 14:35:10 +00:00
|
|
|
};
|
|
|
|
|
2014-06-30 15:01:31 +00:00
|
|
|
struct gic_chip_data {
|
2016-04-11 08:57:54 +00:00
|
|
|
struct fwnode_handle *fwnode;
|
2014-06-30 15:01:31 +00:00
|
|
|
void __iomem *dist_base;
|
2014-11-24 14:35:10 +00:00
|
|
|
struct redist_region *redist_regions;
|
|
|
|
struct rdists rdists;
|
2014-06-30 15:01:31 +00:00
|
|
|
struct irq_domain *domain;
|
|
|
|
u64 redist_stride;
|
2014-11-24 14:35:10 +00:00
|
|
|
u32 nr_redist_regions;
|
2018-12-10 13:56:32 +00:00
|
|
|
u64 flags;
|
2017-10-06 15:24:00 +00:00
|
|
|
bool has_rss;
|
2019-07-18 10:15:14 +00:00
|
|
|
unsigned int ppi_nr;
|
2019-07-18 12:05:17 +00:00
|
|
|
struct partition_desc **ppi_descs;
|
2014-06-30 15:01:31 +00:00
|
|
|
};
|
|
|
|
|
|
|
|
static struct gic_chip_data gic_data __read_mostly;
|
2018-03-26 21:09:25 +00:00
|
|
|
static DEFINE_STATIC_KEY_TRUE(supports_deactivate_key);
|
2014-06-30 15:01:31 +00:00
|
|
|
|
2019-07-16 14:17:31 +00:00
|
|
|
#define GIC_ID_NR (1U << GICD_TYPER_ID_BITS(gic_data.rdists.gicd_typer))
|
2019-09-18 06:57:30 +00:00
|
|
|
#define GIC_LINE_NR min(GICD_TYPER_SPIS(gic_data.rdists.gicd_typer), 1020U)
|
2019-07-16 14:17:31 +00:00
|
|
|
#define GIC_ESPI_NR GICD_TYPER_ESPIS(gic_data.rdists.gicd_typer)
|
|
|
|
|
2019-01-31 14:58:57 +00:00
|
|
|
/*
|
|
|
|
* The behaviours of RPR and PMR registers differ depending on the value of
|
|
|
|
* SCR_EL3.FIQ, and the behaviour of non-secure priority registers of the
|
|
|
|
* distributor and redistributors depends on whether security is enabled in the
|
|
|
|
* GIC.
|
|
|
|
*
|
|
|
|
* When security is enabled, non-secure priority values from the (re)distributor
|
|
|
|
* are presented to the GIC CPUIF as follow:
|
|
|
|
* (GIC_(R)DIST_PRI[irq] >> 1) | 0x80;
|
|
|
|
*
|
2021-01-21 18:22:52 +00:00
|
|
|
* If SCR_EL3.FIQ == 1, the values written to/read from PMR and RPR at non-secure
|
2019-01-31 14:58:57 +00:00
|
|
|
* EL1 are subject to a similar operation thus matching the priorities presented
|
2020-09-12 15:37:07 +00:00
|
|
|
* from the (re)distributor when security is enabled. When SCR_EL3.FIQ == 0,
|
2021-01-21 18:22:52 +00:00
|
|
|
* these values are unchanged by the GIC.
|
2019-01-31 14:58:57 +00:00
|
|
|
*
|
|
|
|
* see GICv3/GICv4 Architecture Specification (IHI0069D):
|
|
|
|
* - section 4.8.1 Non-secure accesses to register fields for Secure interrupt
|
|
|
|
* priorities.
|
|
|
|
* - Figure 4-7 Secure read of the priority field for a Non-secure Group 1
|
|
|
|
* interrupt.
|
|
|
|
*/
|
|
|
|
static DEFINE_STATIC_KEY_FALSE(supports_pseudo_nmis);
|
|
|
|
|
2019-10-02 09:06:12 +00:00
|
|
|
/*
|
|
|
|
* Global static key controlling whether an update to PMR allowing more
|
|
|
|
* interrupts requires to be propagated to the redistributor (DSB SY).
|
|
|
|
* And this needs to be exported for modules to be able to enable
|
|
|
|
* interrupts...
|
|
|
|
*/
|
|
|
|
DEFINE_STATIC_KEY_FALSE(gic_pmr_sync);
|
|
|
|
EXPORT_SYMBOL(gic_pmr_sync);
|
|
|
|
|
2020-09-12 15:37:07 +00:00
|
|
|
DEFINE_STATIC_KEY_FALSE(gic_nonsecure_priorities);
|
|
|
|
EXPORT_SYMBOL(gic_nonsecure_priorities);
|
|
|
|
|
2021-08-11 17:15:05 +00:00
|
|
|
/*
|
|
|
|
* When the Non-secure world has access to group 0 interrupts (as a
|
|
|
|
* consequence of SCR_EL3.FIQ == 0), reading the ICC_RPR_EL1 register will
|
|
|
|
* return the Distributor's view of the interrupt priority.
|
|
|
|
*
|
|
|
|
* When GIC security is enabled (GICD_CTLR.DS == 0), the interrupt priority
|
|
|
|
* written by software is moved to the Non-secure range by the Distributor.
|
|
|
|
*
|
|
|
|
* If both are true (which is when gic_nonsecure_priorities gets enabled),
|
|
|
|
* we need to shift down the priority programmed by software to match it
|
|
|
|
* against the value returned by ICC_RPR_EL1.
|
|
|
|
*/
|
|
|
|
#define GICD_INT_RPR_PRI(priority) \
|
|
|
|
({ \
|
|
|
|
u32 __priority = (priority); \
|
|
|
|
if (static_branch_unlikely(&gic_nonsecure_priorities)) \
|
|
|
|
__priority = 0x80 | (__priority >> 1); \
|
|
|
|
\
|
|
|
|
__priority; \
|
|
|
|
})
|
|
|
|
|
2019-01-31 14:58:59 +00:00
|
|
|
/* ppi_nmi_refs[n] == number of cpus having ppi[n + 16] set as NMI */
|
2019-07-18 11:53:05 +00:00
|
|
|
static refcount_t *ppi_nmi_refs;
|
2019-01-31 14:58:59 +00:00
|
|
|
|
2021-02-27 10:23:45 +00:00
|
|
|
static struct gic_kvm_info gic_v3_kvm_info __initdata;
|
2017-10-06 15:24:00 +00:00
|
|
|
static DEFINE_PER_CPU(bool, has_rss);
|
2016-04-11 15:32:57 +00:00
|
|
|
|
2017-10-06 15:24:00 +00:00
|
|
|
#define MPIDR_RS(mpidr) (((mpidr) & 0xF0UL) >> 4)
|
2014-11-24 14:35:10 +00:00
|
|
|
#define gic_data_rdist() (this_cpu_ptr(gic_data.rdists.rdist))
|
|
|
|
#define gic_data_rdist_rd_base() (gic_data_rdist()->rd_base)
|
2014-06-30 15:01:31 +00:00
|
|
|
#define gic_data_rdist_sgi_base() (gic_data_rdist_rd_base() + SZ_64K)
|
|
|
|
|
|
|
|
/* Our default, arbitrary priority value. Linux only uses one anyway. */
|
|
|
|
#define DEFAULT_PMR_VALUE 0xf0
|
|
|
|
|
2019-07-16 13:41:40 +00:00
|
|
|
enum gic_intid_range {
|
2020-04-25 14:11:20 +00:00
|
|
|
SGI_RANGE,
|
2019-07-16 13:41:40 +00:00
|
|
|
PPI_RANGE,
|
|
|
|
SPI_RANGE,
|
2019-07-18 12:19:25 +00:00
|
|
|
EPPI_RANGE,
|
2019-07-16 14:17:31 +00:00
|
|
|
ESPI_RANGE,
|
2019-07-16 13:41:40 +00:00
|
|
|
LPI_RANGE,
|
|
|
|
__INVALID_RANGE__
|
|
|
|
};
|
|
|
|
|
|
|
|
static enum gic_intid_range __get_intid_range(irq_hw_number_t hwirq)
|
|
|
|
{
|
|
|
|
switch (hwirq) {
|
2020-04-25 14:11:20 +00:00
|
|
|
case 0 ... 15:
|
|
|
|
return SGI_RANGE;
|
2019-07-16 13:41:40 +00:00
|
|
|
case 16 ... 31:
|
|
|
|
return PPI_RANGE;
|
|
|
|
case 32 ... 1019:
|
|
|
|
return SPI_RANGE;
|
2019-07-18 12:19:25 +00:00
|
|
|
case EPPI_BASE_INTID ... (EPPI_BASE_INTID + 63):
|
|
|
|
return EPPI_RANGE;
|
2019-07-16 14:17:31 +00:00
|
|
|
case ESPI_BASE_INTID ... (ESPI_BASE_INTID + 1023):
|
|
|
|
return ESPI_RANGE;
|
2019-07-16 13:41:40 +00:00
|
|
|
case 8192 ... GENMASK(23, 0):
|
|
|
|
return LPI_RANGE;
|
|
|
|
default:
|
|
|
|
return __INVALID_RANGE__;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static enum gic_intid_range get_intid_range(struct irq_data *d)
|
|
|
|
{
|
|
|
|
return __get_intid_range(d->hwirq);
|
|
|
|
}
|
|
|
|
|
2014-06-30 15:01:31 +00:00
|
|
|
static inline unsigned int gic_irq(struct irq_data *d)
|
|
|
|
{
|
|
|
|
return d->hwirq;
|
|
|
|
}
|
|
|
|
|
2020-04-25 14:11:20 +00:00
|
|
|
static inline bool gic_irq_in_rdist(struct irq_data *d)
|
2014-06-30 15:01:31 +00:00
|
|
|
{
|
2020-04-25 14:11:20 +00:00
|
|
|
switch (get_intid_range(d)) {
|
|
|
|
case SGI_RANGE:
|
|
|
|
case PPI_RANGE:
|
|
|
|
case EPPI_RANGE:
|
|
|
|
return true;
|
|
|
|
default:
|
|
|
|
return false;
|
|
|
|
}
|
2014-06-30 15:01:31 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
static inline void __iomem *gic_dist_base(struct irq_data *d)
|
|
|
|
{
|
2019-07-16 13:41:40 +00:00
|
|
|
switch (get_intid_range(d)) {
|
2020-04-25 14:11:20 +00:00
|
|
|
case SGI_RANGE:
|
2019-07-16 13:41:40 +00:00
|
|
|
case PPI_RANGE:
|
2019-07-18 12:19:25 +00:00
|
|
|
case EPPI_RANGE:
|
2019-07-16 13:41:40 +00:00
|
|
|
/* SGI+PPI -> SGI_base for this CPU */
|
2014-06-30 15:01:31 +00:00
|
|
|
return gic_data_rdist_sgi_base();
|
|
|
|
|
2019-07-16 13:41:40 +00:00
|
|
|
case SPI_RANGE:
|
2019-07-16 14:17:31 +00:00
|
|
|
case ESPI_RANGE:
|
2019-07-16 13:41:40 +00:00
|
|
|
/* SPI -> dist_base */
|
2014-06-30 15:01:31 +00:00
|
|
|
return gic_data.dist_base;
|
|
|
|
|
2019-07-16 13:41:40 +00:00
|
|
|
default:
|
|
|
|
return NULL;
|
|
|
|
}
|
2014-06-30 15:01:31 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
static void gic_do_wait_for_rwp(void __iomem *base)
|
|
|
|
{
|
|
|
|
u32 count = 1000000; /* 1s! */
|
|
|
|
|
|
|
|
while (readl_relaxed(base + GICD_CTLR) & GICD_CTLR_RWP) {
|
|
|
|
count--;
|
|
|
|
if (!count) {
|
|
|
|
pr_err_ratelimited("RWP timeout, gone fishing\n");
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
cpu_relax();
|
|
|
|
udelay(1);
|
2019-10-17 08:25:29 +00:00
|
|
|
}
|
2014-06-30 15:01:31 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/* Wait for completion of a distributor change */
|
|
|
|
static void gic_dist_wait_for_rwp(void)
|
|
|
|
{
|
|
|
|
gic_do_wait_for_rwp(gic_data.dist_base);
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Wait for completion of a redistributor change */
|
|
|
|
static void gic_redist_wait_for_rwp(void)
|
|
|
|
{
|
|
|
|
gic_do_wait_for_rwp(gic_data_rdist_rd_base());
|
|
|
|
}
|
|
|
|
|
2015-10-01 12:47:14 +00:00
|
|
|
#ifdef CONFIG_ARM64
|
2015-09-21 20:58:35 +00:00
|
|
|
|
|
|
|
static u64 __maybe_unused gic_read_iar(void)
|
|
|
|
{
|
2016-11-08 13:56:20 +00:00
|
|
|
if (cpus_have_const_cap(ARM64_WORKAROUND_CAVIUM_23154))
|
2015-09-21 20:58:35 +00:00
|
|
|
return gic_read_iar_cavium_thunderx();
|
|
|
|
else
|
|
|
|
return gic_read_iar_common();
|
|
|
|
}
|
2015-10-01 12:47:14 +00:00
|
|
|
#endif
|
2014-06-30 15:01:31 +00:00
|
|
|
|
2014-08-26 15:03:34 +00:00
|
|
|
static void gic_enable_redist(bool enable)
|
2014-06-30 15:01:31 +00:00
|
|
|
{
|
|
|
|
void __iomem *rbase;
|
|
|
|
u32 count = 1000000; /* 1s! */
|
|
|
|
u32 val;
|
|
|
|
|
2018-12-10 13:56:32 +00:00
|
|
|
if (gic_data.flags & FLAGS_WORKAROUND_GICR_WAKER_MSM8996)
|
|
|
|
return;
|
|
|
|
|
2014-06-30 15:01:31 +00:00
|
|
|
rbase = gic_data_rdist_rd_base();
|
|
|
|
|
|
|
|
val = readl_relaxed(rbase + GICR_WAKER);
|
2014-08-26 15:03:34 +00:00
|
|
|
if (enable)
|
|
|
|
/* Wake up this CPU redistributor */
|
|
|
|
val &= ~GICR_WAKER_ProcessorSleep;
|
|
|
|
else
|
|
|
|
val |= GICR_WAKER_ProcessorSleep;
|
2014-06-30 15:01:31 +00:00
|
|
|
writel_relaxed(val, rbase + GICR_WAKER);
|
|
|
|
|
2014-08-26 15:03:34 +00:00
|
|
|
if (!enable) { /* Check that GICR_WAKER is writeable */
|
|
|
|
val = readl_relaxed(rbase + GICR_WAKER);
|
|
|
|
if (!(val & GICR_WAKER_ProcessorSleep))
|
|
|
|
return; /* No PM support in this redistributor */
|
|
|
|
}
|
|
|
|
|
2016-10-14 07:26:21 +00:00
|
|
|
while (--count) {
|
2014-08-26 15:03:34 +00:00
|
|
|
val = readl_relaxed(rbase + GICR_WAKER);
|
2016-05-11 19:23:17 +00:00
|
|
|
if (enable ^ (bool)(val & GICR_WAKER_ChildrenAsleep))
|
2014-08-26 15:03:34 +00:00
|
|
|
break;
|
2014-06-30 15:01:31 +00:00
|
|
|
cpu_relax();
|
|
|
|
udelay(1);
|
2019-10-17 08:25:29 +00:00
|
|
|
}
|
2014-08-26 15:03:34 +00:00
|
|
|
if (!count)
|
|
|
|
pr_err_ratelimited("redistributor failed to %s...\n",
|
|
|
|
enable ? "wakeup" : "sleep");
|
2014-06-30 15:01:31 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Routines to disable, enable, EOI and route interrupts
|
|
|
|
*/
|
2019-07-16 13:41:40 +00:00
|
|
|
static u32 convert_offset_index(struct irq_data *d, u32 offset, u32 *index)
|
|
|
|
{
|
|
|
|
switch (get_intid_range(d)) {
|
2020-04-25 14:11:20 +00:00
|
|
|
case SGI_RANGE:
|
2019-07-16 13:41:40 +00:00
|
|
|
case PPI_RANGE:
|
|
|
|
case SPI_RANGE:
|
|
|
|
*index = d->hwirq;
|
|
|
|
return offset;
|
2019-07-18 12:19:25 +00:00
|
|
|
case EPPI_RANGE:
|
|
|
|
/*
|
|
|
|
* Contrary to the ESPI range, the EPPI range is contiguous
|
|
|
|
* to the PPI range in the registers, so let's adjust the
|
|
|
|
* displacement accordingly. Consistency is overrated.
|
|
|
|
*/
|
|
|
|
*index = d->hwirq - EPPI_BASE_INTID + 32;
|
|
|
|
return offset;
|
2019-07-16 14:17:31 +00:00
|
|
|
case ESPI_RANGE:
|
|
|
|
*index = d->hwirq - ESPI_BASE_INTID;
|
|
|
|
switch (offset) {
|
|
|
|
case GICD_ISENABLER:
|
|
|
|
return GICD_ISENABLERnE;
|
|
|
|
case GICD_ICENABLER:
|
|
|
|
return GICD_ICENABLERnE;
|
|
|
|
case GICD_ISPENDR:
|
|
|
|
return GICD_ISPENDRnE;
|
|
|
|
case GICD_ICPENDR:
|
|
|
|
return GICD_ICPENDRnE;
|
|
|
|
case GICD_ISACTIVER:
|
|
|
|
return GICD_ISACTIVERnE;
|
|
|
|
case GICD_ICACTIVER:
|
|
|
|
return GICD_ICACTIVERnE;
|
|
|
|
case GICD_IPRIORITYR:
|
|
|
|
return GICD_IPRIORITYRnE;
|
|
|
|
case GICD_ICFGR:
|
|
|
|
return GICD_ICFGRnE;
|
|
|
|
case GICD_IROUTER:
|
|
|
|
return GICD_IROUTERnE;
|
|
|
|
default:
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
break;
|
2019-07-16 13:41:40 +00:00
|
|
|
default:
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
WARN_ON(1);
|
|
|
|
*index = d->hwirq;
|
|
|
|
return offset;
|
|
|
|
}
|
|
|
|
|
2015-03-18 11:01:24 +00:00
|
|
|
static int gic_peek_irq(struct irq_data *d, u32 offset)
|
|
|
|
{
|
|
|
|
void __iomem *base;
|
2019-07-16 13:41:40 +00:00
|
|
|
u32 index, mask;
|
|
|
|
|
|
|
|
offset = convert_offset_index(d, offset, &index);
|
|
|
|
mask = 1 << (index % 32);
|
2015-03-18 11:01:24 +00:00
|
|
|
|
|
|
|
if (gic_irq_in_rdist(d))
|
|
|
|
base = gic_data_rdist_sgi_base();
|
|
|
|
else
|
|
|
|
base = gic_data.dist_base;
|
|
|
|
|
2019-07-16 13:41:40 +00:00
|
|
|
return !!(readl_relaxed(base + offset + (index / 32) * 4) & mask);
|
2015-03-18 11:01:24 +00:00
|
|
|
}
|
|
|
|
|
2014-06-30 15:01:31 +00:00
|
|
|
static void gic_poke_irq(struct irq_data *d, u32 offset)
|
|
|
|
{
|
|
|
|
void (*rwp_wait)(void);
|
|
|
|
void __iomem *base;
|
2019-07-16 13:41:40 +00:00
|
|
|
u32 index, mask;
|
|
|
|
|
|
|
|
offset = convert_offset_index(d, offset, &index);
|
|
|
|
mask = 1 << (index % 32);
|
2014-06-30 15:01:31 +00:00
|
|
|
|
|
|
|
if (gic_irq_in_rdist(d)) {
|
|
|
|
base = gic_data_rdist_sgi_base();
|
|
|
|
rwp_wait = gic_redist_wait_for_rwp;
|
|
|
|
} else {
|
|
|
|
base = gic_data.dist_base;
|
|
|
|
rwp_wait = gic_dist_wait_for_rwp;
|
|
|
|
}
|
|
|
|
|
2019-07-16 13:41:40 +00:00
|
|
|
writel_relaxed(mask, base + offset + (index / 32) * 4);
|
2014-06-30 15:01:31 +00:00
|
|
|
rwp_wait();
|
|
|
|
}
|
|
|
|
|
|
|
|
static void gic_mask_irq(struct irq_data *d)
|
|
|
|
{
|
|
|
|
gic_poke_irq(d, GICD_ICENABLER);
|
|
|
|
}
|
|
|
|
|
2015-08-26 16:00:42 +00:00
|
|
|
static void gic_eoimode1_mask_irq(struct irq_data *d)
|
|
|
|
{
|
|
|
|
gic_mask_irq(d);
|
2015-08-26 16:00:43 +00:00
|
|
|
/*
|
|
|
|
* When masking a forwarded interrupt, make sure it is
|
|
|
|
* deactivated as well.
|
|
|
|
*
|
|
|
|
* This ensures that an interrupt that is getting
|
|
|
|
* disabled/masked will not get "stuck", because there is
|
|
|
|
* noone to deactivate it (guest is being terminated).
|
|
|
|
*/
|
2015-09-15 11:19:16 +00:00
|
|
|
if (irqd_is_forwarded_to_vcpu(d))
|
2015-08-26 16:00:43 +00:00
|
|
|
gic_poke_irq(d, GICD_ICACTIVER);
|
2015-08-26 16:00:42 +00:00
|
|
|
}
|
|
|
|
|
2014-06-30 15:01:31 +00:00
|
|
|
static void gic_unmask_irq(struct irq_data *d)
|
|
|
|
{
|
|
|
|
gic_poke_irq(d, GICD_ISENABLER);
|
|
|
|
}
|
|
|
|
|
2019-01-31 14:58:57 +00:00
|
|
|
static inline bool gic_supports_nmi(void)
|
|
|
|
{
|
|
|
|
return IS_ENABLED(CONFIG_ARM64_PSEUDO_NMI) &&
|
|
|
|
static_branch_likely(&supports_pseudo_nmis);
|
|
|
|
}
|
|
|
|
|
2015-03-18 11:01:24 +00:00
|
|
|
static int gic_irq_set_irqchip_state(struct irq_data *d,
|
|
|
|
enum irqchip_irq_state which, bool val)
|
|
|
|
{
|
|
|
|
u32 reg;
|
|
|
|
|
2020-04-25 14:24:01 +00:00
|
|
|
if (d->hwirq >= 8192) /* SGI/PPI/SPI only */
|
2015-03-18 11:01:24 +00:00
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
switch (which) {
|
|
|
|
case IRQCHIP_STATE_PENDING:
|
|
|
|
reg = val ? GICD_ISPENDR : GICD_ICPENDR;
|
|
|
|
break;
|
|
|
|
|
|
|
|
case IRQCHIP_STATE_ACTIVE:
|
|
|
|
reg = val ? GICD_ISACTIVER : GICD_ICACTIVER;
|
|
|
|
break;
|
|
|
|
|
|
|
|
case IRQCHIP_STATE_MASKED:
|
|
|
|
reg = val ? GICD_ICENABLER : GICD_ISENABLER;
|
|
|
|
break;
|
|
|
|
|
|
|
|
default:
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
|
|
|
|
gic_poke_irq(d, reg);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int gic_irq_get_irqchip_state(struct irq_data *d,
|
|
|
|
enum irqchip_irq_state which, bool *val)
|
|
|
|
{
|
2019-07-16 14:17:31 +00:00
|
|
|
if (d->hwirq >= 8192) /* PPI/SPI only */
|
2015-03-18 11:01:24 +00:00
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
switch (which) {
|
|
|
|
case IRQCHIP_STATE_PENDING:
|
|
|
|
*val = gic_peek_irq(d, GICD_ISPENDR);
|
|
|
|
break;
|
|
|
|
|
|
|
|
case IRQCHIP_STATE_ACTIVE:
|
|
|
|
*val = gic_peek_irq(d, GICD_ISACTIVER);
|
|
|
|
break;
|
|
|
|
|
|
|
|
case IRQCHIP_STATE_MASKED:
|
|
|
|
*val = !gic_peek_irq(d, GICD_ISENABLER);
|
|
|
|
break;
|
|
|
|
|
|
|
|
default:
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2019-01-31 14:58:59 +00:00
|
|
|
static void gic_irq_set_prio(struct irq_data *d, u8 prio)
|
|
|
|
{
|
|
|
|
void __iomem *base = gic_dist_base(d);
|
2019-07-16 13:41:40 +00:00
|
|
|
u32 offset, index;
|
2019-01-31 14:58:59 +00:00
|
|
|
|
2019-07-16 13:41:40 +00:00
|
|
|
offset = convert_offset_index(d, GICD_IPRIORITYR, &index);
|
|
|
|
|
|
|
|
writeb_relaxed(prio, base + offset + index);
|
2019-01-31 14:58:59 +00:00
|
|
|
}
|
|
|
|
|
2021-07-29 17:27:47 +00:00
|
|
|
static u32 __gic_get_ppi_index(irq_hw_number_t hwirq)
|
2019-07-18 11:53:05 +00:00
|
|
|
{
|
2021-07-29 17:27:47 +00:00
|
|
|
switch (__get_intid_range(hwirq)) {
|
2019-07-18 11:53:05 +00:00
|
|
|
case PPI_RANGE:
|
2021-07-29 17:27:47 +00:00
|
|
|
return hwirq - 16;
|
2019-07-18 12:19:25 +00:00
|
|
|
case EPPI_RANGE:
|
2021-07-29 17:27:47 +00:00
|
|
|
return hwirq - EPPI_BASE_INTID + 16;
|
2019-07-18 11:53:05 +00:00
|
|
|
default:
|
|
|
|
unreachable();
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2021-07-29 17:27:47 +00:00
|
|
|
static u32 gic_get_ppi_index(struct irq_data *d)
|
|
|
|
{
|
|
|
|
return __gic_get_ppi_index(d->hwirq);
|
|
|
|
}
|
|
|
|
|
2019-01-31 14:58:59 +00:00
|
|
|
static int gic_irq_nmi_setup(struct irq_data *d)
|
|
|
|
{
|
|
|
|
struct irq_desc *desc = irq_to_desc(d->irq);
|
|
|
|
|
|
|
|
if (!gic_supports_nmi())
|
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
if (gic_peek_irq(d, GICD_ISENABLER)) {
|
|
|
|
pr_err("Cannot set NMI property of enabled IRQ %u\n", d->irq);
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* A secondary irq_chip should be in charge of LPI request,
|
|
|
|
* it should not be possible to get there
|
|
|
|
*/
|
|
|
|
if (WARN_ON(gic_irq(d) >= 8192))
|
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
/* desc lock should already be held */
|
2019-07-18 11:53:05 +00:00
|
|
|
if (gic_irq_in_rdist(d)) {
|
|
|
|
u32 idx = gic_get_ppi_index(d);
|
|
|
|
|
2019-01-31 14:58:59 +00:00
|
|
|
/* Setting up PPI as NMI, only switch handler for first NMI */
|
2019-07-18 11:53:05 +00:00
|
|
|
if (!refcount_inc_not_zero(&ppi_nmi_refs[idx])) {
|
|
|
|
refcount_set(&ppi_nmi_refs[idx], 1);
|
2019-01-31 14:58:59 +00:00
|
|
|
desc->handle_irq = handle_percpu_devid_fasteoi_nmi;
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
desc->handle_irq = handle_fasteoi_nmi;
|
|
|
|
}
|
|
|
|
|
|
|
|
gic_irq_set_prio(d, GICD_INT_NMI_PRI);
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void gic_irq_nmi_teardown(struct irq_data *d)
|
|
|
|
{
|
|
|
|
struct irq_desc *desc = irq_to_desc(d->irq);
|
|
|
|
|
|
|
|
if (WARN_ON(!gic_supports_nmi()))
|
|
|
|
return;
|
|
|
|
|
|
|
|
if (gic_peek_irq(d, GICD_ISENABLER)) {
|
|
|
|
pr_err("Cannot set NMI property of enabled IRQ %u\n", d->irq);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* A secondary irq_chip should be in charge of LPI request,
|
|
|
|
* it should not be possible to get there
|
|
|
|
*/
|
|
|
|
if (WARN_ON(gic_irq(d) >= 8192))
|
|
|
|
return;
|
|
|
|
|
|
|
|
/* desc lock should already be held */
|
2019-07-18 11:53:05 +00:00
|
|
|
if (gic_irq_in_rdist(d)) {
|
|
|
|
u32 idx = gic_get_ppi_index(d);
|
|
|
|
|
2019-01-31 14:58:59 +00:00
|
|
|
/* Tearing down NMI, only switch handler for last NMI */
|
2019-07-18 11:53:05 +00:00
|
|
|
if (refcount_dec_and_test(&ppi_nmi_refs[idx]))
|
2019-01-31 14:58:59 +00:00
|
|
|
desc->handle_irq = handle_percpu_devid_irq;
|
|
|
|
} else {
|
|
|
|
desc->handle_irq = handle_fasteoi_irq;
|
|
|
|
}
|
|
|
|
|
|
|
|
gic_irq_set_prio(d, GICD_INT_DEF_PRI);
|
|
|
|
}
|
|
|
|
|
2014-06-30 15:01:31 +00:00
|
|
|
static void gic_eoi_irq(struct irq_data *d)
|
|
|
|
{
|
|
|
|
gic_write_eoir(gic_irq(d));
|
|
|
|
}
|
|
|
|
|
2015-08-26 16:00:42 +00:00
|
|
|
static void gic_eoimode1_eoi_irq(struct irq_data *d)
|
|
|
|
{
|
|
|
|
/*
|
2015-08-26 16:00:43 +00:00
|
|
|
* No need to deactivate an LPI, or an interrupt that
|
|
|
|
* is is getting forwarded to a vcpu.
|
2015-08-26 16:00:42 +00:00
|
|
|
*/
|
2015-09-15 11:19:16 +00:00
|
|
|
if (gic_irq(d) >= 8192 || irqd_is_forwarded_to_vcpu(d))
|
2015-08-26 16:00:42 +00:00
|
|
|
return;
|
|
|
|
gic_write_dir(gic_irq(d));
|
|
|
|
}
|
|
|
|
|
2014-06-30 15:01:31 +00:00
|
|
|
static int gic_set_type(struct irq_data *d, unsigned int type)
|
|
|
|
{
|
2019-07-18 12:19:25 +00:00
|
|
|
enum gic_intid_range range;
|
2014-06-30 15:01:31 +00:00
|
|
|
unsigned int irq = gic_irq(d);
|
|
|
|
void (*rwp_wait)(void);
|
|
|
|
void __iomem *base;
|
2019-07-16 13:41:40 +00:00
|
|
|
u32 offset, index;
|
2019-07-16 13:35:17 +00:00
|
|
|
int ret;
|
2014-06-30 15:01:31 +00:00
|
|
|
|
2019-07-18 12:19:25 +00:00
|
|
|
range = get_intid_range(d);
|
|
|
|
|
2020-04-25 14:24:01 +00:00
|
|
|
/* Interrupt configuration for SGIs can't be changed */
|
|
|
|
if (range == SGI_RANGE)
|
|
|
|
return type != IRQ_TYPE_EDGE_RISING ? -EINVAL : 0;
|
|
|
|
|
2015-01-20 16:52:59 +00:00
|
|
|
/* SPIs have restrictions on the supported types */
|
2019-07-18 12:19:25 +00:00
|
|
|
if ((range == SPI_RANGE || range == ESPI_RANGE) &&
|
|
|
|
type != IRQ_TYPE_LEVEL_HIGH && type != IRQ_TYPE_EDGE_RISING)
|
2014-06-30 15:01:31 +00:00
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
if (gic_irq_in_rdist(d)) {
|
|
|
|
base = gic_data_rdist_sgi_base();
|
|
|
|
rwp_wait = gic_redist_wait_for_rwp;
|
|
|
|
} else {
|
|
|
|
base = gic_data.dist_base;
|
|
|
|
rwp_wait = gic_dist_wait_for_rwp;
|
|
|
|
}
|
|
|
|
|
2019-07-16 13:41:40 +00:00
|
|
|
offset = convert_offset_index(d, GICD_ICFGR, &index);
|
2019-07-16 13:35:17 +00:00
|
|
|
|
2019-07-16 13:41:40 +00:00
|
|
|
ret = gic_configure_irq(index, type, base + offset, rwp_wait);
|
2019-07-18 12:19:25 +00:00
|
|
|
if (ret && (range == PPI_RANGE || range == EPPI_RANGE)) {
|
2019-07-16 13:35:17 +00:00
|
|
|
/* Misconfigured PPIs are usually not fatal */
|
2019-07-18 12:19:25 +00:00
|
|
|
pr_warn("GIC: PPI INTID%d is secure or misconfigured\n", irq);
|
2019-07-16 13:35:17 +00:00
|
|
|
ret = 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
return ret;
|
2014-06-30 15:01:31 +00:00
|
|
|
}
|
|
|
|
|
2015-08-26 16:00:43 +00:00
|
|
|
static int gic_irq_set_vcpu_affinity(struct irq_data *d, void *vcpu)
|
|
|
|
{
|
2020-04-25 14:24:01 +00:00
|
|
|
if (get_intid_range(d) == SGI_RANGE)
|
|
|
|
return -EINVAL;
|
|
|
|
|
2015-09-15 11:19:16 +00:00
|
|
|
if (vcpu)
|
|
|
|
irqd_set_forwarded_to_vcpu(d);
|
|
|
|
else
|
|
|
|
irqd_clr_forwarded_to_vcpu(d);
|
2015-08-26 16:00:43 +00:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2015-10-01 12:47:15 +00:00
|
|
|
static u64 gic_mpidr_to_affinity(unsigned long mpidr)
|
2014-06-30 15:01:31 +00:00
|
|
|
{
|
|
|
|
u64 aff;
|
|
|
|
|
2015-10-01 12:47:15 +00:00
|
|
|
aff = ((u64)MPIDR_AFFINITY_LEVEL(mpidr, 3) << 32 |
|
2014-06-30 15:01:31 +00:00
|
|
|
MPIDR_AFFINITY_LEVEL(mpidr, 2) << 16 |
|
|
|
|
MPIDR_AFFINITY_LEVEL(mpidr, 1) << 8 |
|
|
|
|
MPIDR_AFFINITY_LEVEL(mpidr, 0));
|
|
|
|
|
|
|
|
return aff;
|
|
|
|
}
|
|
|
|
|
2019-01-31 14:58:58 +00:00
|
|
|
static void gic_deactivate_unhandled(u32 irqnr)
|
|
|
|
{
|
|
|
|
if (static_branch_likely(&supports_deactivate_key)) {
|
|
|
|
if (irqnr < 8192)
|
|
|
|
gic_write_dir(irqnr);
|
|
|
|
} else {
|
|
|
|
gic_write_eoir(irqnr);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void gic_handle_nmi(u32 irqnr, struct pt_regs *regs)
|
|
|
|
{
|
2019-06-11 09:38:09 +00:00
|
|
|
bool irqs_enabled = interrupts_enabled(regs);
|
2019-01-31 14:58:58 +00:00
|
|
|
int err;
|
|
|
|
|
2019-06-11 09:38:09 +00:00
|
|
|
if (irqs_enabled)
|
|
|
|
nmi_enter();
|
|
|
|
|
2019-01-31 14:58:58 +00:00
|
|
|
if (static_branch_likely(&supports_deactivate_key))
|
|
|
|
gic_write_eoir(irqnr);
|
|
|
|
/*
|
|
|
|
* Leave the PSR.I bit set to prevent other NMIs to be
|
|
|
|
* received while handling this one.
|
|
|
|
* PSR.I will be restored when we ERET to the
|
|
|
|
* interrupted context.
|
|
|
|
*/
|
2021-10-20 19:23:09 +00:00
|
|
|
err = generic_handle_domain_nmi(gic_data.domain, irqnr);
|
2019-01-31 14:58:58 +00:00
|
|
|
if (err)
|
|
|
|
gic_deactivate_unhandled(irqnr);
|
2019-06-11 09:38:09 +00:00
|
|
|
|
|
|
|
if (irqs_enabled)
|
|
|
|
nmi_exit();
|
2019-01-31 14:58:58 +00:00
|
|
|
}
|
|
|
|
|
2021-06-10 14:13:46 +00:00
|
|
|
static u32 do_read_iar(struct pt_regs *regs)
|
|
|
|
{
|
|
|
|
u32 iar;
|
|
|
|
|
|
|
|
if (gic_supports_nmi() && unlikely(!interrupts_enabled(regs))) {
|
|
|
|
u64 pmr;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* We were in a context with IRQs disabled. However, the
|
|
|
|
* entry code has set PMR to a value that allows any
|
|
|
|
* interrupt to be acknowledged, and not just NMIs. This can
|
|
|
|
* lead to surprising effects if the NMI has been retired in
|
|
|
|
* the meantime, and that there is an IRQ pending. The IRQ
|
|
|
|
* would then be taken in NMI context, something that nobody
|
|
|
|
* wants to debug twice.
|
|
|
|
*
|
|
|
|
* Until we sort this, drop PMR again to a level that will
|
|
|
|
* actually only allow NMIs before reading IAR, and then
|
|
|
|
* restore it to what it was.
|
|
|
|
*/
|
|
|
|
pmr = gic_read_pmr();
|
|
|
|
gic_pmr_mask_irqs();
|
|
|
|
isb();
|
|
|
|
|
|
|
|
iar = gic_read_iar();
|
|
|
|
|
|
|
|
gic_write_pmr(pmr);
|
|
|
|
} else {
|
|
|
|
iar = gic_read_iar();
|
|
|
|
}
|
|
|
|
|
|
|
|
return iar;
|
|
|
|
}
|
|
|
|
|
2014-06-30 15:01:31 +00:00
|
|
|
static asmlinkage void __exception_irq_entry gic_handle_irq(struct pt_regs *regs)
|
|
|
|
{
|
2015-10-01 12:47:15 +00:00
|
|
|
u32 irqnr;
|
2014-06-30 15:01:31 +00:00
|
|
|
|
2021-06-10 14:13:46 +00:00
|
|
|
irqnr = do_read_iar(regs);
|
2014-06-30 15:01:31 +00:00
|
|
|
|
irqchip/gic-v3: Do not enable irqs when handling spurious interrups
We triggered the following error while running our 4.19 kernel
with the pseudo-NMI patches backported to it:
[ 14.816231] ------------[ cut here ]------------
[ 14.816231] kernel BUG at irq.c:99!
[ 14.816232] Internal error: Oops - BUG: 0 [#1] SMP
[ 14.816232] Process swapper/0 (pid: 0, stack limit = 0x(____ptrval____))
[ 14.816233] CPU: 0 PID: 0 Comm: swapper/0 Tainted: G O 4.19.95.aarch64 #14
[ 14.816233] Hardware name: evb (DT)
[ 14.816234] pstate: 80400085 (Nzcv daIf +PAN -UAO)
[ 14.816234] pc : asm_nmi_enter+0x94/0x98
[ 14.816235] lr : asm_nmi_enter+0x18/0x98
[ 14.816235] sp : ffff000008003c50
[ 14.816235] pmr_save: 00000070
[ 14.816237] x29: ffff000008003c50 x28: ffff0000095f56c0
[ 14.816238] x27: 0000000000000000 x26: ffff000008004000
[ 14.816239] x25: 00000000015e0000 x24: ffff8008fb916000
[ 14.816240] x23: 0000000020400005 x22: ffff0000080817cc
[ 14.816241] x21: ffff000008003da0 x20: 0000000000000060
[ 14.816242] x19: 00000000000003ff x18: ffffffffffffffff
[ 14.816243] x17: 0000000000000008 x16: 003d090000000000
[ 14.816244] x15: ffff0000095ea6c8 x14: ffff8008fff5ab40
[ 14.816244] x13: ffff8008fff58b9d x12: 0000000000000000
[ 14.816245] x11: ffff000008c8a200 x10: 000000008e31fca5
[ 14.816246] x9 : ffff000008c8a208 x8 : 000000000000000f
[ 14.816247] x7 : 0000000000000004 x6 : ffff8008fff58b9e
[ 14.816248] x5 : 0000000000000000 x4 : 0000000080000000
[ 14.816249] x3 : 0000000000000000 x2 : 0000000080000000
[ 14.816250] x1 : 0000000000120000 x0 : ffff0000095f56c0
[ 14.816251] Call trace:
[ 14.816251] asm_nmi_enter+0x94/0x98
[ 14.816251] el1_irq+0x8c/0x180 (IRQ C)
[ 14.816252] gic_handle_irq+0xbc/0x2e4
[ 14.816252] el1_irq+0xcc/0x180 (IRQ B)
[ 14.816253] arch_timer_handler_virt+0x38/0x58
[ 14.816253] handle_percpu_devid_irq+0x90/0x240
[ 14.816253] generic_handle_irq+0x34/0x50
[ 14.816254] __handle_domain_irq+0x68/0xc0
[ 14.816254] gic_handle_irq+0xf8/0x2e4
[ 14.816255] el1_irq+0xcc/0x180 (IRQ A)
[ 14.816255] arch_cpu_idle+0x34/0x1c8
[ 14.816255] default_idle_call+0x24/0x44
[ 14.816256] do_idle+0x1d0/0x2c8
[ 14.816256] cpu_startup_entry+0x28/0x30
[ 14.816256] rest_init+0xb8/0xc8
[ 14.816257] start_kernel+0x4c8/0x4f4
[ 14.816257] Code: 940587f1 d5384100 b9401001 36a7fd01 (d4210000)
[ 14.816258] Modules linked in: start_dp(O) smeth(O)
[ 15.103092] ---[ end trace 701753956cb14aa8 ]---
[ 15.103093] Kernel panic - not syncing: Fatal exception in interrupt
[ 15.103099] SMP: stopping secondary CPUs
[ 15.103100] Kernel Offset: disabled
[ 15.103100] CPU features: 0x36,a2400218
[ 15.103100] Memory Limit: none
which is cause by a 'BUG_ON(in_nmi())' in nmi_enter().
From the call trace, we can find three interrupts (noted A, B, C above):
interrupt (A) is preempted by (B), which is further interrupted by (C).
Subsequent investigations show that (B) results in nmi_enter() being
called, but that it actually is a spurious interrupt. Furthermore,
interrupts are reenabled in the context of (B), and (C) fires with
NMI priority. We end-up with a nested NMI situation, something
we definitely do not want to (and cannot) handle.
The bug here is that spurious interrupts should never result in any
state change, and we should just return to the interrupted context.
Moving the handling of spurious interrupts as early as possible in
the GICv3 handler fixes this issue.
Fixes: 3f1f3234bc2d ("irqchip/gic-v3: Switch to PMR masking before calling IRQ handler")
Acked-by: Mark Rutland <mark.rutland@arm.com>
Signed-off-by: He Ying <heying24@huawei.com>
[maz: rewrote commit message, corrected Fixes: tag]
Signed-off-by: Marc Zyngier <maz@kernel.org>
Link: https://lore.kernel.org/r/20210423083516.170111-1-heying24@huawei.com
Cc: stable@vger.kernel.org
2021-04-23 08:35:16 +00:00
|
|
|
/* Check for special IDs first */
|
|
|
|
if ((irqnr >= 1020 && irqnr <= 1023))
|
|
|
|
return;
|
|
|
|
|
2019-01-31 14:58:58 +00:00
|
|
|
if (gic_supports_nmi() &&
|
2021-08-11 17:15:05 +00:00
|
|
|
unlikely(gic_read_rpr() == GICD_INT_RPR_PRI(GICD_INT_NMI_PRI))) {
|
2019-01-31 14:58:58 +00:00
|
|
|
gic_handle_nmi(irqnr, regs);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
2019-01-31 14:58:44 +00:00
|
|
|
if (gic_prio_masking_enabled()) {
|
|
|
|
gic_pmr_mask_irqs();
|
|
|
|
gic_arch_enable_irqs();
|
|
|
|
}
|
|
|
|
|
2020-04-25 14:24:01 +00:00
|
|
|
if (static_branch_likely(&supports_deactivate_key))
|
2018-08-28 15:51:29 +00:00
|
|
|
gic_write_eoir(irqnr);
|
2020-04-25 14:24:01 +00:00
|
|
|
else
|
|
|
|
isb();
|
|
|
|
|
2021-10-20 19:23:09 +00:00
|
|
|
if (generic_handle_domain_irq(gic_data.domain, irqnr)) {
|
2020-04-25 14:24:01 +00:00
|
|
|
WARN_ONCE(true, "Unexpected interrupt received!\n");
|
|
|
|
gic_deactivate_unhandled(irqnr);
|
2018-08-28 15:51:29 +00:00
|
|
|
}
|
2014-06-30 15:01:31 +00:00
|
|
|
}
|
|
|
|
|
2019-01-31 14:58:54 +00:00
|
|
|
static u32 gic_get_pribits(void)
|
|
|
|
{
|
|
|
|
u32 pribits;
|
|
|
|
|
|
|
|
pribits = gic_read_ctlr();
|
|
|
|
pribits &= ICC_CTLR_EL1_PRI_BITS_MASK;
|
|
|
|
pribits >>= ICC_CTLR_EL1_PRI_BITS_SHIFT;
|
|
|
|
pribits++;
|
|
|
|
|
|
|
|
return pribits;
|
|
|
|
}
|
|
|
|
|
|
|
|
static bool gic_has_group0(void)
|
|
|
|
{
|
|
|
|
u32 val;
|
2019-01-31 14:58:55 +00:00
|
|
|
u32 old_pmr;
|
|
|
|
|
|
|
|
old_pmr = gic_read_pmr();
|
2019-01-31 14:58:54 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Let's find out if Group0 is under control of EL3 or not by
|
|
|
|
* setting the highest possible, non-zero priority in PMR.
|
|
|
|
*
|
|
|
|
* If SCR_EL3.FIQ is set, the priority gets shifted down in
|
|
|
|
* order for the CPU interface to set bit 7, and keep the
|
|
|
|
* actual priority in the non-secure range. In the process, it
|
|
|
|
* looses the least significant bit and the actual priority
|
|
|
|
* becomes 0x80. Reading it back returns 0, indicating that
|
|
|
|
* we're don't have access to Group0.
|
|
|
|
*/
|
|
|
|
gic_write_pmr(BIT(8 - gic_get_pribits()));
|
|
|
|
val = gic_read_pmr();
|
|
|
|
|
2019-01-31 14:58:55 +00:00
|
|
|
gic_write_pmr(old_pmr);
|
|
|
|
|
2019-01-31 14:58:54 +00:00
|
|
|
return val != 0;
|
|
|
|
}
|
|
|
|
|
2014-06-30 15:01:31 +00:00
|
|
|
static void __init gic_dist_init(void)
|
|
|
|
{
|
|
|
|
unsigned int i;
|
|
|
|
u64 affinity;
|
|
|
|
void __iomem *base = gic_data.dist_base;
|
2020-03-04 20:33:08 +00:00
|
|
|
u32 val;
|
2014-06-30 15:01:31 +00:00
|
|
|
|
|
|
|
/* Disable the distributor */
|
|
|
|
writel_relaxed(0, base + GICD_CTLR);
|
|
|
|
gic_dist_wait_for_rwp();
|
|
|
|
|
2016-05-06 18:41:56 +00:00
|
|
|
/*
|
|
|
|
* Configure SPIs as non-secure Group-1. This will only matter
|
|
|
|
* if the GIC only has a single security state. This will not
|
|
|
|
* do the right thing if the kernel is running in secure mode,
|
|
|
|
* but that's not the intended use case anyway.
|
|
|
|
*/
|
2019-07-16 14:17:31 +00:00
|
|
|
for (i = 32; i < GIC_LINE_NR; i += 32)
|
2016-05-06 18:41:56 +00:00
|
|
|
writel_relaxed(~0, base + GICD_IGROUPR + i / 8);
|
|
|
|
|
2019-07-16 14:17:31 +00:00
|
|
|
/* Extended SPI range, not handled by the GICv2/GICv3 common code */
|
|
|
|
for (i = 0; i < GIC_ESPI_NR; i += 32) {
|
|
|
|
writel_relaxed(~0U, base + GICD_ICENABLERnE + i / 8);
|
|
|
|
writel_relaxed(~0U, base + GICD_ICACTIVERnE + i / 8);
|
|
|
|
}
|
|
|
|
|
|
|
|
for (i = 0; i < GIC_ESPI_NR; i += 32)
|
|
|
|
writel_relaxed(~0U, base + GICD_IGROUPRnE + i / 8);
|
|
|
|
|
|
|
|
for (i = 0; i < GIC_ESPI_NR; i += 16)
|
|
|
|
writel_relaxed(0, base + GICD_ICFGRnE + i / 4);
|
|
|
|
|
|
|
|
for (i = 0; i < GIC_ESPI_NR; i += 4)
|
|
|
|
writel_relaxed(GICD_INT_DEF_PRI_X4, base + GICD_IPRIORITYRnE + i);
|
|
|
|
|
|
|
|
/* Now do the common stuff, and wait for the distributor to drain */
|
|
|
|
gic_dist_config(base, GIC_LINE_NR, gic_dist_wait_for_rwp);
|
2014-06-30 15:01:31 +00:00
|
|
|
|
2020-03-04 20:33:08 +00:00
|
|
|
val = GICD_CTLR_ARE_NS | GICD_CTLR_ENABLE_G1A | GICD_CTLR_ENABLE_G1;
|
|
|
|
if (gic_data.rdists.gicd_typer2 & GICD_TYPER2_nASSGIcap) {
|
|
|
|
pr_info("Enabling SGIs without active state\n");
|
|
|
|
val |= GICD_CTLR_nASSGIreq;
|
|
|
|
}
|
|
|
|
|
2014-06-30 15:01:31 +00:00
|
|
|
/* Enable distributor with ARE, Group1 */
|
2020-03-04 20:33:08 +00:00
|
|
|
writel_relaxed(val, base + GICD_CTLR);
|
2014-06-30 15:01:31 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Set all global interrupts to the boot CPU only. ARE must be
|
|
|
|
* enabled.
|
|
|
|
*/
|
|
|
|
affinity = gic_mpidr_to_affinity(cpu_logical_map(smp_processor_id()));
|
2019-07-16 14:17:31 +00:00
|
|
|
for (i = 32; i < GIC_LINE_NR; i++)
|
2015-10-01 12:47:16 +00:00
|
|
|
gic_write_irouter(affinity, base + GICD_IROUTER + i * 8);
|
2019-07-16 14:17:31 +00:00
|
|
|
|
|
|
|
for (i = 0; i < GIC_ESPI_NR; i++)
|
|
|
|
gic_write_irouter(affinity, base + GICD_IROUTERnE + i * 8);
|
2014-06-30 15:01:31 +00:00
|
|
|
}
|
|
|
|
|
2016-12-19 17:00:38 +00:00
|
|
|
static int gic_iterate_rdists(int (*fn)(struct redist_region *, void __iomem *))
|
2014-06-30 15:01:31 +00:00
|
|
|
{
|
2016-12-19 17:00:38 +00:00
|
|
|
int ret = -ENODEV;
|
2014-06-30 15:01:31 +00:00
|
|
|
int i;
|
|
|
|
|
2014-11-24 14:35:10 +00:00
|
|
|
for (i = 0; i < gic_data.nr_redist_regions; i++) {
|
|
|
|
void __iomem *ptr = gic_data.redist_regions[i].redist_base;
|
2016-12-19 17:00:38 +00:00
|
|
|
u64 typer;
|
2014-06-30 15:01:31 +00:00
|
|
|
u32 reg;
|
|
|
|
|
|
|
|
reg = readl_relaxed(ptr + GICR_PIDR2) & GIC_PIDR2_ARCH_MASK;
|
|
|
|
if (reg != GIC_PIDR2_ARCH_GICv3 &&
|
|
|
|
reg != GIC_PIDR2_ARCH_GICv4) { /* We're in trouble... */
|
|
|
|
pr_warn("No redistributor present @%p\n", ptr);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
do {
|
2015-10-01 12:47:16 +00:00
|
|
|
typer = gic_read_typer(ptr + GICR_TYPER);
|
2016-12-19 17:00:38 +00:00
|
|
|
ret = fn(gic_data.redist_regions + i, ptr);
|
|
|
|
if (!ret)
|
2014-06-30 15:01:31 +00:00
|
|
|
return 0;
|
|
|
|
|
2016-01-19 13:11:16 +00:00
|
|
|
if (gic_data.redist_regions[i].single_redist)
|
|
|
|
break;
|
|
|
|
|
2014-06-30 15:01:31 +00:00
|
|
|
if (gic_data.redist_stride) {
|
|
|
|
ptr += gic_data.redist_stride;
|
|
|
|
} else {
|
|
|
|
ptr += SZ_64K * 2; /* Skip RD_base + SGI_base */
|
|
|
|
if (typer & GICR_TYPER_VLPIS)
|
|
|
|
ptr += SZ_64K * 2; /* Skip VLPI_base + reserved page */
|
|
|
|
}
|
|
|
|
} while (!(typer & GICR_TYPER_LAST));
|
|
|
|
}
|
|
|
|
|
2016-12-19 17:00:38 +00:00
|
|
|
return ret ? -ENODEV : 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int __gic_populate_rdist(struct redist_region *region, void __iomem *ptr)
|
|
|
|
{
|
|
|
|
unsigned long mpidr = cpu_logical_map(smp_processor_id());
|
|
|
|
u64 typer;
|
|
|
|
u32 aff;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Convert affinity to a 32bit value that can be matched to
|
|
|
|
* GICR_TYPER bits [63:32].
|
|
|
|
*/
|
|
|
|
aff = (MPIDR_AFFINITY_LEVEL(mpidr, 3) << 24 |
|
|
|
|
MPIDR_AFFINITY_LEVEL(mpidr, 2) << 16 |
|
|
|
|
MPIDR_AFFINITY_LEVEL(mpidr, 1) << 8 |
|
|
|
|
MPIDR_AFFINITY_LEVEL(mpidr, 0));
|
|
|
|
|
|
|
|
typer = gic_read_typer(ptr + GICR_TYPER);
|
|
|
|
if ((typer >> 32) == aff) {
|
|
|
|
u64 offset = ptr - region->redist_base;
|
2020-03-04 20:33:12 +00:00
|
|
|
raw_spin_lock_init(&gic_data_rdist()->rd_lock);
|
2016-12-19 17:00:38 +00:00
|
|
|
gic_data_rdist_rd_base() = ptr;
|
|
|
|
gic_data_rdist()->phys_base = region->phys_base + offset;
|
|
|
|
|
|
|
|
pr_info("CPU%d: found redistributor %lx region %d:%pa\n",
|
|
|
|
smp_processor_id(), mpidr,
|
|
|
|
(int)(region - gic_data.redist_regions),
|
|
|
|
&gic_data_rdist()->phys_base);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Try next one */
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int gic_populate_rdist(void)
|
|
|
|
{
|
|
|
|
if (gic_iterate_rdists(__gic_populate_rdist) == 0)
|
|
|
|
return 0;
|
|
|
|
|
2014-06-30 15:01:31 +00:00
|
|
|
/* We couldn't even deal with ourselves... */
|
2015-10-01 12:47:15 +00:00
|
|
|
WARN(true, "CPU%d: mpidr %lx has no re-distributor!\n",
|
2016-12-19 17:00:38 +00:00
|
|
|
smp_processor_id(),
|
|
|
|
(unsigned long)cpu_logical_map(smp_processor_id()));
|
2014-06-30 15:01:31 +00:00
|
|
|
return -ENODEV;
|
|
|
|
}
|
|
|
|
|
2019-07-18 10:15:14 +00:00
|
|
|
static int __gic_update_rdist_properties(struct redist_region *region,
|
|
|
|
void __iomem *ptr)
|
2016-12-19 17:01:52 +00:00
|
|
|
{
|
|
|
|
u64 typer = gic_read_typer(ptr + GICR_TYPER);
|
2019-12-24 11:10:24 +00:00
|
|
|
|
2016-12-19 17:01:52 +00:00
|
|
|
gic_data.rdists.has_vlpis &= !!(typer & GICR_TYPER_VLPIS);
|
2019-12-24 11:10:24 +00:00
|
|
|
|
|
|
|
/* RVPEID implies some form of DirectLPI, no matter what the doc says... :-/ */
|
|
|
|
gic_data.rdists.has_rvpeid &= !!(typer & GICR_TYPER_RVPEID);
|
|
|
|
gic_data.rdists.has_direct_lpi &= (!!(typer & GICR_TYPER_DirectLPIS) |
|
|
|
|
gic_data.rdists.has_rvpeid);
|
2020-04-10 10:13:26 +00:00
|
|
|
gic_data.rdists.has_vpend_valid_dirty &= !!(typer & GICR_TYPER_DIRTY);
|
2019-12-24 11:10:24 +00:00
|
|
|
|
|
|
|
/* Detect non-sensical configurations */
|
|
|
|
if (WARN_ON_ONCE(gic_data.rdists.has_rvpeid && !gic_data.rdists.has_vlpis)) {
|
|
|
|
gic_data.rdists.has_direct_lpi = false;
|
|
|
|
gic_data.rdists.has_vlpis = false;
|
|
|
|
gic_data.rdists.has_rvpeid = false;
|
|
|
|
}
|
|
|
|
|
2019-07-18 12:19:25 +00:00
|
|
|
gic_data.ppi_nr = min(GICR_TYPER_NR_PPIS(typer), gic_data.ppi_nr);
|
2016-12-19 17:01:52 +00:00
|
|
|
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
2019-07-18 10:15:14 +00:00
|
|
|
static void gic_update_rdist_properties(void)
|
2016-12-19 17:01:52 +00:00
|
|
|
{
|
2019-07-18 10:15:14 +00:00
|
|
|
gic_data.ppi_nr = UINT_MAX;
|
|
|
|
gic_iterate_rdists(__gic_update_rdist_properties);
|
|
|
|
if (WARN_ON(gic_data.ppi_nr == UINT_MAX))
|
|
|
|
gic_data.ppi_nr = 0;
|
|
|
|
pr_info("%d PPIs implemented\n", gic_data.ppi_nr);
|
2020-04-10 10:13:26 +00:00
|
|
|
if (gic_data.rdists.has_vlpis)
|
|
|
|
pr_info("GICv4 features: %s%s%s\n",
|
|
|
|
gic_data.rdists.has_direct_lpi ? "DirectLPI " : "",
|
|
|
|
gic_data.rdists.has_rvpeid ? "RVPEID " : "",
|
|
|
|
gic_data.rdists.has_vpend_valid_dirty ? "Valid+Dirty " : "");
|
2016-12-19 17:01:52 +00:00
|
|
|
}
|
|
|
|
|
2019-01-31 14:58:57 +00:00
|
|
|
/* Check whether it's single security state view */
|
|
|
|
static inline bool gic_dist_security_disabled(void)
|
|
|
|
{
|
|
|
|
return readl_relaxed(gic_data.dist_base + GICD_CTLR) & GICD_CTLR_DS;
|
|
|
|
}
|
|
|
|
|
2014-08-26 15:03:35 +00:00
|
|
|
static void gic_cpu_sys_reg_init(void)
|
|
|
|
{
|
2017-10-06 15:24:00 +00:00
|
|
|
int i, cpu = smp_processor_id();
|
|
|
|
u64 mpidr = cpu_logical_map(cpu);
|
|
|
|
u64 need_rss = MPIDR_RS(mpidr);
|
2018-03-20 09:46:42 +00:00
|
|
|
bool group0;
|
2019-01-31 14:58:54 +00:00
|
|
|
u32 pribits;
|
2017-10-06 15:24:00 +00:00
|
|
|
|
2015-09-30 10:48:01 +00:00
|
|
|
/*
|
|
|
|
* Need to check that the SRE bit has actually been set. If
|
|
|
|
* not, it means that SRE is disabled at EL2. We're going to
|
|
|
|
* die painfully, and there is nothing we can do about it.
|
|
|
|
*
|
|
|
|
* Kindly inform the luser.
|
|
|
|
*/
|
|
|
|
if (!gic_enable_sre())
|
|
|
|
pr_err("GIC: unable to set SRE (disabled at EL2), panic ahead\n");
|
2014-08-26 15:03:35 +00:00
|
|
|
|
2019-01-31 14:58:54 +00:00
|
|
|
pribits = gic_get_pribits();
|
2018-03-20 09:46:42 +00:00
|
|
|
|
2019-01-31 14:58:54 +00:00
|
|
|
group0 = gic_has_group0();
|
2018-03-20 09:46:42 +00:00
|
|
|
|
2014-08-26 15:03:35 +00:00
|
|
|
/* Set priority mask register */
|
2019-01-31 14:58:57 +00:00
|
|
|
if (!gic_prio_masking_enabled()) {
|
2019-01-31 14:58:55 +00:00
|
|
|
write_gicreg(DEFAULT_PMR_VALUE, ICC_PMR_EL1);
|
2020-09-12 15:37:07 +00:00
|
|
|
} else if (gic_supports_nmi()) {
|
2019-01-31 14:58:57 +00:00
|
|
|
/*
|
|
|
|
* Mismatch configuration with boot CPU, the system is likely
|
|
|
|
* to die as interrupt masking will not work properly on all
|
|
|
|
* CPUs
|
2020-09-12 15:37:07 +00:00
|
|
|
*
|
|
|
|
* The boot CPU calls this function before enabling NMI support,
|
|
|
|
* and as a result we'll never see this warning in the boot path
|
|
|
|
* for that CPU.
|
2019-01-31 14:58:57 +00:00
|
|
|
*/
|
2020-09-12 15:37:07 +00:00
|
|
|
if (static_branch_unlikely(&gic_nonsecure_priorities))
|
|
|
|
WARN_ON(!group0 || gic_dist_security_disabled());
|
|
|
|
else
|
|
|
|
WARN_ON(group0 && !gic_dist_security_disabled());
|
2019-01-31 14:58:57 +00:00
|
|
|
}
|
2014-08-26 15:03:35 +00:00
|
|
|
|
2016-08-19 16:13:09 +00:00
|
|
|
/*
|
|
|
|
* Some firmwares hand over to the kernel with the BPR changed from
|
|
|
|
* its reset value (and with a value large enough to prevent
|
|
|
|
* any pre-emptive interrupts from working at all). Writing a zero
|
|
|
|
* to BPR restores is reset value.
|
|
|
|
*/
|
|
|
|
gic_write_bpr1(0);
|
|
|
|
|
2018-03-26 21:09:25 +00:00
|
|
|
if (static_branch_likely(&supports_deactivate_key)) {
|
2015-08-26 16:00:42 +00:00
|
|
|
/* EOI drops priority only (mode 1) */
|
|
|
|
gic_write_ctlr(ICC_CTLR_EL1_EOImode_drop);
|
|
|
|
} else {
|
|
|
|
/* EOI deactivates interrupt too (mode 0) */
|
|
|
|
gic_write_ctlr(ICC_CTLR_EL1_EOImode_drop_dir);
|
|
|
|
}
|
2014-08-26 15:03:35 +00:00
|
|
|
|
2018-03-20 09:46:42 +00:00
|
|
|
/* Always whack Group0 before Group1 */
|
|
|
|
if (group0) {
|
|
|
|
switch(pribits) {
|
|
|
|
case 8:
|
|
|
|
case 7:
|
|
|
|
write_gicreg(0, ICC_AP0R3_EL1);
|
|
|
|
write_gicreg(0, ICC_AP0R2_EL1);
|
2020-08-23 22:36:59 +00:00
|
|
|
fallthrough;
|
2018-03-20 09:46:42 +00:00
|
|
|
case 6:
|
|
|
|
write_gicreg(0, ICC_AP0R1_EL1);
|
2020-08-23 22:36:59 +00:00
|
|
|
fallthrough;
|
2018-03-20 09:46:42 +00:00
|
|
|
case 5:
|
|
|
|
case 4:
|
|
|
|
write_gicreg(0, ICC_AP0R0_EL1);
|
|
|
|
}
|
|
|
|
|
|
|
|
isb();
|
|
|
|
}
|
2018-03-09 14:53:19 +00:00
|
|
|
|
2018-03-20 09:46:42 +00:00
|
|
|
switch(pribits) {
|
2018-03-09 14:53:19 +00:00
|
|
|
case 8:
|
|
|
|
case 7:
|
|
|
|
write_gicreg(0, ICC_AP1R3_EL1);
|
|
|
|
write_gicreg(0, ICC_AP1R2_EL1);
|
2020-08-23 22:36:59 +00:00
|
|
|
fallthrough;
|
2018-03-09 14:53:19 +00:00
|
|
|
case 6:
|
|
|
|
write_gicreg(0, ICC_AP1R1_EL1);
|
2020-08-23 22:36:59 +00:00
|
|
|
fallthrough;
|
2018-03-09 14:53:19 +00:00
|
|
|
case 5:
|
|
|
|
case 4:
|
|
|
|
write_gicreg(0, ICC_AP1R0_EL1);
|
|
|
|
}
|
|
|
|
|
|
|
|
isb();
|
|
|
|
|
2014-08-26 15:03:35 +00:00
|
|
|
/* ... and let's hit the road... */
|
|
|
|
gic_write_grpen1(1);
|
2017-10-06 15:24:00 +00:00
|
|
|
|
|
|
|
/* Keep the RSS capability status in per_cpu variable */
|
|
|
|
per_cpu(has_rss, cpu) = !!(gic_read_ctlr() & ICC_CTLR_EL1_RSS);
|
|
|
|
|
|
|
|
/* Check all the CPUs have capable of sending SGIs to other CPUs */
|
|
|
|
for_each_online_cpu(i) {
|
|
|
|
bool have_rss = per_cpu(has_rss, i) && per_cpu(has_rss, cpu);
|
|
|
|
|
|
|
|
need_rss |= MPIDR_RS(cpu_logical_map(i));
|
|
|
|
if (need_rss && (!have_rss))
|
|
|
|
pr_crit("CPU%d (%lx) can't SGI CPU%d (%lx), no RSS\n",
|
|
|
|
cpu, (unsigned long)mpidr,
|
|
|
|
i, (unsigned long)cpu_logical_map(i));
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* GIC spec says, when ICC_CTLR_EL1.RSS==1 and GICD_TYPER.RSS==0,
|
|
|
|
* writing ICC_ASGI1R_EL1 register with RS != 0 is a CONSTRAINED
|
|
|
|
* UNPREDICTABLE choice of :
|
|
|
|
* - The write is ignored.
|
|
|
|
* - The RS field is treated as 0.
|
|
|
|
*/
|
|
|
|
if (need_rss && (!gic_data.has_rss))
|
|
|
|
pr_crit_once("RSS is required but GICD doesn't support it\n");
|
2014-08-26 15:03:35 +00:00
|
|
|
}
|
|
|
|
|
2018-02-25 11:27:04 +00:00
|
|
|
static bool gicv3_nolpi;
|
|
|
|
|
|
|
|
static int __init gicv3_nolpi_cfg(char *buf)
|
|
|
|
{
|
|
|
|
return strtobool(buf, &gicv3_nolpi);
|
|
|
|
}
|
|
|
|
early_param("irqchip.gicv3_nolpi", gicv3_nolpi_cfg);
|
|
|
|
|
2014-11-24 14:35:18 +00:00
|
|
|
static int gic_dist_supports_lpis(void)
|
|
|
|
{
|
2018-07-27 13:51:04 +00:00
|
|
|
return (IS_ENABLED(CONFIG_ARM_GIC_V3_ITS) &&
|
|
|
|
!!(readl_relaxed(gic_data.dist_base + GICD_TYPER) & GICD_TYPER_LPIS) &&
|
|
|
|
!gicv3_nolpi);
|
2014-11-24 14:35:18 +00:00
|
|
|
}
|
|
|
|
|
2014-06-30 15:01:31 +00:00
|
|
|
static void gic_cpu_init(void)
|
|
|
|
{
|
|
|
|
void __iomem *rbase;
|
2019-07-18 10:15:14 +00:00
|
|
|
int i;
|
2014-06-30 15:01:31 +00:00
|
|
|
|
|
|
|
/* Register ourselves with the rest of the world */
|
|
|
|
if (gic_populate_rdist())
|
|
|
|
return;
|
|
|
|
|
2014-08-26 15:03:34 +00:00
|
|
|
gic_enable_redist(true);
|
2014-06-30 15:01:31 +00:00
|
|
|
|
2019-07-25 14:30:51 +00:00
|
|
|
WARN((gic_data.ppi_nr > 16 || GIC_ESPI_NR != 0) &&
|
|
|
|
!(gic_read_ctlr() & ICC_CTLR_EL1_ExtRange),
|
|
|
|
"Distributor has extended ranges, but CPU%d doesn't\n",
|
|
|
|
smp_processor_id());
|
|
|
|
|
2014-06-30 15:01:31 +00:00
|
|
|
rbase = gic_data_rdist_sgi_base();
|
|
|
|
|
2016-05-06 18:41:56 +00:00
|
|
|
/* Configure SGIs/PPIs as non-secure Group-1 */
|
2019-07-18 10:15:14 +00:00
|
|
|
for (i = 0; i < gic_data.ppi_nr + 16; i += 32)
|
|
|
|
writel_relaxed(~0, rbase + GICR_IGROUPR0 + i / 8);
|
2016-05-06 18:41:56 +00:00
|
|
|
|
2019-07-18 10:15:14 +00:00
|
|
|
gic_cpu_config(rbase, gic_data.ppi_nr + 16, gic_redist_wait_for_rwp);
|
2014-06-30 15:01:31 +00:00
|
|
|
|
2014-08-26 15:03:35 +00:00
|
|
|
/* initialise system registers */
|
|
|
|
gic_cpu_sys_reg_init();
|
2014-06-30 15:01:31 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
#ifdef CONFIG_SMP
|
2016-07-13 17:16:05 +00:00
|
|
|
|
2017-10-06 15:24:00 +00:00
|
|
|
#define MPIDR_TO_SGI_RS(mpidr) (MPIDR_RS(mpidr) << ICC_SGI1R_RS_SHIFT)
|
|
|
|
#define MPIDR_TO_SGI_CLUSTER_ID(mpidr) ((mpidr) & ~0xFUL)
|
|
|
|
|
2016-07-13 17:16:05 +00:00
|
|
|
static int gic_starting_cpu(unsigned int cpu)
|
2014-06-30 15:01:31 +00:00
|
|
|
{
|
2016-07-13 17:16:05 +00:00
|
|
|
gic_cpu_init();
|
2018-07-27 13:51:04 +00:00
|
|
|
|
|
|
|
if (gic_dist_supports_lpis())
|
|
|
|
its_cpu_init();
|
|
|
|
|
2016-07-13 17:16:05 +00:00
|
|
|
return 0;
|
2014-06-30 15:01:31 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
static u16 gic_compute_target_list(int *base_cpu, const struct cpumask *mask,
|
2015-10-01 12:47:15 +00:00
|
|
|
unsigned long cluster_id)
|
2014-06-30 15:01:31 +00:00
|
|
|
{
|
2016-09-19 17:29:15 +00:00
|
|
|
int next_cpu, cpu = *base_cpu;
|
2015-10-01 12:47:15 +00:00
|
|
|
unsigned long mpidr = cpu_logical_map(cpu);
|
2014-06-30 15:01:31 +00:00
|
|
|
u16 tlist = 0;
|
|
|
|
|
|
|
|
while (cpu < nr_cpu_ids) {
|
|
|
|
tlist |= 1 << (mpidr & 0xf);
|
|
|
|
|
2016-09-19 17:29:15 +00:00
|
|
|
next_cpu = cpumask_next(cpu, mask);
|
|
|
|
if (next_cpu >= nr_cpu_ids)
|
2014-06-30 15:01:31 +00:00
|
|
|
goto out;
|
2016-09-19 17:29:15 +00:00
|
|
|
cpu = next_cpu;
|
2014-06-30 15:01:31 +00:00
|
|
|
|
|
|
|
mpidr = cpu_logical_map(cpu);
|
|
|
|
|
2017-10-06 15:24:00 +00:00
|
|
|
if (cluster_id != MPIDR_TO_SGI_CLUSTER_ID(mpidr)) {
|
2014-06-30 15:01:31 +00:00
|
|
|
cpu--;
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
out:
|
|
|
|
*base_cpu = cpu;
|
|
|
|
return tlist;
|
|
|
|
}
|
|
|
|
|
2014-11-12 13:46:06 +00:00
|
|
|
#define MPIDR_TO_SGI_AFFINITY(cluster_id, level) \
|
|
|
|
(MPIDR_AFFINITY_LEVEL(cluster_id, level) \
|
|
|
|
<< ICC_SGI1R_AFFINITY_## level ##_SHIFT)
|
|
|
|
|
2014-06-30 15:01:31 +00:00
|
|
|
static void gic_send_sgi(u64 cluster_id, u16 tlist, unsigned int irq)
|
|
|
|
{
|
|
|
|
u64 val;
|
|
|
|
|
2014-11-12 13:46:06 +00:00
|
|
|
val = (MPIDR_TO_SGI_AFFINITY(cluster_id, 3) |
|
|
|
|
MPIDR_TO_SGI_AFFINITY(cluster_id, 2) |
|
|
|
|
irq << ICC_SGI1R_SGI_ID_SHIFT |
|
|
|
|
MPIDR_TO_SGI_AFFINITY(cluster_id, 1) |
|
2017-10-06 15:24:00 +00:00
|
|
|
MPIDR_TO_SGI_RS(cluster_id) |
|
2014-11-12 13:46:06 +00:00
|
|
|
tlist << ICC_SGI1R_TARGET_LIST_SHIFT);
|
2014-06-30 15:01:31 +00:00
|
|
|
|
2018-02-02 14:20:29 +00:00
|
|
|
pr_devel("CPU%d: ICC_SGI1R_EL1 %llx\n", smp_processor_id(), val);
|
2014-06-30 15:01:31 +00:00
|
|
|
gic_write_sgi1r(val);
|
|
|
|
}
|
|
|
|
|
2020-04-25 14:24:01 +00:00
|
|
|
static void gic_ipi_send_mask(struct irq_data *d, const struct cpumask *mask)
|
2014-06-30 15:01:31 +00:00
|
|
|
{
|
|
|
|
int cpu;
|
|
|
|
|
2020-04-25 14:24:01 +00:00
|
|
|
if (WARN_ON(d->hwirq >= 16))
|
2014-06-30 15:01:31 +00:00
|
|
|
return;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Ensure that stores to Normal memory are visible to the
|
|
|
|
* other CPUs before issuing the IPI.
|
|
|
|
*/
|
2018-02-01 00:03:42 +00:00
|
|
|
wmb();
|
2014-06-30 15:01:31 +00:00
|
|
|
|
2015-03-05 00:19:16 +00:00
|
|
|
for_each_cpu(cpu, mask) {
|
2017-10-06 15:24:00 +00:00
|
|
|
u64 cluster_id = MPIDR_TO_SGI_CLUSTER_ID(cpu_logical_map(cpu));
|
2014-06-30 15:01:31 +00:00
|
|
|
u16 tlist;
|
|
|
|
|
|
|
|
tlist = gic_compute_target_list(&cpu, mask, cluster_id);
|
2020-04-25 14:24:01 +00:00
|
|
|
gic_send_sgi(cluster_id, tlist, d->hwirq);
|
2014-06-30 15:01:31 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/* Force the above writes to ICC_SGI1R_EL1 to be executed */
|
|
|
|
isb();
|
|
|
|
}
|
|
|
|
|
2020-04-22 11:28:57 +00:00
|
|
|
static void __init gic_smp_init(void)
|
2014-06-30 15:01:31 +00:00
|
|
|
{
|
2020-04-25 14:24:01 +00:00
|
|
|
struct irq_fwspec sgi_fwspec = {
|
|
|
|
.fwnode = gic_data.fwnode,
|
|
|
|
.param_count = 1,
|
|
|
|
};
|
|
|
|
int base_sgi;
|
|
|
|
|
2016-12-21 19:19:56 +00:00
|
|
|
cpuhp_setup_state_nocalls(CPUHP_AP_IRQ_GIC_STARTING,
|
2016-12-21 19:19:54 +00:00
|
|
|
"irqchip/arm/gicv3:starting",
|
|
|
|
gic_starting_cpu, NULL);
|
2020-04-25 14:24:01 +00:00
|
|
|
|
|
|
|
/* Register all 8 non-secure SGIs */
|
|
|
|
base_sgi = __irq_domain_alloc_irqs(gic_data.domain, -1, 8,
|
|
|
|
NUMA_NO_NODE, &sgi_fwspec,
|
|
|
|
false, NULL);
|
|
|
|
if (WARN_ON(base_sgi <= 0))
|
|
|
|
return;
|
|
|
|
|
|
|
|
set_smp_ipi_range(base_sgi, 8);
|
2014-06-30 15:01:31 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
static int gic_set_affinity(struct irq_data *d, const struct cpumask *mask_val,
|
|
|
|
bool force)
|
|
|
|
{
|
2017-07-04 09:56:35 +00:00
|
|
|
unsigned int cpu;
|
2019-07-16 13:41:40 +00:00
|
|
|
u32 offset, index;
|
2014-06-30 15:01:31 +00:00
|
|
|
void __iomem *reg;
|
|
|
|
int enabled;
|
|
|
|
u64 val;
|
|
|
|
|
2017-07-04 09:56:35 +00:00
|
|
|
if (force)
|
|
|
|
cpu = cpumask_first(mask_val);
|
|
|
|
else
|
|
|
|
cpu = cpumask_any_and(mask_val, cpu_online_mask);
|
|
|
|
|
2017-06-30 09:58:28 +00:00
|
|
|
if (cpu >= nr_cpu_ids)
|
|
|
|
return -EINVAL;
|
|
|
|
|
2014-06-30 15:01:31 +00:00
|
|
|
if (gic_irq_in_rdist(d))
|
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
/* If interrupt was enabled, disable it first */
|
|
|
|
enabled = gic_peek_irq(d, GICD_ISENABLER);
|
|
|
|
if (enabled)
|
|
|
|
gic_mask_irq(d);
|
|
|
|
|
2019-07-16 13:41:40 +00:00
|
|
|
offset = convert_offset_index(d, GICD_IROUTER, &index);
|
|
|
|
reg = gic_dist_base(d) + offset + (index * 8);
|
2014-06-30 15:01:31 +00:00
|
|
|
val = gic_mpidr_to_affinity(cpu_logical_map(cpu));
|
|
|
|
|
2015-10-01 12:47:16 +00:00
|
|
|
gic_write_irouter(val, reg);
|
2014-06-30 15:01:31 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* If the interrupt was enabled, enabled it again. Otherwise,
|
|
|
|
* just wait for the distributor to have digested our changes.
|
|
|
|
*/
|
|
|
|
if (enabled)
|
|
|
|
gic_unmask_irq(d);
|
|
|
|
else
|
|
|
|
gic_dist_wait_for_rwp();
|
|
|
|
|
2017-08-18 08:39:17 +00:00
|
|
|
irq_data_update_effective_affinity(d, cpumask_of(cpu));
|
|
|
|
|
2016-02-19 15:22:43 +00:00
|
|
|
return IRQ_SET_MASK_OK_DONE;
|
2014-06-30 15:01:31 +00:00
|
|
|
}
|
|
|
|
#else
|
|
|
|
#define gic_set_affinity NULL
|
2020-04-25 14:24:01 +00:00
|
|
|
#define gic_ipi_send_mask NULL
|
2014-06-30 15:01:31 +00:00
|
|
|
#define gic_smp_init() do { } while(0)
|
|
|
|
#endif
|
|
|
|
|
2020-07-30 17:03:20 +00:00
|
|
|
static int gic_retrigger(struct irq_data *data)
|
|
|
|
{
|
|
|
|
return !gic_irq_set_irqchip_state(data, IRQCHIP_STATE_PENDING, true);
|
|
|
|
}
|
|
|
|
|
2014-08-26 15:03:35 +00:00
|
|
|
#ifdef CONFIG_CPU_PM
|
|
|
|
static int gic_cpu_pm_notifier(struct notifier_block *self,
|
|
|
|
unsigned long cmd, void *v)
|
|
|
|
{
|
|
|
|
if (cmd == CPU_PM_EXIT) {
|
2016-08-17 12:49:19 +00:00
|
|
|
if (gic_dist_security_disabled())
|
|
|
|
gic_enable_redist(true);
|
2014-08-26 15:03:35 +00:00
|
|
|
gic_cpu_sys_reg_init();
|
2016-08-17 12:49:19 +00:00
|
|
|
} else if (cmd == CPU_PM_ENTER && gic_dist_security_disabled()) {
|
2014-08-26 15:03:35 +00:00
|
|
|
gic_write_grpen1(0);
|
|
|
|
gic_enable_redist(false);
|
|
|
|
}
|
|
|
|
return NOTIFY_OK;
|
|
|
|
}
|
|
|
|
|
|
|
|
static struct notifier_block gic_cpu_pm_notifier_block = {
|
|
|
|
.notifier_call = gic_cpu_pm_notifier,
|
|
|
|
};
|
|
|
|
|
|
|
|
static void gic_cpu_pm_init(void)
|
|
|
|
{
|
|
|
|
cpu_pm_register_notifier(&gic_cpu_pm_notifier_block);
|
|
|
|
}
|
|
|
|
|
|
|
|
#else
|
|
|
|
static inline void gic_cpu_pm_init(void) { }
|
|
|
|
#endif /* CONFIG_CPU_PM */
|
|
|
|
|
2014-06-30 15:01:31 +00:00
|
|
|
static struct irq_chip gic_chip = {
|
|
|
|
.name = "GICv3",
|
|
|
|
.irq_mask = gic_mask_irq,
|
|
|
|
.irq_unmask = gic_unmask_irq,
|
|
|
|
.irq_eoi = gic_eoi_irq,
|
|
|
|
.irq_set_type = gic_set_type,
|
|
|
|
.irq_set_affinity = gic_set_affinity,
|
2020-07-30 17:03:20 +00:00
|
|
|
.irq_retrigger = gic_retrigger,
|
2015-03-18 11:01:24 +00:00
|
|
|
.irq_get_irqchip_state = gic_irq_get_irqchip_state,
|
|
|
|
.irq_set_irqchip_state = gic_irq_set_irqchip_state,
|
2019-01-31 14:58:59 +00:00
|
|
|
.irq_nmi_setup = gic_irq_nmi_setup,
|
|
|
|
.irq_nmi_teardown = gic_irq_nmi_teardown,
|
2020-04-25 14:24:01 +00:00
|
|
|
.ipi_send_mask = gic_ipi_send_mask,
|
2018-08-17 08:18:01 +00:00
|
|
|
.flags = IRQCHIP_SET_TYPE_MASKED |
|
|
|
|
IRQCHIP_SKIP_SET_WAKE |
|
|
|
|
IRQCHIP_MASK_ON_SUSPEND,
|
2014-06-30 15:01:31 +00:00
|
|
|
};
|
|
|
|
|
2015-08-26 16:00:42 +00:00
|
|
|
static struct irq_chip gic_eoimode1_chip = {
|
|
|
|
.name = "GICv3",
|
|
|
|
.irq_mask = gic_eoimode1_mask_irq,
|
|
|
|
.irq_unmask = gic_unmask_irq,
|
|
|
|
.irq_eoi = gic_eoimode1_eoi_irq,
|
|
|
|
.irq_set_type = gic_set_type,
|
|
|
|
.irq_set_affinity = gic_set_affinity,
|
2020-07-30 17:03:20 +00:00
|
|
|
.irq_retrigger = gic_retrigger,
|
2015-08-26 16:00:42 +00:00
|
|
|
.irq_get_irqchip_state = gic_irq_get_irqchip_state,
|
|
|
|
.irq_set_irqchip_state = gic_irq_set_irqchip_state,
|
2015-08-26 16:00:43 +00:00
|
|
|
.irq_set_vcpu_affinity = gic_irq_set_vcpu_affinity,
|
2019-01-31 14:58:59 +00:00
|
|
|
.irq_nmi_setup = gic_irq_nmi_setup,
|
|
|
|
.irq_nmi_teardown = gic_irq_nmi_teardown,
|
2020-04-25 14:24:01 +00:00
|
|
|
.ipi_send_mask = gic_ipi_send_mask,
|
2018-08-17 08:18:01 +00:00
|
|
|
.flags = IRQCHIP_SET_TYPE_MASKED |
|
|
|
|
IRQCHIP_SKIP_SET_WAKE |
|
|
|
|
IRQCHIP_MASK_ON_SUSPEND,
|
2015-08-26 16:00:42 +00:00
|
|
|
};
|
|
|
|
|
2014-06-30 15:01:31 +00:00
|
|
|
static int gic_irq_domain_map(struct irq_domain *d, unsigned int irq,
|
|
|
|
irq_hw_number_t hw)
|
|
|
|
{
|
2015-08-26 16:00:42 +00:00
|
|
|
struct irq_chip *chip = &gic_chip;
|
2020-07-30 17:03:21 +00:00
|
|
|
struct irq_data *irqd = irq_desc_get_irq_data(irq_to_desc(irq));
|
2015-08-26 16:00:42 +00:00
|
|
|
|
2018-03-26 21:09:25 +00:00
|
|
|
if (static_branch_likely(&supports_deactivate_key))
|
2015-08-26 16:00:42 +00:00
|
|
|
chip = &gic_eoimode1_chip;
|
|
|
|
|
2019-07-16 13:41:40 +00:00
|
|
|
switch (__get_intid_range(hw)) {
|
2020-04-25 14:11:20 +00:00
|
|
|
case SGI_RANGE:
|
2019-07-16 13:41:40 +00:00
|
|
|
case PPI_RANGE:
|
2019-07-18 12:19:25 +00:00
|
|
|
case EPPI_RANGE:
|
2014-06-30 15:01:31 +00:00
|
|
|
irq_set_percpu_devid(irq);
|
2015-08-26 16:00:42 +00:00
|
|
|
irq_domain_set_info(d, irq, hw, chip, d->host_data,
|
2014-11-24 14:35:09 +00:00
|
|
|
handle_percpu_devid_irq, NULL, NULL);
|
2019-07-16 13:41:40 +00:00
|
|
|
break;
|
|
|
|
|
|
|
|
case SPI_RANGE:
|
2019-07-16 14:17:31 +00:00
|
|
|
case ESPI_RANGE:
|
2015-08-26 16:00:42 +00:00
|
|
|
irq_domain_set_info(d, irq, hw, chip, d->host_data,
|
2014-11-24 14:35:09 +00:00
|
|
|
handle_fasteoi_irq, NULL, NULL);
|
2015-08-29 23:01:22 +00:00
|
|
|
irq_set_probe(irq);
|
2020-07-30 17:03:21 +00:00
|
|
|
irqd_set_single_target(irqd);
|
2019-07-16 13:41:40 +00:00
|
|
|
break;
|
|
|
|
|
|
|
|
case LPI_RANGE:
|
2014-11-24 14:35:18 +00:00
|
|
|
if (!gic_dist_supports_lpis())
|
|
|
|
return -EPERM;
|
2015-08-26 16:00:42 +00:00
|
|
|
irq_domain_set_info(d, irq, hw, chip, d->host_data,
|
2014-11-24 14:35:18 +00:00
|
|
|
handle_fasteoi_irq, NULL, NULL);
|
2019-07-16 13:41:40 +00:00
|
|
|
break;
|
|
|
|
|
|
|
|
default:
|
|
|
|
return -EPERM;
|
2014-11-24 14:35:18 +00:00
|
|
|
}
|
|
|
|
|
2020-07-30 17:03:21 +00:00
|
|
|
/* Prevents SW retriggers which mess up the ACK/EOI ordering */
|
|
|
|
irqd_set_handle_enforce_irqctx(irqd);
|
2014-06-30 15:01:31 +00:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2015-10-13 11:51:33 +00:00
|
|
|
static int gic_irq_domain_translate(struct irq_domain *d,
|
|
|
|
struct irq_fwspec *fwspec,
|
|
|
|
unsigned long *hwirq,
|
|
|
|
unsigned int *type)
|
2014-06-30 15:01:31 +00:00
|
|
|
{
|
2020-04-25 14:24:01 +00:00
|
|
|
if (fwspec->param_count == 1 && fwspec->param[0] < 16) {
|
|
|
|
*hwirq = fwspec->param[0];
|
|
|
|
*type = IRQ_TYPE_EDGE_RISING;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2015-10-13 11:51:33 +00:00
|
|
|
if (is_of_node(fwspec->fwnode)) {
|
|
|
|
if (fwspec->param_count < 3)
|
|
|
|
return -EINVAL;
|
2014-06-30 15:01:31 +00:00
|
|
|
|
2015-10-14 11:27:16 +00:00
|
|
|
switch (fwspec->param[0]) {
|
|
|
|
case 0: /* SPI */
|
|
|
|
*hwirq = fwspec->param[1] + 32;
|
|
|
|
break;
|
|
|
|
case 1: /* PPI */
|
|
|
|
*hwirq = fwspec->param[1] + 16;
|
|
|
|
break;
|
2019-07-16 14:17:31 +00:00
|
|
|
case 2: /* ESPI */
|
|
|
|
*hwirq = fwspec->param[1] + ESPI_BASE_INTID;
|
|
|
|
break;
|
2019-07-18 12:19:25 +00:00
|
|
|
case 3: /* EPPI */
|
|
|
|
*hwirq = fwspec->param[1] + EPPI_BASE_INTID;
|
|
|
|
break;
|
2015-10-14 11:27:16 +00:00
|
|
|
case GIC_IRQ_TYPE_LPI: /* LPI */
|
|
|
|
*hwirq = fwspec->param[1];
|
|
|
|
break;
|
2019-07-18 12:19:25 +00:00
|
|
|
case GIC_IRQ_TYPE_PARTITION:
|
|
|
|
*hwirq = fwspec->param[1];
|
|
|
|
if (fwspec->param[1] >= 16)
|
|
|
|
*hwirq += EPPI_BASE_INTID - 16;
|
|
|
|
else
|
|
|
|
*hwirq += 16;
|
|
|
|
break;
|
2015-10-14 11:27:16 +00:00
|
|
|
default:
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
2015-10-13 11:51:33 +00:00
|
|
|
|
|
|
|
*type = fwspec->param[2] & IRQ_TYPE_SENSE_MASK;
|
2018-03-16 14:35:17 +00:00
|
|
|
|
2018-03-20 13:44:09 +00:00
|
|
|
/*
|
|
|
|
* Make it clear that broken DTs are... broken.
|
2021-03-22 03:21:30 +00:00
|
|
|
* Partitioned PPIs are an unfortunate exception.
|
2018-03-20 13:44:09 +00:00
|
|
|
*/
|
|
|
|
WARN_ON(*type == IRQ_TYPE_NONE &&
|
|
|
|
fwspec->param[0] != GIC_IRQ_TYPE_PARTITION);
|
2015-10-13 11:51:33 +00:00
|
|
|
return 0;
|
2014-06-30 15:01:31 +00:00
|
|
|
}
|
|
|
|
|
2016-01-19 13:11:15 +00:00
|
|
|
if (is_fwnode_irqchip(fwspec->fwnode)) {
|
|
|
|
if(fwspec->param_count != 2)
|
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
*hwirq = fwspec->param[0];
|
|
|
|
*type = fwspec->param[1];
|
2018-03-16 14:35:17 +00:00
|
|
|
|
|
|
|
WARN_ON(*type == IRQ_TYPE_NONE);
|
2016-01-19 13:11:15 +00:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2015-10-13 11:51:33 +00:00
|
|
|
return -EINVAL;
|
2014-06-30 15:01:31 +00:00
|
|
|
}
|
|
|
|
|
2014-11-24 14:35:09 +00:00
|
|
|
static int gic_irq_domain_alloc(struct irq_domain *domain, unsigned int virq,
|
|
|
|
unsigned int nr_irqs, void *arg)
|
|
|
|
{
|
|
|
|
int i, ret;
|
|
|
|
irq_hw_number_t hwirq;
|
|
|
|
unsigned int type = IRQ_TYPE_NONE;
|
2015-10-13 11:51:33 +00:00
|
|
|
struct irq_fwspec *fwspec = arg;
|
2014-11-24 14:35:09 +00:00
|
|
|
|
2015-10-13 11:51:33 +00:00
|
|
|
ret = gic_irq_domain_translate(domain, fwspec, &hwirq, &type);
|
2014-11-24 14:35:09 +00:00
|
|
|
if (ret)
|
|
|
|
return ret;
|
|
|
|
|
2017-07-04 09:56:33 +00:00
|
|
|
for (i = 0; i < nr_irqs; i++) {
|
|
|
|
ret = gic_irq_domain_map(domain, virq + i, hwirq + i);
|
|
|
|
if (ret)
|
|
|
|
return ret;
|
|
|
|
}
|
2014-11-24 14:35:09 +00:00
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void gic_irq_domain_free(struct irq_domain *domain, unsigned int virq,
|
|
|
|
unsigned int nr_irqs)
|
|
|
|
{
|
|
|
|
int i;
|
|
|
|
|
|
|
|
for (i = 0; i < nr_irqs; i++) {
|
|
|
|
struct irq_data *d = irq_domain_get_irq_data(domain, virq + i);
|
|
|
|
irq_set_handler(virq + i, NULL);
|
|
|
|
irq_domain_reset_irq_data(d);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2021-07-29 17:27:48 +00:00
|
|
|
static bool fwspec_is_partitioned_ppi(struct irq_fwspec *fwspec,
|
|
|
|
irq_hw_number_t hwirq)
|
|
|
|
{
|
|
|
|
enum gic_intid_range range;
|
|
|
|
|
|
|
|
if (!gic_data.ppi_descs)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
if (!is_of_node(fwspec->fwnode))
|
|
|
|
return false;
|
|
|
|
|
|
|
|
if (fwspec->param_count < 4 || !fwspec->param[3])
|
|
|
|
return false;
|
|
|
|
|
|
|
|
range = __get_intid_range(hwirq);
|
|
|
|
if (range != PPI_RANGE && range != EPPI_RANGE)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2016-04-11 08:57:54 +00:00
|
|
|
static int gic_irq_domain_select(struct irq_domain *d,
|
|
|
|
struct irq_fwspec *fwspec,
|
|
|
|
enum irq_domain_bus_token bus_token)
|
|
|
|
{
|
2021-07-29 17:27:48 +00:00
|
|
|
unsigned int type, ret, ppi_idx;
|
|
|
|
irq_hw_number_t hwirq;
|
|
|
|
|
2016-04-11 08:57:54 +00:00
|
|
|
/* Not for us */
|
|
|
|
if (fwspec->fwnode != d->fwnode)
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
/* If this is not DT, then we have a single domain */
|
|
|
|
if (!is_of_node(fwspec->fwnode))
|
|
|
|
return 1;
|
|
|
|
|
2021-07-29 17:27:48 +00:00
|
|
|
ret = gic_irq_domain_translate(d, fwspec, &hwirq, &type);
|
|
|
|
if (WARN_ON_ONCE(ret))
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
if (!fwspec_is_partitioned_ppi(fwspec, hwirq))
|
|
|
|
return d == gic_data.domain;
|
|
|
|
|
2016-04-11 08:57:54 +00:00
|
|
|
/*
|
|
|
|
* If this is a PPI and we have a 4th (non-null) parameter,
|
|
|
|
* then we need to match the partition domain.
|
|
|
|
*/
|
2021-07-29 17:27:48 +00:00
|
|
|
ppi_idx = __gic_get_ppi_index(hwirq);
|
|
|
|
return d == partition_get_domain(gic_data.ppi_descs[ppi_idx]);
|
2016-04-11 08:57:54 +00:00
|
|
|
}
|
|
|
|
|
2014-06-30 15:01:31 +00:00
|
|
|
static const struct irq_domain_ops gic_irq_domain_ops = {
|
2015-10-13 11:51:33 +00:00
|
|
|
.translate = gic_irq_domain_translate,
|
2014-11-24 14:35:09 +00:00
|
|
|
.alloc = gic_irq_domain_alloc,
|
|
|
|
.free = gic_irq_domain_free,
|
2016-04-11 08:57:54 +00:00
|
|
|
.select = gic_irq_domain_select,
|
|
|
|
};
|
|
|
|
|
|
|
|
static int partition_domain_translate(struct irq_domain *d,
|
|
|
|
struct irq_fwspec *fwspec,
|
|
|
|
unsigned long *hwirq,
|
|
|
|
unsigned int *type)
|
|
|
|
{
|
2021-07-29 17:27:48 +00:00
|
|
|
unsigned long ppi_intid;
|
2016-04-11 08:57:54 +00:00
|
|
|
struct device_node *np;
|
2021-07-29 17:27:48 +00:00
|
|
|
unsigned int ppi_idx;
|
2016-04-11 08:57:54 +00:00
|
|
|
int ret;
|
|
|
|
|
2019-07-18 12:05:17 +00:00
|
|
|
if (!gic_data.ppi_descs)
|
|
|
|
return -ENOMEM;
|
|
|
|
|
2016-04-11 08:57:54 +00:00
|
|
|
np = of_find_node_by_phandle(fwspec->param[3]);
|
|
|
|
if (WARN_ON(!np))
|
|
|
|
return -EINVAL;
|
|
|
|
|
2021-07-29 17:27:48 +00:00
|
|
|
ret = gic_irq_domain_translate(d, fwspec, &ppi_intid, type);
|
|
|
|
if (WARN_ON_ONCE(ret))
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
ppi_idx = __gic_get_ppi_index(ppi_intid);
|
|
|
|
ret = partition_translate_id(gic_data.ppi_descs[ppi_idx],
|
2016-04-11 08:57:54 +00:00
|
|
|
of_node_to_fwnode(np));
|
|
|
|
if (ret < 0)
|
|
|
|
return ret;
|
|
|
|
|
|
|
|
*hwirq = ret;
|
|
|
|
*type = fwspec->param[2] & IRQ_TYPE_SENSE_MASK;
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static const struct irq_domain_ops partition_domain_ops = {
|
|
|
|
.translate = partition_domain_translate,
|
|
|
|
.select = gic_irq_domain_select,
|
2014-06-30 15:01:31 +00:00
|
|
|
};
|
|
|
|
|
2018-12-10 13:56:32 +00:00
|
|
|
static bool gic_enable_quirk_msm8996(void *data)
|
|
|
|
{
|
|
|
|
struct gic_chip_data *d = data;
|
|
|
|
|
|
|
|
d->flags |= FLAGS_WORKAROUND_GICR_WAKER_MSM8996;
|
|
|
|
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2020-03-11 11:56:49 +00:00
|
|
|
static bool gic_enable_quirk_cavium_38539(void *data)
|
|
|
|
{
|
|
|
|
struct gic_chip_data *d = data;
|
|
|
|
|
|
|
|
d->flags |= FLAGS_WORKAROUND_CAVIUM_ERRATUM_38539;
|
|
|
|
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2019-07-31 16:29:33 +00:00
|
|
|
static bool gic_enable_quirk_hip06_07(void *data)
|
|
|
|
{
|
|
|
|
struct gic_chip_data *d = data;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* HIP06 GICD_IIDR clashes with GIC-600 product number (despite
|
|
|
|
* not being an actual ARM implementation). The saving grace is
|
|
|
|
* that GIC-600 doesn't have ESPI, so nothing to do in that case.
|
|
|
|
* HIP07 doesn't even have a proper IIDR, and still pretends to
|
|
|
|
* have ESPI. In both cases, put them right.
|
|
|
|
*/
|
|
|
|
if (d->rdists.gicd_typer & GICD_TYPER_ESPI) {
|
|
|
|
/* Zero both ESPI and the RES0 field next to it... */
|
|
|
|
d->rdists.gicd_typer &= ~GENMASK(9, 8);
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
static const struct gic_quirk gic_quirks[] = {
|
|
|
|
{
|
|
|
|
.desc = "GICv3: Qualcomm MSM8996 broken firmware",
|
|
|
|
.compatible = "qcom,msm8996-gic-v3",
|
|
|
|
.init = gic_enable_quirk_msm8996,
|
|
|
|
},
|
|
|
|
{
|
|
|
|
.desc = "GICv3: HIP06 erratum 161010803",
|
|
|
|
.iidr = 0x0204043b,
|
|
|
|
.mask = 0xffffffff,
|
|
|
|
.init = gic_enable_quirk_hip06_07,
|
|
|
|
},
|
|
|
|
{
|
|
|
|
.desc = "GICv3: HIP07 erratum 161010803",
|
|
|
|
.iidr = 0x00000000,
|
|
|
|
.mask = 0xffffffff,
|
|
|
|
.init = gic_enable_quirk_hip06_07,
|
|
|
|
},
|
2020-03-11 11:56:49 +00:00
|
|
|
{
|
|
|
|
/*
|
|
|
|
* Reserved register accesses generate a Synchronous
|
|
|
|
* External Abort. This erratum applies to:
|
|
|
|
* - ThunderX: CN88xx
|
|
|
|
* - OCTEON TX: CN83xx, CN81xx
|
|
|
|
* - OCTEON TX2: CN93xx, CN96xx, CN98xx, CNF95xx*
|
|
|
|
*/
|
|
|
|
.desc = "GICv3: Cavium erratum 38539",
|
|
|
|
.iidr = 0xa000034c,
|
|
|
|
.mask = 0xe8f00fff,
|
|
|
|
.init = gic_enable_quirk_cavium_38539,
|
|
|
|
},
|
2019-07-31 16:29:33 +00:00
|
|
|
{
|
|
|
|
}
|
|
|
|
};
|
|
|
|
|
2019-01-31 14:58:57 +00:00
|
|
|
static void gic_enable_nmi_support(void)
|
|
|
|
{
|
2019-01-31 14:58:59 +00:00
|
|
|
int i;
|
|
|
|
|
2019-07-18 11:53:05 +00:00
|
|
|
if (!gic_prio_masking_enabled())
|
|
|
|
return;
|
|
|
|
|
|
|
|
ppi_nmi_refs = kcalloc(gic_data.ppi_nr, sizeof(*ppi_nmi_refs), GFP_KERNEL);
|
|
|
|
if (!ppi_nmi_refs)
|
|
|
|
return;
|
|
|
|
|
|
|
|
for (i = 0; i < gic_data.ppi_nr; i++)
|
2019-01-31 14:58:59 +00:00
|
|
|
refcount_set(&ppi_nmi_refs[i], 0);
|
|
|
|
|
2019-10-02 09:06:12 +00:00
|
|
|
/*
|
|
|
|
* Linux itself doesn't use 1:N distribution, so has no need to
|
|
|
|
* set PMHE. The only reason to have it set is if EL3 requires it
|
|
|
|
* (and we can't change it).
|
|
|
|
*/
|
|
|
|
if (gic_read_ctlr() & ICC_CTLR_EL1_PMHE_MASK)
|
|
|
|
static_branch_enable(&gic_pmr_sync);
|
|
|
|
|
2020-09-12 15:37:06 +00:00
|
|
|
pr_info("Pseudo-NMIs enabled using %s ICC_PMR_EL1 synchronisation\n",
|
|
|
|
static_branch_unlikely(&gic_pmr_sync) ? "forced" : "relaxed");
|
2019-10-02 09:06:12 +00:00
|
|
|
|
2020-09-12 15:37:07 +00:00
|
|
|
/*
|
|
|
|
* How priority values are used by the GIC depends on two things:
|
|
|
|
* the security state of the GIC (controlled by the GICD_CTRL.DS bit)
|
|
|
|
* and if Group 0 interrupts can be delivered to Linux in the non-secure
|
|
|
|
* world as FIQs (controlled by the SCR_EL3.FIQ bit). These affect the
|
|
|
|
* the ICC_PMR_EL1 register and the priority that software assigns to
|
|
|
|
* interrupts:
|
|
|
|
*
|
|
|
|
* GICD_CTRL.DS | SCR_EL3.FIQ | ICC_PMR_EL1 | Group 1 priority
|
|
|
|
* -----------------------------------------------------------
|
|
|
|
* 1 | - | unchanged | unchanged
|
|
|
|
* -----------------------------------------------------------
|
|
|
|
* 0 | 1 | non-secure | non-secure
|
|
|
|
* -----------------------------------------------------------
|
|
|
|
* 0 | 0 | unchanged | non-secure
|
|
|
|
*
|
|
|
|
* where non-secure means that the value is right-shifted by one and the
|
|
|
|
* MSB bit set, to make it fit in the non-secure priority range.
|
|
|
|
*
|
|
|
|
* In the first two cases, where ICC_PMR_EL1 and the interrupt priority
|
|
|
|
* are both either modified or unchanged, we can use the same set of
|
|
|
|
* priorities.
|
|
|
|
*
|
|
|
|
* In the last case, where only the interrupt priorities are modified to
|
|
|
|
* be in the non-secure range, we use a different PMR value to mask IRQs
|
|
|
|
* and the rest of the values that we use remain unchanged.
|
|
|
|
*/
|
|
|
|
if (gic_has_group0() && !gic_dist_security_disabled())
|
|
|
|
static_branch_enable(&gic_nonsecure_priorities);
|
|
|
|
|
2019-01-31 14:58:57 +00:00
|
|
|
static_branch_enable(&supports_pseudo_nmis);
|
2019-01-31 14:58:59 +00:00
|
|
|
|
|
|
|
if (static_branch_likely(&supports_deactivate_key))
|
|
|
|
gic_eoimode1_chip.flags |= IRQCHIP_SUPPORTS_NMI;
|
|
|
|
else
|
|
|
|
gic_chip.flags |= IRQCHIP_SUPPORTS_NMI;
|
2019-01-31 14:58:57 +00:00
|
|
|
}
|
|
|
|
|
2016-01-19 13:11:14 +00:00
|
|
|
static int __init gic_init_bases(void __iomem *dist_base,
|
|
|
|
struct redist_region *rdist_regs,
|
|
|
|
u32 nr_redist_regions,
|
|
|
|
u64 redist_stride,
|
|
|
|
struct fwnode_handle *handle)
|
2014-06-30 15:01:31 +00:00
|
|
|
{
|
2014-11-24 14:35:10 +00:00
|
|
|
u32 typer;
|
2014-06-30 15:01:31 +00:00
|
|
|
int err;
|
|
|
|
|
2015-08-26 16:00:42 +00:00
|
|
|
if (!is_hyp_mode_available())
|
2018-03-26 21:09:25 +00:00
|
|
|
static_branch_disable(&supports_deactivate_key);
|
2015-08-26 16:00:42 +00:00
|
|
|
|
2018-03-26 21:09:25 +00:00
|
|
|
if (static_branch_likely(&supports_deactivate_key))
|
2015-08-26 16:00:42 +00:00
|
|
|
pr_info("GIC: Using split EOI/Deactivate mode\n");
|
|
|
|
|
2016-04-11 08:57:54 +00:00
|
|
|
gic_data.fwnode = handle;
|
2014-06-30 15:01:31 +00:00
|
|
|
gic_data.dist_base = dist_base;
|
2014-11-24 14:35:10 +00:00
|
|
|
gic_data.redist_regions = rdist_regs;
|
|
|
|
gic_data.nr_redist_regions = nr_redist_regions;
|
2014-06-30 15:01:31 +00:00
|
|
|
gic_data.redist_stride = redist_stride;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Find out how many interrupts are supported.
|
|
|
|
*/
|
2014-11-24 14:35:10 +00:00
|
|
|
typer = readl_relaxed(gic_data.dist_base + GICD_TYPER);
|
2018-05-30 16:29:52 +00:00
|
|
|
gic_data.rdists.gicd_typer = typer;
|
2019-07-31 16:29:33 +00:00
|
|
|
|
|
|
|
gic_enable_quirks(readl_relaxed(gic_data.dist_base + GICD_IIDR),
|
|
|
|
gic_quirks, &gic_data);
|
|
|
|
|
2019-07-16 14:17:31 +00:00
|
|
|
pr_info("%d SPIs implemented\n", GIC_LINE_NR - 32);
|
|
|
|
pr_info("%d Extended SPIs implemented\n", GIC_ESPI_NR);
|
2019-12-24 11:10:25 +00:00
|
|
|
|
2020-03-11 11:56:49 +00:00
|
|
|
/*
|
|
|
|
* ThunderX1 explodes on reading GICD_TYPER2, in violation of the
|
|
|
|
* architecture spec (which says that reserved registers are RES0).
|
|
|
|
*/
|
|
|
|
if (!(gic_data.flags & FLAGS_WORKAROUND_CAVIUM_ERRATUM_38539))
|
|
|
|
gic_data.rdists.gicd_typer2 = readl_relaxed(gic_data.dist_base + GICD_TYPER2);
|
2019-12-24 11:10:25 +00:00
|
|
|
|
2016-01-19 13:11:14 +00:00
|
|
|
gic_data.domain = irq_domain_create_tree(handle, &gic_irq_domain_ops,
|
|
|
|
&gic_data);
|
2014-11-24 14:35:10 +00:00
|
|
|
gic_data.rdists.rdist = alloc_percpu(typeof(*gic_data.rdists.rdist));
|
2019-12-24 11:10:24 +00:00
|
|
|
gic_data.rdists.has_rvpeid = true;
|
2016-12-19 17:01:52 +00:00
|
|
|
gic_data.rdists.has_vlpis = true;
|
|
|
|
gic_data.rdists.has_direct_lpi = true;
|
2020-04-10 10:13:26 +00:00
|
|
|
gic_data.rdists.has_vpend_valid_dirty = true;
|
2014-06-30 15:01:31 +00:00
|
|
|
|
2014-11-24 14:35:10 +00:00
|
|
|
if (WARN_ON(!gic_data.domain) || WARN_ON(!gic_data.rdists.rdist)) {
|
2014-06-30 15:01:31 +00:00
|
|
|
err = -ENOMEM;
|
|
|
|
goto out_free;
|
|
|
|
}
|
|
|
|
|
2020-03-12 03:20:55 +00:00
|
|
|
irq_domain_update_bus_token(gic_data.domain, DOMAIN_BUS_WIRED);
|
|
|
|
|
2017-10-06 15:24:00 +00:00
|
|
|
gic_data.has_rss = !!(typer & GICD_TYPER_RSS);
|
|
|
|
pr_info("Distributor has %sRange Selector support\n",
|
|
|
|
gic_data.has_rss ? "" : "no ");
|
|
|
|
|
2018-05-08 12:14:36 +00:00
|
|
|
if (typer & GICD_TYPER_MBIS) {
|
|
|
|
err = mbi_init(handle, gic_data.domain);
|
|
|
|
if (err)
|
|
|
|
pr_err("Failed to initialize MBIs\n");
|
|
|
|
}
|
|
|
|
|
2014-06-30 15:01:31 +00:00
|
|
|
set_handle_irq(gic_handle_irq);
|
|
|
|
|
2019-07-18 10:15:14 +00:00
|
|
|
gic_update_rdist_properties();
|
2016-12-19 17:01:52 +00:00
|
|
|
|
2014-06-30 15:01:31 +00:00
|
|
|
gic_dist_init();
|
|
|
|
gic_cpu_init();
|
2020-04-25 14:24:01 +00:00
|
|
|
gic_smp_init();
|
2014-08-26 15:03:35 +00:00
|
|
|
gic_cpu_pm_init();
|
2014-06-30 15:01:31 +00:00
|
|
|
|
2018-07-27 13:51:04 +00:00
|
|
|
if (gic_dist_supports_lpis()) {
|
|
|
|
its_init(handle, &gic_data.rdists, gic_data.domain);
|
|
|
|
its_cpu_init();
|
2019-06-10 10:52:01 +00:00
|
|
|
} else {
|
|
|
|
if (IS_ENABLED(CONFIG_ARM_GIC_V2M))
|
|
|
|
gicv2m_init(handle, gic_data.domain);
|
2018-07-27 13:51:04 +00:00
|
|
|
}
|
|
|
|
|
2019-07-18 11:53:05 +00:00
|
|
|
gic_enable_nmi_support();
|
2019-01-31 14:58:57 +00:00
|
|
|
|
2014-06-30 15:01:31 +00:00
|
|
|
return 0;
|
|
|
|
|
|
|
|
out_free:
|
|
|
|
if (gic_data.domain)
|
|
|
|
irq_domain_remove(gic_data.domain);
|
2014-11-24 14:35:10 +00:00
|
|
|
free_percpu(gic_data.rdists.rdist);
|
2016-01-19 13:11:14 +00:00
|
|
|
return err;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int __init gic_validate_dist_version(void __iomem *dist_base)
|
|
|
|
{
|
|
|
|
u32 reg = readl_relaxed(dist_base + GICD_PIDR2) & GIC_PIDR2_ARCH_MASK;
|
|
|
|
|
|
|
|
if (reg != GIC_PIDR2_ARCH_GICv3 && reg != GIC_PIDR2_ARCH_GICv4)
|
|
|
|
return -ENODEV;
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2016-04-11 08:57:54 +00:00
|
|
|
/* Create all possible partitions at boot time */
|
Small release overall.
- x86: miscellaneous fixes, AVIC support (local APIC virtualization,
AMD version)
- s390: polling for interrupts after a VCPU goes to halted state is
now enabled for s390; use hardware provided information about facility
bits that do not need any hypervisor activity, and other fixes for
cpu models and facilities; improve perf output; floating interrupt
controller improvements.
- MIPS: miscellaneous fixes
- PPC: bugfixes only
- ARM: 16K page size support, generic firmware probing layer for
timer and GIC
Christoffer Dall (KVM-ARM maintainer) says:
"There are a few changes in this pull request touching things outside
KVM, but they should all carry the necessary acks and it made the
merge process much easier to do it this way."
though actually the irqchip maintainers' acks didn't make it into the
patches. Marc Zyngier, who is both irqchip and KVM-ARM maintainer,
later acked at http://mid.gmane.org/573351D1.4060303@arm.com
"more formally and for documentation purposes".
-----BEGIN PGP SIGNATURE-----
Version: GnuPG v2.0.22 (GNU/Linux)
iQEcBAABAgAGBQJXPJjyAAoJEL/70l94x66DhioH/j4fwQ0FmfPSM9PArzaFHQdx
LNE3tU4+bobbsy1BJr4DiAaOUQn3DAgwUvGLWXdeLiOXtoWXBiFHKaxlqEsCA6iQ
xcTH1TgfxsVoqGQ6bT9X/2GCx70heYpcWG3f+zqBy7ZfFmQykLAC/HwOr52VQL8f
hUFi3YmTHcnorp0n5Xg+9r3+RBS4D/kTbtdn6+KCLnPJ0RcgNkI3/NcafTemoofw
Tkv8+YYFNvKV13qlIfVqxMa0GwWI3pP6YaNKhaS5XO8Pu16HuuF1JthJsUBDzwBa
RInp8R9MoXgsBYhLpz3jc9vWG7G9yDl5LehsD9KOUGOaFYJ7sQN+QZOusa6jFgA=
=llO5
-----END PGP SIGNATURE-----
Merge tag 'for-linus' of git://git.kernel.org/pub/scm/virt/kvm/kvm
Pull KVM updates from Paolo Bonzini:
"Small release overall.
x86:
- miscellaneous fixes
- AVIC support (local APIC virtualization, AMD version)
s390:
- polling for interrupts after a VCPU goes to halted state is now
enabled for s390
- use hardware provided information about facility bits that do not
need any hypervisor activity, and other fixes for cpu models and
facilities
- improve perf output
- floating interrupt controller improvements.
MIPS:
- miscellaneous fixes
PPC:
- bugfixes only
ARM:
- 16K page size support
- generic firmware probing layer for timer and GIC
Christoffer Dall (KVM-ARM maintainer) says:
"There are a few changes in this pull request touching things
outside KVM, but they should all carry the necessary acks and it
made the merge process much easier to do it this way."
though actually the irqchip maintainers' acks didn't make it into the
patches. Marc Zyngier, who is both irqchip and KVM-ARM maintainer,
later acked at http://mid.gmane.org/573351D1.4060303@arm.com ('more
formally and for documentation purposes')"
* tag 'for-linus' of git://git.kernel.org/pub/scm/virt/kvm/kvm: (82 commits)
KVM: MTRR: remove MSR 0x2f8
KVM: x86: make hwapic_isr_update and hwapic_irr_update look the same
svm: Manage vcpu load/unload when enable AVIC
svm: Do not intercept CR8 when enable AVIC
svm: Do not expose x2APIC when enable AVIC
KVM: x86: Introducing kvm_x86_ops.apicv_post_state_restore
svm: Add VMEXIT handlers for AVIC
svm: Add interrupt injection via AVIC
KVM: x86: Detect and Initialize AVIC support
svm: Introduce new AVIC VMCB registers
KVM: split kvm_vcpu_wake_up from kvm_vcpu_kick
KVM: x86: Introducing kvm_x86_ops VCPU blocking/unblocking hooks
KVM: x86: Introducing kvm_x86_ops VM init/destroy hooks
KVM: x86: Rename kvm_apic_get_reg to kvm_lapic_get_reg
KVM: x86: Misc LAPIC changes to expose helper functions
KVM: shrink halt polling even more for invalid wakeups
KVM: s390: set halt polling to 80 microseconds
KVM: halt_polling: provide a way to qualify wakeups during poll
KVM: PPC: Book3S HV: Re-enable XICS fast path for irqfd-generated interrupts
kvm: Conditionally register IRQ bypass consumer
...
2016-05-19 18:27:09 +00:00
|
|
|
static void __init gic_populate_ppi_partitions(struct device_node *gic_node)
|
2016-04-11 08:57:54 +00:00
|
|
|
{
|
|
|
|
struct device_node *parts_node, *child_part;
|
|
|
|
int part_idx = 0, i;
|
|
|
|
int nr_parts;
|
|
|
|
struct partition_affinity *parts;
|
|
|
|
|
2017-11-11 16:51:25 +00:00
|
|
|
parts_node = of_get_child_by_name(gic_node, "ppi-partitions");
|
2016-04-11 08:57:54 +00:00
|
|
|
if (!parts_node)
|
|
|
|
return;
|
|
|
|
|
2019-07-18 12:05:17 +00:00
|
|
|
gic_data.ppi_descs = kcalloc(gic_data.ppi_nr, sizeof(*gic_data.ppi_descs), GFP_KERNEL);
|
|
|
|
if (!gic_data.ppi_descs)
|
|
|
|
return;
|
|
|
|
|
2016-04-11 08:57:54 +00:00
|
|
|
nr_parts = of_get_child_count(parts_node);
|
|
|
|
|
|
|
|
if (!nr_parts)
|
2017-11-11 16:51:25 +00:00
|
|
|
goto out_put_node;
|
2016-04-11 08:57:54 +00:00
|
|
|
|
treewide: kzalloc() -> kcalloc()
The kzalloc() function has a 2-factor argument form, kcalloc(). This
patch replaces cases of:
kzalloc(a * b, gfp)
with:
kcalloc(a * b, gfp)
as well as handling cases of:
kzalloc(a * b * c, gfp)
with:
kzalloc(array3_size(a, b, c), gfp)
as it's slightly less ugly than:
kzalloc_array(array_size(a, b), c, gfp)
This does, however, attempt to ignore constant size factors like:
kzalloc(4 * 1024, gfp)
though any constants defined via macros get caught up in the conversion.
Any factors with a sizeof() of "unsigned char", "char", and "u8" were
dropped, since they're redundant.
The Coccinelle script used for this was:
// Fix redundant parens around sizeof().
@@
type TYPE;
expression THING, E;
@@
(
kzalloc(
- (sizeof(TYPE)) * E
+ sizeof(TYPE) * E
, ...)
|
kzalloc(
- (sizeof(THING)) * E
+ sizeof(THING) * E
, ...)
)
// Drop single-byte sizes and redundant parens.
@@
expression COUNT;
typedef u8;
typedef __u8;
@@
(
kzalloc(
- sizeof(u8) * (COUNT)
+ COUNT
, ...)
|
kzalloc(
- sizeof(__u8) * (COUNT)
+ COUNT
, ...)
|
kzalloc(
- sizeof(char) * (COUNT)
+ COUNT
, ...)
|
kzalloc(
- sizeof(unsigned char) * (COUNT)
+ COUNT
, ...)
|
kzalloc(
- sizeof(u8) * COUNT
+ COUNT
, ...)
|
kzalloc(
- sizeof(__u8) * COUNT
+ COUNT
, ...)
|
kzalloc(
- sizeof(char) * COUNT
+ COUNT
, ...)
|
kzalloc(
- sizeof(unsigned char) * COUNT
+ COUNT
, ...)
)
// 2-factor product with sizeof(type/expression) and identifier or constant.
@@
type TYPE;
expression THING;
identifier COUNT_ID;
constant COUNT_CONST;
@@
(
- kzalloc
+ kcalloc
(
- sizeof(TYPE) * (COUNT_ID)
+ COUNT_ID, sizeof(TYPE)
, ...)
|
- kzalloc
+ kcalloc
(
- sizeof(TYPE) * COUNT_ID
+ COUNT_ID, sizeof(TYPE)
, ...)
|
- kzalloc
+ kcalloc
(
- sizeof(TYPE) * (COUNT_CONST)
+ COUNT_CONST, sizeof(TYPE)
, ...)
|
- kzalloc
+ kcalloc
(
- sizeof(TYPE) * COUNT_CONST
+ COUNT_CONST, sizeof(TYPE)
, ...)
|
- kzalloc
+ kcalloc
(
- sizeof(THING) * (COUNT_ID)
+ COUNT_ID, sizeof(THING)
, ...)
|
- kzalloc
+ kcalloc
(
- sizeof(THING) * COUNT_ID
+ COUNT_ID, sizeof(THING)
, ...)
|
- kzalloc
+ kcalloc
(
- sizeof(THING) * (COUNT_CONST)
+ COUNT_CONST, sizeof(THING)
, ...)
|
- kzalloc
+ kcalloc
(
- sizeof(THING) * COUNT_CONST
+ COUNT_CONST, sizeof(THING)
, ...)
)
// 2-factor product, only identifiers.
@@
identifier SIZE, COUNT;
@@
- kzalloc
+ kcalloc
(
- SIZE * COUNT
+ COUNT, SIZE
, ...)
// 3-factor product with 1 sizeof(type) or sizeof(expression), with
// redundant parens removed.
@@
expression THING;
identifier STRIDE, COUNT;
type TYPE;
@@
(
kzalloc(
- sizeof(TYPE) * (COUNT) * (STRIDE)
+ array3_size(COUNT, STRIDE, sizeof(TYPE))
, ...)
|
kzalloc(
- sizeof(TYPE) * (COUNT) * STRIDE
+ array3_size(COUNT, STRIDE, sizeof(TYPE))
, ...)
|
kzalloc(
- sizeof(TYPE) * COUNT * (STRIDE)
+ array3_size(COUNT, STRIDE, sizeof(TYPE))
, ...)
|
kzalloc(
- sizeof(TYPE) * COUNT * STRIDE
+ array3_size(COUNT, STRIDE, sizeof(TYPE))
, ...)
|
kzalloc(
- sizeof(THING) * (COUNT) * (STRIDE)
+ array3_size(COUNT, STRIDE, sizeof(THING))
, ...)
|
kzalloc(
- sizeof(THING) * (COUNT) * STRIDE
+ array3_size(COUNT, STRIDE, sizeof(THING))
, ...)
|
kzalloc(
- sizeof(THING) * COUNT * (STRIDE)
+ array3_size(COUNT, STRIDE, sizeof(THING))
, ...)
|
kzalloc(
- sizeof(THING) * COUNT * STRIDE
+ array3_size(COUNT, STRIDE, sizeof(THING))
, ...)
)
// 3-factor product with 2 sizeof(variable), with redundant parens removed.
@@
expression THING1, THING2;
identifier COUNT;
type TYPE1, TYPE2;
@@
(
kzalloc(
- sizeof(TYPE1) * sizeof(TYPE2) * COUNT
+ array3_size(COUNT, sizeof(TYPE1), sizeof(TYPE2))
, ...)
|
kzalloc(
- sizeof(TYPE1) * sizeof(THING2) * (COUNT)
+ array3_size(COUNT, sizeof(TYPE1), sizeof(TYPE2))
, ...)
|
kzalloc(
- sizeof(THING1) * sizeof(THING2) * COUNT
+ array3_size(COUNT, sizeof(THING1), sizeof(THING2))
, ...)
|
kzalloc(
- sizeof(THING1) * sizeof(THING2) * (COUNT)
+ array3_size(COUNT, sizeof(THING1), sizeof(THING2))
, ...)
|
kzalloc(
- sizeof(TYPE1) * sizeof(THING2) * COUNT
+ array3_size(COUNT, sizeof(TYPE1), sizeof(THING2))
, ...)
|
kzalloc(
- sizeof(TYPE1) * sizeof(THING2) * (COUNT)
+ array3_size(COUNT, sizeof(TYPE1), sizeof(THING2))
, ...)
)
// 3-factor product, only identifiers, with redundant parens removed.
@@
identifier STRIDE, SIZE, COUNT;
@@
(
kzalloc(
- (COUNT) * STRIDE * SIZE
+ array3_size(COUNT, STRIDE, SIZE)
, ...)
|
kzalloc(
- COUNT * (STRIDE) * SIZE
+ array3_size(COUNT, STRIDE, SIZE)
, ...)
|
kzalloc(
- COUNT * STRIDE * (SIZE)
+ array3_size(COUNT, STRIDE, SIZE)
, ...)
|
kzalloc(
- (COUNT) * (STRIDE) * SIZE
+ array3_size(COUNT, STRIDE, SIZE)
, ...)
|
kzalloc(
- COUNT * (STRIDE) * (SIZE)
+ array3_size(COUNT, STRIDE, SIZE)
, ...)
|
kzalloc(
- (COUNT) * STRIDE * (SIZE)
+ array3_size(COUNT, STRIDE, SIZE)
, ...)
|
kzalloc(
- (COUNT) * (STRIDE) * (SIZE)
+ array3_size(COUNT, STRIDE, SIZE)
, ...)
|
kzalloc(
- COUNT * STRIDE * SIZE
+ array3_size(COUNT, STRIDE, SIZE)
, ...)
)
// Any remaining multi-factor products, first at least 3-factor products,
// when they're not all constants...
@@
expression E1, E2, E3;
constant C1, C2, C3;
@@
(
kzalloc(C1 * C2 * C3, ...)
|
kzalloc(
- (E1) * E2 * E3
+ array3_size(E1, E2, E3)
, ...)
|
kzalloc(
- (E1) * (E2) * E3
+ array3_size(E1, E2, E3)
, ...)
|
kzalloc(
- (E1) * (E2) * (E3)
+ array3_size(E1, E2, E3)
, ...)
|
kzalloc(
- E1 * E2 * E3
+ array3_size(E1, E2, E3)
, ...)
)
// And then all remaining 2 factors products when they're not all constants,
// keeping sizeof() as the second factor argument.
@@
expression THING, E1, E2;
type TYPE;
constant C1, C2, C3;
@@
(
kzalloc(sizeof(THING) * C2, ...)
|
kzalloc(sizeof(TYPE) * C2, ...)
|
kzalloc(C1 * C2 * C3, ...)
|
kzalloc(C1 * C2, ...)
|
- kzalloc
+ kcalloc
(
- sizeof(TYPE) * (E2)
+ E2, sizeof(TYPE)
, ...)
|
- kzalloc
+ kcalloc
(
- sizeof(TYPE) * E2
+ E2, sizeof(TYPE)
, ...)
|
- kzalloc
+ kcalloc
(
- sizeof(THING) * (E2)
+ E2, sizeof(THING)
, ...)
|
- kzalloc
+ kcalloc
(
- sizeof(THING) * E2
+ E2, sizeof(THING)
, ...)
|
- kzalloc
+ kcalloc
(
- (E1) * E2
+ E1, E2
, ...)
|
- kzalloc
+ kcalloc
(
- (E1) * (E2)
+ E1, E2
, ...)
|
- kzalloc
+ kcalloc
(
- E1 * E2
+ E1, E2
, ...)
)
Signed-off-by: Kees Cook <keescook@chromium.org>
2018-06-12 21:03:40 +00:00
|
|
|
parts = kcalloc(nr_parts, sizeof(*parts), GFP_KERNEL);
|
2016-04-11 08:57:54 +00:00
|
|
|
if (WARN_ON(!parts))
|
2017-11-11 16:51:25 +00:00
|
|
|
goto out_put_node;
|
2016-04-11 08:57:54 +00:00
|
|
|
|
|
|
|
for_each_child_of_node(parts_node, child_part) {
|
|
|
|
struct partition_affinity *part;
|
|
|
|
int n;
|
|
|
|
|
|
|
|
part = &parts[part_idx];
|
|
|
|
|
|
|
|
part->partition_id = of_node_to_fwnode(child_part);
|
|
|
|
|
2018-08-28 00:56:15 +00:00
|
|
|
pr_info("GIC: PPI partition %pOFn[%d] { ",
|
|
|
|
child_part, part_idx);
|
2016-04-11 08:57:54 +00:00
|
|
|
|
|
|
|
n = of_property_count_elems_of_size(child_part, "affinity",
|
|
|
|
sizeof(u32));
|
|
|
|
WARN_ON(n <= 0);
|
|
|
|
|
|
|
|
for (i = 0; i < n; i++) {
|
|
|
|
int err, cpu;
|
|
|
|
u32 cpu_phandle;
|
|
|
|
struct device_node *cpu_node;
|
|
|
|
|
|
|
|
err = of_property_read_u32_index(child_part, "affinity",
|
|
|
|
i, &cpu_phandle);
|
|
|
|
if (WARN_ON(err))
|
|
|
|
continue;
|
|
|
|
|
|
|
|
cpu_node = of_find_node_by_phandle(cpu_phandle);
|
|
|
|
if (WARN_ON(!cpu_node))
|
|
|
|
continue;
|
|
|
|
|
2018-01-02 11:25:29 +00:00
|
|
|
cpu = of_cpu_node_to_id(cpu_node);
|
|
|
|
if (WARN_ON(cpu < 0))
|
2016-04-11 08:57:54 +00:00
|
|
|
continue;
|
|
|
|
|
2017-07-18 21:43:10 +00:00
|
|
|
pr_cont("%pOF[%d] ", cpu_node, cpu);
|
2016-04-11 08:57:54 +00:00
|
|
|
|
|
|
|
cpumask_set_cpu(cpu, &part->mask);
|
|
|
|
}
|
|
|
|
|
|
|
|
pr_cont("}\n");
|
|
|
|
part_idx++;
|
|
|
|
}
|
|
|
|
|
2019-07-18 12:05:17 +00:00
|
|
|
for (i = 0; i < gic_data.ppi_nr; i++) {
|
2016-04-11 08:57:54 +00:00
|
|
|
unsigned int irq;
|
|
|
|
struct partition_desc *desc;
|
|
|
|
struct irq_fwspec ppi_fwspec = {
|
|
|
|
.fwnode = gic_data.fwnode,
|
|
|
|
.param_count = 3,
|
|
|
|
.param = {
|
2018-03-20 13:44:09 +00:00
|
|
|
[0] = GIC_IRQ_TYPE_PARTITION,
|
2016-04-11 08:57:54 +00:00
|
|
|
[1] = i,
|
|
|
|
[2] = IRQ_TYPE_NONE,
|
|
|
|
},
|
|
|
|
};
|
|
|
|
|
|
|
|
irq = irq_create_fwspec_mapping(&ppi_fwspec);
|
|
|
|
if (WARN_ON(!irq))
|
|
|
|
continue;
|
|
|
|
desc = partition_create_desc(gic_data.fwnode, parts, nr_parts,
|
|
|
|
irq, &partition_domain_ops);
|
|
|
|
if (WARN_ON(!desc))
|
|
|
|
continue;
|
|
|
|
|
|
|
|
gic_data.ppi_descs[i] = desc;
|
|
|
|
}
|
2017-11-11 16:51:25 +00:00
|
|
|
|
|
|
|
out_put_node:
|
|
|
|
of_node_put(parts_node);
|
2016-04-11 08:57:54 +00:00
|
|
|
}
|
|
|
|
|
2016-04-11 15:32:57 +00:00
|
|
|
static void __init gic_of_setup_kvm_info(struct device_node *node)
|
|
|
|
{
|
|
|
|
int ret;
|
|
|
|
struct resource r;
|
|
|
|
u32 gicv_idx;
|
|
|
|
|
|
|
|
gic_v3_kvm_info.type = GIC_V3;
|
|
|
|
|
|
|
|
gic_v3_kvm_info.maint_irq = irq_of_parse_and_map(node, 0);
|
|
|
|
if (!gic_v3_kvm_info.maint_irq)
|
|
|
|
return;
|
|
|
|
|
|
|
|
if (of_property_read_u32(node, "#redistributor-regions",
|
|
|
|
&gicv_idx))
|
|
|
|
gicv_idx = 1;
|
|
|
|
|
|
|
|
gicv_idx += 3; /* Also skip GICD, GICC, GICH */
|
|
|
|
ret = of_address_to_resource(node, gicv_idx, &r);
|
|
|
|
if (!ret)
|
|
|
|
gic_v3_kvm_info.vcpu = r;
|
|
|
|
|
2017-06-25 13:10:46 +00:00
|
|
|
gic_v3_kvm_info.has_v4 = gic_data.rdists.has_vlpis;
|
2020-03-04 20:33:13 +00:00
|
|
|
gic_v3_kvm_info.has_v4_1 = gic_data.rdists.has_rvpeid;
|
2021-02-27 10:23:45 +00:00
|
|
|
vgic_set_kvm_info(&gic_v3_kvm_info);
|
2016-04-11 15:32:57 +00:00
|
|
|
}
|
|
|
|
|
2016-01-19 13:11:14 +00:00
|
|
|
static int __init gic_of_init(struct device_node *node, struct device_node *parent)
|
|
|
|
{
|
|
|
|
void __iomem *dist_base;
|
|
|
|
struct redist_region *rdist_regs;
|
|
|
|
u64 redist_stride;
|
|
|
|
u32 nr_redist_regions;
|
|
|
|
int err, i;
|
|
|
|
|
|
|
|
dist_base = of_iomap(node, 0);
|
|
|
|
if (!dist_base) {
|
2017-07-18 21:43:10 +00:00
|
|
|
pr_err("%pOF: unable to map gic dist registers\n", node);
|
2016-01-19 13:11:14 +00:00
|
|
|
return -ENXIO;
|
|
|
|
}
|
|
|
|
|
|
|
|
err = gic_validate_dist_version(dist_base);
|
|
|
|
if (err) {
|
2017-07-18 21:43:10 +00:00
|
|
|
pr_err("%pOF: no distributor detected, giving up\n", node);
|
2016-01-19 13:11:14 +00:00
|
|
|
goto out_unmap_dist;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (of_property_read_u32(node, "#redistributor-regions", &nr_redist_regions))
|
|
|
|
nr_redist_regions = 1;
|
|
|
|
|
treewide: kzalloc() -> kcalloc()
The kzalloc() function has a 2-factor argument form, kcalloc(). This
patch replaces cases of:
kzalloc(a * b, gfp)
with:
kcalloc(a * b, gfp)
as well as handling cases of:
kzalloc(a * b * c, gfp)
with:
kzalloc(array3_size(a, b, c), gfp)
as it's slightly less ugly than:
kzalloc_array(array_size(a, b), c, gfp)
This does, however, attempt to ignore constant size factors like:
kzalloc(4 * 1024, gfp)
though any constants defined via macros get caught up in the conversion.
Any factors with a sizeof() of "unsigned char", "char", and "u8" were
dropped, since they're redundant.
The Coccinelle script used for this was:
// Fix redundant parens around sizeof().
@@
type TYPE;
expression THING, E;
@@
(
kzalloc(
- (sizeof(TYPE)) * E
+ sizeof(TYPE) * E
, ...)
|
kzalloc(
- (sizeof(THING)) * E
+ sizeof(THING) * E
, ...)
)
// Drop single-byte sizes and redundant parens.
@@
expression COUNT;
typedef u8;
typedef __u8;
@@
(
kzalloc(
- sizeof(u8) * (COUNT)
+ COUNT
, ...)
|
kzalloc(
- sizeof(__u8) * (COUNT)
+ COUNT
, ...)
|
kzalloc(
- sizeof(char) * (COUNT)
+ COUNT
, ...)
|
kzalloc(
- sizeof(unsigned char) * (COUNT)
+ COUNT
, ...)
|
kzalloc(
- sizeof(u8) * COUNT
+ COUNT
, ...)
|
kzalloc(
- sizeof(__u8) * COUNT
+ COUNT
, ...)
|
kzalloc(
- sizeof(char) * COUNT
+ COUNT
, ...)
|
kzalloc(
- sizeof(unsigned char) * COUNT
+ COUNT
, ...)
)
// 2-factor product with sizeof(type/expression) and identifier or constant.
@@
type TYPE;
expression THING;
identifier COUNT_ID;
constant COUNT_CONST;
@@
(
- kzalloc
+ kcalloc
(
- sizeof(TYPE) * (COUNT_ID)
+ COUNT_ID, sizeof(TYPE)
, ...)
|
- kzalloc
+ kcalloc
(
- sizeof(TYPE) * COUNT_ID
+ COUNT_ID, sizeof(TYPE)
, ...)
|
- kzalloc
+ kcalloc
(
- sizeof(TYPE) * (COUNT_CONST)
+ COUNT_CONST, sizeof(TYPE)
, ...)
|
- kzalloc
+ kcalloc
(
- sizeof(TYPE) * COUNT_CONST
+ COUNT_CONST, sizeof(TYPE)
, ...)
|
- kzalloc
+ kcalloc
(
- sizeof(THING) * (COUNT_ID)
+ COUNT_ID, sizeof(THING)
, ...)
|
- kzalloc
+ kcalloc
(
- sizeof(THING) * COUNT_ID
+ COUNT_ID, sizeof(THING)
, ...)
|
- kzalloc
+ kcalloc
(
- sizeof(THING) * (COUNT_CONST)
+ COUNT_CONST, sizeof(THING)
, ...)
|
- kzalloc
+ kcalloc
(
- sizeof(THING) * COUNT_CONST
+ COUNT_CONST, sizeof(THING)
, ...)
)
// 2-factor product, only identifiers.
@@
identifier SIZE, COUNT;
@@
- kzalloc
+ kcalloc
(
- SIZE * COUNT
+ COUNT, SIZE
, ...)
// 3-factor product with 1 sizeof(type) or sizeof(expression), with
// redundant parens removed.
@@
expression THING;
identifier STRIDE, COUNT;
type TYPE;
@@
(
kzalloc(
- sizeof(TYPE) * (COUNT) * (STRIDE)
+ array3_size(COUNT, STRIDE, sizeof(TYPE))
, ...)
|
kzalloc(
- sizeof(TYPE) * (COUNT) * STRIDE
+ array3_size(COUNT, STRIDE, sizeof(TYPE))
, ...)
|
kzalloc(
- sizeof(TYPE) * COUNT * (STRIDE)
+ array3_size(COUNT, STRIDE, sizeof(TYPE))
, ...)
|
kzalloc(
- sizeof(TYPE) * COUNT * STRIDE
+ array3_size(COUNT, STRIDE, sizeof(TYPE))
, ...)
|
kzalloc(
- sizeof(THING) * (COUNT) * (STRIDE)
+ array3_size(COUNT, STRIDE, sizeof(THING))
, ...)
|
kzalloc(
- sizeof(THING) * (COUNT) * STRIDE
+ array3_size(COUNT, STRIDE, sizeof(THING))
, ...)
|
kzalloc(
- sizeof(THING) * COUNT * (STRIDE)
+ array3_size(COUNT, STRIDE, sizeof(THING))
, ...)
|
kzalloc(
- sizeof(THING) * COUNT * STRIDE
+ array3_size(COUNT, STRIDE, sizeof(THING))
, ...)
)
// 3-factor product with 2 sizeof(variable), with redundant parens removed.
@@
expression THING1, THING2;
identifier COUNT;
type TYPE1, TYPE2;
@@
(
kzalloc(
- sizeof(TYPE1) * sizeof(TYPE2) * COUNT
+ array3_size(COUNT, sizeof(TYPE1), sizeof(TYPE2))
, ...)
|
kzalloc(
- sizeof(TYPE1) * sizeof(THING2) * (COUNT)
+ array3_size(COUNT, sizeof(TYPE1), sizeof(TYPE2))
, ...)
|
kzalloc(
- sizeof(THING1) * sizeof(THING2) * COUNT
+ array3_size(COUNT, sizeof(THING1), sizeof(THING2))
, ...)
|
kzalloc(
- sizeof(THING1) * sizeof(THING2) * (COUNT)
+ array3_size(COUNT, sizeof(THING1), sizeof(THING2))
, ...)
|
kzalloc(
- sizeof(TYPE1) * sizeof(THING2) * COUNT
+ array3_size(COUNT, sizeof(TYPE1), sizeof(THING2))
, ...)
|
kzalloc(
- sizeof(TYPE1) * sizeof(THING2) * (COUNT)
+ array3_size(COUNT, sizeof(TYPE1), sizeof(THING2))
, ...)
)
// 3-factor product, only identifiers, with redundant parens removed.
@@
identifier STRIDE, SIZE, COUNT;
@@
(
kzalloc(
- (COUNT) * STRIDE * SIZE
+ array3_size(COUNT, STRIDE, SIZE)
, ...)
|
kzalloc(
- COUNT * (STRIDE) * SIZE
+ array3_size(COUNT, STRIDE, SIZE)
, ...)
|
kzalloc(
- COUNT * STRIDE * (SIZE)
+ array3_size(COUNT, STRIDE, SIZE)
, ...)
|
kzalloc(
- (COUNT) * (STRIDE) * SIZE
+ array3_size(COUNT, STRIDE, SIZE)
, ...)
|
kzalloc(
- COUNT * (STRIDE) * (SIZE)
+ array3_size(COUNT, STRIDE, SIZE)
, ...)
|
kzalloc(
- (COUNT) * STRIDE * (SIZE)
+ array3_size(COUNT, STRIDE, SIZE)
, ...)
|
kzalloc(
- (COUNT) * (STRIDE) * (SIZE)
+ array3_size(COUNT, STRIDE, SIZE)
, ...)
|
kzalloc(
- COUNT * STRIDE * SIZE
+ array3_size(COUNT, STRIDE, SIZE)
, ...)
)
// Any remaining multi-factor products, first at least 3-factor products,
// when they're not all constants...
@@
expression E1, E2, E3;
constant C1, C2, C3;
@@
(
kzalloc(C1 * C2 * C3, ...)
|
kzalloc(
- (E1) * E2 * E3
+ array3_size(E1, E2, E3)
, ...)
|
kzalloc(
- (E1) * (E2) * E3
+ array3_size(E1, E2, E3)
, ...)
|
kzalloc(
- (E1) * (E2) * (E3)
+ array3_size(E1, E2, E3)
, ...)
|
kzalloc(
- E1 * E2 * E3
+ array3_size(E1, E2, E3)
, ...)
)
// And then all remaining 2 factors products when they're not all constants,
// keeping sizeof() as the second factor argument.
@@
expression THING, E1, E2;
type TYPE;
constant C1, C2, C3;
@@
(
kzalloc(sizeof(THING) * C2, ...)
|
kzalloc(sizeof(TYPE) * C2, ...)
|
kzalloc(C1 * C2 * C3, ...)
|
kzalloc(C1 * C2, ...)
|
- kzalloc
+ kcalloc
(
- sizeof(TYPE) * (E2)
+ E2, sizeof(TYPE)
, ...)
|
- kzalloc
+ kcalloc
(
- sizeof(TYPE) * E2
+ E2, sizeof(TYPE)
, ...)
|
- kzalloc
+ kcalloc
(
- sizeof(THING) * (E2)
+ E2, sizeof(THING)
, ...)
|
- kzalloc
+ kcalloc
(
- sizeof(THING) * E2
+ E2, sizeof(THING)
, ...)
|
- kzalloc
+ kcalloc
(
- (E1) * E2
+ E1, E2
, ...)
|
- kzalloc
+ kcalloc
(
- (E1) * (E2)
+ E1, E2
, ...)
|
- kzalloc
+ kcalloc
(
- E1 * E2
+ E1, E2
, ...)
)
Signed-off-by: Kees Cook <keescook@chromium.org>
2018-06-12 21:03:40 +00:00
|
|
|
rdist_regs = kcalloc(nr_redist_regions, sizeof(*rdist_regs),
|
|
|
|
GFP_KERNEL);
|
2016-01-19 13:11:14 +00:00
|
|
|
if (!rdist_regs) {
|
|
|
|
err = -ENOMEM;
|
|
|
|
goto out_unmap_dist;
|
|
|
|
}
|
|
|
|
|
|
|
|
for (i = 0; i < nr_redist_regions; i++) {
|
|
|
|
struct resource res;
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
ret = of_address_to_resource(node, 1 + i, &res);
|
|
|
|
rdist_regs[i].redist_base = of_iomap(node, 1 + i);
|
|
|
|
if (ret || !rdist_regs[i].redist_base) {
|
2017-07-18 21:43:10 +00:00
|
|
|
pr_err("%pOF: couldn't map region %d\n", node, i);
|
2016-01-19 13:11:14 +00:00
|
|
|
err = -ENODEV;
|
|
|
|
goto out_unmap_rdist;
|
|
|
|
}
|
|
|
|
rdist_regs[i].phys_base = res.start;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (of_property_read_u64(node, "redistributor-stride", &redist_stride))
|
|
|
|
redist_stride = 0;
|
|
|
|
|
2018-12-10 13:56:31 +00:00
|
|
|
gic_enable_of_quirks(node, gic_quirks, &gic_data);
|
|
|
|
|
2016-01-19 13:11:14 +00:00
|
|
|
err = gic_init_bases(dist_base, rdist_regs, nr_redist_regions,
|
|
|
|
redist_stride, &node->fwnode);
|
2016-04-11 08:57:54 +00:00
|
|
|
if (err)
|
|
|
|
goto out_unmap_rdist;
|
|
|
|
|
|
|
|
gic_populate_ppi_partitions(node);
|
2016-12-06 21:00:52 +00:00
|
|
|
|
2018-03-26 21:09:25 +00:00
|
|
|
if (static_branch_likely(&supports_deactivate_key))
|
2016-12-06 21:00:52 +00:00
|
|
|
gic_of_setup_kvm_info(node);
|
2016-04-11 08:57:54 +00:00
|
|
|
return 0;
|
2016-01-19 13:11:14 +00:00
|
|
|
|
2014-06-30 15:01:31 +00:00
|
|
|
out_unmap_rdist:
|
2014-11-24 14:35:10 +00:00
|
|
|
for (i = 0; i < nr_redist_regions; i++)
|
|
|
|
if (rdist_regs[i].redist_base)
|
|
|
|
iounmap(rdist_regs[i].redist_base);
|
|
|
|
kfree(rdist_regs);
|
2014-06-30 15:01:31 +00:00
|
|
|
out_unmap_dist:
|
|
|
|
iounmap(dist_base);
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
|
|
|
|
IRQCHIP_DECLARE(gic_v3, "arm,gic-v3", gic_of_init);
|
2016-01-19 13:11:15 +00:00
|
|
|
|
|
|
|
#ifdef CONFIG_ACPI
|
2016-04-11 15:32:56 +00:00
|
|
|
static struct
|
|
|
|
{
|
|
|
|
void __iomem *dist_base;
|
|
|
|
struct redist_region *redist_regs;
|
|
|
|
u32 nr_redist_regions;
|
|
|
|
bool single_redist;
|
2019-12-16 11:24:57 +00:00
|
|
|
int enabled_rdists;
|
2016-04-11 15:32:57 +00:00
|
|
|
u32 maint_irq;
|
|
|
|
int maint_irq_mode;
|
|
|
|
phys_addr_t vcpu_base;
|
2016-04-11 15:32:56 +00:00
|
|
|
} acpi_data __initdata;
|
2016-01-19 13:11:16 +00:00
|
|
|
|
|
|
|
static void __init
|
|
|
|
gic_acpi_register_redist(phys_addr_t phys_base, void __iomem *redist_base)
|
|
|
|
{
|
|
|
|
static int count = 0;
|
|
|
|
|
2016-04-11 15:32:56 +00:00
|
|
|
acpi_data.redist_regs[count].phys_base = phys_base;
|
|
|
|
acpi_data.redist_regs[count].redist_base = redist_base;
|
|
|
|
acpi_data.redist_regs[count].single_redist = acpi_data.single_redist;
|
2016-01-19 13:11:16 +00:00
|
|
|
count++;
|
|
|
|
}
|
2016-01-19 13:11:15 +00:00
|
|
|
|
|
|
|
static int __init
|
2019-03-11 20:55:57 +00:00
|
|
|
gic_acpi_parse_madt_redist(union acpi_subtable_headers *header,
|
2016-01-19 13:11:15 +00:00
|
|
|
const unsigned long end)
|
|
|
|
{
|
|
|
|
struct acpi_madt_generic_redistributor *redist =
|
|
|
|
(struct acpi_madt_generic_redistributor *)header;
|
|
|
|
void __iomem *redist_base;
|
|
|
|
|
|
|
|
redist_base = ioremap(redist->base_address, redist->length);
|
|
|
|
if (!redist_base) {
|
|
|
|
pr_err("Couldn't map GICR region @%llx\n", redist->base_address);
|
|
|
|
return -ENOMEM;
|
|
|
|
}
|
|
|
|
|
2016-01-19 13:11:16 +00:00
|
|
|
gic_acpi_register_redist(redist->base_address, redist_base);
|
2016-01-19 13:11:15 +00:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2016-01-19 13:11:16 +00:00
|
|
|
static int __init
|
2019-03-11 20:55:57 +00:00
|
|
|
gic_acpi_parse_madt_gicc(union acpi_subtable_headers *header,
|
2016-01-19 13:11:16 +00:00
|
|
|
const unsigned long end)
|
|
|
|
{
|
|
|
|
struct acpi_madt_generic_interrupt *gicc =
|
|
|
|
(struct acpi_madt_generic_interrupt *)header;
|
2016-04-11 15:32:56 +00:00
|
|
|
u32 reg = readl_relaxed(acpi_data.dist_base + GICD_PIDR2) & GIC_PIDR2_ARCH_MASK;
|
2016-01-19 13:11:16 +00:00
|
|
|
u32 size = reg == GIC_PIDR2_ARCH_GICv4 ? SZ_64K * 4 : SZ_64K * 2;
|
|
|
|
void __iomem *redist_base;
|
|
|
|
|
2017-12-05 19:16:21 +00:00
|
|
|
/* GICC entry which has !ACPI_MADT_ENABLED is not unusable so skip */
|
|
|
|
if (!(gicc->flags & ACPI_MADT_ENABLED))
|
|
|
|
return 0;
|
|
|
|
|
2016-01-19 13:11:16 +00:00
|
|
|
redist_base = ioremap(gicc->gicr_base_address, size);
|
|
|
|
if (!redist_base)
|
|
|
|
return -ENOMEM;
|
|
|
|
|
|
|
|
gic_acpi_register_redist(gicc->gicr_base_address, redist_base);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int __init gic_acpi_collect_gicr_base(void)
|
|
|
|
{
|
|
|
|
acpi_tbl_entry_handler redist_parser;
|
|
|
|
enum acpi_madt_type type;
|
|
|
|
|
2016-04-11 15:32:56 +00:00
|
|
|
if (acpi_data.single_redist) {
|
2016-01-19 13:11:16 +00:00
|
|
|
type = ACPI_MADT_TYPE_GENERIC_INTERRUPT;
|
|
|
|
redist_parser = gic_acpi_parse_madt_gicc;
|
|
|
|
} else {
|
|
|
|
type = ACPI_MADT_TYPE_GENERIC_REDISTRIBUTOR;
|
|
|
|
redist_parser = gic_acpi_parse_madt_redist;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Collect redistributor base addresses in GICR entries */
|
|
|
|
if (acpi_table_parse_madt(type, redist_parser, 0) > 0)
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
pr_info("No valid GICR entries exist\n");
|
|
|
|
return -ENODEV;
|
|
|
|
}
|
|
|
|
|
2019-03-11 20:55:57 +00:00
|
|
|
static int __init gic_acpi_match_gicr(union acpi_subtable_headers *header,
|
2016-01-19 13:11:15 +00:00
|
|
|
const unsigned long end)
|
|
|
|
{
|
|
|
|
/* Subtable presence means that redist exists, that's it */
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2019-03-11 20:55:57 +00:00
|
|
|
static int __init gic_acpi_match_gicc(union acpi_subtable_headers *header,
|
2016-01-19 13:11:16 +00:00
|
|
|
const unsigned long end)
|
|
|
|
{
|
|
|
|
struct acpi_madt_generic_interrupt *gicc =
|
|
|
|
(struct acpi_madt_generic_interrupt *)header;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* If GICC is enabled and has valid gicr base address, then it means
|
|
|
|
* GICR base is presented via GICC
|
|
|
|
*/
|
2019-12-16 11:24:57 +00:00
|
|
|
if ((gicc->flags & ACPI_MADT_ENABLED) && gicc->gicr_base_address) {
|
|
|
|
acpi_data.enabled_rdists++;
|
2016-01-19 13:11:16 +00:00
|
|
|
return 0;
|
2019-12-16 11:24:57 +00:00
|
|
|
}
|
2016-01-19 13:11:16 +00:00
|
|
|
|
2017-12-05 19:16:21 +00:00
|
|
|
/*
|
|
|
|
* It's perfectly valid firmware can pass disabled GICC entry, driver
|
|
|
|
* should not treat as errors, skip the entry instead of probe fail.
|
|
|
|
*/
|
|
|
|
if (!(gicc->flags & ACPI_MADT_ENABLED))
|
|
|
|
return 0;
|
|
|
|
|
2016-01-19 13:11:16 +00:00
|
|
|
return -ENODEV;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int __init gic_acpi_count_gicr_regions(void)
|
|
|
|
{
|
|
|
|
int count;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Count how many redistributor regions we have. It is not allowed
|
|
|
|
* to mix redistributor description, GICR and GICC subtables have to be
|
|
|
|
* mutually exclusive.
|
|
|
|
*/
|
|
|
|
count = acpi_table_parse_madt(ACPI_MADT_TYPE_GENERIC_REDISTRIBUTOR,
|
|
|
|
gic_acpi_match_gicr, 0);
|
|
|
|
if (count > 0) {
|
2016-04-11 15:32:56 +00:00
|
|
|
acpi_data.single_redist = false;
|
2016-01-19 13:11:16 +00:00
|
|
|
return count;
|
|
|
|
}
|
|
|
|
|
|
|
|
count = acpi_table_parse_madt(ACPI_MADT_TYPE_GENERIC_INTERRUPT,
|
|
|
|
gic_acpi_match_gicc, 0);
|
2019-12-16 11:24:57 +00:00
|
|
|
if (count > 0) {
|
2016-04-11 15:32:56 +00:00
|
|
|
acpi_data.single_redist = true;
|
2019-12-16 11:24:57 +00:00
|
|
|
count = acpi_data.enabled_rdists;
|
|
|
|
}
|
2016-01-19 13:11:16 +00:00
|
|
|
|
|
|
|
return count;
|
|
|
|
}
|
|
|
|
|
2016-01-19 13:11:15 +00:00
|
|
|
static bool __init acpi_validate_gic_table(struct acpi_subtable_header *header,
|
|
|
|
struct acpi_probe_entry *ape)
|
|
|
|
{
|
|
|
|
struct acpi_madt_generic_distributor *dist;
|
|
|
|
int count;
|
|
|
|
|
|
|
|
dist = (struct acpi_madt_generic_distributor *)header;
|
|
|
|
if (dist->version != ape->driver_data)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
/* We need to do that exercise anyway, the sooner the better */
|
2016-01-19 13:11:16 +00:00
|
|
|
count = gic_acpi_count_gicr_regions();
|
2016-01-19 13:11:15 +00:00
|
|
|
if (count <= 0)
|
|
|
|
return false;
|
|
|
|
|
2016-04-11 15:32:56 +00:00
|
|
|
acpi_data.nr_redist_regions = count;
|
2016-01-19 13:11:15 +00:00
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2019-03-11 20:55:57 +00:00
|
|
|
static int __init gic_acpi_parse_virt_madt_gicc(union acpi_subtable_headers *header,
|
2016-04-11 15:32:57 +00:00
|
|
|
const unsigned long end)
|
|
|
|
{
|
|
|
|
struct acpi_madt_generic_interrupt *gicc =
|
|
|
|
(struct acpi_madt_generic_interrupt *)header;
|
|
|
|
int maint_irq_mode;
|
|
|
|
static int first_madt = true;
|
|
|
|
|
|
|
|
/* Skip unusable CPUs */
|
|
|
|
if (!(gicc->flags & ACPI_MADT_ENABLED))
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
maint_irq_mode = (gicc->flags & ACPI_MADT_VGIC_IRQ_MODE) ?
|
|
|
|
ACPI_EDGE_SENSITIVE : ACPI_LEVEL_SENSITIVE;
|
|
|
|
|
|
|
|
if (first_madt) {
|
|
|
|
first_madt = false;
|
|
|
|
|
|
|
|
acpi_data.maint_irq = gicc->vgic_interrupt;
|
|
|
|
acpi_data.maint_irq_mode = maint_irq_mode;
|
|
|
|
acpi_data.vcpu_base = gicc->gicv_base_address;
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* The maintenance interrupt and GICV should be the same for every CPU
|
|
|
|
*/
|
|
|
|
if ((acpi_data.maint_irq != gicc->vgic_interrupt) ||
|
|
|
|
(acpi_data.maint_irq_mode != maint_irq_mode) ||
|
|
|
|
(acpi_data.vcpu_base != gicc->gicv_base_address))
|
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static bool __init gic_acpi_collect_virt_info(void)
|
|
|
|
{
|
|
|
|
int count;
|
|
|
|
|
|
|
|
count = acpi_table_parse_madt(ACPI_MADT_TYPE_GENERIC_INTERRUPT,
|
|
|
|
gic_acpi_parse_virt_madt_gicc, 0);
|
|
|
|
|
|
|
|
return (count > 0);
|
|
|
|
}
|
|
|
|
|
2016-01-19 13:11:15 +00:00
|
|
|
#define ACPI_GICV3_DIST_MEM_SIZE (SZ_64K)
|
2016-04-11 15:32:57 +00:00
|
|
|
#define ACPI_GICV2_VCTRL_MEM_SIZE (SZ_4K)
|
|
|
|
#define ACPI_GICV2_VCPU_MEM_SIZE (SZ_8K)
|
|
|
|
|
|
|
|
static void __init gic_acpi_setup_kvm_info(void)
|
|
|
|
{
|
|
|
|
int irq;
|
|
|
|
|
|
|
|
if (!gic_acpi_collect_virt_info()) {
|
|
|
|
pr_warn("Unable to get hardware information used for virtualization\n");
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
gic_v3_kvm_info.type = GIC_V3;
|
|
|
|
|
|
|
|
irq = acpi_register_gsi(NULL, acpi_data.maint_irq,
|
|
|
|
acpi_data.maint_irq_mode,
|
|
|
|
ACPI_ACTIVE_HIGH);
|
|
|
|
if (irq <= 0)
|
|
|
|
return;
|
|
|
|
|
|
|
|
gic_v3_kvm_info.maint_irq = irq;
|
|
|
|
|
|
|
|
if (acpi_data.vcpu_base) {
|
|
|
|
struct resource *vcpu = &gic_v3_kvm_info.vcpu;
|
|
|
|
|
|
|
|
vcpu->flags = IORESOURCE_MEM;
|
|
|
|
vcpu->start = acpi_data.vcpu_base;
|
|
|
|
vcpu->end = vcpu->start + ACPI_GICV2_VCPU_MEM_SIZE - 1;
|
|
|
|
}
|
|
|
|
|
2017-06-25 13:10:46 +00:00
|
|
|
gic_v3_kvm_info.has_v4 = gic_data.rdists.has_vlpis;
|
2020-03-04 20:33:13 +00:00
|
|
|
gic_v3_kvm_info.has_v4_1 = gic_data.rdists.has_rvpeid;
|
2021-02-27 10:23:45 +00:00
|
|
|
vgic_set_kvm_info(&gic_v3_kvm_info);
|
2016-04-11 15:32:57 +00:00
|
|
|
}
|
2016-01-19 13:11:15 +00:00
|
|
|
|
|
|
|
static int __init
|
2020-05-30 14:34:29 +00:00
|
|
|
gic_acpi_init(union acpi_subtable_headers *header, const unsigned long end)
|
2016-01-19 13:11:15 +00:00
|
|
|
{
|
|
|
|
struct acpi_madt_generic_distributor *dist;
|
|
|
|
struct fwnode_handle *domain_handle;
|
2016-04-11 15:32:56 +00:00
|
|
|
size_t size;
|
2016-01-19 13:11:16 +00:00
|
|
|
int i, err;
|
2016-01-19 13:11:15 +00:00
|
|
|
|
|
|
|
/* Get distributor base address */
|
|
|
|
dist = (struct acpi_madt_generic_distributor *)header;
|
2016-04-11 15:32:56 +00:00
|
|
|
acpi_data.dist_base = ioremap(dist->base_address,
|
|
|
|
ACPI_GICV3_DIST_MEM_SIZE);
|
|
|
|
if (!acpi_data.dist_base) {
|
2016-01-19 13:11:15 +00:00
|
|
|
pr_err("Unable to map GICD registers\n");
|
|
|
|
return -ENOMEM;
|
|
|
|
}
|
|
|
|
|
2016-04-11 15:32:56 +00:00
|
|
|
err = gic_validate_dist_version(acpi_data.dist_base);
|
2016-01-19 13:11:15 +00:00
|
|
|
if (err) {
|
2017-11-13 13:53:49 +00:00
|
|
|
pr_err("No distributor detected at @%p, giving up\n",
|
2016-04-11 15:32:56 +00:00
|
|
|
acpi_data.dist_base);
|
2016-01-19 13:11:15 +00:00
|
|
|
goto out_dist_unmap;
|
|
|
|
}
|
|
|
|
|
2016-04-11 15:32:56 +00:00
|
|
|
size = sizeof(*acpi_data.redist_regs) * acpi_data.nr_redist_regions;
|
|
|
|
acpi_data.redist_regs = kzalloc(size, GFP_KERNEL);
|
|
|
|
if (!acpi_data.redist_regs) {
|
2016-01-19 13:11:15 +00:00
|
|
|
err = -ENOMEM;
|
|
|
|
goto out_dist_unmap;
|
|
|
|
}
|
|
|
|
|
2016-01-19 13:11:16 +00:00
|
|
|
err = gic_acpi_collect_gicr_base();
|
|
|
|
if (err)
|
2016-01-19 13:11:15 +00:00
|
|
|
goto out_redist_unmap;
|
|
|
|
|
2019-07-31 15:13:42 +00:00
|
|
|
domain_handle = irq_domain_alloc_fwnode(&dist->base_address);
|
2016-01-19 13:11:15 +00:00
|
|
|
if (!domain_handle) {
|
|
|
|
err = -ENOMEM;
|
|
|
|
goto out_redist_unmap;
|
|
|
|
}
|
|
|
|
|
2016-04-11 15:32:56 +00:00
|
|
|
err = gic_init_bases(acpi_data.dist_base, acpi_data.redist_regs,
|
|
|
|
acpi_data.nr_redist_regions, 0, domain_handle);
|
2016-01-19 13:11:15 +00:00
|
|
|
if (err)
|
|
|
|
goto out_fwhandle_free;
|
|
|
|
|
|
|
|
acpi_set_irq_model(ACPI_IRQ_MODEL_GIC, domain_handle);
|
2016-12-06 21:00:52 +00:00
|
|
|
|
2018-03-26 21:09:25 +00:00
|
|
|
if (static_branch_likely(&supports_deactivate_key))
|
2016-12-06 21:00:52 +00:00
|
|
|
gic_acpi_setup_kvm_info();
|
2016-04-11 15:32:57 +00:00
|
|
|
|
2016-01-19 13:11:15 +00:00
|
|
|
return 0;
|
|
|
|
|
|
|
|
out_fwhandle_free:
|
|
|
|
irq_domain_free_fwnode(domain_handle);
|
|
|
|
out_redist_unmap:
|
2016-04-11 15:32:56 +00:00
|
|
|
for (i = 0; i < acpi_data.nr_redist_regions; i++)
|
|
|
|
if (acpi_data.redist_regs[i].redist_base)
|
|
|
|
iounmap(acpi_data.redist_regs[i].redist_base);
|
|
|
|
kfree(acpi_data.redist_regs);
|
2016-01-19 13:11:15 +00:00
|
|
|
out_dist_unmap:
|
2016-04-11 15:32:56 +00:00
|
|
|
iounmap(acpi_data.dist_base);
|
2016-01-19 13:11:15 +00:00
|
|
|
return err;
|
|
|
|
}
|
|
|
|
IRQCHIP_ACPI_DECLARE(gic_v3, ACPI_MADT_TYPE_GENERIC_DISTRIBUTOR,
|
|
|
|
acpi_validate_gic_table, ACPI_MADT_GIC_VERSION_V3,
|
|
|
|
gic_acpi_init);
|
|
|
|
IRQCHIP_ACPI_DECLARE(gic_v4, ACPI_MADT_TYPE_GENERIC_DISTRIBUTOR,
|
|
|
|
acpi_validate_gic_table, ACPI_MADT_GIC_VERSION_V4,
|
|
|
|
gic_acpi_init);
|
|
|
|
IRQCHIP_ACPI_DECLARE(gic_v3_or_v4, ACPI_MADT_TYPE_GENERIC_DISTRIBUTOR,
|
|
|
|
acpi_validate_gic_table, ACPI_MADT_GIC_VERSION_NONE,
|
|
|
|
gic_acpi_init);
|
|
|
|
#endif
|