[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [Xen-devel] [PATCH v6 06/31] xen/arm: ITS: Port ITS driver to Xen
From: Vijaya Kumar K <Vijaya.Kumar@xxxxxxxxxxxxxxxxxx> The linux driver is based on 4.1 with below commit id 3ad2a5f57656a14d964b673a5a0e4ab0e583c870 Only following code from Linux ITS driver is ported and compiled - LPI initialization - ITS configuration code - Physical command queue management - ITS command building Also redistributor information is split into rdist and rdist_prop structures. The rdist_prop struct holds the redistributor common information for all re-distributor and rdist struct holds the per-cpu specific information. This per-cpu rdist is defined as global and shared with physical ITS driver. Signed-off-by: Vijaya Kumar K <Vijaya.Kumar@xxxxxxxxxxxxxxxxxx> --- v6: - made dump_cmd() static and const parameter - Include commit 591e5bec13f15feb13fc445b6c9c59954711c4ac "irqchip/gicv3-its: Fix mapping of LPIs to collections". - reodered its_alloc_lpi_tables() and its_lpi_init() v5: - dump_cmd is called from its_flush_cmd - Added its_lpi_alloc_chunks, lpi_free, its_send_inv, its_send_mapd, its_send_mapvi to this patch as these functions are ported from linux and more logical to be in this patch. For now these functions are non-static. Will be made static when used. - Used this_cpu instead of per_cpu - Moved GITS_* definitions to git-its.h v4: Major changes - Redistributor refactoring patch is merged - Fixed comments from v3 related to coding style and removing duplicate code. - Target address is stored from bits[48:16] to avoid shifting of target address while building ITS commands - Removed non-static functions - Removed usage of command builder functions - Changed its_cmd_block union to include mix of bit and unsigned variable types to define ITS command structure v3: - Only required changes from Linux ITS driver is ported - Xen coding style is followed. --- xen/arch/arm/gic-v3-its.c | 1011 +++++++++++++++++++++++++++++++++++++ xen/arch/arm/gic-v3.c | 15 +- xen/include/asm-arm/gic-its.h | 282 +++++++++++ xen/include/asm-arm/gic.h | 1 + xen/include/asm-arm/gic_v3_defs.h | 42 +- 5 files changed, 1344 insertions(+), 7 deletions(-) diff --git a/xen/arch/arm/gic-v3-its.c b/xen/arch/arm/gic-v3-its.c new file mode 100644 index 0000000..4ad8d8e --- /dev/null +++ b/xen/arch/arm/gic-v3-its.c @@ -0,0 +1,1011 @@ +/* + * Copyright (C) 2013, 2014 ARM Limited, All Rights Reserved. + * Author: Marc Zyngier <marc.zyngier@xxxxxxx> + * + * Xen changes: + * Vijaya Kumar K <Vijaya.Kumar@xxxxxxxxxxxxxxxxxx> + * Copyright (C) 2014, 2015 Cavium Inc. + * + * This program is free software; you can redistribute it and/or modify + * it under the terms of the GNU General Public License version 2 as + * published by the Free Software Foundation. + * + * This program is distributed in the hope that it will be useful, + * but WITHOUT ANY WARRANTY; without even the implied warranty of + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the + * GNU General Public License for more details. + * + * You should have received a copy of the GNU General Public License + * along with this program. If not, see <http://www.gnu.org/licenses/>. + */ + +#include <xen/config.h> +#include <xen/bitops.h> +#include <xen/init.h> +#include <xen/mm.h> +#include <xen/irq.h> +#include <xen/sched.h> +#include <xen/errno.h> +#include <xen/delay.h> +#include <xen/list.h> +#include <xen/sizes.h> +#include <xen/vmap.h> +#include <asm/p2m.h> +#include <asm/domain.h> +#include <asm/io.h> +#include <asm/device.h> +#include <asm/gic.h> +#include <asm/gic_v3_defs.h> +#include <asm/gic-its.h> +#include <xen/log2.h> + +#define its_print(lvl, fmt, ...) \ + printk(lvl "GIC-ITS:" fmt, ## __VA_ARGS__) + +#define its_err(fmt, ...) its_print(XENLOG_ERR, fmt, ## __VA_ARGS__) +/* TODO: ratelimit for Xen messages */ +#define its_err_ratelimited(fmt, ...) \ + its_print(XENLOG_G_ERR, fmt, ## __VA_ARGS__) + +#define its_dbg(fmt, ...) \ + its_print(XENLOG_DEBUG, fmt, ## __VA_ARGS__) + +#define its_info(fmt, ...) \ + its_print(XENLOG_INFO, fmt, ## __VA_ARGS__) + +#define its_warn(fmt, ...) \ + its_print(XENLOG_WARNING, fmt, ## __VA_ARGS__) + +//#define DEBUG_GIC_ITS + +#ifdef DEBUG_GIC_ITS +# define DPRINTK(fmt, args...) printk(XENLOG_DEBUG fmt, ##args) +#else +# define DPRINTK(fmt, args...) do {} while ( 0 ) +#endif + +#define ITS_FLAGS_CMDQ_NEEDS_FLUSHING (1 << 0) +#define RDIST_FLAGS_PROPBASE_NEEDS_FLUSHING (1 << 0) + +/* + * The ITS structure - contains most of the infrastructure, with the + * msi_controller, the command queue, the collections, and the list of + * devices writing to it. + */ +struct its_node { + spinlock_t lock; + struct list_head entry; + void __iomem *base; + paddr_t phys_base; + paddr_t phys_size; + its_cmd_block *cmd_base; + its_cmd_block *cmd_write; + void *tables[GITS_BASER_NR_REGS]; + u32 order[GITS_BASER_NR_REGS]; + struct its_collection *collections; + u64 flags; + u32 ite_size; + struct dt_device_node *dt_node; +}; + +#define ITS_ITT_ALIGN SZ_256 + +static LIST_HEAD(its_nodes); +static DEFINE_SPINLOCK(its_lock); +static struct rdist_prop *gic_rdists; + +#define gic_data_rdist() (this_cpu(rdist)) + +#ifdef DEBUG_GIC_ITS +static void dump_cmd(const its_cmd_block *cmd) +{ + printk(XENLOG_DEBUG + "ITS: CMD[0] = 0x%lx CMD[1] = 0x%lx CMD[2] = 0x%lx CMD[3] = 0x%lx\n", + cmd->bits[0], cmd->bits[1], cmd->bits[2], cmd->bits[3]); +} +#else +static void dump_cmd(const its_cmd_block *cmd) { } +#endif + +static struct its_collection *dev_event_to_col(struct its_device *dev, + u32 event) +{ + struct its_node *its = dev->its; + + return its->collections + dev->event_map.col_map[event]; +} + +#define ITS_CMD_QUEUE_SZ SZ_64K +#define ITS_CMD_QUEUE_NR_ENTRIES (ITS_CMD_QUEUE_SZ / sizeof(its_cmd_block)) + +static u64 its_cmd_ptr_to_offset(struct its_node *its, its_cmd_block *ptr) +{ + return (ptr - its->cmd_base) * sizeof(*ptr); +} + +static int its_queue_full(struct its_node *its) +{ + int widx; + int ridx; + + widx = its->cmd_write - its->cmd_base; + ridx = readl_relaxed(its->base + GITS_CREADR) / sizeof(its_cmd_block); + + /* This is incredibly unlikely to happen, unless the ITS locks up. */ + if ( ((widx + 1) % ITS_CMD_QUEUE_NR_ENTRIES) == ridx ) + return 1; + + return 0; +} + +static its_cmd_block *its_allocate_entry(struct its_node *its) +{ + its_cmd_block *cmd; + u32 count = 1000000; /* 1s! */ + + while ( its_queue_full(its) ) + { + count--; + if ( !count ) + { + its_err_ratelimited("ITS queue not draining\n"); + return NULL; + } + cpu_relax(); + udelay(1); + } + + cmd = its->cmd_write++; + + /* Handle queue wrapping */ + if (its->cmd_write == (its->cmd_base + ITS_CMD_QUEUE_NR_ENTRIES)) + its->cmd_write = its->cmd_base; + + return cmd; +} + +static its_cmd_block *its_post_commands(struct its_node *its) +{ + u64 wr = its_cmd_ptr_to_offset(its, its->cmd_write); + + writel_relaxed(wr, its->base + GITS_CWRITER); + + return its->cmd_write; +} + +static void its_flush_cmd(struct its_node *its, its_cmd_block *cmd) +{ + /* + * Make sure the commands written to memory are observable by + * the ITS. + */ + if ( its->flags & ITS_FLAGS_CMDQ_NEEDS_FLUSHING ) + clean_and_invalidate_dcache_va_range(cmd, sizeof(*cmd)); + else + dsb(ishst); + + dump_cmd(cmd); +} + +static void its_wait_for_range_completion(struct its_node *its, + its_cmd_block *from, + its_cmd_block *to) +{ + u64 rd_idx, from_idx, to_idx; + u32 count = 1000000; /* 1s! */ + + from_idx = its_cmd_ptr_to_offset(its, from); + to_idx = its_cmd_ptr_to_offset(its, to); + + while ( 1 ) + { + rd_idx = readl_relaxed(its->base + GITS_CREADR); + if ( rd_idx >= to_idx || rd_idx < from_idx ) + break; + + count--; + if ( !count ) + { + its_err_ratelimited("ITS queue timeout\n"); + return; + } + cpu_relax(); + udelay(1); + } +} + +static void its_send_single_command(struct its_node *its, + its_cmd_block *src, + struct its_collection *sync_col) +{ + its_cmd_block *cmd, *sync_cmd, *next_cmd; + unsigned long flags; + + BUILD_BUG_ON(sizeof(its_cmd_block) != 32); + + spin_lock_irqsave(&its->lock, flags); + + cmd = its_allocate_entry(its); + if ( !cmd ) + { /* We're soooooo screewed... */ + its_err_ratelimited("ITS can't allocate, dropping command\n"); + spin_unlock_irqrestore(&its->lock, flags); + return; + } + + memcpy(cmd, src, sizeof(its_cmd_block)); + its_flush_cmd(its, cmd); + + if ( sync_col ) + { + sync_cmd = its_allocate_entry(its); + if ( !sync_cmd ) + { + its_err_ratelimited("ITS can't SYNC, skipping\n"); + goto post; + } + sync_cmd->sync.cmd = GITS_CMD_SYNC; + sync_cmd->sync.ta = sync_col->target_address; + + its_flush_cmd(its, sync_cmd); + } + +post: + next_cmd = its_post_commands(its); + spin_unlock_irqrestore(&its->lock, flags); + + its_wait_for_range_completion(its, cmd, next_cmd); +} + +static void its_send_mapd(struct its_device *dev, int valid) +{ + its_cmd_block cmd; + unsigned long itt_addr; + u8 size; + + size = ilog2(dev->event_map.nr_lpis); + itt_addr = __pa(dev->itt_addr); + itt_addr = ROUNDUP(itt_addr, ITS_ITT_ALIGN); + + memset(&cmd, 0x0, sizeof(its_cmd_block)); + cmd.mapd.cmd = GITS_CMD_MAPD; + cmd.mapd.devid = dev->device_id; + cmd.mapd.size = size - 1; + /* + * ITT address field of MAPD command holds bit[48:8] of + * itt address. Hence shift by 8. + */ + cmd.mapd.itt = itt_addr >> 8; + cmd.mapd.valid = !!valid; + + its_send_single_command(dev->its, &cmd, NULL); +} + +static void its_send_mapc(struct its_node *its, struct its_collection *col, + int valid) +{ + its_cmd_block cmd; + + memset(&cmd, 0x0, sizeof(its_cmd_block)); + cmd.mapc.cmd = GITS_CMD_MAPC; + cmd.mapc.col = col->col_id; + cmd.mapc.ta = col->target_address; + cmd.mapc.valid = !!valid; + + its_send_single_command(its, &cmd, col); +} + +static void its_send_mapvi(struct its_device *dev, u32 phys_id, u32 event) +{ + its_cmd_block cmd; + struct its_collection *col = dev_event_to_col(dev, event); + + memset(&cmd, 0x0, sizeof(its_cmd_block)); + cmd.mapvi.cmd = GITS_CMD_MAPVI; + cmd.mapvi.devid = dev->device_id; + cmd.mapvi.event = event; + cmd.mapvi.phy_id = phys_id; + cmd.mapvi.col = col->col_id; + + its_send_single_command(dev->its, &cmd, col); +} + +static void its_send_invall(struct its_node *its, struct its_collection *col) +{ + its_cmd_block cmd; + + memset(&cmd, 0x0, sizeof(its_cmd_block)); + cmd.invall.cmd = GITS_CMD_INVALL; + cmd.invall.col = col->col_id; + + its_send_single_command(its, &cmd, NULL); +} + +static void its_send_discard(struct its_device *dev, u32 event) +{ + its_cmd_block cmd; + struct its_collection *col = dev_event_to_col(dev, event); + + memset(&cmd, 0x0, sizeof(its_cmd_block)); + cmd.discard.devid = dev->device_id; + cmd.discard.event = event; + + its_send_single_command(dev->its, &cmd, col); +} + +/* + * How we allocate LPIs: + * + * The GIC has id_bits bits for interrupt identifiers. From there, we + * must subtract 8192 which are reserved for SGIs/PPIs/SPIs. Then, as + * we allocate LPIs by chunks of 32, we can shift the whole thing by 5 + * bits to the right. + * + * This gives us (((1UL << id_bits) - 8192) >> 5) possible allocations. + */ +#define IRQS_PER_CHUNK_SHIFT 5 +#define IRQS_PER_CHUNK (1 << IRQS_PER_CHUNK_SHIFT) + +static unsigned long *lpi_bitmap; +static u32 lpi_chunks; +static DEFINE_SPINLOCK(lpi_lock); + +static int its_lpi_to_chunk(int lpi) +{ + return (lpi - 8192) >> IRQS_PER_CHUNK_SHIFT; +} + +static int its_chunk_to_lpi(int chunk) +{ + return (chunk << IRQS_PER_CHUNK_SHIFT) + 8192; +} + +static int its_lpi_init(u32 id_bits) +{ + lpi_chunks = its_lpi_to_chunk(1UL << id_bits); + + lpi_bitmap = xzalloc_bytes(BITS_TO_LONGS(lpi_chunks) * sizeof(long)); + if ( !lpi_bitmap ) + { + lpi_chunks = 0; + return -ENOMEM; + } + + its_info("ITS: Allocated %d chunks for LPIs\n", (int)lpi_chunks); + + return 0; +} + +static unsigned long *its_lpi_alloc_chunks(int nirqs, int *base) +{ + unsigned long *bitmap = NULL; + int chunk_id, nr_chunks, nr_ids, i; + + nr_chunks = DIV_ROUND_UP(nirqs, IRQS_PER_CHUNK); + + spin_lock(&lpi_lock); + + do { + chunk_id = bitmap_find_next_zero_area(lpi_bitmap, lpi_chunks, + 0, nr_chunks, 0); + if ( chunk_id < lpi_chunks ) + break; + + nr_chunks--; + } while ( nr_chunks > 0 ); + + if ( !nr_chunks ) + goto out; + + bitmap = xzalloc_bytes(BITS_TO_LONGS(nr_chunks * IRQS_PER_CHUNK) * + sizeof (long)); + if ( !bitmap ) + goto out; + + for ( i = 0; i < nr_chunks; i++ ) + set_bit(chunk_id + i, lpi_bitmap); + + *base = its_chunk_to_lpi(chunk_id); + nr_ids = nr_chunks * IRQS_PER_CHUNK; + + if ( nr_ids < nirqs ) + { + xfree(bitmap); + bitmap = NULL; + } + +out: + spin_unlock(&lpi_lock); + + return bitmap; +} + +static void its_lpi_free(struct its_device *dev) +{ + int lpi; + + spin_lock(&lpi_lock); + + for ( lpi = dev->event_map.lpi_base; + lpi < (dev->event_map.lpi_base + dev->event_map.nr_lpis); + lpi += IRQS_PER_CHUNK ) + { + int chunk = its_lpi_to_chunk(lpi); + + if (chunk > lpi_chunks) + its_err("Bad LPI chunk %d\n", chunk); + if ( test_bit(chunk, lpi_bitmap) ) + clear_bit(chunk, lpi_bitmap); + } + + spin_unlock(&lpi_lock); + + xfree(dev->event_map.lpi_map); +} + +/* + * We allocate 64kB for PROPBASE. That gives us at most 64K LPIs to + * deal with (one configuration byte per interrupt). PENDBASE has to + * be 64kB aligned (one bit per LPI, plus 8192 bits for SPI/PPI/SGI). + */ +#define LPI_PROPBASE_SZ SZ_64K +#define LPI_PENDBASE_SZ (LPI_PROPBASE_SZ / 8 + SZ_1K) + +/* + * This is how many bits of ID we need, including the useless ones. + */ +#define LPI_NRBITS ilog2(LPI_PROPBASE_SZ + SZ_8K) + +static int __init its_alloc_lpi_tables(void) +{ + paddr_t paddr; + + gic_rdists->prop_page = + alloc_xenheap_pages(get_order_from_bytes(LPI_PROPBASE_SZ), 0); + + if ( !gic_rdists->prop_page ) + { + its_err("Failed to allocate PROPBASE\n"); + return -ENOMEM; + } + + paddr = __pa(gic_rdists->prop_page); + its_info("GIC: using LPI property table @%pa\n", &paddr); + + /* Set LPI priority and Group-1, but disabled */ + memset(gic_rdists->prop_page, + GIC_PRI_IRQ | LPI_PROP_GROUP1, + LPI_PROPBASE_SZ); + + /* Make sure the GIC will observe the written configuration */ + clean_and_invalidate_dcache_va_range(gic_rdists->prop_page, + LPI_PROPBASE_SZ); + + return 0; +} + +static const char *its_base_type_string[] = { + [GITS_BASER_TYPE_DEVICE] = "Devices", + [GITS_BASER_TYPE_VCPU] = "Virtual CPUs", + [GITS_BASER_TYPE_CPU] = "Physical CPUs", + [GITS_BASER_TYPE_COLLECTION] = "Interrupt Collections", + [GITS_BASER_TYPE_RESERVED5] = "Reserved (5)", + [GITS_BASER_TYPE_RESERVED6] = "Reserved (6)", + [GITS_BASER_TYPE_RESERVED7] = "Reserved (7)", +}; + +static void its_free_tables(struct its_node *its) +{ + int i; + + for ( i = 0; i < GITS_BASER_NR_REGS; i++ ) + { + if ( its->tables[i] ) + { + free_xenheap_pages(its->tables[i], its->order[i]); + its->tables[i] = NULL; + its->order[i] = 0; + } + } +} + +static int its_alloc_tables(struct its_node *its) +{ + int err; + int i; + int psz = SZ_64K; + u64 shr = GITS_BASER_InnerShareable; + u64 cache = GITS_BASER_WaWb; + + for ( i = 0; i < GITS_BASER_NR_REGS; i++ ) + { + u64 val = readq_relaxed(its->base + GITS_BASER + i * 8); + u64 type = GITS_BASER_TYPE(val); + u64 entry_size = GITS_BASER_ENTRY_SIZE(val); + unsigned int order = get_order_from_bytes(psz); + int alloc_size; + u64 tmp; + void *base; + + if ( type == GITS_BASER_TYPE_NONE ) + continue; + + /* + * Allocate as many entries as required to fit the + * range of device IDs that the ITS can grok... The ID + * space being incredibly sparse, this results in a + * massive waste of memory. + * + * For other tables, only allocate a single page. + */ + if ( type == GITS_BASER_TYPE_DEVICE ) + { + u64 typer = readq_relaxed(its->base + GITS_TYPER); + u32 ids = GITS_TYPER_DEVBITS(typer); + + order = max(get_order_from_bytes((1UL << ids) * entry_size), order); + if (order >= MAX_ORDER) + { + order = MAX_ORDER - 1; + its_warn("Device Table too large,reduce its page order to %u\n", + order); + } + } + + alloc_size = (1 << order) * PAGE_SIZE; + base = alloc_xenheap_pages(order, 0); + if ( !base ) + { + err = -ENOMEM; + goto out_free; + } + memset(base, 0, alloc_size); + its->tables[i] = base; + its->order[i] = order; + +retry_baser: + val = (__pa(base) | + (type << GITS_BASER_TYPE_SHIFT) | + ((entry_size - 1) << GITS_BASER_ENTRY_SIZE_SHIFT) | + cache | + shr | + GITS_BASER_VALID); + + switch (psz) { + case SZ_4K: + val |= GITS_BASER_PAGE_SIZE_4K; + break; + case SZ_16K: + val |= GITS_BASER_PAGE_SIZE_16K; + break; + case SZ_64K: + val |= GITS_BASER_PAGE_SIZE_64K; + break; + } + + val |= (alloc_size / psz) - 1; + + writeq_relaxed(val, its->base + GITS_BASER + i * 8); + tmp = readq_relaxed(its->base + GITS_BASER + i * 8); + + if ( (val ^ tmp) & GITS_BASER_SHAREABILITY_MASK ) + { + /* + * Shareability didn't stick. Just use + * whatever the read reported, which is likely + * to be the only thing this redistributor + * supports. + */ + shr = tmp & GITS_BASER_SHAREABILITY_MASK; + if ( !shr ) + cache = GITS_BASER_nC; + goto retry_baser; + } + + if ( (val ^ tmp) & GITS_BASER_PAGE_SIZE_MASK ) + { + /* + * Page size didn't stick. Let's try a smaller + * size and retry. If we reach 4K, then + * something is horribly wrong... + */ + switch (psz) { + case SZ_16K: + psz = SZ_4K; + goto retry_baser; + case SZ_64K: + psz = SZ_16K; + goto retry_baser; + } + } + + if ( val != tmp ) + { + its_err("ITS: GITS_BASER%d doesn't stick: %lx %lx\n", + i, (unsigned long) val, (unsigned long) tmp); + err = -ENXIO; + goto out_free; + } + + its_info("ITS: allocated %d %s @%lx (psz %dK, shr %d)\n", + (int)(alloc_size / entry_size), + its_base_type_string[type], + (unsigned long)__pa(base), + psz / SZ_1K, (int)shr >> GITS_BASER_SHAREABILITY_SHIFT); + } + + return 0; + +out_free: + its_free_tables(its); + + return err; +} + +static int its_alloc_collections(struct its_node *its) +{ + its->collections = xzalloc_array(struct its_collection, nr_cpu_ids); + if ( !its->collections ) + return -ENOMEM; + + return 0; +} + +static void its_cpu_init_lpis(void) +{ + void __iomem *rbase = gic_data_rdist().rbase; + void *pend_page; + u64 val, tmp; + + /* If we didn't allocate the pending table yet, do it now */ + pend_page = gic_data_rdist().pend_page; + if ( !pend_page ) + { + paddr_t paddr; + u32 order; + + /* + * The pending pages have to be at least 64kB aligned, + * hence the 'max(LPI_PENDBASE_SZ, SZ_64K)' below. + */ + order = get_order_from_bytes(max(LPI_PENDBASE_SZ, SZ_64K)); + pend_page = alloc_xenheap_pages(order, 0); + if ( !pend_page ) + { + its_err("Failed to allocate PENDBASE for CPU%d with order %d\n", + smp_processor_id(), order); + return; + } + + memset(pend_page, 0, max(LPI_PENDBASE_SZ, SZ_64K)); + /* Make sure the GIC will observe the zero-ed page */ + clean_and_invalidate_dcache_va_range(pend_page, LPI_PENDBASE_SZ); + + paddr = __pa(pend_page); + + its_info("CPU%d: using LPI pending table @%pa\n", + smp_processor_id(), &paddr); + + gic_data_rdist().pend_page = pend_page; + } + + /* Disable LPIs */ + val = readl_relaxed(rbase + GICR_CTLR); + val &= ~GICR_CTLR_ENABLE_LPIS; + writel_relaxed(val, rbase + GICR_CTLR); + + /* + * Make sure any change to the table is observable by the GIC. + */ + dsb(sy); + + /* set PROPBASE */ + val = (__pa(gic_rdists->prop_page) | + GICR_PROPBASER_InnerShareable | + GICR_PROPBASER_WaWb | + ((LPI_NRBITS - 1) & GICR_PROPBASER_IDBITS_MASK)); + + writeq_relaxed(val, rbase + GICR_PROPBASER); + tmp = readq_relaxed(rbase + GICR_PROPBASER); + + if ( (tmp ^ val) & GICR_PROPBASER_SHAREABILITY_MASK ) + { + if ( !(tmp & GICR_PROPBASER_SHAREABILITY_MASK) ) + { + /* + * The HW reports non-shareable, we must + * remove the cacheability attributes as well. + */ + val &= ~(GICR_PROPBASER_SHAREABILITY_MASK | + GICR_PROPBASER_CACHEABILITY_MASK); + val |= GICR_PROPBASER_nC; + writeq_relaxed(val, rbase + GICR_PROPBASER); + } + + its_info("GIC: using cache flushing for LPI property table\n"); + gic_rdists->flags |= RDIST_FLAGS_PROPBASE_NEEDS_FLUSHING; + } + + /* set PENDBASE */ + val = (__pa(pend_page) | + GICR_PROPBASER_InnerShareable | + GICR_PROPBASER_WaWb); + + writeq_relaxed(val, rbase + GICR_PENDBASER); + tmp = readq_relaxed(rbase + GICR_PENDBASER); + + if ( !(tmp & GICR_PENDBASER_SHAREABILITY_MASK) ) + { + /* + * The HW reports non-shareable, we must remove the + * cacheability attributes as well. + */ + val &= ~(GICR_PENDBASER_SHAREABILITY_MASK | + GICR_PENDBASER_CACHEABILITY_MASK); + val |= GICR_PENDBASER_nC; + writeq_relaxed(val, rbase + GICR_PENDBASER); + } + + /* Enable LPIs */ + val = readl_relaxed(rbase + GICR_CTLR); + val |= GICR_CTLR_ENABLE_LPIS; + writel_relaxed(val, rbase + GICR_CTLR); + + /* Make sure the GIC has seen the above */ + dsb(sy); +} + +static void its_cpu_init_collection(void) +{ + struct its_node *its; + int cpu; + + spin_lock(&its_lock); + cpu = smp_processor_id(); + + list_for_each_entry(its, &its_nodes, entry) + { + u64 target; + + /* + * We now have to bind each collection to its target + * redistributor. + */ + if ( readq_relaxed(its->base + GITS_TYPER) & GITS_TYPER_PTA ) + { + target = gic_data_rdist().phys_base; + /* ITS command considers only [48:16] of the GICR address */ + target >>= 16; + } + else + { + /* + * This ITS wants a linear CPU number. + */ + target = readq_relaxed(gic_data_rdist().rbase + GICR_TYPER); + target = GICR_TYPER_CPU_NUMBER(target); + } + + /* Perform collection mapping */ + its->collections[cpu].col_id = cpu; + its->collections[cpu].target_address = target; + + its_send_mapc(its, &its->collections[cpu], 1); + its_send_invall(its, &its->collections[cpu]); + } + + spin_unlock(&its_lock); +} + +static int its_force_quiescent(void __iomem *base) +{ + u32 count = 1000000; /* 1s */ + u32 val; + + val = readl_relaxed(base + GITS_CTLR); + if ( val & GITS_CTLR_QUIESCENT ) + return 0; + + /* Disable the generation of all interrupts to this ITS */ + val &= ~GITS_CTLR_ENABLE; + writel_relaxed(val, base + GITS_CTLR); + + /* Poll GITS_CTLR and wait until ITS becomes quiescent */ + while ( 1 ) + { + val = readl_relaxed(base + GITS_CTLR); + if ( val & GITS_CTLR_QUIESCENT ) + return 0; + + count--; + if ( !count ) + return -EBUSY; + + cpu_relax(); + udelay(1); + } +} + +static int its_probe(struct dt_device_node *node) +{ + paddr_t its_addr, its_size; + struct its_node *its; + void __iomem *its_base; + u32 val, typer; + u64 baser, tmp; + int err; + + if ( !dt_get_property(node, "msi-controller", NULL) ) + { + its_warn("%s: not a msi-controller\n", node->full_name); + return -ENXIO; + } + + err = dt_device_get_address(node, 0, &its_addr, &its_size); + if ( err ) + { + its_warn("%s: no regs?\n", node->full_name); + return -ENXIO; + } + + its_base = ioremap_nocache(its_addr, its_size); + if ( !its_base ) + { + its_warn("%s: unable to map registers\n", node->full_name); + return -ENOMEM; + } + + val = readl_relaxed(its_base + GITS_PIDR2) & GIC_PIDR2_ARCH_REV_MASK; + if ( val != 0x30 && val != 0x40 ) + { + its_warn("%s: no ITS detected, giving up\n", node->full_name); + err = -ENODEV; + goto out_unmap; + } + + err = its_force_quiescent(its_base); + if ( err ) + { + its_warn("%s: failed to quiesce, giving up\n", + node->full_name); + goto out_unmap; + } + + its_info("ITS: %s\n", node->full_name); + + its = xzalloc(struct its_node); + if ( !its ) + { + err = -ENOMEM; + goto out_unmap; + } + + spin_lock_init(&its->lock); + INIT_LIST_HEAD(&its->entry); + its->dt_node = node; + its->base = its_base; + its->phys_base = its_addr; + its->phys_size = its_size; + typer = readl_relaxed(its_base + GITS_TYPER); + its->ite_size = ((typer >> 4) & 0xf) + 1; + + its->cmd_base = xzalloc_bytes(ITS_CMD_QUEUE_SZ); + if ( !its->cmd_base ) + { + err = -ENOMEM; + goto out_free_its; + } + its->cmd_write = its->cmd_base; + + err = its_alloc_tables(its); + if ( err ) + goto out_free_cmd; + + err = its_alloc_collections(its); + if ( err ) + goto out_free_tables; + + baser = (__pa(its->cmd_base) | + GITS_CBASER_WaWb | + GITS_CBASER_InnerShareable | + (ITS_CMD_QUEUE_SZ / SZ_4K - 1) | + GITS_CBASER_VALID); + + writeq_relaxed(baser, its->base + GITS_CBASER); + tmp = readq_relaxed(its->base + GITS_CBASER); + if ( (tmp ^ baser) & GITS_CBASER_SHAREABILITY_MASK ) + { + if (!(tmp & GITS_CBASER_SHAREABILITY_MASK)) + { + /* + * The HW reports non-shareable, we must + * remove the cacheability attributes as + * well. + */ + baser &= ~(GITS_CBASER_SHAREABILITY_MASK | + GITS_CBASER_CACHEABILITY_MASK); + baser |= GITS_CBASER_nC; + writeq_relaxed(baser, its->base + GITS_CBASER); + } + + its_info("ITS: using cache flushing for cmd queue\n"); + its->flags |= ITS_FLAGS_CMDQ_NEEDS_FLUSHING; + } + + writeq_relaxed(0, its->base + GITS_CWRITER); + writel_relaxed(GITS_CTLR_ENABLE, its->base + GITS_CTLR); + + spin_lock(&its_lock); + list_add(&its->entry, &its_nodes); + spin_unlock(&its_lock); + + return 0; + +out_free_tables: + its_free_tables(its); +out_free_cmd: + xfree(its->cmd_base); +out_free_its: + xfree(its); +out_unmap: + iounmap(its_base); + its_err("ITS: failed probing %s (%d)\n", node->full_name, err); + return err; +} + +static bool gic_rdists_supports_plpis(void) +{ + return !!(readl_relaxed(gic_data_rdist().rbase + GICR_TYPER) & GICR_TYPER_PLPIS); +} + +int its_cpu_init(void) +{ + if ( !list_empty(&its_nodes) ) + { + if ( !gic_rdists_supports_plpis() ) + { + its_info("CPU%d: LPIs not supported\n", smp_processor_id()); + return -ENXIO; + } + its_cpu_init_lpis(); + its_cpu_init_collection(); + } + + return 0; +} + +int __init its_init(struct rdist_prop *rdists) +{ + struct dt_device_node *np = NULL; + + static const struct dt_device_match its_device_ids[] __initconst = + { + DT_MATCH_GIC_ITS, + { /* sentinel */ }, + }; + + for (np = dt_find_matching_node(NULL, its_device_ids); np; + np = dt_find_matching_node(np, its_device_ids)) + its_probe(np); + + if ( list_empty(&its_nodes) ) + { + its_warn("ITS: No ITS available, not enabling LPIs\n"); + return -ENXIO; + } + + gic_rdists = rdists; + its_alloc_lpi_tables(); + its_lpi_init(rdists->id_bits); + + return 0; +} + +/* + * Local variables: + * mode: C + * c-file-style: "BSD" + * c-basic-offset: 4 + * indent-tabs-mode: nil + * End: + */ diff --git a/xen/arch/arm/gic-v3.c b/xen/arch/arm/gic-v3.c index d1db1ce..5076706 100644 --- a/xen/arch/arm/gic-v3.c +++ b/xen/arch/arm/gic-v3.c @@ -45,6 +45,7 @@ /* Global state */ static struct { void __iomem *map_dbase; /* Mapped address of distributor registers */ + struct rdist_prop rdist_data; struct rdist_region *rdist_regions; uint32_t rdist_stride; unsigned int rdist_count; /* Number of rdist regions count */ @@ -55,10 +56,10 @@ static struct { static struct gic_info gicv3_info; /* per-cpu re-distributor base */ -static DEFINE_PER_CPU(void __iomem*, rbase); +DEFINE_PER_CPU(struct rdist, rdist); #define GICD (gicv3.map_dbase) -#define GICD_RDIST_BASE (this_cpu(rbase)) +#define GICD_RDIST_BASE (this_cpu(rdist).rbase) #define GICD_RDIST_SGI_BASE (GICD_RDIST_BASE + SZ_64K) /* @@ -618,6 +619,7 @@ static int __init gicv3_populate_rdist(void) uint32_t aff; uint32_t reg; uint64_t typer; + uint64_t offset; uint64_t mpidr = cpu_logical_map(smp_processor_id()); /* @@ -653,9 +655,12 @@ static int __init gicv3_populate_rdist(void) if ( (typer >> 32) == aff ) { - this_cpu(rbase) = ptr; - printk("GICv3: CPU%d: Found redistributor in region %d @%p\n", - smp_processor_id(), i, ptr); + offset = ptr - gicv3.rdist_regions[i].map_base; + this_cpu(rdist).rbase = ptr; + this_cpu(rdist).phys_base = gicv3.rdist_regions[i].base + offset; + printk("GICv3: CPU%d: Found redistributor in region %d @%"PRIpaddr"\n", + smp_processor_id(), i, + this_cpu(rdist).phys_base); return 0; } if ( gicv3.rdist_stride ) diff --git a/xen/include/asm-arm/gic-its.h b/xen/include/asm-arm/gic-its.h new file mode 100644 index 0000000..259b0fe --- /dev/null +++ b/xen/include/asm-arm/gic-its.h @@ -0,0 +1,282 @@ +/* + * Copyright (C) 2015 Cavium Inc. + * Vijaya Kumar K <Vijaya.Kumar@xxxxxxxxxxxxxxxxxx> + * + * This program is free software; you can redistribute it and/or modify + * it under the terms of the GNU General Public License version 2 as + * published by the Free Software Foundation. + * + * This program is distributed in the hope that it will be useful, + * but WITHOUT ANY WARRANTY; without even the implied warranty of + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the + * GNU General Public License for more details. + * + * You should have received a copy of the GNU General Public License + * along with this program. If not, see <http://www.gnu.org/licenses/>. + */ + +#ifndef __ASM_ARM_GIC_ITS_H__ +#define __ASM_ARM_GIC_ITS_H__ + +#include <asm/gic_v3_defs.h> + +/* + * ITS registers, offsets from ITS_base + */ +#define GITS_CTLR 0x0000 +#define GITS_IIDR 0x0004 +#define GITS_TYPER 0x0008 +#define GITS_CBASER 0x0080 +#define GITS_CWRITER 0x0088 +#define GITS_CREADR 0x0090 +#define GITS_BASER0 0x0100 +#define GITS_BASER1 0x0108 +#define GITS_BASER 0x0100 +#define GITS_BASERN 0x013c +#define GITS_PIDR0 GICR_PIDR0 +#define GITS_PIDR1 GICR_PIDR1 +#define GITS_PIDR2 GICR_PIDR2 +#define GITS_PIDR3 GICR_PIDR3 +#define GITS_PIDR4 GICR_PIDR4 +#define GITS_PIDR5 GICR_PIDR5 +#define GITS_PIDR7 GICR_PIDR7 + +#define GITS_TRANSLATER 0x10040 +#define GITS_CTLR_QUIESCENT (1U << 31) +#define GITS_CTLR_ENABLE (1U << 0) + +#define GITS_TYPER_DEVBITS_SHIFT 13 +#define GITS_TYPER_DEVBITS(r) ((((r) >> GITS_TYPER_DEVBITS_SHIFT) & 0x1f) + 1) +#define GITS_TYPER_IDBITS_SHIFT 8 +#define GITS_TYPER_IDBITS(r) ((((r) >> GITS_TYPER_IDBITS_SHIFT) & 0x1f) + 1) +#define GITS_TYPER_PTA (1UL << 19) +#define GITS_TYPER_HCC_SHIFT (24) + +#define GITS_CBASER_VALID (1UL << 63) +#define GITS_CBASER_nC (1UL << 59) +#define GITS_CBASER_WaWb (5UL << 59) +#define GITS_CBASER_InnerShareable (1UL << 10) +#define GITS_CBASER_SHAREABILITY_MASK (3UL << 10) +#define GITS_CBASER_CACHEABILITY_MASK (7UL << 59) + +#define GITS_BASER_NR_REGS 8 + +#define GITS_BASER_VALID (1UL << 63) +#define GITS_BASER_nC (1UL << 59) +#define GITS_BASER_WaWb (5UL << 59) +#define GITS_BASER_TYPE_SHIFT (56) +#define GITS_BASER_TYPE_MASK (0x7) +#define GITS_BASER_TYPE(r) (((r) >> GITS_BASER_TYPE_SHIFT) & 7) +#define GITS_BASER_ENTRY_SIZE_SHIFT (48) +#define GITS_BASER_ENTRY_SIZE(r) ((((r) >> GITS_BASER_ENTRY_SIZE_SHIFT) & 0xff) + 1) +#define GITS_BASER_InnerShareable (1UL << 10) +#define GITS_BASER_SHAREABILITY_SHIFT (10) +#define GITS_BASER_SHAREABILITY_MASK (3UL << GITS_BASER_SHAREABILITY_SHIFT) +#define GITS_BASER_PAGE_SIZE_SHIFT (8) +#define GITS_BASER_PAGE_SIZE_4K (0UL << GITS_BASER_PAGE_SIZE_SHIFT) +#define GITS_BASER_PAGE_SIZE_16K (1UL << GITS_BASER_PAGE_SIZE_SHIFT) +#define GITS_BASER_PAGE_SIZE_64K (2UL << GITS_BASER_PAGE_SIZE_SHIFT) +#define GITS_BASER_PAGE_SIZE_MASK (3UL << GITS_BASER_PAGE_SIZE_SHIFT) +#define GITS_BASER_TYPE_NONE 0 +#define GITS_BASER_TYPE_DEVICE 1 +#define GITS_BASER_TYPE_VCPU 2 +#define GITS_BASER_TYPE_CPU 3 +#define GITS_BASER_TYPE_COLLECTION 4 +#define GITS_BASER_TYPE_RESERVED5 5 +#define GITS_BASER_TYPE_RESERVED6 6 +#define GITS_BASER_TYPE_RESERVED7 7 + +/* + * ITS commands + */ +#define GITS_CMD_MAPD 0x08 +#define GITS_CMD_MAPC 0x09 +#define GITS_CMD_MAPVI 0x0a +#define GITS_CMD_MAPI 0x0b +#define GITS_CMD_MOVI 0x01 +#define GITS_CMD_DISCARD 0x0f +#define GITS_CMD_INV 0x0c +#define GITS_CMD_MOVALL 0x0e +#define GITS_CMD_INVALL 0x0d +#define GITS_CMD_INT 0x03 +#define GITS_CMD_CLEAR 0x04 +#define GITS_CMD_SYNC 0x05 + +#define LPI_PROP_ENABLED (1 << 0) +#define LPI_PROP_GROUP1 (1 << 1) + +/* + * Collection structure - just an ID, and a redistributor address to + * ping. We use one per CPU as collection of interrupts assigned to this + * CPU. + */ +struct its_collection { + u64 target_address; + u16 col_id; +}; + +/* ITS command structure */ +typedef union { + u64 bits[4]; + struct __packed { + uint8_t cmd; + uint8_t pad[7]; + } hdr; + struct __packed { + u8 cmd; + u8 res1[3]; + u32 devid; + u64 size:5; + u64 res2:59; + /* XXX: Though itt is 40 bit. Keep it 48 to avoid shift */ + u64 res3:8; + u64 itt:40; + u64 res4:15; + u64 valid:1; + u64 res5; + } mapd; + struct __packed { + u8 cmd; + u8 res1[7]; + u64 res2; + u16 col; + u32 ta; + u16 res3:15; + u16 valid:1; + u64 res4; + } mapc; + struct __packed { + u8 cmd; + u8 res1[3]; + u32 devid; + u32 event; + u32 res2; + u16 col; + u8 res3[6]; + u64 res4; + } mapi; + struct __packed { + u8 cmd; + u8 res1[3]; + u32 devid; + u32 event; + u32 phy_id; + u16 col; + u8 res2[6]; + u64 res3; + } mapvi; + struct __packed { + u8 cmd; + u8 res1[3]; + u32 devid; + u32 event; + u32 res2; + u16 col; + u8 res3[6]; + u64 res4; + } movi; + struct __packed { + u8 cmd; + u8 res1[3]; + u32 devid; + u32 event; + u32 res2; + u64 res3; + u64 res4; + } discard; + struct __packed { + u8 cmd; + u8 res1[3]; + u32 devid; + u32 event; + u32 res2; + u64 res3; + u64 res4; + } inv; + struct __packed { + u8 cmd; + u8 res1[7]; + u64 res2; + u16 res3; + u32 ta1; + u16 res4; + u16 res5; + u32 ta2; + u16 res6; + } movall; + struct __packed { + u8 cmd; + u8 res1[7]; + u64 res2; + u16 col; + u8 res3[6]; + u64 res4; + } invall; + struct __packed { + u8 cmd; + u8 res1[3]; + u32 devid; + u32 event; + u32 res2; + u64 res3; + u64 res4; + } int_cmd; + struct __packed { + u8 cmd; + u8 res1[3]; + u32 devid; + u32 event; + u32 res2; + u64 res3; + u64 res4; + } clear; + struct __packed { + u8 cmd; + u8 res1[7]; + u64 res2; + u16 res3; + u32 ta; + u16 res4; + u64 res5; + } sync; +} its_cmd_block; + +struct event_lpi_map { + /* Physical LPI map */ + unsigned long *lpi_map; + /* Collection map */ + u16 *col_map; + /* First Physical LPI number assigned */ + u32 lpi_base; + /* Number of ITES/Physical LPIs assigned */ + u32 nr_lpis; +}; + +/* + * The ITS view of a device. + */ +struct its_device { + /* Physical ITS */ + struct its_node *its; + /* Device ITT address */ + paddr_t *itt_addr; + /* Device ITT size */ + unsigned long itt_size; + /* LPI and event mapping */ + struct event_lpi_map event_map; + /* Physical Device id */ + u32 device_id; +}; + +int its_init(struct rdist_prop *rdists); +int its_cpu_init(void); + +#endif /* __ASM_ARM_GIC_ITS_H__ */ +/* + * Local variables: + * mode: C + * c-file-style: "BSD" + * c-basic-offset: 4 + * indent-tabs-mode: nil + * End: + */ diff --git a/xen/include/asm-arm/gic.h b/xen/include/asm-arm/gic.h index d343abf..e330fe3 100644 --- a/xen/include/asm-arm/gic.h +++ b/xen/include/asm-arm/gic.h @@ -161,6 +161,7 @@ DT_MATCH_COMPATIBLE("arm,gic-400") #define DT_MATCH_GIC_V3 DT_MATCH_COMPATIBLE("arm,gic-v3") +#define DT_MATCH_GIC_ITS DT_MATCH_COMPATIBLE("arm,gic-v3-its") #ifdef HAS_GICV3 /* diff --git a/xen/include/asm-arm/gic_v3_defs.h b/xen/include/asm-arm/gic_v3_defs.h index bf7b239..2c322da 100644 --- a/xen/include/asm-arm/gic_v3_defs.h +++ b/xen/include/asm-arm/gic_v3_defs.h @@ -48,6 +48,7 @@ /* Additional bits in GICD_TYPER defined by GICv3 */ #define GICD_TYPE_ID_BITS_SHIFT 19 +#define GICD_TYPER_LPIS_SUPPORTED (1U << 17) #define GICD_CTLR_RWP (1UL << 31) #define GICD_CTLR_ARE_NS (1U << 4) #define GICD_CTLR_ENABLE_G1A (1U << 1) @@ -59,11 +60,12 @@ #define GICR_WAKER_ProcessorSleep (1U << 1) #define GICR_WAKER_ChildrenAsleep (1U << 2) -#define GICD_PIDR2_ARCH_REV_MASK (0xf0) +#define GIC_PIDR2_ARCH_REV_MASK (0xf0) +#define GICD_PIDR2_ARCH_REV_MASK GIC_PIDR2_ARCH_REV_MASK #define GICD_PIDR2_ARCH_REV_SHIFT (0x4) #define GICD_PIDR2_ARCH_GICV3 (0x3) -#define GICR_PIDR2_ARCH_REV_MASK GICD_PIDR2_ARCH_REV_MASK +#define GICR_PIDR2_ARCH_REV_MASK GIC_PIDR2_ARCH_REV_MASK #define GICR_PIDR2_ARCH_REV_SHIFT GICD_PIDR2_ARCH_REV_SHIFT #define GICR_PIDR2_ARCH_GICV3 GICD_PIDR2_ARCH_GICV3 @@ -113,10 +115,24 @@ #define GICR_ICFGR1 (0x0C04) #define GICR_NSACR (0x0E00) +#define GICR_CTLR_ENABLE_LPIS (1UL << 0) +#define GICR_TYPER_CPU_NUMBER(r) (((r) >> 8) & 0xffff) + +#define GICR_PROPBASER_InnerShareable (1U << 10) +#define GICR_PROPBASER_SHAREABILITY_MASK (3UL << 10) +#define GICR_PROPBASER_nC (1U << 7) +#define GICR_PROPBASER_WaWb (5U << 7) +#define GICR_PROPBASER_CACHEABILITY_MASK (7U << 7) +#define GICR_PROPBASER_IDBITS_MASK (0x1f) #define GICR_TYPER_PLPIS (1U << 0) #define GICR_TYPER_VLPIS (1U << 1) #define GICR_TYPER_LAST (1U << 4) +#define GICR_PENDBASER_InnerShareable (1U << 10) +#define GICR_PENDBASER_SHAREABILITY_MASK (3UL << 10) +#define GICR_PENDBASER_nC (1U << 7) +#define GICR_PENDBASER_CACHEABILITY_MASK (7U << 7) + #define DEFAULT_PMR_VALUE 0xff #define GICH_VMCR_EOI (1 << 9) @@ -161,6 +177,28 @@ struct rdist_region { void __iomem *map_base; }; +/* Re-distributor per-cpu information */ +struct rdist { + /* CPU Re-distributor address */ + void __iomem *rbase; + /* CPU LPI pending table */ + void *pend_page; + /* CPU Re-distributor physical address */ + paddr_t phys_base; +}; + +/* Common Re-distributor information */ +struct rdist_prop { + /* CPUs LPI configuration table */ + void *prop_page; + /* Number of ID bits */ + int id_bits; + /* Flags to store rdist attributes */ + uint64_t flags; +}; + +DECLARE_PER_CPU(struct rdist, rdist); + #endif /* __ASM_ARM_GIC_V3_DEFS_H__ */ /* -- 1.7.9.5 _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxx http://lists.xen.org/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |