[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [Xen-devel] [PATCH 24/28] ARM: vITS: handle INV command
The INV command instructs the ITS to update the configuration data for a given LPI by re-reading its entry from the property table. We don't need to care so much about the priority value, but enabling or disabling an LPI has some effect: We remove or push virtual LPIs to their VCPUs, also check the virtual pending bit if an LPI gets enabled. Signed-off-by: Andre Przywara <andre.przywara@xxxxxxx> --- xen/arch/arm/vgic-v3-its.c | 57 ++++++++++++++++++++++++++++++++++++++++++++++ 1 file changed, 57 insertions(+) diff --git a/xen/arch/arm/vgic-v3-its.c b/xen/arch/arm/vgic-v3-its.c index 8747890..82f7bcc 100644 --- a/xen/arch/arm/vgic-v3-its.c +++ b/xen/arch/arm/vgic-v3-its.c @@ -262,6 +262,60 @@ static int its_handle_int(struct virt_its *its, uint64_t *cmdptr) return 0; } +/* + * For a given virtual LPI read the enabled bit from the virtual property + * table and update the virtual IRQ's state. + * This takes care of removing or pushing of virtual LPIs to their VCPUs. + */ +static void update_lpi_enabled_status(struct virt_its* its, + struct vcpu *vcpu, uint32_t vlpi) +{ + struct pending_irq *pirq = lpi_to_pending(vcpu, vlpi, false); + uint8_t property = its->d->arch.vgic.proptable[vlpi - LPI_OFFSET]; + + if ( property & LPI_PROP_ENABLED ) + { + if ( pirq ) + { + unsigned long flags; + + set_bit(GIC_IRQ_GUEST_ENABLED, &pirq->status); + spin_lock_irqsave(&vcpu->arch.vgic.lock, flags); + if ( !list_empty(&pirq->inflight) && + !test_bit(GIC_IRQ_GUEST_VISIBLE, &pirq->status) ) + gic_raise_guest_irq(vcpu, vlpi, property & LPI_PROP_PRIO_MASK); + spin_unlock_irqrestore(&vcpu->arch.vgic.lock, flags); + } + + /* Check whether the LPI has fired while the guest had it disabled. */ + if ( test_and_clear_bit(vlpi - LPI_OFFSET, vcpu->arch.vgic.pendtable) ) + vgic_vcpu_inject_irq(vcpu, vlpi); + } + else + { + if ( pirq ) + { + clear_bit(GIC_IRQ_GUEST_ENABLED, &pirq->status); + gic_remove_from_queues(vcpu, vlpi); + } + } +} + +static int its_handle_inv(struct virt_its *its, uint64_t *cmdptr) +{ + uint32_t devid = its_cmd_get_deviceid(cmdptr); + uint32_t eventid = its_cmd_get_id(cmdptr); + struct vcpu *vcpu; + uint32_t vlpi; + + if ( !read_itte(its, devid, eventid, &vcpu, &vlpi) ) + return -1; + + update_lpi_enabled_status(its, vcpu, vlpi); + + return 0; +} + static int its_handle_mapc(struct virt_its *its, uint64_t *cmdptr) { uint32_t collid = its_cmd_get_collection(cmdptr); @@ -405,6 +459,9 @@ static int vgic_its_handle_cmds(struct domain *d, struct virt_its *its, case GITS_CMD_INT: its_handle_int(its, cmdptr); break; + case GITS_CMD_INV: + its_handle_inv(its, cmdptr); + break; case GITS_CMD_MAPC: its_handle_mapc(its, cmdptr); break; -- 2.9.0 _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxx https://lists.xen.org/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |