[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] [PATCH 2/5] xen/arm: Keep count of inflight interrupts
On 06/25/2013 05:58 PM, Stefano Stabellini wrote: > On Tue, 25 Jun 2013, Ian Campbell wrote: >> On Tue, 2013-06-25 at 00:04 +0100, Julien Grall wrote: >>> From: Stefano Stabellini <stefano.stabellini@xxxxxxxxxx> >>> >>> For guest's timers (both virtual and physical), Xen will inject virtual >>> interrupt. Linux handles timer interrupt as: >> >> We should be wary of developing things based on the way which Linux >> happens to do things. On x86 we have several time modes, which can be >> selected based upon guest behaviour (described in >> docs/man/xl.cfg.pod.5). Do we need to do something similar here? >> >>> 1) Receive the interrupt and ack it >>> 2) Handle the current event timer >>> 3) Set the next event timer >>> 4) EOI the interrupt >>> >>> It's unlikely possible to reinject another interrupt before >> >> I can't parse this sentence. "unlikely to be possible" perhaps? but I'm >> not sure if that is what you meant. >> >>> the previous one is EOIed because the next deadline is shorter than the time >>> to execute code until EOI it. >> >> If we get into this situation once is there any danger that we will get >> into it repeatedly and overflow the count? >> >> Overall I'm not convinced this is the right approach to get the >> behaviour we want. Isn't this interrupt level triggered, with the level >> being determined by a comparison of two registers? IOW can't we >> determine whether to retrigger the interrupt or not by examining the >> state of our emulated versions of those registers? A generic mechanism >> to callback into the appropriate emulator on EOI plus a little bit of >> logic in the vtimer code is all it ought to take. > > AFAICT this what could happen: > > - vtimer fires > - xen mask the vtimer > - xen adds the vtimer to the LR for the guest > - xen EOIs the vtimer > - the guest receive the vtimer interrupt > - the guest set the next deadline in the past > - the guest enables the vtimer > ## an unexpected vtimer interrupt is received by Xen but the current > ## one is still being serviced > - the guest eoi the vtimer > > as a result the guest looses an interrupt. > Julien, is that correct? Yes. > If that is the case, this can only happen once, right? In that case > rather than an atomic_t we could just have a bit in the status field I > proposed before. It should be enough for us to keep track of the case > when the irq is supposed to stay high even after the guest EOIs it. (Of > course that means that we need to re-inject it into the guest). For the timer yes. I wonder, what happen if Xen receive an SGI (for instance SCHEDULE ipi) twice, does Xen must re-inject it multiple times? >>> Signed-off-by: Stefano Stabellini <stefano.stabellini@xxxxxxxxxx> >>> Signed-off-by: Julien Grall <julien.grall@xxxxxxxxxx> >>> --- >>> xen/arch/arm/gic.c | 35 +++++++++++++++++++++++------------ >>> xen/arch/arm/vgic.c | 1 + >>> xen/include/asm-arm/domain.h | 2 ++ >>> 3 files changed, 26 insertions(+), 12 deletions(-) >>> >>> diff --git a/xen/arch/arm/gic.c b/xen/arch/arm/gic.c >>> index 0fee3f2..21575df 100644 >>> --- a/xen/arch/arm/gic.c >>> +++ b/xen/arch/arm/gic.c >>> @@ -817,7 +817,7 @@ static void maintenance_interrupt(int irq, void >>> *dev_id, struct cpu_user_regs *r >>> >>> while ((i = find_next_bit((const long unsigned int *) &eisr, >>> 64, i)) < 64) { >>> - struct pending_irq *p; >>> + struct pending_irq *p, *n; >>> int cpu, eoi; >>> >>> cpu = -1; >>> @@ -826,13 +826,32 @@ static void maintenance_interrupt(int irq, void >>> *dev_id, struct cpu_user_regs *r >>> spin_lock_irq(&gic.lock); >>> lr = GICH[GICH_LR + i]; >>> virq = lr & GICH_LR_VIRTUAL_MASK; >>> + >>> + p = irq_to_pending(v, virq); >>> + if ( p->desc != NULL ) { >>> + p->desc->status &= ~IRQ_INPROGRESS; > > Now that I am thinking about this, shouldn't this be protected by taking > the desc->lock? Right. I don't think desc->lock is enough if we want to protect from release_irq. If this function is called, it will wait until IRQ_INPROGRESS is removed. How about moving ~IRQ_INPROGRESS at then end of the block and add a dmb() before? >>> + /* Assume only one pcpu needs to EOI the irq */ >>> + cpu = p->desc->arch.eoi_cpu; >>> + eoi = 1; >>> + pirq = p->desc->irq; >>> + } >>> + if ( !atomic_dec_and_test(&p->inflight_cnt) ) >>> + { >>> + /* Physical IRQ can't be reinject */ >>> + WARN_ON(p->desc != NULL); >>> + gic_set_lr(i, p->irq, GICH_LR_PENDING, p->priority); >>> + spin_unlock_irq(&gic.lock); >>> + i++; >>> + continue; >>> + } >>> + >>> GICH[GICH_LR + i] = 0; >>> clear_bit(i, &this_cpu(lr_mask)); >>> >>> if ( !list_empty(&v->arch.vgic.lr_pending) ) { >>> - p = list_entry(v->arch.vgic.lr_pending.next, typeof(*p), >>> lr_queue); >>> - gic_set_lr(i, p->irq, GICH_LR_PENDING, p->priority); >>> - list_del_init(&p->lr_queue); >>> + n = list_entry(v->arch.vgic.lr_pending.next, typeof(*n), >>> lr_queue); >>> + gic_set_lr(i, n->irq, GICH_LR_PENDING, n->priority); >>> + list_del_init(&n->lr_queue); >>> set_bit(i, &this_cpu(lr_mask)); >>> } else { >>> gic_inject_irq_stop(); >>> @@ -840,14 +859,6 @@ static void maintenance_interrupt(int irq, void >>> *dev_id, struct cpu_user_regs *r >>> spin_unlock_irq(&gic.lock); >>> >>> spin_lock_irq(&v->arch.vgic.lock); >>> - p = irq_to_pending(v, virq); >>> - if ( p->desc != NULL ) { >>> - p->desc->status &= ~IRQ_INPROGRESS; >>> - /* Assume only one pcpu needs to EOI the irq */ >>> - cpu = p->desc->arch.eoi_cpu; >>> - eoi = 1; >>> - pirq = p->desc->irq; >>> - } >>> list_del_init(&p->inflight); >>> spin_unlock_irq(&v->arch.vgic.lock); >>> > _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxx http://lists.xen.org/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |