[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] [PATCH v4 1/2] arm: support fewer LR registers than virtual irqs
On Tue, 28 Feb 2012, Ian Campbell wrote: > On Thu, 2012-02-23 at 17:00 +0000, Stefano Stabellini wrote: > > If the vgic needs to inject a virtual irq into the guest, but no free > > LR registers are available, add the irq to a list and return. > > There is an ordering constraint on this list. It should be mentioned > somewhere in a comment. right > > Whenever an LR register becomes available we add the queued irq to it > > and remove it from the list. > > We use the gic lock to protect the list and the bitmask. > > > > Signed-off-by: Stefano Stabellini <stefano.stabellini@xxxxxxxxxxxxx> > > --- > > xen/arch/arm/gic.c | 95 > > ++++++++++++++++++++++++++++++++---------- > > xen/include/asm-arm/domain.h | 1 + > > 2 files changed, 74 insertions(+), 22 deletions(-) > > > > diff --git a/xen/arch/arm/gic.c b/xen/arch/arm/gic.c > > index adc10bb..2ff7bce 100644 > > --- a/xen/arch/arm/gic.c > > +++ b/xen/arch/arm/gic.c > > @@ -25,6 +25,7 @@ > > #include <xen/sched.h> > > #include <xen/errno.h> > > #include <xen/softirq.h> > > +#include <xen/list.h> > > #include <asm/p2m.h> > > #include <asm/domain.h> > > > > @@ -45,6 +46,8 @@ static struct { > > unsigned int lines; > > unsigned int cpus; > > spinlock_t lock; > > + uint64_t lr_mask; > > + struct list_head lr_pending; > > } gic; > > > > irq_desc_t irq_desc[NR_IRQS]; > > @@ -247,6 +250,8 @@ static void __cpuinit gic_hyp_init(void) > > > > GICH[GICH_HCR] = GICH_HCR_EN; > > GICH[GICH_MISR] = GICH_MISR_EOI; > > + gic.lr_mask = 0ULL; > > + INIT_LIST_HEAD(&gic.lr_pending); > > } > > > > /* Set up the GIC */ > > @@ -345,16 +350,50 @@ int __init setup_irq(unsigned int irq, struct > > irqaction *new) > > return rc; > > } > > > > -void gic_set_guest_irq(unsigned int virtual_irq, > > +static inline void gic_set_lr(int lr, unsigned int virtual_irq, > > unsigned int state, unsigned int priority) > > { > > - BUG_ON(virtual_irq > nr_lrs); > > - GICH[GICH_LR + virtual_irq] = state | > > + BUG_ON(lr > nr_lrs); > > + GICH[GICH_LR + lr] = state | > > GICH_LR_MAINTENANCE_IRQ | > > ((priority >> 3) << GICH_LR_PRIORITY_SHIFT) | > > ((virtual_irq & GICH_LR_VIRTUAL_MASK) << GICH_LR_VIRTUAL_SHIFT); > > } > > > > +void gic_set_guest_irq(unsigned int virtual_irq, > > + unsigned int state, unsigned int priority) > > +{ > > + int i; > > + struct pending_irq *iter, *n; > > + > > + spin_lock(&gic.lock); > > + > > + if ( list_empty(&gic.lr_pending) ) > > This could really do with some comments. > > The logic here is that if the "lr_pending" list has nothing in it then > there are no other IRQs queued waiting for an LR to become available and > therefore we can take a fast path and inject this IRQ directly? Yes: if there are no IRQs waiting for an LR it means that an LR might actually be free. If we have IRQs queued in lr_pending then there is no point in checking for a free LR. > > + { > > + i = find_first_zero_bit(&gic.lr_mask, sizeof(uint64_t)); > > + if (i < sizeof(uint64_t)) { > > What does find_first_zero_bit(0xffff.fff, 64) return? 0 > > + set_bit(i, &gic.lr_mask); > > + gic_set_lr(i, virtual_irq, state, priority); > > + goto out; > > + } > > + } > > + > > + n = irq_to_pending(current, virtual_irq); > > + list_for_each_entry ( iter, &gic.lr_pending, lr_link ) > > + { > > + if ( iter->priority < priority ) > > + { > > + list_add_tail(&n->lr_link, &iter->lr_link); > > + goto out; > > + } > > + } > > + list_add(&n->lr_link, &gic.lr_pending); > > Should this be list_add_tail? yes > Lower numbers are higher priority and therefore the list should be > ordered from low->high, since we want to pull the next interrupt off the > front of the list. I don't think the above implements that though. Yes, the priority test is inverted. I think we want: iter->priority > priority > If we imagine a list containing priorities [2,4,6] into which we are > inserting a priority 5 interrupt. > > On the first iteration of the loop iter->priority == 2 so "if > (iter->priority < priority)" is true and we insert 5 after 2 in the list > resulting in [2,5,4,6]. Actually list_add_tail adds the entry before the head, not after. So if we have the right priority check and the list is [2,4,6], adding 5 should result in [2,4,5,6]. > > + > > +out: > > + spin_unlock(&gic.lock); > > + return; > > +} > > + > > void gic_inject_irq_start(void) > > { > > uint32_t hcr; > > @@ -431,30 +470,42 @@ void gicv_setup(struct domain *d) > > > > static void maintenance_interrupt(int irq, void *dev_id, struct > > cpu_user_regs *regs) > > { > > - int i, virq; > > + int i = 0, virq; > > uint32_t lr; > > uint64_t eisr = GICH[GICH_EISR0] | (((uint64_t) GICH[GICH_EISR1]) << > > 32); > > > > - for ( i = 0; i < 64; i++ ) { > > - if ( eisr & ((uint64_t)1 << i) ) { > > - struct pending_irq *p; > > - > > - lr = GICH[GICH_LR + i]; > > - virq = lr & GICH_LR_VIRTUAL_MASK; > > - GICH[GICH_LR + i] = 0; > > - > > - spin_lock(¤t->arch.vgic.lock); > > - p = irq_to_pending(current, virq); > > - if ( p->desc != NULL ) { > > - p->desc->status &= ~IRQ_INPROGRESS; > > - GICC[GICC_DIR] = virq; > > - } > > + while ((i = find_next_bit((const long unsigned int *) &eisr, > > + sizeof(eisr), i)) < sizeof(eisr)) { > > > + struct pending_irq *p; > > + > > + spin_lock(&gic.lock); > > + lr = GICH[GICH_LR + i]; > > + virq = lr & GICH_LR_VIRTUAL_MASK; > > + GICH[GICH_LR + i] = 0; > > + clear_bit(i, &gic.lr_mask); > > + > > + if ( !list_empty(gic.lr_pending.next) ) { > > This seems odd, why not "list_empty(gic.lr_pending)"? > > Otherwise won't you fail to inject anything if there is exactly one > entry on lr_pending? You are correct, that is a mistake. > > + p = list_entry(gic.lr_pending.next, typeof(*p), lr_link); > > + gic_set_lr(i, p->irq, GICH_LR_PENDING, p->priority); > > + list_del_init(&p->lr_link); > > + set_bit(i, &gic.lr_mask); > > + } else { > > gic_inject_irq_stop(); > > - list_del(&p->link); > > - INIT_LIST_HEAD(&p->link); > > - cpu_raise_softirq(current->processor, VGIC_SOFTIRQ); > > - spin_unlock(¤t->arch.vgic.lock); > > } > > + spin_unlock(&gic.lock); > > + > > + spin_lock(¤t->arch.vgic.lock); > > + p = irq_to_pending(current, virq); > > + if ( p->desc != NULL ) { > > + p->desc->status &= ~IRQ_INPROGRESS; > > + GICC[GICC_DIR] = virq; > > + } > > + list_del(&p->link); > > + INIT_LIST_HEAD(&p->link); > > + cpu_raise_softirq(current->processor, VGIC_SOFTIRQ); > > + spin_unlock(¤t->arch.vgic.lock); > > + > > + i++; > > Does find_next_bit include or exclude the bit which you give it as the > 3rd argument? include > > } > > } > > > > diff --git a/xen/include/asm-arm/domain.h b/xen/include/asm-arm/domain.h > > index 3372d14..75095ff 100644 > > --- a/xen/include/asm-arm/domain.h > > +++ b/xen/include/asm-arm/domain.h > > @@ -21,6 +21,7 @@ struct pending_irq > > struct irq_desc *desc; /* only set it the irq corresponds to a > > physical irq */ > > uint8_t priority; > > struct list_head link; > > + struct list_head lr_link; > > Calling this "link" or "lr_link" when it is used in the context or link > registers (e.g. link-register_link) just adds to the confusion around > these two lists IMHO, as does having one just called link and the other > prefix_link. Calling them foo_list, both with a descriptive prefix, > would be better, e.g. active_list and queued_list (or whatever you deem > appropriate to their semantics) > > Even better would be if the invariant "always on either active or > pending lists, never both" were true -- in which case only one list_head > would be needed here. I'll try to come up with better descriptive names and comments. > What do we actually use "link" for? It is chained off vgic.inflight_irqs > but we seem to only use it for anything other than manipulating itself > in vgic_softirq where it is used as a binary "something to inject" flag > -- we could just as well maintain a nr_inflight variable if that were > the case, couldn't we? It is used by the vgic to keep track of what IRQs have been injected from its point of view. These IRQs might have been injected and currently resident in an LR register, or they might be queued in lr_pending. I'll write a comment to better the explain the life cycle of an IRQ. _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxx http://lists.xen.org/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |