[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] Interrupt to CPU routing in HVM domains - again
John Levon wrote: On Fri, Sep 05, 2008 at 11:15:43AM -0400, Steve Ofsthun wrote:While the event channel delivery code "binds" HVM event channel interrupts to VCPU0, the interrupt is delivered via the emulated IOAPIC. The guest OS may program this "hardware" to deliver the interrupt to other VCPUs. For linux, this gets done by the irqbalance code among others. Xen overrides this routing for the timer 0 interrupt path in vioapic.c under the #define IRQ0_SPECIAL_ROUTING. We hacked our version of Xen to piggyback on this code to force all event channel interrupts for HVM guests to also avoid any guest rerouting:#ifdef IRQ0_SPECIAL_ROUTING /* Force round-robin to pick VCPU 0 */ if ( ((irq == hvm_isa_irq_to_gsi(0)) && pit_channel0_enabled()) || is_hvm_callback_irq(vioapic, irq) ) deliver_bitmask = (uint32_t)1; #endifYes, please - Solaris 10 PV drivers are buggy in that they use the current VCPUs vcpu_info. I just found this bug, and it's getting fixed, but if this makes sense anyway, it'd be good. I can submit a patch for this, but we feel this is something of a hack. We'd like to provide a more general mechanism for allowing event channel binding to "work" for HVM guests. But to do this, we are trying to address conflicting goals. Either we honor the event channel binding by circumventing the IOAPIC emulation, or we faithfully emulate the IOAPIC and circumvent the event channel binding. Our driver writers would like to see support for multiple callback IRQs. Then particular event channel interrupts could be bound to particular IRQs. This would allow PV device interrupts to be distributed intelligently. It would also allow net and block interrupts to be disentangled for Windows PV drivers. We deal pretty much exclusively with HVM guests, do SMP PV environments selectively bind device interrupts to different VCPUs? Steve This routing override provides a significant performance boost [or rather avoids the performance penalty] for SMP PV drivers up until the time that VCPU0 is saturated with interrupts. You can probably achieve the sameOf course there's no requirement that the evtchn is actually dealt with on the same CPU, just the callback IRQ and the evtchn "ack" (clearing evtchn_upcall_pending). regards john _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxxxxxxxx http://lists.xensource.com/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |