[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] RE: [Xen-devel] IRQ SMP affinity problems in domU with vcpus > 4 on HP ProLiant G6 with dual Xeon 5540 (Nehalem)
I switched over to Xen 3.5-unstable (changeset 20303) and pv_ops dom0 2.6.31.1 hoping that this would resolve the IRQ SMP affinity problem. I had to use pci-stub to hide the PCI devices since pciback wasn't working. With vcpus=16 (APIC routing is physical flat), the interrupts were working in domU and being routed to CPU0 with the default smp_affinity (ffff) but as soon as I changed it to any 16-bit one-hot value or even setting it to the same default value resulted in a complete loss of interrupts (even in the devices that didn't have any change to smp_affinity). With vcpus=4 (APIC routing is logical flat), I can see the interrupts being load balanced across all CPUs but as soon as I changed smp_affinity to any value, the interrupts stopped. This used to work reliably with the non-pv_ops kernel. I attached the logs in case anyone wants to take a look. I did see the MSI message address/data change in both domU and dom0 (using "lspci -vv"): vcpus=16: domU MSI message address/data with default smp_affinity: Address: 00000000fee00000 Data: 40a9 domU MSI message address/data after smp_affinity=0010: Address: 00000000fee08000 Data: 40b1 (8 is APIC ID of CPU4) dom0 MSI message address/data with default smp_affinity: Address: 00000000fee00000 Data: 4094 dom0 MSI message address/data after smp_affinity=0010: Address: 00000000fee00000 Data: 409c Aside from "lspci -vv" what other means are there to track down this problem? Is there some way to print the interrupt vector table? I'm considering adding printk's to the code that Qing mentioned in his previous email (see below). Any suggestions on where in the code to add the printk's? Thanks. Dante -----Original Message----- From: Qing He [mailto:qing.he@xxxxxxxxx] Sent: Sunday, October 11, 2009 10:55 PM To: Cinco, Dante Cc: Keir Fraser; xen-devel@xxxxxxxxxxxxxxxxxxx; xiantao.zhang@xxxxxxxxx Subject: Re: [Xen-devel] IRQ SMP affinity problems in domU with vcpus > 4 on HP ProLiant G6 with dual Xeon 5540 (Nehalem) On Mon, 2009-10-12 at 13:25 +0800, Cinco, Dante wrote: > With vcpus < 4, logical flat mode works fine (no error message). I can > change smp_affinity to any value > 0 and < 16 and the interrupts go to > the proper CPU(s). Could you point me to the code that handles MSI so > that I can better understand the MSI implementation? There are two parts: 1) init or changing the data and address of MSI: 1) qemu-xen: hw/passthrough.c: pt_msg.*_write, MSI access are trapped here first. And then pt_update_msi in hw/pt-msi.c is called to update the MSI binding. 2) xen: drivers/passthrough/io.c: pt_irq_create_bind_vtd, where MSI is actually bound to the guest. 2) on MSI reception: In drivers/passthrough/io.c, hvm_do_IRQ_dpci and hvm_dirq_assist are the routines responsible for handling all assigned irqs (including MSI), and if an MSI is received, vmsi_deliver in arch/x86/vmsi.c get called to deliver MSI to the corresponding vlapic. And I just learned from Xiantao Zhang that for the guest Linux kernel, it enables per-cpu vector if it's in physical mode, and that looks more likely relevant to this problem. It had problem in the older xen to handle this, and changeset 20253 is supposed to fix it, although I noticed your xen version is 20270. Thanks, Qing Attachment:
irq_smp_affinity_problem_pv_ops_dom0_2.6.31.1.tar.gz _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxxxxxxxx http://lists.xensource.com/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |