[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] xen-hypervisor-4.3-amd64 4.3.0-3+b1 -> Intel M/B, I/O virt. enabled, start vm -> Kernel panic
Hi Konrad et al, I have had no reply from the e1000-devel list about my "e1000 Tx Hang" problem, but I may have stumbled on something relevant. Today I created a new VM (wheezy) running on a Wheezy dom0 - my node blue in the previous e-mail trail. I decided to use the e1000e device at 0000:01:00.0 via pciback as this device is currently free (I'm building a replacement for an out of date DMZ VM). Just to recap, dom0 has the following devices: eth0 via-rhine 0000:04:00.0 assigned for zone LOC xenbr0 eth1 via-rhine 0000:04:01.0 assigned for zone DMZ (pci-passthrough) eth2 e1000 0000:04:02.0 used for DBRB eth3 e1000e 0000:01:00.0 planned for Windows client domU eth3 is the one I have "pinched" to build my new DMZ domu. When I create this domu (which uses a DRBD device) I get the following error (copied from dmesg) and eth0 in the domu is clearly unstable. [ 2.703909] input: PC Speaker as /devices/platform/pcspkr/input/input0 [ 2.780858] e1000e: Intel(R) PRO/1000 Network Driver - 2.3.2-k [ 2.780868] e1000e: Copyright(c) 1999 - 2013 Intel Corporation. [ 2.781119] e1000e 0000:00:00.0: enabling device (0000 -> 0002) [ 2.781500] e1000e 0000:00:00.0: Xen PCI mapped GSI16 to IRQ26 [ 2.781933] e1000e 0000:00:00.0: setting latency timer to 64 [ 2.782684] Error: Driver 'pcspkr' is already registered, aborting... [ 2.784207] e1000e 0000:00:00.0: Interrupt Throttling Rate (ints/sec) set to dynamic conservative mode [ 2.895529] e1000e 0000:00:00.0: eth0: (PCI Express:2.5GT/s:Width x1) 68:05:ca:21:80:2c [ 2.895542] e1000e 0000:00:00.0: eth0: Intel(R) PRO/1000 Network Connection [ 2.895557] e1000e 0000:00:00.0: eth0: MAC: 3, PHY: 8, PBA No: E46981-008 [ 3.153496] Adding 524284k swap on /dev/xvda1. Priority:-1 extents:1 across:524284k SS [ 3.531185] EXT3-fs (xvda2): using internal journal [ 5.026608] ADDRCONF(NETDEV_UP): eth0: link is not ready [ 8.121021] e1000e: eth0 NIC Link is Up 1000 Mbps Full Duplex, Flow Control: None [ 8.121454] ADDRCONF(NETDEV_CHANGE): eth0: link becomes ready [ 13.824073] ------------[ cut here ]------------ [ 13.824087] WARNING: at /build/linux-X2rDfB/linux-3.2.57/net/sched/sch_generic.c:256 dev_watchdog+0xf2/0x151() [ 13.824099] NETDEV WATCHDOG: eth0 (e1000e): transmit queue 0 timed out [ 13.824104] Modules linked in: evdev e1000e snd_pcm snd_page_alloc snd_timer snd soundcore pcspkr xen_pcifront coretemp ext3 mbcache jbd xen_blkfront [ 13.824130] Pid: 0, comm: swapper/0 Not tainted 3.2.0-4-amd64 #1 Debian 3.2.57-3 [ 13.824138] Call Trace: [ 13.824141] <IRQ> [<ffffffff81046cd9>] ? warn_slowpath_common+0x78/0x8c [ 13.824157] [<ffffffff81046d85>] ? warn_slowpath_fmt+0x45/0x4a [ 13.824163] [<ffffffff812a7705>] ? netif_tx_lock+0x40/0x75 [ 13.824171] [<ffffffff812a7875>] ? dev_watchdog+0xf2/0x151 [ 13.824179] [<ffffffff810524f8>] ? run_timer_softirq+0x19a/0x261 [ 13.824186] [<ffffffff8109124c>] ? handle_irq_event_percpu+0x15f/0x17d [ 13.824194] [<ffffffff812a7783>] ? netif_tx_unlock+0x49/0x49 [ 13.824203] [<ffffffff8104c36e>] ? __do_softirq+0xb9/0x177 [ 13.824209] [<ffffffff8121c0bd>] ? __xen_evtchn_do_upcall+0x24a/0x287 [ 13.824219] [<ffffffff81356c6c>] ? call_softirq+0x1c/0x30 [ 13.824227] [<ffffffff8100fa21>] ? do_softirq+0x3c/0x7b [ 13.824233] [<ffffffff8104c5d6>] ? irq_exit+0x3c/0x99 [ 13.824240] [<ffffffff8121d47d>] ? xen_evtchn_do_upcall+0x27/0x32 [ 13.824249] [<ffffffff81356cbe>] ? xen_do_hypervisor_callback+0x1e/0x30 [ 13.824254] <EOI> [<ffffffff810013aa>] ? hypercall_page+0x3aa/0x1000 [ 13.824264] [<ffffffff810013aa>] ? hypercall_page+0x3aa/0x1000 [ 13.824273] [<ffffffff8100675a>] ? xen_safe_halt+0xc/0x13 [ 13.824279] [<ffffffff81014614>] ? default_idle+0x47/0x7f [ 13.824286] [<ffffffff8100d24c>] ? cpu_idle+0xaf/0xf2 [ 13.824294] [<ffffffff816abb36>] ? start_kernel+0x3b8/0x3c3 [ 13.824301] [<ffffffff816ad4df>] ? xen_start_kernel+0x412/0x418 [ 13.824308] ---[ end trace c3ec188c56467b6a ]--- [ 13.824328] e1000e 0000:00:00.0: eth0: Reset adapter unexpectedly [ 17.613023] e1000e: eth0 NIC Link is Up 1000 Mbps Full Duplex, Flow Control: None [ 18.736090] eth0: no IPv6 routers present [ 22.832151] e1000e 0000:00:00.0: eth0: Reset adapter unexpectedly [ 26.741021] e1000e: eth0 NIC Link is Up 1000 Mbps Full Duplex, Flow Control: None [ 71.824169] e1000e 0000:00:00.0: eth0: Reset adapter unexpectedly [ 75.645020] e1000e: eth0 NIC Link is Up 1000 Mbps Full Duplex, Flow Control: None [ 365.824077] e1000e 0000:00:00.0: eth0: Reset adapter unexpectedly [ 369.629024] e1000e: eth0 NIC Link is Up 1000 Mbps Full Duplex, Flow Control: None [ 379.824070] e1000e 0000:00:00.0: eth0: Reset adapter unexpectedly [ 383.709023] e1000e: eth0 NIC Link is Up 1000 Mbps Full Duplex, Flow Control: None [ 393.824070] e1000e 0000:00:00.0: eth0: Reset adapter unexpectedly [ 397.657026] e1000e: eth0 NIC Link is Up 1000 Mbps Full Duplex, Flow Control: None Below is the vm config file in case there is anything useful in there. # # Configuration file for the Xen instance vm-server-22.maniscorse, created # by xen-tools 4.3.1 on Sun Jun 1 15:31:46 2014. # # # Kernel + memory size # #bootloader = '/usr/lib/xen-default/bin/pygrub' bootloader = 'pygrub' vcpus = '1' memory = '1024' # # Disk device(s). # root = '/dev/xvda2 ro' #disk = [ # 'phy:/dev/blue/vm-server-22.maniscorse-disk,xvda2,w', # 'phy:/dev/blue/vm-server-22.maniscorse-swap,xvda1,w', # ] # # Add support for drbd # disk = [ 'drbd:vm-22-disk,xvda2,w', 'phy:/dev/blue/vm-server-22.maniscorse-swap,xvda1,w', ] # # Physical volumes # # # Hostname # name = 'vm-server-22.maniscorse' # # Networking # #vif = [ 'ip=192.168.2.222 ,mac=00:16:3e:de:02:00' ] # # # Add support for looped through pci NIC # (Same device number on both blue and green) # pci = [ '01:00.0' ] # # Behaviour # on_poweroff = 'destroy' on_reboot = 'restart' on_crash = 'restart' Is it possible that there is some horrible interaction between the e1000 device and the e1000e device? It just seems quite a co-incidence. Best regards, Jo. _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxx http://lists.xen.org/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |