[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] RE: [Xen-devel] kernel BUG at net/core/dev.c:1133!
> -----Original Message----- > From: Tim Post [mailto:tim.post@xxxxxxxxxxxxxxx] > Sent: 07 July 2006 16:06 > To: Herbert Xu > Cc: Petersson, Mats; netdev@xxxxxxxxxxxxxxx; > xen-devel@xxxxxxxxxxxxxxxxxxx; kaber@xxxxxxxxx; davem@xxxxxxxxxxxxx > Subject: Re: [Xen-devel] kernel BUG at net/core/dev.c:1133! > > I got the exact same thing when attempting to use BOINC on a > single node > supporting a 5 node open SSI cluster, (5 guests) and yes the problem > went away when I flushed the rules. > > I attributed this to a quirk with the cluster CVIP, because I had also > assigned each node its own outbound IP in addition to the > incoming CVIP. > > Since I felt it was due to my tendency to over-tinker, I > didn't mention > it on the lists, was a few months ago. > > Thought I would chime in as it sounds like the same experience, up to > and including BOINC. I haven't been tinkering with anything [on purpose, at least] - the system is a default installation of FC4, with the latest Xen-unstable [bar the last dozen or so changesets - I don't pull the latest every half-hour]. -- Mats > > HTH > > --Tim > > On Sat, 2006-07-08 at 00:39 +1000, Herbert Xu wrote: > > Petersson, Mats <Mats.Petersson@xxxxxxx> wrote: > > > Looks like the GSO is involved? > > > > It's certainly what crashed your machine :) It's probably not the > > guilty party though. Someone is passing through a TSO packet with > > checksum set to something other than CHECKSUM_HW. > > > > I bet it's netfilter and we just never noticed before because real > > NICS would simply corrupt the checksum silently. > > > > Could you confirm that you have netfilter rules (in particular NAT > > rules) and that this goes away if you flush all your > netfilter tables? > > > > Patrick, do we really have to zap the checksum on outbound > NAT? Could > > we update it instead? > > > > > I got this while running Dom0 only (no guests), with a > > > BOINC/Rosetta@home application running on all 4 cores. > > > > > > changeset: 10649:8e55c5c11475 > > > > > > Build: x86_32p (pae). > > > > > > ------------[ cut here ]------------ > > > kernel BUG at net/core/dev.c:1133! > > > invalid opcode: 0000 [#1] > > > SMP > > > CPU: 0 > > > EIP: 0061:[<c04dceb0>] Not tainted VLI > > > EFLAGS: 00210297 (2.6.16.13-xen #12) > > > EIP is at skb_gso_segment+0xf0/0x110 > > > eax: 00000000 ebx: 00000003 ecx: 00000002 edx: c06e2e00 > > > esi: 00000008 edi: cd9e32e0 ebp: c63a7900 esp: c0de5ad0 > > > ds: 007b es: 007b ss: 0069 > > > Process rosetta_5.25_i6 (pid: 8826, threadinfo=c0de4000 > task=cb019560) > > > Stack: <0>c8f69060 00000000 ffffffa3 00000003 cd9e32e0 > 00000002 c63a7900 > > > c04dcfb0 > > > cd9e32e0 00000003 00000000 cd9e32e0 cf8e3000 > cf8e3140 c04dd07e > > > cd9e32e0 > > > cf8e3000 00000000 cd9e32e0 cf8e3000 c04ec07e > cd9e32e0 cf8e3000 > > > c0895140 > > > Call Trace: > > > [<c04dcfb0>] dev_gso_segment+0x30/0xb0 > > > [<c04dd07e>] dev_hard_start_xmit+0x4e/0x110 > > > [<c04ec07e>] __qdisc_run+0xbe/0x280 > > > [<c04dd4b9>] dev_queue_xmit+0x379/0x380 > > > [<c05bbe44>] br_dev_queue_push_xmit+0xa4/0x140 > > > [<c05c2402>] br_nf_post_routing+0x102/0x1d0 > > > [<c05c22b0>] br_nf_dev_queue_xmit+0x0/0x50 > > > [<c05bbda0>] br_dev_queue_push_xmit+0x0/0x140 > > > [<c04f0eab>] nf_iterate+0x6b/0xa0 > > > [<c05bbda0>] br_dev_queue_push_xmit+0x0/0x140 > > > [<c05bbda0>] br_dev_queue_push_xmit+0x0/0x140 > > > [<c04f0f4e>] nf_hook_slow+0x6e/0x120 > > > [<c05bbda0>] br_dev_queue_push_xmit+0x0/0x140 > > > [<c05bbf40>] br_forward_finish+0x60/0x70 > > > [<c05bbda0>] br_dev_queue_push_xmit+0x0/0x140 > > > [<c05c1b71>] br_nf_forward_finish+0x71/0x130 > > > [<c05bbee0>] br_forward_finish+0x0/0x70 > > > [<c05c1d20>] br_nf_forward_ip+0xf0/0x1a0 > > > [<c05c1b00>] br_nf_forward_finish+0x0/0x130 > > > [<c05bbee0>] br_forward_finish+0x0/0x70 > > > [<c04f0eab>] nf_iterate+0x6b/0xa0 > > > [<c05bbee0>] br_forward_finish+0x0/0x70 > > > [<c05bbee0>] br_forward_finish+0x0/0x70 > > > [<c04f0f4e>] nf_hook_slow+0x6e/0x120 > > > [<c05bbee0>] br_forward_finish+0x0/0x70 > > > [<c05bc044>] __br_forward+0x74/0x80 > > > [<c05bbee0>] br_forward_finish+0x0/0x70 > > > [<c05bceb1>] br_handle_frame_finish+0xd1/0x160 > > > [<c05bcde0>] br_handle_frame_finish+0x0/0x160 > > > [<c05c0e0b>] br_nf_pre_routing_finish+0xfb/0x480 > > > [<c05bcde0>] br_handle_frame_finish+0x0/0x160 > > > [<c05c0d10>] br_nf_pre_routing_finish+0x0/0x480 > > > [<c054fe13>] ip_nat_in+0x43/0xc0 > > > [<c05c0d10>] br_nf_pre_routing_finish+0x0/0x480 > > > [<c04f0eab>] nf_iterate+0x6b/0xa0 > > > [<c05c0d10>] br_nf_pre_routing_finish+0x0/0x480 > > > [<c05c0d10>] br_nf_pre_routing_finish+0x0/0x480 > > > [<c04f0f4e>] nf_hook_slow+0x6e/0x120 > > > [<c05c0d10>] br_nf_pre_routing_finish+0x0/0x480 > > > [<c05c1914>] br_nf_pre_routing+0x404/0x580 > > > [<c05c0d10>] br_nf_pre_routing_finish+0x0/0x480 > > > [<c04f0eab>] nf_iterate+0x6b/0xa0 > > > [<c05bcde0>] br_handle_frame_finish+0x0/0x160 > > > [<c05bcde0>] br_handle_frame_finish+0x0/0x160 > > > [<c04f0f4e>] nf_hook_slow+0x6e/0x120 > > > [<c05bcde0>] br_handle_frame_finish+0x0/0x160 > > > [<c05bd124>] br_handle_frame+0x1e4/0x250 > > > [<c05bcde0>] br_handle_frame_finish+0x0/0x160 > > > [<c04ddae5>] netif_receive_skb+0x165/0x2a0 > > > [<c04ddcdf>] process_backlog+0xbf/0x180 > > > [<c04ddebf>] net_rx_action+0x11f/0x1d0 > > > [<c01262e6>] __do_softirq+0x86/0x120 > > > [<c01263f5>] do_softirq+0x75/0x90 > > > [<c0106cef>] do_IRQ+0x1f/0x30 > > > [<c04271d0>] evtchn_do_upcall+0x90/0x100 > > > [<c0105315>] hypervisor_callback+0x3d/0x48 > > > Code: c2 2b 57 24 29 d0 8d 14 2a 89 87 94 00 00 00 89 57 > 60 8b 44 24 08 > > > 83 c4 0c 5b 5e 5f 5d c3 0f 0 > > > b 69 03 fe 8c 66 c0 e9 69 ff ff ff <0f> 0b 6d 04 e8 ab 6c > c0 e9 3a ff ff > > > ff 0f 0b 6c 04 e8 ab 6c c0 > > > <0>Kernel panic - not syncing: Fatal exception in interrupt > > > > Cheers, > > > _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxxxxxxxx http://lists.xensource.com/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |