[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] RE: [Xen-devel] [xen-unstable test] 1959: FAIL [and 1 more messages]
I think its safer to explicitly stop the timer. Might there also be a race with a pending timer when setting up the new one on reconnect? J "Ian Campbell" <Ian.Campbell@xxxxxxxxxxxxx> wrote: >On Fri, 2010-07-30 at 05:58 +0100, Xu, Dongxiao wrote: >> Hi Ian, >> >> Can you try this patch to see if it fixes your problem? >> >> I tried it on my local machine, VM is pinging another host. >> After save/restore/migration, the ping is still ongoing. >> >> Thanks, >> Dongxiao >> >> >> From: Dongxiao Xu <dongxiao.xu@xxxxxxxxx> >> >> When saving guest, the shared ring will be set to NULL, and >> in this case the polling timer interrupt should stop. >> >> Thanks Ian J for reporting the problem. >> >> Signed-off-by: Dongxiao Xu <dongxiao.xu@xxxxxxxxx> >> --- >> drivers/net/xen-netfront.c | 4 ++++ >> 1 files changed, 4 insertions(+), 0 deletions(-) >> >> diff --git a/drivers/net/xen-netfront.c b/drivers/net/xen-netfront.c >> index 328fe40..394bbc8 100644 >> --- a/drivers/net/xen-netfront.c >> +++ b/drivers/net/xen-netfront.c >> @@ -1340,6 +1340,10 @@ static enum hrtimer_restart >> smart_poll_function(struct hrtimer *timer) >> np = netdev_priv(dev); >> >> spin_lock_irqsave(&np->tx_lock, flags); >> + >> + if (!np->rx.sring) >> + goto end; >> + > >Isn't there a period until the end of xennet_disconnect_backend is >reached on resume where rx.sring will still point to the old shared >ring? If so is it safe to drop through in that case? > >Would it be safer to add an explicit suspend handler which stopped the >timer? > >Ian. > >> np->smart_poll.counter++;xennet_disconnect_backend >> >> if (likely(netif_carrier_ok(dev))) { >> -- >> 1.6.3 >> >> >> >> Jeremy Fitzhardinge wrote: >> > On 07/29/2010 09:33 AM, Ian Jackson wrote: >> >> Jeremy Fitzhardinge writes ("Re: [Xen-devel] [xen-unstable test] >> >> 1959: FAIL [and 1 more messages]"): >> >>> On 07/29/2010 08:30 AM, Ian Jackson wrote: >> >>>> Is save/restore supposed to work in pvops ? (Using your kernel for >> >>>> both dom0 and domU.) That would seem to be the next thing to pick >> >>>> off the list ... >> >>> Yes. IanC has been tracking down a bug where it fails after a few >> >>> thousand iterations (now fixed?), but aside from that its >> >>> apparently OK. >> >> Well, I was able to reproduce the failure that the automatic test was >> >> getting. There's a problem with it not getting a copy of the console >> >> output for some reason but I was able to get this oops from the >> >> guest. >> >> >> >> Just after the restore I was able to connect to the PV console and it >> >> echoed a few of my CRs (before saving I had logged in on the >> >> console), >> >> and then it produced the oops. Now it's apparently completely >> >> wedged. >> >> >> > >> > Ah, it looks like the netfront smartpoll stuff isn't coping with >> > save/restore. >> > >> > J >> > >> >> Ian. >> >> >> >> >> >> >> >> [ 63.681260] BUG: unable to handle kernel NULL pointer dereference >> >> at 00000010 [ 63.681293] IP: [<c1300381>] >> >> smart_poll_function+0xbb/0xf2 [ 63.681320] *pdpt = >> >> 000000001fee5027 *pde = 0000000000000000 [ 63.681344] Oops: 0000 >> >> [#1] SMP [ 63.681362] last sysfs file: /sys/kernel/uevent_seqnum >> >> [ 63.681376] Modules linked in: [last unloaded: scsi_wait_scan] [ >> >> 63.681398] [ 63.681410] Pid: 5, comm: events/0 Not tainted >> >> (2.6.32.16 #1) [ 63.681424] EIP: 0061:[<c1300381>] EFLAGS: >> >> 00010002 CPU: 0 [ 63.681438] EIP is at >> >> smart_poll_function+0xbb/0xf2 [ 63.681451] EAX: 00000000 EBX: >> >> dfea8320 ECX: 00000001 EDX: 00000062 [ 63.681465] ESI: 00000064 >> >> EDI: 00000000 EBP: df849cfc ESP: df849cdc [ 63.681479] DS: 007b >> >> ES: 007b FS: 00d8 GS: 0000 SS: 0069 [ 63.681493] Process events/0 >> >> (pid: 5, ti=df848000 task=df839480 task.ti=df848000) [ 63.681508] >> >> Stack: [ 63.681516] dfeac828 00000002 dfeac828 dfea8368 dfea0068 >> >> 0000002c dfeac828 c13002c6 [ 63.681555]<0> df849d1c c1079ab8 >> >> df849d48 c2386358 c2386328 0000002c 00000000 c2386328 [ >> >> 63.681598]<0> df849d5c c1079ce8 000028cd 88bbcad9 d3b2cae7 0000000e >> >> d3b2cae7 0000000e [ 63.681646] Call Trace: [ 63.681662] >> >> [<c13002c6>] ? smart_poll_function+0x0/0xf2 [ 63.681683] >> >> [<c1079ab8>] ? __run_hrtimer+0xa9/0xf6 [ 63.681701] [<c1079ce8>] >> >> ? hrtimer_interrupt+0xcd/0x1c8 [ 63.681719] [<c102cef0>] ? >> >> xen_timer_interrupt+0x2b/0x224 [ 63.681737] [<c102cb3f>] ? >> >> xen_force_evtchn_callback+0xf/0x14 [ 63.681755] [<c102d2ac>] ? >> >> check_events+0x8/0xc [ 63.681776] [<c102d2a3>] ? >> >> xen_restore_fl_direct_end+0x0/0x1 [ 63.681795] [<c14d1471>] ? >> >> _spin_unlock_irqrestore+0x2f/0x31 [ 63.681814] [<c105eb9e>] ? >> >> try_to_wake_up+0x2fa/0x304 [ 63.681832] [<c102cb3f>] ? >> >> xen_force_evtchn_callback+0xf/0x14 [ 63.681850] [<c10a0cc4>] ? >> >> handle_IRQ_event+0x5f/0x122 [ 63.681867] [<c10a233f>] ? >> >> handle_level_irq+0x58/0xa9 [ 63.681886] [<c121983d>] ? >> >> __xen_evtchn_do_upcall+0xab/0x131 [ 63.681904] [<c1219c71>] ? >> >> xen_evtchn_do_upcall+0x20/0x30 [ 63.682179] [<c102ffe7>] ? >> >> xen_do_upcall+0x7/0xc [ 63.682179] [<c102007b>] ? >> >> apic_reg_write+0xa5/0x52f [ 63.682179] [<c1002227>] ? >> >> hypercall_page+0x227/0x1005 [ 63.682179] [<c102cb3f>] ? >> >> xen_force_evtchn_callback+0xf/0x14 [ 63.682179] [<c102d2ac>] ? >> >> check_events+0x8/0xc [ 63.682179] [<c102d26b>] ? >> >> xen_irq_enable_direct_end+0x0/0x1 [ 63.682179] [<c105cd0d>] ? >> >> finish_task_switch+0x4f/0xa6 [ 63.682179] [<c14d0186>] ? >> >> schedule+0x7dd/0x861 [ 63.682179] [<c106dd9d>] ? >> >> __mod_timer+0x135/0x140 [ 63.682179] [<c102d2ac>] ? >> >> check_events+0x8/0xc [ 63.682179] [<c102d2a3>] ? >> >> xen_restore_fl_direct_end+0x0/0x1 [ 63.682179] [<c14d1471>] ? >> >> _spin_unlock_irqrestore+0x2f/0x31 [ 63.682179] [<c107750b>] ? >> >> prepare_to_wait+0x43/0x48 [ 63.682179] [<c10742a2>] ? >> >> worker_thread+0x94/0x1d2 [ 63.682179] [<c10ccd2d>] ? >> >> vmstat_update+0x0/0x2f [ 63.682179] [<c1077357>] ? >> >> autoremove_wake_function+0x0/0x33 [ 63.682179] [<c107420e>] ? >> >> worker_thread+0x0/0x1d2 [ 63.682179] [<c1077120>] ? >> >> kthread+0x61/0x66 [ 63.682179] [<c10770bf>] ? kthread+0x0/0x66 [ >> >> 63.682179] [<c102ff97>] ? kernel_thread_helper+0x7/0x10 [ >> >> 63.682179] Code: c6 89 d0 31 d2 f7 f6 85 d2 75 1a 85 c9 75 0c 8b 83 >> >> 90 08 00 00 c6 40 10 00 eb 32 c7 83 54 45 00 00 00 00 00 00 8b 83 90 >> >> 08 00 00<80> 78 10 00 74 1c 8b 4d e8 b8 00 ca 9a 3b 31 d2 f7 71 44 >> >> 31 c9 [ 63.682179] EIP: [<c1300381>] smart_poll_function+0xbb/0xf2 >> >> SS:ESP 0069:df849cdc [ 63.682179] CR2: 0000000000000010 [ >> >> 63.682179] ---[ end trace 760037e75e5675c8 ]--- [ 63.682179] >> >> Kernel panic - not syncing: Fatal exception in interrupt [ >> >> 63.682179] Pid: 5, comm: events/0 Tainted: G D 2.6.32.16 #1 >> >> [ 63.682179] Call Trace: [ 63.682179] [<c14cf7b1>] ? >> >> printk+0xf/0x11 [ 63.682179] [<c14cf6ee>] panic+0x39/0xed [ >> >> 63.682179] [<c14d2085>] oops_end+0xa1/0xb0 [ 63.682179] >> >> [<c104b426>] no_context+0x137/0x141 [ 63.682179] [<c104b56f>] >> >> __bad_area_nosemaphore+0x13f/0x147 [ 63.682179] [<c104b584>] >> >> bad_area_nosemaphore+0xd/0x10 [ 63.682179] [<c14d3275>] >> >> do_page_fault+0x1c6/0x32b [ 63.682179] [<c14d30af>] ? >> >> do_page_fault+0x0/0x32b [ 63.682179] [<c14d16c6>] >> >> error_code+0x66/0x6c [ 63.682179] [<c14d30af>] ? >> >> do_page_fault+0x0/0x32b [ 63.682179] [<c1300381>] ? >> >> smart_poll_function+0xbb/0xf2 [ 63.682179] [<c13002c6>] ? >> >> smart_poll_function+0x0/0xf2 [ 63.682179] [<c1079ab8>] >> >> __run_hrtimer+0xa9/0xf6 [ 63.682179] [<c1079ce8>] >> >> hrtimer_interrupt+0xcd/0x1c8 [ 63.682179] [<c102cef0>] >> >> xen_timer_interrupt+0x2b/0x224 [ 63.682179] [<c102cb3f>] ? >> >> xen_force_evtchn_callback+0xf/0x14 [ 63.682179] [<c102d2ac>] ? >> >> check_events+0x8/0xc [ 63.682179] [<c102d2a3>] ? >> >> xen_restore_fl_direct_end+0x0/0x1 [ 63.682179] [<c14d1471>] ? >> >> _spin_unlock_irqrestore+0x2f/0x31 [ 63.682179] [<c105eb9e>] ? >> >> try_to_wake_up+0x2fa/0x304 [ 63.682179] [<c102cb3f>] ? >> >> xen_force_evtchn_callback+0xf/0x14 [ 63.682179] [<c10a0cc4>] >> >> handle_IRQ_event+0x5f/0x122 [ 63.682179] [<c10a233f>] >> >> handle_level_irq+0x58/0xa9 [ 63.682179] [<c121983d>] >> >> __xen_evtchn_do_upcall+0xab/0x131 [ 63.682179] [<c1219c71>] >> >> xen_evtchn_do_upcall+0x20/0x30 [ 63.682179] [<c102ffe7>] >> >> xen_do_upcall+0x7/0xc [ 63.682179] [<c102007b>] ? >> >> apic_reg_write+0xa5/0x52f [ 63.682179] [<c1002227>] ? >> >> hypercall_page+0x227/0x1005 [ 63.682179] [<c102cb3f>] ? >> >> xen_force_evtchn_callback+0xf/0x14 [ 63.682179] [<c102d2ac>] >> >> check_events+0x8/0xc [ 63.682179] [<c102d26b>] ? >> >> xen_irq_enable_direct_end+0x0/0x1 [ 63.682179] [<c105cd0d>] ? >> >> finish_task_switch+0x4f/0xa6 [ 63.682179] [<c14d0186>] >> >> schedule+0x7dd/0x861 [ 63.682179] [<c106dd9d>] ? >> >> __mod_timer+0x135/0x140 [ 63.682179] [<c102d2ac>] ? >> >> check_events+0x8/0xc [ 63.682179] [<c102d2a3>] ? >> >> xen_restore_fl_direct_end+0x0/0x1 [ 63.682179] [<c14d1471>] ? >> >> _spin_unlock_irqrestore+0x2f/0x31 [ 63.682179] [<c107750b>] ? >> >> prepare_to_wait+0x43/0x48 [ 63.682179] [<c10742a2>] >> >> worker_thread+0x94/0x1d2 [ 63.682179] [<c10ccd2d>] ? >> >> vmstat_update+0x0/0x2f [ 63.682179] [<c1077357>] ? >> >> autoremove_wake_function+0x0/0x33 [ 63.682179] [<c107420e>] ? >> >> worker_thread+0x0/0x1d2 [ 63.682179] [<c1077120>] >> >> kthread+0x61/0x66 [ 63.682179] [<c10770bf>] ? kthread+0x0/0x66 [ >> >> 63.682179] [<c102ff97>] kernel_thread_helper+0x7/0x10 >> > > -- Sent from my Android phone with K-9 Mail. Please excuse my brevity. _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxxxxxxxx http://lists.xensource.com/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |