[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: [Xen-devel] [PATCH v4] x86/apicv: fix RTC periodic timer and apicv issue



sent too quick. I meant please add your tested-by here. :-)

> From: Tian, Kevin
> Sent: Thursday, December 22, 2016 3:48 PM
> 
> Thanks a lot!
> 
> > From: Gao, Chao
> > Sent: Thursday, December 22, 2016 4:02 AM
> >
> > Hi, xuquan.
> > I have tested it on my skylake server. W/o this patch the inaccurate
> > wall clock time issue only exists in Win7-32 guest. Win7-64, Win8-32, 
> > Win8-64,
> > Win10-32 ,Win10-64 and linux-4.8.0+ guests don't have this issue.
> > W/ this v4 patch, the issue disappears in Win7-32 guest and no wall lock 
> > time
> > related regression is found on Win7-64, Win8-32, Win8-64, Win10-32, Win10-64
> > and linux-4.8.0+ guest.
> >
> > In windows guest, the test procedure is
> > 1. Create a windows guest with 2 vCPU
> > 2. run the following .bat in guest
> >     :abcd
> >     echo 111111
> >     goto abcd
> >
> > 3. Start a stop-watch outside the guest and monitor the clock at the lower 
> > right
> >    corner in guest. After 120 seconds according the guest clock, stop the 
> > stop-watch.
> >    If the time shows in the stop-watch is about 120 seconds, then I think
> >    there is no the above issue in the guest. Otherwise, the time is 
> > inaccurate.
> >
> > In Win7-32 case, the stop-watch time is about 70 seconds, so the clock in 
> > guest is
> > obviously inaccurate.
> >
> > In linux guest, the test procedure is
> > 1. Create a linux guest with 4 vCPU
> > 2. insmod the following linux module
> >    (through output of /proc/interrupt, about 850000 ipis in 13 seconds)
> > 3. use date command to get guest time, others are same as test in windows 
> > guest
> >
> > #include <linux/init.h>
> > #include <linux/module.h>
> > #include <linux/kthread.h>
> > #include <linux/sched.h>
> > #include <asm/delay.h>
> > MODULE_LICENSE("GPL");
> > void workload(void *info)
> > {
> >         asm volatile("nop");
> > }
> >
> > void msleep(unsigned int msecs);
> > static int ipi_generator(void * info)
> > {
> >         int i;
> >         while (!kthread_should_stop()) {
> >                 for(i=0; i< 5 * 10000; i++)
> >                 {
> >                         smp_call_function(workload, NULL,1);
> >                 }
> >                 msleep(1);
> >         }
> >         return 0;
> > }
> > struct task_struct *thread;
> > static int __init ipi_init(void)
> > {
> >         thread = kthread_run(ipi_generator, NULL, "IPI");
> >         if (IS_ERR(thread))
> >                 return PTR_ERR(thread);
> >         return 0;
> > }
> >
> > static void __exit ipi_exit(void)
> > {
> >         kthread_stop(thread);
> > }
> > module_init(ipi_init);
> > module_exit(ipi_exit);
> >
> > Are these tests sufficient? Please let me know if you have any other 
> > thoughts.
> >
> > On Wed, Dec 21, 2016 at 05:44:08AM +0000, Xuquan (Quan Xu) wrote:
> > >When Xen apicv is enabled, wall clock time is faster on Windows7-32
> > >guest with high payload (with 2vCPU, captured from xentrace, in
> > >high payload, the count of IPI interrupt increases rapidly between
> > >these vCPUs).
> > >
> > >If IPI intrrupt (vector 0xe1) and periodic timer interrupt (vector 0xd1)
> > >are both pending (index of bit set in vIRR), unfortunately, the IPI
> > >intrrupt is high priority than periodic timer interrupt. Xen updates
> > >IPI interrupt bit set in vIRR to guest interrupt status (RVI) as a high
> > >priority and apicv (Virtual-Interrupt Delivery) delivers IPI interrupt
> > >within VMX non-root operation without a VM-Exit. Within VMX non-root
> > >operation, if periodic timer interrupt index of bit is set in vIRR and
> > >highest, the apicv delivers periodic timer interrupt within VMX non-root
> > >operation as well.
> > >
> > >But in current code, if Xen doesn't update periodic timer interrupt bit
> > >set in vIRR to guest interrupt status (RVI) directly, Xen is not aware
> > >of this case to decrease the count (pending_intr_nr) of pending periodic
> > >timer interrupt, then Xen will deliver a periodic timer interrupt again.
> > >
> > >And that we update periodic timer interrupt in every VM-entry, there is
> > >a chance that already-injected instance (before EOI-induced exit happens)
> > >will incur another pending IRR setting if there is a VM-exit happens
> > >between virtual interrupt injection (vIRR->0, vISR->1) and EOI-induced
> > >exit (vISR->0), since pt_intr_post hasn't been invoked yet, then the
> > >guest receives more periodic timer interrupt.
> > >
> > >So we set eoi_exit_bitmap for intack.vector when it's higher than
> > >pending periodic time interrupts. This way we can guarantee there's
> > >always a chance to post periodic time interrupts when periodic time
> > >interrupts becomes the highest one.
> > >
> > >Signed-off-by: Quan Xu <xuquan8@xxxxxxxxxx>
> > >---
> > > xen/arch/x86/hvm/vmx/intr.c | 9 +++++++--
> > > 1 file changed, 7 insertions(+), 2 deletions(-)
> > >
> > >diff --git a/xen/arch/x86/hvm/vmx/intr.c b/xen/arch/x86/hvm/vmx/intr.c
> > >index 639a705..0cf26b4 100644
> > >--- a/xen/arch/x86/hvm/vmx/intr.c
> > >+++ b/xen/arch/x86/hvm/vmx/intr.c
> > >@@ -315,9 +315,13 @@ void vmx_intr_assist(void)
> > >         * Set eoi_exit_bitmap for periodic timer interrup to cause 
> > > EOI-induced VM
> > >         * exit, then pending periodic time interrups have the chance to 
> > > be injected
> > >         * for compensation
> > >+        * Set eoi_exit_bitmap for intack.vector when it's higher than 
> > >pending
> > >+        * periodic time interrupts. This way we can guarantee there's 
> > >always a chance
> > >+        * to post periodic time interrupts when periodic time interrupts 
> > >becomes the
> > >+        * highest one
> > >         */
> > >         if (pt_vector != -1)
> > >-            vmx_set_eoi_exit_bitmap(v, pt_vector);
> > >+            vmx_set_eoi_exit_bitmap(v, intack.vector);
> > >
> > >         /* we need update the RVI field */
> > >         __vmread(GUEST_INTR_STATUS, &status);
> > >@@ -334,7 +338,8 @@ void vmx_intr_assist(void)
> > >             __vmwrite(EOI_EXIT_BITMAP(i), 
> > > v->arch.hvm_vmx.eoi_exit_bitmap[i]);
> > >         }
> > >
> > >-        pt_intr_post(v, intack);
> > >+        if ( intack.vector == pt_vector )
> > >+            pt_intr_post(v, intack);
> > >     }
> > >     else
> > >     {
> > >--
> > >1.8.3.4

_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxx
https://lists.xen.org/xen-devel

 


Rackspace

Lists.xenproject.org is hosted with RackSpace, monitoring our
servers 24x7x365 and backed by RackSpace's Fanatical Support®.