[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

RE: [Xen-devel] [PATCH] Add a timer mode that disables pending missed ticks



Anyone make measurements on the final patch?

I just ran a 64-bit RHEL5.1 pvm kernel and saw a loss of about 0.2% with no 
load.  This was xen-unstable tip today with no options specified.  32-bit was 
about 0.01%.

I think I missed something... how do I run the various accounting choices and 
which ones are known to be appropriate for which kernels?

Thanks,
Dan

> -----Original Message-----
> From: xen-devel-bounces@xxxxxxxxxxxxxxxxxxx
> [mailto:xen-devel-bounces@xxxxxxxxxxxxxxxxxxx]On Behalf Of Keir Fraser
> Sent: Thursday, December 06, 2007 4:57 AM
> To: Dave Winchell
> Cc: Shan, Haitao; xen-devel@xxxxxxxxxxxxxxxxxxx; Dong, Eddie; Jiang,
> Yunhong
> Subject: Re: [Xen-devel] [PATCH] Add a timer mode that 
> disables pending
> missed ticks
> 
> 
> Please take a look at xen-unstable changeset 16545.
> 
>  -- Keir
> 
> On 26/11/07 20:57, "Dave Winchell" <dwinchell@xxxxxxxxxxxxxxx> wrote:
> 
> > Keir,
> >
> > The accuracy data I've collected for i/o loads for the
> > various time protocols follows. In addition, the data
> > for cpu loads is shown.
> >
> > The loads labeled cpu and i/o-8 are on an 8 processor AMD box.
> > Two guests, red hat and sles 64 bit, 8 vcpu each.
> > The cpu load is usex -e36 on each guest.
> > (usex is available at http://people.redhat.com/anderson/usex.)
> > i/o load is 8 instances of dd if=/dev/hda6 of=/dev/null.
> >
> > The loads labeled i/o-32 are 32 instances of dd.
> > Also, these are run on 4 cpu AMD box.
> > In addition, there is an idle rh-32bit guest.
> > All three guests are 8vcpu.
> >
> > The loads labeled i/o-4/32 are the same as i/o-32
> > except that the redhat-64 guest has 4 instances of dd.
> >
> > Date Duration Protocol sles, rhat error load
> >
> > 11/07 23 hrs 40 min ASYNC -4.96 sec, +4.42 sec -.006%, +.005% cpu
> > 11/09 3 hrs 19 min ASYNC -.13 sec, +1.44 sec, -.001%, +.012% cpu
> >
> > 11/08 2 hrs 21 min SYNC -.80 sec, -.34 sec, -.009%, -.004% cpu
> > 11/08 1 hr 25 min SYNC -.24 sec, -.26 sec, -.005%, -.005% cpu
> > 11/12 65 hrs 40 min SYNC -18 sec, -8 sec, -.008%, -.003% cpu
> >
> > 11/08 28 min MIXED -.75 sec, -.67 sec -.045%, -.040% cpu
> > 11/08 15 hrs 39 min MIXED -19. sec,-17.4 sec, -.034%, -.031% cpu
> >
> >
> > 11/14 17 hrs 17 min ASYNC -6.1 sec,-55.7 sec, -.01%, -.09% i/o-8
> > 11/15 2 hrs 44 min ASYNC -1.47 sec,-14.0 sec, -.015% -.14% i/o-8
> >
> > 11/13 15 hrs 38 min SYNC -9.7 sec,-12.3 sec, -.017%, -.022% i/o-8
> > 11/14 48 min SYNC - .46 sec, - .48 sec, -.017%, -.018% i/o-8
> >
> > 11/14 4 hrs 2 min MIXED -2.9 sec, -4.15 sec, -.020%, -.029% i/o-8
> > 11/20 16 hrs 2 min MIXED -13.4 sec,-18.1 sec, -.023%, -.031% i/o-8
> >
> >
> >
> > 11/21 28 min MIXED -2.01 sec, -.67 sec, -.12%, -.04% i/o-32
> > 11/21 2 hrs 25 min SYNC -.96 sec, -.43 sec, -.011%, -.005% i/o-32
> > 11/21 40 min ASYNC -2.43 sec, -2.77 sec -.10%, -.11% i/o-32
> >
> > 11/26 113 hrs 46 min MIXED -297. sec, 13. sec -.07%, .003% i/o-4/32
> > 11/26 4 hrs 50 min SYNC -3.21 sec, 1.44 sec, -.017%, .01% i/o-4/32
> >
> >
> > Overhead measurements:
> >
> > Progress in terms of number of passes through a fixed 
> system workload
> > on an 8 vcpu red hat with an 8 vcpu sles idle.
> > The workload was usex -b48.
> >
> >
> > ASYNC 167 min 145 passes .868 passes/min
> > SYNC 167 min 144 passes .862 passes/min
> > SYNC 1065 min 919 passes .863 passes/min
> > MIXED 221 min 196 passes .887 passes/min
> >
> >
> > Conclusions:
> >
> > The only protocol which meets the .05% accuracy requirement for ntp
> > tracking under the loads
> > above is the SYNC protocol. The worst case accuracies for 
> SYNC, MIXED,
> > and ASYNC
> > are .022%, .12%, and .14%, respectively.
> >
> > We could reduce the cost of the SYNC method by only 
> scheduling the extra
> > wakeups if a certain number
> > of ticks are missed.
> >
> > Regards,
> > Dave
> >
> > Keir Fraser wrote:
> >
> >> On 9/11/07 19:22, "Dave Winchell" 
> <dwinchell@xxxxxxxxxxxxxxx> wrote:
> >>
> >>
> >>
> >>> Since I had a high error (~.03%) for the ASYNC method a 
> couple of days ago,
> >>> I ran another ASYNC test. I think there may have been something
> >>> wrong with the code I used a couple of days ago for 
> ASYNC. It may have been
> >>> missing the immediate delivery of interrupt after context 
> switch in.
> >>>
> >>> My results indicate that either SYNC or ASYNC give 
> acceptable accuracy,
> >>> each running consistently around or under .01%. MIXED has 
> a fairly high
> >>> error of
> >>> greater than .03%. Probably too close to .05% ntp 
> threshold for comfort.
> >>> I don't have an overnight run with SYNC. I plan to leave 
> SYNC running
> >>> over the weekend. If you'd rather I can leave MIXED 
> running instead.
> >>>
> >>> It may be too early to pick the protocol and I can run 
> more overnight tests
> >>> next week.
> >>>
> >>>
> >>
> >> I'm a bit worried about any unwanted side effects of the 
> SYNC+run_timer
> >> approach -- e.g., whether timer wakeups will cause higher 
> system-wide CPU
> >> contention. I find it easier to think through the 
> implications of ASYNC. I'm
> >> surprised that MIXED loses time, and is less accurate than 
> ASYNC. Perhaps it
> >> delivers more timer interrupts than the other approaches, 
> and each interrupt
> >> event causes a small accumulated error?
> >>
> >> Overall I would consider MIXED and ASYNC as favourites and 
> if the latter is
> >> actually more accurate then I can simply revert the changeset that
> >> implemented MIXED.
> >>
> >> Perhaps rather than running more of the same workloads you 
> could try idle
> >> VCPUs and I/O bound VCPUs (e.g., repeated large disc reads 
> to /dev/null)? We
> >> don't have any data on workloads that aren't CPU bound, so 
> that's really an
> >> obvious place to put any further effort imo.
> >>
> >> -- Keir
> >>
> >>
> >>
> >>
> >
> 
> 
> 
> _______________________________________________
> Xen-devel mailing list
> Xen-devel@xxxxxxxxxxxxxxxxxxx
> http://lists.xensource.com/xen-devel
> 
>


_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-devel


 


Rackspace

Lists.xenproject.org is hosted with RackSpace, monitoring our
servers 24x7x365 and backed by RackSpace's Fanatical Support®.