[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: [Xen-users] network CPU consumption question



On Fri, 2 Jun 2006 10:41:31 -0500
Tim Freeman <tfreeman@xxxxxxxxxxx> wrote:

> On Fri, 2 Jun 2006 08:51:14 -0400
> "Tim Wood" <twwood@xxxxxxxxx> wrote:
> 
> > Sorry I don't have full concrete answers, but you might be able to get some
> > ideas from either of these papers if you haven't seen them:
> > 
> > XenMon - http://www.hpl.hp.com/techreports/2005/HPL-2005-187.html
> > Measuring CPU Overhead for IO -
> > http://www.hpl.hp.com/personal/Lucy_Cherkasova/projects/papers/final-perf-study-usenix.pdf
> > 
> > They show how the scheduling parameters can have a pretty large effect on
> > performance mainly due to wasted context switches between domains.
> 
> Thanks, I am aware of both the papers.  Since I have the XenMon output
> I looked at the ex/s and did not see anything to explain the leap with
> context switching.  The two I/O bound guests together are 'charged'
> roughly 19% of the CPU usage to process the same amt of packets as the
> one I/O guest could process with roughly 11.5%. 
> 
> It's my assumption that context switching between any two domains
> results in CPU usage being increased for both of those domains.  Isn't
> that true?  (I remember being roughly convinced of this by data at one
> point, but it would take me some time to find that). 
> 
> If true, in the situation I reported I think it would mean that the two
> I/O guest domains (dom1/dom2) in scenario 2 would together have to have
> had a relative increase in ex/s over dom0 to account for the extra CPU
> usage seen for dom1/dom2 but not dom0.  (i.e., maybe dom1 and dom2 are
> being switched between themselves far more often and that is causing
> extra load for just these two domains). 
> 
> But in scenario one, dom0 and dom1 read ~1300 each, in scenario two
> dom0, dom1 and dom2 each read ~380.  So it's far less in fact...
> (totals including cpu domain: scenario 1 = ~4000 ex/s, scenario 2 =
> ~1900 ex/s). 

And this is intuititive to me, the more domains that are running (and
assuming somewhat fair scheduling parameters) the less frequently they
have a chance for the CPU and the more virtual interrupts are stored
up, i.e., more work to do = domain doesn't block as quickly (when it's
run out of work).  Less blocking means less context switching given the
same workloads.  In the data, the overall blocked % does indeed
plummet in scenario 2. 

Tim 

> 
> Thanks,
> Tim
> 
> 
> >
> > On 5/31/06, Tim Freeman <tfreeman@xxxxxxxxxxx> wrote:
> > >
> > > Running into a mysterious situation here...
> > >
> > > I have a Xen3 setup running SEDF measuring bandwidth offnode on a
> > > non-VMM machine. Running one guest domain bridged through domain
> > > 0, transfers from memory (/dev/zero) to the non-VMM machine.  Another
> > > domain runs pure CPU bound activity, as much as it can (this avoids the
> > > non-interrupt-batching context-switching-heavy situation that exists
> > > otherwise).
> > >
> > > On these computers, getting roughly 12.3 MB/s with this distribution
> > > of CPU consumption (measured with XenMon):
> > >
> > > dom0: 17.79%, dom1 (net): 11.41%, dom2 (cpu): 69.48%
> > >
> > > (these are means from many runs)
> > >
> > > I get an extremely similar bandwidth reading in another scenario with
> > > two net transferring domains (their bandwidths are halved, talking
> > > about the *sum* of the two).  However, the CPU consumption did not
> > > scale very well:
> > >
> > > dom0: 19.38%, dom1 (net): 9.61%, dom2 (net): 9.55%, dom3 (cpu): 60.18%
> > >
> > > In the previous scenario dom1 used around 11.5 percent of the CPU to
> > > accomplish what here takes roughly 19 percent.
> > >
> > > Any ideas what this extra CPU work is?  There is extra context
> > > switching in the second situation (which could account for the dom0
> > > increase), but the network bound guest disparity seems a little extreme
> > > to only be that?
> > >
> > > Thanks,
> > > Tim
> > >
> > >
> > > _______________________________________________
> > > Xen-users mailing list
> > > Xen-users@xxxxxxxxxxxxxxxxxxx
> > > http://lists.xensource.com/xen-users
> > >
> > 
> 
> _______________________________________________
> Xen-users mailing list
> Xen-users@xxxxxxxxxxxxxxxxxxx
> http://lists.xensource.com/xen-users
> 

_______________________________________________
Xen-users mailing list
Xen-users@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-users


 


Rackspace

Lists.xenproject.org is hosted with RackSpace, monitoring our
servers 24x7x365 and backed by RackSpace's Fanatical Support®.