[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: [Xen-devel] [PATCH] turn off writable page tables

  • To: "Andrew Theurer" <habanero@xxxxxxxxxx>
  • From: "Jacob Gorm Hansen" <jacobg@xxxxxxx>
  • Date: Tue, 25 Jul 2006 22:31:37 -0700
  • Cc: Ian Pratt <m+Ian.Pratt@xxxxxxxxxxxx>, xen-devel@xxxxxxxxxxxxxxxxxxx
  • Delivery-date: Tue, 25 Jul 2006 22:31:59 -0700
  • Domainkey-signature: a=rsa-sha1; q=dns; c=nofws; s=beta; d=gmail.com; h=received:message-id:date:from:sender:to:subject:cc:in-reply-to:mime-version:content-type:content-transfer-encoding:content-disposition:references:x-google-sender-auth; b=IXMSxAKutswBu6tZEFTCQD2m7ctPqKCazRvjK3irYPVnO2Adgl2MK1ERHsQiN6kVI2Xg81ucHZOcoi9Ydd6hLrFy0KEFbb0emItapYT+dQKSnRjH32DIvaA1m9MReq9kRoiyeHbF+cY5UPcBaz79C7rI/nFEgy6TVqHuysjfKJw=
  • List-id: Xen developer discussion <xen-devel.lists.xensource.com>

On 7/25/06, Andrew Theurer <habanero@xxxxxxxxxx> wrote:
Ian Pratt wrote:
>> on Xeon MP processor, uniprocessor dom0 kernel, pae=y:
>> benchmark                c/s 10729 force_emulate
>> ------------------------ --------- -------------
>> lmbench fork+exit:       469.5833  470.3913   usec, lower is better
>> lmbench fork+execve:     1241.0000 1225.7778  usec, lower is better
>> lmbench fork+/sbin/bash: 12190.000 12119.000  usec, lower is better
> It's kinda weird that these scores are so close -- I guess its just
> coincidence that we must be getting something like an average of 10-20
> pte's updated per pagetable page and the cost of doing multiple emulates
> perfectly balances the cost of unhooking/rehooking.

Just a silly question; is the old batched update mechanism totally out
of the picture here? Is it the cost of taking additional faults that
makes writable ptes as slow as emulation (which I suppose just means
shadow p.t.s)? Is there tension between shadow pt cache size inside
Xen and runtime performance?


Xen-devel mailing list



Lists.xenproject.org is hosted with RackSpace, monitoring our
servers 24x7x365 and backed by RackSpace's Fanatical Support®.