[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] Xen 4.2.2 / KVM / VirtualBox benchmark on Haswell
> > IMO it's a bad thing because it's far from a representative > benchmark, > > which can lead to wrong conclusions when evaluation I/O performance. > > Ancient doesn't mean non-representative. A good file-system benchmark In this particular case it is: PostMark is a single-threaded application that performs read and write operations on a fixed set of files, at an unrealistically low directory depth; modern I/O workloads exhibit much more complicated behaviour than this. > is a tricky one to come up with because most FS-es are good at some > things and bad at others. If you really want to test the virtualization > overhead on FS I/O, the only sane way to test it is by putting the FS > on the host's RAM disk and testing from there. That should expose the > full extent of the overhead, subject to the same caveat about > different FS-es being better at different load types. > > Personally I'm in favour of redneck-benchmarks that easily push the > whole stack to saturation point (e.g. highly parallel kernel > compile) since those cannot be cheated. But generically speaking, the > only way to get a worthwhile measure is to create a custom benchmark > that tests your specific application to saturation point. Any > generic/synthetic benchmark will provide results that are almost > certainly going to be misleading for any specific real-world load you > are planning to run on your system. > > For example, on a read-only MySQL load (read-only because it > simplified testing, no need to rebuild huge data sets between runs, > just drop all the caches), in custom application performance test that > I carried out for a client, ESX showed a ~40% throughput degradation > over bare metal (8 cores/server, 16 SQL threads cat-ing select- > filtered general-log extracts, load generator running in same VM). And > the test machines (both physical and virtual had enough RAM in them > that they were both only disk I/O bound for the first 2-3 minutes of > the test (which took the best part of an hour to complete); which goes > to show that disk I/O bottlenecks are good at covering up overheads > elsewhere. > > Gordan _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxx http://lists.xen.org/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |