[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] RE: [Xen-users] Re: Xen Disk I/O performance vs native performance: Xen I/O is definitely super super super slow
>-----Message d'origine----- >De : Fajar A. Nugraha [mailto:fajar@xxxxxxxxx] >Envoyé : jeudi 12 février 2009 15:03 >À : DOGUET Emmanuel >Cc : xen-users@xxxxxxxxxxxxxxxxxxx >Objet : Re: [Xen-users] Re: Xen Disk I/O performance vs native >performance: Xen I/O is definitely super super super slow > >On Thu, Feb 12, 2009 at 8:37 PM, DOGUET Emmanuel ><Emmanuel.DOGUET@xxxxxxxx> wrote: >> >> Oops sorry! >> >> We use only phy: with LVM. PV only (Linux on domU,Linux form dom0). >> LVM is on hardware RAID. > >That's better :) Now for more questions : >What kind of test did you run? How did you determine that "domU was 2x >slower than dom0"? >How much memory did you assign to domU and dom0? Are other programs >running? What were the results (how many seconds, how many MBps, etc.) With : - Oracle (Create table space : time to create and iostat) and - dd if=/dev/zero of=TEST bs=4k count=1250000 (5Gb for avoid mem cache). New platform with : Dom0: 4Go (Quad Core) DomU1: 4Go 2 VCPUS DomU2: 10Go 4 VCPUS Trying two with only one DomU. this problem is only with 2 platform. Example with configuration with 2 RAID (HP ML370, 32 bits): dom0: 5120000000 bytes (5.1 GB) copied, 139.492 seconds, 36.7 MB/s domU 5120000000 bytes (5.1 GB) copied, 279.251 seconds, 18.3 MB/s release : 2.6.18-53.1.21.el5xen version : #1 SMP Wed May 7 09:10:58 EDT 2008 machine : i686 nr_cpus : 4 nr_nodes : 1 sockets_per_node : 2 cores_per_socket : 1 threads_per_core : 2 cpu_mhz : 3051 hw_caps : bfebfbff:00000000:00000000:00000080:00004400 total_memory : 4863 free_memory : 1 xen_major : 3 xen_minor : 1 xen_extra : .0-53.1.21.el5 xen_caps : xen-3.0-x86_32p xen_pagesize : 4096 platform_params : virt_start=0xf5800000 xen_changeset : unavailable cc_compiler : gcc version 4.1.2 20070626 (Red Hat 4.1.2-14) cc_compile_by : brewbuilder cc_compile_domain : build.redhat.com cc_compile_date : Wed May 7 08:39:04 EDT 2008 xend_config_format : 2 Example with configuration with 1 RAID (HP DL360, 64bits): dom0: 5120000000 bytes (5.1 GB) copied, 170.3 seconds, 30.1 MB/s domU: 5120000000 bytes (5.1 GB) copied, 666.184 seconds, 7.7 MB/s release : 2.6.18-128.el5xen version : #1 SMP Wed Dec 17 12:01:40 EST 2008 machine : x86_64 nr_cpus : 8 nr_nodes : 1 sockets_per_node : 2 cores_per_socket : 4 threads_per_core : 1 cpu_mhz : 2666 hw_caps : bfebfbff:20000800:00000000:00000140:000ce3bd:00000000:00000001 total_memory : 18429 free_memory : 0 node_to_cpu : node0:0-7 xen_major : 3 xen_minor : 1 xen_extra : .2-128.el5 xen_caps : xen-3.0-x86_64 xen-3.0-x86_32p xen_pagesize : 4096 platform_params : virt_start=0xffff800000000000 xen_changeset : unavailable cc_compiler : gcc version 4.1.2 20080704 (Red Hat 4.1.2-44) cc_compile_by : mockbuild cc_compile_domain : redhat.com cc_compile_date : Wed Dec 17 11:37:15 EST 2008 xend_config_format : 2 PS: I don't use virt-install but generate myself the xmdomain.cfg. So PV or HVM???? Bye bye. > >I've had good results so far, with domU's disk I/O performance is >similar or equal to dom0. A simple > >time dd if=/dev/zero of=test1G bs=1M count=1024 > >took about 5 seconds and give me about 200 MB/s on idle dom0 and domU. >This is on IBM, hardware RAID, 7x144GB RAID5 + 1 hot spare 2.5" SAS >disk. Both dom0 and domU has 512MB memory. > >> >> For the RAID my question was (I'm bad in English): >> >> It's better to have : >> >> *case 1* >> Dom0 and DomU on hard-drive 1 (with HP raid: c0d0) >> >> Or >> >> *case 2* >> Dom0 on hard-drive 1 (if HP raid: c0d0) >> DomU on hard-drive 2 (if HP raid: c0d1) >> >> > >Depending on how you use it, it might not matter :) >General rule-of-thumb, more disks should provide higher I/O throughput >when setup properly. In general (like when all disks are the same, for >general-purpose domUs) I'd simply put all available disks in a RAID5 >(or multiple RAID5s for lots of disks) and put them all in a single >VG. > >Regards, > >Fajar > _______________________________________________ Xen-users mailing list Xen-users@xxxxxxxxxxxxxxxxxxx http://lists.xensource.com/xen-users
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |