[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: [Xen-users] Xen 3.4.3 - bad IO performance on drbd devices



On Fri, Jul 30, 2010 at 12:28:58PM +0200, Felix Botner wrote:
> Hi
> 
> some more information about the xen drbd performance problem (two servers, 
> each server has two drbd devices (protocol c, ext4) and is primary for one of 
> them). Any test without memory restriction for dom0. The network connection 
> between the drbd servers limits the troughput to ~230 MB for each drbd device 
> (2 x 1000 Mbps cards, bond mode 0). 
> 
> The local drive is a hardware raid10 with 12 300gb (3gbps 10k SAS) harddiscs. 
> The kernel is the latest debian/squeeze based kernel 2.6.32-5-amd64 (with 
> http://svn.debian.org/wsvn/kernel/?op=comp&compare[]=%2Fdists%2Fsid@15998&compare[]=%2Fdists%2Fsid@16001)
> 
> bonnie on the local drive with hypevisor
> ~ 400 MB/sec
> 
> bonnie on the connected drbd device with hypevisor
> ~ 170 MB/sec (bonnie on one drbd device/server)
> ~ 110 MB/sec (bonnie on both drbd devices/servers)
> 
> bonnie on the connected drbd device without hypevisor
> ~ 230 MB/sec (bonnie on one drbd device/server)
> ~ 200 MB/sec (bonnie on both drbd devices/servers)
> 
> bonnie on the disconnected drbd device with hypevisor
> ~ 300 MB/sec
> 
> bonnie on the disconnected drbd device without hypevisor
> ~ 360 MB/sec
> 
> What interests me is the throughput when writing to both servers. With the xen
> hypevisor i get 110 MB/sec on each drbd device (220 MB/sec io throughput on 
> each server because drbd writes "locale" and "remote"). Without the 
> hypervisor i get 200 MB/sec on each drbd device (400 MB/sec io throughput on 
> each server, the maximum of what the io backend allows).
> 
> But even with only one drbd device the io is without the hypevisor much 
> better 
> (230 MB/sec without and 170 MB/sec with hypervisor)
> 
> The strange thing is, a drbd resync gives me with or without hypervisor ~ 230 
> MB/sec. And when i start one server without hypervisor, bonnie gives me ~ 230 
> MB/sec on that server and on the remote server with the hypervisor (because 
> drbd also writes remote).
> 
> Any hints?
> 

I guess DRBD performance mainly depends on the latency you have,
because of the synchronization needed between the hosts.

Have you tried pinning the dom0 vcpus?

-- Pasi


> Thanks,
> Felix
> 
> xm info
> host                   : samla
> release                : 2.6.32-ucs9-xen-amd64
> version                : #1 SMP Thu Jul 22 04:32:22 UTC 2010
> machine                : x86_64
> nr_cpus                : 8
> nr_nodes               : 1
> cores_per_socket       : 4
> threads_per_core       : 1
> cpu_mhz                : 2133
> hw_caps                : 
> bfebfbff:28100800:00000000:00000340:009ce3bd:00000000:00000001:00000000
> virt_caps              : hvm
> total_memory           : 24564
> free_memory            : 22230
> node_to_cpu            : node0:0-7
> node_to_memory         : node0:22230
> xen_major              : 3
> xen_minor              : 4
> xen_extra              : .3
> xen_caps               : xen-3.0-x86_64 xen-3.0-x86_32p hvm-3.0-x86_32 
> hvm-3.0-x86_32p hvm-3.0-x86_64
> xen_scheduler          : credit
> xen_pagesize           : 4096
> platform_params        : virt_start=0xffff800000000000
> xen_changeset          : unavailable
> cc_compiler            : gcc version 4.3.2 (Debian 4.3.2-1.1.13.200909082302)
> cc_compile_by          : root
> cc_compile_domain      : [unknown]
> cc_compile_date        : Thu Jun 17 14:54:34 UTC 2010
> xend_config_format     : 4
> 
> Am Dienstag 20 Juli 2010 15:54:21 schrieb Felix Botner:
> > Am Dienstag 20 Juli 2010 14:35:30 schrieb Pasi Kärkkäinen:
> > > On Tue, Jul 20, 2010 at 10:46:34AM +0200, Felix Botner wrote:
> > > > Hi everyone,
> > > >
> > > > i have two servers installed with a debian/lenny based os (64bit), a
> > > > debian/sid based kernel 2.6.32-xen-amd64 and xen 3.4.3-4. Each server
> > > > has two drbd devices (protocol c, formatted with ext4) and is primary
> > > > for one of them. Every drbd pair has a dedicated network interface (a
> > > > bond mode 0 interface with two 1000 Mbps cards).
> > >
> > > <snip>
> > >
> > > > The io performance on the connected drbd devices is significantly worse
> > > > if i start the kernel with the xen hypervisor (with "kernel
> > > > /boot/xen-3.4.3.gz"). Without the hypervisor (but the same kernel) the
> > > > systems are about 50% faster.
> > >
> > > Are you measuring from dom0 or from a guest?
> >
> > From dom0, there are no guests at the moment.
> >
> > > > Why is there such a difference?
> > > > Can i optimize my xend (i already added dom0_mem=2048M dom0_max_vcpus=2
> > > > dom0_vcpus_pin as boot parameter with no effect)?
> > > > Are there any known issues using XEN and bonding/drbd?
> > > >
> > > > Feel free to ask for more information about the system or the setup.
> > >
> > > How much memory does your server have?
> > > ie. how much ram do you have when you boot it baremetal without Xen.
> >
> > ~20GB without Xen. Now i removed the hypervisor param dom0_mem=2048M from
> > menu.lst (in xend-config set (dom0-min-mem 196) and (enable-dom0-ballooning
> > yes)), rebooted and as far as i know there should be no memory restriction
> > for dom0. "xm list" shows me the complete memory for dom0:
> >
> > server1-> xm list
> > Name                                        ID   Mem VCPUs      State
> > Time(s)
> > Domain-0                                     0 18323     2     r-----  
> > 1785.0
> >
> > server2-> xm list
> > Name                                        ID   Mem VCPUs      State
> > Time(s)
> > Domain-0                                     0 22375     2     r-----  
> > 1754.2
> >
> > But bonnie++ gives me still bad results:
> >
> > server1,60000M,,,113444,30,55569,13,,,231555,23,622.6,0,16,25634,70,+++++,+
> >++, +++++,+++,+++++,+++,+++++,+++,+++++,+++
> > server2,60000M,,,114014,31,53864,13,,,243541,27,617.4,0,16,+++++,+++,+++++,
> > +++,+++++,+++,+++++,+++,+++++,+++,+++++,+++
> >
> > So i don't think caching is the issue, or?
> >
> > bye
> >
> > Felix
> >
> > > Remember all the free memory of the host/dom0 will be used by linux
> > > pagecache.. So if you limit dom0 to 2GB, it'll have less cache than the
> > > baremetal case.
> > >
> > > -- Pasi
> 
> 
> -- 
> Felix Botner
> 
> Open Source Software Engineer
> 
> Univention GmbH
> Linux for your business
> Mary-Somerville-Str.1
> 28359 Bremen
> Tel. : +49 421 22232-0
> Fax : +49 421 22232-99
> 
> botner@xxxxxxxxxxxxx>
> http://www.univention.de
> 
> Geschäftsführer: Peter H. Ganten
> HRB 20755 Amtsgericht Bremen
> Steuer-Nr.: 71-597-02876 
> 
> _______________________________________________
> Xen-users mailing list
> Xen-users@xxxxxxxxxxxxxxxxxxx
> http://lists.xensource.com/xen-users

_______________________________________________
Xen-users mailing list
Xen-users@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-users


 


Rackspace

Lists.xenproject.org is hosted with RackSpace, monitoring our
servers 24x7x365 and backed by RackSpace's Fanatical Support®.