[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: [Xen-users] My future plan



Hi Jonathan,

I use a DRBD bases IET install. It syncs between the nodes with two bonded 
Intel e1000 NICs. I use the same network cards to connect to the Xen 
hypervisors. 
MIND YOU: I use dual port NICs  (two in total on the storga servers) but I 
CROSS the connections: that is, I connect one port of one card to the Xen 
nodes, but I use the other for the DRBD sync; And the other way around of 
course. This way, if a card breaks, I still have things running. To be able to 
use two switches in between Xen hosts and the storage, I use multipathing to 
connect to the iSCSI LUNs. This results in higher speed and network 
redundancy. It would make no sense to use more than 2 ports since DRBD cannot 
sync faster, but also, as mentioned before, it seems that bonding more than 2 
does not result in higher speeds. 
This however is easily tested with netperf . I would be happy to hear someones 
testresults about this.

O yes, if you don't get the expected speeds with bonded cards in mode 0, try 
looking at tcp_reordering in /proc/sys/net/ipv4 something ...


On Wednesday 09 June 2010 14:53:28 Jonathan Tripathy wrote:
> So should I just go with 2 NICs for the storage server then?
> 
> In your future setup, how many NICs are you using for the storage server
>  and how many for the nodes? I take it you're using software iSCSI?
> 
> ________________________________
> 
> From: Bart Coninckx [mailto:bart.coninckx@xxxxxxxxxx]
> Sent: Wed 09/06/2010 11:25
> To: xen-users@xxxxxxxxxxxxxxxxxxx
> Cc: Jonathan Tripathy; Michael Schmidt
> Subject: Re: [Xen-users] My future plan
> 
> 
> 
> On the DRBD mailing lists I've seen a couple of times that they did tests
>  with bonding and they claim that a bond with more than 2 NICs will
>  actually decrease performance because of the TCP reordering that needs to
>  be done.
> 
> That's the reason why I limit the storage connection to two NICs. I have a
> very similar to yours in the making by the way.
> 
> On Tuesday 08 June 2010 15:55:47 Jonathan Tripathy wrote:
> > Hi Michael,
> >
> > Thanks for the tips using SSD for the node OS drives.
> >
> > Regarding the NIC, I was thinking about using this for the nodes:
> >
> > http://www.intel.com/products/server/adapters/pro1000pt-dualport/pro1000p
> >t- dualport-overview.htm
> >
> > and this for the server:
> >
> > http://www.intel.com/products/server/adapters/pro1000pt-quadport-low-prof
> >il e/pro1000pt-quadport-low-profile-overview.htm
> >
> > Are those the cards you were talking about? They are very cheap on ebay
> > you see...
> >
> > Think 4 port bonding for the server is good enough for 8 nodes?
> >
> > Thanks
> >
> > ________________________________
> >
> > From: Michael Schmidt [mailto:michael.schmidt@xxxxxxxxxx]
> > Sent: Tue 08/06/2010 14:49
> > To: Jonathan Tripathy; Xen-users@xxxxxxxxxxxxxxxxxxx
> > Subject: Re: [Xen-users] My future plan
> >
> >
> > Hi Jonathan,
> >
> > you should think about flash or SD cards as xen-boot-drive.
> > This provides you lower costs and higher energy efficiency.
> > If you mount /tmp and /var/log to an tmpfs, this disks works very well
> > and long.
> >
> > If you dont need so much disk space for your storage, use sas disks.
> > SAS (10k/15k) disks provides you many more IOPs than sata disks (more
> > IOPS per $/EUR as well). And very important: A very large cache for your
> > raid controller.
> >
> > Intel e1000e is a pretty good choice. This cards have a large buffer and
> >  generates just a few interrupts on your CPUs (in comparison to the
> >  Broadcom NICs).
> >
> > Best Regards
> >
> > Michael Schmidt
> > Am 08.06.10 14:55, schrieb Jonathan Tripathy:
> >
> >       My future plan currently looks like this for my VPS hosting
> > solution, so any feedback would be appreciated:
> >
> >       Each Node:
> >       Dell R210 Intel X3430 Quad Core 8GB RAM
> >       Intel PT 1Gbps Server Dual Port NIC using linux "bonding"
> >       Small pair of HDDs for OS (Probably in RAID1)
> >       Each node will run about 10 - 15 customer guests
> >
> >
> >       Storage Server:
> >       Some Intel Quad Core Chip
> >       2GB RAM (Maybe more?)
> >       LSI 8704EM2 RAID Controller (Think this controller does 3 Gbps)
> >       Battery backup for the above RAID controller
> >       4 X RAID10 Arrays (4 X 1.5TB disks per array, 16 disks in total)
> >       Each RAID10 array will connect to 2 nodes (8 nodes per storage
> > server) Intel PT 1Gbps Quad port NIC using Linux bonding
> >       Exposes 8 X 1.5GB iSCSI targets (each node will use one of these)
> >
> >       HP Procurve 1800-24G switch to create 1 X 4 port trunk (for storage
> >  server), and 8 X 2 port trunk (for the nodes)
> >
> >       What you think? Any tips?
> >
> >       Thanks
> >
> >
> >       _______________________________________________
> >       Xen-users mailing list
> >       Xen-users@xxxxxxxxxxxxxxxxxxx
> >       http://lists.xensource.com/xen-users
> 

_______________________________________________
Xen-users mailing list
Xen-users@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-users


 


Rackspace

Lists.xenproject.org is hosted with RackSpace, monitoring our
servers 24x7x365 and backed by RackSpace's Fanatical Support®.