[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: [Xen-users] My future plan




----- Original message -----
> Hi Jonathan,
>
> I use a DRBD bases IET install. It syncs between the nodes with two bonded
> Intel e1000 NICs. I use the same network cards to connect to the Xen
> hypervisors.

Correction, the same kind of cards

> MIND YOU: I use dual port NICs  (two in total on the storga servers) but I
> CROSS the connections: that is, I connect one port of one card to the Xen
> nodes, but I use the other for the DRBD sync; And the other way around of
> course. This way, if a card breaks, I still have things running. To be able to
> use two switches in between Xen hosts and the storage, I use multipathing to
> connect to the iSCSI LUNs. This results in higher speed and network
> redundancy. It would make no sense to use more than 2 ports since DRBD cannot
> sync faster, but also, as mentioned before, it seems that bonding more than 2
> does not result in higher speeds.
> This however is easily tested with netperf . I would be happy to hear someones
> testresults about this.
>
> O yes, if you don't get the expected speeds with bonded cards in mode 0, try
> looking at tcp_reordering in /proc/sys/net/ipv4 something ...
>
>
> On Wednesday 09 June 2010 14:53:28 Jonathan Tripathy wrote:
> > So should I just go with 2 NICs for the storage server then?
> >
> > In your future setup, how many NICs are you using for the storage server
> >  and how many for the nodes? I take it you're using software iSCSI?
> >
> > ________________________________
> >
> > From: Bart Coninckx [mailto:bart.coninckx@xxxxxxxxxx]
> > Sent: Wed 09/06/2010 11:25
> > To: xen-users@xxxxxxxxxxxxxxxxxxx
> > Cc: Jonathan Tripathy; Michael Schmidt
> > Subject: Re: [Xen-users] My future plan
> >
> >
> >
> > On the DRBD mailing lists I've seen a couple of times that they did tests
> >  with bonding and they claim that a bond with more than 2 NICs will
> >  actually decrease performance because of the TCP reordering that needs to
> >  be done.
> >
> > That's the reason why I limit the storage connection to two NICs. I have a
> > very similar to yours in the making by the way.
> >
> > On Tuesday 08 June 2010 15:55:47 Jonathan Tripathy wrote:
> > > Hi Michael,
> > >
> > > Thanks for the tips using SSD for the node OS drives.
> > >
> > > Regarding the NIC, I was thinking about using this for the nodes:
> > >
> > > http://www.intel.com/products/server/adapters/pro1000pt-dualport/pro1000p
> > > t- dualport-overview.htm
> > >
> > > and this for the server:
> > >
> > > http://www.intel.com/products/server/adapters/pro1000pt-quadport-low-prof
> > > il e/pro1000pt-quadport-low-profile-overview.htm
> > >
> > > Are those the cards you were talking about? They are very cheap on ebay
> > > you see...
> > >
> > > Think 4 port bonding for the server is good enough for 8 nodes?
> > >
> > > Thanks
> > >
> > > ________________________________
> > >
> > > From: Michael Schmidt [mailto:michael.schmidt@xxxxxxxxxx]
> > > Sent: Tue 08/06/2010 14:49
> > > To: Jonathan Tripathy; Xen-users@xxxxxxxxxxxxxxxxxxx
> > > Subject: Re: [Xen-users] My future plan
> > >
> > >
> > > Hi Jonathan,
> > >
> > > you should think about flash or SD cards as xen-boot-drive.
> > > This provides you lower costs and higher energy efficiency.
> > > If you mount /tmp and /var/log to an tmpfs, this disks works very well
> > > and long.
> > >
> > > If you dont need so much disk space for your storage, use sas disks.
> > > SAS (10k/15k) disks provides you many more IOPs than sata disks (more
> > > IOPS per $/EUR as well). And very important: A very large cache for your
> > > raid controller.
> > >
> > > Intel e1000e is a pretty good choice. This cards have a large buffer and
> > >  generates just a few interrupts on your CPUs (in comparison to the
> > >  Broadcom NICs).
> > >
> > > Best Regards
> > >
> > > Michael Schmidt
> > > Am 08.06.10 14:55, schrieb Jonathan Tripathy:
> > >
> > >            My future plan currently looks like this for my VPS hosting
> > > solution, so any feedback would be appreciated:
> > >
> > >            Each Node:
> > >            Dell R210 Intel X3430 Quad Core 8GB RAM
> > >            Intel PT 1Gbps Server Dual Port NIC using linux "bonding"
> > >            Small pair of HDDs for OS (Probably in RAID1)
> > >            Each node will run about 10 - 15 customer guests
> > >
> > >
> > >            Storage Server:
> > >            Some Intel Quad Core Chip
> > >            2GB RAM (Maybe more?)
> > >            LSI 8704EM2 RAID Controller (Think this controller does 3 Gbps)
> > >            Battery backup for the above RAID controller
> > >            4 X RAID10 Arrays (4 X 1.5TB disks per array, 16 disks in total)
> > >            Each RAID10 array will connect to 2 nodes (8 nodes per storage
> > > server) Intel PT 1Gbps Quad port NIC using Linux bonding
> > >            Exposes 8 X 1.5GB iSCSI targets (each node will use one of these)
> > >
> > >            HP Procurve 1800-24G switch to create 1 X 4 port trunk (for storage
> > >  server), and 8 X 2 port trunk (for the nodes)
> > >
> > >            What you think? Any tips?
> > >
> > >            Thanks
> > >
> > >
> > >            _______________________________________________
> > >            Xen-users mailing list
> > >            Xen-users@xxxxxxxxxxxxxxxxxxx
> > >            http://lists.xensource.com/xen-users
> >
>
> _______________________________________________
> Xen-users mailing list
> Xen-users@xxxxxxxxxxxxxxxxxxx
> http://lists.xensource.com/xen-users

_______________________________________________
Xen-users mailing list
Xen-users@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-users

 


Rackspace

Lists.xenproject.org is hosted with RackSpace, monitoring our
servers 24x7x365 and backed by RackSpace's Fanatical Support®.