[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: [Xen-users] My future plan



On the DRBD mailing lists I've seen a couple of times that they did tests with 
bonding and they claim that a bond with more than 2 NICs will actually 
decrease performance because of the TCP reordering that needs to be done. 

That's the reason why I limit the storage connection to two NICs. I have a 
very similar to yours in the making by the way.


On Tuesday 08 June 2010 15:55:47 Jonathan Tripathy wrote:
> Hi Michael,
> 
> Thanks for the tips using SSD for the node OS drives.
> 
> Regarding the NIC, I was thinking about using this for the nodes:
> 
> http://www.intel.com/products/server/adapters/pro1000pt-dualport/pro1000pt-
> dualport-overview.htm
> 
> and this for the server:
> 
> http://www.intel.com/products/server/adapters/pro1000pt-quadport-low-profil
> e/pro1000pt-quadport-low-profile-overview.htm
> 
> Are those the cards you were talking about? They are very cheap on ebay you
>  see...
> 
> Think 4 port bonding for the server is good enough for 8 nodes?
> 
> Thanks
> 
> ________________________________
> 
> From: Michael Schmidt [mailto:michael.schmidt@xxxxxxxxxx]
> Sent: Tue 08/06/2010 14:49
> To: Jonathan Tripathy; Xen-users@xxxxxxxxxxxxxxxxxxx
> Subject: Re: [Xen-users] My future plan
> 
> 
> Hi Jonathan,
> 
> you should think about flash or SD cards as xen-boot-drive.
> This provides you lower costs and higher energy efficiency.
> If you mount /tmp and /var/log to an tmpfs, this disks works very well and
>  long.
> 
> If you dont need so much disk space for your storage, use sas disks.
> SAS (10k/15k) disks provides you many more IOPs than sata disks (more IOPS
>  per $/EUR as well). And very important: A very large cache for your raid
>  controller.
> 
> Intel e1000e is a pretty good choice. This cards have a large buffer and
>  generates just a few interrupts on your CPUs (in comparison to the
>  Broadcom NICs).
> 
> Best Regards
> 
> Michael Schmidt
> Am 08.06.10 14:55, schrieb Jonathan Tripathy:
> 
>       My future plan currently looks like this for my VPS hosting solution, so
>  any feedback would be appreciated:
> 
>       Each Node:
>       Dell R210 Intel X3430 Quad Core 8GB RAM
>       Intel PT 1Gbps Server Dual Port NIC using linux "bonding"
>       Small pair of HDDs for OS (Probably in RAID1)
>       Each node will run about 10 - 15 customer guests
> 
> 
>       Storage Server:
>       Some Intel Quad Core Chip
>       2GB RAM (Maybe more?)
>       LSI 8704EM2 RAID Controller (Think this controller does 3 Gbps)
>       Battery backup for the above RAID controller
>       4 X RAID10 Arrays (4 X 1.5TB disks per array, 16 disks in total)
>       Each RAID10 array will connect to 2 nodes (8 nodes per storage server)
>       Intel PT 1Gbps Quad port NIC using Linux bonding
>       Exposes 8 X 1.5GB iSCSI targets (each node will use one of these)
> 
>       HP Procurve 1800-24G switch to create 1 X 4 port trunk (for storage
>  server), and 8 X 2 port trunk (for the nodes)
> 
>       What you think? Any tips?
> 
>       Thanks
> 
> 
>       _______________________________________________
>       Xen-users mailing list
>       Xen-users@xxxxxxxxxxxxxxxxxxx
>       http://lists.xensource.com/xen-users
> 

_______________________________________________
Xen-users mailing list
Xen-users@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-users


 


Rackspace

Lists.xenproject.org is hosted with RackSpace, monitoring our
servers 24x7x365 and backed by RackSpace's Fanatical Support®.