[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

RE: [Xen-users] My future plan

  • To: "Bart Coninckx" <bart.coninckx@xxxxxxxxxx>, <xen-users@xxxxxxxxxxxxxxxxxxx>
  • From: "Jonathan Tripathy" <jonnyt@xxxxxxxxxxx>
  • Date: Wed, 9 Jun 2010 13:53:28 +0100
  • Cc:
  • Delivery-date: Wed, 09 Jun 2010 05:58:52 -0700
  • List-id: Xen user discussion <xen-users.lists.xensource.com>
  • Thread-index: AcsHvhizXTStFiehR9+fCmlwyxC1rAAFKkqZ
  • Thread-topic: [Xen-users] My future plan

So should I just go with 2 NICs for the storage server then?
In your future setup, how many NICs are you using for the storage server and how many for the nodes? I take it you're using software iSCSI?

From: Bart Coninckx [mailto:bart.coninckx@xxxxxxxxxx]
Sent: Wed 09/06/2010 11:25
To: xen-users@xxxxxxxxxxxxxxxxxxx
Cc: Jonathan Tripathy; Michael Schmidt
Subject: Re: [Xen-users] My future plan

On the DRBD mailing lists I've seen a couple of times that they did tests with
bonding and they claim that a bond with more than 2 NICs will actually
decrease performance because of the TCP reordering that needs to be done.

That's the reason why I limit the storage connection to two NICs. I have a
very similar to yours in the making by the way.

On Tuesday 08 June 2010 15:55:47 Jonathan Tripathy wrote:
> Hi Michael,
> Thanks for the tips using SSD for the node OS drives.
> Regarding the NIC, I was thinking about using this for the nodes:
> http://www.intel.com/products/server/adapters/pro1000pt-dualport/pro1000pt-
> dualport-overview.htm
> and this for the server:
> http://www.intel.com/products/server/adapters/pro1000pt-quadport-low-profil
> e/pro1000pt-quadport-low-profile-overview.htm
> Are those the cards you were talking about? They are very cheap on ebay you
>  see...
> Think 4 port bonding for the server is good enough for 8 nodes?
> Thanks
> ________________________________
> From: Michael Schmidt [mailto:michael.schmidt@xxxxxxxxxx]
> Sent: Tue 08/06/2010 14:49
> To: Jonathan Tripathy; Xen-users@xxxxxxxxxxxxxxxxxxx
> Subject: Re: [Xen-users] My future plan
> Hi Jonathan,
> you should think about flash or SD cards as xen-boot-drive.
> This provides you lower costs and higher energy efficiency.
> If you mount /tmp and /var/log to an tmpfs, this disks works very well and
>  long.
> If you dont need so much disk space for your storage, use sas disks.
> SAS (10k/15k) disks provides you many more IOPs than sata disks (more IOPS
>  per $/EUR as well). And very important: A very large cache for your raid
>  controller.
> Intel e1000e is a pretty good choice. This cards have a large buffer and
>  generates just a few interrupts on your CPUs (in comparison to the
>  Broadcom NICs).
> Best Regards
> Michael Schmidt
> Am 08.06.10 14:55, schrieb Jonathan Tripathy:
>       My future plan currently looks like this for my VPS hosting solution, so
>  any feedback would be appreciated:
>       Each Node:
>       Dell R210 Intel X3430 Quad Core 8GB RAM
>       Intel PT 1Gbps Server Dual Port NIC using linux "bonding"
>       Small pair of HDDs for OS (Probably in RAID1)
>       Each node will run about 10 - 15 customer guests
>       Storage Server:
>       Some Intel Quad Core Chip
>       2GB RAM (Maybe more?)
>       LSI 8704EM2 RAID Controller (Think this controller does 3 Gbps)
>       Battery backup for the above RAID controller
>       4 X RAID10 Arrays (4 X 1.5TB disks per array, 16 disks in total)
>       Each RAID10 array will connect to 2 nodes (8 nodes per storage server)
>       Intel PT 1Gbps Quad port NIC using Linux bonding
>       Exposes 8 X 1.5GB iSCSI targets (each node will use one of these)
>       HP Procurve 1800-24G switch to create 1 X 4 port trunk (for storage
>  server), and 8 X 2 port trunk (for the nodes)
>       What you think? Any tips?
>       Thanks
>       _______________________________________________
>       Xen-users mailing list
>       Xen-users@xxxxxxxxxxxxxxxxxxx
>       http://lists.xensource.com/xen-users

Xen-users mailing list



Lists.xenproject.org is hosted with RackSpace, monitoring our
servers 24x7x365 and backed by RackSpace's Fanatical Support®.