[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-users] iscsi vs nfs for xen VMs
On Sat, Jan 29, 2011 at 06:12:34PM +0100, Bart Coninckx wrote: > On 01/29/11 16:30, Pasi Kärkkäinen wrote: > > On Sat, Jan 29, 2011 at 04:27:52PM +0100, Bart Coninckx wrote: > > > >> On 01/29/11 16:09, Pasi Kärkkäinen wrote: > >> > >>> On Thu, Jan 27, 2011 at 09:35:38AM +0100, Adi Kriegisch wrote: > >>> > >>> > >>>> Hi! > >>>> > >>>> > >>>> > >>>>>> iSCSI tipically has a quite big overhead due to the protocol, FC, SAS, > >>>>>> native infiniband, AoE have very low overhead. > >>>>>> > >>>>>> > >>>>>> > >>>>> For iSCSI vs AoE, that isn't as true as you might think. TCP offload can > >>>>> take care of a lot of the overhead. Any server class network adapter > >>>>> these days should allow you to send 60kb packets to the network adapter > >>>>> and it will take care of the segmentation, while AoE would be limited to > >>>>> MTU sized packets. With AoE you need to checksum every packet yourself > >>>>> while with iSCSI it is taken care of by the network adapter. > >>>>> > >>>>> > >>>> What AoE actually does is sending a frame per block. Block size is 4K -- > >>>> so > >>>> no need for fragmentation. The overhead is pretty low, because we're > >>>> talking about Ethernet frames. > >>>> Most iSCSI issues I have seen are with reordering of packages due to > >>>> transmission across several interfaces. So what most people recommend is > >>>> to > >>>> keep the number of interfaces to two. To keep performance up this means > >>>> you > >>>> have to use 10G, FC or similar which is quite expensive -- especially if > >>>> you'd like to have a HA SAN network (HSRP and stuff like that is > >>>> required). > >>>> > >>>> AoE does not suffer from those issues: Using 6 GBit interfaces is no > >>>> problem at all, load balancing will happen automatically, as the load is > >>>> distributed equally across all available interfaces. HA is very simple: > >>>> just use two switches and connect one half of the interfaces to one > >>>> switch > >>>> and the other half to the other switch. (It is recommended to use > >>>> switches > >>>> that can do jumbo frames and flow control) > >>>> IMHO most of the current recommendations and practises surrounding iSCSI > >>>> are there to overcome the shortcomings of the protocol. AoE is way more > >>>> robust and easier to handle. > >>>> > >>>> > >>>> > >>> iSCSI does not have problems using multiple gige interfaces. > >>> Just setup multipathing properly. > >>> > >>> -- Pasi > >>> > >>> > >>> _______________________________________________ > >>> Xen-users mailing list > >>> Xen-users@xxxxxxxxxxxxxxxxxxx > >>> http://lists.xensource.com/xen-users > >>> > >>> > >> On this subject: am using multipathing to iSCSI too, hoping to have > >> aggregated speed on top of path redundancy but the speed seems not to > >> surpass the one of a single interface. > >> > >> Is anyone successful at doing this? > >> > >> > > You're benchmarking sequential/linear IO, using big blocksizes, right? > > > > Some questions: > > - Are you using multipath round robin path policy? > > - After how many IOs do you switch paths? You might need to lower the > > min_ios. > > > > > > -- Pasi > > > > > Hi Pasi, > > the benchmarking was intuitively done, with just dd and bonnie++. > with dd, use "bs=1024k" or so to force big blocks. > It is indeed rr, this is a part of my multipath.conf: > > defaults { > udev_dir /dev > polling_interval 10 > selector "round-robin 0" > path_grouping_policy multibus > getuid_callout "/lib/udev/scsi_id --whitelisted > --device=/dev/%n" > prio const > path_checker directio > rr_min_io 100 > max_fds 8192 > rr_weight priorities > failback immediate > no_path_retry 5 > user_friendly_names no > } > > should the "100" go down a bit? > Yeah, for example Equallogic enterprise iSCSI storage manuals recommends rr_min_io of "3". That should allow utilizing both paths. -- Pasi _______________________________________________ Xen-users mailing list Xen-users@xxxxxxxxxxxxxxxxxxx http://lists.xensource.com/xen-users
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |