[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-users] Release 0.8.0 of GPL PV Drivers for Windows
On Sun, Mar 02, 2008 at 11:52:07AM -0500, jim burns wrote: > On Sunday 02 March 2008 04:51:47 am Pasi Kärkkäinen wrote: > > Yep. I asked because of the "bad" 80 Mbit/sec result on a 100 Mbit > > network.. > > My guess is my cpu(s) don't have enough raw power to saturate the nic, but > let's see what happens after your suggestions are implemented. > I don't think that.. saturating 100 Mbit NIC/link with a single thread was/is possible with 10 years old CPUs.. > > That's because with some version of xen and/or drivers (I'm not sure > > actually) it was a known fact that performance got bad when you had hw > > checksum calculations turned on.. > > > > So just to see if that's the case here.. I guess this was mostly for domU.. > > Ahh, because there's no hardware - makes sense. > > Alright - let's try one change/set of related changes at a time to isolate > their effect. > Yep. > > > rx-checksumming: off > > > tx-checksumming: off > > > scatter-gather: off > > > tcp segmentation offload: off > > > udp fragmentation offload: off > > > generic segmentation offload: off > > > > Maybe try turning on offloading/checksumming settings here? > > Ok - before any changes, 'iperf -c insp6400 -t 60' gives 78.7 Mbps. > > On SuSE: > [830] > sudo ethtool -K eth0 tx on > Cannot set device tx csum settings: Operation not supported > [2] 23551 exit 85 sudo ethtool -K eth0 tx on > jimb@Dell4550 03/02/08 10:36AM:~ > [831] > sudo ethtool -K eth0 tso on > Cannot set device tcp segmentation offload settings: Operation not supported > [2] 23552 exit 88 sudo ethtool -K eth0 tso on > > on fc8: > [742] > sudo ethtool -K peth0 tx on > Password: > Cannot set device tx csum settings: Operation not supported > zsh: exit 85 sudo ethtool -K peth0 tx on > jimb@Insp6400 03/02/08 10:38AM:~ > [743] > sudo ethtool -K peth0 tso on > Cannot set device tcp segmentation offload settings: Operation not supported > zsh: exit 88 sudo ethtool -K peth0 tso on > > Ahem - moving on! > > > I usually use at least 256k window sizes :) > > Trying adding '-w 262144' on both server and client side, iperf gets 72.1 > Mpbs. Worse. > OK. > > Did you try with multiple threads at the same time? Did it have any effect? > > Adding '-P 4' to client, iperf gets an aggregate rate of 74.5 Mpbs. Worse. > OK. > > "ifconfig eth0 txqueuelen <value>" on linux.. I don't know how to do that > > in windows. > > > > But it's important to do that on dom0 for vifX.Y devices.. those are the > > dom0 sides of the virtual machine virtual NICs. > > [740] > ifconfig > [...] > peth0 Link encap:Ethernet HWaddr 00:15:C5:04:7D:4F > inet6 addr: fe80::215:c5ff:fe04:7d4f/64 Scope:Link > UP BROADCAST RUNNING MULTICAST MTU:1492 Metric:1 > RX packets:11545320 errors:0 dropped:237 overruns:0 frame:0 > TX packets:13476839 errors:0 dropped:0 overruns:0 carrier:0 > collisions:0 txqueuelen:1000 > RX bytes:421343991 (401.8 MiB) TX bytes:4224204231 (3.9 GiB) > Interrupt:22 > > tap0 Link encap:Ethernet HWaddr 0E:92:BB:CA:D8:DA > inet6 addr: fe80::c92:bbff:feca:d8da/64 Scope:Link > UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1 > RX packets:0 errors:0 dropped:0 overruns:0 frame:0 > TX packets:14090 errors:0 dropped:0 overruns:0 carrier:0 > collisions:0 txqueuelen:500 > RX bytes:0 (0.0 b) TX bytes:7860525 (7.4 MiB) > > vif4.0 Link encap:Ethernet HWaddr FE:FF:FF:FF:FF:FF > inet6 addr: fe80::fcff:ffff:feff:ffff/64 Scope:Link > UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1 > RX packets:2595441 errors:0 dropped:0 overruns:0 frame:0 > TX packets:923150 errors:0 dropped:795 overruns:0 carrier:0 > collisions:0 txqueuelen:32 > RX bytes:3366016758 (3.1 GiB) TX bytes:166299319 (158.5 MiB) > > vif16.0 Link encap:Ethernet HWaddr FE:FF:FF:FF:FF:FF > inet6 addr: fe80::fcff:ffff:feff:ffff/64 Scope:Link > UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1 > RX packets:9848 errors:0 dropped:0 overruns:0 frame:0 > TX packets:20828 errors:0 dropped:4698 overruns:0 carrier:0 > collisions:0 txqueuelen:32 > RX bytes:834219 (814.6 KiB) TX bytes:16017488 (15.2 MiB) > > Yowza! 32? Those *are* small! > 500 isn't much better for tap0, either. > Yep.. > > > > - Check sysctl net.core.netdev_max_backlog setting.. it should be at > > > > least 1000, possibly even more.. this applies to dom0 and linux domU. > > > > > > Where is this set, and what do I have to restart to make it > > > effective? /etc/sysctl.conf? > > > > Yep, modify /etc/sysctl.conf and run "sysctl -p /etc/sysctl.conf". > > > > > In general, are there any downsides in changing these values? > > > > http://kb.pert.geant2.net/PERTKB/InterfaceQueueLength > > Interesting link - thanx. > > Ok - setting 'ifconfig eth0 txqueuelen 2500' (peth0 on fc8) and > net.core.netdev_max_backlog = 2500 on both machines, iperf gets 65.9 Mpbs. > Worse. Probably only useful for Gpbs links. Removing changes, as in all cases > above. > Ok. Well, it was worth checking and testing :) > > There's something about these settings > > > > btw. what was the CPU usage for dom0 and for domU when you did these iperf > > tests? > > About 75%. I've noticed, at least on my SuSE box, that multimedia playback > suffers over 50%. (wine) > > On my WIndows guest, setting tap0's txqueuelen to 1000 had no effect (it > probably wouldn't since it's receiving); setting window size to 256k hung my > guest; after rebooting the guest, it had no effect on 2nd try; and changing > sysctl had no effect. Cpu % was about 80-85% w/o any changes (negligible on > fc8) averaged over 2 vcpus. With any of the changes above, cpu % went down to > 65-75% - the only change noticed. > > Have fun digesting this :-) > Yeah it's interesting.. I've measured 600+ Mbit/sec throughput from linux domU to external network.. on a 3 GHz P4. So I don't think CPU is the problem here.. -- Pasi _______________________________________________ Xen-users mailing list Xen-users@xxxxxxxxxxxxxxxxxxx http://lists.xensource.com/xen-users
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |