[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [Xen-devel] [PATCH 0/6] xen,xend,tools: NUMA support for Xen
Reposting the latest patches, no significant changes since May when I last received feedback. I've done some simple overhead and performance numbers for these patches. Measuring NUMA allocator patch overhead via balloon driver and DOM0 involved starting DOM0 with all memory (4G) and ballooning down (256M) as a starting point, [1]tracking the time it takes to balloon back up to full memory. The test was done with and without NUMA patches. With NUMA: Try1: 911ms Try2: 907ms Try3: 910ms Without NUMA: Try1: 606ms Try2: 604ms Try3: 608ms Measuring NUMA allocator patch overhead via increase_reservation memory op during domain creation we [2]modified the increase_reservation op to track the time using get_s_time() from start to finish and sampled the times for various memory sizes. With NUMA: MemSize 128M 512M 1G 2G 3G ------------------------------------- Try1: 6ms 26ms 53ms 221ms 390ms Try2: 6ms 26ms 48ms 212ms 390ms Try3: 6ms 26ms 48ms 212ms 390ms Without NUMA: MemSize 128M 512M 1G 2G 3G ------------------------------------- Try1: 4ms 16ms 25ms 70ms 100ms Try2: 3ms 14ms 28ms 56ms 109ms Try3: 3ms 14ms 23ms 56ms 95ms Using a microbenchmark which mallocs memory and touches each byte, we can observe the affects of local memory versus remote. The domain is created with 1GB memory, and 1 VCPU coming from the same node. The microbenchmark forks off one child per-VCPU and malloc/memsets 512M buffer each. We then compare worst-case (All non-local memory) and best-case (all-local). Machine Topology: node_to_cpu : node0:0 node1:1 Domain's memory placement (bebop) tmp # xen_numastat -d 8 DOM8: NODE0: PAGES: 5 DOM8: NODE1: PAGES: 262144 Domain's vcpu placement (bebop) tmp # xm vcpu-list 8 Name ID VCPU CPU State Time(s) CPU Affinity hungerforce 8 0 0 -b- 12.0 0 All-remote memory: root@amd64-domU:/usr/src # while true; do ./memwrite -j1 -m512M; sleep 1; done Time to write '0' to 512.000 MiB 1168711 usecs. Throughput: 438.090 MiB/sec Time to write '0' to 512.000 MiB 1175179 usecs. Throughput: 435.678 MiB/sec Time to write '0' to 512.000 MiB 1172454 usecs. Throughput: 436.691 MiB/sec Time to write '0' to 512.000 MiB 1170378 usecs. Throughput: 437.466 MiB/sec Domain's vcpu placement (bebop) tmp # xm vcpu-list 8 Name ID VCPU CPU State Time(s) CPU Affinity hungerforce 8 0 0 -b- 15.9 1 All-local memory: root@amd64-domU:/usr/src # while true; do ./memwrite -j1 -m512M; sleep 1; done Time to write '0' to 512.000 MiB 759186 usecs. Throughput: 674.406 MiB/sec Time to write '0' to 512.000 MiB 765143 usecs. Throughput: 669.156 MiB/sec Time to write '0' to 512.000 MiB 768462 usecs. Throughput: 666.266 MiB/sec Time to write '0' to 512.000 MiB 763406 usecs. Throughput: 670.679 MiB/sec 1. diff -r ae245d35457b linux-2.6-xen-sparse/drivers/xen/balloon/balloon.c --- a/linux-2.6-xen-sparse/drivers/xen/balloon/balloon.c Wed Jun 28 12:59:29 2006 +++ b/linux-2.6-xen-sparse/drivers/xen/balloon/balloon.c Mon Jun 26 14:54:10 2006 @@ -44,6 +44,7 @@ #include <linux/bootmem.h> #include <linux/highmem.h> #include <linux/vmalloc.h> +#include <linux/time.h> #include <xen/xen_proc.h> #include <asm/hypervisor.h> #include <xen/balloon.h> @@ -63,6 +64,9 @@ #endif static DECLARE_MUTEX(balloon_mutex); + +static struct timeval start, stop, delay; +static int calc_delay = 0; /* * Protects atomic reservation decrease/increase against concurrent increases. @@ -337,6 +341,14 @@ return need_sleep; } +/** Convert to milliseconds */ +static inline __u64 tv_to_ms(const struct timeval* tv) +{ + __u64 ms = tv->tv_usec / 1000; + ms += (__u64)tv->tv_sec * (__u64)1000; + return ms; +} + /* * We avoid multiple worker processes conflicting via the balloon mutex. * We may of course race updates of the target counts (which are protected @@ -350,6 +362,11 @@ down(&balloon_mutex); + if (calc_delay) { + do_gettimeofday(&delay); + calc_delay = 0; + } + do { credit = current_target() - current_pages; if (credit > 0) @@ -366,6 +383,13 @@ /* Schedule more work if there is some still to be done. */ if (current_target() != current_pages) mod_timer(&balloon_timer, jiffies + HZ); + else { + /* We've hit target, notify completion */ + do_gettimeofday(&stop); + printk(KERN_WARNING "Ballooning complete. startup delay: %lums", + tv_to_ms(&delay)-tv_to_ms(&start)); + printk(" total time: %lums\n", tv_to_ms(&stop)-tv_to_ms(&start)); + } up(&balloon_mutex); } @@ -376,6 +400,11 @@ /* No need for lock. Not read-modify-write updates. */ hard_limit = ~0UL; target_pages = target; + + /* note start time of balloon process */ + do_gettimeofday(&start); + calc_delay = 1; + schedule_work(&balloon_worker); } 2. diff -r c257ac74b5c7 xen/common/memory.c --- a/xen/common/memory.c Tue Jul 4 16:31:13 2006 +++ b/xen/common/memory.c Wed Jul 5 12:12:43 2006 @@ -27,6 +27,10 @@ * high-order bits of the @cmd parameter, which are otherwise unused and zero. */ #define START_EXTENT_SHIFT 4 /* cmd[:4] == start_extent */ + +static int calc_start = 1; +static unsigned long initial_extent; +static s_time_t start, stop; static long increase_reservation( @@ -574,6 +578,13 @@ switch ( op ) { case XENMEM_increase_reservation: + if ( calc_start ) { + printk("WARK: calcing start time on IR of %lu pages\n", + reservation.nr_extents); + initial_extent = reservation.nr_extents; + start = get_s_time(); + calc_start = 0; + } rc = increase_reservation( d, reservation.extent_start, @@ -612,6 +623,12 @@ __HYPERVISOR_memory_op, "lh", op | (rc << START_EXTENT_SHIFT), arg); + if ( op == XENMEM_increase_reservation ) { + stop = get_s_time(); + printk("WARK: increase_reservation of %lu pages took->%lums\n", + initial_extent, ((u64)stop-(u64)start)/1000000 ); + calc_start = 1; + } break; case XENMEM_exchange: -- Ryan Harper Software Engineer; Linux Technology Center IBM Corp., Austin, Tx (512) 838-9253 T/L: 678-9253 ryanh@xxxxxxxxxx _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxxxxxxxx http://lists.xensource.com/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |