[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [Xen-devel] [PATCH] Memory pool for paravirtual guest in memory over 128G system
32-bit paravirtual domains can only consume physical memory below 168GB. On systems with memory beyond that address, they'll be confined to memory below 128GB. this patch define a memory pool for 32-bit paravirtual guest used. Using total_available_memory (in GB) in xend-config.sxp to specify the amount of memory reserved,in the memory pool exclusively for 32-bit paravirtual domains. Additionally you should use dom0_mem = <-Value> as a parameter in xen kernel to reserve the memory for 32-bit paravirtual domains, default is "0" (0GB). Value: the total memory for 32-bit paravirtual domains the main idea of this patch is: 1) The admin sets aside some memory below 128G for 32-bit paravirtual domain creation (via dom0_mem=-<value> in kernel comand line). 2) The admin also explicitly states to the tools (i..e xend) how much memory is supposed to be left untouched by 64-bit domains 3) If a 32-bit pv DomU gets created, no ballooning ought to be necessary (since if it is, no guarantee can be made about the address range of the memory ballooned out), and memory gets allocated from the reserved range. 4) Upon 64-bit (or 32-bit HVM or HVM) DomU creation, the tools determine the amount of memory to be ballooned out of Dom0 by adding the amount needed for the new guest and the amount still in the reserved pool (and then of course subtracting the total amount of memory the hypervisor has available for guest use). Signed-off-by: james song (wei)<jsong@xxxxxxxxxx> diff -r c8caa281a75a tools/examples/xend-config.sxp --- a/tools/examples/xend-config.sxp Mon Nov 23 08:06:54 2009 +0000 +++ b/tools/examples/xend-config.sxp Tue Nov 24 14:37:56 2009 +0800 @@ -192,6 +192,16 @@ # If enable-dom0-ballooning = no, dom0 will never balloon out. (enable-dom0-ballooning yes) +# 32-bit paravirtual domains can only consume physical +# memory below 168GB. On systems with memory beyond that address, +# they'll be confined to memory below 128GB. +# Using total_available_memory (in GB) to specify the amount of memory reserved +# in the memory pool exclusively for 32-bit paravirtual domains. +# Additionally you should use dom0_mem = <-Value> as a parameter in +# xen kernel to reserve the memory for 32-bit paravirtual domains, default +# is "0" (0GB). +(total_available_memory 0) + # In SMP system, dom0 will use dom0-cpus # of CPUS # If dom0-cpus = 0, dom0 will take all cpus available (dom0-cpus 0) diff -r c8caa281a75a tools/libxc/xc_dom_compat_linux.c --- a/tools/libxc/xc_dom_compat_linux.c Mon Nov 23 08:06:54 2009 +0000 +++ b/tools/libxc/xc_dom_compat_linux.c Tue Nov 24 14:37:56 2009 +0800 @@ -122,6 +122,28 @@ xc_dom_release(dom); return rc; } +int xc_get_bit_size(const char *image_name, const char *cmdline, + const char *features, int *bit_size) +{ + struct xc_dom_image *dom; + int rc; + *bit_size = 0; + dom = xc_dom_allocate(cmdline, features); + if ( (rc = xc_dom_kernel_file(dom, image_name)) != 0 ) + goto out; + if ( (rc = xc_dom_parse_image(dom)) != 0 ) + goto out; + if( dom->guest_type != NULL){ + if(strstr(dom->guest_type, "x86_64") != NULL) + *bit_size = X86_64_B_SIZE; //64bit Guest + if(strstr(dom->guest_type, "x86_32") != NULL) + *bit_size = X86_32_B_SIZE; //32bit Guest + } + +out: + xc_dom_release(dom); + return rc; +} int xc_dom_linux_build(int xc_handle, struct xc_dom_image *dom, diff -r c8caa281a75a tools/libxc/xenguest.h --- a/tools/libxc/xenguest.h Mon Nov 23 08:06:54 2009 +0000 +++ b/tools/libxc/xenguest.h Tue Nov 24 14:37:56 2009 +0800 @@ -13,6 +13,8 @@ #define XCFLAGS_DEBUG 2 #define XCFLAGS_HVM 4 #define XCFLAGS_STDVGA 8 +#define X86_64_B_SIZE 64 +#define X86_32_B_SIZE 32 /* callbacks provided by xc_domain_save */ struct save_callbacks { @@ -161,6 +163,9 @@ int xc_await_suspend(int xce, int suspend_evtchn); +int xc_get_bit_size(const char *image_name, const char *cmdline, + const char *features, int *type); + int xc_mark_page_online(int xc, unsigned long start, unsigned long end, uint32_t *status); diff -r c8caa281a75a tools/python/xen/lowlevel/xc/xc.c --- a/tools/python/xen/lowlevel/xc/xc.c Mon Nov 23 08:06:54 2009 +0000 +++ b/tools/python/xen/lowlevel/xc/xc.c Tue Nov 24 14:37:56 2009 +0800 @@ -397,6 +397,25 @@ PyDict_SetItemString(info_dict, "cpumap", cpulist); Py_DECREF(cpulist); return info_dict; +} + +static PyObject *pyxc_getBitSize(XcObject *self, + PyObject *args, + PyObject *kwds) +{ + PyObject *info_type; + char *image = NULL, *cmdline = "", *features = NULL; + int type = 0; + static char *kwd_list[] = { "image", "cmdline", "features"}; + if ( !PyArg_ParseTupleAndKeywords(args, kwds, "sss", kwd_list, + â, &cmdline, &features) ) + return NULL; + xc_get_bit_size(image, cmdline, features, &type); + if (type < 0) + return pyxc_error_to_exception(); + info_type = Py_BuildValue("{s:i}", + "type", type); + return info_type; } static PyObject *pyxc_linux_build(XcObject *self, @@ -1777,6 +1796,13 @@ " vcpus [int, 1]: Number of Virtual CPUS in domain.\n\n" "Returns: [int] 0 on success; -1 on error.\n" }, + {"getBitSize", + (PyCFunction)pyxc_getBitSize, + METH_VARARGS | METH_KEYWORDS, "\n" + "Get the bitsize of a guest OS.\n" + " image [str]: Name of kernel image file. May be gzipped.\n" + " cmdline [str, n/a]: Kernel parameters, if any.\n\n"}, + { "hvm_build", (PyCFunction)pyxc_hvm_build, METH_VARARGS | METH_KEYWORDS, "\n" diff -r c8caa281a75a tools/python/xen/xend/MemoryPool.py --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/tools/python/xen/xend/MemoryPool.py Tue Nov 24 14:37:56 2009 +0800 @@ -0,0 +1,100 @@ +import xen.lowlevel.xc +import XendDomain +import XendOptions +from XendLogging import log +from XendError import VmError + +class MemoryPool: + + def init(self): + xoptions = XendOptions.instance() + self.default_reserved_memory = xoptions.get_reserved_memory() * 1024 * 1024 #KiB + if self.default_reserved_memory <= 0: + return + self.enable_memory_pool = 1 + self.dom0_ballooning = xoptions.get_enable_dom0_ballooning() + if not self.dom0_ballooning: + return + self.reserve_memory = 0 + self.untouched_memory = 0 + #init reserved memory + #if not reserve_memory_size: + xc = xen.lowlevel.xc.xc() + physinfo = xc.physinfo() + total_mem = physinfo['total_memory'] + if total_mem < self.reserve_memory: + self.default_reserved_memory = total_mem + self.reserve_memory = self.default_reserved_memory + self.untouched_memory = self.default_reserved_memory + log.debug("MemoryPool: init reserved_memory %d KiB" %self.reserve_memory) + + def __init__(self): + self.reserve_memory = 0 + self.untouched_memory = 0 + self.default_reserved_memory = 0 + self.enable_memory_pool = 0 + self.dom0_ballooning = 0 + def available_memory_check(self, need_mem): + return self.is_enabled() and self.reserved_memory > need_mem + + def decrease_memory(self, value): + if not self.is_enabled() or value <= 4096: #4M for PV guest kernel and ramdisk unzip + return + elif self.reserve_memory < value: + raise VMError(('I need %d KiB, but only have %d KiB in Memory Pool') %(value,self.reserve_memory)) + else: + self.reserve_memory -= value + log.debug("MemoryPool: decrease_memory: decrease: %d reserved_memory %d KiB" %(value,self.reserve_memory)) + return + + def decrease_untouched_memory(self, value): + if not self.is_enabled(): + return + elif self.untouched_memory < value: + raise VmError(('I need %d KiB untouch mem, but only have %d KiB untouched mem in Memory Pool') %(value,self.reserve_memory)) + else: + self.untouched_memory -= value + log.debug("MemoryPool: decrease_untouched_memory: untouched_memory %d KiB" %self.untouched_memory) + return + + def increase_memory(self, value): + if not self.is_enabled(): + return + else: + self.reserve_memory += value + if self.reserve_memory > self.default_reserved_memory: + raise VmError(('the maxsize of memory pool is %d KiB, but current is %d KiB') %(value,self.reserve_memory)) + log.debug("MemoryPool: increase_memory:%d, reserved_memory %d KiB" %(value,self.reserve_memory)) + return + + def is_enabled(self): + return self.enable_memory_pool and self.dom0_ballooning + + def get_pool_size(self): + if self.is_enabled(): + return self.default_reserved_memory + else: + return 0 + + def get_left_memory(self): + if self.is_enabled(): + return self.reserve_memory + else: + return 0 + + def get_untouched_memory(self): + if self.is_enabled(): + return self.untouched_memory + else: + return 0 + +def instance(): + """Singleton constructor. Use this instead of the class constructor. + """ + global MP_inst + try: + MP_inst + except: + MP_inst = MemoryPool() + MP_inst.init() + return MP_inst diff -r c8caa281a75a tools/python/xen/xend/XendDomainInfo.py --- a/tools/python/xen/xend/XendDomainInfo.py Mon Nov 23 08:06:54 2009 +0000 +++ b/tools/python/xen/xend/XendDomainInfo.py Tue Nov 24 14:37:56 2009 +0800 @@ -326,6 +326,10 @@ @type info: dictionary @ivar domid: Domain ID (if VM has started) @type domid: int or None + @ivar guest_bitsize: the bitsize of guest + @type guest_bitsize: int or None + @ivar alloc_mem: the memory domain allocated when booting + @type alloc_mem: int or None @ivar vmpath: XenStore path to this VM. @type vmpath: string @ivar dompath: XenStore path to this Domain. @@ -383,6 +387,8 @@ self.domid = self.info.get('domid') else: self.domid = domid + self.guest_bitsize = None + self.alloc_mem = None #REMOVE: uuid is now generated in XendConfig #if not self._infoIsSet('uuid'): @@ -2715,6 +2721,8 @@ self._configureBootloader() try: + #xstransact.Mkdir("/local/domain/%u/used_reserved_memory" + # % self.getDomid()) self.image = image.create(self, self.info) # repin domain vcpus if a restricted cpus list is provided @@ -2750,6 +2758,7 @@ # Round vtd_mem up to a multiple of a MiB. vtd_mem = ((vtd_mem + 1023) / 1024) * 1024 + self.guest_bitsize = self.image.getBitSize() # Make sure there's enough RAM available for the domain balloon.free(memory + shadow + vtd_mem, self) @@ -2940,9 +2949,9 @@ if self.domid is None: return - from xen.xend import XendDomain log.debug("XendDomainInfo.destroy: domid=%s", str(self.domid)) + #released_mem = xstransact.Read(self.dompath, "memory/target") paths = self._prepare_phantom_paths() @@ -2960,7 +2969,11 @@ XendDomain.instance().remove_domain(self) self.cleanupDomain() - + if self.info.is_hvm() or self.guest_bitsize != 32: + if self.alloc_mem: + import MemoryPool + log.debug("%s KiB need to add to Memory pool" %self.alloc_mem) + MemoryPool.instance().increase_memory(self.alloc_mem) self._cleanup_phantom_devs(paths) self._cleanupVm() diff -r c8caa281a75a tools/python/xen/xend/XendOptions.py --- a/tools/python/xen/xend/XendOptions.py Mon Nov 23 08:06:54 2009 +0000 +++ b/tools/python/xen/xend/XendOptions.py Tue Nov 24 14:37:56 2009 +0800 @@ -104,6 +104,8 @@ xend_unix_path_default = '/var/lib/xend/xend-socket' dom0_min_mem_default = 0 + + reserved_memory_default = 0 dom0_vcpus_default = 0 @@ -364,6 +366,13 @@ return self.get_config_bool('enable-dom0-ballooning', enable_dom0_ballooning_default) + def get_reserved_memory(self): + if not self.get_enable_dom0_ballooning(): + return 0 #no ballooning of dom0 will close this item + else: + return self.get_config_int('total_available_memory', self.reserved_memory_default) + + def get_dom0_vcpus(self): return self.get_config_int('dom0-cpus', self.dom0_vcpus_default) diff -r c8caa281a75a tools/python/xen/xend/balloon.py --- a/tools/python/xen/xend/balloon.py Mon Nov 23 08:06:54 2009 +0000 +++ b/tools/python/xen/xend/balloon.py Tue Nov 24 14:37:56 2009 +0800 @@ -23,6 +23,7 @@ import XendDomain import XendOptions +import MemoryPool from XendLogging import log from XendError import VmError import osdep @@ -97,10 +98,22 @@ xoptions = XendOptions.instance() dom0 = XendDomain.instance().privilegedDomain() xc = xen.lowlevel.xc.xc() - + memory_pool = MemoryPool.instance() try: dom0_min_mem = xoptions.get_dom0_min_mem() * 1024 dom0_ballooning = xoptions.get_enable_dom0_ballooning() + guest_size = 0 + hvm = dominfo.info.is_hvm() + if memory_pool.is_enabled() and dominfo.domid: + if not hvm : + if need_mem <= 4 * 1024: + guest_size = 32 + else: + guest_size = dominfo.image.getBitSize() + if guest_size == 32: + dom0_ballooning = 0 + else: #no ballooning as memory pool enabled + dom0_ballooning = xoptions.get_enable_dom0_ballooning() dom0_alloc = get_dom0_current_alloc() retries = 0 @@ -109,6 +122,11 @@ last_new_alloc = None last_free = None rlimit = RETRY_LIMIT + mem_need_balloon = 0 + left_memory_pool = 0 + mem_target = 0 + untouched_memory_pool = 0 + real_need_mem = need_mem # stop tmem from absorbing any more memory (must THAW when done!) xc.tmem_control(0,TMEMC_FREEZE,-1, 0, 0, 0, "") @@ -119,6 +137,25 @@ free_mem = physinfo['free_memory'] scrub_mem = physinfo['scrub_memory'] total_mem = physinfo['total_memory'] + if memory_pool.is_enabled() and dominfo.domid: + if guest_size != 32 or hvm: + if need_mem > 4 * 1024: + dominfo.alloc_mem = need_mem + left_memory_pool = memory_pool.get_left_memory() + if need_mem > left_memory_pool: + dominfo.alloc_mem = 0 + raise VmError(('Not enough free memory' + ' so I cannot release any more. ' + 'I need %d KiB but only have %d in the pool.') % + (need_mem, memory_pool.get_left_memory())) + else: + untouched_memory_pool = memory_pool.get_untouched_memory() + if (left_memory_pool - untouched_memory_pool) > need_mem: + dom0_ballooning = 0 + else: + mem_need_balloon = need_mem - left_memory_pool + untouched_memory_pool + need_mem = free_mem + scrub_mem + mem_need_balloon + if dom0_ballooning: max_free_mem = total_mem - dom0_min_mem else: @@ -170,15 +207,17 @@ retries = 0 sleep_time = SLEEP_TIME_GROWTH - while retries < rlimit: physinfo = xc.physinfo() free_mem = physinfo['free_memory'] scrub_mem = physinfo['scrub_memory'] - if free_mem >= need_mem: - log.debug("Balloon: %d KiB free; need %d; done.", - free_mem, need_mem) + if (guest_size != 32 or hvm) and dominfo.domid: + memory_pool.decrease_untouched_memory(mem_need_balloon) + memory_pool.decrease_memory(real_need_mem) + else: + log.debug("Balloon: %d KiB free; need %d; done.", + free_mem, need_mem) return if retries == 0: @@ -189,7 +228,6 @@ if dom0_ballooning: dom0_alloc = get_dom0_current_alloc() new_alloc = dom0_alloc - (need_mem - free_mem - scrub_mem) - if free_mem + scrub_mem >= need_mem: if last_new_alloc == None: log.debug("Balloon: waiting on scrubbing") @@ -213,11 +251,13 @@ # Not enough memory; diagnose the problem. if not dom0_ballooning: + dominfo.alloc_mem = 0 raise VmError(('Not enough free memory and enable-dom0-ballooning ' 'is False, so I cannot release any more. ' 'I need %d KiB but only have %d.') % (need_mem, free_mem)) elif new_alloc < dom0_min_mem: + dominfo.alloc_mem = 0 raise VmError( ('I need %d KiB, but dom0_min_mem is %d and shrinking to ' '%d KiB would leave only %d KiB free.') % @@ -226,6 +266,7 @@ else: dom0_start_alloc_mb = get_dom0_current_alloc() / 1024 dom0.setMemoryTarget(dom0_start_alloc_mb) + dominfo.alloc_mem = 0 raise VmError( ('Not enough memory is available, and dom0 cannot' ' be shrunk any further')) diff -r c8caa281a75a tools/python/xen/xend/image.py --- a/tools/python/xen/xend/image.py Mon Nov 23 08:06:54 2009 +0000 +++ b/tools/python/xen/xend/image.py Tue Nov 24 14:37:56 2009 +0800 @@ -736,6 +736,12 @@ vhpt = self.vhpt, superpages = self.superpages) + def getBitSize(self): + return xc.getBitSize(image = self.kernel, + cmdline = self.cmdline, + features = self.vm.getFeatures() + ).get('type') + def getRequiredAvailableMemory(self, mem_kb): if self.is_stubdom : mem_kb += self.vramsize @@ -768,6 +774,9 @@ ImageHandler.__init__(self, vm, vmConfig) self.shutdownWatch = None self.rebootFeatureWatch = None + + def getBitSize(self): + return None def configure(self, vmConfig): ImageHandler.configure(self, vmConfig) @@ -1011,6 +1020,9 @@ rc = HVMImageHandler.buildDomain(self) self.setCpuid() return rc + + def getBitSize(self): + return None def getRequiredAvailableMemory(self, mem_kb): return mem_kb + self.vramsize http://old.nabble.com/file/p26570558/memory_pool.patch memory_pool.patch -- View this message in context: http://old.nabble.com/-PATCH--Memory-pool-for-paravirtual-guest-in-memory-over-128G-system-tp26570558p26570558.html Sent from the Xen - Dev mailing list archive at Nabble.com. _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxxxxxxxx http://lists.xensource.com/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |