[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] RE: [Xen-devel] Strange (???) xl behavior for save, migrate and migrate-receive
> From: Daniel Kiper [mailto:dkiper@xxxxxxxxxxxx] > Subject: [Xen-devel] Strange (???) xl behavior for save, migrate and > migrate-receive > > During work on memory hotplug for Xen I have received some notices > that it breaks machine migration. I had some time and I done some > tests a few days ago. It looks that source of this problem is > xl command itself. I discovered that generic save/restore mechanism > is used for machine migration. xl save store machine config which > was used at machine startup with current machine state. It means > that it does not take into account any config changes which were made > during machine run. This behavior does not allow migrating domain, > on which memory hotplug was used, to restore on destination host > because current size of memory allocated for machine is larger than > size of memory allocated at startup by memory option. Yes, it is > memory option not maxmem option. However, it is not important here > because I think that generic behavior of xl save, migrate and migrate-receive > should be changed (fix for memory hotplug case is workaround for the > generic problem which will return sooner or later). I think that xl save, > migrate and migrate-receive should use current machine state and __CURRENT__ > config (from xenstore ???) to do their tasks. However, I am aware that > this change could have large impact on current users. That is why I decided > to ask you about your opinion and suggested solutions in that case > (in general not memory hotplug only). > > Currently, these problems could be workaround by passing > path to config file with current config to xl command. > > I have done tests on Xen Ver. 4.1.2-rc3. I have not done tests > on xm command, however, I suppose that it has similar behavior. Hi Daniel -- In a recent internal discussion at Oracle, we were thinking about whether to enable hotplug functionality in a guest kernel and it raised some concerns about manageability. I think right now the system administrator of the guest can arbitrarily increase memory size beyond maxmem... that is really the whole point of your implementation, right? But this may be unacceptable to the "data center administrator" (the admin who runs the "cloud" and determines such things as vcpus and maxmem across all guests) since multiple guests may try to do this semi-maliciously to grab as much RAM as they can. And Xen has no way to discourage this, so will just hand out the RAM first-come-first-serve, right? I was thinking one way to handle this problem would be to have a new vm.cfg parameter, e.g. "maxmem_hotplug". If unspecified (or zero), there are no constraints placed on the guest. If specified (in MB), Xen/xl will disallow hotplug memory requests beyond this maximum. I suspect, if implemented properly, this might also eliminate your live migration issue. Apologies if something like this was previously discussed or is already working in your implementation. Dan P.S. Also FYI, selfballooning is implemented in Oracle's kernel so we should work to ensure that selfballooning and hotplug work properly together. _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxxxxxxxx http://lists.xensource.com/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |