[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] RE: [Xen-devel] Error restoring DomU when using GPLPV
On further debugging, it appears that the p2m_size may be OK, but there's something about those 24 "magic" gpfns that isn't quite right. > -----Original Message----- > From: Dan Magenheimer > Sent: Friday, September 04, 2009 3:29 PM > To: Wayne Gong; Annie Li; Keir Fraser > Cc: Joshua West; James Harper; xen-devel@xxxxxxxxxxxxxxxxxxx > Subject: RE: [Xen-devel] Error restoring DomU when using GPLPV > > > I think I've tracked down the cause of this problem > in the hypervisor, but am unsure how to best fix it. > > In tools/libxc/xc_domain_save.c, the static variable p2m_size > is said to be "number of pfns this guest has (i.e. number of > entries in the P2M)". But apparently p2m_size is getting > set to a very large number (0x100000) regardless of the > maximum psuedophysical memory for the hvm guest. As a result, > some "magic" pages in the 0xf0000-0xfefff range are getting > placed in the save file. But since they are not "real" > pages, the restore process runs beyond the maximum number > of physical pages allowed for the domain and fails. > (The gpfn of the last 24 pages saved are f2020, fc000-fc012, > feffb, feffc, feffd, feffe.) > > p2m_size is set in "save" with a call to a memory_op hypercall > (XENMEM_maximum_gpfn) which for an hvm domain returns > d->arch.p2m->max_mapped_pfn. I suspect that the meaning > of max_mapped_pfn changed at some point to more match > its name, but this changed the semantics of the hypercall > as used by xc_domain_restore, resulting in this curious > problem. > > Any thoughts on how to fix this? > > > -----Original Message----- > > From: Annie Li > > Sent: Tuesday, September 01, 2009 10:27 PM > > To: Keir Fraser > > Cc: Joshua West; James Harper; xen-devel@xxxxxxxxxxxxxxxxxxx > > Subject: Re: [Xen-devel] Error restoring DomU when using GPLPV > > > > > > > > > It seems this problem is connected with gnttab, not shareinfo. > > > I changed some code about grant table in winpv driver (not using > > > balloon down shinfo+gnttab method), > save/restore/migration can work > > > properly on Xen3.4 now. > > > > > > What i changed is winpv driver use hypercall > > XENMEM_add_to_physmap to > > > map corresponding grant tables which devices require, instead of > > > mapping all 32 pages grant table during initialization. It seems > > > those extra grant table mapping cause this problem. > > > > Wondering whether those extra grant table mapping is the root > > cause of > > the migration problem? or by luck as linux PVHVM too? > > > > Thanks > > Annie. > > > > > > _______________________________________________ > > Xen-devel mailing list > > Xen-devel@xxxxxxxxxxxxxxxxxxx > > http://lists.xensource.com/xen-devel > > > > _______________________________________________ > Xen-devel mailing list > Xen-devel@xxxxxxxxxxxxxxxxxxx > http://lists.xensource.com/xen-devel > _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxxxxxxxx http://lists.xensource.com/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |