[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] [PATCH RESEND v7 7/9] libxc: allocate domain memory for vnuma enabled
On Thu, Aug 21, 2014 at 01:10:27AM -0400, Elena Ufimtseva wrote: > vNUMA-aware domain memory allocation based on provided > vnode to pnode map. If this map is not defined, use > default allocation. Default allocation will not specify > any physical node when allocating memory. > Domain creation will fail if at least one node was not defined. > > Signed-off-by: Elena Ufimtseva <ufimtseva@xxxxxxxxx> > --- > tools/libxc/xc_dom.h | 13 ++++++++ > tools/libxc/xc_dom_x86.c | 76 > ++++++++++++++++++++++++++++++++++++++-------- > 2 files changed, 77 insertions(+), 12 deletions(-) > > diff --git a/tools/libxc/xc_dom.h b/tools/libxc/xc_dom.h > index 6ae6a9f..61c2a06 100644 > --- a/tools/libxc/xc_dom.h > +++ b/tools/libxc/xc_dom.h > @@ -164,6 +164,16 @@ struct xc_dom_image { > > /* kernel loader */ > struct xc_dom_arch *arch_hooks; > + > + /* > + * vNUMA topology and memory allocation structure. > + * Defines the way to allocate memory on per NUMA > + * physical defined by vnode_to_pnode. > + */ > + uint32_t vnodes; > + uint64_t *numa_memszs; > + unsigned int *vnode_to_pnode; > + > /* allocate up to virt_alloc_end */ > int (*allocate) (struct xc_dom_image * dom, xen_vaddr_t up_to); > }; > @@ -385,6 +395,9 @@ static inline xen_pfn_t xc_dom_p2m_guest(struct > xc_dom_image *dom, > int arch_setup_meminit(struct xc_dom_image *dom); > int arch_setup_bootearly(struct xc_dom_image *dom); > int arch_setup_bootlate(struct xc_dom_image *dom); > +int arch_boot_alloc(struct xc_dom_image *dom); > + > +#define LIBXC_VNUMA_NO_NODE ~((unsigned int)0) > > /* > * Local variables: > diff --git a/tools/libxc/xc_dom_x86.c b/tools/libxc/xc_dom_x86.c > index bf06fe4..3d29c38 100644 > --- a/tools/libxc/xc_dom_x86.c > +++ b/tools/libxc/xc_dom_x86.c > @@ -759,7 +759,7 @@ static int x86_shadow(xc_interface *xch, domid_t domid) > int arch_setup_meminit(struct xc_dom_image *dom) > { > int rc; > - xen_pfn_t pfn, allocsz, i, j, mfn; > + xen_pfn_t pfn, i, j, mfn; > > rc = x86_compat(dom->xch, dom->guest_domid, dom->guest_type); > if ( rc ) > @@ -811,25 +811,77 @@ int arch_setup_meminit(struct xc_dom_image *dom) > /* setup initial p2m */ > for ( pfn = 0; pfn < dom->total_pages; pfn++ ) > dom->p2m_host[pfn] = pfn; > + > + /* > + * Any PV domain should have at least one vNUMA node. > + * If no config was defined, one default vNUMA node > + * will be set. Panic if otherwise. Could you explain you rationale of why we want to crash please? > + */ > + if ( dom->vnodes == 0 ) { > + xc_dom_panic(dom->xch, XC_INTERNAL_ERROR, > + "%s: Cannot construct vNUMA topology with 0 > vnodes\n", > + __FUNCTION__); > + return -EINVAL; > + } > > /* allocate guest memory */ > - for ( i = rc = allocsz = 0; > - (i < dom->total_pages) && !rc; > - i += allocsz ) > - { > - allocsz = dom->total_pages - i; > - if ( allocsz > 1024*1024 ) > - allocsz = 1024*1024; > - rc = xc_domain_populate_physmap_exact( > - dom->xch, dom->guest_domid, allocsz, > - 0, 0, &dom->p2m_host[i]); > - } > + rc = arch_boot_alloc(dom); > + if ( rc ) > + return rc; > > /* Ensure no unclaimed pages are left unused. > * OK to call if hadn't done the earlier claim call. */ > (void)xc_domain_claim_pages(dom->xch, dom->guest_domid, > 0 /* cancels the claim */); > } > + return rc; > +} > + > +/* > + * Allocates domain memory taking into account > + * defined vnuma topology and vnode_to_pnode map. > + * Any pv guest will have at least one vnuma node > + * with vnuma_memszs[0] = domain memory and the rest > + * topology initialized with default values. > + */ > +int arch_boot_alloc(struct xc_dom_image *dom) > +{ > + int rc; > + unsigned int n, memflags; > + unsigned long long vnode_pages; > + unsigned long long allocsz = 0, node_pfn_base, i; > + > + rc = allocsz = node_pfn_base = n = 0; > + > + for ( n = 0; n < dom->vnodes; n++ ) > + { > + memflags = 0; > + if ( dom->vnode_to_pnode[n] != LIBXC_VNUMA_NO_NODE ) > + { > + memflags |= XENMEMF_exact_node(dom->vnode_to_pnode[n]); > + memflags |= XENMEMF_exact_node_request; > + } > + /* memeszs are in megabytes, calc pages from it for this node. */ > + vnode_pages = (dom->numa_memszs[n] << 20) >> PAGE_SHIFT_X86; > + for ( i = 0; i < vnode_pages; i += allocsz ) > + { > + allocsz = vnode_pages - i; > + if ( allocsz > 1024*1024 ) > + allocsz = 1024*1024; > + > + rc = xc_domain_populate_physmap_exact(dom->xch, dom->guest_domid, > + allocsz, 0, memflags, > + &dom->p2m_host[node_pfn_base + > i]); > + if ( rc ) > + { > + xc_dom_panic(dom->xch, XC_INTERNAL_ERROR, > + "%s: Failed allocation of %Lu pages for vnode %d on > pnode %d out of %lu\n", > + __FUNCTION__, vnode_pages, n, > dom->vnode_to_pnode[n], dom->total_pages); > + return rc; > + } > + } > + node_pfn_base += i; > + } > > return rc; > } > -- > 1.7.10.4 > _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxx http://lists.xen.org/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |