|
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] [PATCH v6 06/23] libxc: allocate memory with vNUMA information for PV guest
On Thu, 2015-02-26 at 15:55 +0000, Wei Liu wrote:
> From libxc's point of view, it only needs to know vnode to pnode mapping
> and size of each vnode to allocate memory accordingly. Add these fields
> to xc_dom structure.
>
> The caller might not pass in vNUMA information. In that case, a dummy
> layout is generated for the convenience of libxc's allocation code. The
> upper layer (libxl etc) still sees the domain has no vNUMA
> configuration.
>
> Note that for this patch on PV x86 guest can have multiple regions of
> ram allocated.
>
> Signed-off-by: Wei Liu <wei.liu2@xxxxxxxxxx>
Acked-by: Ian Campbell <ian.campbell@xxxxxxxxxx>
> Cc: Ian Jackson <ian.jackson@xxxxxxxxxxxxx>
> Cc: Dario Faggioli <dario.faggioli@xxxxxxxxxx>
> Cc: Elena Ufimtseva <ufimtseva@xxxxxxxxx>
> ---
> Changes in v6:
> 1. Ditch XC_VNUMA_NO_NODE and use XEN_NUMA_NO_NODE.
> 2. Update comment in xc_dom.h.
>
> Changes in v5:
> 1. Ditch xc_vnuma_info.
>
> Changes in v4:
> 1. Pack fields into a struct.
> 2. Use "page" as unit.
> 3. __FUNCTION__ -> __func__.
> 4. Don't print total_pages.
> 5. Improve comment.
>
> Changes in v3:
> 1. Rewrite commit log.
> 2. Shorten some error messages.
> ---
> tools/libxc/include/xc_dom.h | 12 ++++-
> tools/libxc/xc_dom_x86.c | 101
> +++++++++++++++++++++++++++++++++++++------
> 2 files changed, 97 insertions(+), 16 deletions(-)
>
> diff --git a/tools/libxc/include/xc_dom.h b/tools/libxc/include/xc_dom.h
> index 6b8ddf4..a7d059a 100644
> --- a/tools/libxc/include/xc_dom.h
> +++ b/tools/libxc/include/xc_dom.h
> @@ -119,8 +119,10 @@ struct xc_dom_image {
>
> /* physical memory
> *
> - * An x86 PV guest has a single contiguous block of physical RAM,
> - * consisting of total_pages starting at rambase_pfn.
> + * An x86 PV guest has one or more blocks of physical RAM,
> + * consisting of total_pages starting at rambase_pfn. The start
> + * address and size of each block is controlled by vNUMA
> + * structures.
> *
> * An ARM guest has GUEST_RAM_BANKS regions of RAM, with
> * rambank_size[i] pages in each. The lowest RAM address
> @@ -168,6 +170,12 @@ struct xc_dom_image {
> struct xc_dom_loader *kernel_loader;
> void *private_loader;
>
> + /* vNUMA information */
> + xen_vmemrange_t *vmemranges;
> + unsigned int nr_vmemranges;
> + unsigned int *vnode_to_pnode;
> + unsigned int nr_vnodes;
> +
> /* kernel loader */
> struct xc_dom_arch *arch_hooks;
> /* allocate up to virt_alloc_end */
> diff --git a/tools/libxc/xc_dom_x86.c b/tools/libxc/xc_dom_x86.c
> index bea54f2..268d4db 100644
> --- a/tools/libxc/xc_dom_x86.c
> +++ b/tools/libxc/xc_dom_x86.c
> @@ -760,7 +760,8 @@ static int x86_shadow(xc_interface *xch, domid_t domid)
> int arch_setup_meminit(struct xc_dom_image *dom)
> {
> int rc;
> - xen_pfn_t pfn, allocsz, i, j, mfn;
> + xen_pfn_t pfn, allocsz, mfn, total, pfn_base;
> + int i, j;
>
> rc = x86_compat(dom->xch, dom->guest_domid, dom->guest_type);
> if ( rc )
> @@ -811,26 +812,98 @@ int arch_setup_meminit(struct xc_dom_image *dom)
> if ( rc )
> return rc;
> }
> - /* setup initial p2m */
> - dom->p2m_size = dom->total_pages;
> +
> + /* Setup dummy vNUMA information if it's not provided. Note
> + * that this is a valid state if libxl doesn't provide any
> + * vNUMA information.
> + *
> + * The dummy values make libxc allocate all pages from
> + * arbitrary physical nodes. This is the expected behaviour if
> + * no vNUMA configuration is provided to libxc.
> + *
> + * Note that the following hunk is just for the convenience of
> + * allocation code. No defaulting happens in libxc.
> + */
> + if ( dom->nr_vmemranges == 0 )
> + {
> + dom->nr_vmemranges = 1;
> + dom->vmemranges = xc_dom_malloc(dom, sizeof(*dom->vmemranges));
> + dom->vmemranges[0].start = 0;
> + dom->vmemranges[0].end = dom->total_pages << PAGE_SHIFT;
> + dom->vmemranges[0].flags = 0;
> + dom->vmemranges[0].nid = 0;
> +
> + dom->nr_vnodes = 1;
> + dom->vnode_to_pnode = xc_dom_malloc(dom,
> + sizeof(*dom->vnode_to_pnode));
> + dom->vnode_to_pnode[0] = XEN_NUMA_NO_NODE;
> + }
> +
> + total = dom->p2m_size = 0;
> + for ( i = 0; i < dom->nr_vmemranges; i++ )
> + {
> + total += ((dom->vmemranges[i].end - dom->vmemranges[i].start)
> + >> PAGE_SHIFT);
> + dom->p2m_size =
> + dom->p2m_size > (dom->vmemranges[i].end >> PAGE_SHIFT) ?
> + dom->p2m_size : (dom->vmemranges[i].end >> PAGE_SHIFT);
> + }
> + if ( total != dom->total_pages )
> + {
> + xc_dom_panic(dom->xch, XC_INTERNAL_ERROR,
> + "%s: vNUMA page count mismatch (0x%"PRIpfn" !=
> 0x%"PRIpfn")\n",
> + __func__, total, dom->total_pages);
> + return -EINVAL;
> + }
> +
> dom->p2m_host = xc_dom_malloc(dom, sizeof(xen_pfn_t) *
> dom->p2m_size);
> if ( dom->p2m_host == NULL )
> return -EINVAL;
> - for ( pfn = 0; pfn < dom->total_pages; pfn++ )
> - dom->p2m_host[pfn] = pfn;
> + for ( pfn = 0; pfn < dom->p2m_size; pfn++ )
> + dom->p2m_host[pfn] = INVALID_P2M_ENTRY;
>
> /* allocate guest memory */
> - for ( i = rc = allocsz = 0;
> - (i < dom->total_pages) && !rc;
> - i += allocsz )
> + for ( i = 0; i < dom->nr_vmemranges; i++ )
> {
> - allocsz = dom->total_pages - i;
> - if ( allocsz > 1024*1024 )
> - allocsz = 1024*1024;
> - rc = xc_domain_populate_physmap_exact(
> - dom->xch, dom->guest_domid, allocsz,
> - 0, 0, &dom->p2m_host[i]);
> + unsigned int memflags;
> + uint64_t pages;
> + unsigned int pnode = dom->vnode_to_pnode[dom->vmemranges[i].nid];
> +
> + memflags = 0;
> + if ( pnode != XEN_NUMA_NO_NODE )
> + memflags |= XENMEMF_exact_node(pnode);
> +
> + pages = (dom->vmemranges[i].end - dom->vmemranges[i].start)
> + >> PAGE_SHIFT;
> + pfn_base = dom->vmemranges[i].start >> PAGE_SHIFT;
> +
> + for ( pfn = pfn_base; pfn < pfn_base+pages; pfn++ )
> + dom->p2m_host[pfn] = pfn;
> +
> + for ( j = 0; j < pages; j += allocsz )
> + {
> + allocsz = pages - j;
> + if ( allocsz > 1024*1024 )
> + allocsz = 1024*1024;
> +
> + rc = xc_domain_populate_physmap_exact(dom->xch,
> + dom->guest_domid, allocsz, 0, memflags,
> + &dom->p2m_host[pfn_base+j]);
> +
> + if ( rc )
> + {
> + if ( pnode != XEN_NUMA_NO_NODE )
> + xc_dom_panic(dom->xch, XC_INTERNAL_ERROR,
> + "%s: failed to allocate 0x%"PRIx64"
> pages (v=%d, p=%d)\n",
> + __func__, pages, i, pnode);
> + else
> + xc_dom_panic(dom->xch, XC_INTERNAL_ERROR,
> + "%s: failed to allocate 0x%"PRIx64"
> pages\n",
> + __func__, pages);
> + return rc;
> + }
> + }
> }
>
> /* Ensure no unclaimed pages are left unused.
_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxx
http://lists.xen.org/xen-devel
|
![]() |
Lists.xenproject.org is hosted with RackSpace, monitoring our |