[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [Xen-devel] [RFC XEN PATCH v4 40/41] tools/libxl: initiate PMEM mapping via QMP callback
The base guest physical address of each vNVDIMM device is decided by QEMU. Add a QMP callback to get the base address from QEMU and query Xen hypervisor to map host PMEM pages to that address. Signed-off-by: Haozhong Zhang <haozhong.zhang@xxxxxxxxx> --- Cc: Ian Jackson <ian.jackson@xxxxxxxxxxxxx> Cc: Wei Liu <wei.liu2@xxxxxxxxxx> --- tools/libxl/libxl_internal.h | 6 ++ tools/libxl/libxl_nvdimm.c | 29 ++++++++++ tools/libxl/libxl_qmp.c | 129 +++++++++++++++++++++++++++++++++++++++++++ 3 files changed, 164 insertions(+) diff --git a/tools/libxl/libxl_internal.h b/tools/libxl/libxl_internal.h index bfa95d8619..8c66b0e93b 100644 --- a/tools/libxl/libxl_internal.h +++ b/tools/libxl/libxl_internal.h @@ -4349,6 +4349,12 @@ static inline bool libxl__string_is_default(char **s) { return *s == NULL; } + +#if defined(__linux__) +int libxl_vnvdimm_add_pages(libxl__gc *gc, uint32_t domid, + xen_pfn_t mfn, xen_pfn_t gpfn, xen_pfn_t nr_pages); +#endif /* __linux__ */ + #endif /* diff --git a/tools/libxl/libxl_nvdimm.c b/tools/libxl/libxl_nvdimm.c index 1863d76bbc..a2c70fc7e6 100644 --- a/tools/libxl/libxl_nvdimm.c +++ b/tools/libxl/libxl_nvdimm.c @@ -196,3 +196,32 @@ int libxl_vnvdimm_copy_config(libxl_ctx *ctx, GC_FREE; return rc; } + +#if defined(__linux__) + +int libxl_vnvdimm_add_pages(libxl__gc *gc, uint32_t domid, + xen_pfn_t mfn, xen_pfn_t gpfn, xen_pfn_t nr_pages) +{ + unsigned int nr; + int ret; + + while (nr_pages) { + nr = min(nr_pages, (unsigned long)UINT_MAX); + + ret = xc_domain_populate_pmem_map(CTX->xch, domid, mfn, gpfn, nr); + if (ret && ret != -ERESTART) { + LOG(ERROR, "failed to map PMEM pages, mfn 0x%" PRI_xen_pfn ", " + "gpfn 0x%" PRI_xen_pfn ", nr_pages %u, err %d", + mfn, gpfn, nr, ret); + break; + } + + nr_pages -= nr; + mfn += nr; + gpfn += nr; + } + + return ret; +} + +#endif /* __linux__ */ diff --git a/tools/libxl/libxl_qmp.c b/tools/libxl/libxl_qmp.c index e1eb47c1d2..d471f36870 100644 --- a/tools/libxl/libxl_qmp.c +++ b/tools/libxl/libxl_qmp.c @@ -1170,6 +1170,127 @@ int libxl_qemu_monitor_command(libxl_ctx *ctx, uint32_t domid, return rc; } +#if defined(__linux__) + +static int qmp_register_vnvdimm_callback(libxl__qmp_handler *qmp, + const libxl__json_object *o, + void *arg) +{ + GC_INIT(qmp->ctx); + const libxl_domain_config *guest_config = arg; + const libxl_device_vnvdimm *vnvdimm; + const libxl__json_object *obj, *sub_map, *sub_obj; + const char *id, *expected_id; + unsigned int i, slot; + unsigned long gpa, size, mfn, gpfn, nr_pages; + int rc = 0; + + for (i = 0; (obj = libxl__json_array_get(o, i)); i++) { + if (!libxl__json_object_is_map(obj)) + continue; + + sub_map = libxl__json_map_get("data", obj, JSON_MAP); + if (!sub_map) + continue; + + sub_obj = libxl__json_map_get("slot", sub_map, JSON_INTEGER); + slot = libxl__json_object_get_integer(sub_obj); + if (slot > guest_config->num_vnvdimms) { + LOG(ERROR, + "Invalid QEMU memory device slot %u, expecting less than %u", + slot, guest_config->num_vnvdimms); + rc = -ERROR_INVAL; + goto out; + } + vnvdimm = &guest_config->vnvdimms[slot]; + + /* + * Double check whether it's a NVDIMM memory device, through + * all memory devices in QEMU on Xen are for vNVDIMM. + */ + expected_id = libxl__sprintf(gc, "xen_nvdimm%u", slot + 1); + if (!expected_id) { + LOG(ERROR, "Cannot build device id"); + rc = -ERROR_FAIL; + goto out; + } + sub_obj = libxl__json_map_get("id", sub_map, JSON_STRING); + id = libxl__json_object_get_string(sub_obj); + if (!id || strncmp(id, expected_id, strlen(expected_id))) { + LOG(ERROR, + "Invalid QEMU memory device id %s, expecting %s", + id, expected_id); + rc = -ERROR_FAIL; + goto out; + } + + sub_obj = libxl__json_map_get("addr", sub_map, JSON_INTEGER); + gpa = libxl__json_object_get_integer(sub_obj); + sub_obj = libxl__json_map_get("size", sub_map, JSON_INTEGER); + size = libxl__json_object_get_integer(sub_obj); + if ((gpa | size) & ~XC_PAGE_MASK) { + LOG(ERROR, + "Invalid address 0x%lx or size 0x%lx of QEMU memory device %s, " + "not aligned to 0x%lx", + gpa, size, id, XC_PAGE_SIZE); + rc = -ERROR_INVAL; + goto out; + } + gpfn = gpa >> XC_PAGE_SHIFT; + + nr_pages = size >> XC_PAGE_SHIFT; + if (nr_pages > vnvdimm->nr_pages) { + LOG(ERROR, + "Invalid size 0x%lx of QEMU memory device %s, " + "expecting no larger than 0x%lx", + size, id, vnvdimm->nr_pages << XC_PAGE_SHIFT); + rc = -ERROR_INVAL; + goto out; + } + + switch (vnvdimm->backend_type) { + case LIBXL_VNVDIMM_BACKEND_TYPE_MFN: + mfn = vnvdimm->u.mfn; + break; + + default: + LOG(ERROR, "Invalid NVDIMM backend type %u", vnvdimm->backend_type); + rc = -ERROR_INVAL; + goto out; + } + + rc = libxl_vnvdimm_add_pages(gc, qmp->domid, mfn, gpfn, nr_pages); + if (rc) { + LOG(ERROR, + "Cannot map PMEM pages for QEMU memory device %s, " + "mfn 0x%lx, gpfn 0x%lx, nr 0x%lx, rc %d", + id, mfn, gpfn, nr_pages, rc); + rc = -ERROR_FAIL; + goto out; + } + } + + out: + GC_FREE; + return rc; +} + +static int libxl__qmp_query_vnvdimms(libxl__qmp_handler *qmp, + const libxl_domain_config *guest_config) +{ + int rc; + GC_INIT(qmp->ctx); + + rc = qmp_synchronous_send(qmp, "query-memory-devices", NULL, + qmp_register_vnvdimm_callback, + (void *)guest_config, qmp->timeout); + + GC_FREE; + return rc; +} + +#endif /* __linux__ */ + int libxl__qmp_initializations(libxl__gc *gc, uint32_t domid, const libxl_domain_config *guest_config) { @@ -1189,6 +1310,14 @@ int libxl__qmp_initializations(libxl__gc *gc, uint32_t domid, if (!ret) { ret = qmp_query_vnc(qmp); } + +#if defined(__linux__) + if (!ret && guest_config->num_vnvdimms) { + ignore_error = false; + ret = libxl__qmp_query_vnvdimms(qmp, guest_config); + } +#endif /* __linux__ */ + libxl__qmp_close(qmp); out: -- 2.15.1 _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxxxxxxxxx https://lists.xenproject.org/mailman/listinfo/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |