[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]
Re: [Xen-devel] [RFC XEN PATCH 16/16] tools/libxl: initiate pmem mapping via qmp callback
On 01/27/17 17:13 -0500, Konrad Rzeszutek Wilk wrote:
On Mon, Oct 10, 2016 at 08:32:35AM +0800, Haozhong Zhang wrote:
QMP command 'query-nvdimms' is used by libxl to get the backend, the
guest SPA and size of each vNVDIMM device, and then libxl starts mapping
backend to guest for each vNVDIMM device.
Signed-off-by: Haozhong Zhang <haozhong.zhang@xxxxxxxxx>
---
Cc: Ian Jackson <ian.jackson@xxxxxxxxxxxxx>
Cc: Wei Liu <wei.liu2@xxxxxxxxxx>
---
tools/libxl/libxl_qmp.c | 64 +++++++++++++++++++++++++++++++++++++++++++++++++
1 file changed, 64 insertions(+)
diff --git a/tools/libxl/libxl_qmp.c b/tools/libxl/libxl_qmp.c
index f8addf9..02edd09 100644
--- a/tools/libxl/libxl_qmp.c
+++ b/tools/libxl/libxl_qmp.c
@@ -26,6 +26,7 @@
#include "_libxl_list.h"
#include "libxl_internal.h"
+#include "libxl_nvdimm.h"
/* #define DEBUG_RECEIVED */
@@ -1146,6 +1147,66 @@ out:
return rc;
}
+static int qmp_register_nvdimm_callback(libxl__qmp_handler *qmp,
+ const libxl__json_object *o,
+ void *unused)
+{
+ GC_INIT(qmp->ctx);
+ const libxl__json_object *obj = NULL;
+ const libxl__json_object *sub_obj = NULL;
+ int i = 0;
unsigned int.
will fix
Thanks,
Haozhong
+ const char *mem_path;
+ uint64_t slot, spa, length;
+ int ret = 0;
+
+ for (i = 0; (obj = libxl__json_array_get(o, i)); i++) {
+ if (!libxl__json_object_is_map(obj))
+ continue;
+
+ sub_obj = libxl__json_map_get("slot", obj, JSON_INTEGER);
+ slot = libxl__json_object_get_integer(sub_obj);
+
+ sub_obj = libxl__json_map_get("mem-path", obj, JSON_STRING);
+ mem_path = libxl__json_object_get_string(sub_obj);
+ if (!mem_path) {
+ LOG(ERROR, "No mem-path is specified for NVDIMM #%" PRId64, slot);
+ ret = -EINVAL;
+ goto out;
+ }
+
+ sub_obj = libxl__json_map_get("spa", obj, JSON_INTEGER);
+ spa = libxl__json_object_get_integer(sub_obj);
+
+ sub_obj = libxl__json_map_get("length", obj, JSON_INTEGER);
+ length = libxl__json_object_get_integer(sub_obj);
+
+ LOG(DEBUG,
+ "vNVDIMM #%" PRId64 ": %s, spa 0x%" PRIx64 ", length 0x%" PRIx64,
+ slot, mem_path, spa, length);
+
+ ret = libxl_nvdimm_add_device(gc, qmp->domid, mem_path, spa, length);
+ if (ret) {
+ LOG(ERROR,
+ "Failed to add NVDIMM #%" PRId64
+ "(mem_path %s, spa 0x%" PRIx64 ", length 0x%" PRIx64 ") "
+ "to domain %d (err = %d)",
+ slot, mem_path, spa, length, qmp->domid, ret);
+ goto out;
+ }
+ }
+
+ out:
+ GC_FREE;
+ return ret;
+}
+
+static int libxl__qmp_query_nvdimms(libxl__qmp_handler *qmp)
+{
+ return qmp_synchronous_send(qmp, "query-nvdimms", NULL,
+ qmp_register_nvdimm_callback,
+ NULL, qmp->timeout);
+}
+
int libxl__qmp_hmp(libxl__gc *gc, int domid, const char *command_line,
char **output)
{
@@ -1187,6 +1248,9 @@ int libxl__qmp_initializations(libxl__gc *gc, uint32_t
domid,
if (!ret) {
ret = qmp_query_vnc(qmp);
}
+ if (!ret && guest_config->num_vnvdimms) {
+ ret = libxl__qmp_query_nvdimms(qmp);
+ }
libxl__qmp_close(qmp);
return ret;
}
--
2.10.1
_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxx
https://lists.xen.org/xen-devel
_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxx
https://lists.xen.org/xen-devel
|