[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [Xen-devel] [RFC XEN PATCH v2 00/15] Add vNVDIMM support to HVM domains
This is v2 RFC patch series to add vNVDIMM support to HVM domains. v1 can be found at https://lists.xenproject.org/archives/html/xen-devel/2016-10/msg00424.html. No label and no _DSM except function 0 "query implemented functions" is supported by this version, but they will be added by future patches. The corresponding Qemu patch series is sent in another thread "[RFC QEMU PATCH v2 00/10] Implement vNVDIMM for Xen HVM guest". All patch series can be found at Xen: https://github.com/hzzhan9/xen.git nvdimm-rfc-v2 Qemu: https://github.com/hzzhan9/qemu.git xen-nvdimm-rfc-v2 Changes in v2 ============== - One of the primary changes in v2 is dropping the linux kernel patches, which were used to reserve on host pmem for placing its frametable and M2P table. In v2, we add a management tool xen-ndctl which is used in Dom0 to notify Xen hypervisor of which storage can be used to manage the host pmem. For example, 1. xen-ndctl setup 0x240000 0x380000 0x380000 0x3c0000 tells Xen hypervisor to use host pmem pages at MFN 0x380000 ~ 0x3c0000 to manage host pmem pages at MFN 0x240000 ~ 0x380000. I.e. the former is used to place the frame table and M2P table of both ranges of pmem pages. 2. xen-ndctl setup 0x240000 0x380000 tells Xen hypervisor to use the regular RAM to manage the host pmem pages at MFN 0x240000 ~ 0x380000. I.e the regular RMA is used to place the frame table and M2P table. - Another primary change in v2 is dropping the support to map files on the host pmem to HVM domains as virtual NVDIMMs, as I cannot find a stable to fix the fiemap of host files. Instead, we can rely on the ability added in Linux kernel v4.9 that enables creating multiple pmem namespaces on a single nvdimm interleave set. - Other changes are logged in each patch separately. How to Test ============== 0. This patch series can be tested either on the real hardware with NVDIMM, or in the nested virtualization environment on KVM. The latter requires QEMU 2.9 or newer with, for example, following commands and options, # dd if=/dev/zero of=nvm-8G.img bs=1G count=8 # rmmod kvm-intel; modprobe kvm-intel nested=1 # qemu-system-x86_64 -enable-kvm -smp 4 -cpu host,+vmx \ -hda DISK_IMG_OF_XEN \ -machine pc,nvdimm \ -m 8G,slots=4,maxmem=128G \ -object memory-backend-file,id=mem1,mem-path=nvm-8G,size=8G \ -device nvdimm,id=nv1,memdev=mem1,label-size=2M \ ... Above will create a nested virtualization environment with a 8G pmem mode NVDIMM device (whose last 2MB is used as the label storage area). 1. Check out Xen and QEMU from above repositories and branches. Build and install Xen with qemu-xen replaced by above QEMU. 2. Build and install Linux kernel 4.9 or later as Dom0 kernel with the following configs selected: CONFIG_ACPI_NFIT CONFIG_LIBNVDIMM CONFIG_BLK_DEV_PMEM CONFIG_NVDIMM_PFN CONFIG_FS_DAX 3. Check out ndctl from https://github.com/pmem/ndctl.git. Build and install ndctl in Dom0. 4. Boot to Xen Dom0. 5. Create pmem namespaces on the host pmem region. # ndctl disable-region region0 # ndctl zero-labels nmem0 // clear existing labels # ndctl init-labels nmem0 // initialize the label area # ndctl enable-region region0 # ndctl create-namespace -r region0 -s 4G -m raw // create one 4G pmem namespace # ndctl create-namespace -r region0 -s 1G -m raw // create one 1G pmem namespace # ndctl list --namespaces [ { "dev":"namespace0.0", "mode":"raw", "size":4294967296, "uuid":"bbfbedbd-3ada-4f55-9484-01f2722c651b", "blockdev":"pmem0" }, { "dev":"namespace0.1", "mode":"raw", "size":1073741824, "uuid":"dd4d3949-6887-417b-b819-89a7854fcdbd", "blockdev":"pmem0.1" } ] 6. Ask Xen hypervisor to use namespace0.1 to manage namespace0.0. # grep namespace /proc/iomem 240000000-33fffffff : namespace0.0 340000000-37fffffff : namespace0.1 # xen-ndctl setup 0x240000 0x340000 0x340000 0x380000 7. Start a HVM domain with "vnvdimms=[ '/dev/pmem0' ]" in its xl config. If ndctl is installed in HVM domain, "ndctl list" should be able to list a 4G pmem namespace, e.g. { "dev":"namespace0.0", "mode":"raw", "size":4294967296, "blockdev":"pmem0" } Haozhong Zhang (15): xen/common: add Kconfig item for pmem support xen: probe pmem regions via ACPI NFIT xen/x86: allow customizing locations of extended frametable & M2P xen/x86: add XEN_SYSCTL_nvdimm_pmem_setup to setup host pmem xen/x86: add XENMEM_populate_pmemmap to map host pmem pages to HVM domain tools: reserve guest memory for ACPI from device model tools/libacpi: expose the minimum alignment used by mem_ops.alloc tools/libacpi: add callback acpi_ctxt.p2v to get a pointer from physical address tools/libacpi: add callbacks to access XenStore tools/libacpi: add a simple AML builder tools/libacpi: load ACPI built by the device model tools/libxl: build qemu options from xl vNVDIMM configs tools/libxl: add support to map host pmem device to guests tools/libxl: initiate pmem mapping via qmp callback tools/misc: add xen-ndctl .gitignore | 1 + docs/man/xl.cfg.pod.5.in | 6 + tools/firmware/hvmloader/Makefile | 3 +- tools/firmware/hvmloader/util.c | 75 ++++++ tools/firmware/hvmloader/util.h | 10 + tools/firmware/hvmloader/xenbus.c | 44 +++- tools/flask/policy/modules/dom0.te | 2 +- tools/flask/policy/modules/xen.if | 2 +- tools/libacpi/acpi2_0.h | 2 + tools/libacpi/aml_build.c | 326 +++++++++++++++++++++++ tools/libacpi/aml_build.h | 116 +++++++++ tools/libacpi/build.c | 311 ++++++++++++++++++++++ tools/libacpi/libacpi.h | 21 ++ tools/libxc/include/xc_dom.h | 1 + tools/libxc/include/xenctrl.h | 36 +++ tools/libxc/xc_dom_x86.c | 7 + tools/libxc/xc_domain.c | 15 ++ tools/libxc/xc_misc.c | 17 ++ tools/libxl/Makefile | 7 +- tools/libxl/libxl_create.c | 4 +- tools/libxl/libxl_dm.c | 109 +++++++- tools/libxl/libxl_dom.c | 22 ++ tools/libxl/libxl_nvdimm.c | 182 +++++++++++++ tools/libxl/libxl_nvdimm.h | 42 +++ tools/libxl/libxl_qmp.c | 116 ++++++++- tools/libxl/libxl_types.idl | 8 + tools/libxl/libxl_x86_acpi.c | 41 +++ tools/misc/Makefile | 4 + tools/misc/xen-ndctl.c | 227 ++++++++++++++++ tools/xl/xl_parse.c | 16 ++ xen/arch/x86/acpi/boot.c | 4 + xen/arch/x86/domain.c | 7 + xen/arch/x86/sysctl.c | 22 ++ xen/arch/x86/x86_64/mm.c | 191 ++++++++++++-- xen/common/Kconfig | 9 + xen/common/Makefile | 1 + xen/common/compat/memory.c | 1 + xen/common/domain.c | 3 + xen/common/memory.c | 43 +++ xen/common/pmem.c | 448 ++++++++++++++++++++++++++++++++ xen/drivers/acpi/Makefile | 2 + xen/drivers/acpi/nfit.c | 116 +++++++++ xen/include/acpi/actbl.h | 1 + xen/include/acpi/actbl1.h | 42 +++ xen/include/public/hvm/hvm_xs_strings.h | 11 + xen/include/public/memory.h | 14 +- xen/include/public/sysctl.h | 29 ++- xen/include/xen/acpi.h | 4 + xen/include/xen/pmem.h | 66 +++++ xen/include/xen/sched.h | 3 + xen/include/xsm/dummy.h | 11 + xen/include/xsm/xsm.h | 12 + xen/xsm/dummy.c | 4 + xen/xsm/flask/hooks.c | 17 ++ xen/xsm/flask/policy/access_vectors | 4 + 55 files changed, 2795 insertions(+), 43 deletions(-) create mode 100644 tools/libacpi/aml_build.c create mode 100644 tools/libacpi/aml_build.h create mode 100644 tools/libxl/libxl_nvdimm.c create mode 100644 tools/libxl/libxl_nvdimm.h create mode 100644 tools/misc/xen-ndctl.c create mode 100644 xen/common/pmem.c create mode 100644 xen/drivers/acpi/nfit.c create mode 100644 xen/include/xen/pmem.h -- 2.12.0 _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxx https://lists.xen.org/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |