|
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] [PATCH 0/4] [HVM] NUMA support in HVM guests
Anthony, thanks for looking into the patches, I appreciate your comments. This is true, I encountered this before, but didn't want to wait longer
for sending up the patches. Actually the "numanodes=n" config file
option shouldn't specify the number of nodes, but a list of specific
nodes to use, like "numanodes=0,2" to pin the domain on the first and
the third node.
This sounds interesting. It shouldn't be that hard to do this in libxc, but we should think about a semantic to specify this behavior in the config file (if we change the semantic from the number to specific node like I described above).It may be nice to pin node from the lightest overhead node. >than the number of pcpus on pnode. Otherwise vcpus belonging to a domain runWe also need to add some limitations for numanodes. The number of vcpus on vnode should not be larger > on the same pcpu, which is not what we want.Would be nice, but in the moment I would push this into the sysadmin's responsibility. In setup_numa_mem, each node has even memory size, if the memory allocation fails, >the domain creation fails. This may be too "rude", I think we can support guest> NUMA with each node has different memory size, even more, and maybe some node doesn't have memory. What we need guarantee is guest see physical topology. Sound reasonable. I will look into this. >the workload on the platform may be very imbalanced, we need a method to dynamically balance workload.In your patch, when create NUMA guest, vnode is pinned to pnode. While after some creations and destroys domain operation, You are right, this may become a problem. I think the second solution is easier to implement. A NUMA-aware scheduler would be nice, but my idea was that the guest OS can better schedule (more fine-grained on a per-process base than on a per-machine base) things. Changing the processing node without moving the memory along should be an exception (as it changes NUMA topology and in the moment I don't see methods to propagate this nicely to the (HVM) guest), so I think a kind of "real-emergency balancer" which includes page-migration (quite expensive with bigger memory sizes!) would be more appropriate.There are two methods IMO. 1. Implement NUMA-aware scheduler and page migration 2. Run a daemon in dom0, this daemon monitors workload, and use live-migration to balance workload if necessary. After all my patches were more a discussion base than a final solution, so I see there is more work to do. In the moment I am working on including PV guests. Regards, Andre. -- Andre Przywara AMD-Operating System Research Center (OSRC), Dresden, Germany Tel: +49 351 277-84917 ----to satisfy European Law for business letters: AMD Saxony Limited Liability Company & Co. KGSitz (Geschäftsanschrift): Wilschdorfer Landstr. 101, 01109 Dresden, Deutschland Registergericht Dresden: HRA 4896vertretungsberechtigter Komplementär: AMD Saxony LLC (Sitz Wilmington, Delaware, USA) Geschäftsführer der AMD Saxony LLC: Dr. Hans-R. Deppe, Thomas McCoy _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxxxxxxxx http://lists.xensource.com/xen-devel
|
![]() |
Lists.xenproject.org is hosted with RackSpace, monitoring our |