|
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [Xen-devel] [PATCHv1 2/4] x86/domain: Compile with lock_profile=y enabled.
From: Konrad Rzeszutek Wilk <konrad.wilk@xxxxxxxxxx>
Our 'struct domain' has when lock profiling is enabled is bigger than
one page.
We can't use vmap nor vzalloc as both of those stash the
physical address in struct page which makes the assumptions
in 'arch_init_memory' trip over ASSERTs.
Signed-off-by: Konrad Rzeszutek Wilk <konrad.wilk@xxxxxxxxxx>
Signed-off-by: David Vrabel <david.vrabel@xxxxxxxxxx>
---
xen/arch/x86/domain.c | 13 +++++++++++--
1 file changed, 11 insertions(+), 2 deletions(-)
diff --git a/xen/arch/x86/domain.c b/xen/arch/x86/domain.c
index 2c3bb09..40d9d7c 100644
--- a/xen/arch/x86/domain.c
+++ b/xen/arch/x86/domain.c
@@ -231,6 +231,7 @@ static unsigned int __init noinline
_domain_struct_bits(void)
struct domain *alloc_domain_struct(void)
{
struct domain *d;
+ unsigned int order = get_order_from_bytes(sizeof(*d));
#ifdef CONFIG_BIGMEM
const unsigned int bits = 0;
#else
@@ -244,10 +245,18 @@ struct domain *alloc_domain_struct(void)
bits = _domain_struct_bits();
#endif
+
+#ifndef LOCK_PROFILE
BUILD_BUG_ON(sizeof(*d) > PAGE_SIZE);
- d = alloc_xenheap_pages(0, MEMF_bits(bits));
+#endif
+ d = alloc_xenheap_pages(order, MEMF_bits(bits));
if ( d != NULL )
- clear_page(d);
+ {
+ unsigned int sz;
+
+ for ( sz = 0; sz < (PAGE_SIZE << order); sz += PAGE_SIZE )
+ clear_page((void *)d + sz);
+ }
return d;
}
--
2.1.4
_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxx
http://lists.xen.org/xen-devel
|
![]() |
Lists.xenproject.org is hosted with RackSpace, monitoring our |