|
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] multicalls.c warning in xen_mc_flush
On Tue, May 29, 2012 at 01:39:39PM +0200, William Dauchy wrote:
> On Fri, May 25, 2012 at 11:01 PM, Konrad Rzeszutek Wilk
> <konrad.wilk@xxxxxxxxxx> wrote:
> > Not yet. Could you ping me in week say please?
>
> ping.
Pls try the attached patch.
>From e4c315c0c3d842712ae64ec95c099fd44e65291a Mon Sep 17 00:00:00 2001
From: Konrad Rzeszutek Wilk <konrad.wilk@xxxxxxxxxx>
Date: Tue, 29 May 2012 21:38:22 -0400
Subject: [PATCH] x86/i386: Check PSE bit before using PAGE_KERNEL_LARGE.
During bootup we would unconditionally do this on non-NUMA machines:
setup_arch
\-initmem_init
\-x86_numa_init (with dummy_init as callback)
\- init_alloc_remap
\- set_pmd_pfn (with PAGE_PSE)
without checking to see if the CPU supports PSE. This
patch adds that and also allows the init_alloc_remap function
to properly work by falling back on PTEs.
This bug has been observed when running an i386 PV Xen
guest with CONFIG_NUMA built in - but it should be also
easily observed on other CPUs which do not expose the PSE support.
We would get this in the guest:
memblock_reserve: [0x0000002ac00000-0x0000002be00000]
init_alloc_remap+0x195/0x251
------------[ cut here ]------------
WARNING: at /home/konrad/ssd/linux/arch/x86/xen/multicalls.c:129
xen_mc_flush+0x160/0x1e0()
Modules linked in:
Pid: 0, comm: swapper Not tainted 3.4.0-08268-gc0b1dd2 #1
Call Trace:
[<c107b62d>] warn_slowpath_common+0x6d/0xa0
[<c10380a0>] ? xen_mc_flush+0x160/0x1e0
[<c10380a0>] ? xen_mc_flush+0x160/0x1e0
[<c107b67d>] warn_slowpath_null+0x1d/0x20
[<c10380a0>] xen_mc_flush+0x160/0x1e0
[<c103a46d>] xen_set_pmd_hyper+0xad/0x170
[<c103896d>] ? pte_pfn_to_mfn+0xad/0xc0
[<c1074b2e>] set_pmd_pfn+0x9e/0xf0
[<c172290d>] init_alloc_remap+0x1e3/0x251
[<c1722325>] x86_numa_init+0x340/0x65e
[<c103c7fe>] ? __raw_callee_save_xen_restore_fl+0x6/0x8
[<c172265f>] initmem_init+0xb/0xd6
[<c1719428>] ? acpi_boot_table_init+0x10/0x7d
[<c1712dd1>] setup_arch+0xb9c/0xc8a
[<c103c7fe>] ? __raw_callee_save_xen_restore_fl+0x6/0x8
[<c170c8fb>] start_kernel+0xbe/0x395
[<c170c306>] i386_start_kernel+0xa9/0xb0
[<c170f86c>] xen_start_kernel+0x632/0x63a
[<c1409078>] ? tmem_objnode_alloc+0x28/0xa0
---[ end trace a7919e7f17c0a725 ]---
------------[ cut here ]------------
with the hypervisor telling us:
(XEN) mm.c:943:d0 Attempt to map superpage without allowsuperpage flag in
hypervisor
Signed-off-by: Konrad Rzeszutek Wilk <konrad.wilk@xxxxxxxxxx>
---
arch/x86/mm/pgtable_32.c | 29 ++++++++++++++++++++++++++++-
1 files changed, 28 insertions(+), 1 deletions(-)
diff --git a/arch/x86/mm/pgtable_32.c b/arch/x86/mm/pgtable_32.c
index a69bcb8..32085ec 100644
--- a/arch/x86/mm/pgtable_32.c
+++ b/arch/x86/mm/pgtable_32.c
@@ -86,7 +86,34 @@ void set_pmd_pfn(unsigned long vaddr, unsigned long pfn,
pgprot_t flags)
}
pud = pud_offset(pgd, vaddr);
pmd = pmd_offset(pud, vaddr);
- set_pmd(pmd, pfn_pmd(pfn, flags));
+
+ if (cpu_has_pse)
+ set_pmd(pmd, pfn_pmd(pfn, flags));
+ else {
+ pgprot_t new_flag = PAGE_KERNEL;
+ pte_t *pte;
+ int i;
+
+ /*
+ * This is run _after_ initial memory mapped so the
+ * PTE page are allocated - but we check it just in case.
+ */
+ if (pmd_none(*pmd)) {
+ printk(KERN_WARNING "set_pmd_pfn: pmd_none\n");
+ return;
+ }
+
+ pte = (pte_t *)pmd_page_vaddr(*pmd);
+ for (i = 0; i < PTRS_PER_PTE; i++) {
+ if (pte_none(*pte)) {
+ printk(KERN_WARNING "set_pmd_pfn: pte_none\n");
+ return;
+ }
+ set_pte(pte, pfn_pte(pfn + i, new_flag));
+ pte++;
+ }
+ }
+
/*
* It's enough to flush this one mapping.
* (PGE mappings get flushed as well)
--
1.7.7.6
_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxx
http://lists.xen.org/xen-devel
|
![]() |
Lists.xenproject.org is hosted with RackSpace, monitoring our |