[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] [BAND-AID PATCH] x86: partially undo (disable) d639e6a05a
At 12:35 +0000 on 27 Mar (1427459701), Jan Beulich wrote: > >>> On 27.03.15 at 13:28, <tim@xxxxxxx> wrote: > > At 12:19 +0000 on 27 Mar (1427458745), Jan Beulich wrote: > >> As from osstest results it looks like commit ("x86: allow 64-bit PV > >> guest kernels to suppress user mode exposure of M2P") is guilty in > >> causing migration failures, comment out the meat of it without fully > >> reverting, until it is understood what is causing the issue. > > > > I'd prefer to just revert it unless there's a benfit to keeping some > > parts still enabled. > > I'd like to avoid reverting 2e4e0d4efc as a prereq , and putting > together a revert without also reverting that other one I predicted > (maybe wrongly) to take more time than creating this one, and I'm > pretty short on time today. So yes, if you or Andrew want to put > together a full revert of just that one commit, you may consider it > pre-acked. OK, here you go. I'll leave it to you to decide when/if to apply, as I won't be around later. From 7ae18ee3c980f750dbcb91be3fe452e99c39d1a7 Mon Sep 17 00:00:00 2001 From: Tim Deegan <tim@xxxxxxx> Date: Fri, 27 Mar 2015 12:43:12 +0000 Subject: [PATCH] Revert "x86: allow 64-bit PV guest kernels to suppress user mode exposure of M2P" This reverts commit d639e6a05a0f8ee0e61c6cc4eebba78934ef3648. Signed-off-by: Tim Deegan <tim@xxxxxxx> Acked-by: Jan Beulich <JBeulich@xxxxxxxx> Conflicts: xen/arch/x86/domain.c xen/arch/x86/mm.c xen/arch/x86/mm/shadow/multi.c --- xen/arch/x86/domain.c | 9 +-------- xen/arch/x86/domain_build.c | 2 +- xen/arch/x86/mm.c | 28 ++-------------------------- xen/arch/x86/mm/shadow/multi.c | 16 ---------------- xen/arch/x86/x86_64/mm.c | 6 +++--- xen/include/asm-x86/config.h | 3 +-- xen/include/asm-x86/mm.h | 5 +---- xen/include/public/xen.h | 12 ------------ 8 files changed, 9 insertions(+), 72 deletions(-) diff --git a/xen/arch/x86/domain.c b/xen/arch/x86/domain.c index 7bae90e..393aa26 100644 --- a/xen/arch/x86/domain.c +++ b/xen/arch/x86/domain.c @@ -339,7 +339,7 @@ static int setup_compat_l4(struct vcpu *v) l4tab = __map_domain_page(pg); clear_page(l4tab); - init_guest_l4_table(l4tab, v->domain, 1); + init_guest_l4_table(l4tab, v->domain); unmap_domain_page(l4tab); v->arch.guest_table = pagetable_from_page(pg); @@ -970,11 +970,7 @@ int arch_set_info_guest( case -EINTR: rc = -ERESTART; case -ERESTART: - break; case 0: - if ( !compat && !VM_ASSIST(d, m2p_strict) && - !paging_mode_refcounts(d) ) - fill_ro_mpt(cr3_gfn); break; default: if ( cr3_page == current->arch.old_guest_table ) @@ -1009,10 +1005,7 @@ int arch_set_info_guest( default: if ( cr3_page == current->arch.old_guest_table ) cr3_page = NULL; - break; case 0: - if ( VM_ASSIST(d, m2p_strict) ) - zap_ro_mpt(cr3_gfn); break; } } diff --git a/xen/arch/x86/domain_build.c b/xen/arch/x86/domain_build.c index 287b932..e5c845c 100644 --- a/xen/arch/x86/domain_build.c +++ b/xen/arch/x86/domain_build.c @@ -1203,7 +1203,7 @@ int __init construct_dom0( l3start = __va(mpt_alloc); mpt_alloc += PAGE_SIZE; } clear_page(l4tab); - init_guest_l4_table(l4tab, d, 0); + init_guest_l4_table(l4tab, d); v->arch.guest_table = pagetable_from_paddr(__pa(l4start)); if ( is_pv_32on64_domain(d) ) v->arch.guest_table_user = v->arch.guest_table; diff --git a/xen/arch/x86/mm.c b/xen/arch/x86/mm.c index c92ac90..3e38761 100644 --- a/xen/arch/x86/mm.c +++ b/xen/arch/x86/mm.c @@ -1380,8 +1380,7 @@ static int alloc_l3_table(struct page_info *page) return rc > 0 ? 0 : rc; } -void init_guest_l4_table(l4_pgentry_t l4tab[], const struct domain *d, - bool_t zap_ro_mpt) +void init_guest_l4_table(l4_pgentry_t l4tab[], const struct domain *d) { /* Xen private mappings. */ memcpy(&l4tab[ROOT_PAGETABLE_FIRST_XEN_SLOT], @@ -1396,25 +1395,6 @@ void init_guest_l4_table(l4_pgentry_t l4tab[], const struct domain *d, l4e_from_pfn(domain_page_map_to_mfn(l4tab), __PAGE_HYPERVISOR); l4tab[l4_table_offset(PERDOMAIN_VIRT_START)] = l4e_from_page(d->arch.perdomain_l3_pg, __PAGE_HYPERVISOR); - if ( zap_ro_mpt || is_pv_32on64_domain(d) || paging_mode_refcounts(d) ) - l4tab[l4_table_offset(RO_MPT_VIRT_START)] = l4e_empty(); -} - -void fill_ro_mpt(unsigned long mfn) -{ - l4_pgentry_t *l4tab = map_domain_page(mfn); - - l4tab[l4_table_offset(RO_MPT_VIRT_START)] = - idle_pg_table[l4_table_offset(RO_MPT_VIRT_START)]; - unmap_domain_page(l4tab); -} - -void zap_ro_mpt(unsigned long mfn) -{ - l4_pgentry_t *l4tab = map_domain_page(mfn); - - l4tab[l4_table_offset(RO_MPT_VIRT_START)] = l4e_empty(); - unmap_domain_page(l4tab); } static int alloc_l4_table(struct page_info *page) @@ -1464,7 +1444,7 @@ static int alloc_l4_table(struct page_info *page) adjust_guest_l4e(pl4e[i], d); } - init_guest_l4_table(pl4e, d, !VM_ASSIST(d, m2p_strict)); + init_guest_l4_table(pl4e, d); unmap_domain_page(pl4e); return rc > 0 ? 0 : rc; @@ -2774,8 +2754,6 @@ int new_guest_cr3(unsigned long mfn) invalidate_shadow_ldt(curr, 0); - if ( !VM_ASSIST(d, m2p_strict) && !paging_mode_refcounts(d) ) - fill_ro_mpt(mfn); curr->arch.guest_table = pagetable_from_pfn(mfn); update_cr3(curr); @@ -3133,8 +3111,6 @@ long do_mmuext_op( op.arg1.mfn); break; } - if ( VM_ASSIST(d, m2p_strict) && !paging_mode_refcounts(d) ) - zap_ro_mpt(op.arg1.mfn); } curr->arch.guest_table_user = pagetable_from_pfn(op.arg1.mfn); diff --git a/xen/arch/x86/mm/shadow/multi.c b/xen/arch/x86/mm/shadow/multi.c index bf8a469..c82aa96 100644 --- a/xen/arch/x86/mm/shadow/multi.c +++ b/xen/arch/x86/mm/shadow/multi.c @@ -1435,9 +1435,6 @@ void sh_install_xen_entries_in_l4(struct domain *d, mfn_t gl4mfn, mfn_t sl4mfn) shadow_l4e_from_mfn(page_to_mfn(d->arch.perdomain_l3_pg), __PAGE_HYPERVISOR); - if ( !VM_ASSIST(d, m2p_strict) ) - sl4e[shadow_l4_table_offset(RO_MPT_VIRT_START)] = shadow_l4e_empty(); - /* Shadow linear mapping for 4-level shadows. N.B. for 3-level * shadows on 64-bit xen, this linear mapping is later replaced by the * monitor pagetable structure, which is built in make_monitor_table @@ -3978,19 +3975,6 @@ sh_update_cr3(struct vcpu *v, int do_locking) /* PAGING_LEVELS==4 implies 64-bit, which means that * map_domain_page_global can't fail */ BUG_ON(v->arch.paging.shadow.guest_vtable == NULL); - if ( !shadow_mode_external(d) && !is_pv_32on64_domain(d) ) - { - shadow_l4e_t *sl4e = v->arch.paging.shadow.guest_vtable; - - if ( (v->arch.flags & TF_kernel_mode) && - !VM_ASSIST(d, m2p_strict) ) - sl4e[shadow_l4_table_offset(RO_MPT_VIRT_START)] = - idle_pg_table[l4_table_offset(RO_MPT_VIRT_START)]; - else if ( !(v->arch.flags & TF_kernel_mode) && - VM_ASSIST(d, m2p_strict) ) - sl4e[shadow_l4_table_offset(RO_MPT_VIRT_START)] = - shadow_l4e_empty(); - } } else v->arch.paging.shadow.guest_vtable = __linear_l4_table; diff --git a/xen/arch/x86/x86_64/mm.c b/xen/arch/x86/x86_64/mm.c index 5c70061..6875c92 100644 --- a/xen/arch/x86/x86_64/mm.c +++ b/xen/arch/x86/x86_64/mm.c @@ -480,7 +480,7 @@ static int setup_m2p_table(struct mem_hotadd_info *info) l2_ro_mpt += l2_table_offset(va); } - /* NB. Cannot be GLOBAL: guest user mode should not see it. */ + /* NB. Cannot be GLOBAL as shadow_mode_translate reuses this area. */ l2e_write(l2_ro_mpt, l2e_from_pfn(mfn, /*_PAGE_GLOBAL|*/_PAGE_PSE|_PAGE_USER|_PAGE_PRESENT)); } @@ -583,7 +583,7 @@ void __init paging_init(void) 0x77, 1UL << L3_PAGETABLE_SHIFT); ASSERT(!l2_table_offset(va)); - /* NB. Cannot be GLOBAL: guest user mode should not see it. */ + /* NB. Cannot be GLOBAL as shadow_mode_translate reuses this area. */ l3e_write(&l3_ro_mpt[l3_table_offset(va)], l3e_from_page(l1_pg, /*_PAGE_GLOBAL|*/_PAGE_PSE|_PAGE_USER|_PAGE_PRESENT)); @@ -621,7 +621,7 @@ void __init paging_init(void) l3e_from_page(l2_pg, __PAGE_HYPERVISOR | _PAGE_USER)); ASSERT(!l2_table_offset(va)); } - /* NB. Cannot be GLOBAL: guest user mode should not see it. */ + /* NB. Cannot be GLOBAL as shadow_mode_translate reuses this area. */ if ( l1_pg ) l2e_write(l2_ro_mpt, l2e_from_page( l1_pg, /*_PAGE_GLOBAL|*/_PAGE_PSE|_PAGE_USER|_PAGE_PRESENT)); diff --git a/xen/include/asm-x86/config.h b/xen/include/asm-x86/config.h index 7700c77..3569753 100644 --- a/xen/include/asm-x86/config.h +++ b/xen/include/asm-x86/config.h @@ -346,8 +346,7 @@ extern unsigned long xen_phys_start; #define NATIVE_VM_ASSIST_VALID ((1UL << VMASST_TYPE_4gb_segments) | \ (1UL << VMASST_TYPE_4gb_segments_notify) | \ (1UL << VMASST_TYPE_writable_pagetables) | \ - (1UL << VMASST_TYPE_pae_extended_cr3) | \ - (1UL << VMASST_TYPE_m2p_strict)) + (1UL << VMASST_TYPE_pae_extended_cr3)) #define VM_ASSIST_VALID NATIVE_VM_ASSIST_VALID #define COMPAT_VM_ASSIST_VALID (NATIVE_VM_ASSIST_VALID & \ ((1UL << COMPAT_BITS_PER_LONG) - 1)) diff --git a/xen/include/asm-x86/mm.h b/xen/include/asm-x86/mm.h index d1f95c8..136f1c3 100644 --- a/xen/include/asm-x86/mm.h +++ b/xen/include/asm-x86/mm.h @@ -318,10 +318,7 @@ static inline void *__page_to_virt(const struct page_info *pg) int free_page_type(struct page_info *page, unsigned long type, int preemptible); -void init_guest_l4_table(l4_pgentry_t[], const struct domain *, - bool_t zap_ro_mpt); -void fill_ro_mpt(unsigned long mfn); -void zap_ro_mpt(unsigned long mfn); +void init_guest_l4_table(l4_pgentry_t[], const struct domain *); int is_iomem_page(unsigned long mfn); diff --git a/xen/include/public/xen.h b/xen/include/public/xen.h index 17ecb94..dd52a50 100644 --- a/xen/include/public/xen.h +++ b/xen/include/public/xen.h @@ -486,18 +486,6 @@ DEFINE_XEN_GUEST_HANDLE(mmuext_op_t); /* x86/PAE guests: support PDPTs above 4GB. */ #define VMASST_TYPE_pae_extended_cr3 3 -/* - * x86/64 guests: strictly hide M2P from user mode. - * This allows the guest to control respective hypervisor behavior: - * - when not set, L4 tables get created with the respective slot blank, - * and whenever the L4 table gets used as a kernel one the missing - * mapping gets inserted, - * - when set, L4 tables get created with the respective slot initialized - * as before, and whenever the L4 table gets used as a user one the - * mapping gets zapped. - */ -#define VMASST_TYPE_m2p_strict 32 - #if __XEN_INTERFACE_VERSION__ < 0x00040600 #define MAX_VMASST_TYPE 3 #endif -- 2.1.0 _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxx http://lists.xen.org/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |